Skip to content

Commit 130e756

Browse files
committed
adding pages with rendered mermaid diagrams
1 parent 1ca74d8 commit 130e756

File tree

78 files changed

+3312
-495
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

78 files changed

+3312
-495
lines changed

docs/quix-connectors/quix-streams/sources/coming-soon/AmazonGlue-source.md

Lines changed: 2 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -15,17 +15,7 @@ Use the Quix-made AWS Glue Kafka source connector to consume data from AWS Glue
1515

1616
After data is ingested from AWS Glue, process and transform it on the fly with Quix Streams, an open-source, Kafka-based Python library. Quix Streams offers an intuitive Streaming DataFrame API (similar to pandas DataFrame) for real-time data processing. It supports aggregations, windowing, filtering, group-by operations, branching, merging, serialization, and more, allowing you to shape your data to fit your needs.
1717

18-
```mermaid
19-
graph LR
20-
source["AWS Glue<br>(source)"] -->|source connector| raw
21-
subgraph Quix Platform
22-
raw["Kafka topic<br>(source data)"]
23-
process["Quix Streams<br>(stream processing)"]
24-
processed["Kafka topic<br>(processed data)"]
25-
end
26-
raw --> process
27-
process --> processed
28-
```
18+
![Diagram](images/AmazonGlue-source_diagram_1.png)
2919

3020
## Quix Kafka connectors — a simpler, better alternative to Kafka Connect
3121

@@ -50,7 +40,7 @@ By using Quix as your central data hub, you can:
5040
* Benefit from managed data integration infrastructure, thus reducing complexity and operational burden
5141
* Use a flexible, comprehensive toolkit to build data integration pipelines, including CI/CD and IaC support, environment management features, observability and monitoring capabilities, an online code editor, Python code templates, a CLI tool, and 130+ Kafka source and sink connectors
5242

53-
[Explore the Quix platform](https://portal.demo.quix.io/pipeline?workspace=demo-gametelemetrytemplate-prod) [Book a demo](https://share.hsforms.com/1iW0TmZzKQMChk0lxd_tGiw4yjw2)
43+
[Explore the Quix platform](https://portal.demo.quix.io/pipeline?workspace=demo-gametelemetrytemplate-prod)  |  [Book a demo](https://share.hsforms.com/1iW0TmZzKQMChk0lxd_tGiw4yjw2)
5444

5545
## FAQs
5646

Lines changed: 46 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,46 @@
1+
<!--[tech-name]-->
2+
# Apache Pulsar
3+
4+
<!--[blurb-about-tech]-->
5+
Apache Pulsar is a cloud-native, distributed messaging and streaming platform originally created at Yahoo, which supports multi-tenancy, seamless scalability, and low latency message delivery.
6+
7+
Quix enables you to sync to Apache Kafka <span id="to_or_from">from</span> <span id="techname">Apache Pulsar</span>, in seconds.
8+
9+
## Speak to us
10+
11+
Get a personal guided tour of the Quix Platform, SDK and API's to help you get started with assessing and using Quix, without wasting your time and without pressuring you to signup or purchase. Guaranteed!
12+
13+
[Book here!](https://quix.io/book-a-demo)
14+
15+
## Explore
16+
17+
If you prefer to explore the platform in your own time then have a look at our readonly environment
18+
19+
👉[https://portal.demo.quix.io/pipeline?workspace=demo-gametelemetrytemplate-prod](https://portal.demo.quix.io/pipeline?workspace=demo-gametelemetrytemplate-prod&token=pat-0e3c85cd4fc5436998718c120dbd6df5&_ga=2.25371390.276140621.1730716142-1628354139.1730474801)
20+
21+
## FAQ
22+
23+
### How can I use this connector?
24+
25+
Contact us to find out how to access this connector.
26+
27+
[Book here!](https://quix.io/book-a-demo)
28+
29+
### Real-time data
30+
31+
Now that data volumes are increasing exponentially, the ability to process data in real-time is crucial for industries such as finance, healthcare, and e-commerce, where timely information can significantly impact outcomes. By utilizing advanced stream processing frameworks and in-memory computing solutions, organizations can achieve seamless data integration and analysis, enhancing their operational efficiency and customer satisfaction.
32+
33+
## What is <span id="techname">Apache Pulsar</span>?
34+
35+
<!--[tech-seo-text]-->
36+
Apache Pulsar is an open-source, distributed messaging and streaming platform that supports a wide variety of use cases, including real-time messaging and data streaming. It is designed to provide a unified messaging model and low-latency message delivery in a multi-tenant environment.
37+
38+
## What data is <span id="techname">Apache Pulsar</span> good for?
39+
40+
<!--[tech-data-seo-text]-->
41+
Apache Pulsar is ideal for real-time analytics, stream processing, and microservice-based architecture communications due to its ability to handle high throughput and low latency message delivery. It excels in scenarios requiring multi-tenancy and flexible topic management at scale.
42+
43+
## What challenges do organizations have with <span id="techname">Apache Pulsar</span> and real-time data?
44+
45+
<!--[tech-challenges-seo-text]-->
46+
Organizations often face challenges with Apache Pulsar and real-time data due to the complexity of configuring the platform for optimal performance in large-scale environments. Ensuring consistent low-latency across geographically distributed systems can also present significant challenges, alongside managing schema evolution and data consistency for real-time analytics.

docs/quix-connectors/quix-streams/sources/coming-soon/AmazonS3-source.md

Lines changed: 2 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -15,17 +15,7 @@ Use the Quix-made Amazon S3 Kafka source connector to publish messages from Amaz
1515

1616
After data is ingested from Amazon S3, process and transform it on the fly with Quix Streams, an open-source, Kafka-based Python library. Quix Streams offers an intuitive Streaming DataFrame API (similar to pandas DataFrame) for real-time data processing. It supports aggregations, windowing, filtering, group-by operations, branching, merging, serialization, value processing, timestamp management, and more, allowing you to shape your data to fit your needs.
1717

18-
```mermaid
19-
graph LR
20-
source["Amazon S3<br>(source)"] -->|source connector| raw
21-
subgraph Quix Platform
22-
raw["Kafka topic<br>(source data)"]
23-
process["Quix Streams<br>(stream processing)"]
24-
processed["Kafka topic<br>(processed data)"]
25-
end
26-
raw --> process
27-
process --> processed
28-
```
18+
![Diagram](images/AmazonS3-source_diagram_1.png)
2919

3020
## Quix Kafka connectors — a simpler, better alternative to Kafka Connect
3121

@@ -50,7 +40,7 @@ By using Quix as your central data hub, you can:
5040
* Benefit from managed data integration infrastructure, thus reducing complexity and operational burden
5141
* Use a flexible, comprehensive toolkit to build data integration pipelines, including CI/CD and IaC support, environment management features, observability and monitoring capabilities, an online code editor, Python code templates, a CLI tool, and 130+ Kafka source and sink connectors
5242

53-
[Explore the Quix platform](https://portal.demo.quix.io/pipeline?workspace=demo-gametelemetrytemplate-prod) [Book a demo](https://share.hsforms.com/1iW0TmZzKQMChk0lxd_tGiw4yjw2)
43+
[Explore the Quix platform](https://portal.demo.quix.io/pipeline?workspace=demo-gametelemetrytemplate-prod)  |  [Book a demo](https://share.hsforms.com/1iW0TmZzKQMChk0lxd_tGiw4yjw2)
5444

5545
## FAQs
5646

Lines changed: 86 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -1,46 +1,106 @@
1-
<!--[tech-name]-->
2-
# Apache Pulsar
1+
<!--- BEGIN MARKDOWN --->
2+
# Integrate Pulsar with Kafka using the source Pulsar Kafka connector
33

4-
<!--[blurb-about-tech]-->
5-
Apache Pulsar is a cloud-native, distributed messaging and streaming platform originally created at Yahoo, which supports multi-tenancy, seamless scalability, and low latency message delivery.
4+
Quix enables you to publish data from Apache Pulsar to Apache Kafka and then process it. All of this in real time, using pure Python, and at any scale.
65

7-
Quix enables you to sync to Apache Kafka <span id="to_or_from">from</span> <span id="techname">Apache Pulsar</span>, in seconds.
6+
[Book a demo](https://share.hsforms.com/1iW0TmZzKQMChk0lxd_tGiw4yjw2)
87

9-
## Speak to us
8+
## Move Pulsar data to Kafka and process it in two simple steps
109

11-
Get a personal guided tour of the Quix Platform, SDK and API's to help you get started with assessing and using Quix, without wasting your time and without pressuring you to signup or purchase. Guaranteed!
10+
1. ### Ingest data from Pulsar into Kafka
1211

13-
[Book here!](https://quix.io/book-a-demo)
12+
Use the Quix-made Pulsar Kafka source connector to publish data from Pulsar into Quix-managed Apache Kafka topics. The connector enables you to stream data in a scalable, fault-tolerant manner, with consistently low latencies and strong ordering guarantees.
1413

15-
## Explore
14+
2. ### Process and transform data with Python
1615

17-
If you prefer to explore the platform in your own time then have a look at our readonly environment
16+
After data is ingested from Pulsar, process and transform it on the fly with Quix Streams, an open-source, Kafka-based Python library. Quix Streams offers an intuitive Streaming DataFrame API (similar to pandas DataFrame) for real-time data processing and message transformation. It supports aggregations, windowing, filtering, group-by operations, branching, merging, serialization, and more, allowing you to shape your data to fit your needs.
1817

19-
👉[https://portal.demo.quix.io/pipeline?workspace=demo-gametelemetrytemplate-prod](https://portal.demo.quix.io/pipeline?workspace=demo-gametelemetrytemplate-prod&token=pat-0e3c85cd4fc5436998718c120dbd6df5&_ga=2.25371390.276140621.1730716142-1628354139.1730474801)
18+
![Diagram](images/ApachePulsar-source_diagram_1.png)
2019

21-
## FAQ
20+
## Quix Kafka connectors — a simpler, better alternative to Kafka Connect
2221

23-
### How can I use this connector?
22+
Quix offers a Python-native, developer-friendly approach to data integration that eliminates the complexity associated with Kafka Connect deployment, configuration, and management.
2423

25-
Contact us to find out how to access this connector.
24+
With Quix Kafka connectors, there's no need to wrestle with complex connector configurations, worker scaling, or infrastructure management that typically come with traditional messaging systems like Kafka Connect.
2625

27-
[Book here!](https://quix.io/book-a-demo)
26+
Quix fully manages the entire Kafka connectors lifecycle, from deployment to monitoring. This means faster development, easier debugging, and lower operational overhead compared to traditional Kafka Connect implementations.
2827

29-
### Real-time data
28+
## Quix, your solution to simplify real-time data integration
3029

31-
Now that data volumes are increasing exponentially, the ability to process data in real-time is crucial for industries such as finance, healthcare, and e-commerce, where timely information can significantly impact outcomes. By utilizing advanced stream processing frameworks and in-memory computing solutions, organizations can achieve seamless data integration and analysis, enhancing their operational efficiency and customer satisfaction.
30+
As a Kafka-based platform, Quix streamlines real-time data integration across your entire tech stack, empowering you to effortlessly collect data from disparate sources into Kafka, transform and process it with Python, and send it to your chosen destination(s).
3231

33-
## What is <span id="techname">Apache Pulsar</span>?
32+
By using Quix as your central data hub, you can:
3433

35-
<!--[tech-seo-text]-->
36-
Apache Pulsar is an open-source, distributed messaging and streaming platform that supports a wide variety of use cases, including real-time messaging and data streaming. It is designed to provide a unified messaging model and low-latency message delivery in a multi-tenant environment.
34+
* Accelerate time to insights from your data to drive informed business decisions
35+
* Ensure data accuracy, quality, and consistency across your organization
36+
* Automate data integration pipelines and eliminate manual tasks
37+
* Ensure guaranteed message delivery and manage and protect sensitive data with robust security measures
38+
* Handle data with scalability, fault tolerance, sub-second latencies, and exactly-once processing guarantees
39+
* Reduce your data integration TCO to a fraction of the typical cost
40+
* Benefit from managed data integration infrastructure, thus reducing complexity and operational burden
41+
* Use a flexible, comprehensive toolkit to build data integration pipelines, including CI/CD and IaC support, environment management features, observability and monitoring capabilities, an online code editor, Python code templates, a CLI tool, and 130+ Kafka source and sink connectors
3742

38-
## What data is <span id="techname">Apache Pulsar</span> good for?
43+
[Explore the Quix platform](https://portal.demo.quix.io/pipeline?workspace=demo-gametelemetrytemplate-prod)  |  [Book a demo](https://share.hsforms.com/1iW0TmZzKQMChk0lxd_tGiw4yjw2)
3944

40-
<!--[tech-data-seo-text]-->
41-
Apache Pulsar is ideal for real-time analytics, stream processing, and microservice-based architecture communications due to its ability to handle high throughput and low latency message delivery. It excels in scenarios requiring multi-tenancy and flexible topic management at scale.
45+
## FAQs
4246

43-
## What challenges do organizations have with <span id="techname">Apache Pulsar</span> and real-time data?
47+
### What is Pulsar?
4448

45-
<!--[tech-challenges-seo-text]-->
46-
Organizations often face challenges with Apache Pulsar and real-time data due to the complexity of configuring the platform for optimal performance in large-scale environments. Ensuring consistent low-latency across geographically distributed systems can also present significant challenges, alongside managing schema evolution and data consistency for real-time analytics.
49+
Apache Pulsar is a distributed messaging and streaming data platform that offers multi-tenancy, high throughput, low latency, and global data replication. It is designed for handling high throughput workloads and provides features like multiple consumers in a single subscription, message durability, and advanced processing capabilities. Pulsar is suitable for stream processing, pub-sub messaging, and real-time analytics.
50+
51+
### What is Apache Kafka?
52+
53+
Apache Kafka is a scalable, reliable, and fault-tolerant event streaming platform that enables global data replication, real-time integration, and data exchange between different systems. Kafka and Pulsar both utilize a publish-subscribe model that ensures any source system can write data to a central pipeline, while destination systems can read that streaming data instantly as it arrives. In essence, Kafka acts as a central nervous system for data. It helps organizations unify their data architecture and provide a continuous, real-time flow of information across disparate components.
54+
55+
### What are Kafka connectors?
56+
57+
Kafka connectors are pre-built components that help integrate Apache Kafka with external systems. They allow you to reliably transfer data in and out of a Kafka cluster without writing custom integration code. There are two main types of Kafka connectors:
58+
59+
* Source connectors. These are used to pull data from source systems into Kafka topics.
60+
61+
* Sink connectors. These are used to push data from Kafka topics to destination systems.
62+
63+
### What is real-time data, and why is it important?
64+
65+
Real-time data is information that’s made available for use as soon as it's generated. It’s passed from source to destination systems with minimal latency, enabling rapid decision-making, immediate insights, and instant actions. Real-time data is crucial for industries like finance, logistics, manufacturing, healthcare, game development, information technology, and e-commerce. It empowers businesses to improve operational efficiency, increase revenue, enhance customer satisfaction, quickly respond to changing conditions, and gain a competitive advantage.
66+
67+
### What data can you publish from Pulsar to Kafka in real time?
68+
69+
* Event logs, e.g., application logs, audit logs, and user activity logs with timestamps and metadata
70+
* Metrics and performance data, including CPU usage, memory usage, and request latency
71+
* Sensor and IoT data like temperature, humidity, and pressure readings
72+
* Financial transactions showing real-time updates and approvals with associated timestamps
73+
* User interaction events such as button clicks, form submissions, and page views
74+
* Streaming media data including video playback progress, resolution changes, and buffer status
75+
* System notifications, such as alerts, warnings, and status changes
76+
77+
### What are key factors to consider when publishing Pulsar data to Kafka in real time?
78+
79+
* Multi-cloud deployments may affect global data replication and require careful configuration for robust geo-replication capabilities.
80+
* Message ordering and deduplication can be challenging to maintain during schema evolution, particularly with large messages and complex retention policies.
81+
* Defining optimal Pulsar topic partitioning strategies is crucial to avoid uneven data distribution and scaling issues.
82+
* Careful management of producer and consumer configurations is essential to handle high throughput workloads without impacting performance or causing resource contention.
83+
* Implementing security features such as encryption, authentication, and authorization is vital to protecting data streams when transferring between Pulsar and Kafka environments.
84+
* Network reliability is essential to ensure guaranteed message delivery and avoid data loss during transmission.
85+
* Aligning data schemas between Pulsar and Kafka can be complex, requiring synchronization tools and processes to ensure consistency and integrity of streaming data.
86+
87+
### How does the Pulsar Kafka source connector offered by Quix work?
88+
89+
The source Pulsar Kafka connector provided by Quix is fully managed and written in Python.
90+
91+
The connector continuously retrieves data from Pulsar and publishes it to designated Quix-managed Kafka topics.
92+
93+
The connector provides strong data delivery guarantees (ordering and exactly-once semantics) to ensure data is reliably ingested into Kafka. You can customize its write performance and choose between several serialization formats (such as JSON, Avro, and Protobuf).
94+
95+
To find out more about the source Pulsar Kafka connector offered by Quix, [book a demo](https://share.hsforms.com/1iW0TmZzKQMChk0lxd_tGiw4yjw2).
96+
97+
### Does Quix offer a sink Pulsar Kafka connector too?
98+
99+
Yes, Quix also provides a sink Pulsar connector for Kafka.
100+
101+
Learn more about it.
102+
103+
In fact, Quix offers 130+ Kafka sink and source connectors, enabling you to move data from a variety of sources into Kafka, process it, and then send messages to your desired destination(s). All in real time.
104+
105+
[Explore the library of Quix Kafka connectors](https://quix.io/connectors)
106+
<!--- END MARKDOWN --->

docs/quix-connectors/quix-streams/sources/coming-soon/Cassandra-source.md

Lines changed: 2 additions & 12 deletions
Original file line numberDiff line numberDiff line change
@@ -15,17 +15,7 @@ Use the Quix-made Cassandra Kafka source connector to publish data from Apache C
1515

1616
After data is ingested from Cassandra, process and transform it on the fly with Quix Streams, an open-source, Kafka-based Python library that performs computations. Quix Streams offers an intuitive Streaming DataFrame API (similar to pandas DataFrame) for real-time data processing. It supports aggregations, windowing, filtering, group-by operations, branching, merging, serialization, and more, allowing you to shape your data to fit your needs.
1717

18-
```mermaid
19-
graph LR
20-
source["Cassandra<br>(source)"] -->|source connector| raw
21-
subgraph Quix Platform
22-
raw["Kafka topic<br>(source data)"]
23-
process["Quix Streams<br>(stream processing)"]
24-
processed["Kafka topic<br>(processed data)"]
25-
end
26-
raw --> process
27-
process --> processed
28-
```
18+
![Diagram](images/Cassandra-source_diagram_1.png)
2919

3020
## Quix Kafka connectors — a simpler, better alternative to Kafka Connect
3121

@@ -50,7 +40,7 @@ By using Quix as your central data hub, you can:
5040
* Benefit from managed data integration infrastructure, thus reducing complexity and operational burden
5141
* Use a flexible, comprehensive toolkit to build data integration pipelines, including CI/CD and IaC support, environment management features, observability and monitoring capabilities, an online code editor, Python code templates, a CLI tool, and 130+ Kafka source and sink connectors
5242

53-
[Explore the Quix platform](https://portal.demo.quix.io/pipeline?workspace=demo-gametelemetrytemplate-prod) [Book a demo](https://share.hsforms.com/1iW0TmZzKQMChk0lxd_tGiw4yjw2)
43+
[Explore the Quix platform](https://portal.demo.quix.io/pipeline?workspace=demo-gametelemetrytemplate-prod)  |  [Book a demo](https://share.hsforms.com/1iW0TmZzKQMChk0lxd_tGiw4yjw2)
5444

5545
## FAQs
5646

0 commit comments

Comments
 (0)