Streaming ETL demo - Enriching event stream data with CDC data from MySQL, stream into Elasticsearch
This is designed to be run as a step-by-step demo. The ksqldb-statements.sql
should match those run in this doc end-to-end and in theory you can just run the file, but I have not tested it. PRs welcome for a one-click script that just demonstrates the end-to-end running demo :)
The slides that accompany this demo can be found here: https://go.rmoff.net/devoxxuk19
Start the environment
docker-compose up -d
Run Postman if you’re going to use that for the pull query demo at the end (see ksqlDB.postman_collection.json).
bash -c ' \
echo -e "\n\n=============\nWaiting for Kafka Connect to start listening on localhost ⏳\n=============\n"
while [ $(curl -s -o /dev/null -w %{http_code} http://localhost:8083/connectors) -ne 200 ] ; do
echo -e "\t" $(date) " Kafka Connect listener HTTP state: " $(curl -s -o /dev/null -w %{http_code} http://localhost:8083/connectors) " (waiting for 200)"
sleep 5
done
echo -e $(date) "\n\n--------------\n\o/ Kafka Connect is ready! Listener HTTP state: " $(curl -s -o /dev/null -w %{http_code} http://localhost:8083/connectors) "\n--------------\n"
'
Make sure that the Elasticsearch, Debezium, and DataGen connectors are available:
+
curl -s localhost:8083/connector-plugins|jq '.[].class'|egrep 'DatagenConnector|MySqlConnector|ElasticsearchSinkConnector'
+ Expect:
+
"io.confluent.connect.elasticsearch.ElasticsearchSinkConnector"
"io.confluent.kafka.connect.datagen.DatagenConnector"
"io.debezium.connector.mysql.MySqlConnector"
Optionally, use something like screen
or tmux
to have these both easily to hand. Or multiple Terminal tabs. Whatever works for you :)
-
ksqlDB CLI:
docker exec -it ksqldb bash -c 'echo -e "\n\n⏳ Waiting for ksqlDB to be available before launching CLI\n"; while : ; do curl_status=$(curl -s -o /dev/null -w %{http_code} http://ksqldb:8088/info) ; echo -e $(date) " ksqlDB server listener HTTP state: " $curl_status " (waiting for 200)" ; if [ $curl_status -eq 200 ] ; then break ; fi ; sleep 5 ; done ; ksql http://ksqldb:8088'
SHOW TOPICS;
ksql> SHOW TOPICS;
Kafka Topic | Partitions | Partition Replicas
-------------------------------------------------------------------------
confluent_rmoff_02ksql_processing_log | 1 | 1
ratings | 1 | 1
-------------------------------------------------------------------------
CREATE STREAM RATINGS WITH (KAFKA_TOPIC='ratings',VALUE_FORMAT='AVRO');
CREATE STREAM POOR_RATINGS AS
SELECT USER_ID, STARS, CHANNEL, MESSAGE FROM RATINGS WHERE STARS<3;
SELECT * FROM POOR_RATINGS EMIT CHANGES LIMIT 5;
DESCRIBE EXTENDED POOR_RATINGS;
Optionally, bring up a second ksqlDB prompt and show live ratings / live filtered ratings:
-- Live stream of ratings data
SET 'auto.offset.reset' = 'latest';
PRINT 'ratings';
-- You can use SELECT too, but PRINT makes it clearer that it's coming from a topic
SELECT TIMESTAMPTOSTRING(ROWTIME, 'yyyy-MM-dd HH:mm:ss') AS RATING_TIMESTAMP, STARS, CHANNEL FROM RATINGS EMIT CHANGES;
-- Just the ratings with star rating less than 3:
SET 'auto.offset.reset' = 'latest';
PRINT 'POOR_RATINGS';
-- You can use SELECT too, but PRINT makes it clearer that it's coming from a topic
SELECT TIMESTAMPTOSTRING(ROWTIME, 'yyyy-MM-dd HH:mm:ss') AS RATING_TIMESTAMP, STARS, CHANNEL FROM POOR_RATINGS EMIT CHANGES;
CREATE SINK CONNECTOR SINK_ES_POOR_RATINGS WITH (
'connector.class' = 'io.confluent.connect.elasticsearch.ElasticsearchSinkConnector',
'topics' = 'POOR_RATINGS',
'connection.url' = 'http://elasticsearch:9200',
'type.name' = '_doc',
'key.ignore' = 'false',
'schema.ignore' = 'true',
'transforms'= 'ExtractTimestamp',
'transforms.ExtractTimestamp.type'= 'org.apache.kafka.connect.transforms.InsertField$Value',
'transforms.ExtractTimestamp.timestamp.field' = 'RATING_TS'
);
Show in Kibana (http://localhost:5601/app/kibana)
Launch the MySQL CLI:
docker exec -it mysql bash -c 'mysql -u $MYSQL_USER -p$MYSQL_PASSWORD demo'
SHOW TABLES;
+----------------+
| Tables_in_demo |
+----------------+
| CUSTOMERS |
+----------------+
1 row in set (0.00 sec)
SELECT ID, FIRST_NAME, LAST_NAME, EMAIL, CLUB_STATUS FROM CUSTOMERS LIMIT 5;
+----+-------------+------------+------------------------+-------------+
| ID | FIRST_NAME | LAST_NAME | EMAIL | CLUB_STATUS |
+----+-------------+------------+------------------------+-------------+
| 1 | Rica | Blaisdell | rblaisdell0@rambler.ru | bronze |
| 2 | Ruthie | Brockherst | rbrockherst1@ow.ly | platinum |
| 3 | Mariejeanne | Cocci | mcocci2@techcrunch.com | bronze |
| 4 | Hashim | Rumke | hrumke3@sohu.com | platinum |
| 5 | Hansiain | Coda | hcoda4@senate.gov | platinum |
+----+-------------+------------+------------------------+-------------+
5 rows in set (0.00 sec)
In ksqlDB:
CREATE SOURCE CONNECTOR SOURCE_MYSQL_01 WITH (
'connector.class' = 'io.debezium.connector.mysql.MySqlConnector',
'database.hostname' = 'mysql',
'database.port' = '3306',
'database.user' = 'debezium',
'database.password' = 'dbz',
'database.server.id' = '42',
'database.server.name' = 'asgard',
'table.whitelist' = 'demo.customers',
'database.history.kafka.bootstrap.servers' = 'kafka:29092',
'database.history.kafka.topic' = 'dbhistory.demo' ,
'include.schema.changes' = 'false',
'transforms'= 'unwrap,extractkey',
'transforms.unwrap.type'= 'io.debezium.transforms.ExtractNewRecordState',
'transforms.extractkey.type'= 'org.apache.kafka.connect.transforms.ExtractField$Key',
'transforms.extractkey.field'= 'id',
'key.converter'= 'org.apache.kafka.connect.storage.StringConverter',
'value.converter'= 'io.confluent.connect.avro.AvroConverter',
'value.converter.schema.registry.url'= 'http://schema-registry:8081'
);
Check that it’s running:
ksql> SHOW CONNECTORS;
Connector Name | Type | Class | Status
----------------------------------------------------------------------------------------------------------------
source-datagen-01 | SOURCE | io.confluent.kafka.connect.datagen.DatagenConnector | RUNNING (1/1 tasks RUNNING)
SOURCE_MYSQL_01 | SOURCE | io.debezium.connector.mysql.MySqlConnector | RUNNING (1/1 tasks RUNNING)
----------------------------------------------------------------------------------------------------------------
SHOW TOPICS;
ksql> SHOW TOPICS;
Kafka Topic | Partitions | Partition Replicas
-------------------------------------------------------------------------
POOR_RATINGS | 1 | 1
asgard.demo.CUSTOMERS | 1 | 1
confluent_rmoff_02ksql_processing_log | 1 | 1
dbhistory.demo | 1 | 1
ratings | 1 | 1
-------------------------------------------------------------------------
Show topic contents
PRINT 'asgard.demo.CUSTOMERS' FROM BEGINNING;
Key format: KAFKA (STRING)
Value format: AVRO rowtime: 3/3/20 1:35:52 PM UTC, key: 1, value: {"id": 1, "first_name": "Rica", "last_name": "Blaisdell", "email": "[email protected]", "gender": "Female", "club_status": "bronze", "comments": "Universal optimal hierarchy", "create_ts": "2020-03-03T13:32:03Z", "update_ts": "202
0-03-03T13:32:03Z"} rowtime: 3/3/20 1:35:52 PM UTC, key: 2, value: {"id": 2, "first_name": "Ruthie", "last_name": "Brockherst", "email": "[email protected]", "gender": "Female", "club_status": "platinum", "comments": "Reverse-engineered tangible interface", "create_ts": "2020-03-03T13:32:03Z", "updat
e_ts": "2020-03-03T13:32:03Z"}
…
Create ksqlDB stream and table
CREATE TABLE CUSTOMERS WITH (KAFKA_TOPIC='asgard.demo.CUSTOMERS', VALUE_FORMAT='AVRO');
Query the ksqlDB table:
SET 'auto.offset.reset' = 'earliest';
SELECT ID, FIRST_NAME, LAST_NAME, EMAIL, CLUB_STATUS FROM CUSTOMERS EMIT CHANGES;
MySQL terminal:
INSERT INTO CUSTOMERS (ID,FIRST_NAME,LAST_NAME) VALUES (42,'Rick','Astley');
UPDATE CUSTOMERS SET EMAIL = '[email protected]' where ID=42;
UPDATE CUSTOMERS SET CLUB_STATUS = 'bronze' where ID=42;
UPDATE CUSTOMERS SET CLUB_STATUS = 'platinum' where ID=42;
Check the data in ksqlDB:
Here’s the table - the latest value for a given key
SELECT TIMESTAMPTOSTRING(ROWTIME, 'HH:mm:ss') AS EVENT_TS,
ID,
FIRST_NAME,
LAST_NAME,
EMAIL,
CLUB_STATUS
FROM CUSTOMERS WHERE ID=42
EMIT CHANGES;
+----------+----+-----------+----------+-----------------+------------+
|EVENT_TS |ID |FIRST_NAME |LAST_NAME |EMAIL |CLUB_STATUS |
+----------+----+-----------+----------+-----------------+------------+
|15:43:58 |42 |Rick |Astley |rick@example.com |platinum |
^CQuery terminated
Here’s the stream - every event, which in this context is every change event on the source database:
CREATE STREAM CUSTOMERS_STREAM WITH (KAFKA_TOPIC='asgard.demo.CUSTOMERS', VALUE_FORMAT='AVRO');
SET 'auto.offset.reset' = 'earliest';
SELECT TIMESTAMPTOSTRING(ROWTIME, 'HH:mm:ss') AS EVENT_TS,
ID,
FIRST_NAME,
LAST_NAME,
EMAIL,
CLUB_STATUS
FROM CUSTOMERS_STREAM WHERE ID=42
EMIT CHANGES;
+---------+----+-----------+----------+------------------+------------+
|TS |ID |FIRST_NAME |LAST_NAME |EMAIL |CLUB_STATUS |
+---------+----+-----------+----------+------------------+------------+
|16:08:49 |42 |Rick |Astley |null |null |
|16:09:30 |42 |Rick |Astley |rick@example.com |null |
|16:09:32 |42 |Rick |Astley |rick@example.com |bronze |
|16:09:35 |42 |Rick |Astley |rick@example.com |platinum |
^CQuery terminated
ksql>
SELECT R.RATING_ID, R.MESSAGE, R.CHANNEL,
C.ID, C.FIRST_NAME + ' ' + C.LAST_NAME AS FULL_NAME,
C.CLUB_STATUS
FROM RATINGS R
LEFT JOIN CUSTOMERS C
ON CAST(R.USER_ID AS STRING) = C.ROWKEY
WHERE C.FIRST_NAME IS NOT NULL
EMIT CHANGES;
+------------+-----------------------------------+-------+--------------------+-------------+
|RATING_ID |MESSAGE |ID |FULL_NAME |CLUB_STATUS |
+------------+-----------------------------------+-------+--------------------+-------------+
|1 |more peanuts please |9 |Even Tinham |silver |
|2 |Exceeded all my expectations. Thank|8 |Patti Rosten |silver |
| | you ! | | | |
|3 |meh |17 |Brianna Paradise |bronze |
|4 |is this as good as it gets? really |14 |Isabelita Talboy |gold |
| |? | | | |
|5 |why is it so difficult to keep the |19 |Josiah Brockett |gold |
| |bathrooms clean ? | | | |
…
Persist this stream of data
SET 'auto.offset.reset' = 'earliest';
CREATE STREAM RATINGS_WITH_CUSTOMER_DATA
WITH (KAFKA_TOPIC='ratings-enriched')
AS
SELECT R.RATING_ID, R.MESSAGE, R.STARS, R.CHANNEL,
C.ID, C.FIRST_NAME + ' ' + C.LAST_NAME AS FULL_NAME,
C.CLUB_STATUS, C.EMAIL
FROM RATINGS R
LEFT JOIN CUSTOMERS C
ON CAST(R.USER_ID AS STRING) = C.ROWKEY
WHERE C.FIRST_NAME IS NOT NULL
EMIT CHANGES;
CUSTOMERS is a ksqlDB table, which means that we have the latest value for a given key.
Check out the ratings for customer id 2 only:
SELECT TIMESTAMPTOSTRING(ROWTIME, 'HH:mm:ss') AS EVENT_TS,
FULL_NAME, CLUB_STATUS, STARS, MESSAGE, CHANNEL
FROM RATINGS_WITH_CUSTOMER_DATA
WHERE ID=2
EMIT CHANGES;
In mysql, make a change to ID 2
UPDATE CUSTOMERS SET CLUB_STATUS = 'bronze' WHERE ID=2;
Observe in the continuous ksqlDB query that the customer name has now changed.
CREATE SINK CONNECTOR SINK_ELASTIC_01 WITH (
'connector.class' = 'io.confluent.connect.elasticsearch.ElasticsearchSinkConnector',
'connection.url' = 'http://elasticsearch:9200',
'type.name' = '',
'behavior.on.malformed.documents' = 'warn',
'errors.tolerance' = 'all',
'errors.log.enable' = 'true',
'errors.log.include.messages' = 'true',
'topics' = 'ratings-enriched,UNHAPPY_PLATINUM_CUSTOMERS',
'key.ignore' = 'true',
'schema.ignore' = 'true',
'key.converter' = 'org.apache.kafka.connect.storage.StringConverter',
'transforms'= 'ExtractTimestamp',
'transforms.ExtractTimestamp.type'= 'org.apache.kafka.connect.transforms.InsertField$Value',
'transforms.ExtractTimestamp.timestamp.field' = 'EXTRACT_TS'
);
Check status
ksql> SHOW CONNECTORS;
Connector Name | Type | Class | Status
--------------------------------------------------------------------------------------------------------------------------
source-datagen-01 | SOURCE | io.confluent.kafka.connect.datagen.DatagenConnector | RUNNING (1/1 tasks RUNNING)
SOURCE_MYSQL_01 | SOURCE | io.debezium.connector.mysql.MySqlConnector | RUNNING (1/1 tasks RUNNING)
SINK_ELASTIC_00 | SINK | io.confluent.connect.elasticsearch.ElasticsearchSinkConnector | RUNNING (1/1 tasks RUNNING)
--------------------------------------------------------------------------------------------------------------------------
Check data in Elasticsearch:
docker exec elasticsearch curl -s "http://localhost:9200/_cat/indices/*?h=idx,docsCount"
unhappy_platinum_customers 1
.kibana_task_manager_1 2
.apm-agent-configuration 0
kafka-ratings-enriched-2018-08 1
.kibana_1 11
ratings-enriched 3699
Tested on Elasticsearch 7.5.0
Simple aggregation - count of ratings per person, per minute:
SELECT TIMESTAMPTOSTRING(WINDOWSTART, 'yyyy-MM-dd HH:mm:ss') AS WINDOW_START_TS,
FULL_NAME,COUNT(*) AS RATINGS_COUNT
FROM RATINGS_WITH_CUSTOMER_DATA
WINDOW TUMBLING (SIZE 1 MINUTE)
GROUP BY FULL_NAME
EMIT CHANGES;
Persist this and show the timestamp:
CREATE TABLE RATINGS_PER_CUSTOMER_PER_MINUTE AS
SELECT FULL_NAME,COUNT(*) AS RATINGS_COUNT
FROM RATINGS_WITH_CUSTOMER_DATA
WINDOW TUMBLING (SIZE 1 MINUTE)
GROUP BY FULL_NAME
EMIT CHANGES;
SELECT TIMESTAMPTOSTRING(WINDOWSTART, 'yyyy-MM-dd HH:mm:ss') AS WINDOW_START_TS,
FULL_NAME,
RATINGS_COUNT
FROM RATINGS_PER_CUSTOMER_PER_MINUTE
WHERE ROWKEY='Rica Blaisdell'
EMIT CHANGES;
SELECT TIMESTAMPTOSTRING(WINDOWSTART, 'yyyy-MM-dd HH:mm:ss') AS WINDOW_START_TS,
FULL_NAME,
RATINGS_COUNT
FROM RATINGS_PER_CUSTOMER_PER_MINUTE
WHERE ROWKEY='Rica Blaisdell'
AND WINDOWSTART > '2020-04-29T18:30:00.000';
Show REST API with Postman or bash:
docker exec -it ksqldb bash
Copy and paste:
# Store the epoch (milliseconds) five minutes ago
PREDICATE=$(date --date '-5 min' +%s)000
# Pull from ksqlDB the aggregate-by-minute for the last five minutes for a given user:
curl -X "POST" "http://ksqldb:8088/query" \
-H "Content-Type: application/vnd.ksql.v1+json; charset=utf-8" \
-d '{"ksql":"SELECT TIMESTAMPTOSTRING(WINDOWSTART, '\''yyyy-MM-dd HH:mm:ss'\'') AS WINDOW_START_TS, FULL_NAME, RATINGS_COUNT FROM RATINGS_PER_CUSTOMER_PER_MINUTE WHERE ROWKEY='\''Rica Blaisdell'\'' AND WINDOWSTART > '$PREDICATE';"}'
Press Ctrl-D to exit the Docker container