Overview | Quick Start | Concepts | Syntax Reference | Demo | Examples | FAQ | Roadmap |
---|
Table of Contents
This documents lists example queries to illustrate the look and feel of the KSQL syntax.
The following examples use a pageviews
stream and a users
table similar to the
quick start, but slightly adapted to showcase more functionality.
Note: The following statements require that the corresponding Kafka topic already exists in your Kafka cluster.
You can start with creating a stream:
CREATE STREAM pageviews \
(viewtime BIGINT, \
userid VARCHAR, \
pageid VARCHAR) \
WITH (kafka_topic='pageviews-topic', \
value_format='DELIMITED');
The above statement creates a stream with three columns on the Kafka topic that is named pageviews
. You should also
tell KSQL the format of the values that are stored in the topic. In this example, the values format is DELIMITED
.
Associating Kafka message keys:
The above statement does not make any assumptions about the Kafka message key in the underlying Kafka topic. However,
if the value of the message key in Kafka is the same as one of the columns defined in the stream in KSQL, you can
provide such information in the WITH clause. For instance, if the Kafka message key has the same value as the pageid
column, you can write the CREATE STREAM statement as follows:
CREATE STREAM pageviews \
(viewtime BIGINT, \
userid VARCHAR, \
pageid VARCHAR) \
WITH (kafka_topic='pageviews-topic', \
value_format='DELIMITED', \
key='pageid');
Associating Kafka message timestamps:
If you want to use the value of one of the columns as the Kafka message timestamp, you can provide
such information to KSQL in the WITH clause. The message timestamp is used in window-based
operations in KSQL (such as windowed aggregations) and to support event-time based processing in KSQL.
For instance, if you want to use the value of the viewtime
column as the message timestamp, you can rewrite the above
statement as follows:
CREATE STREAM pageviews \
(viewtime BIGINT, \
userid VARCHAR, \
pageid VARCHAR) \
WITH (kafka_topic='pageviews-topic', \
value_format='DELIMITED', \
key='pageid', \
timestamp='viewtime');
Note: The following statements require that the corresponding Kafka topic already exists in your Kafka cluster.
Here's how you can create a table with several columns:
CREATE TABLE users \
(registertime BIGINT, \
gender VARCHAR, \
regionid VARCHAR, \
userid VARCHAR, \
interests array<VARCHAR>, \
contact_info map<VARCHAR, VARCHAR>) \
WITH (kafka_topic='users-topic', \
value_format='JSON');
As you can see the above table has, next to columns with primitive data types, a column of array
type and another
column of map
type. We will look at working with arrays and maps later on.
Now that you have the pageviews
stream and users
table, take a look at some example queries that you can write
in KSQL. The focus is on two types of KSQL statements: CREATE STREAM AS SELECT and CREATE TABLE AS SELECT. For these
statements KSQL persists the results of the query in a new stream or table, which is backed by a Kafka topic.
For this example, imagine you want to create a new stream by transforming pageviews
in the following way:
- The
viewtime
column value is used as the Kafka message timestamp in the new stream's underlying Kafka topic. - The new stream's Kafka topic has 5 partitions.
- The data in the new stream is in JSON format.
- Add a new column that shows the message timestamp in human-readable string format.
- The
userid
column is the key for the new stream.
The following statement will generate a new stream, pageviews_transformed
with the above properties:
CREATE STREAM pageviews_transformed \
WITH (timestamp='viewtime', \
partitions=5, \
value_format='JSON') AS \
SELECT viewtime, \
userid, \
pageid, \
TIMESTAMPTOSTRING(viewtime, 'yyyy-MM-dd HH:mm:ss.SSS') AS timestring \
FROM pageviews \
PARTITION BY userid;
The following query creates a new stream by joining the pageviews_transformed
stream with
the users
table:
CREATE STREAM pageviews_enriched AS \
SELECT pv.viewtime, \
pv.userid AS userid, \
pv.pageid, \
pv.timestring, \
u.gender, \
u.regionid, \
u.interests, \
u.contact_info \
FROM pageviews_transformed pv \
LEFT JOIN users u ON pv.userid = users.userid;
Note that by default all the Kafka topics will be read from the current offset (aka the latest available data); however, in a stream-table join, the table topic will be read from the beginning.
Now assume that you want to count the number of pageviews per region. Here is the query that would perform this count:
CREATE TABLE pageviews_per_region AS \
SELECT regionid, \
count(*) \
FROM pageviews_enriched \
GROUP BY regionid;
The above query counts the pageviews from the time you start the query until you terminate the query. Note that we used CREATE TABLE AS SELECT statement here since the result of the query is a KSQL table. The results of aggregate queries in KSQL are always a table because it computes the aggregate for each key (and possibly for each window per key) and updates these results as it processes new input data.
KSQL supports aggregation over WINDOW too. Let's rewrite the above query so that we compute the pageview count per region every 1 minute:
CREATE TABLE pageviews_per_region_per_minute AS \
SELECT regionid, \
count(*) \
FROM pageviews_enriched \
WINDOW TUMBLING (SIZE 1 MINUTE) \
GROUP BY regionid;
If you want to count the pageviews for only "Region_6" by female users for every 30 seconds, you can change the above query as the following:
CREATE TABLE pageviews_per_region_per_30secs AS \
SELECT regionid, \
count(*) \
FROM pageviews_enriched \
WINDOW TUMBLING (SIZE 30 SECONDS) \
WHERE UCASE(gender)='FEMALE' AND LCASE(regionid)='region_6' \
GROUP BY regionid;
UCASE and LCASE functions in KSQL are used to convert the values of gender and regionid columns to upper and lower case, so that you can match them correctly. KSQL also supports LIKE operator for prefix, suffix and substring matching.
KSQL supports HOPPING windows and SESSION windows too. The following query is the same query as above that computes the count for hopping window of 30 seconds that advances by 10 seconds:
CREATE TABLE pageviews_per_region_per_30secs10secs AS \
SELECT regionid, \
count(*) \
FROM pageviews_enriched \
WINDOW HOPPING (SIZE 30 SECONDS, ADVANCE BY 10 SECONDS) \
WHERE UCASE(gender)='FEMALE' AND LCASE (regionid) LIKE '%_6' \
GROUP BY regionid;
The next statement counts the number of pageviews per region for session windows with a session inactivity gap of 60 seconds. In other words, you are sessionizing the input data and then perform the counting/aggregation step per region.
CREATE TABLE pageviews_per_region_per_session AS \
SELECT regionid, \
count(*) \
FROM pageviews_enriched \
WINDOW SESSION (60 SECONDS) \
GROUP BY regionid;
The interests
column in the users
table is an array
of strings that represents the interest of each user. The contact_info
column is a string-to-string map
that represents the following
contact information for each user: phone, city, state, and zipcode.
The following query will create a new stream from pageviews_enriched
that includes the first interest of each
user along with the city and zipcode for each user:
CREATE STREAM pageviews_interest_contact AS \
SELECT interests[0] AS first_interest, \
contact_info['zipcode'] AS zipcode, \
contact_info['city'] AS city, \
viewtime, \
userid, \
pageid, \
timestring, \
gender, \
regionid \
FROM pageviews_enriched;
You can set config properties for KSQL and your queries with the SET statement. This includes configuring settings relating to Kafka's Streams API as well as settings for Kafka's producer and consumer clients.
SET '<property-name>'='<property-value>';
Both property name and property value should be enclosed in single quotes. A property that is set using the SET statement will remain in effect for the remainder of the KSQL CLI session until you issue another SET statement to change it.
Here are some of the common config properties that you might want to change from their default values:
-
auto.offset.reset
: The default value in KSQL islatest
meaning all the Kafka topics will be read from the current offset (aka latest available data). You can change it using the following statement:SET 'auto.offset.reset'='earliest';
-
commit.interval.ms
: The default value is2000
. Here is an example to change the value to5000
:SET 'commit.interval.ms'='5000';
-
cache.max.bytes.buffering
: The default value is10485760
(10 MB);