Protobuf/Avro Schema
Timeplus supports reading or writing messages in Protobuf or Avro format. This document covers how to process data without a Schema Registry. Check this page if your Kafka topics are associated with a Schema Registry.
Create A Schema
Without a Schema Registry, you need to define the Protobuf or Avro schema using SQL.
Protobuf
CREATE OR REPLACE FORMAT SCHEMA schema_name AS '
syntax = "proto3";
message SearchRequest {
string query = 1;
int32 page_number = 2;
int32 results_per_page = 3;
}
' TYPE Protobuf
Then refer to this schema while creating an external stream for Kafka:
CREATE EXTERNAL STREAM stream_name(
query string,
page_number int32,
results_per_page int32)
SETTINGS type='kafka',
brokers='pkc-1234.us-west-2.aws.confluent.cloud:9092',
topic='topic_name',
security_protocol='SASL_SSL',
username='..',
password='..',
data_format='ProtobufSingle',
format_schema='schema_name:SearchRequest'
Please note:
- If you want to ensure there is only a single Protobuf message per Kafka message, please set
data_format
toProtobufSingle
. If you set it toProtobuf
, then there could be multiple Protobuf messages in a single Kafka message. - The
format_schema
setting contains two parts: the registered schema name (in this example: schema_name), and the message type (in this example: SearchRequest). Combining them together with a semicolon. - You can use this external stream to read or write Protobuf messages in the target Kafka/Confluent topics.
- For more advanced use cases, please check the examples for complex schema.
Avro
Available since Proton 1.5.10.
CREATE OR REPLACE FORMAT SCHEMA schema_name AS '{
"namespace": "example.avro",
"type": "record",
"name": "User",
"fields": [
{"name": "name", "type": "string"},
{"name": "favorite_number", "type": ["int", "null"]},
{"name": "favorite_color", "type": ["string", "null"]}
]
}
' TYPE Avro;
Then refer to this schema while creating an external stream for Kafka:
CREATE EXTERNAL STREAM stream_name(
name string,
favorite_number nullable(int32),
favorite_color nullable(string))
SETTINGS type='kafka',
brokers='pkc-1234.us-west-2.aws.confluent.cloud:9092',
topic='topic_name',
security_protocol='SASL_SSL',
username='..',
password='..',
data_format='Avro',
format_schema='schema_name'
List Schemas
List schemas in the current Timeplus deployment:
SHOW FORMAT SCHEMAS
Show Details For A Schema
SHOW CREATE FORMAT SCHEMA schema_name
Drop A Schema
DROP FORMAT SCHEMA <IF EXISTS> schema_name;
Examples For Complex Protobuf Schema
Nested Schema
CREATE FORMAT SCHEMA simple_nested AS '
syntax = "proto3";
message Name {
string first = 1;
string last = 2;
}
message Person {
string email = 1;
Name name = 2;
int32 age = 3;
map<string, int32> skills = 4;
}
' TYPE Protobuf
CREATE EXTERNAL STREAM people(
email string,
name_first string,
name.last string,
skills map(string,int32),
age int32
)
SETTINGS type='kafka'.. data_format='ProtobufSingle',
format_schema='simple_nested:Person'
Please note:
Person
is the top level message type. It refers to theName
message type.- Use
name
as the prefix as the column names. Use either _ or . to connect the prefix with the nested field names. - When you create an external stream to read the Protobuf messages, you don't have to define all possible columns. Only the columns you defined will be read. Other columns/fields are skipped.
Enum
Say in your Protobuf definition, there is an enum type:
enum Level {
LevelOne = 0;
LevelTwo = 1;
}
You can use the enum type in Proton, e.g.
CREATE EXTERNAL STREAM ..(
..
level enum8('LevelOne'=0,'LevelTwo'=1),
..
)
Repeat
Say in your Protobuf definition, there is a repeated type:
repeated string Status
You can use the array type in Proton, e.g.
CREATE EXTERNAL STREAM ..(
..
status array(string),
..
)
Repeated and Nested
Say in your Protobuf definition, there is a field in a custom type and also repeated:
syntax = "proto3";
message DataComponent {
optional string message = 1;
message Params {
message KeyValue {
optional string name = 1;
optional string value = 2;
}
repeated KeyValue Param = 1;
}
optional Params params = 2;
}
You can use the tuple type in Proton, e.g.
CREATE EXTERNAL STREAM ..(
message string,
params tuple(Param nested( name string, value string ))
)
The streaming data will be shown as:
select * from stream_name;
message | params |
---|---|
No. 0 | ([('key_1','value_1'),('key_2','value_2'),('key_3','value_3')]) |
Package
Say in your Protobuf definition, there is a package:
package demo;
message StockRecord {
..
}
If there is only 1 package in the Protobuf definition type, you don't have to include the package name. For example:
CREATE EXTERNAL STREAM ..(
..
)
SETTINGS .. format_schema="schema_name:StockRecord"
If there are multiple packages, you can use the fully qualified name with package, e.g.
CREATE EXTERNAL STREAM ..(
..
)
SETTINGS .. format_schema="schema_name:demo.StockRecord"
Import Schemas
If you have used CREATE FORMAT SCHEMA to register a format schema, say schema_name
, you can create the other schema and import this:
CREATE FORMAT SCHEMA import_example AS '
import "schema_name.proto";
message Test{
required string ID = 1;
optional Level TheLevel = 2;
}
' TYPE Protobuf
Please make sure to add .proto
as the suffix.
Avro Data Types Mapping
Avro Primitive Types
The table below shows supported Avro primitive data types and how they match Timeplus data types in INSERT and SELECT queries.
Avro data type | Timeplus data type |
---|---|
int | int8,int16,int32,uint8,uint16,uint32 |
long | int64,uint64 |
float | float32 |
double | float64 |
bytes,string | string |
fixed(N) | fixed_string(N) |
enum | enum8,enum16 |
array(T) | array(T) |
map(k,v) | map(k,v) |
union(null,T) | nullable(T) |
null | nullable(nothing) |
int(date) | date,date32 |
long (timestamp-millis) | datetime64(3) |
long (timestamp-micros) | datetime64(6) |
string (uuid) | uuid |
record | tuple |
Avro Logical Types
Here is a list of supported logical types:
- UUID: maps to uuid.
- Date: maps to date.
- Timestamp (millisecond precision): maps to datetime64(3).
- Timestamp (microsecond precision): maps to datetime64(6).
Other logical types are not implemented yet.
record
There are two ways to map a record
. The simple one is using tuple
. Here is an example:
First given a Avro schema like this:
{
"name": "Root",
"type": "record",
"fields": [{
"name": "a_record_field",
"type": "record",
"fields": [
{"name": "one", "type": "string"},
{"name": "two", "type": "int"}
]
}]
}
The external stream uses tuple will be like this:
CREATE EXTERNAL STREAM avro (
a_record_field tuple(one string, two int32)
) SETTINGS ...;
The other way is flatting the fields, i.e. we will create a column for each field. The external stream can be defined as:
CREATE EXTERNAL STREAM avro (
`a_record_field.one` string,
`a_record_field.two` int32
) SETTINGS ...;
The column name for each field will be the name of the record field itself (in this case a_record_field) followed by a dot (.), and followed by the field name. This is how "flatten" works.
array of record
To map an array of record, you can use either array(tuple(...))
or nested()
, they are the same. By default, Timeplus will flatten the columns. For example:
Give an Avro schema:
{
"name": "Root",
"type": "record",
"fields": [{
"name": "an_array_of_records",
"type": {
"type": "array",
"items": {
"name": "record_inside_an_array",
"type": "record",
"fields": [
{"name": "one", "type": "string"},
{"name": "two", "type": "int"}
]
}
}
}]
}
We would create a stream like this:
CREATE EXTERNAL STREAM avro (
an_array_of_records array(tuple(one string, two int32))
) SETTINGS ...;
will become:
CREATE EXTERNAL STREAM avro (
`an_array_of_records.one` array(string),
`an_array_of_records.two` array(int32)
) SETTINGS ...;
The Avro output format can handle this properly.
You can use SET flatten_nested = 0
to disable the flatten behavior. The Avro output format can handle it well too.
union
Since Timeplus does not support native union, there is no "perfect" way to handle Avro unions. One stream can only handle one of the union elements (except for null
, more details later). If you need to generate values for different element types, you will need to create multiple streams.
For example. Given an Avro schema:
{
"name": "Root",
"type": "record",
"fields": [{
"name": "int_or_string",
"type": ["int", "string"]
}]
}
When we create the stream, we can only map the int_or_string
field to either int or string, for example:
CREATE EXTERNAL STREAM avro (
int_or_string int32
) SETTINGS ...;
This stream can only write int
values. If you want to write string values, you will need to create another stream like this:
CREATE EXTERNAL STREAM avro (
int_or_string string
) SETTINGS ...;
We can also use the flatten naming convention to map the union field. For this example, the streams will be:
-- using the `int` element
CREATE EXTERNAL STREAM avro (
`int_or_string.int` int32
) SETTINGS ...;
-- using the `string` element
CREATE EXTERNAL STREAM avro (
`int_or_string.string` string
) SETTINGS ...;
For named types (record, fixed, and enum), use the name property instead of the type name. For example, given an Avro schema:
{
"name": "Root",
"type": "record",
"fields": [{
"name": "int_or_record",
"type": ["int", {
"name": "details",
"type": "record",
"fields": [...]
}]
}]
}
In order to map to the record element of the union, we need to use the name details, so the stream definition will be:
CREATE EXTERNAL STREAM avro (
`int_or_record.details` tuple(...) -- do not use `int_or_record.record`, this won't work
) SETTINGS ...;
The Avro input format only supports the flatten naming convention, so, if you create the stream using the first method:
CREATE EXTERNAL STREAM avro (
int_or_string int32
) SETTINGS ...;
then, SELECT * FROM avro
won't work.
nullable
There is a special case for union, which is, when the union has two elements, and one of it is null
, then this union field will be mapped to a nullable column. Example:
Avro schema:
{
"name": "Root",
"type": "record",
"fields": [{
"name": "null_or_int",
"type": ["null", "int"]
}]
}
Stream:
CREATE EXTERNAL STREAM avro (
null_or_int nullable(int32)
) SETTINGS ...;
However, in Timeplus, nullable
cannot be applied on all the types. For instance, nullable(tuple(...))
is invalid. If a field in the Avro schema is "type": ["null", {"type": "record"}]
, you can only map it to a tuple
, and it can't be null
.