Overview
DataHub offers data distribution capabilities. You can distribute CKafka data to TDW for data storage, query, and analysis.
Prerequisites
Currently, this feature depends on the TDW service, and you need to enable the relevant product features to use it.
Directions
2. In the left sidebar, click Connector > Task List , select the right region, and then click Create Task .
3. Fill in the task name, select Data Distribution as the task type, select Data Warehouse (TDW) as the data target type, then click Next .
4. Configure data source information.
Topic Type: Select the data source Topic
Elastic Topic: Select the pre-created elastic Topic, for details see Topic Management. CKafka Instance Topic: Select the instance and Topic created in CKafka. If the instance has ACL policies, you need to ensure the selected topic has read and write permissions. For details, please seeTopic Management. Start Offset: Select how to handle historical messages during dump by setting the Topic offset.
5. After setting the above information, click Next , click Preview Topic Message , and the first message from the source topic will be obtained and parsed.
Note
Currently, message parsing should meet the following requirements:
The message is in JSON string format. JSON keys matching the TDW field names will correspond to the table structure in TDW.
The source data should be in a single-layer JSON format. For nested JSON, you can useData Processing to perform simple message format conversion. 6. (Optional) Toggle on the Data Processing button for source data, for detailed configuration, please see Simple Data Processing. 7. Click Next to configure the data target information.
Source Data: Click to pull data from the source Topic.
TDW BID: Enter the TDW BID.
TDW TID: Enter the TDW TID.
8. Click Submit . You can see the newly created task in the task list and view the task creation progress in the status bar.
Was this page helpful?