Skip to content
This repository has been archived by the owner on Aug 5, 2022. It is now read-only.

Commit

Permalink
Add table of contents (#59)
Browse files Browse the repository at this point in the history
* Add table of contents
  • Loading branch information
ismailsimsek authored Feb 6, 2022
1 parent 8332107 commit 89cc123
Show file tree
Hide file tree
Showing 3 changed files with 20 additions and 7 deletions.
17 changes: 15 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -2,11 +2,17 @@
![contributions welcome](https://img.shields.io/badge/contributions-welcome-brightgreen.svg?style=flat)
![Java CI with Maven](https://github.com/memiiso/debezium-server-batch/workflows/Java%20CI%20with%20Maven/badge.svg?branch=master)

# Table of contents
* [Debezium Batch Consumers](#debezium-batch-consumers)
* [`sparkbatch` Consumer](#sparkbatch-consumer)
* [`sparkbigquerybatch` Consumer](#sparkbigquerybatch-consumer)
* [`bigquerybatch` Consumer](#bigquerybatch-consumer)
* [Install from source](#install-from-source)

# Debezium Batch Consumers

This project adds batch consumers
to [Debezium Server](https://debezium.io/documentation/reference/operations/debezium-server.html). Using batch consumers
its possible to consume CDC events as mini batches
to [Debezium Server](https://debezium.io/documentation/reference/operations/debezium-server.html). Using batch consumers its possible to consume CDC events as mini batches

## `sparkbatch` Consumer
Consumes debezium events using spark
Expand All @@ -25,6 +31,13 @@ Consumes debezium events using spark
| `debezium.sink.batch.destination-regexp-replace` | `` | Regexp Replace part to modify destination |
| `debezium.sink.batch.batch-size-wait` | `NoBatchSizeWait` | Batch size wait strategy to optimize data files and upload interval. explained below. |

## `sparkbigquerybatch` Consumer
Slightly customized spark consumer that delivers the messages into Bigquery.

| Config | Default | Description |
|---------|---------------|--------------------|
| `@TODO` | `@TODO` | Destination bucket |

## `bigquerybatch` Consumer
Consumes debezium events to Bigquery using Bigquery writer api.

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@
import static org.apache.spark.sql.functions.*;

/**
* Implementation of the consumer that delivers the messages into Amazon S3 destination.
* Implementation of the consumer that delivers the messages into Bigquery destination using Spark.
*
* @author Ismail Simsek
*/
Expand Down
8 changes: 4 additions & 4 deletions pom.xml
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@
<skipITs>true</skipITs>

<version.groovy>3.0.9</version.groovy>
<version.awssdk>2.15.79</version.awssdk>
<version.awssdk>2.17.123</version.awssdk>
<version.assembly.plugin>3.1.1</version.assembly.plugin>
<version.jackson>2.12.6</version.jackson>
<version.spark>3.2.1</version.spark>
Expand Down Expand Up @@ -125,7 +125,7 @@
<dependency>
<groupId>com.google.cloud.bigdataoss</groupId>
<artifactId>gcs-connector</artifactId>
<version>hadoop3-2.2.3</version>
<version>hadoop3-2.2.4</version>
</dependency>
<!-- Google Bom -->
<dependency>
Expand All @@ -138,12 +138,12 @@
<dependency>
<groupId>com.google.cloud</groupId>
<artifactId>google-cloud-bigquery</artifactId>
<version>2.6.2</version>
<version>2.8.0</version>
</dependency>
<dependency>
<groupId>com.google.cloud</groupId>
<artifactId>google-cloud-bigquerystorage</artifactId>
<version>2.8.0</version>
<version>2.8.4</version>
</dependency>
<!-- END Google -->

Expand Down

0 comments on commit 89cc123

Please sign in to comment.