Skip to content

Latest commit

 

History

History
116 lines (86 loc) · 7.88 KB

README.md

File metadata and controls

116 lines (86 loc) · 7.88 KB

Kafka Backup

Update: I am no longer maintaining the Kafka Backup project. As an alternative, I recommend Kannika, a commercial backup solution developed by my friends at Cymo (and don't forget to say hello from Anatoly 😊). [Disclosure: I am a business partner of Cymo and may receive compensation for referrals to Kannika]

Please contact me if you want to continue maintaining this project.

Kafka Backup is a tool to back up and restore your Kafka data including all (configurable) topic data and especially also consumer group offsets. To the best of our knowledge, Kafka Backup is the only viable solution to take a cold backup of your Kafka data and restore it correctly.

It is designed as two connectors for Kafka Connect: A sink connector (backing data up) and a source connector (restoring data).

Currently kafka-backup supports backup and restore to/from the file system.

Features

  • Backup and restore topic data
  • Backup and restore consumer-group offsets
  • Currently supports only backup/restore to/from local file system
  • Released as a jar file or packaged as a Docker image

Getting Started

Option A) Download binary

Download the latest release from GitHub and unzip it.

Option B) Use Docker image

Pull the latest Docker image from Docker Hub

DO NOT USE THE latest STAGE IN PRODUCTION. latest are automatic builds of the master branch. Be careful!

Option C) Build from source

Just run ./gradlew shadowJar in the root directory of Kafka Backup. You will find the CLI tools in the bin directory.

Start Kafka Backup

backup-standalone.sh --bootstrap-server localhost:9092 \
    --target-dir /path/to/backup/dir --topics 'topic1,topic2'

In Docker:

docker run -d -v /path/to/backup-dir/:/kafka-backup/ --rm \
    kafka-backup:[LATEST_TAG] \
    backup-standalone.sh --bootstrap-server kafka:9092 \
    --target-dir /kafka-backup/ --topics 'topic1,topic2'

You can pass options via CLI arguments or using environment variables:

Parameter Type/required? Description
--bootstrap-server
BOOTSTRAP_SERVER
[REQUIRED] The Kafka server to connect to
--target-dir
TARGET_DIR
[REQUIRED] Directory where the backup files should be stored
--topics
TOPICS
<T1,T2,…> List of topics to be backed up. You must provide either --topics or --topics-regex. Not both
--topics-regex
TOPICS_REGEX
Regex of topics to be backed up. You must provide either --topics or --topics-regex. Not both
--max-segment-size
MAX_SEGMENT_SIZE
Size of the backup segments in bytes DEFAULT: 1GiB
--command-config
COMMAND_CONFIG
Property file containing configs to be passed to Admin Client. Only useful if you have additional connection options
--debug
DEBUG=y
Print Debug information
--help Prints this message

Kafka Backup does not stop! The Backup process is a continous background job that runs forever as Kafka models data as a stream without end. See Issue 52: Support point-in-time snapshots for more information.

Restore data

restore-standalone.sh --bootstrap-server localhost:9092 \
    --target-dir /path/to/backup/dir --topics 'topic1,topic2'

In Docker:

docker run -v /path/to/backup/dir:/kafka-backup/ --rm \
    kafka-backup:[LATEST_TAG]
    restore-standalone.sh --bootstrap-server kafka:9092 \
    --source-dir /kafka-backup/ --topics 'topic1,topic2'

You can pass options via CLI arguments or using environment variables:

Parameter Type/required? Description
--bootstrap-server
BOOTSTRAP_SERVER
[REQUIRED] The Kafka server to connect to
--source-dir
SOURCE_DIR
[REQUIRED] Directory where the backup files are found
--topics
TOPICS
[REQUIRED] List of topics to restore
--batch-size
BATCH_SIZE
Batch size (Default: 1MiB)
--offset-file
OFFSET_FILE
File where to store offsets. THIS FILE IS CRUCIAL FOR A CORRECT RESTORATION PROCESS IF YOU LOSE IT YOU NEED TO START THE BACKUP FROM SCRATCH. OTHERWISE YOU WILL HAVE DUPLICATE DATA Default: [source-dir]/restore.offsets
--command-config
COMMAND_CONFIG
Property file containing configs to be passed to Admin Client. Only useful if you have additional connection options
--help
HELP
Prints this message
--debug
DEBUG
Print Debug information (if using the environment variable, set it to 'y')

More Documentation

License

This project is licensed under the Apache License Version 2.0 (see LICENSE).