Hi all, I am currently not able to maintain this Project on my own. If you are interested in supporting me, please let me know for example by opening an issue.
Kafka Backup is a tool to back up and restore your Kafka data including all (configurable) topic data and especially also consumer group offsets. To the best of our knowledge, Kafka Backup is the only viable solution to take a cold backup of your Kafka data and restore it correctly.
It is designed as two connectors for Kafka Connect: A sink connector (backing data up) and a source connector (restoring data).
Currently kafka-backup
supports backup and restore to/from the file
system.
- Backup and restore topic data
- Backup and restore consumer-group offsets
- Currently supports only backup/restore to/from local file system
- Released as a jar file or packaged as a Docker image
Option A) Download binary
Download the latest release from GitHub and unzip it.
Option B) Use Docker image
Pull the latest Docker image from Docker Hub
DO NOT USE THE latest
STAGE IN PRODUCTION. latest
are automatic builds of the master branch. Be careful!
Option C) Build from source
Just run ./gradlew shadowJar
in the root directory of Kafka Backup. You will find the CLI tools in the bin
directory.
backup-standalone.sh --bootstrap-server localhost:9092 \
--target-dir /path/to/backup/dir --topics 'topic1,topic2'
In Docker:
docker run -d -v /path/to/backup-dir/:/kafka-backup/ --rm \
kafka-backup:[LATEST_TAG] \
backup-standalone.sh --bootstrap-server kafka:9092 \
--target-dir /kafka-backup/ --topics 'topic1,topic2'
You can pass options via CLI arguments or using environment variables:
Parameter | Type/required? | Description |
---|---|---|
--bootstrap-server BOOTSTRAP_SERVER |
[REQUIRED] | The Kafka server to connect to |
--target-dir TARGET_DIR |
[REQUIRED] | Directory where the backup files should be stored |
--topics TOPICS |
<T1,T2,…> | List of topics to be backed up. You must provide either --topics or --topics-regex . Not both |
--topics-regex TOPICS_REGEX |
Regex of topics to be backed up. You must provide either --topics or --topics-regex . Not both |
|
--max-segment-size MAX_SEGMENT_SIZE |
Size of the backup segments in bytes DEFAULT: 1GiB | |
--command-config COMMAND_CONFIG |
Property file containing configs to be passed to Admin Client. Only useful if you have additional connection options | |
--debug DEBUG=y |
Print Debug information | |
--help |
Prints this message |
Kafka Backup does not stop! The Backup process is a continous background job that runs forever as Kafka models data as a stream without end. See Issue 52: Support point-in-time snapshots for more information.
restore-standalone.sh --bootstrap-server localhost:9092 \
--target-dir /path/to/backup/dir --topics 'topic1,topic2'
In Docker:
docker run -v /path/to/backup/dir:/kafka-backup/ --rm \
kafka-backup:[LATEST_TAG]
restore-standalone.sh --bootstrap-server kafka:9092 \
--source-dir /kafka-backup/ --topics 'topic1,topic2'
You can pass options via CLI arguments or using environment variables:
Parameter | Type/required? | Description |
---|---|---|
--bootstrap-server BOOTSTRAP_SERVER |
[REQUIRED] | The Kafka server to connect to |
--source-dir SOURCE_DIR |
[REQUIRED] | Directory where the backup files are found |
--topics TOPICS |
[REQUIRED] | List of topics to restore |
--batch-size BATCH_SIZE |
Batch size (Default: 1MiB) | |
--offset-file OFFSET_FILE |
File where to store offsets. THIS FILE IS CRUCIAL FOR A CORRECT RESTORATION PROCESS IF YOU LOSE IT YOU NEED TO START THE BACKUP FROM SCRATCH. OTHERWISE YOU WILL HAVE DUPLICATE DATA Default: [source-dir]/restore.offsets | |
--command-config COMMAND_CONFIG |
Property file containing configs to be passed to Admin Client. Only useful if you have additional connection options | |
--help HELP |
Prints this message | |
--debug DEBUG |
Print Debug information (if using the environment variable, set it to 'y') |
This project is licensed under the Apache License Version 2.0 (see LICENSE).