Reaper for Apache Cassandra® web interface for Cassandra repairs
Repairs are a critical anti-entropy operation in Apache Cassandra®. In the past, there have been many custom solutions to manage them outside of your main Cassandra installation. K8ssandra provides the Reaper web interface that eliminates the need for a custom solution. Just like K8ssandra makes Cassandra setup easy, Reaper makes configuration of repairs even easier.
TipThe requirement for your environment may vary considerably, however the general recommendation is to run a repair operation on your Cassandra clusters about once a week.
- Web Browser
values.yamlconfiguration, or use
--setflags on the command line
- Kubernetes cluster with the following elements deployed:
- DNS name configured for the Reaper web interface
As an alternative to configuring an Ingress, consider port forwarding. It’s another way to provide external access to resources that have been deployed by K8ssandra in your Kubernetes environment. Those resources could include Prometheus metrics, pre-configured Grafana dashboards, and the Reaper web interface for repairs of Cassandra® data. The
kubectl port-forward command does not require an Ingress/Traefik to work.
Access the Reaper web interface
With the prerequisites satisfied the Reaper web interface should be available at the following address:
For example, to upgrade a previously installed
k8ssandra that’s running locally, where the
releaseName used on the prior
helm install command was
helm upgrade k8ssandra k8ssandra/k8ssandra --set reaper.ingress.enabled=true,reaper.ingress.host=localhost
TipNotice how in the
helm upgradeexample above, with Ingress/Traefik access, the DNS hostname is specified on the command line as
localhost. If you are not running locally, specify a hostname other than
After about five or more minutes, check that the pods are running. Example:
kubectl get pods
NAME READY STATUS RESTARTS AGE k8ssandra-cass-operator-65cc657-fq6bc 1/1 Running 0 10m k8ssandra-dc1-default-sts-0 3/3 Running 0 10m k8ssandra-dc1-stargate-bb47877d5-54sdt 1/1 Running 0 10m k8ssandra-grafana-7f84d96d47-xd79s 2/2 Running 0 10m k8ssandra-kube-prometheus-stack-operator-76b984f9f4-pp745 1/1 Running 0 10m k8ssandra-medusa-operator-6888946787-qwzsx 1/1 Running 2 10m k8ssandra-reaper-k8ssandra-656f5b77cc-nqfzv 1/1 Running 0 10m k8ssandra-reaper-k8ssandra-schema-88cpx 0/1 Completed 0 10m k8ssandra-reaper-operator-5b8c4c66b8-8cf86 1/1 Running 2 10m prometheus-k8ssandra-kube-prometheus-stack-prometheus-0 2/2 Running 1 10m
What can I do in Reaper?
To access Reaper, navigate to http://localhost:8080/webui/.
TipAgain, if you are not running locally, use the hostname value in the URL that you specified on the
--set reaper.ingress.hostflag above.
Check the cluster’s health
In the Reaper UI, notice how the nodes are displayed inside the datacenter for the cluster.
The color of the nodes indicates the overall load the nodes are experiencing at the current moment.
Schedule a cluster repair
On the UI’s left sidebar, notice the Schedule option.
Click Add schedule and fill out the details when you are done click the final add schedule to apply the new repair job. A Cassandra best practice is to have one repair complete per week to prevent zombie data from coming back after a deletion.
Enter values for the keyspace, tables, owner, and other fields. Then click Add Schedule. The details for adding a schedule are similar to the details for the Repair form, except the “Clause” field is replaced with two fields:
- “Start time”
- “Interval in days”
After creating a scheduled repair, the page is updated with a list of Active and Paused repair schedules.
ImportantWhen choosing to add a new repair schedule, we recommended that you limit the repair schedules to specific tables, instead of scheduling repairs for an entire keyspace. Creating different repair schedules will allow for simpler scheduling, fine-grain tuning for more valuable data, and easily grouping tables with smaller data load into different repair cycles. For example, if there are certain tables that contain valuable data or a business requirement for high consistency and high availability, they could be scheduled for repairs during low-traffic periods.
For additional information, see Schedule a cluster repair on the Reaper site.
WarningUsers with access to the Reaper web interface can pause or delete scheduled repairs. To add authentication security in the UI, see Reaper authentication.
When you enable the autoscheduling feature, Reaper dynamically schedules repairs for all non-system keyspaces in a cluster. A cluster’s keyspaces are monitored and any modifications (additions or removals) are detected. When a new keyspace is created, a new repair schedule is created automatically for that keyspace. Conversely, when a keyspace is removed, the corresponding repair schedule is deleted.
To enable autoscheduling, set the property
Run a cluster repair
On the repair job you just configured, click Run now.
Notice the repair job kicking off.
Was this page helpful?
Glad to hear it! Please tell us how we can improve.
Sorry to hear that. Please tell us how we can improve.