Skip to content

Latest commit

 

History

History
56 lines (40 loc) · 3.36 KB

maintenance.live-monitoring.md

File metadata and controls

56 lines (40 loc) · 3.36 KB

Live (Online) Monitoring

Table of Contents:

Monitoring Your Cluster with Datastax MCAC

If Datastax MCAC is configured to be installed by our toolkit, Cassandra.toolkit's ansible playbook will generate some artifacts for Datastax MCAC for you, including:

  • Datastax MCAC grafana dashboards
  • Adding a new Cassandra javaagent
  • Generating a docker compose file

You can find these in the artifacts dir (src/ansible/artifacts/datastax-mcac) of your ansible controller node. If the datastax-mcac/ dir is not there, it means that it was not configured correctly, and you'll want to double check your configuration and try running the ansible playbook again.

The docker compose file can be ran using:

# assuming: 
#   - calling from project root
#   - Datastax MCAC version 0.1.10
docker-compose -f ./src/ansible/artifacts/datastax-mcac/datastax-mcac-dashboards-0.1.10/docker-compose.yaml up -d

Grafana should now be viewable at http://localhost:3000, and you can use admin as username and admin as password for credentials. The Prometheus GUI should be available at http://localhost:9090 as well.

Monitoring Your Cluster with Cassandra Exporter

Cassandra Exporter (With Prometheus and Grafana)

Note that as of May 2021, Cassandra Exporter is still in Beta.

One difference between that documentation and what you can do now that you setup cassandra.toolkit on your cluster is that we have generated a docker-compose.yml file for you to use, assuming you set install_cassandra_exporter=True when setting ansible variables. This docker-compose file makes it easy to start running prometheus and grafana on your ansible control node.

# assuming calling from project root
docker-compose -f ./src/ansible/artifacts/docker/docker-compose.yml up -d 

Grafana should now be viewable at http://localhost:3000.

Live Monitoring with Filebeat, Elasticsearch, and Kibana

For more on this process, checkout the following resources:

One difference between that documentation and what you can do now that you setup cassandra.toolkit on your cluster is that we have generated a docker-compose.yml file for you to use, assuming you set install_cassandra_exporter=True when setting ansible variables. This docker-compose file makes it easy to start running Elasticsearch and Kibana on your ansible control node.

# assuming calling from project root
docker-compose -f ./src/ansible/artifacts/docker/docker-compose.yml up -d 

Kibana should now be viewable at http://localhost:5601.

Note that although Filebeat, Elasticsearch, and Kibana can be used for online monitoring, they can also be used for offline monitoring. Click here for documentation on offline monitoring with Filebeat, Elasticsearch, and Kibana.