1 Reply Latest reply on Oct 28, 2015 3:55 AM by Thomas Heute

    hawkular sizing

    Juergen Hoffmann Newbie

      Hi Folks,


      I am pretty new to hawkular and cassandra.


      I have a question around sizing my environment for the monitoring of wildfly using hawkular. Do make a good estimation, I would like to know how much data is stored for a typical wildfly instance. If I look at Hawkular - Quick start I see some metrics being gathered every 30 seconds.

      • Is that the default / recommended collection of metrics to monitor a wildfly instance?
      • How much is space is used for a typical wildfly instance over the period of a year? Is that 1GB, 5GB or more? It would be great if you could share some information about the sizing
      • AFAIK a cassandra cluster is the best recommended approach for a data storage. Can you elaborate on how many read/write requests a typical cassandra node is able to handle?
      • How can I monitor cassandra performance and are there any official sizing recommendations of cassandra.


      Any help is highly appreciated



        • 1. Re: hawkular sizing
          Thomas Heute Master


          thanks for the question, they definitely make sense. I wished we had the numbers yet, but so far we are still under heavy development working on Alpha releases. Those numbers are likely frequently changing.


          As a beginning of an answer, the collection of metrics will really depend on the infrastructure you are ready to put in place, by experience with RHQ, some will prefer to decrease collection or metrics to monitor to limit the management solution to a single host, some will want to have more precise datapoints (or be alerted sooner) and will invest in a larger deployment, IMO in general the collection would be between 5s to 5min for metrics that change often and 30m or 1h for metrics that barely change.


          Question about disk usage and Cassandra cluster size are questions I really want to be able to answer ASAP, your question will help prioritize this. We need to tell the maximum ingestion rate for a single "average" server and how much disk space does it use over a year (extrapolating), unfortunately I don't have those numbers yet and we need to make this automated since it will change rapidly in the coming months.