Monitoring your JHipster Applications
JHipster provides several options to monitor your applications at runtime.
Summary
- Generated dashboards
- JHipster Registry
- JHipster Console
- Forwarding metrics to a supported third party monitoring system
- Zipkin
- Alerting with Elastalert
Generated dashboards
For monoliths and gateways, JHipster generates several dashboards to monitor each application. Those dashboards are available at runtime, and are the easiest way to do some simple monitoring.
The metrics dashboard
The metrics dashboard uses Dropwizard metrics to give a detailed view of the application performance.
It gives metrics on:
- the JVM
- HTTP requests
- methods used in Spring Beans (using the
@Timed
annotation) - database connection pool
By clicking on the eye next to the JVM thread metrics, you will get a stacktrace of the running application, which is very useful to find out blocked threads.
The health dashboard
The health dashboard uses Spring Boot Actuator’s health endpoint to give health information on various parts of the application. Many health checks are provided out-of-the-box by Spring Boot Actuator, and it’s also very easy to add application-specific health checks.
The logs dashboard
The logs dashboard allows to manage at runtime the Logback configuration of the running application. Changing the log level of a Java package is as simple as clicking on a button, which is very convenient both in development and in production.
JHipster Registry
The JHipster Registry has its own documentation page here.
It mostly provides the same monitoring dashboards as in the previous section, but it works on a separate server. As such, it is a bit more complex to set up, but it is highly recommended to have dashboards running outside of the running application: otherwise, they won’t be available when there is an application failure.
JHipster Console
The dashboards described in the previous sections only show the current value of application metrics, when advanced users want to monitor the evolution of those values over time.
Therefore JHipster applications can be configured to forward their metrics to an external monitoring system where they can be graphed over time and analyzed.
To achieve this, JHipster provide the JHipster Console, a custom monitoring solution based on the ELK stack and fully integrated with JHipster.
Forwarding logs to the JHipster Console
To configure a JHipster application to forward their logs to JHipster Console, enable logstash logging in their application-dev.yml
or application-prod.yml
:
jhipster:
logging:
logstash:
enabled: true
host: localhost # If using a Virtual Machine on Mac OS X or Windows with docker-machine, use the Docker's host IP here
port: 5000
queueSize: 512
To configure metrics monitoring, enable metrics log reporting in your JHipster application:
jhipster:
metrics:
logs:
enabled: true
reportFrequency: 60 # seconds
Setting those properties will enrich your forwarded logs with metrics coming from Dropwizard metrics.
Overview of the JHipster Console
The JHipster Console is a monitoring tool based on the ELK Stack. It provides ready-to-use dashboards and analytics tools to have a real-time overview of your infrastructure’s performance.
It is an Open Source application, available on GitHub at jhipster/jhipster-console.
The ELK stack is composed of:
- Elasticsearch for indexing the data (logs and metrics)
- Logstash to manage and process the logs received from the applications
- Kibana to visualize the logs with a nice interface
The JHipster Console is a Docker-based project that adds features on top of the official Elasticsearch, Logstash and Kibana Docker images. We have made a few visual changes to Kibana and set up useful dashboards, so that you can get started to monitor your JHipster applications in minutes instead of the hours that would be needed to set up your own monitoring infrastructure.
Monitoring a JHipster microservice architecture
The JHipster Console fully supports the monitoring of a JHipster microservice architecture and even provides the following microservice specific features :
- Distributed tracing with Zipkin
- Log enriching with service name, instance ID, Zipkin correlation IDs
- Zipkin server and UI to visualize traces and spans
- Linking between the Zipkin UI and Kibana so that you can jump to the logs corresponding to a particular trace ID (to use this, click on the Logs icon in the trace page)
Setting up JHipster Console
If you already have a JHipster microservice architecture set up with the Docker Compose workflow, the JHipster Console can be automatically set up by the Docker Compose sub-generator.
If you are using the monolithic version of JHipster, you can get the JHipster Console’s Docker-Compose file from GitHub or with the following command:
curl -O https://raw.githubusercontent.com/jhipster/jhipster-console/master/bootstrap/docker-compose.yml
Then you will be able to start the console with:
docker-compose up -d
It will start Elasticsearch, Logstash, Kibana and ElastAlert all at once. You will then be able to access the JHipster Console at http://localhost:5601. It should automatically receive logs from your applications if they have been correctly configured to forward their logs and metrics to Logstash.
To stop everything, run:
docker-compose stop
Once stopped, you can remove the containers if you don’t intend to start them again:
docker-compose rm
You can combine the two previous commands in one by running: docker-compose down
.
Using JHipster Console
Once your application is running with logs and metrics forwarding enabled, you can view a dashboards by clicking on the Load Saved Dashboards icon ( ) in the Dashboard tab.
You can also use Kibana’s Discover and Visualize tabs to explore your data and create new visualizations. To understand how to use Kibana’s interface effectively please refer to its official documentation in particular the Discover, Visualize and Dashboard sections of the Kibana User Guide.
Data persistence with docker volumes
When using JHipster Console you can enable docker volumes in the docker-compose.yml
file by uncommenting the appropriate lines. Those volumes are used to share data between containers and the host. They will persist data and configuration even if containers are removed from your system.
- Elasticsearch has its data saved to
log-data/
- Logstash loads its configuration from
log-conf/logstash.conf
, you can edit this file to add new parsing rules for data received by logstash on UDP port 5000. - Kibana loads dashboards description files in
dashboards/
on each startup.
Save your custom searches, visualizations and dashboards as JSON for auto import
Searches, visualization and dashboards created in Kibana can be exported using the Management > Saved Objects menu.
You can then extract the JSON description of a specific object under the _source
field of the export.json file.
You can then put this data in a JSON file in one of the jhipster-console/dashboards
sub-folder for auto-import.
If you have created useful dashboards and visualizations for your JHipster applications please consider contributing those back to the community by submitting a Pull Request on the JHipster Console’s GitHub project.
Forwarding metrics to a supported third party monitoring system (JMX, Prometheus)
JHipster also provides a Metrics exporters for JMX.
Forwarding metrics to alternative systems is also supported and can also simply be enabled in your YAML configuration files.
jhipster:
metrics:
jmx:
enabled: true
Note: Unlike in previous JHipster versions, JHipster 5 metrics reporting only support JMX out of the box. Please have a look to the Metrics official documentation for instructions on how to setup other reporter like Graphite.
JHipster also supports Prometheus as a Metrics exporter. Prometheus is disabled by default and can be enabled in your YAML configuration file.
jhipster:
metrics:
prometheus:
enabled: true
This will export your metrics under /prometheusMetrics
. As this endpoint is not secured, you can protect it with basic auth, such that
prometheus can still scrape the endpoint by creating a new configuration file (e.g. BasicAuthConfiguration.java
).
@Configuration
@Order(1)
@ConditionalOnProperty(prefix = "jhipster", name = "metrics.prometheus.enabled")
public class BasicAuthConfiguration extends WebSecurityConfigurerAdapter {
@Override
protected void configure(HttpSecurity http) throws Exception {
http
.antMatcher("/prometheusMetrics/**")
.authorizeRequests()
.anyRequest().hasAuthority(AuthoritiesConstants.ADMIN)
.and()
.httpBasic().realmName("jhipster")
.and()
.sessionManagement()
.sessionCreationPolicy(SessionCreationPolicy.STATELESS)
.and().csrf().disable();
}
}
You can login with the default admin/admin
. You must add following configuration to you prometheus configuration such that prometheus can still scrape your application.
basic_auth:
[ username: "admin" ]
[ password: "admin" ]
Zipkin
JHipster applications can integrate with Zipkin through Spring Cloud Sleuth to provide distributed tracing for your microservice architecture. To enable Zipkin tracing, package your application with the zipkin
maven/gradle profile and set the spring.zipkin.enabled
property to true. This will trigger span reporting to the Zipkin server and also add correlation IDs (TraceId, SpanId and ParentId) to request headers and logs. The Zipkin server and UI is provided as part of the JHipster Console and integrates with the Kibana dashboard.
Zipkin also provide a service dependency graph feature that lets you visualize the dependencies between microservices over time.
The Zipkin instance should be available on http://127.0.0.1:9411/ if you run it on your machine, or on http://<docker-host-ip>:9411 if you run it with Docker.
Alerting with Elastalert
JHipster Console comes with built-in alerting by integrating Elastalert, an alerting system that can generate alerts from data in Elasticsearch. Elastalert is simple to use and able to define complex alerting rules to detect failures, spikes or any pattern based on an Elasticsearch Query.
Enable alerting
To enable alerting, setup the jhipster-alerter
container by adding the following lines docker-compose.yml
.
jhipster-alerter:
image: jhipster/jhipster-alerter
#volumes:
# - ../jhipster-alerter/rules/:/opt/elastalert/rules/
# - ../alerts/config.yaml:/opt/elastalert/config.yaml
Configure alerting
Elastalert configuration can be modified in alerts/config.yaml
. For example, you can configure the alerting frequency and the buffer period, by changing the following properties:
run_every:
minutes: 1
buffer_time:
minutes: 5
Then you will need to write some rules that define when alerts will be thrown.
Write alertings rules
To define new alerts, add new Yaml rule files in alerts/rules
and then test them over past data with:
./test-alerting-rule.sh rule.yaml
Note that those Yaml files should have a .yaml
file extension. Read more on how to write rules at Elastalert’s official documentation.