Hasura GraphQL Engine is an awesome piece of software, check it out! There is also a great cloud offering and a pro version, which holds a lot more very useful features.
For small projects, where I wanted to self host the graphql engine, it is laking visibility in terms of metrics. This project can help with this, since it reads and parses the hasura logs and collects and provides prometheus metrics from it.
The logs can expected to be read from a log file.
The application will start a web server on ${LISTEN_ADDR}
, where the metrics
can be scraped from the /metrics
path.
Hasura should be configured to at least emit http-log
, webhook-log
and query-log
log types.
Other envvars needed for configuration:
HASURA_GRAPHQL_ENDPOINT
The hasura endpoint (defaults to http://localhost:8080
)
HASURA_GRAPHQL_ADMIN_SECRET
The hasura admin secret is optional, if its not provided,
some collectors are disabled.
metrics 0.1.7
A prometheus metric generator for Hasura based on the log stream
USAGE:
metrics [OPTIONS] --logfile <logfile>
OPTIONS:
--collect-interval <collect-interval>
[env: COLLECT_INTERVAL=] [default: 15000]
--concurrency-limit <concurrency-limit>
[env: CONCURRENCY_LIMIT=] [default: 0]
--exclude_collectors <collector>[;collector...]
[env: EXCLUDE_COLLECTORS=] [possible values: cron-triggers, event-triggers,
scheduled-events, metadata-inconsistency]
-h, --help
Print help information
--hasura-admin-secret <hasura-admin-secret>
[env: HASURA_GRAPHQL_ADMIN_SECRET=]
--hasura-endpoint <hasura-endpoint>
[env: HASURA_GRAPHQL_ENDPOINT=] [default: http://localhost:8080]
--histogram-buckets <histogram-buckets>
[env: HISTOGRAM_BUCKETS=]
-l, --common-labels <common-labels>
[env: COMMON_LABELS=]
--listen <listen>
[env: LISTEN_ADDR=] [default: 0.0.0.0:9090]
--logfile <logfile>
[env: LOG_FILE=]
--sleep <sleep>
[env: SLEEP_TIME=] [default: 1000]
-V, --version
Print version information
If you want to provide multiple values for some key in ENVIROMENT VARIABLE, they should be separated by ;
, for example:
EXCLUDE_COLLECTORS=cron-triggers;event-triggers;scheduled-events
-
hasura_log_lines_counter
This is a counter the counts all parsed log lines. The labels include the log type. -
hasura_log_lines_counter_total
This is a counter that is the sum of all counted log lines. -
hasura_query_execution_seconds
This is a histogram, that stores the query execution time in seconds. The labels are:
operation
which holds the operation name of the graphql query or nothing if none is provided.error
which holds the error code if an error was detected or nothing if this was successful
For each label setting there are the following entries:
hasura_query_execution_seconds_bucket
hasura_query_execution_seconds_sum
hasura_query_execution_seconds_count
-
hasura_request_counter
This is a counter that counts the number of http requests. It provides
status
the http status code andurl
the path that was called. -
hasura_request_query_counter
This is a counter that counts the number of queries. The labels are:
operation
which holds the operation name of the graphql query or nothing if none is provided.error
which holds the error code if an error was detected or nothing if this was successful
-
hasura_websockets_active
This is a gauge that holds the currently active websocket connections.
-
hasura_websockets_operations
This is a counter that counts the operations (queries) over websockets. The labels are:
operation
which holds the operation name of the graphql query or nothing if none is provided.error
which holds the error code if an error was detected or nothing if this was successful
-
hasura_websockets_operations_active
This is a gauge that holds the currently active websocket operations.
-
hasura_healthy
This is a gauge that is 1 if the instance is healthy or 0 otherwise
-
hasura_metadata_version
This is a gauge, that holds a
hasura_version
label, with the hasura version and the value of1
if that version was detected.
The following metrics are the same as in the project (https://github.com/zolamk/hasura-exporter), also the idea on how to access them is based on it. So all credit for these need to go to @zolamk, I just ported them here. These metrics are disabled if no admin secret is provided. Cron triggers and one off events won't work if the postgres database with the metadata is not accessible as a data source with the 'default' name.
-
hasura_pending_cron_triggers
,hasura_processed_cron_triggers
,hasura_successful_cron_triggers
,hasura_failed_cron_triggers
These are gauges, that shows the number of (pending, processed, successful, failed) cron triggers labeled with the trigger name
-
hasura_pending_event_triggers
,hasura_processed_event_triggers
,hasura_successful_event_triggers
,hasura_failed_event_triggers
These are gauges, that shows the number of (pending, processed, successful, failed) event triggers labeled with the trigger name
-
hasura_pending_one_off_events
,hasura_processed_one_off_events
,hasura_successful_one_off_events
,hasura_failed_one_off_events
These are gauges, that shows the number of (pending, processed, successful, failed) one off events
-
hasura_metadata_consistency_status
This is a gauge that is 1 if the instance metadata is consistent or 0 otherwise
Hasura logs are sent via opentelemetry gRPC. In the example provided, the logs can be viewed via jaeger on http://localhost:16686/search. In order to build a new docker image, in the main directory run:
docker build -t metric .
docker compose run
Don't use version v0.1.0
its broken.
The docker image ghcr.io/afitzek/hasura-metric-adapter:v0.1.6
needs four environment variables to be configured.
LISTEN_ADDR
: The listen address for the metric endpoint
LOG_FILE
: The log file, which will hold the hasura logs
HASURA_GRAPHQL_ENDPOINT
The hasura endpoint (defaults to http://localhost:8080
)
OPENTEL_ENDPOINT
The Open Telemetry collector endpoint (defaults to http://localhost:4317
)
HASURA_GRAPHQL_ADMIN_SECRET
The hasura admin secret this is required
An example k8s setup can be found in k8s_example.yaml.
The important pieces of the example are:
-
The adapter runs in sidecar mode with the hasura container
-
The containers use a shared namespace, so that a named pipe can be accessed in both containers
shareProcessNamespace: true
-
The hasura container and the adapter share a common volume called
logs
, which is mounted in both containers as/tmp/log
. -
Overwrite the hasura command to:
"/bin/sh", "-c", "rm -f /tmp/log/stdout.log && mkfifo /tmp/log/stdout.log && /bin/graphql-engine serve | tee /tmp/log/stdout.log"
This creates a named pipe, and pipes the logs of graphql-engine to the stdout for logging and to the named pipe for the metric adapter to collect.
(an alternative if you can't have shared process namespaces in the pod, is to use a file, but as @Hongbo-Miao pointed out in #11 the log file can become very big)
"/bin/sh", "-c", ": > /tmp/log/stdout.log && /bin/graphql-engine serve | tee /tmp/log/stdout.log"
This truncates the log file, to not count metrics on container restarts, starts the graphql-engine and pipes the stdout to stdout and the file
/tmp/log/stdout.log
. -
HASURA_GRAPHQL_ENABLED_LOG_TYPES
includeshttp-log
,webhook-log
andquery-log
. -
The metric adapter is set up to listen on port
9999
and read the log from the shared volume/tmp/log/stdout.log
.
A docker-compose.yml is provided, ready to be used, following the same principles used on the construction of the k8s_example.yaml.
In order to run it, execute:
docker-compose build
docker-compose up
- Event Log errors:
If you see constant errors that the
event_log
is not existing, than you probably don't use events. Hasura might only create theevent_log
table when needed, in this case disable the appropriate collector. For exampleEXCLUDE_COLLECTORS=event-triggers