Humio is configured by setting environment variables. The example configuration file below contains comments describing each individual option.

Docker Tip
When running Humio in Docker you can pass set the --env-file= flag and keep your configuration in a file. For a quick intro to setting configuration options see the documentation for installation on docker.

Docker only loads the environment file when the container is initially created. As such, if you make changes to the settings in your environment file, simply stopping and starting the container will not work. You need to docker rm the container and docker run it again to pick up changes.

Example configuration

# The stacksize should be at least 2M.

# Make Humio write a backup of the data files:
# Backup files are written to mount point "/backup".

# ID to choose for this server when starting up the first time.
# Leave commented out to autoselect the next available ID.
# If set, the server refuses to run unless the ID matches the state in data.
# If set, must be a (small) positive integer.

# The URL that other humio hosts in the cluster can use to reach this server.
# Required for clustering. Examples:  or  http://humio01:8080
# Security: We recommend using a TLS endpoint.
# If all servers in the Humio cluster share a closed LAN, using those endpoints may be okay.

# The URL which users/browsers will use to reach the server
# This URL is used to create links to the server
# It is important to set this property when using OAuth authentication or alerts.

## For how long should dashboard queries be kept running if they are not polled.
## When opening a dashboard, results will be immediately ready if queries are running.
## Default is 3 days 

## Warn when ingest is delayed
## How much should the ingest delay fall behind before a warning is shown in the search UI

# Specify the replication factor for the Kafka ingest queue

# Kafka bootstrap servers list. Used as `bootstrap.servers` towards kafka.
# should be set to a comma separated host:port pairs string.
# Example: `my-kafka01:9092` or `kafkahost01:9092,kafkahost02:9092`

# By default Humio will create topics and manage number of replica in Kafka for the topics being used.
# If you run Humio on top of an existing Kafka or want to manage this outside of Humio, set this to false.

# Deletes events from the ingest queue when they have been saved in Humio.
# Still it is important to configure Kafka retention on the ingest queue. 
# The Kafka retentionn defines how long data can be kept on the ingest queue and thus how much time Humio has to read the data and store it internally.

# It is possible to add extra Kafka configuration properties. by creating a properties file and pointing to it.
# These properties are added to all Kafka producers and consumers in Humio.
# For example, this enables Humio to connect to a Kafka cluster using SSL and SASL.
# Note the file must be mapped into Humio's Docker container - if running Humio as a Docker container

#Add a prefix to the topic names in Kafka. 
#Adding a prefix is recommended if you share the Kafka installation with applications other than Humio
# The default is not to add a prefix.

# Zookeeper servers.
# Defaults to "localhost:2181", which is okay for a single server system, but
# should be set to a comma separated host:port pairs string.
# Example: zoohost01:2181,zoohost02:2181,zoohost03:2181
# Note, there is NO security on the zookeeper connections. Keep inside trusted LAN.

# Maximum number of datasources (unique tag combinations) in a repo.
# There will be a sub-directory for each combination that exists.
# (Since v1.1.10)

# Compresions level for data in segment files. Defaults to 9, range is [1 ; 17]

# (Approximate) limit on the number of hours a segment file can be open for writing
# before being flushed even if it is not full. (Full is set using BLOCKS_PER_SEGMENT)
# Default: version < 1.4.x had 720, 1.4.x has 24

# How long can a mini-segment stay open. How long back is a fail-over likely to go?

# Desired number of blocks (each ~1MB before compression) in a final segment after merge
# Segments will get closed earlier if expired due to MAX_HOURS_SEGMENT_OPEN.
# Defaults to 2000.

# Desired number of blocks (each ~1MB before compression)
# in a mini-segment before merge. Defaults to 64.
# Mini-segments will get closed earlier if expired due to FLUSH_BLOCK_SECONDS

# Select roles for node, with current options being "all" or
# "httponly". The latter allows the node to avoid spending cpu time on
# tasks that are irrelevant to a nodes that has never had any local
# segments files and that will never any assigned either. Leave as
# "all" unless the node is a stateless http frontend or ingest
# listener only.

# How long should the digest worker thread keep working on
# flushing the contents of in-memory buffers when Humio is told to shut down
# using "sigterm" (normal shutdown). Default to 300 seconds as millis.
# If too low, then the next startup will need to start further back in
# time on the ingest queue.

# Let Humio send emails using the Postmark service
# Create a Postmark account and insert the token here

# Let Humio send emails using an SMTP server. ONLY put a password here
# if you also enable starttls. Otherwise you will expose your password.
# Example using GMail:
# Example using a local clear-text non-authenticated SMTP server

# Use a HTTP proxy for sending alert notifications
# This can be usefull if Humio is not allowed direct access to the internet

# Select the TCP port to listen for http.

# Select the TCP port for ElasticSearch Bulk API

# Select the TCP port for exporting Prometheus metrics. Disabled by default

# Select the IP to bind the udp/tcp/http listening sockets to.
# Each listener entity has a listen-configuration. This ENV is used when that is not set.

# Select the IP to bind the http listening socket to. (Defaults to HUMIO_SOCKET_BIND)

# Verify checksum of segments files when reading them. Default to true. Allows detecting partial and malformed files.
# (Since v1.1.16)

# S3 access keys for archiving of ingested logs

# Users need to be created in Humio before they can login with external authentication methods like SAML/LDAP/OAUTH etc.
# set this parameter to true - then users are automatically created in Humio when successfully logging with external authentication methods.
# Users will not have access to any existing repositories except for a personal sandbox repository when they are created.
# if false - users must be explicitly created in Humio before they can login.

# Allows disabling use of personal API tokens. This may be relevant when e.g.
# ldap or saml is set as authentication mechanism, as the personl API tokens
# never expire and thus allows a user to access Humio even when the ldap/saml
# account has been closed or deleted. Defaults to true.

# Initial partition count for storage partitions.
# Has effect ONLY on first start of first node in the cluster.

# Initial partition count for digest partitions.
# Has effect ONLY on first start of first node in the cluster.

# How big a backlog of events in Humio is allowed before Humio starts responding
# http-status=503 on the http interfaces and reject ingesting messages on http?
# Measured in seconds worth of latency from an event arrive at Humio until it has
# been fully processed.
# (Note that typical latency in normal conditions is is zero to one second.)
# Set to a large number, such as 31104000 (~1 year as seconds) to avoid
# having this kind of back pressure towards the ingest clients.
# Range: Min=300, Max=2147483647.

# A configuration flag to limit state in Humio searches. 
# For example this is used to limit the number of groups in the groupBy function.
# This is necessary to limit how much memory searches can use and avoid out of memory etc. 

# The maximum allowed value for the "limit" parameter on timechart (and bucket)

# Maximum allowed file size that can be uploaded to Humio, when uploading csv or json files.
# Used to set a limit on how big files can be.

# Limits how many entries are allowed when using the match and lookup function

# The maximum allowed number of points in a timechart (or bucket result)
# When this is hit the result will become approximate and discard input.

# SECONDARY_DATA_DIRECTORY enables using a secondary file system to
# store segment files. When to move the files is controlled by
# Secondary storage is not enabled by default.
# Note! When using docker, make sure to mount the volume
# into the container as well.
# See the page on "Secondary storage" for more information.

# Humio will write threaddumps to humio-threaddumps.log with the interval specified here
# If not specified Humio will write threaddumps every 10 seconds

Java virtual machine parameters

You can supplement or tune the Java virtual machine parameters used when running Humio with the HUMIO_JVM_ARGS environment variable. The defaults are:


Number of CPU Cores

You can specify the number of processors for the machine running Humio by setting the CORES property. Humio uses this number when parallelizing queries and other internal tasks.

By default, Humio uses the Java available processors function to get the number of CPU cores. This is usually the optimal number. Be aware that the auto-detected number can be too high when running in a containerized environment where JVM does not always detect the proper number of cores.

Derived from the number of CPU cores, Humio internally sets QUERY_EXECUTOR_CORES and DIGEST_EXECUTOR_CORES to half that number (but minimum of 2) to reduce pressure on context switching due to hyper-threading since the numer of CPU cores usually include hyper-threads. If the number of cores set through CORES is number of actual physical cores and not hyperthreads, you may want to set these to the same number as CORES. Note that raising this number above the default may lead to an unstable and slow system due to context switching costs growing to a point where no real work gets done when the system gets loaded, while it may appear to work fine when not fully utilized.

Configuring Authentication

Humio supports different ways of authentication users. Read more in the dedicated Authentication Documentation.

Public URL

PUBLIC_URL is the URL where the Humio instance is reachable from a browser. Leave out trailing slashes.

This property is only important if you plan to use OAuth Federated Login, Auth0 Login or if you want to be able to have Alert Notifications have consistent links back to the Humio UI.

The URL might only be reachable behind a VPN but that is no problem, as the user’s browser can access it.