Lumi concepts for Splunk users
AI summary
About AI summaries.
Imply Lumi is designed for compatibility with the Splunk® platform in your observability applications. If you use Lumi in conjunction with Splunk, it's important to understand how their similar concepts align and translate functionality from one system to another. You can ensure compatibility and continuity when making changes to your observability workflow.
This topic describes analogous concepts in Lumi and Splunk, focusing on the technical aspects of sending and searching events. The information is geared towards a Splunk administrator who's familiar with configuring universal or heavy forwarders in Splunk to ingest, process, and forward data.
Refer to the Splunk documentation for more information on configuring Splunk.
How to use this topic
This topic is organized by the flow of data through the Splunk pipeline. Depending on the topology of your deployment, you can apply the configuration settings on different Splunk processing components, whether the forwarder, indexer, or search head.
-
When you send events to Lumi, your forwarder configuration largely remains the same with the addition of defining a destination for Lumi. See Inputs and Outputs in this topic.
-
If you define event processing on the indexers, be sure to apply the equivalent processing settings in Lumi. For details, see Props and transforms and Splunk default fields.
-
You can search events in Lumi either directly in the explore view or through a federated search from Splunk. Commands in the query are processed by the federated search head (Splunk) or the remote search head (Lumi). For more information, see Search.
See the Splunk documentation on how the data pipeline relates to Splunk processing components.
The following diagram presents an overview of the relationship between Splunk and Lumi concepts:

Updating your configuration
Splunk stores configuration across multiple layers of .conf files for default, local, and app settings.
Where you make changes depends on the type of change and your deployment architecture.
For example, to configure local inputs on a forwarder, you edit $SPLUNK_HOME/etc/system/local/inputs.conf.
See the Splunk documentation to learn about the file structure and configuration precedence.
Some Splunk deployments use deployment servers to manage groups of Splunk Enterprise instances.
Deployment server updates can override any changes you make to forwarders' local .conf files.
If you use a deployment server, you need to update the appropriate deployment application rather than the local system files.
Alternatively, if you prefer to edit the .conf files locally, you can configure Splunk to preserve your local configuration.
Restart Splunk after you make any configuration changes.
Inputs
Inputs define the source of events you plan to ingest.
You set these in the Splunk configuration file inputs.conf.
In many cases, you don't need to update your existing inputs.
The batch input source in Splunk is similar to the file upload integration in Lumi.
File upload in Lumi is intended for you to ingest data quickly to evaluate how to use the product in your observability workflows.
Note that it's not intended for backfill scenarios to retroactively process historical data.
Receiver
The inputs.conf file also defines the Splunk receivers—how Splunk listens for incoming data.
You don't need to configure receivers in Lumi, but take note of the port Lumi uses to listen for incoming data.
For the port listed in an integration's prerequisites, ensure that it's open for outbound traffic on the machine forwarding the events to Lumi.
Outputs
In Splunk, the outputs.conf configuration file determines where to forward events, whether to a Splunk instance or a third-party system.
In Lumi, ingestion integrations are the routes available for event forwarding.
For example, if you already use the S2S data protocol from Splunk, you might have configured processors of type [tcpout] or [httpout] in your outputs configuration.
To integrate with Lumi, you add a target group to the existing output processor with configuration details for one of the S2S integrations in Lumi.
For the list of ingestion integrations in Lumi, see Send events to Lumi.
To configure event forwarding destinations in Splunk, see Configure forwarding with outputs.conf.
Props and transforms
Before you store or analyze observability data, you might choose to perform field extractions or mask sensitive data.
In Splunk, you configure this processing in the props.conf and transforms.conf configuration files.
Splunk has an add-on ecosystem of preconfigured settings to process specific types of logs. For example, the Splunk Add-on for Microsoft Windows handles Windows event logs.
In Lumi, you can apply some of these settings using pipelines and IAM key settings.
Pipelines
Pipelines transform events entering Lumi. They can add, transform, or remove metadata, which become user attributes on Lumi events.
A predefined pipeline in Lumi processes events with a preset series of transformations, analogous to a Splunk technology add-on (TA). For example, see the predefined pipeline for Windows event logs.
You can also create your own pipeline to parse events.
For an example of how a Splunk configuration from props.conf and transforms.conf maps to a pipeline, see Compare to Splunk configuration.
IAM key attributes
For events in the S2S protocol, you configure event parsing in the props.conf configuration file.
In Lumi, you apply event parsing settings on the IAM key.
For more information, see S2S attributes
and IAM key attributes.
Splunk default fields
Splunk default fields are metadata fields on the events, such as index, source, and sourcetype.
They provide context that you can use to drill down in your observability analysis.
The forwarders assign these metadata fields based on the properties in inputs.conf.
For the Splunk HTTP event collector, you can assign these fields on the HEC token or provide them in the JSON payload.
Lumi stores the Splunk default fields as user attributes on the incoming events. In Lumi, you can assign Splunk default fields by the following:
- Assign them as IAM key attributes. Applies to Splunk HEC, S2S over HTTP, and S3 pull.
- Send them in the JSON payload for Splunk HEC.
- Set them by a forwarding agent that sends the events.
- Assign them in a Lumi pipeline.
For details on how Lumi prioritizes assignment of user attributes, see Event model.
Index
An index in Splunk is the repository that stores the data. It's similar to a database table in terms of how you organize and scope your data. An index stores structured events and associated metadata after Splunk parses the events and applies any specified filters or transforms. In Splunk you also use an index for:
- Maintenance purposes, such as to define a retention rule or to limit user access.
- Filtering your searches, such as for drilling down into specific event features or optimizing query performance.
In Lumi, the index is a user attribute on an event. Treat it as any other metadata field to understand an event's context or to filter your searches. The Lumi index also applies to Splunk federated search. The Splunk federated index directly corresponds to an index in Lumi. On the Lumi IAM key, you can also set allowed indexes that determine what data is made available to Splunk.
For example, your events in Lumi store demo, test, or main for the index.
You create a federated index in Splunk and enter demo for the remote dataset.
When you search the federated index, Splunk only returns events that have the Lumi user attribute index: demo.
For details on assigning the index user attribute in Lumi, see Index user attribute.
In both Splunk and Lumi, the default index name is main.
Search
You can issue queries on Lumi events in Splunk through Splunk federated search. In federated searches, you use the Splunk Search Processing Language (SPL) commands and functions supported for use in Lumi.
When you perform a federated search in Splunk, some commands run on the federated search head (Splunk) and others run on the remote search head (Lumi). See Federated search reference for the commands that run on the federated search head by design and commands Lumi has implemented on the remote search head.
To learn how to search Lumi events in Splunk, see Search events with Splunk.
Learn more
See the following topics for more information:
- Send events to Lumi for integrations for sending events to Lumi.
- Search events with Lumi for how to search events in Lumi.
- Forward data to third-party systems for configuring Splunk to forward data to non-Splunk systems.
- List of configuration files for configuration files available in Splunk.