Azure Logs Ingestion API
Send logs to Azure Log Analytics using Logs Ingestion API
Azure Logs Ingestion plugin lets you ingest your records using Logs Ingestion API in Azure Monitor to supported Azure tables or to custom tables that you create.
The Logs ingestion API requires the following components:
A Data Collection Endpoint (DCE)
A Data Collection Rule (DCR) and
A Log Analytics Workspace
To visualize the basic logs ingestion operation, see the following image:

To get more details about how to set up these components, refer to the following documentation:
Configuration parameters
tenant_id
The tenant ID of the Azure Active Directory (AAD) application.
none
client_id
The client ID of the AAD application.
none
dce_url
Data Collection Endpoint(DCE) URL.
none
table_name
The name of the custom log table (include the _CL
suffix as well if applicable)
none
time_key
Optional. Specify the key name where the timestamp will be stored.
@timestamp
time_generated
Optional. If enabled, will generate a timestamp and append it to JSON. The key name is set by the time_key
parameter.
true
compress
Optional. Enable HTTP payload gzip compression.
true
Get started
To send records into an Azure Log Analytics using Logs Ingestion API the following resources needs to be created:
A Data Collection Endpoint (DCE) for ingestion
A Data Collection Rule (DCR) for data transformation
Either an Azure tables or custom tables
An app registration with client secrets (for DCR access).
Follow this guideline to set up the DCE, DCR, app registration and a custom table.
Configuration file
Use this configuration file to get started:
pipeline:
inputs:
- name: tail
path: /path/to/your/sample.log
tag: sample
key: RawData
# Or use other plugins
#- name: cpu
# tag: sample
filters:
- name: modify
match: sample
# Add a json key named "Application":"fb_log"
add: Application fb_log
outputs:
# Enable this section to see your json-log format
#- name: stdout
# match: '*'
- name: azure_logs_ingestion
match: sample
client_id: XXXXXXXX-xxxx-yyyy-zzzz-xxxxyyyyzzzzxyzz
client_secret: some.secret.xxxzzz
tenant_id: XXXXXXXX-xxxx-yyyy-zzzz-xxxxyyyyzzzzxyzz
dce_url: https://log-analytics-dce-XXXX.region-code.ingest.monitor.azure.com
dcr_id: dcr-xxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx
table_name: ladcr_CL
time_generated: true
time_key: Time
compress: true
Set up your DCR transformation based on the JSON output from the Fluent Bit pipeline (input, parser, filter, output).
Last updated
Was this helpful?