CrowdStrike Falcon LogScale
Forwarding data from Apica Ascent to CrowdStrike Falcon LogScale (formerly Humio) is highly efficient because LogScale is built on a high-speed, index-free architecture that natively supports OpenTelemetry (OTLP) and the CrowdStrike Parsing Standard (CPS).
By using Apica Flow as the forwarder, you can normalize your telemetry to the CPS schema before it hits your LogScale ingest quota.
1. Prerequisites in Falcon LogScale
You must first create a "landing zone" in your LogScale repository to receive the OTLP stream.
Select a Repository: Open the Falcon Console and navigate to Next-Gen SIEM > Log Management. Select the repository where you want the data to reside.
Create an Ingest Token: * Go to Settings > Ingest Tokens.
Click + Add Token.
Give it a name (e.g.,
Apica_Forwarder).Crucial: Assign a Parser to this token (e.g.,
jsonor a customCPS-compliantparser). This ensures LogScale knows how to handle the incoming OTLP payload.
Capture the Ingest URL: Your endpoint depends on your region:
US-1:
https://cloud.us.humio.com/api/v1/ingest/otlpEU-1:
https://cloud.community.humio.com/api/v1/ingest/otlp
2. Configuration Strategy: The Forwarder
In the Apica Flow (Ascent) UI, you will create a new target destination using the OTLP/HTTP protocol.
Field
Value
Destination Name
CrowdStrike_LogScale_Forwarder
Endpoint
https://<your-region-url>/api/v1/ingest/otlp
Protocol
http/protobuf
Auth Header Key
Authorization
Auth Header Value
Bearer <Your-LogScale-Ingest-Token>
3. Detailed Reference: Metadata & Enrichment (CPS)
CrowdStrike uses the CPS (CrowdStrike Parsing Standard), which is based on the Elastic Common Schema (ECS). To make your Apica data "search-ready" in LogScale, you should map your OTel attributes to these specific keys.
SQL
4. Implementation Reference: Exporter Configuration
If you are managing the forwarder via a standalone OTel Collector or the Apica Fleet Agent:
YAML
5. Key Implementation Notes
Index-Free Speed: LogScale does not require you to pre-define schemas. However, using the
#Vendorand#Producttags in Apica Flow will allow LogScale to automatically apply Marketplace Apps (dashboards and alerts) to your data.Batching for Performance: LogScale thrives on large batches. Ensure the
batchprocessor in Apica is set to a high timeout (e.g.,5s) or size (e.g.,1000spans) to maximize ingestion throughput.Timestamping: Ensure you are sending the
Timestampattribute. LogScale uses this for its "Live Tail" view. If omitted, LogScale will use the "arrival time," which can cause issues with out-of-order events.
Last updated
Was this helpful?