Azure Databricks
Last updated
Last updated
This guide take you through how you can forward your logs from an Azure Databricks cluster to Apica Ascent. Before you proceed with this setup, ensure that you meet the following prerequisites.
Private VNI
An Azure Databricks cluster in private VNI
Apica Ascent endpoint
Note: The Databricks cluster must be launched in your own private EMI failing which the default deployment of the Databricks cluster will be fully managed by Azure, the resource group will be locked, and SSH connections to the node will be disabled.
For more information on deploying Azure Databricks in your own private EMI, refer to Deploy Azure Databricks in your Azure virtual network (VNet injection).
To configure your Azure Databricks cluster to forward logs to your Apica Ascent endpoint, do the following.
Navigate to the Compute section on your Azure portal.
Click Create Cluster.
Choose your cluster size.
Click Advanced options > SSH. Paste your public key under SSH public key. You can generate a public key by running the command ssh-keygen -t rsa -b 4096 -C "email-id”
. You will use the private key to log into the machine later on.
Next, on the Azure portal, under Network security group, add port 2200
in the Inbound ports section for the machines that the Databricks cluster spun up.
To install and configure Fluent Bit on your Databricks cluster, do the following.
Log into the machine using the following command.
Install Fluent Bit as per the version of Ubuntu OS running on the machine. For detailed installation instructions, refer to the Fluent Bit documentation.
Use the following Fluent Bit configuration file.
In the Fluent Bit configuration file above, substitute the following details based on your implementation.
logiq-endpoint
TOKEN
Databricks-worker
Next, replace the existing configuration at /etc/td-agent-bit/td-agent-bit.conf
with the modified file.
Finally, restart Fluent Bit by running the following command.
Now, when you log into your Apica Ascent UI, you should see the logs from your Azure Databricks cluster being ingested. See the Explore Section to view the logs.