Versions Compared

Key

  • This line was added.
  • This line was removed.
  • Formatting was changed.
Info

In order to accelerate telemetry analysis and processing, LHO can be configured to use an external Databricks Job instead of using In-App processing.

Databricks Analyzer Job (DAJ) Tasks

Databricks Analyzer Job is used for the following tasks

  1. Processing and analyzing telemetry data

  2. Processing cluster-events (including tags and estimated costs)

  3. Persisting data in the LHO database

Step 1

...

  1. Download the JAR file for bplm-analyser
    from blobstorage location for LHO libraries

...

  1. Download from LHO libraries from analysis-module the scala files which contain the notebook files that will be used to configure the Databricks Jobs

Step 2

For each create DAJ Task create a Databricks Job.

...

Code Block
  Recommended cluster specs:
  - for analyses: depending on load (4/8 moderate nodes)
  - for cluster events: SINGLENODE cluster of moderate instance type
  - for persisting (storage): SINGLENODE cluster of moderate instance type

Step 3 Link DAJ Persistence Execution

(1) Configure Analyzer Processor Job

...

copy ID of persisting job
got to Analyzer Cluster Events job
set as Parameter the storageJobId

Step 4 Configure LHO Environment Variables

LHO environment variables

...

Code Block
METRIC_PROCESSOR_ENABLED=false
DAJ_WORKSPACE_HOST=dbc-f4095c7c-e857.cloud.databricks.com
DAJ_JOB_ID=967951742258704,931886104401413
DAJ_SCHEDULE_CRON_EXPRESSION="0 30 */3 * * ?"

...

Step 5 Validate that DAJ is enabled

...