HCL SW Blogs
Select Page

Let us begin with understanding of Azure what it is all about before moving to our Azure Databricks plugin and how it benefits our workload automation users.


“Azure is an open and flexible cloud platform that enables you to quickly build, deploy and manage applications across a global network of Microsoft-managed datacentres. You can build applications using any language, tool, or framework. And you can integrate your public cloud applications with your existing IT environment.”

Azure is incredibly flexible, and allows you to use multiple languages, frameworks, and tools to create the customised applications that you need. As a platform, it also allows you to scale applications up with unlimited servers and storage.

What is an Azure Databricks?

Azure Databricks provides the latest versions of Apache Spark and allows you to seamlessly integrate with open-source libraries. Spin up clusters and build quickly in a fully managed Apache Spark environment with the global scale and availability of Azure.

Azure Databricks is a data analytics platform optimized for the Microsoft Azure cloud services platform. Azure Databricks offers three environments for developing data intensive applications: Databricks SQL, Databricks Data Science & Engineering, and Databricks Machine Learning.

  • Azure databricks plugin supports the latest analytic services for Azure data lake storage Gen1, Azure data lake storage Gen2, Azure blob storage, and other storage accounts.
  • It provides an easy-to-use platform for analysts to read data from multiple data sources and turn it into easily understandable data.
  • It provides faster performance with various optimizations at the I/O layer and processing layer (Databricks I/O).

Azure Databricks Plugin

Log in to the Dynamic Workload Console and open the Workload Designer. Choose to create a new job and select “Azure Databricks Plugin” job type in the Cloud section.

Fig1: Job DefinitionFig1: Job Definition

Connection Tab

Establishing connection to the Azure Databricks Workspace.

Workspace InstanceA unique instance name (per-workspace URL), that is assigned to each Azure Databricks deployment. It is the fully-qualified domain name used to log into your Azure Databricks deployment and make API requests.

Example: adb-<workspace-id>.<random-number>.azuredatabricks.net. The workspace ID appears immediately after adb- and before the “dot” (.). For the per-workspace URL https://adb-5555555555555555.19.azuredatabricks.net/

Access TokenEnter the access token that is generated in the Azure cloud, to authenticate to and access Databricks REST APIs.

We can generate token under User settings in workspace.

Fig 2Fig 2

Test Connection – Click to verify that the connection to the Azure server works correctly.

Connection TabFig 3: Connection Tab

Action Tab

Use this section to define the operation details.


  • Run the selected job
  • Cancel the pending/running job

SelectIt gives the list of available jobs present in the workspace to select.

Details – It gives more information on selected job.

Action TabFig 4: Action Tab

Submitting your job

It is time to Submit your job into the current plan. You can add your job to the job stream that automates your business process flow. Select the action menu in the top-left corner of the job definition panel and click on Submit Job into Current Plan. A confirmation message is displayed, and you can switch to the Monitoring view to see what is going on.

Monitor page with extra propertiesFig 5: Monitor page with extra properties

Once we submit the job, we can cancel the particular job by Kill option.

Fig 6Fig 6

Job Log

Fig 7Fig 7

WA Logo

Are you curious to try out the Azure Storage plugin? Download the integrations from the Automation Hub and get started or drop a line at santhoshkumar.kumar@hcl.com



Comment wrap
Further Reading
Workload Automation Administrator
Automation | September 26, 2022
Easy Websphere’s Liberty Management for the Workload Automation Administrator
Learn about how HWA main components are deployed on WebSphere liberty, that was designed to be highly efficient and optimized for modern cloud technologies.
Retail SuperMarket UseCases over Self Service Catalogue
Automation | September 8, 2022
CASE STUDY : Retail SuperMarket UseCases over Self Service Catalogue
A comprehensive solution to manage the workload automation process with HCL for Retail Industry. Read the case study to figure out complete solutions.
Filters result by