Discover how a bimodal integration strategy can address the major data management challenges facing your organization today.
Get the Report →Integrate Databricks Data into Power Automate Desktop using CData Connect
CData Connect for Databricks Data enables you to integrate Databricks data into workflows built using Microsoft Power Automate Desktop.
CData Connect enables you to access live Databricks data in workflow automation tools like Power Automate. This article shows how to integrate Databricks data into a simple workflow, saving Databricks data into a CSV file.
CData Connect provides a live interface for Databricks, allowing you to integrate with live Databricks data in Power Automate — without replicating the data. Connect uses optimized data processing out of the box to push all supported SQL operations (filters, JOINs, etc) directly to Databricks, leveraging server-side processing to quickly return Databricks data.
About Databricks Data Integration
Accessing and integrating live data from Databricks has never been easier with CData. Customers rely on CData connectivity to:
- Access all versions of Databricks from Runtime Versions 9.1 - 13.X to both the Pro and Classic Databricks SQL versions.
- Leave Databricks in their preferred environment thanks to compatibility with any hosting solution.
- Secure authenticate in a variety of ways, including personal access token, Azure Service Principal, and Azure AD.
- Upload data to Databricks using Databricks File System, Azure Blog Storage, and AWS S3 Storage.
While many customers are using CData's solutions to migrate data from different systems into their Databricks data lakehouse, several customers use our live connectivity solutions to federate connectivity between their databases and Databricks. These customers are using SQL Server Linked Servers or Polybase to get live access to Databricks from within their existing RDBMs.
Read more about common Databricks use-cases and how CData's solutions help solve data problems in our blog: What is Databricks Used For? 6 Use Cases.
Getting Started
Configure Databricks Connectivity for Power Automate
Connectivity to Databricks from Power Automate is made possible through CData Connect Cloud. To work with Databricks data from Power Automate, we start by creating and configuring a Databricks connection.
- Log into Connect Cloud, click Connections and click Add Connection
- Select "Databricks" from the Add Connection panel
-
Enter the necessary authentication properties to connect to Databricks.
To connect to a Databricks cluster, set the properties as described below.
Note: The needed values can be found in your Databricks instance by navigating to Clusters, and selecting the desired cluster, and selecting the JDBC/ODBC tab under Advanced Options.
- Server: Set to the Server Hostname of your Databricks cluster.
- HTTPPath: Set to the HTTP Path of your Databricks cluster.
- Token: Set to your personal access token (this value can be obtained by navigating to the User Settings page of your Databricks instance and selecting the Access Tokens tab).
- Click Create & Test
- Navigate to the Permissions tab in the Add Databricks Connection page and update the User-based permissions.
Add a Personal Access Token
If you are connecting from a service, application, platform, or framework that does not support OAuth authentication, you can create a Personal Access Token (PAT) to use for authentication. Best practices would dictate that you create a separate PAT for each service, to maintain granularity of access.
- Click on your username at the top right of the Connect Cloud app and click User Profile.
- On the User Profile page, scroll down to the Personal Access Tokens section and click Create PAT.
- Give your PAT a name and click Create.
- The personal access token is only visible at creation, so be sure to copy it and store it securely for future use.
With the connection configured, you are ready to connect to Databricks data from Power Automate Desktop.
Integrate Databricks Data into Power Automate Workflows
After configuring CData Connect with Databricks, you are ready to integrate Databricks data into your Power Automate workflows. Open Microsoft Power Automate, add a new flow, and name the flow.
In the flow editor, you can add the options to connect to Databricks, query Databricks using SQL, and write the query results to a CSV document.
Add an Open SQL Connection Action
Add an "Open SQL connection" action (Action -> Database) and click the option to build the Connection string. In the Data Link Properties wizard:
- On the Provider tab: select Microsoft OLE DB Driver for SQL Server
- On the Connection tab:
- Select or enter a server name: set to tds.cdata.com,14333
- Enter information to lag onto the server: select "Use a specific username and password"
- Set User name to your CData Connect Cloud user name (e.g. [email protected])
- Set Password to your PAT
- Select the database: use the database configured above (e.g. Databricks1)
- Click "Test Connection" to ensure the connection is configured properly
- Click "OK"
After building the connection string in the Data Link Properties wizard, save the action.
Add an Execute SQL Statement Action
Add an "Execute SQL statement" action (Action -> Database) and configure the properties.
- Get connection by: SQL connection variable
- SQL connection: %SQLConnection% (the variable from the "Open SQL connection" action above)
- SQL statement: SELECT * FROM Customers
After configuring the properties, save the action.
Add a Write to CSV File Action
Add a "Write to CSV file" action (Action -> File) and configure the properties.
- Variable to write to: %QueryResult% (the variable from the "Execute SQL statement" action above)
- File path: set to a file on disk
- Configure Advanced settings as needed.
After configuring the properties, save the action.
Add a Close SQL Connection Action
Add a "Close SQL connection" action (Action -> Database) and configure the properties.
- SQL Connection: %SQLConnection% (the variable from the "Open SQL connection" action above)
After configuring the properties, save the action.
Save & Run the Flow
Once you have configured all the options for the flow, click the disk icon to save the flow. Click the play icon to run the flow.
Now you have a workflow to save Databricks data into a CSV file.
With CData Connect Cloud, you get live connectivity to Databricks data within your Microsoft Power Automate workflows.
SQL Access to Databricks Data from Cloud Applications
Now you have a direct connection to live Databricks data from Power Automate tasks. You can create more connections and workflows to drive business — all without replicating Databricks data.
To get SQL data access to 100+ SaaS, Big Data, and NoSQL sources directly from your cloud applications, sign up for a free trial of CData Connect Cloud.
Related Power Automate Articles
This article walks through using CData Connect Cloud with Power Automate Desktop. Check out our other articles for more ways to work with Power Automate (Desktop & Online):