Ready to get started?

Download a free trial of the Azure Data Lake Storage ODBC Driver to get started:

 Download Now

Learn more:

Azure Data Lake Storage Icon Azure Data Lake Storage ODBC Driver

The Azure Data Lake Storage ODBC Driver is a powerful tool that allows you to connect with live data from Azure Data Lake Storage, directly from any applications that support ODBC connectivity.

Access Azure Data Lake Storage data like you would a database - read, write, and update Azure Data Lake Storage ADLSData, etc. through a standard ODBC Driver interface.

Establish a Live Connection with Azure Data Lake Storage Data using Tableau Bridge



The CData ODBC Driver for Azure Data Lake Storage enables you to integrate live Azure Data Lake Storage data into Tableau Cloud dashboards using the Tableau Bridge.

The Tableau Bridge enables you to publish dashboards to Tableau Cloud while maintaining live connectivity with any data source. In this article, you will use the Tableau Bridge to maintain data freshness in a published workbook by listening for changes in the underlying Azure Data Lake Storage data.

The CData ODBC drivers offer unmatched performance for interacting with live Azure Data Lake Storage data in Tableau Cloud due to optimized data processing built into the driver. When you issue complex SQL queries from Tableau Cloud to Azure Data Lake Storage, the driver pushes supported SQL operations, like filters and aggregations, directly to Azure Data Lake Storage and utilizes the embedded SQL engine to process unsupported operations (often SQL functions and JOIN operations) client-side. With built-in dynamic metadata querying, you can visualize and analyze Azure Data Lake Storage data using native Tableau data types.

Connect to Azure Data Lake Storage as an ODBC Data Source

If you have not already, first specify connection properties in an ODBC DSN (data source name). This is the last step of the driver installation. You can use the Microsoft ODBC Data Source Administrator to create and configure ODBC DSNs.

Authenticating to a Gen 1 DataLakeStore Account

Gen 1 uses OAuth 2.0 in Azure AD for authentication.

For this, an Active Directory web application is required. You can create one as follows:

  1. Sign in to your Azure Account through the .
  2. Select "Azure Active Directory".
  3. Select "App registrations".
  4. Select "New application registration".
  5. Provide a name and URL for the application. Select Web app for the type of application you want to create.
  6. Select "Required permissions" and change the required permissions for this app. At a minimum, "Azure Data Lake" and "Windows Azure Service Management API" are required.
  7. Select "Key" and generate a new key. Add a description, a duration, and take note of the generated key. You won't be able to see it again.

To authenticate against a Gen 1 DataLakeStore account, the following properties are required:

  • Schema: Set this to ADLSGen1.
  • Account: Set this to the name of the account.
  • OAuthClientId: Set this to the application Id of the app you created.
  • OAuthClientSecret: Set this to the key generated for the app you created.
  • TenantId: Set this to the tenant Id. See the property for more information on how to acquire this.
  • Directory: Set this to the path which will be used to store the replicated file. If not specified, the root directory will be used.

Authenticating to a Gen 2 DataLakeStore Account

To authenticate against a Gen 2 DataLakeStore account, the following properties are required:

  • Schema: Set this to ADLSGen2.
  • Account: Set this to the name of the account.
  • FileSystem: Set this to the file system which will be used for this account.
  • AccessKey: Set this to the access key which will be used to authenticate the calls to the API. See the property for more information on how to acquire this.
  • Directory: Set this to the path which will be used to store the replicated file. If not specified, the root directory will be used.

When you configure the DSN, you may also want to set the Max Rows connection property. This will limit the number of rows returned, which is especially helpful for improving performance when designing reports and visualizations.

Add Azure Data Lake Storage Data to a Dashboard

  1. From a new workbook, click Data -> New Data Source -> Other Databases (ODBC).
    Select the CData Data Source Name (for example: CData ADLS Source).
  2. In the Database menu, select CData.
  3. In the Table box, enter a table name or click New Custom SQL to enter an SQL query. This article retrieves the Resources table.
  4. Drag the table onto the join area. At this point, you can include multiple tables, leveraging the built-in SQL engine to process complex data requests.
  5. Click the tab for your worksheet. Columns are listed as Dimensions and Measures, which you can drag and drop onto the dashboard to create visualizations.

Set Up Tableau Bridge as a Service

  1. In the Server menu, select Start Tableau Bridge Client.
  2. Sign in to the Tableau Bridge using a site admin level account.
  3. If prompted, select the Tableau Cloud site where you want to publish live data. The bridge client will open and is accessible from the system tray.
  4. By default, the Tableau Bridge client is set to Application mode. Select 'Switch to service' to enable Tableau Bridge to handle live connections.
  5. Log in to your Tableau Cloud site as an administrator.
  6. From your site, click Settings, then Bridge.
  7. In the Bridge settings, under Enable Clients to Maintain Live Connections, check the box labeled 'Enable Tableau Bridge clients to maintain live connections to on-premises data.'

Publish a Dashboard Containing the Live Data Source

Having configured both the Tableau Bridge and Tableau Cloud to enable live data connections, you can now publish your workbook to Tableau Cloud. From the Server menu, select Publish Workbook.

After choosing the workbook name and project that you wish to publish to, configure the deployment so that the CData ODBC driver for Azure Data Lake Storage is embedded in your workbook as a separate, live data source.

  1. Under Data Sources, select the option to Edit the embedded data sources in the workbook.
  2. Change Publish Type to 'Published separately,' then select a desired means of authentication.
  3. Last, select 'Maintain connection to a live data source' and click the green Publish Workbook button.

The published workbook now updates alongside the underlying Azure Data Lake Storage data. From a published dashboard, simply click the Refresh button to reflect the most recent changes.