Query federation for Amazon Redshift in Databricks SQL (Experimental)

Experimental

The configurations described in this article are Experimental. Experimental features are provided as-is and are not supported by Databricks through customer technical support. To get full query federation support, you should instead use Lakehouse Federation, which enables your Databricks users to take advantage of Unity Catalog syntax and data governance tools.

This article describes how to configure read-only query federation to Redshift on serverless and pro SQL warehouses. For information about configuring Redshift S3 credentials, see Query Amazon Redshift using Databricks.

You configure connections to Redshift at the table level. You can use secrets to store and access text credentials without displaying them in plaintext. See the following example:

DROP TABLE IF EXISTS redshift_table;
CREATE TABLE redshift_table
USING redshift
OPTIONS (
  dbtable '<table-name>',
  tempdir 's3a://<bucket>/<directory-path>',
  url 'jdbc:redshift://<database-host-url>',
  user secret('redshift_creds', 'my_username'),
  password secret('redshift_creds', 'my_password'),
  forward_spark_s3_credentials 'true'
);