Skip to main content

Ingest data from NetSuite

Preview

This feature is in Public Preview.

This page shows how to create a managed NetSuite ingestion pipeline using Lakeflow Connect.

Requirements

  • To create an ingestion pipeline, you must first meet the following requirements:

    • Your workspace must be enabled for Unity Catalog.

    • Serverless compute must be enabled for your workspace. See Serverless compute requirements.

    • If you plan to create a new connection: You must have CREATE CONNECTION privileges on the metastore.

      If the connector supports UI-based pipeline authoring, an admin can create the connection and the pipeline at the same time by completing the steps on this page. However, if the users who create pipelines use API-based pipeline authoring or are non-admin users, an admin must first create the connection in Catalog Explorer. See Connect to managed ingestion sources.

    • If you plan to use an existing connection: You must have USE CONNECTION privileges or ALL PRIVILEGES on the connection object.

    • You must have USE CATALOG privileges on the target catalog.

    • You must have USE SCHEMA and CREATE TABLE privileges on an existing schema or CREATE SCHEMA privileges on the target catalog.

  • To ingest from NetSuite, you must first complete the steps in Configure NetSuite for ingestion into Databricks.

Download the SuiteAnalytics Connect JDBC driver

The NetSuite connector requires that you upload your SuiteAnalytics Connect JDBC driver JAR file to a volume in Databricks. The following driver versions are supported: 8.10.147.0, 8.10.170.0, and 8.10.184.0.

  1. Log in to your NetSuite account.
  2. In the Settings portlet, click Set up Analytics Connect.
  3. Download the JAR file to your local machine.

Create a volume

Create a volume to store the JAR file.

See Create a volume.

Users who create NetSuite ingestion pipelines must have access to this volume.

Upload the JAR file to the volume

  1. In the Catalog pane, navigate to the volume you created.
  2. Click Upload to volume.
  3. Select the JAR file you downloaded and click Upload.

Create an ingestion pipeline

Each source table is ingested into a streaming table.

  1. In the sidebar of the Databricks workspace, click Data Ingestion.
  2. On the Add data page, under Databricks connectors, click NetSuite.
  3. On the Connection page of the ingestion wizard, select the connection that stores your NetSuite access credentials. If you have the CREATE CONNECTION privilege on the metastore, you can click Plus icon. Create connection to create a new connection with the authentication details in Configure NetSuite for ingestion into Databricks.
  4. Click Next.
  5. On the Ingestion setup page, enter a unique name for the pipeline.
  6. Select a catalog and a schema to write event logs to. If you have USE CATALOG and CREATE SCHEMA privileges on the catalog, you can click Plus icon. Create schema in the drop-down menu to create a new schema.
  7. Click Create pipeline and continue.
  8. On the Source page, select the tables to ingest.
  9. Click Save and continue.
  10. On the Destination page, select a catalog and a schema to load data into. If you have USE CATALOG and CREATE SCHEMA privileges on the catalog, you can click Plus icon. Create schema in the drop-down menu to create a new schema.
  11. Click Save and continue.
  12. (Optional) On the Schedules and notifications page, click Plus icon. Create schedule. Set the frequency to refresh the destination tables.
  13. (Optional) Click Plus icon. Add notification to set email notifications for pipeline operation success or failure, then click Save and run pipeline.

Examples

Use these examples to configure your pipeline.

Ingest a single source table

The following pipeline definition file ingests a single source table:

YAML
variables:
dest_catalog:
default: main
dest_schema:
default: ingest_destination_schema

# The main pipeline for netsuite_dab
resources:
pipelines:
pipeline_netsuite:
name: netsuite_pipeline
catalog: ${var.dest_catalog}
schema: ${var.dest_schema}
ingestion_definition:
connection_name: <netsuite-connection>
objects:
# An array of objects to ingest from NetSuite. This example ingests the transaction table.
- table:
source_schema: default
source_table: transaction
destination_catalog: ${var.dest_catalog}
destination_schema: ${var.dest_schema}
table_configuration:
netsuite_jar_path: /Volumes/<catalog>/<schema>/<volume>/NQjc.jar

Ingest multiple source tables

The following pipeline definition file ingests multiple source tables:

YAML
variables:
dest_catalog:
default: main
dest_schema:
default: ingest_destination_schema

# The main pipeline for netsuite_dab
resources:
pipelines:
pipeline_netsuite:
name: netsuite_pipeline
catalog: ${var.dest_catalog}
schema: ${var.dest_schema}
ingestion_definition:
connection_name: <netsuite-connection>
objects:
# An array of objects to ingest from NetSuite. This example ingests the transaction and account tables.
- table:
source_schema: default
source_table: transaction
destination_catalog: ${var.dest_catalog}
destination_schema: ${var.dest_schema}
table_configuration:
netsuite_jar_path: /Volumes/<catalog>/<schema>/<volume>/NQjc.jar
- table:
source_schema: default
source_table: account
destination_catalog: ${var.dest_catalog}
destination_schema: ${var.dest_schema}
table_configuration:
netsuite_jar_path: /Volumes/<catalog>/<schema>/<volume>/NQjc.jar

Bundle job definition file

The following is an example job definition file to use with Declarative Automation Bundles. The job runs every day, exactly one day from the last run.

YAML
resources:
jobs:
netsuite_dab_job:
name: netsuite_dab_job

trigger:
periodic:
interval: 1
unit: DAYS

email_notifications:
on_failure:
- <email-address>

tasks:
- task_key: refresh_pipeline
pipeline_task:
pipeline_id: ${resources.pipelines.pipeline_netsuite.id}

Common patterns

For advanced pipeline configurations, see Common patterns for managed ingestion pipelines.

Next steps

Start, schedule, and set alerts on your pipeline. See Common pipeline maintenance tasks.

Additional resources