Skip To Content

Use Snowflake records

Use records from a Snowflake table as input to ArcGIS Data Pipelines.

Usage notes

Keep the following in mind when working with Snowflake:

  • To use a dataset from Snowflake, you must first create a data store item. Data store items securely store credentials and connection information so the data can be read by Data Pipelines. To create a data store, follow the steps in the Connect to Snowflake section below.
  • To change the data store item you configured, use the Data store item parameter to remove the currently selected item, and choose one of the following options:
    • Add data store—Create a new data store item.
    • Select item—Browse your content to select an existing data store item.
  • Use the Table parameter to specify the dataset you want to use. Table names are case sensitive. For example, if the table is named MyTable, you must specify it using the exact capitalization.
  • To improve the performance of reading input datasets, consider the following options:
    • Use the Use caching parameter to store a copy of the dataset. The cached copy is only maintained while at least one browser tab open to the editor is connected. This may make it faster to access the data during processing. If the source data has been updated since it was cached, uncheck this parameter and preview or run the tool again.
    • After configuring an input dataset, configure any of the following tools that limit the amount of data being processed:

Connect to Snowflake

To use data stored in Snowflake, complete the following steps to create a data store item in the Data Pipelines editor:

  1. On the Data Pipelines editor toolbar, click Inputs and choose Snowflake.

    The Select a data store connection dialog box appears.

  2. Choose Add a new data store.
  3. Click Next.

    The Add a connection to a data store dialog box appears.

  4. Provide the server URL to the Snowflake source. Note that validation may fail if you specify https:// in the server URL.
    The following is an example: my_account.snowflakecomputing.com.
  5. Provide the username and password for the Snowflake account.
  6. Provide the name of the database that contains the datasets to connect to.
  7. Provide the name of the schema used for the database.
  8. Provide the name of the warehouse that will provide the compute resources for the connection.

    The role must have at least USAGE and CREATE STAGE privileges on the schema that contains the table that you will read from.

  9. Provide the name of the role that defines the set of permissions for the connection.
  10. Click Next.

    The item details pane appears.

  11. Provide a title for the new data store item.

    This title will appear in your portal content. You can also store the item in a specific folder, and provide item tags or a summary.

  12. Click Save to create the data store item.

    A Snowflake element that you can configure for a certain dataset is added to the canvas.

Limitations

The following are known limitations:

  • User authentication is the only authentication type supported for connecting to Snowflake in Data Pipelines.
  • To use a data store item to connect to external data sources, you must be the owner of the data store item. Data store items that are shared with you are not supported as input.

Licensing requirements

The following licensing and configurations are required:

  • Creator or Professional user type
  • Publisher, Facilitator, or Administrator role, or an equivalent custom role

To learn more about Data Pipelines requirements, see Requirements.