Intercom Extract
    • Dark
      Light

    Intercom Extract

    • Dark
      Light

    Article Summary

    This article is specific to the following platforms - Snowflake - Redshift - BigQuery.

    Intercom Extract

    The Intercom Extract component calls the Intercom API to retrieve and store data to be either referenced by an external table or loaded into a table, depending on the user's cloud data warehouse. Users can then transform their data with the Matillion ETL library of transformation components.

    Using this component may return structured data that requires flattening. For help with flattening such data, we recommend using the Nested Data Load Component for Amazon Redshift and the Extract Nested Data Component for Snowflake or Google BigQuery.

    Properties

    Snowflake Properties

    PropertySettingDescription
    NameStringA human-readable name for the component.
    Data SourceSelectSelect a data source. Once configured, one or more properties specific to that data source will become available; these properties are mandatory and must be configured.
    OAuthSelectThe name of the OAuth entry that has been configured for this service. For help creating and authorising an OAuth entry, please refer to the Intercom Extract Authentication Guide.
    Tag IDStringThe unique identifier for the tag (available when selecting "Companies" as data source).
    Segment IDStringThe unique identifier for the segment (available when selecting "Companies" as data source).
    User IDStringThe unique identifier for the user (available when selecting "Data Events" as data source).
    EmailStringThe email address associated with the Intercom account (available when selecting "Data Events" as data source).
    Intercom User IDStringThe unique identifier of the Intercom user associated with the Data Events (available when selecting "Data Events" as data source).
    Page LimitIntegerSpecify the maximum number of pages to stage.
    LocationStorage LocationProvide an S3 bucket path (AWS only), GCS bucket path (GCP only), or Azure Blob Storage path (Azure only) that will be used to store the data. Once on an S3 bucket, GCS bucket or Azure Blob, the data can be referenced by an external table. A folder will be created at this location with the same name as the Target Table.
    IntegrationSelect(GCP only) Choose your Google Cloud Storage Integration. Integrations are required to permit Snowflake to read data from and write to a Google Cloud Storage bucket. Integrations must be set up in advance of selecting them in Matillion ETL. To learn more about setting up a storage integration, read our Storage Integration Setup Guide.
    WarehouseSelectChoose a Snowflake warehouse that will run the load.
    DatabaseSelectSelect a database. A database is a logical grouping of schemas. Each database belongs to a single Snowflake account.
    SchemaSelectSelect the schema. A schema is a logical grouping of database "objects" (tables, views, etc.). Each schema belongs to a single database. The special value, [Environment Default], will use the schema defined in the environment. For more information on using multiple schemas, see this article.
    Target TableStringSpecify the external table to be used.
    Warning: This table will be recreated and any existing table of the same name will be dropped.

    Redshift Properties

    PropertySettingDescription
    NameStringA human-readable name for the component.
    Data SourceSelectSelect a data source. Once configured, one or more properties specific to that data source will become available; these properties are mandatory and must be configured.
    OAuthSelectThe name of the OAuth entry that has been configured for this service. For help creating and authorising an OAuth entry, please refer to the Intercom Extract Authentication Guide.
    Tag IDStringThe unique identifier for the tag (available when selecting "Companies" as data source).
    Segment IDStringThe unique identifier for the segment (available when selecting "Companies" as data source).
    User IDStringThe unique identifier for the user (available when selecting "Data Events" as data source).
    EmailStringThe email address associated with the Intercom account (available when selecting "Data Events" as data source).
    Intercom User IDStringThe unique identifier of the Intercom user associated with the Data Events (available when selecting "Data Events" as data source).
    Page LimitIntegerSpecify the maximum number of pages to stage.
    LocationStorage LocationSpecify the S3 bucket. Users can click through the tree structure to locate the preferred S3 bucket, or specify the URL of the S3 bucket in the URL field, following the template: s3://<bucket>/<path>
    TypeDropdownSelect between a standard table and an external table.
    Standard SchemaDropdownSelect the Redshift schema. The special value, [Environment Default], will use the schema defined in the Matillion ETL environment.
    External SchemaSelectSelect the table's external schema. To learn more about external schema, please consult the Configuring The Matillion ETL Client section of the Getting Started With Amazon Redshift Spectrum documentation.
    For more information on using multiple schema, see Schema Support.
    Target TableStringSpecify the external table to be used.
    Warning: This table will be recreated and any existing table of the same name will be dropped.

    BigQuery Properties

    PropertySettingDescription
    NameStringA human-readable name for the component.
    Data SourceSelectSelect a data source. Once configured, one or more properties specific to that data source will become available; these properties are mandatory and must be configured.
    OAuthSelectThe name of the OAuth entry that has been configured for this service. For help creating and authorising an OAuth entry, please refer to the Intercom Extract Authentication Guide.
    Tag IDStringThe unique identifier for the tag (available when selecting "Companies" as data source).
    Segment IDStringThe unique identifier for the segment (available when selecting "Companies" as data source).
    User IDStringThe unique identifier for the user (available when selecting "Data Events" as data source).
    EmailStringThe email address associated with the Intercom account (available when selecting "Data Events" as data source).
    Intercom User IDStringThe unique identifier of the Intercom user associated with the Data Events (available when selecting "Data Events" as data source).
    Page LimitIntegerSpecify the maximum number of pages to stage.
    Table TypeSelectSelect whether the table is Native (by default in BigQuery) or an external table.
    ProjectSelectSelect the Google Bigquery project. The special value, [Environment Default], will use the project defined in the environment.
    For more information, refer to the BigQuery documentation.
    DatasetSelectSelect the Google Bigquery dataset to load data into. The special value, [Environment Default], will use the dataset defined in the environment.
    For more information, refer to the BigQuery documentation.
    Target TableStringA name for the table.
    Warning: This table will be recreated and will drop any existing table of the same name.
    Only available when the table type is Native.
    New Target TableStringA name for the new external table.
    Only available when the table type is External.
    Cloud Storage Staging AreaCloud Storage BucketSpecify the target Google Cloud Storage bucket to be used for staging the queried data. Users can either:
    1. Input the URL string of the Cloud Storage bucket following the template provided: gs://<bucket>/<path>
    2. Navigate through the file structure to select the target bucket.

    Only available when the table type is Native.
    LocationCloud Storage BucketSpecify the target Google Cloud Storage bucket to be used for staging the queried data. Users can either:
    1. Input the URL string of the Cloud Storage bucket following the template provided: gs://<bucket>/<path>
    2. Navigate through the file structure to select the target bucket.
    Only available when the table type is Native.
    Load OptionsMultiple SelectClean Cloud Storage Files: Destroy staged files on Cloud Storage after loading data. Default is On.
    Cloud Storage File Prefix: Give staged file names a prefix of your choice. The default setting is an empty field.
    Recreate Target Table: Choose whether the component recreates its target table before the data load. If Off, the component will use an existing table or create one if it does not exist. Default is On.
    Use Grid Variable: Check this checkbox to use a grid variable. This box is unchecked by default.