Data Connections and Integration
Complete guide for connecting Atlan to your data sources, managing integrations, and troubleshooting connection issues.
How can I set up on-prem databases if I don't use Docker Compose?
To use Atlan's metadata-extractor tool to extract metadata from on-premises databases with Kubernetes deployment architecture, refer to the How to connect on-premises databases to Kubernetes documentation.
Does Atlan integrate with MongoDB?
Atlan currently supports native integration with MongoDB and Microsoft Azure Cosmos DB for MongoDB deployments.
Can Atlan work with the IBM Informix database?
Atlan doesn't currently offer native support for IBM Informix database. Atlan is built on an Open API architecture, so you can catalog your IBM Informix assets programmatically. Refer to the developer documentation to publish database, schema, table, and column objects.
Is Atlan compatible with data quality tools?
Yes, Atlan integrates with several data quality and observability tools:
- Native integrations with tools like Great Expectations, Monte Carlo, and other data quality platforms
- You can view data quality metrics and alerts directly within asset profiles
- Custom integrations can be built using Atlan APIs to connect additional data quality tools
For the most current list of supported integrations, check the connections documentation.
Does Atlan integrate with Talend or Matillion?
Atlan currently only supports native integration with Matillion.
If Talend uses the ELT (Extract, Load, Transform) paradigm, this means that the SQL transformation queries are pushed down into the warehouse. In that case, Atlan automatically generates the lineage for these transformations by parsing the SQL statements extracted from the query history of the data warehouse. Lineage for the extract and load steps isn't automated for Talend but is supported for Fivetran.
However, if the transformation is taking place outside the warehouse, Atlan needs access to the transformation queries to generate the lineage. This access can either be gained via a Git repository if utilized or by sharing the queries or stored procedures in a shared location, such as an S3 bucket.
Why does Atlan require the site administrator explorer role in Tableau?
To learn why Atlan requires the Site Administrator Explorer role in Tableau, refer to this guide.
Why is the view button for Tableau unavailable?
The view button for Tableau assets may be unavailable due to:
- Insufficient permissions to access the Tableau content
- The Tableau server being inaccessible from your current network
- Authentication issues between Atlan and Tableau
- The specific asset being moved or deleted in Tableau
Check your Tableau permissions and network connectivity, or contact your Tableau administrator.
Who is a source owner in Microsoft Power BI?
Microsoft Power BI provides metadata for who generated or configured reports, datasets, and dataflows, which is then mapped to the source owner field. This has no connection to Microsoft Power BI workspace admins. You can refer to the API response schema to learn more:
createdBy
in WorkspaceInfoReportconfiguredBy
in WorkspaceInfoDatasetconfiguredBy
in WorkspaceInfoDataflow
The API doesn't return source owner metadata for other asset types due to limitations at source, see Microsoft Power BI documentation.
The value of the source owner is a string. While for sources like Looker, Atlan displays the username based on the metadata received, Microsoft Power BI APIs only return the user's email address, which is what Atlan displays for supported assets.
In certain cases, assets that are connected to the same Microsoft Power BI workspace and have the same source owner may display a different source owner or none at all. This is because Atlan maps it to the metadata returned by the APIs. For example, if the response is null, Atlan won't display a source owner.
How does Atlan work with dbt single tenant vs multi-tenant?
There are multiple deployment options available for dbt Cloud (see dbt documentation here and here).
When integrating with dbt Cloud, Atlan uses the following APIs to fetch metadata from all deployment options:
- dbt Cloud Administrative API - to fetch account, project, environment, job, and run metadata.
- dbt Cloud Discovery API - to fetch models, sources, and tests from each dbt environment.
Are there any dbt assets that can't be viewed in dbt?
Atlan displays the View in dbt link for newly created or synced dbt assets only — including models, sources, and tests. Atlan doesn't display this link for assets without target URLs.
How can I reuse my documentation from dbt?
Atlan's dbt connectivity provides two ways to ingest documentation from dbt:
- Automatically load details like descriptions, from your existing dbt project details.
- Any other (Atlan-specific) details you want to document, through dbt's meta field.
What happens to asset metadata in Atlan if I switch to a new server?
If the assets in your new server retain the same qualifiedName
, then these are recreated in Atlan with all the metadata attached. The qualifiedName
determines asset uniqueness in Atlan (and influences the GUID of the asset). Following this logic, any asset recreated with the same name remains the same asset in Atlan - considering that the qualifiedName
has remained the same.
By contrast, if an asset is recreated with a new name, it becomes a new asset in Atlan - considering that the qualifiedName
has changed. This means that the attached metadata is no longer available.
If an asset is removed from the data source or no longer available, it'll be archived in Atlan as part of the crawler's cleanup policy.
Why is the metadata getting lost when migrating from Snowflake to dbt?
Metadata loss during Snowflake to dbt migration typically occurs because:
- The
qualifiedName
changes between the source Snowflake table and the dbt model - Asset lineage connections aren't properly maintained during migration
- Custom metadata and tags aren't transferred in the migration process
To preserve metadata:
- Make sure to use consistent naming conventions between Snowflake and dbt
- Use dbt's meta fields to preserve custom metadata
- Re-run Atlan crawlers after migration to rebuild lineage connections
What happens when an asset is removed from Redshift?
If an asset such as a table or a schema is removed from Redshift or any other source, it'll also be automatically removed from Atlan during the next workflow run. Such assets are archived in Atlan (soft-deleted), so that they don't appear in search results by default. You can only access archived assets through discovery.
What data is Atlan actually bringing in?
Atlan extracts metadata only, not the actual data:
- Schema information: Table structures, column names, data types
- Usage statistics: Query patterns, user access logs (when available)
- Lineage data: Data flow and transformation logic
- Custom metadata: Tags, descriptions, business glossary terms
- Data profiles: example data previews (configurable, with privacy controls)
Atlan never stores your actual business data - it only catalogues metadata to help you understand and govern your data landscape.
Can offline extraction fail if there are spaces in the path?
Atlan currently doesn't support spaces in folder names for S3. The offline extraction workflow fails if you include any spaces in the folder name in S3. To follow documented guidelines for safe characters, refer to Amazon S3 documentation.
What does API only mean?
In the Where in Atlan column of What does Atlan crawl from (connector name)? documentation, API only indicates source properties that have been crawled in Atlan but aren't published or discoverable on the product UI.
Can you integrate with Jupyter notebook?
Yes, Atlan supports integration with Jupyter notebooks through:
- Python SDK for programmatic access to Atlan APIs
- Jupyter extensions for data discovery and cataloguing
- Ability to document and version notebook-based analyses
- Integration with notebook-based data science workflows
Check the developer documentation for specific integration examples and code samples.
Can I integrate Atlan with any web application?
Yes, Atlan provides comprehensive APIs and SDKs that enable integration with virtually any web application:
- REST APIs for all Atlan functionality
- Python, Java, and JavaScript SDKs
- Webhook support for real-time notifications
- OAuth2 and API key authentication methods
Visit the developer portal for integration guides, API documentation, and code examples.
What does the Snowflake workflow follow when gathering information?
The Snowflake workflow or any other connector workflow has built-in differential crawling capabilities, which means it will crawl updates that have been made across assets in the Snowflake system and sync them to Atlan. This way, any incremental metadata changes happening in Snowflake are made available in Atlan after each workflow run.
What's a Snowflake process?
Snowflake process (in Atlan) = Snowflake transformation
Data processing in a warehouse involves a combination of these three processes:
- Extract (from source).
- Load (to warehouse).
- Transformation (of the data between source format and a consolidated format in the warehouse).
Whether run through a Python script, directly from Snowflake's UI, or via third-party programs like Matillion, these transformations occur within Snowflake through SQL. Regardless of what ran this transformation SQL, a Snowflake process in Atlan captures the transformation logic for use in lineage.