
Atlas
PRODRequirements
1. Create Database & Messaging Services
You need to create at least a Database Service before ingesting the metadata from Atlas. Make sure to note down the name, since we will use it to create Atlas Service. For example, to create a Hive Service you can follow these steps:1
Visit the Services Page
Click `Settings` in the side navigation bar and then `Services`. The first step is to ingest the metadata from your sources. To do that, you first need to create a Service connection first. This Service will be the bridge between OpenMetadata and your source system. Once a Service is created, it can be used to configure your ingestion workflows.

2
Create a New Service
Click on _Add New Service_ to start the Service creation.

3
Select the Service Type
Select Atlas as the Service type and click _Next_.

4
Name and Describe your Service
Provide a name and description for your Service.
Service Name
OpenMetadata uniquely identifies Services by their **Service Name**. Provide a name that distinguishes your deployment from other Services, including the other Atlas Services that you might be ingesting metadata from. Note that when the name is set, it cannot be changed.
5
Configure the Service Connection
In this step, we will configure the connection settings required for Atlas. Please follow the instructions below to properly configure the Service to read from your sources. You will also find helper documentation on the right-hand side panel in the UI.

2. Atlas Metadata Ingestion
Then, prepare the Atlas Service and configure the Ingestion:1
Visit the Services Page
Click `Settings` in the side navigation bar and then `Services`. The first step is to ingest the metadata from your sources. To do that, you first need to create a Service connection first. This Service will be the bridge between OpenMetadata and your source system. Once a Service is created, it can be used to configure your ingestion workflows.

2
Create a New Service
Click on _Add New Service_ to start the Service creation.

3
Select the Service Type
Select Atlas as the Service type and click _Next_.

4
Name and Describe your Service
Provide a name and description for your Service.
Service Name
OpenMetadata uniquely identifies Services by their **Service Name**. Provide a name that distinguishes your deployment from other Services, including the other Atlas Services that you might be ingesting metadata from. Note that when the name is set, it cannot be changed.
5
Configure the Service Connection
In this step, we will configure the connection settings required for Atlas. Please follow the instructions below to properly configure the Service to read from your sources. You will also find helper documentation on the right-hand side panel in the UI.

Connection Details
1
Connection Details
- Host and Port: Host and port of the Atlas service.
- Username: username to connect to the Atlas. This user should have privileges to read all the metadata in Atlas.
- Password: password to connect to the Atlas.
- databaseServiceName: source database of the data source. This is the service we created before: e.g.,
local_hive) - messagingServiceName: messaging service source of the data source.
- Entity Type: Name of the entity type in Atlas.
2
Test the Connection
Once the credentials have been added, click on Test Connection and Save the changes.

3
Configure Metadata Ingestion
In this step we will configure the metadata ingestion pipeline,
Please follow the instructions below

4
Schedule the Ingestion and Deploy
Scheduling can be set up at an hourly, daily, weekly, or manual cadence. The
timezone is in UTC. Select a Start Date to schedule for ingestion. It is
optional to add an End Date.Review your configuration settings. If they match what you intended,
click Deploy to create the service and schedule metadata ingestion.If something doesn’t look right, click the Back button to return to the
appropriate step and change the settings as needed.After configuring the workflow, you can click on Deploy to create the
pipeline.

5
View the Ingestion Pipeline
Once the workflow has been successfully deployed, you can view the
Ingestion Pipeline running from the Service Page.
