Subscribe

UiPath AI Center

UiPath AI Center

AI Center Single node

🚧

Before proceeding, make sure you meet the requirements detailed here

At a high level, the installation of AI Center needs to execute these steps:

Step

Active Time

Waiting Time

  1. Provision AIF Machine

< 5 min

--

  1. Configure Database

< 1 min

--

  1. Configure Orchestrator

< 5 min

--

  1. Run AI Center Infrastructure Installer

< 1 min

~20 mins

  1. Run AI Center application Installer

< 5 min

~20 mins

  1. Verify Installation

~5 min

--

Network Configuration

  • The Linux machine where AI Center will be installed needs to be able to connect to the Orchestrator machine (domain and port).
  • The Linux machine where AI Center will be installed needs to be able to connect to the SQL Server (domain/IP and port).
  • Robots/Studio that will make use of AI Center need connectivity to the AI Center Linux machine.

For peripheral Document Understanding components (Data Manager and OCR Engines):

  • Data Manager needs access to AI Center on prem :<port_number> or to public SaaS endpoints like https://invoices.uipath.com in case Prelabelling is needed (prelabeling is optional).
  • Data Manager needs access to OCR engine :<port_number>. OCR engine might be UiPath Document OCR on premises, Omnipage OCR on premises, Google Cloud Vision OCR, Microsoft Read Azure, Microsoft Read on premises.
  • Robots need access to OCR :<port_number>. Same OCR options as above, except for Omnipage, which is available in the Robots directly as an Activity Pack.

Connectivity Requirements - Online Install

The AI Center Online install refers to an on-premises installation that downloads AI Center application and all related artifacts (e.g. machine learning models) from the internet.

Endpoints the Installer Connects

The AI Center installer downloads container images and machine learning models to populate your AI Center instance with ready-to-use machine learning (this includes Document Understanding models). For this reason, at installation time, the Linux machine needs access to these endpoints over https (port 443):

Host Name

Purpose

registry.replicated.com

Upstream Docker images are pulled via registry.replicated.com. This domain is owned by Replicated, Inc which is headquartered in Los Angeles, CA

proxy.replicated.com

Upstream Docker images are pulled via proxy.replicated.com. The on-prem docker client uses a license ID to authenticate to proxy.replicated.com. This domain is owned by Replicated, Inc which is headquartered in Los Angeles, CA

replicated.app

Upstream application YAML and metadata is pulled from replicated.app. The current running version of the application (if any) will be sent, in addition to a license ID and an application IDs are sent to replicated.app to authenticate and receive these YAML files. This domain is owned by Replicated, Inc which is headquartered in Los Angeles, CA

get.replicated.com

Sync artifacts from replicated. This domain is owned by Replicated, Inc which is headquartered in Los Angeles, CA

api.replicated.com

API requests to infrastructure installer. This domain is owned by Replicated, Inc which is headquartered in Los Angeles, CA

k8s.kurl.sh

Kubernetes cluster installation scripts and artifacts are served from kurl.sh. An application identifier is sent in a URL path, and bash scripts and binary executables are served from kurl.sh. This domain is owned by Replicated, Inc which is headquartered in Los Angeles, CA

kurl-sh.s3.amazonaws.com

tar.gz packages are downloaded from Amazon S3 during embedded cluster installations. The IP ranges to whitelist for accessing these can be scraped dynamically from the AWS IP Address Ranges1 documentation.

*.docker.io

Upstream Docker images are pulled from docker.io. There may be multiple subdomains like registry-1.docker.io thus the pattern to the right should be allowed.

*.docker.com

Other upstream Docker images are pulled from docker.com. There may be multiple subdomains thus the pattern to the right should be allowed.

raw.githubusercontent.com

For scripts to create persistent volume claim deployment.

Endpoints Connected To at Runtime

At runtime, an AI Center that was installed via the online installer connects to these endpoints:

Host Name

Purpose

du-metering.uipath.com

To account for, and validate Document Understanding licenses.

registry.replicated.com

Upstream Docker images are pulled via a private docker registry from registry.replicated.com. This domain is owned by Replicated, Inc which is headquartered in Los Angeles, CA.

raw.githubusercontent.com

For scripts to update OOB models daily.

custom (optional)

This depends on how the persona using AI Center has chosen to build their ml packages. AI Center dynamically builds a container image. The dependencies for that image can be bundled within the ML Package itself (in which case no extra outbound network calls are done), or can be specified in a requirements.txt file. This file can specify the location from which dependencies will be downloaded.

Connectivity Requirements - Airgapped Install

The AI Center airgapped install refers to an on-premises installation triggered after a one-time download from a uipath domain.

No internet connection is required at installation time (note: if the node has a GPU, this assumes that NVIDA driver version 450.51.06 and nvidia-container-runtime have been installed as detailed in the prerequisite for an airgapped install).

At application tuntime, whether connectivity is require or not is completely up to the user of AI Center . An AI Center user creates a ml packages that can be deployed and trained on AI Center . AI Center dynamically builds a container image from that ML Package. The dependencies for that image can be bundled within the ML Package itself (in which case no extra outbound network calls are done), or can be specified in a requirements.txt file. This file can specify the location from which dependencies will be downloaded such as an internal, secure python package dependency repository.

Updated 23 days ago


AI Center Single node


Suggested Edits are limited on API Reference Pages

You can only suggest edits to Markdown body content, but not to the API spec.