- Overview
- Requirements
- Installation
- Post-installation
- Cluster administration
- Managing products
- Managing the cluster in ArgoCD
- Setting up the external NFS server
- Automated: Enabling the Backup on the Cluster
- Automated: Disabling the Backup on the Cluster
- Automated, Online: Restoring the Cluster
- Automated, Offline: Restoring the Cluster
- Manual: Enabling the Backup on the Cluster
- Manual: Disabling the Backup on the Cluster
- Manual, Online: Restoring the Cluster
- Manual, Offline: Restoring the Cluster
- Additional configuration
- Migrating objectstore from persistent volume to raw disks
- Monitoring and alerting
- Migration and upgrade
- Migration options
- Step 1: Moving the Identity organization data from standalone to Automation Suite
- Step 2: Restoring the standalone product database
- Step 3: Backing up the platform database in Automation Suite
- Step 4: Merging organizations in Automation Suite
- Step 5: Updating the migrated product connection strings
- Step 6: Migrating standalone Insights
- Step 7: Deleting the default tenant
- B) Single tenant migration
- Product-specific configuration
- Best practices and maintenance
- Troubleshooting
- How to Troubleshoot Services During Installation
- How to Uninstall the Cluster
- How to clean up offline artifacts to improve disk space
- How to clear Redis data
- How to enable Istio logging
- How to manually clean up logs
- How to clean up old logs stored in the sf-logs bucket
- How to disable streaming logs for AI Center
- How to debug failed Automation Suite installations
- How to delete images from the old installer after upgrade
- How to automatically clean up Longhorn snapshots
- How to disable TX checksum offloading
- How to address weak ciphers in TLS 1.2
- Unable to run an offline installation on RHEL 8.4 OS
- Error in Downloading the Bundle
- Offline installation fails because of missing binary
- Certificate issue in offline installation
- First installation fails during Longhorn setup
- SQL connection string validation error
- Prerequisite check for selinux iscsid module fails
- Azure disk not marked as SSD
- Failure After Certificate Update
- Automation Suite not working after OS upgrade
- Automation Suite Requires Backlog_wait_time to Be Set 1
- Volume unable to mount due to not being ready for workloads
- RKE2 fails during installation and upgrade
- Failure to upload or download data in objectstore
- PVC resize does not heal Ceph
- Failure to Resize Objectstore PVC
- Rook Ceph or Looker pod stuck in Init state
- StatefulSet volume attachment error
- Failure to create persistent volumes
- Storage reclamation patch
- Backup failed due to TooManySnapshots error
- All Longhorn replicas are faulted
- Setting a timeout interval for the management portals
- Update the underlying directory connections
- Cannot Log in After Migration
- Kinit: Cannot Find KDC for Realm <AD Domain> While Getting Initial Credentials
- Kinit: Keytab Contains No Suitable Keys for *** While Getting Initial Credentials
- GSSAPI Operation Failed With Error: An Invalid Status Code Was Supplied (Client's Credentials Have Been Revoked).
- Alarm Received for Failed Kerberos-tgt-update Job
- SSPI Provider: Server Not Found in Kerberos Database
- Login Failed for User <ADDOMAIN><aduser>. Reason: The Account Is Disabled.
- ArgoCD login failed
- Failure to get the sandbox image
- Pods not showing in ArgoCD UI
- Redis Probe Failure
- RKE2 Server Fails to Start
- Secret Not Found in UiPath Namespace
- After the Initial Install, ArgoCD App Went Into Progressing State
- MongoDB pods in CrashLoopBackOff or pending PVC provisioning after deletion
- Unexpected Inconsistency; Run Fsck Manually
- Degraded MongoDB or Business Applications After Cluster Restore
- Missing Self-heal-operator and Sf-k8-utils Repo
- Unhealthy Services After Cluster Restore or Rollback
- RabbitMQ pod stuck in CrashLoopBackOff
- Prometheus in CrashloopBackoff state with out-of-memory (OOM) error
- Missing Ceph-rook metrics from monitoring dashboards
- Pods cannot communicate with FQDN in a proxy environment
- Using the Automation Suite Diagnostics Tool
- Using the Automation Suite support bundle
- Exploring Logs

Automation Suite Installation Guide
Manual: Advanced Installation Experience
cluster_config.json
configuration file.
cluster_config.json
file defines the parameters, settings, and preferences applied to the UiPath products deployed via Automation Suite. You
need to update this file if you want to change defaults and use any advanced configuration for your cluster.
cluster_config.json
, you can use either:
- a Linux text editor, such as vi or GNU nano, directly on the Linux machine via SSH (e.g., command:
vi cluster_config.json
); - your favorite text editor and then copy paste the file on the machine.
cluster_config.json
file allows you to configure the UiPath products you want to deploy. There are two types of UiPath products:
- mandatory products: these are installed by default and do not have the
enabled
flag available on them; - optional products: these are not required to complete the install. However, be aware that products may have dependencies. Please see Product dependencies for more details on this.
cluster_config.json
file, use true
or false
for the enabled
flag.
{
"fqdn": "PLACEHOLDER",
"fixed_rke_address": "PLACEHOLDER",
"multinode": "false",
"admin_username": "PLACEHOLDER",
"admin_password": "PLACEHOLDER",
"profile": "ha",
"telemetry_optout": "true",
"rke_token": "PLACEHOLDER",
"server_certificate": {
"ca_cert_file": "/absolute/path/to/rootCA.crt",
"tls_cert_file": "/absolute/path/to/server.crt",
"tls_key_file": "/absolute/path/to/server.key"
},
"infra": {
"docker_registry": {
"username": "PLACEHOLDER",
"password": "PLACEHOLDER"
},
"custom_dns_resolver": "/path/to/custom-resolv.conf"
},
"identity_certificate": {
"token_signing_cert_file": "/absolute/path/to/identity.pfx",
"token_signing_cert_pass": ""
},
"sql": {
"server_url": "PLACEHOLDER",
"username": "PLACEHOLDER",
"password": "PLACEHOLDER",
"port": "PLACEHOLDER",
"create_db": "PLACEHOLDER"
},
"sql_connection_string_template": "PLACEHOLDER",
"sql_connection_string_template_jdbc": "PLACEHOLDER",
"sql_connection_string_template_odbc": "PLACEHOLDER",
"orchestrator": {
"testautomation": {
"enabled": true
},
"updateserver": {
"enabled": true
}
},
"aicenter": {
"enabled": true,
"sql_connection_str": "PLACEHOLDER"
},
"documentunderstanding": {
"enabled": true,
"datamanager": {
"sql_connection_str": "PLACEHOLDER"
},
"handwriting": {
"enabled": true,
"max_cpu_per_pod": 2
}
},
"insights": {
"enabled": true
},
"test_manager": {
"enabled": true
},
"automation_ops": {
"enabled": true
},
"automation_hub": {
"enabled": true
},
"apps": {
"enabled": true
},
"action_center": {
"enabled": true
},
"task_mining": {
"enabled": true
},
"dataservice": {
"enabled": true
}
}
{
"fqdn": "PLACEHOLDER",
"fixed_rke_address": "PLACEHOLDER",
"multinode": "false",
"admin_username": "PLACEHOLDER",
"admin_password": "PLACEHOLDER",
"profile": "ha",
"telemetry_optout": "true",
"rke_token": "PLACEHOLDER",
"server_certificate": {
"ca_cert_file": "/absolute/path/to/rootCA.crt",
"tls_cert_file": "/absolute/path/to/server.crt",
"tls_key_file": "/absolute/path/to/server.key"
},
"infra": {
"docker_registry": {
"username": "PLACEHOLDER",
"password": "PLACEHOLDER"
},
"custom_dns_resolver": "/path/to/custom-resolv.conf"
},
"identity_certificate": {
"token_signing_cert_file": "/absolute/path/to/identity.pfx",
"token_signing_cert_pass": ""
},
"sql": {
"server_url": "PLACEHOLDER",
"username": "PLACEHOLDER",
"password": "PLACEHOLDER",
"port": "PLACEHOLDER",
"create_db": "PLACEHOLDER"
},
"sql_connection_string_template": "PLACEHOLDER",
"sql_connection_string_template_jdbc": "PLACEHOLDER",
"sql_connection_string_template_odbc": "PLACEHOLDER",
"orchestrator": {
"testautomation": {
"enabled": true
},
"updateserver": {
"enabled": true
}
},
"aicenter": {
"enabled": true,
"sql_connection_str": "PLACEHOLDER"
},
"documentunderstanding": {
"enabled": true,
"datamanager": {
"sql_connection_str": "PLACEHOLDER"
},
"handwriting": {
"enabled": true,
"max_cpu_per_pod": 2
}
},
"insights": {
"enabled": true
},
"test_manager": {
"enabled": true
},
"automation_ops": {
"enabled": true
},
"automation_hub": {
"enabled": true
},
"apps": {
"enabled": true
},
"action_center": {
"enabled": true
},
"task_mining": {
"enabled": true
},
"dataservice": {
"enabled": true
}
}
Mandatory parameters |
Description |
---|---|
|
The load balancer (multi-node HA-ready production mode) or machine (single-node evaluation mode) domain name. |
|
Fixed address used to load balance node registration and kube API requests. This should be fqdn if load balancer is configured as recommended. Otherwise FQDN of 1st Server Node. Refer to Configuring the load balancer. Can be the IP/FQDN of the first rke2 server in your setup. |
|
Set to
true when choosing a multi-node HA-ready production profile. The value of this flag is set automatically by the interactive install wizard. It is used for internal purposes only and should
not be modified manually. |
|
The username that you would like to set as admin (such as: admin ) for the host organization. |
|
The host admin password to be set. |
|
Use a newly generated GUID here. This is a pre-shared, cluster-specific secret. It is needed for all the nodes joining the cluster. |
|
Sets the profile of the installation. The available profiles are:
The value of this flag is set automatically by the interactive install wizard. It is used for internal purposes only and should not be modified manually. |
|
The username that you would like to set for the docker registry installation. |
|
The password that you would like to set for the docker registry installation. |
Optional parameters |
Description |
---|---|
|
true or false - used to opt-out of sending telemetry back to UiPath®. It is set to false by default.
If you wish to opt out, then set to
true .
|
Please refer to the prerequisite documents to obtain certificate:
- Single-node evaluation mode: Configuring the certificates
- Multi-node HA-ready production: Configuring the certificates
If no certificate is provided at the time of installation, the installer creates a self-issued certificate and configures it in the cluster. The validity of the certificate is 90 days.
pwd
to get the path of the directory where files are placed and append the certificate file name to the cluster_config.json
.
In multi-node HA-ready production installations, a certificate is required only on the first node.
Parameter |
Description |
---|---|
|
Absolute path to the Certificate Authority (CA) certificate. This CA is the authority that signs the TLS certificate. A CA bundle should contain only the chain certificates used to sign the TLS certificate. The chain limit is up to nine certificates. If you are using a self-signed certificate, you need to specify the path to the
rootCA.crt , which you previously created. Leave blank if you want the installer to generate.
|
|
Absolute path to TLS certificate (
server.crt for self-signed created in earlier steps). Leave blank if you want the installer to generate.
|
|
Absolute path to certificate key (
server.key for self-signed created in earlier steps). Leave blank if you want the installer to generate.
|
|
Absolute path to the Identity Service certificate used to sign tokens (
identity.pfx for self-signed created in earlier steps). Leave blank if you want the installer to generate an identity certificate using
the server certificate.
|
|
Plain text password set when it was exported. |
|
Absolute path to the file containing additional CA certificates that you want to be trusted by all the products running as
part of Automation Suite. All certificates in the file should be valid
PEM format.
For example, you need to provide the file containing the SQL server CA certificate if the certificate is not issued by a public certificate authority. |
The interactive installer automatically creates databases using the following workflow:
-
The interactive installer script checks the value of the
sql.create_db
parameter in thecluster_config.json
file.-
If the
sql.create_db
parameter is set totrue
, the installer automatically generates all the databases on your behalf. In this case, the installer uses the default database names and default templates, and ignores any custom database names you provided.For details, see Automatically create the necessary databases.
-
If
sql.create_db
is set tofalse
, you must bring your own databases. In this case, you must manually set up your databases. Note that you can use custom database names, provided that you follow the provided naming conventions. This step is critical because we use the database name in conjunction with the connection template to form the database connection string. If you do not follow the recommended naming convention, you must provide the SQL connection strings yourself.For details, see Bring your own databases.
-
-
The interactive installer generates the connection strings as follows:
-
If you did not define a connection string for your service, the installer uses the connection template to generate all database connection strings.
-
If you defined a connection string for your service, the installer uses the provided connection string for your database.
-
If you want the installer to create the databases, fill in the following fields:
Parameter |
Description |
---|---|
|
Set to
true to allow the installer to create the databases. Note that the installer uses the default database names and default templates,
and ignores any custom database names you provided.
|
|
FQDN of the SQL server where you want the installer to configure database. |
|
Port number on which a database instance should be hosted in the SQL server. |
|
Username / userid to connect to the SQL server. |
|
Password of the username provided earlier to connect to the SQL server. |
dbcreator
role. This grants them permission to create the database in SQL Server. Otherwise the installation fails.
ODBC connection does not support usernames that contain special characters. For database usernames for AI Center and Document Understanding, use only uppercase and lowercase letters.
If you choose to bring your own databases for a new Automation Suite installation, we strongly recommend setting up new databases rather than using existing ones. This precaution is necessary to prevent any conflicts with the operation of Automation Suite that might occur due to leftover metadata from old databases.
If you bring your own database, you must provide the SQL connection strings for every database. Automation Suite supports the following SQL connection string formats:
Parameter |
Description |
Products |
---|---|---|
|
Full ADO.NET connection string where Catalog name is set to DB_NAME_PLACEHOLDER. The installer will replace this placeholder with the default database names for the installed suite services. |
Platform, Orchestrator, Test Manager, Automation Hub, Automation Ops, Insights, Task Mining, Data Service, Document Understanding |
|
Full JDBC connection string where database name is set to DB_NAME_PLACEHOLDER. The installer will replace this placeholder with the default database names for the installed suite services. |
AI Center |
|
Full ODBC connection string where database name is set to DB_NAME_PLACEHOLDER. The installer will replace this placeholder with the default database names for the installed suite services. |
Document Understanding |
db_securityadmin
and db_owner
roles for all Automation Suite databases. If security restrictions do not allow the use of db_owner
, then the SQL account should have the following roles and permissions on all databases:
db_ddladmin
db_datawriter
db_datareader
EXECUTE
permission on dbo schema
If you manually set the connection strings in the configuration file, you can escape SQL, JDBC, or ODBC passwords as follows:
- for SQL: add
'
at the beginning and end of the password, and double any other'
. - for JDBC/ODBC: add
{
at the beginning of the password and}
at the end, and double any other}
.
TrustServerCertificate
is set to False
, and you must provide an additional CA certificate for the SQL Server. This is required if the SQL Server certificate is
self-signed or signed by an internal CA.
See Certificate configuration for more details.
sql_connection_string_template example
Server=tcp:sfdev1804627-c83f074b-sql.database.windows.net:1433;Initial Catalog=DB_NAME_PLACEHOLDER;Persist Security Info=False;User Id=testadmin@sfdev1804627-c83f074b-sql.database.windows.net;Password=***;MultipleActiveResultSets=False;Encrypt=True;TrustServerCertificate=False;Connection Timeout=30;Max Pool Size=100;
Server=tcp:sfdev1804627-c83f074b-sql.database.windows.net:1433;Initial Catalog=DB_NAME_PLACEHOLDER;Persist Security Info=False;User Id=testadmin@sfdev1804627-c83f074b-sql.database.windows.net;Password=***;MultipleActiveResultSets=False;Encrypt=True;TrustServerCertificate=False;Connection Timeout=30;Max Pool Size=100;
sql_connection_string_template_jdbc example
jdbc:sqlserver://sfdev1804627-c83f074b-sql.database.windows.net:1433;database=DB_NAME_PLACEHOLDER;user=testadmin;password=***;encrypt=true;trustServerCertificate=false;Connection Timeout=30;hostNameInCertificate=sfdev1804627-c83f074b-sql.database.windows.net"
jdbc:sqlserver://sfdev1804627-c83f074b-sql.database.windows.net:1433;database=DB_NAME_PLACEHOLDER;user=testadmin;password=***;encrypt=true;trustServerCertificate=false;Connection Timeout=30;hostNameInCertificate=sfdev1804627-c83f074b-sql.database.windows.net"
sql_connection_string_template_odbc example
SERVER=sfdev1804627-c83f074b-sql.database.windows.net,1433;DATABASE=DB_NAME_PLACEHOLDER;DRIVER={ODBC Driver 17 for SQL Server};UID=testadmin;PWD=***;MultipleActiveResultSets=False;Encrypt=YES;TrustServerCertificate=NO;Connection Timeout=30;"
SERVER=sfdev1804627-c83f074b-sql.database.windows.net,1433;DATABASE=DB_NAME_PLACEHOLDER;DRIVER={ODBC Driver 17 for SQL Server};UID=testadmin;PWD=***;MultipleActiveResultSets=False;Encrypt=YES;TrustServerCertificate=NO;Connection Timeout=30;"
Default and optional DB names for Automation Suite services
{
"orchestrator": "AutomationSuite_Orchestrator",
"orchestrator_ta": "AutomationSuite_Orchestrator",
"orchestrator_upd": "AutomationSuite_Platform",
"platform": "AutomationSuite_Platform",
"test_manager": "AutomationSuite_Test_Manager",
"automation_ops": "AutomationSuite_Platform",
"automation_hub": "AutomationSuite_Automation_Hub",
"insights": "AutomationSuite_Insights",
"task_mining": "AutomationSuite_Task_Mining",
"dataservice": "AutomationSuite_DataService",
"aicenter": "AutomationSuite_AICenter",
"documentunderstanding": "AutomationSuite_DU_Datamanager",
}
{
"orchestrator": "AutomationSuite_Orchestrator",
"orchestrator_ta": "AutomationSuite_Orchestrator",
"orchestrator_upd": "AutomationSuite_Platform",
"platform": "AutomationSuite_Platform",
"test_manager": "AutomationSuite_Test_Manager",
"automation_ops": "AutomationSuite_Platform",
"automation_hub": "AutomationSuite_Automation_Hub",
"insights": "AutomationSuite_Insights",
"task_mining": "AutomationSuite_Task_Mining",
"dataservice": "AutomationSuite_DataService",
"aicenter": "AutomationSuite_AICenter",
"documentunderstanding": "AutomationSuite_DU_Datamanager",
}
sql_connection_str
for that specific service.
You still have to manually create these databases before running the installer.
Overriding the default connection string for Orchestrator and the platform
{
"orchestrator": {
"sql_connection_str": "Server=tcp:sfdev1804627-c83f074b-sql.database.windows.net,1433;Initial Catalog=CustomOrchDB;Persist Security Info=False;User Id=testadmin@sfdev1804627-c83f074b-sql.database.windows.net;Password=***;MultipleActiveResultSets=False;Encrypt=True;TrustServerCertificate=False;Connection Timeout=30;Max Pool Size=100;"
},
"platform": {
"sql_connection_str": "Server=tcp:sfdev1804627-c83f074b-sql.database.windows.net,1433;Initial Catalog=CustomIDDB;Persist Security Info=False;User Id=testadmin@sfdev1804627-c83f074b-sql.database.windows.net;Password=***;MultipleActiveResultSets=False;Encrypt=True;TrustServerCertificate=False;Connection Timeout=30;Max Pool Size=100;"
}
}
{
"orchestrator": {
"sql_connection_str": "Server=tcp:sfdev1804627-c83f074b-sql.database.windows.net,1433;Initial Catalog=CustomOrchDB;Persist Security Info=False;User Id=testadmin@sfdev1804627-c83f074b-sql.database.windows.net;Password=***;MultipleActiveResultSets=False;Encrypt=True;TrustServerCertificate=False;Connection Timeout=30;Max Pool Size=100;"
},
"platform": {
"sql_connection_str": "Server=tcp:sfdev1804627-c83f074b-sql.database.windows.net,1433;Initial Catalog=CustomIDDB;Persist Security Info=False;User Id=testadmin@sfdev1804627-c83f074b-sql.database.windows.net;Password=***;MultipleActiveResultSets=False;Encrypt=True;TrustServerCertificate=False;Connection Timeout=30;Max Pool Size=100;"
}
}
sql_connection_str
in the corresponding product blocks. The connection string should have a format supported by the respective product.
Example for setting database connection string for AI Center
Parameter |
Description |
---|---|
|
AICenter JDBC connection string (Refer below for the JDBC format) |
"aicenter": {
"enabled": true,
"sql_connection_str": "jdbc:sqlserver://sfdev1804627-c83f074b-sql.database.windows.net;database=aicenter;user=testadmin@sfdev1804627-c83f074b-sql.database.windows.net;password=TFgID_9GsE7_P@srCQp0WemXX_euHQZJ"
}
"aicenter": {
"enabled": true,
"sql_connection_str": "jdbc:sqlserver://sfdev1804627-c83f074b-sql.database.windows.net;database=aicenter;user=testadmin@sfdev1804627-c83f074b-sql.database.windows.net;password=TFgID_9GsE7_P@srCQp0WemXX_euHQZJ"
}
Sample Document Understanding connection string
"documentUnderstanding": {
"datamanager": {
"sql_connection_str": "SERVER=sql-server.database.windows.net;DATABASE=datamanager;DRIVER={ODBC Driver 17 for SQL Server};UID=testadmin;PWD=myPassword"
},
"sql_connection_str": "Server=tcp:database.example.com,1433;Initial Catalog=db;Persist Security Info=False;User Id=testadmin@example.com;Password=***;MultipleActiveResultSets=False;Encrypt=True;TrustServerCertificate=False;Connection Timeout=30;Max Pool Size=100;",
"cjkOcr":
{
"enabled": true
}
}
"documentUnderstanding": {
"datamanager": {
"sql_connection_str": "SERVER=sql-server.database.windows.net;DATABASE=datamanager;DRIVER={ODBC Driver 17 for SQL Server};UID=testadmin;PWD=myPassword"
},
"sql_connection_str": "Server=tcp:database.example.com,1433;Initial Catalog=db;Persist Security Info=False;User Id=testadmin@example.com;Password=***;MultipleActiveResultSets=False;Encrypt=True;TrustServerCertificate=False;Connection Timeout=30;Max Pool Size=100;",
"cjkOcr":
{
"enabled": true
}
}
Handwriting is always enabled for online installation.
max_cpu_per-pod
is 2
, but you can adjust it according to your needs. For more information, check the Document Understanding configuration file.
orchestrator.orchestrator_robot_logs_elastic
section. If not provided, robot logs are saved to Orchestrator's database.
orchestrator.orchestrator_robot_logs_elastic
fields:
Parameter |
Description |
---|---|
|
The address of the Elasticsearch instance that should be used. It should be provided in the form of a URI. If provided, then username and password are also required. |
|
The Elasticsearch username, used for authentication. |
|
The Elasticsearch password, used for authentication. |
Example:
"orchestrator": {
"orchestrator_robot_logs_elastic": {
"elastic_uri": "https://elastic.example.com:9200",
"elastic_auth_username": "elastic-user",
"elastic_auth_password": "elastic-password"
}
}
"orchestrator": {
"orchestrator_robot_logs_elastic": {
"elastic_uri": "https://elastic.example.com:9200",
"elastic_auth_username": "elastic-user",
"elastic_auth_password": "elastic-password"
}
}
If enabling Insights, users can include SMTP server configuration that will be used to send scheduled emails/alert emails. If not provided, scheduled emails and alert emails will not function.
insights.smtp_configuration
fields details:
Parameter |
Description |
---|---|
|
Valid values are
TLSv1_2 , TLSv1_1 , SSLv23 . Omit key altogether if not using TLS.
|
|
Address that alert/scheduled emails will be sent from. |
|
Hostname of the SMTP server. |
|
Port of the SMTP server. |
|
Username for SMTP server authentication. |
|
Password for SMTP server authentication. |
Example:
"insights": {
"enabled": true,
"smtp_configuration": {
"tls_version": "TLSv1_2",
"from_email": "test@test.com",
"host": "smtp.sendgrid.com",
"port": 587,
"username": "login",
"password": "password123"
}
}
"insights": {
"enabled": true,
"smtp_configuration": {
"tls_version": "TLSv1_2",
"from_email": "test@test.com",
"host": "smtp.sendgrid.com",
"port": 587,
"username": "login",
"password": "password123"
}
}
To provision enough resources for monitoring (see Using the monitoring stack), you should consider the number of vCPUs in the cluster and the amount of desired metric retention. See below for how to set the following monitoring resource configurations.
The following table describes the monitoring field details:
Parameter |
Description |
---|---|
|
In days. This is the amount of days that metrics will be retained for the purpose of visualization in Grafana and manual querying via the Prometheus console. Default value is
7 .
|
|
In GiB. Amount of storage space to reserve per Prometheus replica. A good rule of thumb is to set this value to:
Example: If you set
prometheus_retention to 14 days, and your cluster has 80 cores spread across 5 machines, this becomes:
Default value is
45 and should not be set lower.
If Prometheus starts to run out of storage space, an alert will be triggered with specific remediation instructions. See here. |
|
In MiB. Amount of memory to limit each Prometheus replica to. A good rule of thumb is to set this value to:
Example: If you've set
prometheus_retention to 14 days, and your cluster has 80 cores spread across 5 machines, this becomes:
Default value is
3200 for the single-node evaluation mode, and 6000 for the multi-node HA-ready production mode, and should not be set lower.
If Prometheus starts to run out of memory, an alert will be triggered with specific remediation instructions. See here. |
Example:
"monitoring": {
"prometheus_retention": 14,
"prometheus_memory_limit": 16000,
"prometheus_storage_size": 104
}
"monitoring": {
"prometheus_retention": 14,
"prometheus_memory_limit": 16000,
"prometheus_storage_size": 104
}
Make sure you meet the proxy server requirements before configuring the proxy server during installation.
For details, see Step 2: Adding proxy configuration to each node.
cluster_config.json
during the advanced configuration step.
You need to add the following to the configuration file using vim or your favorite editor to the configuration file:
"proxy": {
"enabled": true,
"http_proxy": "http://<PROXY-SERVER-IP>:<PROXY-PORT>",
"https_proxy": "http://<PROXY-SERVER-IP>:<PROXY-PORT>",
"no_proxy": "alm.<fqdn>,<fixed_rke_address>:9345,<fixed_rke_address>:6443,<named server address>,<metadata server address>,<k8s address range>,<private_subnet_ip>,<sql server host>,<Comma separated list of ips that should not go through proxy server>"
}
"proxy": {
"enabled": true,
"http_proxy": "http://<PROXY-SERVER-IP>:<PROXY-PORT>",
"https_proxy": "http://<PROXY-SERVER-IP>:<PROXY-PORT>",
"no_proxy": "alm.<fqdn>,<fixed_rke_address>:9345,<fixed_rke_address>:6443,<named server address>,<metadata server address>,<k8s address range>,<private_subnet_ip>,<sql server host>,<Comma separated list of ips that should not go through proxy server>"
}
Mandatory parameters |
Description |
---|---|
|
Use
true or false to enable or disable proxy settings.
|
|
Used to route HTTP outbound requests from the cluster. This should be the proxy server FQDN and port. |
|
Used to route HTTPS outbound requests from the cluster. This should be the proxy server FQDN and port. |
|
Comma-separated list of hosts, IP addresses, or IP ranges in CIDR format that you do not want to route via the proxy server.
This should be a private subnet range, sql server host, named server address, metadata server address:
*.<fqdn>,<fixed_rke_address>:9345,<fixed_rke2_address>:6443 .
|
To enable resilience to zonal failures in a multi-node cluster, take the following steps:
cluster_config.json
using the interactive installation wizard, exit at advanced configuration step and add the following to the configuration
file using vim or your favorite editor:
"zone_resilience": true
"zone_resilience": true
Mandatory parameters |
Description |
---|---|
|
Use
true or false to enable or disable resilience to zonal failure.
|
--zone
and --region
arguments is:
- recommended if you provision your machines on AWS, Azure, or GCP, and metadata services are enabled as the installer populates the zone and region details.
- mandatory if you provision your machines on AWS, Azure, or GCP, and metadata services are disabled, or if you opt for a different cloud provider.
/etc/resolv.conf
. Kubernetes does not work with local DNS resolvers (127.0.0.1 or 127.0.0.0/8), so if you have such name servers configured
in /etc/resolv.conf
file, you need to pass a file reference with the correct nameserver entries accessible from anywhere on the VM in the .infra.custom_dns_resolver
parameter in cluster_config.json
.
For details on a known limitation, see Kubernetes documentation.
Optional Parameters |
Description |
---|---|
|
Path to the file with correct name server entries that can be accessed from anywhere on the VM. These name server entries
must not be from
127.0.0.0/8 .
|
- Cluster_config.json Sample
- General configuration
- Certificate configuration
- Database configuration
- Database creation workflow
- Automatically Create the Necessary Databases
- Bring Your Own Database
- Orchestrator-specific Configuration
- Insights-specific Configuration
- Monitoring Configuration
- Optional: Configuring the proxy server
- Optional: Enabling resilience to zonal failures in a multi-node HA-ready production cluster
- Optional: Passing custom resolv.conf