- Overview
- About the UI Automation activities
- UI Automation
- Applications and technologies automated with UI Automation
- Project compatibility
- UI-ANA-016 - Pull Open Browser URL
- UI-ANA-017 - ContinueOnError True
- UI-ANA-018 - List OCR/Image Activities
- UI-DBP-006 - Container Usage
- UI-DBP-013 - Excel Automation Misuse
- UI-DBP-030 - Forbidden Variables Usage In Selectors
- UI-PRR-001 - Simulate Click
- UI-PRR-002 - Simulate Type
- UI-PRR-003 - Open Application Misuse
- UI-PRR-004 - Hardcoded Delays
- UI-REL-001 - Large Idx in Selectors
- UI-SEC-004 - Selector Email Data
- UI-SEC-010 - App/Url Restrictions
- UI-USG-011 - Non Allowed Attributes
- UX-SEC-010 - App/Url Restrictions
- UX-DBP-029 - Insecure Password Use
- UI-PST-001 - Audit Log Level in Project Settings
- UiPath Browser Migration Tool
- Clipping region
- Computer Vision Recorder
- Activate
- Anchor Base
- Attach Browser
- Attach Window
- Block User Input
- Callout
- Check
- Click
- Click Image
- Click Image Trigger
- Click OCR Text
- Click Text
- Click Trigger
- Close Application
- Close Tab
- Close Window
- Context Aware Anchor
- Copy Selected Text
- Element Attribute Change Trigger
- Element Exists
- Element Scope
- Element State Change Trigger
- Export UI Tree
- Extract Structured Data
- Find Children
- Find Element
- Find Image
- Find Image Matches
- Find OCR Text Position
- Find Relative Element
- Find Text Position
- Get Active Window
- Get Ancestor
- Get Attribute
- Get Event Info
- Get From Clipboard
- Get Full Text
- Get OCR Text
- Get Password
- Get Position
- Get Source Element
- Get Text
- Get Visible Text
- Go Back
- Go Forward
- Go Home
- Google Cloud Vision OCR
- Hide Window
- Highlight
- Hotkey Trigger
- Hover
- Hover Image
- Hover OCR Text
- Hover Text
- Image Exists
- Indicate On Screen
- Inject .NET Code
- Inject Js Script
- Invoke ActiveX Method
- Key Press Trigger
- Load Image
- Maximize Window
- Microsoft Azure Computer Vision OCR
- Microsoft OCR
- Microsoft Project Oxford Online OCR
- Minimize Window
- Monitor Events
- Mouse Trigger
- Move Window
- Navigate To
- OCR Text Exists
- On Element Appear
- On Element Vanish
- On Image Appear
- On Image Vanish
- Open Application
- Open Browser
- Refresh Browser
- Replay User Event
- Restore Window
- Save Image
- Select Item
- Select Multiple Items
- Send Hotkey
- Set Clipping Region
- Set Focus
- Set Text
- Set To Clipboard
- Set Web Attribute
- Show Window
- Start Process
- System Trigger
- Take Screenshot
- Tesseract OCR
- Text Exists
- Tooltip
- Type Into
- Type Secure Text
- Use Foreground
- Wait Attribute
- Wait Element Vanish
- Wait Image Vanish
- Application Event Trigger
- Check/Uncheck
- Check App State
- Check Element
- Click
- Click Event Trigger
- Drag and Drop
- Extract Table Data
- For Each UI Element
- Get Attribute
- Get Browser Data
- Get Text
- Get URL
- Go to URL
- Highlight
- Hover
- Inject Js Script
- Keyboard Shortcuts
- Keypress Event Trigger
- Mouse Scroll
- Navigate Browser
- Select Item
- Set Browser Data
- Set Runtime Browser
- Set Text
- Take Screenshot
- Type Into
- Use Application/Browser
- Perform browser search and retrieve results using UI Automation APIs
- Web Browsing
- Find Images
- Click Images
- Trigger and Monitor Events
- Create and Override Files
- HTML Pages: Extract and Manipulate Information
- Window Manipulation
- Automated List Selection
- Find and Manipulate Window Elements
- Manage Text Automation
- Load and Process Images
- Manage Mouse Activated Actions
- Automate Application Runtime
- Automated Run of a Local Application
- Browser Navigation
- Web Automation
- Trigger Scope Example
- Enable UI Automation support in DevExpress
- Computer Vision Local Server
- Mobile Automation
- Release notes
- Project compatibility
- Get Log Types
- Get Logs
- Get Page Source
- Get Device Orientation
- Get Session Identifier
- Install App
- Manage Current App
- Manage Other App
- Open DeepLink
- Open URL
- Mobile Device Connection
- Directional Swipe
- Draw Pattern
- Positional Swipe
- Press Hardware Button
- Set Device Orientation
- Take Screenshot
- Take Screenshot Part
- Element Exists
- Execute Command
- Get Attribute
- Get Selected Item
- Get Text
- Set Selected Item
- Set Text
- Swipe
- Tap
- Type Text
- Terminal
UI Automation Activities
About the UI Automation activities
UI automation activities enable you to interact with graphical user interface elements in software applications. These activities include clicking buttons, typing into text boxes, selecting from dropdown menus, and more. This is done by using UiPath Robots that perform tasks the same way a human user would, making it easier to automate complex business processes and repetitive tasks.
When automating user inferfaces, the initial phase consists in understanding the logical sequence of steps that a human user would take. Then the next phase consists in translating these steps into activities and configuring them. UI automation activities can serve as containers, input, output, and synchronization activities, while the activity properties determine how the UiPath Robots perform an action.
The UI Automation activities include the following packages:
- UI Automation
- Computer Vision Local Server
- Mobile Automation
- Terminal
Use the following table to identify the compatibility between the platform and the activity packages:
Activity package | Windows - Legacy Studio projects | Windows (Modern) Studio projects | Cross-platform Studio projects |
---|---|---|---|
UI Automation |
|
|
|
Computer Vision Local Server |
|
|
|
Mobile Automation |
|
|
|
Terminal |
|
|
|
The UI Automation activity package offers an extensive range of capabilities that simulate human behavior. It provides powerful and comprehensive features designed to enhance productivity by building user interface automations, saving time and minimizing errors. Some of the most common use cases are:
- Business process automation, such as customer service, accounting, order processing, employee onboarding, etc.
- Web scraping, such as customer feedback analysis, price comparison, lead generation, product review scraping, etc.
- Automated testing on user interfaces to ensure the applications function correctly
To learn more, go to the UI Automation package section.
The Computer Vision Local Server package contains no activities, but once installed in a project, enables you to use a local AI Computer Vision server.
AI Computer Vision is an AI skill that allows all UiPath Robots to capture the screen and visually identify all the UI elements, rather than relying on selectors or images. It is an algorithm that enables human-like recognition of user interfaces, using a mix of AI, OCR, text fuzzy-matching, and an anchoring system to tie it all together.
To learn more, go to the Computer Vision Local Server package section.
The Mobile Automation activity package offers a robust set of activities for automating mobile application testing and interaction. It provides tools for testing Android and iOS applications, allowing actions like tapping, swiping, entering text, and validating mobile elements, among others. Key features of the activity package include:
- Multi-platform mobile automation - automate mobile applications across Android and iOS platforms
- Mobile element interaction - perform actions like tapping, swiping, scrolling, and entering text on target mobile elements
- Mobile validation and verification - validate mobile element properties, verify expected values, and handle specific scenarios
- Mobile device management - capture device information, manage device orientation, control network connectivity, install and uninstall applications, and handle device-specific settings
Get started with Mobile Automation
The Terminal activity package enables you to connect to a wide range of terminal environments using a direct connection or terminal emulators.
You can then use activities to perform actions such as retrieving text, colors, fields, or screen positions, sending keys, moving the cursor, or waiting for certain text or fields to appear as triggers.
To learn more, go to the Terminal package section.