- collaboration
- Invite Team Members
- Assign Projects
- Users & Role Management
- Review Management [Test Cases]
- Review Management [Elements]
- Execution Controls
- Manage Test Cases
- Test Case List Actions
- Import and Export Test Cases
- Import Test Project Test Cases
- Importing Postman Collections and Environments
- Update Test Case result in a Test Plan
- Test Cases (Mobile Web App)
- Test Step Types
- Type: Natural Language
- Type: REST API
- Type: Step Group
- Type: For Loop
- Type: While Loop
- Type: Block
- Type: If Condition
- Nested Step Groups
- Image Injection
- Cross-application testing
- Test Data Types
- Raw
- Parameter
- Runtime
- Random
- Data Generator
- Phone Number
- Mail Box
- Environment
- Concat Test Data
- Create Test Data [Parameter]
- Update Test Data Profile
- Updating Value in TDP
- Import TDP
- Bulk Deletion of a Test Data Profile
- Create Test Data [Environment]
- Elements (Objects)
- Web Applications
- Record Single Element
- Record Multiple Elements
- Create Elements
- Supported Locator Types
- Formulating Elements
- Shadow DOM Elements
- Verifying elements in Chrome DevTools
- Handling iframe Elements?
- Dynamic Locators using Parameter
- Dynamic Locators using Runtime
- Using Environment Test Data for Dynamic Locators
- Locating Dynamic Elements in Date Widget
- Freeze & Inspect Dynamic Elements (WebPage)
- Locating Dynamic Elements in Tables
- Import/Export Elements
- AI Enabled Auto-Healing
- Locator Precedence (Web Apps)
- Verify Elements from Test Recorder
- test step recorder
- Install Chrome Extension
- Install Firefox Extension
- Install Edge Extension
- Exclude Attributes/Classes
- test plans
- Add, Edit, Delete Test Machines
- Add, Edit, Delete Test Suites
- Schedule Test Plans
- Run Test Suites In Parallel
- Cross Browser Testing
- Distributed Testing
- Headless Testing
- Test Lab Types
- Disabling Test Cases in Test Plans
- AfterTest Case
- Post Plan Hook
- AfterTest Suite
- Email Configuration in Test Plan
- Execute Partial Test Plans via API
- Ad-hoc Run
- Test Plan Executions
- Dry Runs on Local Devices
- Run Tests on Private Grid
- Run Tests on Vendor Platforms
- Run Test Plans on Local Devices
- Test Locally Hosted Applications
- Debug Test Case Failures
- Parallel and Allowed queues
- debugging
- Debug results on local devices (Web applications)
- Debug Results on Local Devices
- Launch Debugger in the Same Window
- Testsigma Agent
- Pre-requisites
- Setup: Windows, Mac, Linux
- Setup: Android Local Devices
- Setting up iOS Local Devices
- Update Agent Manually
- Update Drivers Manually
- Delete Corrupted Agent
- Delete Agents: Soft & Permanent
- Triggering Tests on Local Devices
- troubleshooting
- Agent - Startup and Registration Errors
- Agent Logs
- Upgrade Testsigma Agent Automatically
- Specify Max Sessions for Agents
- Testsigma Agent - FAQs
- continuous integration
- Test Plan Details
- REST API (Generic)
- Jenkins
- Azure DevOps
- AWS DevOps
- AWS Lambda
- Circle CI
- Bamboo CI
- Travis CI
- CodeShip CI
- Shell Script(Generic)
- Bitrise CI
- GitHub CICD
- Bitbucket CICD
- GitLab CI/CD
- desired capabilities
- Most Common Desired Capabilities
- Browser Console Debug Logs
- Geolocation Emulation
- Bypass Unsafe Download Prompt
- Geolocation for Chrome & Firefox
- Custom User Profile in Chrome
- Emulate Mobile Devices (Chrome)
- Add Chrome Extension
- Network Throttling
- Network Logs
- Biometric Authentication
- Enable App Resigning in iOS
- Enable Capturing Screenshots (Android & iOS)
- Configure Android WebViews
- Incognito/Private mode
- Set Google Play Store Credentials
- addons
- What is an Addon?
- Addons Community Marketplace
- Install Community Addon
- Prerequisites(Create/Update Addon)
- Create an Addon
- Update Addon
- Addon Types
- Create a Post Plan Hook add-on in Testsigma
- Create OCR Text Extraction Addon
- configuration
- API Keys
- Security(SSO)
- Setting Up Google Single Sign-On(SSO) Login in Testsigma
- Setting Up Okta Single Sign-On Integration with SAML Login in Testsigma
- Setting up SAML-based SSO login for Testsigma in Azure
- iOS Settings
- Creating WDA File for iOS App Testing
- uploads
- Upload Files
- Upload Android and iOS Apps
- How to generate mobile builds for Android/iOS applications?
- Testsigma REST APIs
- Environments
- Elements
- Test Plans
- Upload Files
- Get Project Wide Information
- Upload & Update Test Data Profile
- Fetch Test Results (All Levels)
- Trigger Multiple Test Plans
- Trigger Test Plans Remotely & Wait Until Completion
- Run the Same Test Plan Multiple Times in Parallel
- Schedule, Update & Delete a Test Plan Using API
- Update Test Case Results Using API
- Create and update values of Test Data Profile using REST API
- Rerun Test Cases from Run Results using API
- open source dev environment setup
- macOS and IntelliJ Community Edition
- macOS and IntelliJ Ultimate Edition
- Windows and IntelliJ Ultimate Edition
- Setup Dev Environment [Addons]
- NLPs
- Retrieve Value in Text Element
- Capture Dropdown Elements
- Unable to Select Radiobutton
- Unable to Click Checkbox
- Clearing the Session or Cookies
- UI Identifier NLP
- Drag & Drop NLP
- Uploading Files NLP
- setup
- Server Docker Deployment Errors
- Secured Business Application Support
- Troubleshooting Restricted Access to Testsigma
- Why mobile device not displayed in Testsigma Mobile Test Recorder?
- Unable to Create New Test Session
- Agent Startup Failure Due to Used Ports
- Tests Permanently Queued in Local Executions
- Fix Testsigma Agent Registration Failures
- Testsigma Agent Cleanup
- Need of Apache Tomcat for Testsigma Agent
- web apps
- URL not accessible
- Test Queued for a Long Time
- Issues with UI Identifiers
- Missing Elements in Recorder
- Collecting HAR File
- Errors with Browser Session
- Page Loading Issues
- mobile apps
- Failed to Start Mobile Test Recorder
- Troubleshooting “Failed to perform action Mobile Test Recorder” error
- Test Execution State is Queued for a Long Time
- Mobile app keeps stopping after successful launch
- More pre-requisite settings
- Unable to start WDA Process on iPhone
- Most Common causes for Click/Tap NLP failure
- Finding App Package & App Activity (Android)
- Cross-environment Compatible ID Locators (Android)
- Why Accessibility IDs Over other Locators?
- Common Android Issues & Proposed Solutions
- Finding the App Bundle ID for iOS
- on premise setup
- On-Premise Setup Prerequisites
- On-Premise Setup with Docker-compose File
- Post-Installation Checklist for On-Premise Setup
- Install Docker on an Unix OS in Azure Infrastructure
- SMTP Configuration in Testsigma
- Configure Custom Domains
- salesforce testing
- Intro: Testsigma for Salesforce Testing
- Creating a Connected App
- Creating a Salesforce Project
- Creating Metadata Connections
- Adding User Connections
- Build Test Cases: Manual+Live
- Salesforce Element Repositories
- Intro: Testsigma Special NLPs
- Error Handling On Metadata Refresh
- Automating Listview Table NLPs
- windows automation
- Intro: Windows Desktop Automation
- Windows (Adv) Project & Application
- Object Learning (Using UFT One)
- Converting TSR Files to TSRx
- Importing/Updating TSRx Files
- Test Cases for Windows Automation
- Error Handling Post TSRx File Update
Random Test Data in Testsigma
Many test cases need random inputs such as inputting an account number, ID number, phone number, etc. With Testsigma, you can create random alphanumeric data blocks through its Random Test Data.
Prerequisites
- You should be familiar with test step creation using NLPs.
Using Random Test Data
You can use the Random Test Data feature in your test cases by specifying a number ranging from 1 to 256. This will generate a random alphanumeric string with a length as specified by the number during the execution of the test case.
For example, you have a test case that requires entering a passport number, and you want to generate a random 9-digit passport number for every execution. You can achieve this using Random test data type in your Test Case.
Here is a quick GIF demonstrating how to add Random Data to a Test Case in Testsigma.
Using random data may lead to unrepeatable results. You can mitigate this by logging (or otherwise recording) every random choice you make and then playing those choices back. That could be as easy as recording the initial seed to your random number generator, assuming your data does not change over time.
Writing tests in a general enough way to deal with arbitrary, randomly selected data can be hard. This is the harder problem. Choosing a different random integer is one thing; choosing a different user (with properties that vary from one user to the next) is something else. Consider how a test works in the abstract: it chooses some inputs, applies them to a function, and verifies the result.
There are a few ways to approach that:
- The developer precalculates the expected result and hardcodes it into the test. That only works if the developer decides on the inputs ahead of time.
- The test uses the inputs to calculate the expected result and then compares that to the actual result. If the test can use arbitrary (random inputs), this means the test needs to replicate a lot of the logic in the system under test. This is almost certainly not what you want, especially for a complicated system. You are just as likely as the developer to implement a complex algorithm in a buggy way.
- The test uses some other means of verifying that the result matches the input. Sometimes, there are shortcuts or at least alternatives you can take to verify a result. As a naive example, if a system sums a list of numbers, you might verify the result by subtracting the list from the sum and then checking whether the new result is zero. Most systems cannot be tested this way.
A good place to use randomly selected data is in a comparator test, where you compare two versions of the same system using the same inputs. A comparator test will not tell you whether a system is correct, but it will help you find changes in behavior. That might be something for you to consider.