Skip to main content
No Result Found

Test Insights - View advanced analytics

Use Test Insights for analyzing and exploring your test data on BrowserStack.

Test Insights provides visibility into your testing at scale. An interactive dashboard that provides actionable insights to help you improve testing efficiency, identify high impact issues or bottlenecks so you can release quality software at speed.

Visit the following sections to learn more:

Test Insights interface

Test Insights includes visualizations, filters and metrics (like errors, failures, dropped tests and more) to help you get an overall view of your test suite performance. You can also view the desktop and device coverage to optimize your overall testing strategy.

Automate test insights

Get insights about your automated test runs on Browser and Mobile devices. The dashboard gives you insights into the following data:

Test Coverage

This section captures the breadth and width of your test suite, giving you insights into the types of devices, desktops and operating systems that you have been testing.

Desktop & Device Coverage

The overall picture of the tests that you are running on desktops & mobile devices. If you are not running device tests, then the fields will be empty.

  1. Desktop vs Device Coverage (Visualization):The visualization of the ratio of the tests run on Desktops & Mobile Devices. Hovering over the visualization will give you absolute numbers of test sessions run on each platform.

  2. Desktop vs Device Performance (Visualization):The visualization of the performance of tests, to ease the effort of comparing test performances between the platforms. Hovering over the visualization will give you absolute values of the division of the visual.
    Desktop vs Device Performance comparison on different platforms

Desktop Coverage

  1. Unique Browser + OS Combinations (Tile):The number of unique Browser & Operating System combinations that you have tested. For example,Chrome 77 with Windows 8 & Chrome 75 with Windows 8 are counted as two different browser & OS combinations.

  2. Unique OS Versions (Tile): The number of unique Operating System versions that you have tested. For example, Windows 7 and Windows 10 are counted as two unique OS versions.

  3. Unique Browsers Versions (Tile): The number of unique Browsers versions that you have tested. For example Chrome 78, Chrome 76 and Chrome 77 are counted as three unique browser versions.

  4. Operating Systems (Visualization): The overall ratio of all tests run on the two Operating System platforms, i.e. OS X & Windows. Hovering over the visualization will give you absolute values of the visual.

  5. Operating System Performance (Visualization): The visualization of the performance of tests divided by the operating system type, to ease the effort of comparing test performances between the platforms. Hovering over the visualization will give you the absolute value of the division.

    Operating System Performance Visualization of tests performed

  6. Browser Family Breakdown (Visualization): The list of most tested browser family in the test suite, as per volume of tests run. Chrome on Windows 10 and Chrome on Windows 7 are considered different browser families. The visual also gives you deep-dive into the performance of the tests on the browser families, to ease the effort of finding devices with maximum failing test cases.

    Test Coverage Bar Graph, browser family vs volume of tests run

  7. Browser Version Breakdown (Visualization): This list of top tested browser versions in the test suite, as per volume of tests run. Chrome on Windows 10 will be a different family compared to Chrome 78 on Windows 7. The visual also gives you deep-dive into the performance of the tests on the browser versions.

    Browser version breakdown visualization in dashboard

Device Coverage

  1. Unique Device + OS Combinations (Tile): The number of unique Device & Operating System combinations that you have tested. For example, iPhone X with iOS 13 & iPhone X with iOS 12 are counted as two unique device & OS combinations.

  2. Unique OS Versions (Tile): The number of unique Operating Systems that you have tested. For example, Android 7.1 and Android 7.0 are counted as two unique OS versions.

  3. Unique Devices (Tile): The number of unique Devices that you have tested. For example, iPhone X and iPhone Xs are counted as two unique devices.

  4. Operating Systems (Visualization): The overall ratio of all tests run on the two Operating System platforms, i.e. iOS & Android. Hovering over the visualization will give you absolute values of the visual.

  5. Operating System Performance (Visualization): The visualization of the performance of tests divided by the operating system type, to ease the effort of comparing test performances between the platforms. Hovering over the visualization will give you the absolute value of the division.

    Test Insights - Tests device coverage

  6. Device-Operating System Breakdown (Visualization): The list of most tested device-operating system combinations in the test suite, as per volume of tests run. Samsung Galaxy S9 with Android 8.0 and Samsung Galaxy S9 with Android 8.1 are unique devices. The visual also gives you deep-dive into the performance of the tests on the browser families, to ease the effort of finding devices with maximum failing test cases.

    Visualization of device-operating system breakdown

Parallel utilization

This section will give you insights on your parallel usage, queue usage and dropped tests related to unavailable parallel consumption.

  1. Parallel Usage (Visualization): Your plan comes with a specific number of parallel tests, allowing you to run automated tests simultaneously, thereby reducing testing time. You can read more about Parallels.
    • Max Parallels Used: The maximum parallels that were actively being used during the period.
    • Max Parallels Allocated: Maximum parallel tests available on your plan, this can be allocated by the Group Admins & Owners at the team level or the group level.

      View max parallels allocated
  2. Queue Usage (Visualization): Our queuing feature enables you to launch additional tests, even after your parallel test limit is hit. Queue limit is as equal to the parallels available in your license. Attempts to launch tests beyond the queuing limit will fail and will be logged as dropped tests.

    Queue Usage Visualization of Tests

  3. Dropped Test Sessions (Visualization): Test sessions may drop due to various reasons, and this visualization will assist you in avoiding dropping tests related to parallels. A dropped test did not run, it gets dropped either before it got added to the queue or it got queued and then dropped.
    • NPA : No Parallels Available can cause tests to drop; it will happen when the parallels and queue are full.
      Visualization of dropped test sessions

Test Stability

This dashboard will give you insights around the health metrics of your test suites. Focusing on health trends, error deep-dives and more.

  1. Total Test Sessions Run (Tile): The total number of Automate test sessions that you have run on BrowserStack.

  2. Longest Session Duration (Tile): The longest Automate test session, in minutes, that you have run on BrowserStack. Ideally, it should not be long.

  3. Longest Session Duration (Tile): The longest Automate test session, in minutes, that you have run on BrowserStack. Ideally, it should not be long.

    Test session stability circle graph

  4. Test Trends (Visualization): The visualization of the overall health of your tests, with a full breakdown of final statuses. Refer to this section for definitions of the statuses, hovering over the visuals will give you details on the absolute values for the particular granularity.

    Screenshot for test stability on dashboard

  5. Error Percentage (Tile): The number of tests which were not marked as pass or fail, and ended up giving an error state at the end of the test session as a percentage of total test sessions run.

  6. Error Reports (Visualization): The breakdown of the tests which faced errors.
    • FAILED-TO-START-BROWSER : This error occurs when BrowserStack is unable to launch the browser specified in your script, preventing test execution. How to identify the cause and resolve the error, here is the link.
    • REQUEST-TIMEDOUT : This error occurs when a socket hangs up error is raised at the hub nodes. The hub waits for a maximum of 240 seconds before an unresponsive browser is killed, changing the session status to ERROR on the dashboard.
    • OTHERS : There are some other reasons like connectivity issues for an error to occur, we have bucketed all of them into others.

      Viewing test stability section in Test Insights
  7. Dropped Test Trends (Visualization): Test sessions may drop due to various reasons, and this visualization will assist you in avoiding dropping tests related to test configuration issues or incorrect local setup. A dropped test did not run - it gets dropped either before it got added to the queue or it got queued and was then dropped.
    • Incorrect Local Setup: The most common reason for these tests getting dropped is that Local Testing was enabled in the session capabilities but it wasn’t enabled on the system. To avoid this, either remove the local capability or enable local testing via bash/bin file or chrome extension.
    • Incorrect Configuration Issues: These tests get dropped because the capabilities mentioned in the test sessions were incorrect or did not match the system expectations. For example, the specified screen resolution is not compatible with the chosen OS version could be a reason. Please note, as these tests were dropped, OS, OS Version, Browser, Browser Version & Device will not get applied to these visualizations.
  8. Incorrect Local Setup (Tile): The absolute number of test sessions which got dropped because of Local Setup Issues. The most common reason for these tests getting dropped is that Local Testing was enabled in the session capabilities, but it wasn’t enabled on the system. To avoid this, either remove the local capability or enable local testing via bash/bin file or chrome extension.

  9. Incorrect Configuration Issues (Tile): The absolute number of test sessions which got dropped because of Incorrect Test Configuration Issues. The most common reason for these tests getting dropped is that Local Testing was enabled in the session capabilities, but it wasn’t enabled on the system. To avoid this, either remove the local capability or enable local testing via bash/bin file or chrome extension.
    incorrect configuration issues

Build Reports

This dashboard provides you with insights into the build related trends, this dashboard is most powerful when combined with the Build Name filter. Take a look at different builds, and unlock insights on the progress of the builds.

  1. Total Test Time (Tile): The total time that you have used the BrowserStack Testing Platform, is the cumulative value of all the session times.

  2. Build Trends (Visualization): The visualization of the progress of your Build Sizes (Average Test Sessions/Build) and Build Timings (Average Build Time). Useful to gauge the overall progress of your test suites and test time optimization strategies.

    Visualization of the progress of your build sizes and build timings

  3. Build Volume Report (Visualization): Total number of builds tested as per the granularity selection. Really useful to find out the usage of the platform on a daily, weekly or monthly basis.

    Automate build volume report

App Automate test insights

Test Coverage

This dashboard captures the breadth and width of your test suite, giving you insights into the types of devices, and operating systems that you have used for testing your applications.

  1. Unique Device-OS Combinations (Tile): The number of unique Device & Operating System combinations that you have tested. For example, iPhone 11 with iOS 12 & iPhone 11 with iOS 13 are counted as two different Device-OS Combinations combinations.

  2. Unique OS Versions (Tile): The number of unique Operating Systems that you have tested. For example, Android 7.1 and Android 7.0 are counted as two unique OS versions.

  3. Unique Devices (Tile): The number of unique Devices that you have tested. For example, iPhone X and iPhone Xs will be considered as different devices.

  4. Operating Systems (Visualization): The overall ratio of all tests run on the two Operating System platforms, i.e. iOS & Android. Hovering over the visualization will give you absolute values of the visual.

    App automate test coverage

  5. Operating System Breakdown (Visualization): The visualization of the performance of tests divided by the operating system type, to ease the effort of comparing test performances between the platforms. Hovering over the visualization will give you the absolute value of the division.

    Test coverage on dashboard showing operating system breakdown

  6. Device-Operating System Breakdown (Visualization): The list of most tested device-operating system combinations in the test suite, as per the volume of tests run. Samsung Galaxy S9 with Android 8.0 and Samsung Galaxy S9 with Android 8.1 are unique devices. The visual also gives you deep-dive into the performance of the tests on the browser families, to ease the effort of finding devices with maximum failing test cases.

    Operating system breakdown graph for App Automate in Test Coverage

Parallel Utilization

This dashboard will give you insights on your parallel usage, and queue usage, as per the app sessions that you have run on BrowserStack.

  1. Parallel Usage (Visualization): Your plan comes with a specific number of parallel tests, allowing you to run automated tests simultaneously, thereby reducing testing time. You can read more about Parallels.
    • Max Parallels Used: The maximum parallels that were actively being used during the period.
    • Max Parallels Allocated: Maximum parallel tests available on your plan, this can be allocated by the Group Admins & Owners at the team level or the group level.

      Graph for parallel utilization
  2. Queue Usage (Visualization): Our queuing feature enables you to launch additional tests, even after your parallel test limit is hit. Queue limit is as equal to the parallels available in your license. Attempts to launch tests beyond the queuing limit will fail and will be logged as dropped tests.

    App automate queue utilization

  3. Dropped Test Sessions (Visualization): Test sessions may drop due to various reasons, and this visualization will assist you in avoiding dropping tests related to parallels. A dropped test did not run, it gets dropped either before it got added to the queue or it got queued and then dropped.
    • NPA: No Parallels Available can cause tests to drop; it will happen when the parallels and queues are full.
      Dropped Test Sessions

Test Stability

This dashboard will give you insights around the health metrics of your test suites. Focusing on health trends, error deep-dives, and more.

  1. Total Test Sessions Run (Tile): The total number of App Automate test sessions that you have run on BrowserStack.

  2. Longest Session Duration (Tile): The longest App Automate test session, in minutes, that you have run on BrowserStack. Ideally, it should be a short duration.

  3. Test Status Breakdown (Visualization): Overall breakdown of all the Automate test sessions that you have run, with an analysis of the final test status. Refer to this section for definitions of the statuses. Hovering over the individual pie-sections will give you the absolute value of the number of tests with the final status.

    Test status breakdown

  4. Error Percentage (Tile): The number of tests which were not marked as pass or fail, and ended up giving an error state at the end of the test session as a percentage of total test sessions run.

  5. Error Reports (Visualization): The breakdown of the tests which faced errors.

    App Automate Tests Insights - Test stability error percentage

  6. Appium/Espresso/XCUITest Test Trends (Visualization): The visualization of the overall health of your tests, with a full breakdown of final statuses. Refer to this section for definitions of the statuses, hovering over the visuals will give you details on the absolute values for the particular granularity.
    • Appium: Each session is counted as an individual test session.

      Graph showing AppAutomate appium test trends
    • Espresso & Xcuitest: Each build can have multiple test cases executed, and each test case is counted individually in the visualization.

      Espresso test trend chart in Test insights

      Visualization of app automate XCUI test trends
  7. Appium/Espresso/XCUITest Dropped Test Trends (Visualization): Test sessions may drop due to various reasons, and this visualization will assist you in avoiding dropping tests related to test configuration issues. A dropped test did not run, it gets dropped either before it got added to the queue or it got queued and then dropped.

  8. Appium/Espresso/XCUITest Dropped Test Count (Visualization): The absolute number of test sessions which got dropped. Test sessions may drop due to various reasons, and this visualization will assist you in avoiding dropping tests related to test configuration issues.
    Viewing Test Stability App Automate Dropped graphs
    teststability-appautomate-espressodroppedtest
    Image showing XCUI dropped test trends and counts when Test Stability is selected under App Automate Test Insights

Build Reports

This dashboard provides you with insights into the build related trends, this dashboard is most powerful when combined with the Build Name filter. Take a look at different builds, and unlock insights on the progress of the builds.

  1. Total Test Time (Tile): The total time that you have used the BrowserStack Testing Platform, is the cumulative value of all the session times.

  2. Build Size Report (Visualization): The average duration of the tests, broken down by frameworks.
    • Xcuitest (Tests): Each session is counted as an individual test session.
    • Espresso (Tests): Each build can have multiple test cases executed, and each test case is counted individually in the visualization.
    • Appium (Test Sessions): Each build can have multiple test cases executed, and each test case is counted individually in the visualization.

      Graphical visualisation of each test in particular build
  3. Build Time Report (Visualization): The average duration of the tests, broken down by frameworks. The visualization of the progress of your Build Timings (Average Build Time). Useful to gauge the overall progress of your test suites and test time optimization strategies.

    Graph for daily average build time

  4. Build Volume Report (Visualization): Total number of builds tested as per the granularity selection. Really useful to find out the usage of the platform on a daily, weekly or monthly basis.

    Screenshot of build reports provided on dashboard
Note: Test Insights doesn’t include data for JS Testing currently.

Live test insights

Get insights about your manual website tests across multiple browsers. The dashboard gives you insights into the following data:

Coverage

This section captures the breadth and width of your test suite, giving you insights into the types of browsers, devices and operating systems tested on.

Graphical resprentation of unique browser and os combinations tested on

Testing frequency

View the testing frequency in any given month - number of days the tested, number of tests run and total minutes used.

View Testing Frequency in a month

App Live test insights

Get insights about your manual app tests across multiple devices. The dashboard gives you insights into the following data:

Coverage

This section captures the breadth and width of your test suite, giving you insights into the types of browsers, devices and operating systems tested on.

Coverage

Testing frequency

View the testing frequency in any given month - number of days the tested, number of tests run and total minutes used.

Testing frequency chart

Filter analytics

Filter options

Each dashboard has its own set of filters relevant to the information that is displayed.

These filters allow restricting the data displayed based on the criteria you specify. For example, you might want to filter the builds to the last two months, or for a specific project.

Note: Each filter on the dashboard will include three sections: the filter field name (e.g., Teams), the operator type (for example, is equal to), and the target filter value. You can also add multiple combinations of the three sections, i.e. Field Name, Operator Type, and Target Value.

The combination of these three sections provides you with hundreds of permutations, which unlocks powerful data visualization options.

Filter names

The filters are dynamic, meaning depending on the data in the visualizations the filters might be different for different dashboards.

Image of Filters like timerange,user,build,os,etc. available under Automate Test Insights when Test Stability is selected

Name Description
Data Granularity The date granularity gives you the option of changing the granularity on timeline-based visualizations. The possibilities range from hourly data to monthly data, really useful to find daily, weekly, or monthly trends.
Date Range The date range filter gives you the option to select the date range for the whole dashboard. Once you choose the same, the entire dashboard will reload data as per the selection. Combining an operator like “is in the past” or “on or after” adds a lot of flexibility to use a simple date range filter.
User(s) The Users filter auto loads the users associated with your group; it uses the user’s name. If you are not sure about the name of a particular user, you can find the list of users on the Organization page.
View Only Group User’s Data To view data for test run by Group level users (Group Owner, Group Admin and Group User), select this filter as “is” and then “yes”. Is any value will show data for everyone, across the group. Is no, will exclude Group Users (Group Owner, Group Admin and Group User).
Team(s) The Teams filter auto loads the teams associated with your group; you will find the list of the teams in the Organization section of your account.
Package Name View data related to your app by using the package name filter, very useful if you are interested in gauging the performance of your app or tests over a timeline.
App ID You can filter out the visualizations according to the App ID which is generated whenever you upload an app for testing. You will find App ID in the test session details or can find it via CLI.
Build Name The Build Name filter loads the names of the builds associated with your account. If your group uses a nomenclature standard for build names, you can use the operator option, for example, if all builds related to regression start with “Reg-”, the “Starts with” operator can be used.
Project The Project filter loads the names of the projects associated with your account. Generally, project capability is used to tag builds for different Engg Projects like App Ver. 2 or New Feature Releases
OS The OS filter allows you to dig into the visualizations and reports at an OS level, it will suggest the operating systems as per the tests that have been run on your account like OS X, Windows, Android or iOS.
OS Version The OS version filter allows you to dig into the visualizations and reports at an OS level, it will suggest the operating systems as per the tests that have been run on your account like OS X Catalina, Windows XP, etc.
Browser The browser filter lists all the browser families that you have run test sessions on, it will contain the browsers families like Chrome, Safari, etc
Browser Version The browser version gives you the option of slicing the data at Browser version level like Chrome 78, Chrome 79, etc.
Device Device filter allows you to filter the data for different devices that you have tested. Example, Samsung Galaxy S9, iPhone XS, iPhone 11, etc.
Session Name The session name filter will allow you to gauge the timeline of sessions which have been run frequently, for example canary tests, regression tests, or test on iPhone, etc. This filter is highly useful if the sessions and builds have been named properly. Another useful use case would be to check flackyness of a test over a timeline (Using Test Trends in Test Stability dashboard)

Operator types

  1. Date Related Operator Types: There are multiple date related operator types available for filtering data, example is in the past, is on the day

    date-related-operator-types

  2. Non-Date Related Operator Types: There are multiple non-date related operator types available for filtering data, example is equal to, contains, starts with

    Multiple options of operator types for filtering data

Advanced operator types

Matches (Advanced)

The filter operator has an option of “Matches (Advanced)”, you can utilize the same to get the most out of the test insights feature.

Example Description
FOO is equal to “FOO”, exactly
FOO,BAR is equal to either “FOO” or “BAR”, exactly
FOO% starts with “FOO”, matches “foolish” and “food” but not “buffoon” or “fast food”
%FOO ends with “FOO”, matches “buffoo” and “fast foo” but not “buffoon” or “fast food”
F%OD starts with an “F” and ends with “OD”, matches “fast food”
EMPTY string is empty (has zero characters) or is null (no value)
NULL value is null
-FOO is not equal to “FOO” (is any value except “FOO”), matches “pizza”, “trash”, “fun” but not “foo”
-FOO,-BAR is not equal to either “FOO” or “BAR”, matches any value except “FOO” and “BAR”
-%FOO% doesn’t contain “FOO”, does not match “buffoon” or “fast food”
-FOO% doesn’t start with “FOO”, does not match “foolish” or “food”
-%FOO doesn’t end with “FOO”, does not match “buffoo” or “fast foo”
-EMPTY string is not empty (has at least one character)
-NULL value of column is not null
FOO%,BAR starts with “FOO” or is “BAR” exactly, matches “food” and matches “bar” but not “barfood”
FOO%,-FOOD starts with “FOO” but is not “FOOD”
_UF has any single character followed by “UF”, matches “buffoon”

Schedule, Send and Download reports

Test Insights allows you to schedule periodic email, delivered directly to email box. Dashboard level email schedules can be set and also shared with your team members who may not be present on BrowserStack.

Reports can be set up as schedules, or send a one-time email via the Send option, available on the top-right menu. You can also download the dashboards in form of PDF or CSV, to share the reports via internal mediums. The schedules can be set up with different filters.

Options for Scheduling, Sending and Downloading Reports

Schedule reports

In order to set a schedule, click settings icon on top-right menu and click Schedule from the top-right menu. Click New to create a new schedule. You get the following options:
Image for creating a new schedule in test insights

  1. Give your schedule a name: The name that will be used as the subject line of the email. And will show up as the schedules name on the left menu.

  2. Who should it be emailed to: Add all the emails that should be sent to the email.Include a custom message that allows you to add a message in the schedules mail body.

  3. Format data as: Three options are available
    • PDF format
    • Visualization which sends an image of the dashboard
    • CSV ZIP File attaches CSV file of the data to the email
  4. Trigger: Select the repeating interval.

  5. Deliver this schedule: Select the frequency and delivery schedules.

  6. Filter: Apply all the filters, as per your need.

  7. Advanced Options: Select the settings for the following:
    • Time zone
    • Single column layout
    • Paper size
    • Formatted data values (for CSV)
      Advanced options

Send reports

Send option gives you the option of sharing the dashboard reports as a one time email. You get all the above options other than scheduling options.

Sending parallel utilization reports from dashboard

Download reports

Download via CSV or PDF option gives allows you to instantly download the full dashboard, and the file is shareable internally via your organisation’s preferred communication medium.

Test Insights - Report download option

Toggle reports

You can also choose to hide or show different data points in each report on the dashboard, simply by turning on or off the toggle. The corresponding data point in the report’s legend works as the toggle.

Test Insights - Graph Toggling option

Test performance status

Performance Breakdowns:

PASS​​​​​ED : The tests marked as Passed by the users.

FAIL​​​ED​​ : The tests marked as Failed by the users.

ERROR​​​​​ : The tests which were not marked pass or failed and there was an error. You can get more details on these in Test Stability section.

TIMEOUT​​​​​ : The tests which faced a timeout, such as a device did not receive the next test command.

UNMARKED*​​​​ : The test sessions that got completed successfully, but were not marked by the user.

IGNORED**​ : The tests that were skipped because of OS or test settings.

SKIPPED**​ : A test case that was never executed by the test runner.

Note: Click to get more details about test status for App Automate tests depending on the framework you are using : Appium, Espresso and XCUITest

*​​​​ Applicable only for Appium test sessions.
** Applicable only for Espresso/XCUI tests.

We're sorry to hear that. Please share your feedback so we can do better

Contact our Support team for immediate help while we work on improving our docs.

We're continuously improving our docs. We'd love to know what you liked





Thank you for your valuable feedback

Is this page helping you?

Yes
No

We're sorry to hear that. Please share your feedback so we can do better

Contact our Support team for immediate help while we work on improving our docs.

We're continuously improving our docs. We'd love to know what you liked





Thank you for your valuable feedback!

Talk to an Expert
Download Copy Check Circle