Skip to main content

Investigate failures

This document explains how to investigate failed tests root cause using TestOps analytics capabilities.

requirements
  • Access to Katalon TestOps with Viewer role or higher.
  • Failed tests in the concerned time frame/release/sprint.

Failure investigation workflow​

Follow these steps to investigate failures.

Step 1: Detect and prioritize failures​

In Executions List, navigate to a failed test execution. You will arrive at a "test results details page", where you typically investigate failure:

testops historical results trend line
tip
  • Use test quality indicators (smart tags) to help you prioritize:
  • Test runs with new failures or with a high pass rate. These are likely due to a recent code change.
  • Test runs with always failed tag.
  • If a test run has flaky tag, first investigate its flakiness status to avoid wasting efforts on troubleshooting. See Investigate Flaky Test for more details.

Step 2: Review test result details to investigate​

The Test Result Details page is the central investigation workspace.

testops test result details

Read failure message for context​

The failure message is your first clue about what went wrong. Below are some common failure messages:

  • "Element not found: #submitButton" (locator issue)
  • "Expected: 100, Actual: 99" (assertion failure)
  • "Timeout after 30 seconds" (timing issue)
  • "Connection refused" (network/environment issue)

Identifying these failure messages helps identify issues quickly and saves investigation time.

Triage with AI​

With full access to execution context, AI analysis will help you understand the failure at a quick glance.

Click Analyze with AI in the Test Result Details page. The agent automatically reviews details like logs, stack traces, screenshots, scripts, and result pattern to triage the failure and suggest an approach:

testops failure ai analyze button

AI analysis also acts as a quick failure documenting tool, to help you communicate with teammates/stakeholders more efficiently.

Analyze the failure/error manually​

Use AI's analysis to focus your manual review.

  1. View the full Execution Logs with info, warnings, errors, and stack traces.
  2. Analyze stack traces for exception messages, error location, framework calls... and cross-reference with AI. E.g., If AI mentions Broken Selector, keep an eye out for NoSuchElementException error.
  3. Check execution environment: If AI suggested Environment Issue, check environment details (browser, device, OS type, and execution profile) for any potential root cause.
  4. Review screenshots/videos: go back to the test run details page from the test results, and review images or videos for visual clues. E.g. wrong page loaded, elements overlaying, ...
  5. Verify test data and configuration: check dataset, inputs, staging/QA/production environment, authentication... to make sure testing data is as expected.

Step 3: Determine Root Cause and Take Action​

Classify the failure before assigning a teammate to take action. For example:

  • Product Defect: Application behaves incorrectly → Create defect, attach evidence. See Linkages to learn how to create and auto-attach defect to a failed result.
  • Test Issue:
  • Faulty assertions, locators, or logic → Assign teammate to fix the test and re-run.
  • Tests with flaky tag → Follow the Investigate Flaky Tests workflow and if the test is truly unstable (), assign a teammate to review.
  • Environment Issue: Infrastructure, configuration, or network failures → Escalate to DevOps / infra team

Step 4: Document findings for future reference​

Documenting your investigation helps you validate resolutions and quickly traces back in case of future recurrence.

Within TestOps, you can configure customizable fields and add details to the run in the test result details page. For example:

  • Configure an "issue type" fields with values env-issue, data_issue,...
  • Configure a known-flaky tag to remind teammates the test used to be flaky in the past
Custom Fields showing on UI - TC

Besides these fields, you can also:

  • Update ALM defect ticket or issue tracker
  • Share findings via Slack or internal communication channels
  • Use AI summary to share findings
Was this page helpful?