Use Exploratory Testing
Learn how to use AI-powered exploratory testing to autonomously discover bugs and edge cases in your application.
What Is Exploratory Testing?
Exploratory testing lets an AI agent autonomously:
- Discover pages — Navigate and find all pages in your app
- Test interactions — Try buttons, forms, links, flows
- Find bugs — Identify broken functionality
- Analyze failures — Categorize why tests failed
- Generate report — Provide actionable bug findings
Key difference from scripted tests:
- ✅ Exploratory: “Explore the app and find bugs”
- ❌ Scripted: “Execute these exact steps”
Running Your First Session
Session Configuration
Basic Settings
Session Name: Any descriptive text
✅ "Checkout Flow Testing"
✅ "Mobile App Smoke Test"
✅ "Auth Edge Cases"Target URL: Full URL where app is hosted
✅ https://example.com
✅ https://staging-app.example.com
✅ https://app.example.com/dashboardProject: Which project to associate session with
Advanced Settings
Persona Selection
-
Guest (default) — Unauthenticated user
- Tests public pages
- Good for landing pages, public docs, auth flows
-
With Persona — Logged-in user
- Tests authenticated pages
- Tests user-specific features
- Requires persona created in advance
Duration: 5-120 Minutes
| Duration | Use Case | Coverage |
|---|---|---|
| 5-10 min | Quick smoke test | Main flows only |
| 20-30 min | Standard testing | Most flows, basic coverage |
| 45-60 min | Deep exploration | All major flows + edge cases |
| 90-120 min | Comprehensive audit | Complete coverage + variations |
Risk Focus Area
Focus AI testing on specific area:
- All — No specific focus, test everything equally
- Authentication — Login, 2FA, session management, account flows
- Payment — Checkout, billing, payment methods, subscriptions
- User Data — Profile, settings, data deletion, preferences
Monitoring Session Progress
Session Phases
Your session goes through 3 phases:
Scouting (20%) → Discovering pages
↓
Exploring (60%) → Testing interactions
↓
Reporting (20%) → Analyzing findings
↓
Completed → Results readyStatus indicators:
- 🟢 Running — Session in progress
- 🟡 Paused — Session paused (click Resume)
- ✅ Completed — Session finished
- ❌ Failed — Error occurred
Live Metrics
Watch in real-time as session runs:
- Pages Discovered: Count of unique pages found
- Interactions Tested: Number of actions attempted
- Bugs Found: Live count increasing
- Current Focus: What the AI is testing right now
Example progression:
Pages: 5 → 12 → 23 → 35
Bugs: 0 → 2 → 5 → 8
Progress: 20% → 45% → 80% → 100%Reviewing Results
After Session Completes
AI-Generated Report
Natural language summary of findings:
"BugBrain discovered 12 bugs during exploration:
🔴 CRITICAL (2):
- Payment form doesn't validate card numbers
- Session expires while filling checkout form
🟠 HIGH (3):
- Search results page missing error handling
- Mobile menu doesn't close after clicking
🟡 MEDIUM (4):
- Form labels not properly centered
- Pagination buttons have low contrast
🟢 LOW (3):
- Typo on products page
- Broken favicon linkBug List
Sortable list of all detected bugs:
Sort by:
- Severity (critical → low)
- Category (selector, timeout, assertion, etc.)
- Page (group by location)
- Confidence (AI’s certainty)
Each bug shows:
- Title and description
- Steps to reproduce
- Expected vs actual behavior
- Screenshot/evidence
- AI-suggested fix
Using Findings
For test case generation:
- Review bug list
- Create test cases for high-severity bugs
- Add to test plan
- Run regularly to catch regressions
For priority queue:
- Export bugs as issues (GitHub/Jira)
- Assign to team
- Schedule in sprint
- Track fixes
For quality metrics:
- Compare sessions over time
- Track bug trends (increasing/decreasing)
- Measure improvement after fixes
- Report to stakeholders
Advanced Usage
Running Sessions Regularly
Schedule weekly exploratory sessions:
- Create session → Save as template
- Day & time → Off-peak hours (e.g., 2 AM)
- Frequency → Weekly or daily
- Notifications → Email on new critical bugs
Benefits:
- Continuous bug detection
- Catch regressions after updates
- Maintain quality baseline
Using Knowledge Maps
After 2-3 sessions on same URL, Knowledge Map reaches 75%+ confidence:
- AI understands app structure better
- Finds more relevant bugs
- Less duplicate findings
- Better recommendations
To leverage knowledge:
- Run multiple sessions (3-4 weeks)
- Wait for high confidence score
- Auto-generate test cases from findings
- Test cases will be much better
Comparing Sessions
Track bugs over time:
Session 1: 15 bugs
Session 2: 12 bugs (improvement ✓)
Session 3: 8 bugs (more improvement ✓)
Session 4: 7 bugs (stabilized, good quality)Use trend charts to show QA progress to stakeholders.
Parallel Sessions
Run multiple simultaneous sessions:
Session 1: Desktop app testing
Session 2: Mobile app testing
Session 3: API edge cases
Session 4: Authentication flowsAll run in parallel, results aggregated.
Troubleshooting
”Session Failed: Target URL Unreachable”
- Verify URL is live
- Check firewall/IP allowlisting
- Ensure app isn’t behind authentication
- Wait for app to restart if needed
”Session Timeout: Max Time Exceeded”
- Session ran for full duration
- Results are still available even if incomplete
- Increase duration for next session if needed
”No Bugs Found”
Possible reasons:
- App is very robust (good news!)
- AI didn’t find edge cases
- Run longer session or with specific persona
- Try different risk focus area
”Too Many False Positives”
- Bugs detected that aren’t really bugs
- Review findings with team
- False positives help identify brittle tests
- Use findings to improve test robustness
Best Practices
✅ DO:
- Run sessions on staging environment
- Disable notifications temporarily (avoid alert spam)
- Start with short sessions (10-20 min) for quick validation
- Run weekly or bi-weekly for continuous discovery
- Review critical bugs within 24 hours
- Create test cases from high/critical bugs
❌ DON’T:
- Run on production (could trigger monitors/alerts)
- Use production user accounts
- Expect 100% coverage
- Ignore exploratory findings
- Set unrealistic expectations (not a replacement for manual QA)
Example: Full Workflow
Week 1:
Monday → Run exploratory session (2 hour)
→ Find 8 bugs
→ Create 3 test cases for critical bugs
→ Assign to dev team
Week 2:
Monday → Run second session
→ Compare to week 1
→ Verify critical bugs fixed
→ Find 5 new edge cases
→ Add to backlog
Week 3:
Monday → Run third session
→ Knowledge map now 85% confident
→ Auto-generate 10 test cases
→ Add to regression test plan
Week 4:
Monday → Run fourth session
→ Bug count stabilized at 4-5
→ Quality improved significantly
→ Report trend to stakeholdersPro Tip: Exploratory testing complements scripted tests perfectly. Use scripted tests for regression prevention, exploratory for new bug discovery.