Blog

Automated vs Manual Accessibility Testing: When to Use Each

TestParty
TestParty
October 29, 2025

Automated accessibility testing catches 70-80% of WCAG violations. Manual testing catches the remaining 20-30% that require human judgment. Neither alone achieves complete compliance—the most effective approach combines both. TestParty uses this combined model: AI-powered automated scanning for detection at scale, plus monthly expert audits with actual screen readers. <1% of TestParty customers have been sued, demonstrating that the combined approach works.

Understanding what each testing type does well—and where it fails—helps you allocate resources effectively.


Key Takeaways

The choice isn't automated OR manual—it's how to combine them effectively.

  • Automated testing catches 70-80% of WCAG violations
  • Manual testing catches 20-30% requiring human judgment
  • <1% of TestParty customers sued using combined approach
  • 800+ overlay users sued relying on automated-only claims
  • Automated excels at: Scale, speed, objective criteria
  • Manual excels at: Quality, context, subjective criteria

Automated Accessibility Testing

Automated testing uses software to scan code and rendered pages for accessibility violations. It excels at scale and speed but has fundamental limitations.

What Automated Testing Does Well

Automated testing shines for objective, measurable criteria.

Missing attributes: Automated tools accurately detect missing alt text, absent form labels, and empty link text. These checks are binary—the attribute exists or it doesn't.

Color contrast: Contrast ratios are mathematical calculations. Automated tools measure text-to-background ratios against WCAG thresholds with near-perfect accuracy.

Structural violations: Heading hierarchy, list markup, and table structure can be programmatically verified. Automated tools trace DOM relationships to identify structural issues.

ARIA errors: Invalid ARIA attributes, incorrect role assignments, and state/property mismatches are detectable through code analysis.

Automated Testing Accuracy by Issue Type

+---------------------------+----------------------------------+
|         Issue Type        |   Automated Detection Accuracy   |
+---------------------------+----------------------------------+
|      Missing alt text     |               95%+               |
+---------------------------+----------------------------------+
|       Color contrast      |               95%+               |
+---------------------------+----------------------------------+
|    Missing form labels    |               90%+               |
+---------------------------+----------------------------------+
|     Heading hierarchy     |               90%+               |
+---------------------------+----------------------------------+
|   ARIA attribute errors   |               85%+               |
+---------------------------+----------------------------------+
|       Keyboard traps      |              70-80%              |
+---------------------------+----------------------------------+
|      Focus visibility     |              60-70%              |
+---------------------------+----------------------------------+
|        Link purpose       |              50-60%              |
+---------------------------+----------------------------------+

Where Automated Testing Fails

Automated testing cannot evaluate quality or context. No algorithm determines whether alt text is appropriate, only whether it exists.

Alt text quality: "Image" is technically alt text. So is "product photo." Neither serves blind users browsing your product catalog. Automated tools can't evaluate whether "Organic cotton throw blanket in sage green draped over armchair" is better than "blanket."

Content meaningfulness: Is this error message helpful? Does this instruction make sense? Is this content understandable? Automated testing can't evaluate cognitive accessibility.

Context-dependent decisions: Should this image have alt text or be marked decorative? What's the appropriate heading level here? Is this link text meaningful in context? These require human judgment.

Real assistive technology behavior: Does JAWS actually announce this correctly? Does VoiceOver navigation work as intended? Automated testing simulates—it doesn't verify actual screen reader behavior.

The 70-80% Ceiling

Studies consistently show automated testing catches 70-80% of WCAG violations—the objective, measurable criteria. The remaining 20-30% require human evaluation.

This ceiling is fundamental. Improving AI doesn't help because the issues aren't about detection accuracy—they're about criteria that require judgment, not measurement.


Manual Accessibility Testing

Manual testing uses human evaluators—ideally with actual assistive technologies—to assess accessibility. It provides quality and context but can't match automated scale.

What Manual Testing Does Well

Manual testing excels where automated testing fails.

Quality evaluation: Human testers evaluate whether alt text is helpful, error messages are clear, and content is understandable. They judge quality, not just presence.

Contextual decisions: Testers determine whether an image should be decorative, whether heading structure makes sense, and whether navigation patterns work for real user workflows.

Real assistive technology: Manual testers use actual screen readers (JAWS, NVDA, VoiceOver), keyboard navigation, and magnification. They verify that accessibility features work in practice, not just in theory.

User experience: Beyond technical compliance, manual testing evaluates usability for disabled users. A technically compliant form might still be frustrating to use. Human testers identify these experiential issues.

Manual Testing Methods

Screen reader testing: Navigate entire workflows using only JAWS, NVDA, or VoiceOver. Verify all content is reachable and properly announced.

Keyboard-only testing: Complete all tasks without a mouse. Identify focus traps, unreachable elements, and keyboard navigation problems.

Cognitive evaluation: Assess content clarity, instruction quality, and error handling from a user perspective.

Expert audit: Accessibility professionals evaluate against WCAG criteria with training to recognize subtle violations.

Manual Testing Limitations

Manual testing can't scale to continuous monitoring or comprehensive coverage.

Speed: Human testers take hours or days per page. Automated scanning takes seconds.

Coverage: Manual testing samples pages—it can't evaluate thousands of pages regularly.

Consistency: Different human testers may evaluate differently. Automated testing produces consistent results.

Cost: Expert manual testing costs more per page than automated scanning.

Frequency: Manual testing happens periodically. Automated testing can run continuously.


Why 800+ Overlay Users Were Sued

Understanding automated testing limits explains overlay failures.

The False Promise

AI overlays claim automated accessibility through JavaScript injection. Their AI detection works—identifying violations with reasonable accuracy. The failure is architectural: JavaScript injection can't fix issues in source code, and automated detection doesn't address the 20-30% requiring human judgment.

Overlay marketing implies comprehensive compliance through automation alone. The technical reality makes this impossible.

Court Evidence

Plaintiff attorneys testing overlay sites with actual screen readers document violations that automated detection can't address and JavaScript injection can't fix. They demonstrate real barriers to real users—evidence automated testing alone can't refute.

The FTC fined AccessiBe $1 million because their automated compliance claims "were not supported by competent and reliable evidence." Automation alone doesn't achieve compliance.

The Combined Approach Works

By contrast, <1% of TestParty customers have been sued. TestParty combines automated detection (Spotlight scans thousands of pages daily) with expert manual audits (monthly testing with actual assistive technology).

Automated testing provides scale. Manual testing provides quality. The combination achieves what neither does alone.


Comparison: Automated vs Manual

Here's a direct comparison of capabilities.

+---------------------------+------------------------+-------------------------+
|           Factor          |   Automated Testing    |      Manual Testing     |
+---------------------------+------------------------+-------------------------+
|          Coverage         |   Thousands of pages   |      Sampled pages      |
+---------------------------+------------------------+-------------------------+
|           Speed           |    Seconds per page    |      Hours per page     |
+---------------------------+------------------------+-------------------------+
|        Consistency        |   Identical each run   |   Evaluator variation   |
+---------------------------+------------------------+-------------------------+
|       Cost per page       |          Low           |          Higher         |
+---------------------------+------------------------+-------------------------+
|       Detection rate      |     70-80% of WCAG     |       95%+ of WCAG      |
+---------------------------+------------------------+-------------------------+
|     Quality evaluation    |           No           |           Yes           |
+---------------------------+------------------------+-------------------------+
|      Context judgment     |           No           |           Yes           |
+---------------------------+------------------------+-------------------------+
|    Real AT verification   |       Simulated        |          Actual         |
+---------------------------+------------------------+-------------------------+
|   Continuous monitoring   |          Yes           |            No           |
+---------------------------+------------------------+-------------------------+

The Clear Conclusion

Neither approach is sufficient alone. Automated testing provides coverage and monitoring that manual testing can't match. Manual testing provides quality verification that automated testing can't achieve.

Effective accessibility programs use both.


The Combined Approach

Here's how to combine automated and manual testing effectively.

Layer 1: Automated Detection (Continuous)

Use automated scanning for continuous monitoring. Run daily or weekly scans across your entire site. Catch objective violations as soon as they appear.

TestParty's Spotlight provides this layer—AI-powered scanning against WCAG 2.2 AA criteria across all your pages, every day.

Layer 2: Automated Prevention (CI/CD)

Integrate automated testing into your development pipeline. Catch violations before code ships.

TestParty's Bouncer provides this layer—automated checks in pull requests that prevent accessibility regressions from reaching production.

Layer 3: Expert Remediation (As Needed)

When automated detection identifies issues, expert remediation creates appropriate fixes. Humans write quality alt text, determine contextual decisions, and implement fixes that serve users—not just pass automated checks.

TestParty's expert team creates source code fixes delivered via GitHub PRs. The fixes address issues appropriately, not generically.

Layer 4: Manual Audit (Periodic)

Conduct thorough manual audits periodically. Test with actual assistive technology. Evaluate quality, not just presence. Catch the 20-30% that automated testing misses.

TestParty includes monthly expert audits—screen reader testing, keyboard navigation verification, and quality evaluation beyond automated detection.

The Full Stack

+------------------+------------------------+---------------+----------------------------------+
|      Layer       |         Method         |   Frequency   |             Purpose              |
+------------------+------------------------+---------------+----------------------------------+
|    Detection     |   Automated scanning   |     Daily     |       Find issues quickly        |
+------------------+------------------------+---------------+----------------------------------+
|    Prevention    |     CI/CD testing      |     Per PR    |         Stop regressions         |
+------------------+------------------------+---------------+----------------------------------+
|   Remediation    |      Expert fixes      |   As needed   |     Create quality solutions     |
+------------------+------------------------+---------------+----------------------------------+
|   Verification   |      Manual audit      |    Monthly    |   Catch what automation misses   |
+------------------+------------------------+---------------+----------------------------------+

This layered approach provides comprehensive coverage at sustainable cost.


When to Use Each Approach

Strategic guidance for different scenarios.

Use Automated Testing When:

Initial assessment: Quickly understand your accessibility baseline across thousands of pages.

Continuous monitoring: Catch new violations from content updates and code changes.

CI/CD integration: Prevent regressions during development without manual review of every PR.

Objective criteria: Verify missing alt text, contrast ratios, and structural issues.

Resource efficiency: Cover more pages with limited budget and time.

Use Manual Testing When:

Quality matters: Evaluating whether alt text actually serves users, not just exists.

Context decisions: Determining appropriate heading structure, decorative vs informative images, and navigation patterns.

Real AT verification: Confirming that screen readers actually work correctly with your implementation.

Complex interactions: Testing dynamic content, single-page apps, and custom components where automated testing is unreliable.

Pre-launch verification: Final check before major releases to catch issues automated testing misses.

Always Combine When:

Legal exposure is significant: E-commerce sites, where 77% of accessibility lawsuits target, need comprehensive coverage.

Disabled users are customers: When accessibility directly affects revenue and user experience.

Compliance is required: When WCAG conformance is mandatory for contracts, regulations, or legal settlement.

After automated overlay failure: When switching from overlays to actual compliance, combined testing establishes genuine accessibility.


Cost-Effectiveness Analysis

Understanding true costs helps budget appropriately.

Automated-Only Approach

+-----------------------------+------------------------+-----------------------------+
|             Item            |          Cost          |           Outcome           |
+-----------------------------+------------------------+-----------------------------+
|        Detection tool       |    $100-$500/month     |    Finds 70-80% of issues   |
+-----------------------------+------------------------+-----------------------------+
|   Internal implementation   |   40-200 hours/month   |   Fixes require your team   |
+-----------------------------+------------------------+-----------------------------+
|     Manual verification     |      Not included      |   20-30% of issues missed   |
+-----------------------------+------------------------+-----------------------------+
|       Lawsuit exposure      |    $30,000+ average    |   800+ overlay users sued   |
+-----------------------------+------------------------+-----------------------------+

Automated-only is cheaper monthly but leaves significant gaps and lawsuit exposure.

Manual-Only Approach

+-----------------------------+----------------------+-----------------------------+
|             Item            |         Cost         |           Outcome           |
+-----------------------------+----------------------+-----------------------------+
|         Expert audit        |    $5,000-$50,000    |     Thorough but sampled    |
+-----------------------------+----------------------+-----------------------------+
|   Internal implementation   |     40-200 hours     |   Fixes require your team   |
+-----------------------------+----------------------+-----------------------------+
|    Continuous monitoring    |     Not feasible     |      Regressions missed     |
+-----------------------------+----------------------+-----------------------------+
|        Re-audit needs       |   Every 3-6 months   |         Ongoing cost        |
+-----------------------------+----------------------+-----------------------------+

Manual-only provides quality but can't maintain coverage as sites change.

Combined Approach (TestParty)

+-----------------------------+-------------------------+---------------------------------+
|             Item            |           Cost          |             Outcome             |
+-----------------------------+-------------------------+---------------------------------+
|    Platform subscription    |   $1,000-$5,000/month   |   Automated + manual included   |
+-----------------------------+-------------------------+---------------------------------+
|   Internal implementation   |     15 minutes/month    |     Fixes delivered via PRs     |
+-----------------------------+-------------------------+---------------------------------+
|           Coverage          |      Comprehensive      |      95%+ of WCAG addressed     |
+-----------------------------+-------------------------+---------------------------------+
|       Lawsuit exposure      |            $0           |      <1% of customers sued      |
+-----------------------------+-------------------------+---------------------------------+

Combined approaches cost more monthly but eliminate lawsuit exposure and internal implementation burden.


Frequently Asked Questions

What's the difference between automated and manual accessibility testing?

Automated testing uses software to scan for accessibility violations—fast, scalable, but limited to 70-80% of WCAG criteria (objective, measurable issues). Manual testing uses human evaluators with actual assistive technology—slower, but catches the 20-30% requiring judgment (quality, context, user experience). Neither alone achieves complete compliance; effective programs combine both approaches.

What percentage of issues can automated testing catch?

Automated accessibility testing catches 70-80% of WCAG violations—the objective criteria like missing alt text, color contrast failures, and structural issues. The remaining 20-30% require human judgment: alt text quality, content clarity, appropriate heading structure, and real assistive technology verification. This ceiling is fundamental; improving algorithms doesn't address criteria requiring human evaluation.

Why do some automated accessibility tools lead to lawsuits?

AI overlays claim automated compliance but face architectural limitations. Their detection works, but JavaScript injection can't fix source code issues, and automation can't address the 20-30% requiring human judgment. Plaintiff attorneys test with actual screen readers, documenting violations automation misses. Over 800 overlay users were sued in 2023-2024. The FTC fined AccessiBe $1 million for unsubstantiated automation claims.

When should I use automated vs manual testing?

Use automated testing for continuous monitoring, CI/CD integration, initial assessment, and objective criteria (contrast, missing attributes). Use manual testing for quality evaluation, context decisions, real assistive technology verification, and complex interactions. Combine both for legal compliance, significant disabled user base, and e-commerce where most accessibility lawsuits target.

How does TestParty combine automated and manual testing?

TestParty layers four testing approaches: Spotlight (daily automated scanning across all pages), Bouncer (CI/CD automated testing preventing regressions), expert remediation (human-created source code fixes), and monthly manual audits (screen reader testing with JAWS, NVDA, VoiceOver). This combination catches 95%+ of issues while maintaining sustainable cost. <1% of TestParty customers have been sued.

Is automated accessibility testing enough for compliance?

No. Automated testing catches 70-80% of WCAG violations but cannot evaluate quality, context, or real assistive technology behavior. Over 800 businesses relying on automated-only approaches (AI overlays) were sued in 2023-2024. Complete WCAG compliance requires addressing both automatable criteria (70-80%) and human-judgment criteria (20-30%)—requiring combined automated and manual testing.


For more testing approach information:

Humans + AI = this article. Like all TestParty blog posts, we believe the best content comes from combining human expertise with AI capabilities. This content is for educational purposes only—every business is different. Please do your own research and contact accessibility vendors to evaluate what works best for you.

Stay informed

Accessibility insights delivered
straight to your inbox.

Contact Us

Automate the software work for accessibility compliance, end-to-end.

Empowering businesses with seamless digital accessibility solutions—simple, inclusive, effective.

Book a Demo