Why Accessibility Audits Don't Age Well (And What to Do About It)
TABLE OF CONTENTS
- Key Takeaways
- What a Manual Audit Is (And Isn't)
- The Decay Problem: Why Audits Drift Toward Irrelevance
- The "Audit Half-Life" Concept
- Why Audits Are Still Valuable
- The Common Failure Pattern
- The Better Pattern: Audit as Calibration
- Practical Operating Model: Audit + Verify
- Evidence and Defensibility
- Standards Support the Continuous Model
- FAQ
- Related Resources
A manual audit is a photograph; accessibility is a movie. The audit captures state at a specific moment—the pages tested, the components reviewed, the assistive technology configurations used. But software changes continuously. Features ship. Content updates. Third parties release new versions. A/B tests introduce variants. The moment the audit is complete, decay begins.
Audits don't "fail"—teams fail when they treat audits as a substitute for operational controls. An audit is a validation tool: it confirms current state and identifies issues for remediation. It's not a prevention mechanism, not a monitoring system, and not a guarantee of future compliance. Organizations that treat an annual audit as "accessibility coverage" discover this when the next audit finds issues that look suspiciously like the last one's findings.
Without CI gates and monitoring, audit findings return like seasonal allergies. You remediate the issues found in March. By September, similar issues have appeared in new code, new content, and new features. The next audit finds them again. WebAIM's 2024 Million report documents this cycle at scale: 95.9% of home pages fail accessibility checks, with the same issue types appearing year after year—low contrast, missing alt text, unlabeled forms. Audits identify these issues; audits don't prevent them.
Key Takeaways
Understanding audit limitations helps organizations use them appropriately while building systems that maintain accessibility between audits.
- Audits are point-in-time validations – They capture state; they don't maintain it; software changes faster than audit cycles
- The "audit half-life" determines decay – The faster you ship, the shorter the time until your audited state becomes outdated
- Audits remain valuable for specific purposes – Catching complex issues automation misses, training teams, creating remediation roadmaps, validating AT usability
- Backlogs without ownership become permanent – Audit findings that enter a generic backlog without code attribution and ownership rarely get fixed
- The solution is audit calibration plus operational controls – Use audits to validate that CI/CD and monitoring work; don't rely on audits alone
What a Manual Audit Is (And Isn't)
Clear definitions prevent misuse of audits.
What It Is
A manual accessibility audit is a structured evaluation of digital content against accessibility criteria (typically WCAG), performed by trained accessibility professionals using assistive technology and expert review. Audits typically include:
+------------------------+----------------------------------------------------+
| Component | What It Covers |
+------------------------+----------------------------------------------------+
| Automated scanning | Machine-detectable issues as baseline |
+------------------------+----------------------------------------------------+
| Manual testing | Keyboard navigation, screen reader verification |
+------------------------+----------------------------------------------------+
| Expert review | Judgment-based evaluation of patterns and UX |
+------------------------+----------------------------------------------------+
| Documentation | Findings, severity ratings, remediation guidance |
+------------------------+----------------------------------------------------+Good audits provide detailed, actionable findings with screenshots, code examples, and specific remediation recommendations.
What It Isn't
An audit is not:
- A guarantee of accessibility – It reflects state at test time only
- A prevention mechanism – It can't stop new issues from appearing
- A monitoring system – It doesn't track changes over time
- Legal immunity – Demonstrating effort, not guaranteeing protection
- A substitute for engineering practices – Detection, not prevention
Organizations often purchase audits expecting ongoing accessibility. What they receive is a snapshot that begins decaying immediately.
The Decay Problem: Why Audits Drift Toward Irrelevance
The gap between audit and reality widens every day.
Releases Keep Shipping
Modern software development is continuous:
- Weekly or daily deployments
- Feature flags enabling new functionality
- Bug fixes changing behavior
- Dependency updates affecting components
Each release is an opportunity to introduce accessibility issues. A modal component works correctly when audited; a developer modifies it three weeks later; the modification breaks focus management. The audit doesn't know.
Content Keeps Changing
Content changes happen outside engineering workflows:
- Marketing uploads new banners
- Product managers add documentation
- Content editors publish articles
- Customer support updates help pages
These changes don't trigger code review. They don't go through CI. A content editor adds an infographic without alt text; the audit completed last month has no visibility.
Third Parties Keep Updating
Embedded third-party services update on their own schedules:
- Chat widget vendors release new versions
- Payment processors change their forms
- Analytics tools modify their overlays
- Social embeds get redesigned
You don't control these updates. An audit validated your checkout flow with Stripe Elements v1; Stripe released v2 last week with different accessibility characteristics.
Small Changes Compound
Individually, each change is minor. Collectively, they transform the site:
+----------------------+-------------------------------------+
| Time Since Audit | Typical State Change |
+----------------------+-------------------------------------+
| 1 week | Minor drift, likely similar |
+----------------------+-------------------------------------+
| 1 month | Notable changes in active areas |
+----------------------+-------------------------------------+
| 3 months | Significant divergence possible |
+----------------------+-------------------------------------+
| 6 months | Audit reflects historical state |
+----------------------+-------------------------------------+
| 12 months | Audit is archaeology |
+----------------------+-------------------------------------+The audit that cost $50,000 is describing a codebase that no longer exists.
The "Audit Half-Life" Concept
A useful mental model: the time until half your audited surfaces have materially changed.
Calculating Half-Life
Half-life depends on:
- Release frequency: Daily deploys = short half-life
- Content velocity: High-volume publishing = short half-life
- Surface area: More pages = more opportunity for change
- Third-party count: More dependencies = more external changes
+-------------------------------------+-----------------------+
| Organization Type | Typical Half-Life |
+-------------------------------------+-----------------------+
| Static marketing site | 6-12 months |
+-------------------------------------+-----------------------+
| E-commerce with active catalog | 1-3 months |
+-------------------------------------+-----------------------+
| SaaS with continuous deployment | 2-6 weeks |
+-------------------------------------+-----------------------+
| High-velocity content publisher | 1-4 weeks |
+-------------------------------------+-----------------------+Implications for Audit Frequency
If your half-life is two months, an annual audit describes a state that has completely turned over six times since testing. Even quarterly audits can't keep pace with weekly deploys.
The math doesn't support audit-only strategies for organizations with meaningful release velocity. You need operational controls (CI/CD, monitoring) that work continuously, with audits as periodic calibration.
Why Audits Are Still Valuable
Audits have real value when used appropriately.
Catching Complex Issues Automation Misses
Automated testing catches 30-40% of WCAG issues. Audits catch what automation misses:
- Complex interaction patterns
- Screen reader user experience quality
- Cognitive accessibility (clarity, predictability)
- Custom widget usability
- Error recovery workflows
- Timeout and session management
These require human judgment that no automated tool provides.
Training and Awareness
Audit findings educate teams:
- Specific examples of what fails and why
- Before/after comparisons when remediated
- Pattern recognition for future development
- Understanding of real AT user experience
Teams that review audit findings with the auditor learn more than teams that just receive a report.
Creating Remediation Roadmaps
Good audits provide actionable prioritization:
- Severity ratings (critical, high, medium, low)
- Affected user populations
- Remediation complexity estimates
- Quick wins vs. architectural fixes
This helps teams allocate limited resources effectively.
Validating AT Usability
Audits with real AT testing verify what automation can't measure:
- Is the screen reader experience actually usable?
- Can a keyboard user complete critical tasks?
- Are error messages understandable?
- Is the cognitive load reasonable?
This qualitative assessment is essential for actual accessibility, not just technical compliance.
Baseline Establishment
For organizations starting accessibility programs, audits establish baseline:
- Current state documentation
- Priority areas identification
- Comparison point for future measurement
You can't improve what you haven't measured.
The Common Failure Pattern
Audits fail when they become the entire accessibility strategy.
Findings Arrive as PDFs
The auditor delivers a PDF or slide deck. It contains detailed findings with screenshots and recommendations. It goes to... someone. Usually compliance or a project manager.
Tickets Are Created Without Code Attribution
Findings become tickets: "Missing alt text on product images." The ticket doesn't specify which component generates product images, which templates use that component, or what code produces the issue.
Teams Argue About Interpretation
Developers receive tickets they didn't create for issues they don't understand. Debates ensue:
- "Is this really a WCAG failure?"
- "The auditor must be wrong"
- "We can't fix this without redesigning the feature"
- "That's a third-party component; we can't change it"
Backlog Grows Faster Than Remediation
While teams debate, new features ship. New content publishes. The backlog grows. Some tickets get fixed; more tickets enter. Net issue count stays flat or increases.
The Next Audit Looks Familiar
Twelve months later, the next audit finds:
- Some original issues fixed
- Similar issues in new code
- Regression of some previously-fixed issues
- Third-party issues unchanged
The cycle repeats.
The Better Pattern: Audit as Calibration
Audits work best as calibration for operational systems, not as primary enforcement.
Audits Validate That Automation Works
Use audit findings to check your automated tools:
- Did CI catch the issues the audit found?
- If not, are your rules configured correctly?
- Are there issue categories you're not testing for?
Audit findings that CI missed indicate CI gaps to address.
Audits Identify Component Patterns
Audit findings should flow to components:
- "Multiple forms have missing labels" → Fix the form input component
- "Modal focus management fails" → Fix the modal component
- "Icon buttons lack names" → Fix the icon button component
Pattern-to-component mapping turns page-level findings into system-level fixes.
Audits Prioritize Journey Remediation
Use audits to prioritize journey investment:
- Critical journeys with many issues → Immediate attention
- Secondary journeys with issues → Scheduled remediation
- Low-traffic areas → Lower priority
Audits Inform Lint Rules
Audit findings suggest lint rules:
- Repeated "missing label" findings → Add jsx-a11y label rule
- Repeated "click without keyboard" → Add click-events-have-key-events rule
- Repeated "invalid ARIA" → Add aria-proptypes rule
Each lint rule prevents future occurrences of the pattern.
Audits Train Teams
Schedule audit walkthroughs:
- Review findings with the team that owns the code
- Discuss why patterns fail and how to fix them
- Demonstrate AT testing techniques
- Build team capability for ongoing accessibility
Practical Operating Model: Audit + Verify
The sustainable model combines periodic audits with continuous verification.
Audit Cadence
+----------------------------------------+-------------------------------+
| Organization Type | Recommended Audit Cadence |
+----------------------------------------+-------------------------------+
| Static site, low change rate | Annually |
+----------------------------------------+-------------------------------+
| Moderate change rate | Semi-annually |
+----------------------------------------+-------------------------------+
| High velocity, critical compliance | Quarterly |
+----------------------------------------+-------------------------------+
| Post-major-redesign | On-demand |
+----------------------------------------+-------------------------------+Continuous Verification
Between audits:
- CI checks on every PR
- Production monitoring weekly
- AT spot-checks monthly on critical journeys
- Regression testing on major releases
The Verification Loop
- Audit: Establish state, identify issues
- Remediate: Fix issues in source code
- Prevent: Add lint rules and tests
- Monitor: Watch for drift and regression
- Re-audit: Validate that the system works
The audit validates the verification system; it doesn't replace it.
Evidence and Defensibility
Audits contribute to defensibility, but evidence requirements extend beyond audit reports.
What Audits Provide
- Point-in-time validation by qualified third party
- Detailed documentation of testing methodology
- Expert attestation of compliance level
- Baseline for remediation tracking
What Audits Don't Provide
- Proof of ongoing compliance
- Evidence of remediation (only the issues, not the fixes)
- Demonstration of process (only the outcome at test time)
- Protection against future issues
Building Complete Evidence
DOJ guidance emphasizes that organizations must ensure accessibility of online services. Complete evidence includes:
+-------------------------+---------------------------------+
| Evidence Type | Source |
+-------------------------+---------------------------------+
| Policy commitment | Internal documentation |
+-------------------------+---------------------------------+
| Testing process | CI logs, monitoring reports |
+-------------------------+---------------------------------+
| Remediation records | Git commits, ticket history |
+-------------------------+---------------------------------+
| Training evidence | Completion records |
+-------------------------+---------------------------------+
| Audit validation | Third-party audit reports |
+-------------------------+---------------------------------+
| Ongoing monitoring | Production scan results |
+-------------------------+---------------------------------+Audits are one component of a complete evidence package.
Standards Support the Continuous Model
W3C and regulatory guidance align with continuous rather than periodic approaches.
W3C Guidance
W3C Planning and Managing Web Accessibility recommends integrating accessibility throughout the process and repeating activities over time. The guidance explicitly positions accessibility as ongoing, not one-time.
W3C Understanding Conformance notes that testing involves a combination of automated testing and human evaluation—supporting mixed methods, not audit-only.
Regulatory Framing
Section 508 emphasizes testing as part of the development lifecycle, not as periodic assessment. The framing is "lifecycle testing" rather than "annual audit."
Industry Practice
Organizations with mature accessibility programs report:
- Decreasing reliance on audits as primary detection
- Increasing investment in CI/CD and monitoring
- Using audits for validation and expert AT testing
- Treating audits as calibration, not coverage
FAQ
How often should we audit?
Depends on change rate and risk tolerance. High-velocity organizations with significant compliance exposure might audit quarterly. Stable organizations with lower risk might audit annually. The key insight: audit frequency should match your ability to maintain state between audits. If state degrades significantly before the next audit, you need operational controls, not more frequent audits.
Are audits worth the cost?
Yes, for the right purposes. Audits catch complex issues automation misses. They provide expert perspective. They create baselines and validate improvement. They're not worth the cost if used as the entire accessibility strategy—that's buying photographs of a moving target. Value audits appropriately and complement with operational controls.
Can automated tools replace audits?
No. Automated tools catch 30-40% of WCAG issues. The rest require human judgment: evaluating alt text quality, testing complex interactions, assessing cognitive accessibility, verifying real AT usability. Use automation for continuous coverage of what it can detect; use audits for expert evaluation of what it can't.
How do we prevent audit findings from entering permanent backlog?
Attribute findings to code locations. Assign ownership to specific teams. Set SLAs for severity levels. Track remediation progress. Decompose findings into component-level issues. Add tests and lint rules that prevent recurrence. The goal is fixing root causes, not just individual instances.
Should we audit before or after major redesigns?
Both. Audit before to identify existing issues that should be addressed in the redesign. Audit after to validate the new state and catch issues introduced during the redesign. The redesign is an opportunity to fix structural issues that would be expensive to retrofit—knowing them before you start enables design-time fixes.
What qualifications should auditors have?
Look for: IAAP certification (CPWA, WAS, CPACC), demonstrated AT expertise (actual screen reader and keyboard testing, not just tool running), WCAG interpretation experience, ability to provide actionable developer-focused guidance, references from similar organizations. Avoid auditors who primarily run automated tools and produce generic reports.
Related Resources
Internal Links
- How Continuous Compliance Works
- Accessibility Testing vs Monitoring
- The Role of CI/CD in Accessibility
- Accessibility Testing Tools: Manual vs Automated vs AI
- What "Shift Left" Means for Accessibility
- Accessibility as Infrastructure, Not a Feature
External Sources
- W3C WAI: Planning and Managing Web Accessibility
- W3C Understanding Conformance
- DOJ ADA Web Guidance
- WebAIM Million 2024 Report
- Section 508: Testing and Validation
- W3C WAI: Evaluating Web Accessibility
This article was written by TestParty's editorial team with AI assistance. All statistics and claims have been verified against primary sources. Last updated: January 2026.
Stay informed
Accessibility insights delivered
straight to your inbox.


Automate the software work for accessibility compliance, end-to-end.
Empowering businesses with seamless digital accessibility solutions—simple, inclusive, effective.
Book a Demo