Accessibility Reports

Hi @steve.green
we had a conversation last year - time flies! I’ve posted a link to the chat below. I would assume that many are in the same boat as me - having good intentions but trying to do more with less.
It’s useful for me to know what ‘good enough’ looks like and what ‘fully compliant’ looks like. You’re clearly saying that the latter is impossible for most generalist test teams.

Hi @sles12
I stumbled on the VPAT recently and the fact that it is a checklist immediately caught my attention.
Noting that although it has a WCAG version it is presumably a little biased towards the US market? Ie the US Rehabilitation Act of 1973. Would you recommend this as a resource for UK-based companies or might there be better options.

Hi @danuk ,
I just learned something about the history of VPATs.
Our VPAT is an Excelsheet with a row for every WCAG 2.2 AA success criterion. So I’m not sure how any US bias would happen. I think Ady gave a good answer.

“Fully compliant” is light years beyond what a generalist testing team can achieve. “Good enough” is difficult to define, and it is the job of the product owner to define it, not the testers. There are potentially legal and other consequences of this decision (such as loss of revenue or reputation) so you should kick it upstairs to someone who gets paid to make decisions like this.

“Good enough” might focus on a few success criteria that provide a worthwhile benefit and are relatively easy to understand and test, such as keyboard navigation and text alternatives for images. Testing the colour contrast of text isn’t difficult, and nor is testing the colour contrast of some non-text items such as focus indicators and the borders of form controls, although a full non-text contrast test will be beyond most generalist testers.

Alternatively, the people upstairs might also define “Good enough” as “Whatever you can do in any spare time” or “Don’t do anything until someone complains”.

1 Like

@steve.green, don’t scare people away from accessibility :pleading_face: If you tell someone they can’t do it, why even try?

My project just got officially assessed by a third party assessment provider (really nit picky) and we are 98% WCAG 2.2 compliant. No one in the team apart from Design had any idea about accessibility when we started improvements. A couple of years later the review didn’t bring many surprises, but it gave us the push to tackle those “we can’t fix it” topics.
I started knowing nothing and the first version wasn’t perfectly tested, but it’s progress over perfection and now I’m QA Accessibility Lead. Teaching others I can see how some training and the right tools and a contact for the tricky questions empowers all testers.

I’m not saying they are not capable of doing it and that they shouldn’t try. I’m saying they can’t do it properly at the moment and that they have a vast amount to learn if they want to do it properly.

Nothing wrong with that - we all have to start at the beginning. But if some people don’t want to put in the effort to learn something that’s extremely difficult, maybe they shouldn’t start. I get the impression that nowadays, testers want to become instant experts in things they know nothing about. They don’t want to put in the years of study needed to genuinely become expert.

I don’t know who did your assessment, but no reputable accessibility consultancy would quote a percentage WCAG conformance figure because it’s meaningless.

In my case, the report is a statutory compliance requirement: it certifies our conformity to the agency that grants the accessibility label. Because of its critical importance, we have automated part of the workflow through APIs in our CI pipeline—at every production release, a monitoring job runs, checks the various categories of accessibility issues, and returns a result aligned with our reporting criteria. The generated reports are then archived on a dedicated server so they can be produced for legal purposes whenever necessary.