Multimedia Content

recommended solution This page discusses what is needed to make the content of the multimedia accessible. If you are looking for an accessible video player, you should check out the Enable Video Player page.

What makes a video accessible is widely misunderstood. Many web professionals know about closed captions. What many don't know is that they absolutely need audio descriptions in order to be WCAG AA compliant.

WCAG video related guidelines
Requirement A AA AAA
Closed Captions required required required
Audio Descriptions Use either audio descriptions or transcripts to be single A compliant. required required
Transcripts Use either audio descriptions or transcripts to be single A compliant. not required required
Good contrast for overlay text not required required required
Can be paused required required required
No Seizure Inducing Sequences required required required
Sign Language not required not required required for prerecorded video
WCAG audio related guidelines
Requirement A AA AAA
Closed Captions/Transcripts required required required
Can be paused or has independent volume control required required required
Has Low Background Noise not required not required required

Captions:

A text version of speech and other important audio content in a video, allowing it to be accessible to people who can't hear all of the audio. They are different from subtitles, which involve translating the video's language into an alternate language — closed captions are in the same language as the audio. Subtitles don't usually contain any information about other non-spoken audio in a video (e.g. ambient sound, music, etc).

How To Edit and Create Caption Files

On the web, captions are usually saved in the WebVTT format. There are a few caption editors that you can use to edit caption files:

Note that these are only two examples of captioning software. They are listed here as examples, and may or may not be useful for your usecase.

Captioning by hand may take a long time, depending on the length and the complexity of the video. You could hire a third party to create them for you if your budget allows, or use AI to create the first pass and then edit the caption file afterwards to ensure accuracy.

Should I Use Auto Generated Captions For My Videos?

Auto-generated captions should always be checked for accuracy. If the captions aren't accurate, they are technically not a pass for WCAG 1.2.2 - Captions (Prerecorded).

While AI has come a long way recently, and can sometimes be very accurate, you still should remediate AI-generated captions since it still produces errors. AI programs such as Whisper.cpp can generate well written captions (and can even translate from other languages into English) with greater accuracy than older technologies (e.g. YouTube's auto-generated captions). Let's discuss the issues with AI caption programs in detail below.

Do Background Sounds and Music Need Captions?

Some video services like YouTube will generate captions automatically using AI that converts audio to text. While this works reasonably well in when it's just one or two people talking in the video, auto-captioning software can fail when there is a lot of background music and noise (this is especially true with YouTube auto-generated captions). Programs like Whisper.cpp are much better in these situations, but can mess up the timing of the captions in certain situations. And while Whisper.cpp can do a good job with music lyrics in some situations, it depends on the music sometimes. Also Whisper.cpp can sometimes hallucinate, which is another reason to double check its work.

Furthermore, non-verbal sound effects in the video can be integral to understanding content or videos for deaf and hard of hearing users. Human caption writers will usually describe these sound effects inside parantheses. For example: in a horror film, the noises made a horror film by something offscreen can be captioned as "(a squeaky door opening can be heard in the distance)". Similar captions could describe a buzzer indicating a wrong answer or end of the round that isn't announced by the host in a game show. These are overt examples, but it is a huge reason auto captions should not be a replacement but rather a starting point for providing accessible captioning (and transcripts).

Audio Descriptions:

An extra audio track with spoken commentary will help users with visual disabilities perceive content that is presented only visually. A narrator usually describes the visual-only content in the video. Audio descriptions can be implemented in several ways:

  1. They can be included in the primary video, available to all users on the video's default audio track.
  2. They can be provided in a different audio track in the video. On broadcast television in North America, this has be implemented using Secondary Audio Programming.
  3. They can be recited by a computer voice speaking over the video content, as shown in our demo of AblePlayer on the Enable Accessible Video Player page.
  4. They can be provided by an alternate cut of the video. You would just need to have a link to that audio described version on the video page. An example of this is on the Killer B Cinema Trailer for 3 Dev Adam (with the audio described version linked in the video description).

Please note: If all of the visual information in described in the video's audio track already, then a separate audio description track is not necessary.

Transcripts:

A transcript is used by users who can neither hear audio nor see video, like deaf/blind users. It is like the script the actors in the video use to know what is going on in the story that they are about to act in. A transcript also should include descriptions of audio information in the video (like laughing) and visual information (such as the creak of a door opening).

All of these concepts are explained in greater detail in WebAIM's excellent article on Captions, Transcripts, and Audio Descriptions.

To find one of the most cost-effective way of implementing all three, I would suggest looking into the Enable Video Player page, which shows how do so using Able Player.

Do YouTube Transcripts Count As WCAG Transcripts?

Note that YouTube does have a "transcript" functionality, but what it basically does is just show all the captions with timing information in a section of the page. Here is a screenshot of where it appears in the video component on YouTube (next to the save button is a menu button with three dots, with a screen reader label of "More actions", that has "Show transcript" as a menu item):

Note that most of the time this is not really a transcript, since it doesn't have any of the visual information conveyed in the video. Below is a screenshot of a video has a YouTube transcript generated from a captions file:

A screenshot of the YouTube transcript component for the video linked above.  The text reads:



0:02
(Eerie, unsettling music plays in the background)
0:14
Somewhere between science and superstition,
0:21
there is another world
0:23
A world of darkness.
0:34
Nobody expected it.
0:39
Nobody believed it.
0:42
Nothing could stop it.
0:47
[Screaming]
0:59
The one hope.
1:01
The only hope.
1:04
The Turkish Exorcist.
1:09
(A very rough version of Tubular Bells from the movie's soundtrack plays in the background)
Figure 1. A screenshot of the YouTube transcript of the trailer for The Turkish Exorcist. Note that the transcript does not have descriptions of the visual actions that happen in the video. It only has information about the audio track (including that of the sound effects and music).

Now compare this with the AblePlayer video in the first example above. The AblePlayer video actually does have a proper transcript; it contains descriptions of what is happening visually that is not given by the spoken text in the video).

A screenshot of the AblePlayer transcript component for the Accessible Video Player example above.  The transcript reads:
    
    (Scene is a title screen, which reads 'Creating Accessible HTML5 Modal Dialogs For Desktop and Mobile. Zoltan Hawryluk - useragentman.com')
In this video, I will be demonstrating how to create an accessible modal dialog using the HTML5 dialog tag. It uses a polyfill for browsers that don't support it yet natively. For details on the polyfill and more information on how this demo was built, please checkout the video's accompanying blog post on useragentman.com.
Figure 2. A screenshot of the AblePlayer example transcript at the bottom of this page. To sighted users, the visual action in the video not covered by the closed captions appear in the red blocks inside the transcript.

Note that AblePlayer has interactive transcripts that, when clicked, will jump to the section of the video that the transcript applies to. This interactivity is a nice to have, and not necessary to satisfy WCAG..

The main takeaway here is that YouTube transcripts aren't really transcripts, since they don't give all users (particularly deafblind users), what is going on in the video.

How to Avoid Seizure Inducing Sequences

All animations, including video, must follow WCAG requirement 2.3.1 - Three Flashes or Below Threshold. This is to prevent animated and video content from causing users from having seizures.

Since I have already written blog posts on how to detect and fix seizure inducing sequences from video, I will just link to that content here:

A young boy wearing a cap, a blue jacket, and gloves is excitedly looking at a small yellow creature with long ears in front of him. They are both facing a large red background. A still from Dennō Senshi Porygon, an episode of the Pokémon TV show that caused some children to have seizures in 1997, due to a huge area of the screen flashing red and white.

Good Contrast On Overlay Text

Just like live HTML text over images, overlayed text in video must have enough contrast to be legible. Unlike live HTML text, it is impossible to enlarge this text or change the colour of overlayed text via CSS, since it is "burned in" to the pixels of the video. An example is in the image below.

Two martial artists are engaged in combat in an outdoor setting with traditional buildings in the background. The text 'PRODUCTION MANAGER: RICHARD CHEUNG' is overlaid in large white letters across the image. Some of this text is hard to read due to the white text blending in with some of the light gray buildings in the background. A screenshot from the opening credits from the film "Dance of the Drunk Mantis".

While you could check foreground and background colours of individual pixels using a tool like the WebAIM Contrast Checker, what I usually do is take a screenshot of the video frame with the burned in text and use a screenshot-based tool like Color Contrast Analyzer Chrome Plugin offered by IT Accessibility at NC State.

Two men fighting in the background with words overlayed on top that read 'Also Starring: Sun Kwei (as Rubber Legs student'), Charlier Shek (as Moneybags)'.  Like the previous screenshot, some of the text is hard to read since the white text sometimes blends into the light background of the ground. Another screenshot from the opening credits from the film "Dance of the Drunk Mantis"..
A screenshot of the WCAG 2.0 Contrast Checker plugin for Chrome testing the previous screenshot. It outlines a lot of the text, but some of it is not outlined, which indicates the text doesn't have proper contrast. A screenshot of the WCAG 2.0 Contrast Checker plugin for Chrome testing the previous screenshot.

Just like any colour contrast issues on web pages, you can fix this by:

A screenshot from the trailer of a film.  There is easy-to-read text overlayed on top of the action in the background that reads 'Killer B Cinema proudly presents'. A screenshot from the trailer for the 1973 Turkish film Karate Girl. Note that the text is easy to read due to the addition of dark text shadowing around the text onscreen.

Fixing Low Background Noise

A lot of background noise can be an issue for hard-of-hearding users who are listening to multimedia content. WCAG 1.4.7: Low or No Background Audio (a AAA requirement) states that recommends that for all multimedia content, one of the following should be true:

While this is an AAA requirement, it would be nice to fix these types of issues when they arise. If you have a piece of multimedia that has a lot of background noise, we recommend fixing that by re-mixing the audio. If you don't have the original multi-channel master of the audio, this problem still can be fixed by separating the speech with the rest of the audio using AI using a tool like Vocal Remover. You can then take these two tracks and remixing them with the vocals boosted in volume using a tool like Audacity. If the vocals aren't clear enough, you could use a tool like Adobe Podcast's Enhance Speech tool.

If the background noise is a persistent hiss or hum, you can Audacity's Noise Reduction tool to remove that from the audio track.

Below are two videos from a VHS copy of the American dub of the Japanese cartoon "8-man" (called "eighth Man" in the U.S.). The first video has the original audio with a lot of bad audio noise from the original over-the-air recording. The second video has removed that noise by extracting the voice data using Vocal Remover, improving its fidelity using Adobe Podcast's Enhance Speech tool and remixing it with the similar music from a different source. The results can be quite impressive.

An English language clip from an over-the-air recording of the 8th-man cartoon "Virus". Note that really bad audio hum in the background that makes the clip's audio almost unlistenable.
The same clip after AI was used to clean up the audio as described above. The differences are night and day.

Sign Language

Watching a video with captions can be more inclusive, but does add to the cognitive load and eye fatigue of the person watching the film. For people who understand sign language, it may be more desireable to have sign language interpretation as part of the mulimedia content.

Not everyone with hearing loss understands sign language. It should be noted that there are many sign languages used throughout the world. American Sign Language, British Sign Lanaguage and Plains Sign Language are just a few examples. For these reasons, Sign Language support in pre-recorded multimedia is a WCAG AAA guideline.

That said, Deaf users may want to have sign language and captions at the same time, especially if the may have trouble reading the captions due to cogntive of vision related disabilities. For that reason, sign language support in multimedia is mandated in a lot of government communications in countries around the world.

If you do wish to have sign language in your media, you will want to keep the following in mind:

  1. Ensure there is good lighting for the sign language video.
  2. Include a high contrast border around the signer that captures a lot of the signing space (transparent background is experimental).
  3. Ensuring high contrast between hands and background/clothes is ideal.
  4. The sign language interpreters face must be large enough for the viewer to see their facial expressions.
  5. The standard placement for a sign language interpreter is in the bottom right of video, but this can be adjusted in order to not obscure important content in a video.
  6. If you are in a multilanguage environment (like Canada), you may need two interpreters (e.g. American Sign Language for English-speaking Canada and Quebec Sign Language) for French-speaking Canada.
  7. Remember just because two countries (like the United States and the United Kingdom) have the same spoken language, they still may have two different forms of sign language (in this case, ASL and BSL).
  8. Some video editors may want to freeze the sign language interpreter to emphasize what is being said.

Considerations when Broadcasting Events Live

Live online events need to be accessible as well! The following should be kept in mind if you want to ensure an accessible experience for all viewers.

  1. Captions should be used throughout your entire broadcast. Human made ones are better than AI ones (or you could also have a human edit the AI ones on the fly).
  2. When posting the video afterwards, your captions should be edited for clarity and accuracy.
  3. Consider using sign language interpreters for your presentation. If you do so, please keep in mind that one interpreter is not enough, since any kind of on-the-fly translation is a cognitively tiring experience for the translator.
A COVID-19 update given by the Government of Canada, which was broadcast live on May 20, 2020. Note the two sign language interpreters for the two different forms of sign language used in Canada. These translators are swapped out just after the 31 minute mark to ensure the translators are not fatigued, which can easily lead to mistakes in the translation. Unfortunately, the original captions are auto-generated by the YouTube A.I. auto-captioner, which results in a less than optimal experience for users that need them. We replaced these captions with those generated by Whisper.cpp, which generates much better results.