3 Innovative Ways to Approach English Descriptive Audio
Updated: January 24, 2019
3.5% of the world’s population live with vision impairment, making audio description – also referred to as English Descriptive Audio – an important component to our society’s fast-growing video content. A major barrier to audio description is cost, as traditional methods require hiring voice actors to record the audio in a studio. We’re sharing three innovative ways to approach English Descriptive Audio.
1. Synthesized speech
Traditional description requires human labor for the entire workflow. Once the description transcript is written, human voice actors are hired to record the description. This method is extremely costly and quite time consuming. Synthesized speech completely removes the need for voice recording and editing. Skipping these tedious steps allows audio description to be created much faster and at a lower cost.
3Play Media takes this unique approach to description, using a combination of humans and technology. Certified human describers write high quality descriptions, then synthesized speech is used to vocalize these descriptions. Using a combination of human editing and advanced technology, the cost of audio description can be significantly decreased without sacrificing quality. In fact, synthesized speech has many benefits.
The sound of synthesized speech is familiar to most blind and low vision users who are accustomed to fast-paced and mechanized audio from using screen readers.
Additionally, with synthesized speech, the user is in complete control, as the synthesized speech will vocalize the exact description written by the human describer. Likewise, if you decide that you would like to make a tweak to the description, you can do so without having to re-hire and pay a voice actor for their time in the studio.
2. Human voice narration
There are a number of ways in which you can create audio description on your own. If you are creating a talking head video with a slide deck, for example, in a classroom lecture, you can narrate the visual information that’s contained in the slides or in the background of the video as they’re happening when you record the original video. If you thoroughly describe what is in the video while you’re recording, it will eliminate the need to go through afterwards and add in descriptions. This is a great way to cut costs.
Note that if you do choose to go this route, it’s important to consider the best practices for creating accessible video for blind and low vision users. These include verbally covering all displayed visual information, identifying the speaker and speaker changes, explaining any participation by audience members, and taking frequent pauses.
3. WebVTT or text file
Even if you’re not creating a talking head video, there are still a few ways that you can create audio description yourself. For instance, you can create a text-only description, or essentially write a text version describing all of the visual information that’s happening in the video. While this is certainly easier, it does lose some of the cinematic detail for the viewer, and it doesn’t include quite the same amount of accommodation.
To make this slightly more accessible, you can create a text-only description that is time-coded, just like you would time code a caption file. When doing this, you want to be sure that the description fits into the natural pauses of the video. Next, you can utilize this document to create a WebVTT file, which is similar to captions, but for description. WebVTT files are supported natively in HTML5.
Similarly, you could also create a text-only merged transcript and description, which is a text doc that contains both the transcript of the audio and your video, and the description of the visual information in your video. This is really helpful for deaf-blind viewers as an accommodation, but it’s not time-synchronized, so it would be the equivalent of providing a transcript only for a video instead of providing captions.
Artificial Intelligence Is Good, but Is It Good Enough for Captions?
“Artificial intelligence (AI) makes it possible for machines to learn from experience, adjust to new inputs, and perform human-like tasks.” In other words, rather than programming machines with a single rule for output, they are taught to recognize patterns and then utilize…
Automatic Captions: Are They Helping or Hindering Students?
In the 2019 State of Captioning report, 37% of respondents indicated that they use automatic captions for educational videos. At the same time, 50% said that they’re aware automatic captions don’t fulfill the legal requirements. These findings seem to conflict, so I…
2015 Harvard Accessibility Lawsuit Will Move Forward in Federal Court
On February 5, 2015, the National Association of the Deaf (NAD) filed a class-action lawsuit against Massachusetts Institute of Technology (MIT) and Harvard University. Harvard was accused of violating both the Americans with Disabilities Act (ADA) and Section 504 of the Rehabilitation…