What Is Closed Captioning?
Updated: June 3, 2019
Closed captions are chunks of time-synchronized text that reflect the audio track and can be read while watching the video. The process of closed captioning involves transcribing the audio to text, dividing the text into chunks known as “caption frames,” and then synchronizing the caption frames with the video. Closed captions are typically located underneath the video or overlaid on top of the video.
Captions should assume that the viewer cannot hear the audio at all. Closed captions should convey not only the spoken content, but also any sound effects, speaker identification, and other non-speech elements. In this way, captions are different from subtitles, which assume that the viewer can hear but cannot understand the language. Learn more about the difference between closed captions, subtitles, and SDH (subtitles for the deaf and hard of hearing).
Closed captions should account for any sound that is not visually apparent but is integral to the plot. For example, the sound of keys jangling should be included in closed captions if it is important to the plot development–like in the case where a person is standing behind a locked door. However, it should not be included if it is the sound of keys jangling in someone’s pocket while they are walking down the street. Learn more about captioning and subtitling standards.
Origin of Closed Captions
Closed captions originated in the early 1980s as a result of an FCC mandate and gradually became a requirement for most programming on broadcast television. Accessibility laws went on to require closed captions for Internet video, and have become available across many different types of devices and media. The original purpose of closed captions was to provide accommodations for people with hearing disabilities. However, it turns out that most people who use closed captions are not deaf or hard of hearing. Closed captions are often used because they improve comprehension and remove language barriers, especially for people who know English as a second language. Closed captions also compensate for poor audio quality or a noisy background, and allow the video to be consumed in sound sensitive environments like a workplace.
Closed Captioning Terminology
Captions vs. Transcript
The difference between captions and a transcript is that a transcript is not synchronized with the media. Transcripts are sufficient for audio-only content like a podcast, but captions are required for videos or presentations with a voice-over.
Closed Captions vs. Open Captions
The difference between closed and open captions is that closed captions allow the viewer to turn the captions on and off. Open captions are burned into the video and cannot be turned off. Closed captions are more common with online video. See a demo of open vs. closed captions.
Pre-Recorded vs. Live Captions
Pre-recorded captions (aka “offline captions”) versus live captions (aka “real-time captions”) refers to the timing of when the captioning process is done. Live captioning is done in real time by stenographers while the event is happening. Pre-recorded captioning is done after the event has taken place.
Closed Captions vs. Subtitles vs. SDH
In the US, “closed captions” are designed for viewers who are deaf and hard of hearing. In addition to the spoken words, closed captions convey sound effects, speaker identification, and other non-speech elements. On the other hand, “subtitles” are intended for viewers who can hear the audio, but cannot understand the language. For this reason, American subtitles communicate the spoken content but not the sound effects. Outside of the U.S., the words “subtitles” and “closed captions” are often used interchangeably. “SDH” means “subtitles for the deaf and hard-of-hearing.” SDH subtitles convey the same information as closed captions but in some cases have technical differences. For more information, read the article How Subtitles for the Deaf and Hard-of-Hearing (SDH) Differ from Closed Captions
Closed Caption Formats
There are many different caption formats that are used depending on the type of media player or device. Below is a list of some of the more common caption formats and their respective applications.
|Format Type||Use Cases|
|SCC||Broadcast, iOS, web media|
|SRT||YouTube and web media|
|SAMI||Microsoft / Windows Media|
|QT||QuickText / QuickTime|
How to Caption Videos in Foreign Languages
There are plenty of reasons why you should caption your content. Increased accessibility, improved user engagement and SEO, and flexibility in sound-sensitive environments are just a few of the many benefits. If you produce video in languages other than English, finding a…
How to Translate Videos into Foreign Languages
As the world continues to become more interconnected, it’s important to create content that has the ability to reach a global audience. Technology has given people around the world access to online content like never before. Video made up 69% of all…
How to Publish via Integration, Manual Upload, and Encoding
Imagine that you created a video you want to publish on your video player. You go through your usual process: film, edit, and add captions. After all your hard work, now it’s time to publish your video. You wonder how you’re going…