Don't forget the Chrome Dev Summit, starting Monday at 10:00am (Pacific) and streaming live on YouTube. Schedule.

From Raw Video to Web Ready

Now that I've introduced you to applications that do media file manipulation, I'm going to take a raw video file off a camera and transform it into an encrypted resource that you can play back using a video library such as Google's Shaka Player. I'm specifically going to show you how to format your video for mobile playback.

To be more specific about the goals, the result of the procedures described here will be media resources with the following characteristics:

  • Audio and video streams are split into separate files
  • Versions of the video file are in mp4 and webm format
  • Versions of the audio file are in m4a and webm format
  • A bitrate of 0.35 Megabits per second (Mbs)
  • Resolution of 640 by 360
  • Encrypted
  • Viewable on all major browsers using appropriate technologies

By "appropriate technologies" I mean DASH and HLS, which are the two primary means of providing video in HTML on the major browsers. What those terms mean and how to use them is a whole topic itself. I won't be getting into those, but by the end of this article, you'll be able to create media files that are ready for use in DASH and HLS.

One final note: selection of the file formats, bitrate and resolution are not arbitrary. I've selected these values for speedy playback on the mobile web.

If you want to play along at home, you'll need a raw video file off a camera, preferably one that contains both audio and video. If you don't have one handy, then here's ten seconds of an mov file that I took of the Rathaus-Glockenspiel in Munich's MarienPlatz.

How are media Files put together?

Before I start manipulating media files, I want to talk a bit about how media files are put together. I think of them as being like an onion. The file that you see in your operating system shell is a container, identified by a file extension (mp4, webm, etc.). The container houses one or more streams. Media files may contain many streams of varying types and even multiple streams of the same type. Most files you'll encounter will only contain a single audio and a single video stream. (Some may also contain captions and data, but I won't be covering those.)

Within the audio and video streams, the actual data is compressed using a codec. As we'll see later, the distinction between a container and a codec is important as files with the same container can have their contents encoded with different codecs.

(If these terms are new to you, I explain them in Media file characteristics.)

The image below illustrates this. On the left is the basic structure. On the right are the specifics of that structure for an mp4. I'll explain file manipulation by moving downward through these layers.

Media container onion

Change the container

Let's start by changing the file container. You'll recall that we're starting with a file that has an mov extension. I'm going to use ffmpeg to change the container type from mov to mp4 and webm. In actual practice, you would likely specify a codec at the same time. For this lesson, I'm letting ffmpeg use its defaults.

ffmpeg -i glocken.mp4

Creating a webm file is a bit more complicated. ffmpeg has no trouble converting a mov file to a webm file that will play in Chrome and Firefox. For whatever reason, the file created using ffmpeg's defaults (at least for the version I used) doesn't quite conform to the webm spec. (For the curious, it sets a DisplayUnit size that isn't defined by the webm spec.) Fortunately, I can fix this using a video filter. Do so with the -vf flag and the setsar filter.

ffmpeg -i -vf setsar=1:1 glocken.webm

Webm takes quite a bit longer to create than mp4. This isn't surprising when you look at the results. While mp4 compresses to about a quarter of the original file's size, webm is down in the single digits, though results may vary.

-rw-r--r--. 1 fr  12M Jun 27 11:48
-rw-rw-r--. 1 fr 9.7M Jun 27 14:47 glocken.mp4
-rw-rw-r--. 1 fr 480K Jun 27 14:50 glocken.webm

Check your work

This is a good place to remind you that you can verify the results of any task in this article using the same applications you're using to do the work. Remember that, as described in the primers, you'll need both ffmpeg and Shaka Packager since neither shows you everything.

packager input=glocken.mp4 --dump_stream_info

ffmpeg -i glocken.mp4

Split the streams

When you package media resources for DASH you'll need to separate the video and audio streams. Splitting the audio and video streams of a file is often referred to as demultiplexing or demuxing for short. Professional content creators often record audio and video to separate files to begin with. If resources come from a consumer-grade device, audio and video will be in the same file and you'll need to split them.

In the context of preparing media for the web, demuxing is basically a file copying operation where I copy one of the streams to a new file. Using Shaka Packager I might, for example, pull out the video stream.

packager input=glocken.mp4,stream=video,output=glocken_video.mp4

Notice that the stream descriptor has an input, an output, and a stream type. Despite the mp4 extension this output file wouldn't have any sound. To get an audio file, I'd rerun this command with stream=audio and output it to a different file name. But, you might remember from Application Primers that Shaka Packager can take multiple stream descriptors. This means that Shaka Packager lets me split the audio and video in a single command.

packager \
  input=glocken.mp4,stream=video,output=glocken_video.mp4 \

A full discussion of audio and video formats is beyond the scope of this article, though in the next section I'll cover what you need for DASH and HLS. One thing to note is that the audio stream of an mp4 file is the advanced audio coding format for which m4a is a common file extension.

Shaka Packager presents demuxing as though you're extracting a stream into a new file. It's a little different in ffmpeg, which presents as though you're stripping the stream you don't want. With ffmpeg, you need two operations.

ffmpeg -i glocken.webm -vcodec copy -an glocken_video.webm
ffmpeg -i glocken.webm -acodec copy -vn glocken_audio.webm

Just as with Shaka Packager, we have both an input and an output file. Another difference from Shaka Packager is that the streams are identified with flags that refer their codecs. The -vcodec copy and -acodec copy portions of the command tell ffmpeg to copy the streams I want while the -an and -vn flags strip the streams I don't want. The keyword copy means I'm moving the streams without changing their codecs.

A list of demuxing commands is provided on the cheat sheet.

Change the codec

Continuing downward, we arrive at the codec. Codec, which is short for coder- decoder, is a compression format for video or audio data. As stated earlier, a codec is not the same thing as a container. Two files of the same container type could hold data compressed using completely different codecs. The webm format for example allows audio to be encoded using either vorbis or opus. To change the codec I need ffmpeg.

In the last section I demuxed the audio and video like this:

ffmpeg -i glocken.webm -vcodec copy -an glocken_video.webm
ffmpeg -i glocken.webm -acodec copy -vn glocken_audio.webm

If I need to change the audio and video codec, I would replace the copy keyword with the name of a codec. For example, this command outputs an audio file encoded with the aac codec.

ffmpeg -i glocken.webm -vn -c:a vorbis glocken.m4a

The cheat sheet lists commands needed to convert codecs. The tables summarize the libraries used in ffmpeg to perform the codec conversions for webm and mp4 files. These are the formats recommended for DASH and HLS respectively.


Extension Codec Library
mp4 H264 libx264
webm VP9 libvpx-vp9


Extension Codec Library
mp4 aac aac
webm vorbis libvorbis
opus libopus

File properties

I've unpeeled the onion. Yet if you compare what I've done to the list of stated goals you'll see that I'm not quite done. Among the remaining items are bitrate and resolution. These properties correlate to the amount of data in a media file. It probably goes without saying, but I'm going to say it anyway, that you can always lower bitrate and resolution, but increasing them is a problem. Without special software and algorithms, quality is going to take a hit.

The first step in changing bitrate and resolution is to display the file characteristics and verify that your source file has a higher bitrate or resolution than your desired result.


Bitrate is the maximum number of bits used to encode one second of a stream. The more bits used to encode a second of stream, the higher the fidelity.

Unsurprisingly, bitrates for the web are low. The table below compares web bitrates with common home video formats. Values are given in megabits per second (Mbs).

Format Bitrate
Blu-ray 20Mbs
DVD 6 Mbs
Desktop web video 2 Mbs
4G mobile video 0.7 Mbs
3G mobile video 0.35 Mbs
2G mobile video Depends on network type.
  • EDGE: 0.4 Mbs
  • GPRS: 0.04Mbs

Which value should I use for video on my web pages? The short answer is at least desktop, 4G, and 3G. If you're serving video in one of the markets referred to as "the next billion users", say India, for example, you'll want to include 2G as well. For demonstration purposes, I'm going to target 3G.

In ffmpeg you set the bitrate with the (surprise!) bitrate (-b) flag.

ffmpeg -i -b:v 350k glocken.mp4
ffmpeg -i -b:a 64k glocken.m4a

Notice that there are two bitrate flags, -b:a and -b:v. One is for audio and the other is for video.


Resolution is the amount of information in a single frame of video, given as the number of logical pixels in each dimension. Youtube recommends the following resolutions for video uploads, all in the 16:9 aspect ratio. There's nothing specific to Youtube about this list. It's just a list of common 16:9 video resolutions.

Abbreviation Dimensions
2160p 3840x2160
1440p 2560x1440
1080p 1920x1080
720p 1280x720
480p 854x480
360p 640x360
240p 426x240

Which one do I use? That depends on your application. For simple embedding you may chose a single resolution. If you're preparing files for DASH or HLS, you may chose one, several, or all. Fortunately, this is one of the simplest transformations you'll make with ffmpeg.

ffmpeg -i glocken.webm -s 640x360 glocken_640x360.webm

It's worth reiterating that you should start from the highest resolution and bitrate file you have available. If you're one of the many who are now converting your sites from using Flash videos to HTML5 videos, you'll want to find your original camera or other high resolution sources and convert from that rather than from flv and f4v files.


If you plan to enforce copyright on your media, you'll want to encrypt them. I'm going to introduce two encryption methods. The first is Widevine, a proprietary encryption method offered by Google. The second, also free, is the Clear Key encryption system , which is supported by a W3C spec.

Clear Key Encryption

Clear Key is a common key system defined in the Encrypted Media Extensions specification that uses unencrypted keys to decrypt MPEG Common Encryption (CENC) content.

In practice this is a simple way to encrypt and decrypt content for playback on the web. What's more, playback can be done with either DASH or HLS. This section will show you how to encrypt using Clear Key.

The first thing you need to do is generate a key. You can use whatever method you want, but I'm going to use OpenSSL to write sixteen random hex digits to a file. This step is necessary when preparing resources for both HLS and DASH.

openssl rand -out media.key 16

Use Shaka Packager to do the actual encryption. Use the content of the media.key file for both the -key and -key-id flags.

packager \
  input=glocken.mp4,stream=audio,output=glocken_audio_encrypted.m4a \
  input=glocken.mp4,stream=video,output=glocken_video_encrypted.mp4 \
  --enable_fixed_key_encryption --enable_fixed_key_decryption \

Widevine Encryption

Unless your company completes the Master License Agreement with Widevine, this type of encryption can really only be used for testing. Since this is covered without much explanation on the Shaka Packager README here it goes.

Everything in this command except the name of your files and the --content_id flag should be copied exactly from the example. The --content_id is 16 or 32 random hex digits. Use the keys provided here instead of your own. (This is how Widevine works.)

packager \
  input=glocken.mp4,stream=audio,output=glocken_audio_encrypted.m4a \
  input=glocken.mp4,stream=video,output=glocken_video_encrypted.mp4 \
  --enable_widevine_encryption \
  --key_server_url "" \
  --content_id "16_Rand_Hex_Chrs" --signer "widevine_test" \
  --aes_signing_key "1ae8ccd0e7985cc0b6203a55855a1034afc252980e970ca90e5202689f947ab9" \
  --aes_signing_iv "d58ce954203b7c9a9a9d467f59839249"


This does not cover everything you could do to a media file before posting it to the web, not by a longshot. To be fair, this subject is one deserving of its own website. I'm hoping this introduction will give you enough to help you find your own answers to questions.