Watch the video.
Update 3 (05/16/2020): Wrote an updated guide to use VMAF through FFmpeg.
Update 2 (01/06/2016): Fixed reference video bitrate unit from Kbps to KBps
When working with videos, you should be focusing all your efforts on best quality of streaming, less bandwidth usage, and low latency in order to deliver the best experience for the users.
This is not an easy task. You often need to test different bitrates, encoder parameters, fine tune your CDN and even try new codecs. You usually run a process of testing a combination of configurations and codecs and check the final renditions with your naked eyes. This process doesn’t scale, can’t we just trust computers to check that?
bit rate (bitrate): is a measure often used in digital video, usually it is assumed the rate of bits per seconds, it is one of the many terms used in video streaming.
We were about to start a new hack day session here at Globo.com and since some of us learned how to measure the noise introduced when encoding and compressing images, we thought we could play with the stuff we learned by applying the methods to measure video quality.
PSNR: is an engineering term for the ratio between the maximum possible power of a signal and the power of corrupting noise.
First, you calculate the MSE which is the average of the squares of the errors and then you normalize it to decibels.
MSE = ∑ ∑ ( [n1[i]-n2[i]] ) ^ 2 / m * n *n1 is the original image, n2 the comparable image, m and n are the image size PSNR = 10 log₁₀ ( MAX ^ 2 / MSE ) *MAX is the maximum possible pixel value of the image
For 3D signals (colored image), your MSE needs to sum all the means for each plane (ie: RGB, YUV and etc) and then divide by 3 (or 3 * MAX ^ 2).
To validate our idea, we downloaded videos (720p, h264) with the bitrate of 3400 kbps from distinct groups like News, Soap Opera and Sports. We called this group of videos the pivots or reference videos. After that, we generated some transrated versions of them with lower bitrates. We created 700 kbps, 900 kbps, 1300 kbps, 1900 kbps and 2800 kbps renditions for each reference video.
Heads Up! Typically the pivot video (most commonly referred to as reference video), uses a truly lossless compression, the bitrate for a YUV420p raw video should be 1280x720x1.5(given the YUV420 format)x24fps /1000 = 33177.6KBps, far more than what we used as reference (3400KBps).
We extracted 25 images for each video and calculate the PSNR comparing the pivot image with the modified ones. Finally, we calculate the mean. Just to help you understand the numbers below, a higher PSNR means that the image is more similar to the pivot.
|700 kbps||900 kbps||1300 kbps||1900 kbps||2800 kbps||3400 kbps|
We defined a PSNR of 38 (from our observations) as the ideal but then we noticed that the News group didn’t meet the goal. When we plotted the News data in the graph we could see what happened.
The issue with the video from the News group is that they’re a combination of different sources: External traffic camera with poor resolution, talking heads in a studio camera with good resolution and quality, some scenes with computer graphics (like the weather report) and others. We suspected that the News average was affected by those outliers but this kind of video is part of our reality.
We needed a better way to measure the quality perception so we searched for alternatives and we reached one of the Netflix’s posts: an approach toward a practical perceptual video quality metric (VMAF). At first, we learned that PSNR does not consistently reflect human perception and that Netflix is creating ways to approach this with the VMAF model.
They created a dataset with several videos including videos that are not part of the Netflix library and put real people to grade it. They called this score of DMOS. Now they could compare how each algorithm scores against DMOS.
They realized that none of them were perfect even though they have some strength in certain situations. They adopted a machine-learning based model to design a metric that seeks to reflect human perception of video quality (a Support Vector Machine (SVM) regressor).
The Netflix approach is much wider than using PSNR alone. They take into account more features like motion, different resolutions and screens and they even allow you train the model with your own video dataset.
“We developed Video Multimethod Assessment Fusion, or VMAF, that predicts subjective quality by combining multiple elementary quality metrics. The basic rationale is that each elementary metric may have its own strengths and weaknesses with respect to the source content characteristics, type of artifacts, and degree of distortion. By ‘fusing’ elementary metrics into a final metric using a machine-learning algorithm – in our case, a Support Vector Machine (SVM) regressor”
The best news (pun intended) is that the VMAF is FOSS by Netflix and you can use it now. The following commands can be executed in the terminal. Basically, with Docker installed, it installs the VMAF, downloads a video, transcodes it (using docker image of FFmpeg) to generate a comparable video and finally checks the VMAF score.
|# clone the project (later they'll push a docker image to dockerhub)|
|git clone –depth 1 https://github.com/Netflix/vmaf.git vmaf|
|# build the image|
|docker build -t vmaf .|
|# get the pivot video (reference video)|
|# generate a new transcoded video (vp9, vcodec:500kbps)|
|docker run –rm -v $(PWD):/files jrottenberg/ffmpeg -i /files/big_buck_bunny_360p_5mb.mp4 -c:v libvpx-vp9 -b:v 500K -c:a libvorbis /files/big_buck_bunny_360p.webm|
|# extract the yuv (yuv420p) color space from them|
|docker run –rm -v $(PWD):/files jrottenberg/ffmpeg -i /files/big_buck_bunny_360p_5mb.mp4 -c:v rawvideo -pix_fmt yuv420p /files/360p_mpeg4-v_1000.yuv|
|docker run –rm -v $(PWD):/files jrottenberg/ffmpeg -i /files/big_buck_bunny_360p.webm -c:v rawvideo -pix_fmt yuv420p /files/360p_vp9_700.yuv|
|# checks VMAF score|
|docker run –rm -v $(PWD):/files vmaf run_vmaf yuv420p 640 368 /files/360p_mpeg4-v_1000.yuv /files/360p_vp9_700.yuv –out-fmt json|
|# and you can even check VMAF score using existent trained model|
|docker run –rm -v $(PWD):/files vmaf run_vmaf yuv420p 640 368 /files/360p_mpeg4-v_1000.yuv /files/360p_vp9_700.yuv –out-fmt json –model /files/resource/model/nflxall_vmafv4.pkl|
You saved around 1.89 MB (37%) and still got the VMAF score 94.
Using a composed solution like VMAF or VQM-VFD proved to be better than using a single metric, there are still issues to be solved but I think it’s reasonable to use such algorithms plus A/B tests given the impractical scenario of hiring people to check video impairments.
A/B tests: For instance, you could use X% of your user base for Y days offering them the newest changes and see how much they would reject it.
A friend and I were extending Nginx by using lua scripts on it. Just in case you don’t know, to enable lua scripting at nginx you can use a lua module you can read more about how to push Nginx to its limits with Lua.
Anyway we were in a cycle:
- We did some lua coding.
- Restart nginx.
- Test it manually.
And this cycle was repeating until we have what we want. This was time consuming and pretty boring as well.
We then thought we could try to do some test unit with the scripts. And it was amazingly simple. We created a file called tests.lua and then we import the code we were using on nginx config.
package.path = package.path .. ";puppet/modules/nginx/functions.lua.erb" require("functions")
We also created a simple assertion handler which outputs function name when it fails or pass.
function should(assertive) local test_name = debug.getinfo(2, "n").name assert(assertive, test_name .. " FAILED!") print(test_name .. " OK!") end
Then we could create test suit to run.
function it_sorts_hls_playlist_by_bitrate() local unsorted_playlist = [[#EXTM3U #EXT-X-STREAM-INF:PROGRAM-ID=1,BANDWIDTH=1277952 stream_1248/playlist.m3u8 #EXT-X-STREAM-INF:PROGRAM-ID=1,BANDWIDTH=356352 stream_348/playlist.m3u8 #EXT-X-STREAM-INF:PROGRAM-ID=1,BANDWIDTH=485376 stream_474/playlist.m3u8]] local expected_sorted = [[#EXTM3U #EXT-X-STREAM-INF:PROGRAM-ID=1,BANDWIDTH=356352 stream_348/playlist.m3u8 #EXT-X-STREAM-INF:PROGRAM-ID=1,BANDWIDTH=485376 stream_474/playlist.m3u8 #EXT-X-STREAM-INF:PROGRAM-ID=1,BANDWIDTH=1277952 stream_1248/playlist.m3u8]] should(sort_bitrates(unsorted_playlist) == expected_sorted) end
I think that helped us speeding up a lot our cycle. Once again, by isolating a component and testing it, it’s a great way to make us productive.
I come from (lately) Ruby, Java and Clojure and currently I’m working with some python projects and I was missing the way I used to test my code
with rspec. After a quick research, I found three great projects that helps to make more readable tests, they are: py.test, Mock and sure.
I was missing a better way to make asserts into my tests. The option about using the plain assert is good but not enough and by using sure I could do some awesome code
looking similar to rspec.
#instead of plain old assert assert add(1, 4) == 5 #I'm empowered by add(1, 4).should.be.equals(5) .should.be.empty ['chip8', 'schip8', 'chip16'].shouldnt.be.empty [1, 2, 3, 4].should.have.length_of(4)
And this makes all the difference, my test code now is more expressive.
Struggling with monkeypatch
The other challenge I was facing was to understand and use monkeypatch for mocks and stubs. I find easier to use the Mock library even though its @patch looks similar to monkeypatch but I could grasp it quickly.
#Stubing def test_simple_math_remotely_stubed(): server = Mock() server.computes_add = Mock(return_value=3) add_remotely(1, 2, server).should.be.equals(3) #Mocking def test_simple_math_remotely_mocked(): server = Mock() add_remotely(1, 2, server) server.computes_add.assert_called_once_with(1, 2) #Stubing an internal dependency @patch('cmath.nasa.random') def test_simple_math_with_randon_generated_by_nasa(nasa_random_generator): nasa_random_generator.configure_mock(return_value=42) add_and_sum_with_rnd(3, 9).should.be.equals(54) #Mocking an internal dependency @patch('cmath.mailer.send') def test_simple_math_that_sends_email(mailer_mock): add_and_sends_email(3, 9) mailer_mock.assert_called_once_with( firstname.lastname@example.org', subject='Complex addition', body='The result was 12')
Make sure you
- Are using virtualenv for better lib version managment
- Have installed pytest, sure and mock
- Git cloned the project above to understand it.
I started programming in Visual Basic and I taste its roots, which are almost all full of procedure commands (bunch of do, goto and end), then I moved to C#, sharper it changes the end’s for }’s and give us a little more power based on some premises: we can treat two different things in the same way, polymorphism. The last static language, but not the least, I used (and I use it) Java, abusing of his new way of treating a set of things equality, the interfaces and using its “powers” on reflections.
Although when I started to use Ruby I saw that I could treat a group of things equality without doing any extra work. I still need to code models and composed types, even though we can create or change them dynamically using “real power” of metaprogramming.
When I start to study and apply the Clojure and its principles, my first reaction was the rejection, how can I go on without my formal objects, how can I design software without a model in the head and so on. I wasn’t thinking about how actually I do software, currently I use TDD to design software and I don’t think what models I need to have, I do think in terms of “what I want”. At minimum, Clojure make me think about, do we really need object to design software?! . A three days ago I saw an amazing video about similar thoughts: Some thoughts on Ruby after 18 months of Clojure.
Summarising: With my limited knowledge of theses languages, let’s suppose we use a function (which we don’t have source code) and we want to do something before that function is executed (intercept) using: VB I’ll need to check every single piece of code which we call this function and call another one, in Java we can use a AOP framework, in Ruby we can use the spells of metaprogramming. It seems that some frameworks, patterns and extra work aren’t needed more because of this dynamic language evolution.
My conclusions using dynamic languages (Clojure/Ruby) for now it’s: I write less code and reuse them more easy, so I don’t see any reason to create/use a new static typed language, would you see any motivation to do that?
PS: When I use C# (.Net Framework 1.3 – 2.0) it was not so super cool as today.