[x264-devel] 4-th annual MSU MPEG-4 AVC/ H.264 codecs comparison

sean darcy seandarcy2 at gmail.com
Sat Jan 26 18:28:24 CET 2008


On Jan 25, 2008 9:56 AM, Dmitriy Vatolin <dmitriy at graphics.cs.msu.ru> wrote:
> Dear all!
>
>   Below is FAQ with answer to very commonly asked question "why so old
>   x264 was used". Really we receive all codecs simultaneously and
>   answer about preparation time is below.
>
> ---------------------------------------------------------------
> FOURTH ANNUAL MSU MPEG-4 AVC/ H.264 CODECS COMPARISON RELEASED!
> ---------------------------------------------------------------
>
> Most interesting features:
> * Codecs were analyzed separately in the following types of application:
>   * "Video Conferences"
>   * "Movies"
>   * "HDTV"
>   (settings for every application area was provided by developers)
> * PSNR/SSIM RD Curves
> * Relative Bitrate/Relative Speed graphs individually for all the usage areas
> * Per-frame analysis
> * New specific analysis methods suggested (on synthetic sequences)
>
> -------------------------------------------------------------------
>   SHORT FAQ OF FOURTH ANNUAL H.264 CODECS COMPARISON
> -------------------------------------------------------------------
>
> Q. What was main goal of the comparison?
>
> A. main idea of this annual comparison - official commercial companies
>    involving and so increasing number of participants. We are already
>    involved Intel, AMD, ATI, Apple (partially) and smaller Fraunhofer,
>    Elecard, Ateme, VSS, MainConcept, Sorenson, Artemis etc.
>
> Q. Why official? You can get codec from a developer's site or (if
>    unavailable) from P2P networks and compare them successfully!
>
> A. This is not true. For example, key feature of our comparisons is
>    comparison in several application areas (videoconferences, SD,
>    HDTV). Different areas require different codec settings. The
>    simplest strategy -- use of default settings -- definitely works,
>    but it is as much definitely nonoptimal (in terms of
>    quality-bitrate balance). So we received more optimal settings from
>    developers to test codec with higher accuracy. Also every year we
>    change sequences in an unpredictable way to prevent tuning to
>    specific sequences by developers.
>
>    Another problem is speed. Default settings of one codec can made it
>    10 time slower than another one and all your quality measurements
>    will not be so correct. Of course we can select suitable parameters
>    with close speed, but everybody can say "you select wrong
>    parameters for our codec, so you received wrong results", that's
>    why we do not use this approach in the annual H.264 comparison. Our
>    current solution is not sophisticated: We just asked companies for
>    optimal settings (as they think of them) for the predefined
>    application area and predefined time limits. The first step of
>    codec qualification to the comparison is time measurement: we
>    report time, measured on our system to developers with comments
>    like "you have 50% overtime, please make your preset faster", or
>    "your time reserve is still 40%, you can make your preset slower,
>    but with better quality/rate results".
>
>    Another advantage of official contacts: we typically receive the
>    newest internal versions, that are planned to be published not very
>    soon (it was up to half a year for one of our participants).
>
>
> Q. Fine. But why you have so long timeout between starting till
>    publication?
>
>    Several reasons:
>
>    1. As we told previously - we have slow speed acceptance procedure,
>       that require correspondence with companies.
>
>    2. Next - if it's possible - we try to get companies chance to fix
>       clear bug. We receive fresh research version, so some bugs inside
>       is possible (especially this was actual for second annual
>       comparison). This can be very simple problem like "unlucky
>       build" with fast fixing. So if we have time reserve, we stop
>       measurement, report bug and get company chance to fix it. Of
>       course we wait very few time (2-3 days) with clear final
>       deadline, but this procedure also add additional delay.
>
>    3. Also we have measurements approval procedure. After we done all
>       measurements and after internal checking we provide measured
>       results of company's codec and test sequences to correspondent
>       companies, so companies are able to check our measurements with
>       there codec. We published our measurement tool (MSU Video Quality
>       Metric) and provide our measurement results in it's format,
>       so all results can be verified.
>
>    4. Next step is reports preparation (short, long versions and
>       versions for private comparison). Report size can be up to 200
>       pages, we prepare several reports, so this is also not so fast
>       step.
>
>    5. And final step is reports verification and reviewing. We provide
>       report draft to all participants for verification. They report
>       defect and commonly good suggestions. If we think, that some
>       suggestion is very important we can add appendix to comparison
>       with additional measurements or some clarifications. This also
>       require time. And final step is independent reviewing by very
>       good independent experts in H.264 codecs.
>
> Q. Ok. Situation with time is clear. Is it possible for a company to
>    remove some results if it ask for this?
>
> A. Now we avoid such situation by using a so-called private participation.
>
>    The motivation is pretty simple: at the start, without knowing its
>    real position against competitors, a developer did want to
>    participate in the public comparison. But after receiving results,
>    if this is not the first place or so, they did not want to see
>    these results published :). So, if we allow to remove, there will
>    be only one codec in the comparison. Maybe two. :) (somebody and
>    x264). So we restrict such removing now, but allow private
>    participation.
>
>    Some companies did not want to participate due to not so good
>    current results which they know beforehand about. In private
>    participation a company pay us and receive its personal report.
>    This company codec is not included in the public comparison at all.
>    So, if company pay us, we prevent any conflicts of interests. A
>    company can understand its place and does further works on codec
>    improvement, but do not get its current results published and
>    well-known (the annual comparisons are very popular - we have more
>    than 200000 total downloads now).
>
> Q. And the main question: what codec is the best?
>
> A. We really have maybe the maximum information about the best codecs in
>    the world and we really do not know. :) At first, please specify
>    usage area, speed and other requirements, etc, etc, etc... We can
>    answer what codec is better for some common usage samples.
>
>    Our another objective is to help codec developers to make there
>    codecs better and better and better. Commonly developers achieve
>    some really good results and believe that this is the absolute
>    limit, but we help them to see new limits. :)
>
> -----------------------------------------------------
>
>    Any suggestions and fixes are welcome!
>
>    Comparison URL is:
>    http://www.compression.ru/video/codec_comparison/mpeg-4_avc_h264_2007_en.html
>
> --
> Best regards,
>  Dr. Vatolin
>  mailto:dmitriy at graphics.cs.msu.ru
>
>
>
>
> _______________________________________________
> x264-devel mailing list
> x264-devel at videolan.org
> http://mailman.videolan.org/listinfo/x264-devel
>

Very interesting.


The report mentions that the presets ( options? ) for x264 were
provided by the x264 devs.


What were they?


Congrats to the developers.


sean



More information about the x264-devel mailing list