[FFmpeg-user] should I shoot the dog?

Michael Koch astroelectronic at t-online.de
Tue Sep 29 16:37:10 EEST 2020

Am 29.09.2020 um 14:58 schrieb Mark Filipak (ffmpeg):
> On 09/29/2020 04:06 AM, Michael Koch wrote:
>> Am 29.09.2020 um 04:28 schrieb Mark Filipak (ffmpeg):
>>> I just want to understand the frame structures that ffmpeg creates, 
>>> and that ffmpeg uses in processing and filtering. Are Y, Cb, Cr 
>>> separate buffers? That would be logical. Or are the Y, Cb, Cr values 
>>> combined and organized similarly to macroblocks? I've found some 
>>> code that supports that. Or are the Y, Cb, Cr values thrown 
>>> together, pixel-by-pixel. That would be logical, too.
>> As far as I understood it, that depends on the pixel format.
>> For example there are "packed" pixel formats rgb24, bgr24, argb, 
>> rgba, abgr, bgra,rgb48be, rgb48le, bgr48be, bgr48le.
>> And there are "planar" pixel formats gbrp, bgrp16be, bgrp16le.
> Hi Michael,
> "Packed" and "planar", eh? What evidence do you have? ...Share the candy!

As far as I know, this is not described in the official documentation. 
You can find it for example here:

> Now, I'm not talking about streams. I'm talking about after decoding. 
> I'm talking about the buffers. I would think that a single, consistent 
> format would be used.

There is no single consistent format used internally. See Gyan's answer 


More information about the ffmpeg-user mailing list