I'm new in AV technology and have been trying to combine FFmpeg with Apple's CoreVideo framework to process webcam captures. First, i have webcam captures from CoreVideo(which could be found from AVCaptureVideoDataOutputSampleBufferDelegate) represented by CMSampleBuffer - (void)captureOutput:(AVCaptureOutput *)captureOutput didOutputSampleBuffer:(CMSampleBufferRef)sampleBuffer fromConnection:(AVC
![FFmpeg: How to get AVPacket from OSX CMSampleBufferRef](https://cdn-ak-scissors.b.st-hatena.com/image/square/98d6f053a97a87156775f60757c60865d0f2c47d/height=288;version=1;width=512/https%3A%2F%2Fcdn.sstatic.net%2FSites%2Fstackoverflow%2FImg%2Fapple-touch-icon%402.png%3Fv%3D73d79a89bded)