[vlc-devel] GSoC 2018 Idea `libplacebo Improvement'

Thomas Guillem thomas at gllm.fr
Wed Mar 21 08:54:00 CET 2018

On Wed, Mar 21, 2018, at 08:33, Steve Lhomme wrote:
> I used to do that with D3D11. It was fine with an integrated (Intel)
> GPU. But with discrete GPU the SW decoders were really slow. Having to
> use memory via the PCI bus. To counter that I pretend to do direct
> rendering and provide CPU memory and during prepare() I copy back to
> the GPU. It's much faster for discrete GPUs and doesn't impact much
> the integrated ones.
With OpenGL 4.x, mapped buffer performances are really good with the
discrete GPU I tested (GeForce GTX 660).For AMD GPUs, I had to map the buffer with the following option:
GL_CLIENT_STORAGE_BIT in order to indicate that we prefer the buffer to
be on the CPU side. Cf.
Don't forget that we also need feature parity with the OpenGL vout of
VLC. This means: - Interop with VAAPI surfaces. VLC use the
   "EGL_EXT_image_dma_buf_import" extension. I guess we can do the same
   with Vulkan ? Mapping some DMA buffers to a texture ? - Bonus: Interop with cuda ? (It's not done yet on the GL side of VLC) - Give VLC a way to specify its own Vertex Shaders (for 360°,
   VR/3D stuff) .-  And for Metal/MoltenVK: Interop with IOSurface/CVPixelBuffer.

> _________________________________________________
> vlc-devel mailing list
> To unsubscribe or modify your subscription options:
> https://mailman.videolan.org/listinfo/vlc-devel

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://mailman.videolan.org/pipermail/vlc-devel/attachments/20180321/4a7d97ac/attachment.html>

More information about the vlc-devel mailing list