
Recherche avancée
Médias (91)
-
#3 The Safest Place
16 octobre 2011, par
Mis à jour : Février 2013
Langue : English
Type : Audio
-
#4 Emo Creates
15 octobre 2011, par
Mis à jour : Février 2013
Langue : English
Type : Audio
-
#2 Typewriter Dance
15 octobre 2011, par
Mis à jour : Février 2013
Langue : English
Type : Audio
-
#1 The Wires
11 octobre 2011, par
Mis à jour : Février 2013
Langue : English
Type : Audio
-
ED-ME-5 1-DVD
11 octobre 2011, par
Mis à jour : Octobre 2011
Langue : English
Type : Audio
-
Revolution of Open-source and film making towards open film making
6 octobre 2011, par
Mis à jour : Juillet 2013
Langue : English
Type : Texte
Autres articles (68)
-
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir -
Encoding and processing into web-friendly formats
13 avril 2011, parMediaSPIP automatically converts uploaded files to internet-compatible formats.
Video files are encoded in MP4, Ogv and WebM (supported by HTML5) and MP4 (supported by Flash).
Audio files are encoded in MP3 and Ogg (supported by HTML5) and MP3 (supported by Flash).
Where possible, text is analyzed in order to retrieve the data needed for search engine detection, and then exported as a series of image files.
All uploaded files are stored online in their original format, so you can (...) -
L’utiliser, en parler, le critiquer
10 avril 2011La première attitude à adopter est d’en parler, soit directement avec les personnes impliquées dans son développement, soit autour de vous pour convaincre de nouvelles personnes à l’utiliser.
Plus la communauté sera nombreuse et plus les évolutions seront rapides ...
Une liste de discussion est disponible pour tout échange entre utilisateurs.
Sur d’autres sites (8033)
-
Use modern avconv syntax for codec selection in documentation and tests
18 octobre 2012, par Diego BiurrunUse modern avconv syntax for codec selection in documentation and tests
- [DBH] doc/encoders.texi
- [DBH] doc/faq.texi
- [DBH] doc/filters.texi
- [DBH] tests/fate-run.sh
- [DBH] tests/fate/demux.mak
- [DBH] tests/fate/h264.mak
- [DBH] tests/fate/microsoft.mak
- [DBH] tests/fate/mp3.mak
- [DBH] tests/fate/mpc.mak
- [DBH] tests/fate/utvideo.mak
- [DBH] tests/fate/video.mak
- [DBH] tests/fate/vqf.mak
- [DBH] tests/lavf-regression.sh
-
Hardware Accelerated H264 Decode using DirectX11 in Unity Plugin for UWP
8 janvier 2019, par rohit nI’ve built an Unity plugin for my UWP app which converts raw h264 packets to RGB data and renders it to a texture. I’ve used FFMPEG to do this and it works fine.
int framefinished = avcodec_send_packet(m_pCodecCtx, &packet);
framefinished = avcodec_receive_frame(m_pCodecCtx, m_pFrame);
// YUV to RGB conversion and render to texture after thisNow, I’m trying to shift to hardware based decoding using DirectX11 DXVA2.0.
Using this : https://docs.microsoft.com/en-us/windows/desktop/medfound/supporting-direct3d-11-video-decoding-in-media-foundation
I was able to create a decoder(ID3D11VideoDecoder) but I don’t know how to supply it the raw H264 packets and get the YUV or NV12 data as output.
(Or if its possible to render the output directly to the texture since I can get the ID3D11Texture2D pointer)so my question is, How do you send the raw h264 packets to this decoder and get the output from it ?
Also, this is for real time operation so I’m trying to achieve minimal latency.
Thanks in advance !
-
How do I create and initialise a DXGI_FORMAT_NV12 resource in DX12 (source is AVFrame)
5 janvier 2023, par mikeI'm trying to create an NV12 resource as source for a video encoder in DX12. While I intend to eventually populate a resource from GPU, what I'm trying to do now is take an ffmpeg
AVFrame
I already have (inAV_PIX_FMT_YUV420P
format) and create a texture inDXGI_FORMAT_NV12
format using that data.

I understand the NV12 format (https://learn.microsoft.com/en-us/windows/win32/medfound/recommended-8-bit-yuv-formats-for-video-rendering#nv12) has U and V interleaved while the
AV_PIX_FMT_YUV420P
doesn't.

My main question is what does the
D3D12_RESOURCE_DESC
look like for an NV12 texture - do I tell it I need more than one array/mip level to make it planar ? Or do I just give it a single memory address with both planes layed out as per the NV12 format, and it figures out subresources for me based on the format ?

I understand that to read the data I define two SRVs, one for Y mapped to the Red channel and a second for U and V, but it's how I initialise it that's confusing me.