
Recherche avancée
Autres articles (105)
-
Des sites réalisés avec MediaSPIP
2 mai 2011, parCette page présente quelques-uns des sites fonctionnant sous MediaSPIP.
Vous pouvez bien entendu ajouter le votre grâce au formulaire en bas de page. -
Automated installation script of MediaSPIP
25 avril 2011, parTo overcome the difficulties mainly due to the installation of server side software dependencies, an "all-in-one" installation script written in bash was created to facilitate this step on a server with a compatible Linux distribution.
You must have access to your server via SSH and a root account to use it, which will install the dependencies. Contact your provider if you do not have that.
The documentation of the use of this installation script is available here.
The code of this (...) -
Les autorisations surchargées par les plugins
27 avril 2010, parMediaspip core
autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs
Sur d’autres sites (13929)
-
avutil/avstring : add a "ALL" entry and the possibility to negate matches to av_match_...
11 février 2016, par Michael Niedermayeravutil/avstring : add a "ALL" entry and the possibility to negate matches to av_match_name()
This will extend the whitelist features to allow blacklisting individual protocols and to
explicitly force everything to be enabled.Reviewed-by : Stefano Sabatini <stefasab@gmail.com>
Signed-off-by : Michael Niedermayer <michael@niedermayer.cc> -
What does FFmpeg expect me to send to the first rawvideo input pipe ?
23 octobre 2023, par SomebodyI'm using two named pipes, in order :


- 

- video_pipe

-f rawvideo
-video_size 1x1
-pix_fmt gray
-r 1



- audio_pipe

-f s16le
-ar 32000
-channels 1









I thought FFmpeg needed to read individual frames from a rawvideo pipe but I must be mistaken cause it doesn't start reading from the second pipe until I feed 11 bytes to the first pipe although, in the example given, a grayscale frame of one pixel is exactly one byte. I have experimenting by increasing video_size and here's the table I could infer :







 Actual frame size in bytes 

Bytes needed to be sent before to move on 







 1 

11 




 2 

17 




 3 

25 




 4 

33 




 5 

41 




 6 

49 









I can't just send multiple frames as I want to output a 1 second video.
I tested most of the parameters in this page : https://github.com/FFmpeg/FFmpeg/blob/ff5a3575fec2d49d5fae4ec1198a939e203314db/libavformat/options_table.h
but none of them solved it. (I also used "-re" with no luck).


This is an example command in case you want to reproduce the issue :

ffmpeg -y -re -f rawvideo -video_size 1x1 -pixel_format gray -framerate 1 -i \\.\pipe\video_pipe -f s16le -ar 32000 -channels 1 -i \\.\pipe\audio_pipe -map 0:v:0 -map 1:a:0 out.mp4


Any idea of how I could send the exact frame bytes amount instead of being forced to send way more bytes ?


- video_pipe

-
How to make WebM screen recording chunks independently processable for audio with FFmpeg ?
3 décembre 2024, par Dinesh KumarI am streaming screen recordings from the browser into 5-second WebM chunks using the
MediaRecorder
API. The first chunk (root chunk) is independently processable because it contains the necessary EBML headers and metadata. However, subsequent chunks are not independently processable, as they lack the required metadata, which prevents me from extracting audio independently from them.

I am unable to extract audio independently from the individual chunks using FFmpeg due to missing headers, resulting in errors like
EBML header parsing failed
. The first chunk works fine on its own, but the subsequent chunks also need to be processed independently for audio extraction.

I am looking for a solution using FFmpeg to fix these chunks so that I can extract audio independently from each chunk.


- 

- Is there a way to repair these chunks post-recording with FFmpeg to include the missing metadata and headers, making them independently processable for audio extraction ?
- Can FFmpeg reinitialize the EBML headers in each chunk, or is there a command that can add the metadata from the first chunk to subsequent chunks to allow for independent audio extraction ?






Additionally, should I consider any changes in the MediaRecorder API to ensure that the chunks are properly formatted for independent processing ? The goal is to make each WebM chunk fully independent, allowing me to extract audio independently from each chunk.