
Recherche avancée
Médias (1)
-
Sintel MP4 Surround 5.1 Full
13 mai 2011, par
Mis à jour : Février 2012
Langue : English
Type : Video
Autres articles (51)
-
Les formats acceptés
28 janvier 2010, parLes commandes suivantes permettent d’avoir des informations sur les formats et codecs gérés par l’installation local de ffmpeg :
ffmpeg -codecs ffmpeg -formats
Les format videos acceptés en entrée
Cette liste est non exhaustive, elle met en exergue les principaux formats utilisés : h264 : H.264 / AVC / MPEG-4 AVC / MPEG-4 part 10 m4v : raw MPEG-4 video format flv : Flash Video (FLV) / Sorenson Spark / Sorenson H.263 Theora wmv :
Les formats vidéos de sortie possibles
Dans un premier temps on (...) -
La file d’attente de SPIPmotion
28 novembre 2010, parUne file d’attente stockée dans la base de donnée
Lors de son installation, SPIPmotion crée une nouvelle table dans la base de donnée intitulée spip_spipmotion_attentes.
Cette nouvelle table est constituée des champs suivants : id_spipmotion_attente, l’identifiant numérique unique de la tâche à traiter ; id_document, l’identifiant numérique du document original à encoder ; id_objet l’identifiant unique de l’objet auquel le document encodé devra être attaché automatiquement ; objet, le type d’objet auquel (...) -
Utilisation et configuration du script
19 janvier 2011, parInformations spécifiques à la distribution Debian
Si vous utilisez cette distribution, vous devrez activer les dépôts "debian-multimedia" comme expliqué ici :
Depuis la version 0.3.1 du script, le dépôt peut être automatiquement activé à la suite d’une question.
Récupération du script
Le script d’installation peut être récupéré de deux manières différentes.
Via svn en utilisant la commande pour récupérer le code source à jour :
svn co (...)
Sur d’autres sites (5644)
-
Is it possible to fetch some key-frames of a video by using the HTTP Range header
9 décembre 2020, par pvdI've read the SO problem , and it seems not applying to my specific case.


Is it possible to fetch some key-frames of a video from web server by the HTTP Range header ? For example, for a 30 seconds duration video, we'd like to analysis the I-frame around 00:00:02, 00:00:15, 00:00:28.


I need to analysis the videos from internal web server to detect if there's specific watermarks in it and some other analysis.


Since the first I-frame might be invalid sometimes(Logo for example), we were planning to extract the I-frame from the 00:00:02, the middle I-frame, and the last 2nd second I-frame.


For example, for a 30 seconds duration video, we'd like to analysis the I-frame around 00:00:02, 00:00:15, 00:00:28.


We could make it works while download the whole video, since most of the data we downloaded from the server are not being used. I was wondering if maybe we could only use the HTTP Range header to download partial data and analysis it ?


-
Livestream WebVTT with HLS
10 novembre 2022, par kltyeI've implemented an HLS service with ffmpeg (which pulls a live stream from nginx-rtmp). That all works fine, but now I'm wondering what kind of programming pattern I should be using to get live captioning to work.


I'm planning on using ffmpeg to output the incoming mp4 stream to multiple WAV chunks (i.e., the same way HLS fMP4 parts are created), and then sending those chunks over to Azure Cognitive Services for speech-to-text recognition. My question is, what do I do when I receive the speech results ? Do I dump that vtt file into the same directory as my HLS chunks, and then serve that up using a single m3u8 file (with audio/video tracks along with the text track) ?


Currently ffmpeg is updating the m3u8 playlist for HLS clients ; would it be possible for me to create the m3u8 playlist just for the vtt files, and serve that concurrently with the "regular" HLS playlist ? Also, time synchronization would seem to be difficult, because I'll be sending discrete WAV files over to Azure, so the vtt timestamps are going to be relative to the chunk I'm sending.


Help ! I've done searches online, and I grasp the various issues, but I'm not sure how to plumb them all together.


-
Audio alternative to FFmpeg - Core Audio IOS
9 juin 2021, par cs guyI have been using FFmpeg Android for a music app I'm working on. I built a custom audio engine from stratch with C++ and FFmpeg and it works amazing and it fulfilled all my needs. However, Due to FFmpeg being Lgpl lisence, it seems to me after some researching it is not possible to use a lgpl lisence due to app stores policy. Im not a lawyer or have the money to hire a lawyer for a commercial advise. So I am thinking to replace ffmpeg with another audio decoder, processor library. I am planning to feed the custom decoded data to audio devices through Apples core audio library.


Here are my needs :


- 

- Need to decode ogg files
- Need to encode pcm data as aac file
- Need to add post process FX to decoded data such as low pass filter etc








So what I am asking for is an answer to one of the following :


- 

- Could FFmpeg really not be used in app store due to lgpl static linking issues ? (I looked at the most famous apps that use FFmpeg on Android, all of them does not use FFmpeg on IOS)
- If I were to use another library for FFmpeg what is the best alternative to work with ? Did anyone actually had experienced the same situation that I am in ?






I also tried using AudioKit but it has a critical problem that does not meet with my requirement so I dropped it.


I am looking for advice here. Thanks !