
Recherche avancée
Autres articles (111)
-
La file d’attente de SPIPmotion
28 novembre 2010, parUne file d’attente stockée dans la base de donnée
Lors de son installation, SPIPmotion crée une nouvelle table dans la base de donnée intitulée spip_spipmotion_attentes.
Cette nouvelle table est constituée des champs suivants : id_spipmotion_attente, l’identifiant numérique unique de la tâche à traiter ; id_document, l’identifiant numérique du document original à encoder ; id_objet l’identifiant unique de l’objet auquel le document encodé devra être attaché automatiquement ; objet, le type d’objet auquel (...) -
Les tâches Cron régulières de la ferme
1er décembre 2010, parLa gestion de la ferme passe par l’exécution à intervalle régulier de plusieurs tâches répétitives dites Cron.
Le super Cron (gestion_mutu_super_cron)
Cette tâche, planifiée chaque minute, a pour simple effet d’appeler le Cron de l’ensemble des instances de la mutualisation régulièrement. Couplée avec un Cron système sur le site central de la mutualisation, cela permet de simplement générer des visites régulières sur les différents sites et éviter que les tâches des sites peu visités soient trop (...) -
Encodage et transformation en formats lisibles sur Internet
10 avril 2011MediaSPIP transforme et ré-encode les documents mis en ligne afin de les rendre lisibles sur Internet et automatiquement utilisables sans intervention du créateur de contenu.
Les vidéos sont automatiquement encodées dans les formats supportés par HTML5 : MP4, Ogv et WebM. La version "MP4" est également utilisée pour le lecteur flash de secours nécessaire aux anciens navigateurs.
Les documents audios sont également ré-encodés dans les deux formats utilisables par HTML5 :MP3 et Ogg. La version "MP3" (...)
Sur d’autres sites (12511)
-
FFMPEG concat leaves audio gapes between clips
14 novembre 2022, par GotCubesI'm writing a python script that uses subprocess to invoke FFMPEG, not using pyffmpeg.



My script generates a variable number of MP4 files using the AAC audio codec, and concatenates them together using FFMPEG. Here is how I'm constructing each clip :



ffmpeg -loop 1 -i image.jpg -i recording.mp3 -tune stillimage -c:a aac -b:a 256k -shortest clip.mp4




The command I'm using to concatenate them is :



ffmpeg -f concat -i clip_names.txt -c copy video_raw.mp4




I then take that resulting video, and mix a looping audio track over it, and adjust the volume. (Sorry for the awful formatting)



ffmpeg -i video_raw -filter_complex
 "amovie=Tracks/Breaktime.mp3:loop=0,
 volume=0.1,
 asetpts=N/SR/TB[aud];
 [0:a][aud]amix[a]"
-map 0:v -map [a] -b:a 256k -shortest final_video.mp4




These commands seem to work as I intend them to. When I play the resulting MP4 from my local machine, everything plays without issue.



However, I uploaded the video to YouTube, and ran into issues. When the video is played from YouTube, there is about a second of silence at every timestamp where two clips were concatenated, before the next clip begins. I've tried this from Chrome, IE, and Firefox, all with the same issues.



Based on what I've looked into so far, I think it could be an issue with how the priming samples of each individual clip are handled. I'm not obligated to keep using MP4 or AAC, so if using a different audio/video codec would work better, feel free to suggest !



Is there some type of manipulation I can do in FFMPEG to get rid of the priming samples, or somehow process them differently ? In the end, I'm looking for each clip to play back to back without the delay that the concat operation seems to insert. Thank you !


-
Convert sequence of h264 frames into video with ffmpeg
12 août 2022, par utnd03I'm testing a hardware video encoder module, and for each input frame (raw YUV in NV12 format) the encoder generates an h264 output frame. How can I "concatenate" these individual h264 output frames into a video (e.g. mp4) with
ffmpeg
?

$ ls
dst0000.h264 dst0002.h264 dst0004.h264 dst0006.h264 dst0008.h264 
dst0001.h264 dst0003.h264 dst0005.h264 dst0007.h264

$ file dst0000.h264
dst0000.h264: JVT NAL sequence, H.264 video, main @ L 31



I tried
ffmpeg -i dst0001.h264 -c:v copy output.mp4
, which allows me to view a single h264 frame (and they look normal, so the encoder is working properly).

I also tried
ffmpeg -f image2 -s 640x480 -r 30 -i dst%04d.h264 -c:v copy out.mp4
, but this gave the following error, which I had no idea how to fix :

[image2 @ 0x55bf32735780] Could not find codec parameters for stream 0 (Video: none, none, 640x480): unknown codec
Consider increasing the value for the 'analyzeduration' (0) and 'probesize' (5000000) options
Input #0, image2, from 'dst%04d.h264':
 Duration: 00:00:00.30, start: 0.000000, bitrate: N/A
 Stream #0:0: Video: none, none, 640x480, 30 fps, 30 tbr, 30 tbn
[mp4 @ 0x55bf3273fc80] Could not find tag for codec none in stream #0, codec not currently supported in container
Could not write header for output file #0 (incorrect codec parameters ?): Invalid argument
Error initializing output stream 0:0 -- 
Stream mapping:
 Stream #0:0 -> #0:0 (copy)
 Last message repeated 1 times



-
RGB-frame encoding - FFmpeg/libav
4 février 2014, par learnerI am learning video encoding & decoding in FFmpeg. I tried the code sample on this page (only the video encoding & decoding part). Here the dummy image being created is in YCbCr format. How do I achieve similar encoding by creating RGB frames ? I am stuck at :
Firstly, how to create this RGB dummy frame ?
Secondly, how to encode it ? Which codec to use ? Most of them work with YUV420p only...
EDIT : I have a YCbCr encoder and decoder as given on the this page. The thing is, I have RGB frame-sequence in my database and I need to encode it. But the encoder is for YCbCr. So, I am wondering to convert RGB frames to YCbCr (or YUV420P) somehow and then encode them.
At decoding end, I get decoded YCbCr frames and I convert them back to RGB. How to go ahead with it ?I did try the swscontext thing, but the converted frames lose color information and also scaling errors. I thought of doing it manually using two for loops and colorspace conversion formulae
but I am not able to access individual pixel of a frame using FFmpeg/libav library ! Like in OpenCV we can easily access it with something like : Mat img(x,y) but no such thing here ! I am totally a newcomer to this area...Someone can help me ?
Many Thanks !