
Recherche avancée
Médias (91)
-
Les Miserables
9 décembre 2019, par
Mis à jour : Décembre 2019
Langue : français
Type : Textuel
-
VideoHandle
8 novembre 2019, par
Mis à jour : Novembre 2019
Langue : français
Type : Video
-
Somos millones 1
21 juillet 2014, par
Mis à jour : Juin 2015
Langue : français
Type : Video
-
Un test - mauritanie
3 avril 2014, par
Mis à jour : Avril 2014
Langue : français
Type : Textuel
-
Pourquoi Obama lit il mes mails ?
4 février 2014, par
Mis à jour : Février 2014
Langue : français
-
IMG 0222
6 octobre 2013, par
Mis à jour : Octobre 2013
Langue : français
Type : Image
Autres articles (112)
-
Personnaliser en ajoutant son logo, sa bannière ou son image de fond
5 septembre 2013, parCertains thèmes prennent en compte trois éléments de personnalisation : l’ajout d’un logo ; l’ajout d’une bannière l’ajout d’une image de fond ;
-
Websites made with MediaSPIP
2 mai 2011, parThis page lists some websites based on MediaSPIP.
-
Des sites réalisés avec MediaSPIP
2 mai 2011, parCette page présente quelques-uns des sites fonctionnant sous MediaSPIP.
Vous pouvez bien entendu ajouter le votre grâce au formulaire en bas de page.
Sur d’autres sites (11316)
-
Python : Passing complex (ffmpeg) arguments to Popen
25 juin 2016, par xaccrocheurThis ffmpeg Popen invocation works :
command = ['ffmpeg', '-y',
'-i', filename,
'-filter_complex', 'showwavespic',
'-colorkey', 'red',
'-frames:v', '1',
'-s', '800:30',
'-vsync', '2',
'/tmp/waveform.png']
process = sp.Popen( command, stdin=sp.PIPE, stderr=sp.PIPE)
process.wait()But I need to use ’compand, showwavespic’ and this comma seems to be blocking the execution. I also need to pass all sorts of strange characters, like columns and, well, all that you can find in a CLI invocation.
How can I pass complex arguments ?
-
Transcode HLS Segments individually using FFMPEG
27 mai 2013, par rayhI am recording a continuous, live stream to a high-bitrate HLS stream. I then want to asynchronously transcode this to different formats/bitrates. I have this working, mostly, except audio artefacts are appearing between each segment (gaps and pops).
Here is an example ffmpeg command line :
ffmpeg -threads 1 -nostdin -loglevel verbose \
-nostdin -y -i input.ts -c:a libfdk_aac \
-ac 2 -b:a 64k -y -metadata -vn output.tsInspecting an example sound file shows that there is a gap at the end of the audio :
And the start of the file looks suspiciously attenuated (although this may not be an issue) :
My suspicion is that these artefacts are happening because transcoding are occurring without the context of the stream as a whole.
Any ideas on how to convince FFMPEG to produce audio that will fit back into a HLS stream ?
** UPDATE 1 **
Here are the start/end of the original segment. As you can see, the start still appears the same, but the end is cleanly ended at 30s. I expect some degree of padding with lossy encoding, but I there is some way that HLS manages to do gapless playback (is this related to iTunes method with custom metadata ?)
** UPDATED 2 **
So, I converted both the original (128k aac in MPEG2 TS) and the transcoded (64k aac in aac/adts container) to WAV and put the two side-by-side. This is the result :
I'm not sure if this is representative of how a client will play it back, but it seems a bit odd that decoding the transcoded one introduces a gap at the start and makes the segment longer. Given they are both lossy encoding, I would have expected padding to be equally present in both (if at all).
** UPDATE 3 **
According to http://en.wikipedia.org/wiki/Gapless_playback - Only a handful of encoders support gapless - for MP3, I've switched to lame in ffmpeg, and the problem, so far, appears to have gone.
For AAC (see http://en.wikipedia.org/wiki/FAAC), I have tried libfaac (as opposed to libfdk_aac) and it also seems to produce gapless audio. However, the quality of the latter isn't that great and I'd rather use libfdk_aac is possible.
-
How to manipulate YUV frames obtained from an ffmpeg video decoder ?
23 juillet 2014, par user3665376I wanted to apply effects on video while rendering on android . I decoded the mpeg4 video using ffmpeg to raw YUV(YUV420P) frames .
I modified the Y(luminescence) bytes using the below snippet and rendered it to screen after converting it to RGB frames .
AVFrame *frame = [YUV420P frame recieved from ffmpeg decoder]
uint8_t *tempPtr = NULL;
tempPtr = frame->data[0];
int j;
for(j = 0 ; j < frame->linesize[0] ; j++){
*(tempPtr++) = 1;
}I applied a Y value of 1 to all pixels of all frames . I expected an output video of low luminescence ie.much darker video . But I was not able to notice any change in luma of the rendered video .
Note : The YCbCr data planes at AVFRame.data[0] , AVFrame.data[1] , Avframe.data[2] of the decoded frame has expected values . The AVFrame.buf pointers were all pointing to NULL (Don’t know whether it has anything to do with data).
Wanted to know whether anybody has encountered this situation before ??Links to related opics also will be very much helpful.
Thanks.