
Recherche avancée
Médias (1)
-
Bug de détection d’ogg
22 mars 2013, par
Mis à jour : Avril 2013
Langue : français
Type : Video
Autres articles (78)
-
MediaSPIP v0.2
21 juin 2013, parMediaSPIP 0.2 est la première version de MediaSPIP stable.
Sa date de sortie officielle est le 21 juin 2013 et est annoncée ici.
Le fichier zip ici présent contient uniquement les sources de MediaSPIP en version standalone.
Comme pour la version précédente, il est nécessaire d’installer manuellement l’ensemble des dépendances logicielles sur le serveur.
Si vous souhaitez utiliser cette archive pour une installation en mode ferme, il vous faudra également procéder à d’autres modifications (...) -
MediaSPIP version 0.1 Beta
16 avril 2011, parMediaSPIP 0.1 beta est la première version de MediaSPIP décrétée comme "utilisable".
Le fichier zip ici présent contient uniquement les sources de MediaSPIP en version standalone.
Pour avoir une installation fonctionnelle, il est nécessaire d’installer manuellement l’ensemble des dépendances logicielles sur le serveur.
Si vous souhaitez utiliser cette archive pour une installation en mode ferme, il vous faudra également procéder à d’autres modifications (...) -
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir
Sur d’autres sites (8418)
-
FFMPEG libraries : Bitstream "h264_mp4toannexb" filter does not work
25 décembre 2013, par user2677612We are using
FFmpeg
libraries git-ee94362libavformat
v55.2.100. Our purpose is to mux two streams (video and audio) into M3U8 playlist using HLS.We are using
AV_CODEC_ID_H264
output encoder,AV_PIX_FMT_YUV420P
output video pixel format andCODEC_FLAG_GLOBAL_HEADER
flag for the encoder.The last causes us to use "h264_mp4toannexb" bit stream filter.
So, here is the code snippet :
AVPacket outpkt = {0};
int isGotVideoPacket = 0;
av_init_packet(&outpkt);
out_video_frame->pts = (int64_t) (video_frame_count * in_video_frame_duration / av_q2d(out_video_stream->time_base));
int ret = avcodec_encode_video2(enc_out_video_ctx, &outpkt, out_video_frame[i], &isGotVideoPacket);
assert(0 <= ret);
if ((1 == isGotVideoPacket) && (0 < outpkt.size)) {
AVPacket new_outpkt = outpkt;
if ((AVBitStreamFilterContext*) 0 != vbsf_ctx) {
AVPacket new_outpkt = outpkt;
ret = av_bitstream_filter_filter(vbsf_ctx, enc_out_video_ctx, (const char*)0, &new_outpkt.data, &new_outpkt.size, outpkt.data, outpkt.size, outpkt.flags & AV_PKT_FLAG_KEY);
if (ret > 0)
{
outpkt = new_outpkt;
}
else
{
// We get ret = -22
char errbuf[128] = "";
// Both the functions get "Error number -22 occurred" that don't explain anything
av_strerror (ret, errbuf, 128);
av_make_error_string (errbuf, 128, ret);
}
assert(0 <= ret);
}
outpkt->stream_index = output_video_stream->index;
// If to comment av_bitstream_filter_filter() and "if-else", then
// At frame #37 we get the following error from av_interleaved_write_frame():
// [mpegts @ 09628140] H.264 bitstream malformed, no startcode found, use the h264_mp4toannexb bitstream filter (-bsf h264_mp4toannexb).
ret = av_interleaved_write_frame(ofmt_ctx, &outpkt);
assert(0 <= ret);
}Our questions :
1. What is the meaning of the "-22" error from av_bitstream_filter_filter()?
2. Where can we get full FFMPEG error code description list?
3. If we are using av_bitstream_filter_filter() right? If no, what is the right way?Andrey Mochenov.
-
Prepare mp4 videos for Media Source Extensions API using ffmpeg
6 février 2018, par lerThis command produce init.mp4 + bunch of m4s files, i’m trying to play them using MSE :
ffmpeg -i <input file="file" /> -f hls -hls_segment_type fmp4 -c:v copy playlist.m3u8
This is the client side code i’m using :
var socket = io();
var video = document.querySelector('video');
var mimeCodec = 'video/mp4; codecs="avc1.64000d,mp4a.40.2"';
if ('MediaSource' in window && MediaSource.isTypeSupported(mimeCodec)) {
var mediaSource = new MediaSource;
video.src = URL.createObjectURL(mediaSource);
mediaSource.addEventListener('sourceopen', sourceOpen);
} else {
console.error('Unsupported MIME type or codec: ', mimeCodec);
}
function sourceOpen (_) {
var mediaSource = this;
var sourceBuffer = mediaSource.addSourceBuffer(mimeCodec);
sourceBuffer.mode = 'sequence';
socket.on('broadcast', function (newPiece) {
// here i'm getting the buffer of the video == buffer
sourceBuffer.addEventListener('updateend', function (_) {
video.play().then(function() { }).catch(function(error) { });
});
sourceBuffer.appendBuffer(buffer); // when the seconde video comes i append it's buffer
})
};Everything works fine when i send
init.mp4
file followed byplaylist0.m4s, playlist1.m4s, playlist2.m4s, ....
.
But when i try to playinit.mp4
file followed immediately 6,7,8 not 0,1,2 meaningplaylist6.m4s, playlist7.m4s, playlist8.m4s, ....
, it didn’t work.
I don’t know why, this supposed to be live video, the viewer that is watching the live from the beginning getsinit.mp4, playlist0.m4s, playlist1.m4s, playlist2.m4s, ....
.
Someone that came after 5 minutes gets something like thisinit.mp4, playlist32.m4s, playlist33.m4s, playlist34.m4s, ....
and so on, but so far it works only for the viewer that get’sinit.mp4, playlist0.m4s, playlist1.m4s, playlist2.m4s, ....
. the video can’t play for the others -
Why are Cb and Cr planes displaced differently from lum by the displace complex filter in ffmpeg ?
6 mai 2017, par NebI have a video encoded with the yuv420p pixel format and I want to displace its pixels. I’m using ffmpeg and its new displace filter. The filter takes as inputs (the video to be displaced and) two displacement maps respectively for X and Y axis. I decided to create the displacement maps directly into ffmpeg using the nullsrc video source filter and the geq filter to specify the value of the three planes : lum, Cb, Cr. The script is the following :
ffmpeg INPUT.mp4 -f lavfi -i nullsrc=size=${WIDTH}x${HEIGHT}:d=0.1,geq='lum=128+30*sin(2*PI*X/400):Cb=128+30*sin(2*PI*X/400):Cr='128+30*sin(2*PI*X/400)' -f lavfi -i nullsrc=size=${WIDTH}x${HEIGHT}:d=0.1,geq='lum=128+30*sin(2*PI*X/400):Cb=128+30*sin(2*PI*X/400):Cr=128+30*sin(2*PI*X/400)' -lavfi '[0][1][2]displace' OUTPUT.mp4
I used the example provided in the documentation of ffmpeg, since the expression used in geq is irrelevant for the purposes of the problem.
At the and of the computation, I get the pixels of the input video not properly displaced, meaning that I can clearly see a sort of ghost carrying-color-information video under a displaced but b/w one.
After some tests, I noticed that the displacemnt map created had only the luma plane displaced correctly while the chrominance planes were displaced, but differently from luma, which is the origin of the planes disalignment in the intput video as you can see in the following extract frames :I also noticed that the video describing the Cb and Cr planes of the displacement maps have half resolution of the luma plane.
My question is : how can i setup correctly the Cr and Cb planes in the geq definition so that they are exactly identical to the luma plane ?
It would be also great if someone could explain me why ffmpeg gives me an output so much different for luma and Cb, Cr planes even if the function provided is the same.
If, it can help, i’m using ffmpeg 3.3-static build.
Thanks for your time.