
Recherche avancée
Médias (3)
-
GetID3 - Bloc informations de fichiers
9 avril 2013, par
Mis à jour : Mai 2013
Langue : français
Type : Image
-
GetID3 - Boutons supplémentaires
9 avril 2013, par
Mis à jour : Avril 2013
Langue : français
Type : Image
-
Collections - Formulaire de création rapide
19 février 2013, par
Mis à jour : Février 2013
Langue : français
Type : Image
Autres articles (45)
-
Supporting all media types
13 avril 2011, parUnlike most software and media-sharing platforms, MediaSPIP aims to manage as many different media types as possible. The following are just a few examples from an ever-expanding list of supported formats : images : png, gif, jpg, bmp and more audio : MP3, Ogg, Wav and more video : AVI, MP4, OGV, mpg, mov, wmv and more text, code and other data : OpenOffice, Microsoft Office (Word, PowerPoint, Excel), web (html, CSS), LaTeX, Google Earth and (...)
-
Contribute to translation
13 avril 2011You can help us to improve the language used in the software interface to make MediaSPIP more accessible and user-friendly. You can also translate the interface into any language that allows it to spread to new linguistic communities.
To do this, we use the translation interface of SPIP where the all the language modules of MediaSPIP are available. Just subscribe to the mailing list and request further informantion on translation.
MediaSPIP is currently available in French and English (...) -
Librairies et binaires spécifiques au traitement vidéo et sonore
31 janvier 2010, parLes logiciels et librairies suivantes sont utilisées par SPIPmotion d’une manière ou d’une autre.
Binaires obligatoires FFMpeg : encodeur principal, permet de transcoder presque tous les types de fichiers vidéo et sonores dans les formats lisibles sur Internet. CF ce tutoriel pour son installation ; Oggz-tools : outils d’inspection de fichiers ogg ; Mediainfo : récupération d’informations depuis la plupart des formats vidéos et sonores ;
Binaires complémentaires et facultatifs flvtool2 : (...)
Sur d’autres sites (3791)
-
ffmpeg can't stream to remote client
4 septembre 2014, par KFLI’m building a simple
ffmpeg
command line on my laptop to stream from its camera. The command line reads (verbose output at the botton) :host1> ffmpeg -v verbose \
-f dshow \
-i video="Camera":audio="Microphone" \
-r 30 -g 0 -vcodec h264 -acodec libmp3lame \
-tune zerolatency \
-preset ultrafast \
-f mpegts udp://12.34.56.78:12345Firstly, it works locally. I.e., I can view the output by using
ffplay
on the same host :host1> ffplay -hide_banner -v udp://12.34.56.78:12345
Now what is NOT working is when I do this from another machine in the same network. It shows a
nan
progress :host2> ffplay -hide_banner -v udp://12.34.56.78:12345
nan : 0.000 fd= 0 aq= 0KB vq= 0KB sq= 0B f=0/0I used
ncat
to dump the raw content. But there’s no output :host2>\ncat\ncat -v -u 12.34.56.78 12345
Ncat: Version 5.59BETA1 ( http://nmap.org/ncat )
Ncat: Connected to 12.34.56.78:12345.
(...and nothing happen...)Note that I can exclude firewall issues as I used
ncat
to communicate with each other across the wire using the same port and protocol (UDP). This works and they can chat to each other :host1> ncat -l -u -p 12345
host2> ncat -u 12.34.56.78 12345Any hint ?
I’m using Windows x64 with FFMPEG 64bit installed from here. Below is the Output of my ffmpeg command :
C:\ffmpeg\bin>ffmpeg -v verbose -f dshow -i video="Integrated Camera":audio="Microphone (Realtek High Definition Audio)" -r 30 -g 0 -vcodec h264 -acodec libmp3lame -tune zerolatency -preset ultrafast -f mpegts udp://12.34.56.78:12345
ffmpeg version N-66012-g97b8809 Copyright (c) 2000-2014 the FFmpeg developers
built on Sep 1 2014 00:21:15 with gcc 4.8.3 (GCC)
configuration: --disable-static --enable-shared --enable-gpl --enable-version3 --disable-w32threads --enable-avisynth --enable-bzlib --enable-fontconfig --enable-frei0r --enable-gnutls --enable-iconv --enable-libass --enable-libbluray --enable-libbs2b --enable-libcaca --enable-libfreetype --enable-libgme --enable-libgsm --enable-libilbc --enable-libmodplug -enable-libmp3lame --enable-libopencore-amrnb --enable-libopencore-amrwb --enable-libopenjpeg --enable-libopus --enable-librtmp --enable-libschroedinger --enable-libsoxr --enable-libspeex --enable-libtheora --enable-libtwolame --enable-libvidstab --enable-libvo-aacenc --enable-libvo-amrwbenc --enable-libvorbis --enable-libvpx --enable-libwavpack --enable-libwebp --enable-libx264 --enable-libx265 --enable-libxavs --enable-libxvid --enable-decklink --enable-zlib
libavutil 54. 7.100 / 54. 7.100
libavcodec 56. 1.100 / 56. 1.100
libavformat 56. 3.100 / 56. 3.100
libavdevice 56. 0.100 / 56. 0.100
libavfilter 5. 0.103 / 5. 0.103
libswscale 3. 0.100 / 3. 0.100
libswresample 1. 1.100 / 1. 1.100
libpostproc 53. 0.100 / 53. 0.100
Guessed Channel Layout for Input Stream #0.1 : stereo
Input #0, dshow, from 'video=Integrated Camera:audio=Microphone (Realtek High Definition Audio)':
Duration: N/A, start: 171840.657000, bitrate: N/A
Stream #0:0: Video: rawvideo, bgr24, 640x480, 30 fps, 30 tbr, 10000k tbn, 30 tbc
Stream #0:1: Audio: pcm_s16le, 44100 Hz, 2 channels, s16, 1411 kb/s
Matched encoder 'libx264' for codec 'h264'.
[graph 0 input from stream 0:0 @ 0000000000470aa0] w:640 h:480 pixfmt:bgr24 tb:1/10000000 fr:10000000/333333 sar:0/1 sws_param:flags=2
[auto-inserted scaler 0 @ 0000000004326d00] w:iw h:ih flags:'0x4' interl:0
[format @ 0000000004325a00] auto-inserting filter 'auto-inserted scaler 0' between the filter 'Parsed_null_0' and the filter 'format'
[auto-inserted scaler 0 @ 0000000004326d00] w:640 h:480 fmt:bgr24 sar:0/1 -> w:640 h:480 fmt:yuv444p sar:0/1 flags:0x4
No pixel format specified, yuv444p for H.264 encoding chosen.
Use -pix_fmt yuv420p for compatibility with outdated media players.
[graph 1 input from stream 0:1 @ 0000000000460c20] tb:1/44100 samplefmt:s16 samplerate:44100 chlayout:0x3
[audio format for output stream 0:1 @ 00000000004601a0] auto-inserting filter 'auto-inserted resampler 0' between the filter 'Parsed_anull_0' and the filter 'audio format for output stream 0:1'
[auto-inserted resampler 0 @ 00000000004604a0] ch:2 chl:stereo fmt:s16 r:44100Hz -> ch:2 chl:stereo fmt:s16p r:44100Hz
[libx264 @ 000000000081bb20] using cpu capabilities: MMX2 SSE2Fast SSSE3 SSE4.2 AVX
[libx264 @ 000000000081bb20] profile High 4:4:4 Intra, level 3.0, 4:4:4 8-bit
[mpegts @ 000000000081abe0] muxrate VBR, pcr every 3 pkts, sdt every 200, pat/pmt every 40 pkts
Output #0, mpegts, to 'udp://12.34.56.78:12345':
Metadata:
encoder : Lavf56.3.100
Stream #0:0: Video: h264 (libx264), yuv444p, 640x480, q=-1--1, 30 fps, 90k tbn, 30 tbc
Metadata:
encoder : Lavc56.1.100 libx264
Stream #0:1: Audio: mp3 (libmp3lame), 44100 Hz, stereo, s16p
Metadata:
encoder : Lavc56.1.100 libmp3lame
Stream mapping:
Stream #0:0 -> #0:0 (rawvideo (native) -> h264 (libx264))
Stream #0:1 -> #0:1 (pcm_s16le (native) -> mp3 (libmp3lame))
Press [q] to stop, [?] for help
*** 1 dup!
frame= 241 fps= 31 q=28.0 Lsize= 3439kB time=00:00:08.03 bitrate=3506.4kbits/s dup=1 drop=0
video:3035kB audio:125kB subtitle:0kB other streams:0kB global headers:0kB muxing overhead: 8.791966%
Input file #0 (video=Integrated Camera:audio=Microphone (Realtek High Definition Audio)):
Input stream #0:0 (video): 240 packets read (221184000 bytes); 240 frames decoded;
Input stream #0:1 (audio): 16 packets read (1411200 bytes); 16 frames decoded (352800 samples);
Total: 256 packets (222595200 bytes) demuxed
Output file #0 (udp://12.34.56.78:12345):
Output stream #0:0 (video): 241 frames encoded; 241 packets muxed (3108187 bytes);
Output stream #0:1 (audio): 306 frames encoded (352512 samples); 307 packets muxed (128313 bytes);
Total: 548 packets (3236500 bytes) muxed
[libx264 @ 000000000081bb20] frame I:241 Avg QP:27.97 size: 12897
[libx264 @ 000000000081bb20] mb I I16..4: 100.0% 0.0% 0.0%
[libx264 @ 000000000081bb20] coded y,u,v intra: 26.3% 0.5% 0.0%
[libx264 @ 000000000081bb20] i16 v,h,dc,p: 19% 28% 21% 31%
[libx264 @ 000000000081bb20] kb/s:3095.29
[dshow @ 0000000000467720] real-time buffer[Integrated Camera] too full (90% of size: 3041280)! frame dropped!
Received signal 2: terminating. (I pressed CTRL-C) -
Ffmpeg and Xing header - iTunes issue and Duration issue
17 mars 2023, par AbsI extract the audio from a bunch of flv files as an MP3. This works great :


ffmpeg -i video.flv -vn -acodec copy audio.mp3



However, some audio that I extract have durations that are longer than they should be and some MP3 files keep looping the audio ! Also in some audio players like WMP, the seekbar gets stuck at one point.


I have found out this is due to Xing headers not being correct or set (?). I was hoping someone could help me edit the above command so this is fixed and maybe some explanation of why this happens so I can learn.


-
ffmpeg decoding through C-API leads to artefacts when input resolution is 1200x1600. Am I doing something wrong ?
26 février 2023, par AntonioUsing the C-API and FFmpeg 5.1 I have been able to encode h264 videos with libx264 on Android.
Now I wanted to replay them on Linux inside my C++ application. These videos can be played correctly on a browser, or on other players that I tried like mplayer or ffplay from ffmpeg. Also, I can unroll the frames with
ffmpeg -i recording.mp4 -start_number 0 -qscale:v 5 %06d.jpg
and the images look alright.

However in my C++ application every now and then, but in a very repeatable way, I get artifacts (like the bright pixels showing up above the monitor). They do not accumulate, even though they are not related to keyframes. So whatever error is going on, it doesn't seem to have an impact on subsequent frames. I use OpenCV to visualize the output, and I am pretty sure the problem is not the conversion to BGR because the artifact is already there if I simply show the y channel (luminance, grayscale).


These artifacts show up in videos that I have recorded with a 1200x1600 resolution. It is to be noted that 1200 is not divisible by 32 so ffmpeg does add some padding, but I am dealing with it and it's not an issue. Videos recorded at 1920x1440 are replayed with no artifacts. Two sample videos can be found here for download.


Here follows the code I am using, on the bottom you can see a picture of my decoded image with the artifact and the same as unrolled by ffmpeg command line. It should be noted that I am working with a custom built version of ffmpeg, out of conan packages, while the unrolling is done with ffmpeg from command line that comes with Ubuntu.


extern "C" {
#include <libavcodec></libavcodec>avcodec.h>
#include <libavformat></libavformat>avformat.h>
}

#include 
#include 
#include 

#include <iostream>
int main(int argc, char** argv) {

 int ret;

 auto pkt = av_packet_alloc();
 if (!pkt) {
 std::cerr << "Failed av_packet_alloc()" << std::endl;
 exit(1);
 }

 AVFormatContext* av_format = avformat_alloc_context();
 ret = avformat_open_input(&av_format, FILE_NAME, nullptr, nullptr);
 if (ret < 0) {
 std::cerr << "Failed avformat_open_input, Error: " << ret << std::endl;
 ///Error codes https://stackoverflow.com/questions/12780931/ffmpeg-exit-status-1094995529
 exit(1);
 }
 av_dump_format(av_format, 0, FILE_NAME, 0);
 auto video_st_number = av_find_best_stream(av_format, AVMEDIA_TYPE_VIDEO, -1, -1, nullptr, 0);
 if (video_st_number < 0) {
 std::cerr << "av_find_best_stream couldn't find video stream" << std::endl;
 exit(1);
 }
 auto video_st = av_format->streams[video_st_number];
 auto codec_id = video_st->codecpar->codec_id;
 std::cout << "Duration " << video_st->duration << std::endl;
 std::cout << "n_frames " << video_st->nb_frames << std::endl;

 auto frame = av_frame_alloc();
 if (!frame) {
 fprintf(stderr, "Could not allocate video frame\n");
 exit(1);
 }

 auto codec = avcodec_find_decoder(codec_id);
 if (!codec) {
 fprintf(stderr, "Codec not found\n");
 exit(1);
 }

 auto c = avcodec_alloc_context3(codec);
 if (!c) {
 fprintf(stderr, "Could not allocate video codec context\n");
 exit(1);
 }


 if ((ret = avcodec_parameters_to_context(c, video_st->codecpar))) {
 fprintf(stderr, "Failed avcodec_parameters_to_context\n");
 exit(1);
 }

 c->pix_fmt = AV_PIX_FMT_YUV420P;///Not really necessary
 c->thread_count = 1;///No impact

 /* open it */
 if (avcodec_open2(c, codec, NULL) < 0) {
 fprintf(stderr, "Could not open codec\n ");
 exit(1);
 }

 std::size_t counter = 0;
 std::size_t n_keyframes = 0;

 while (ret >= 0) {
 ret = av_read_frame(av_format, pkt);
 if (pkt->size == 0) {
 std::cout << "Skipping packet of size zero" << std::endl;
 av_packet_unref(pkt);
 continue;
 }
 while (avcodec_send_packet(c, pkt) != 0) {
 if (avcodec_receive_frame(c, frame) != 0) {
 std::cerr << "Error receiving frame" << std::endl;
 exit(1);
 } else {
 n_keyframes += frame->key_frame;
 std::cout << "Decoded " << ++counter << " frames. Frame No. " << frame->pts / pkt->duration << " "
 << frame->decode_error_flags << " " << frame->key_frame << " " << n_keyframes << " "
 << frame->pkt_dts << std::endl;
 }
 display(frame);
 }
 av_packet_unref(pkt);
 }

 avcodec_send_packet(c, nullptr);
 std::cout << "Flushing decoder" << std::endl;

 while (avcodec_receive_frame(c, frame) == 0) {
 n_keyframes += frame->key_frame;
 std::cout << "Decoded " << ++counter << " frames. Frame No. " << frame->pts << " " << frame->decode_error_flags
 << " " << frame->key_frame << " " << n_keyframes << " " << frame->pkt_dts << std::endl;

 display(frame);
 }

 avcodec_free_context(&c);
 avformat_free_context(av_format);
 av_frame_free(&frame);
 av_packet_free(&pkt);

 return 0;
}
</iostream>




For completeness, this is the
display
function, using openCV

void display(const AVFrame* frame) {
 static std::vector yuv_buffer;
 yuv_buffer.resize(frame->linesize[0] * 3 / 2 * frame->width);
 cv::Mat mYUV(frame->height * 3 / 2, frame->width, CV_8UC1, yuv_buffer.data(), frame->linesize[0]);
 memcpy(mYUV.ptr(), frame->data[0], frame->linesize[0] * frame->height);
 //cv::imshow("grayscale", mYUV.rowRange(0, frame->height));
 //cv::imshow("u", cv::Mat(frame->height / 2, frame->width / 2, CV_8UC1, frame->data[1], frame->linesize[1]));
 //cv::imshow("v", cv::Mat(frame->height / 2, frame->width / 2, CV_8UC1, frame->data[2], frame->linesize[2]));

 int dest_row = frame->height;
 for (int j = 0; j < frame->height / 2; j++) {
 memcpy(mYUV.ptr(dest_row), frame->data[1] + frame->linesize[1] * j, frame->width);
 j++;
 memcpy(mYUV.ptr(dest_row) + frame->width / 2, frame->data[1] + frame->linesize[1] * j, frame->width);
 dest_row++;
 }
 for (int j = 0; j < frame->height / 2; j++) {
 memcpy(mYUV.ptr(dest_row), frame->data[2] + frame->linesize[2] * j, frame->width);
 j++;
 memcpy(mYUV.ptr(dest_row) + frame->width / 2, frame->data[2] + frame->linesize[2] * j, frame->width);
 dest_row++;
 }
 cv::Mat mRGB(frame->height, frame->width, CV_8UC3);
 cvtColor(mYUV, mRGB, cv::COLOR_YUV2BGR_I420, 3);
 cv::imshow("Video", mRGB);
 cv::waitKey(0);
}




Note : The
AVFrame
->cv::Mat
converter is now available in corrected version as answer here.