Recherche avancée

Médias (1)

Mot : - Tags -/pirate bay

Autres articles (98)

  • Personnaliser en ajoutant son logo, sa bannière ou son image de fond

    5 septembre 2013, par

    Certains thèmes prennent en compte trois éléments de personnalisation : l’ajout d’un logo ; l’ajout d’une bannière l’ajout d’une image de fond ;

  • Les autorisations surchargées par les plugins

    27 avril 2010, par

    Mediaspip core
    autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs

  • Support de tous types de médias

    10 avril 2011

    Contrairement à beaucoup de logiciels et autres plate-formes modernes de partage de documents, MediaSPIP a l’ambition de gérer un maximum de formats de documents différents qu’ils soient de type : images (png, gif, jpg, bmp et autres...) ; audio (MP3, Ogg, Wav et autres...) ; vidéo (Avi, MP4, Ogv, mpg, mov, wmv et autres...) ; contenu textuel, code ou autres (open office, microsoft office (tableur, présentation), web (html, css), LaTeX, Google Earth) (...)

Sur d’autres sites (5787)

  • how to create video from ffmpeg ?

    4 mai 2020, par Hong

    I have 10 images, 2 videos(mov, mp4) and 1 audio.

    



    My plan is

    



      

    1. create images to video (images.mp4)
    2. 


    



    ffmpeg -framerate 0.75 -pattern_type glob -i '*.png' \
 -c:v libx264 -vf "format=yuv420p,pad=ceil(iw/2)*2:ceil(ih/2)*2" -r 30 -pix_fmt yuv420p out.mp4  


    



      

    1. concat videos
filelist :
    2. 


    



    file 'images.mp4'
file 'video1.mov'
file 'video2.mp4'


    



    2-1 )
ffmpeg -f concat -safe 0 -i filelist.txt -auto_convert 1 -c copy output.mp4

    



    2-2 ) first. i'm encode mov to mp4. and change filelist.txt

    



    file 'images.mp4'
file 'video1.mp4'
file 'video2.mp4'
ffmpeg -f concat -safe 0 -i filelist.txt -auto_convert 1 -c copy output.mp4


    



      

    1. add background music.
It has not yet reached this stage.
It's blocked in plan 2.
    2. 


    



    this plan is right ?
I'm not sure about this plan.

    



    Step 2 output is strange.
The frame of the video No. 1 and No. 2 is not smooth.

    



    How can I create video from images and meger another video.

    


  • Attempting to compile FFmpeg 4.2.3 statically for Windows 10 (x86_64), but binaries asks for missing DLLs

    29 mai 2020, par Expectator

    I am using Msys MinGW (x86_64) and pulled a snapshot of the latest major release of FFmpeg off of their website. Here is my ./configure options. I plan to use the binaries on both the computer that I compiled it on, and other Windows computers that I own.

    



    ./configure --enable-libaom --enable-avisynth --enable-chromaprint --enable-libdav1d --enable-libdavs2 --enable-libgme --enable-libmfx --enable-libkvazaar --enable-libmp3lame --enable-libilbc --enable-libvpx --enable-libmodplug --enable-version3 --enable-nonfree --enable-libopencore-amrnb --enable-libopencore-amrwb --enable-libvo-amrwbenc --enable-libfdk-aac --enable-libopenh264 --enable-libopenjpeg --enable-nvenc --enable-nvdec --enable-cuda --enable-cuvid --enable-libtwolame --enable-vapoursynth --enable-libwavpack --enable-libx264 --enable-libx265 --enable-libxavs --enable-libxavs2 --enable-gpl --enable-static --disable-shared


    



    Output of configure script (pastebin)

    



    Output of uname -a (in Msys)

    



    MINGW64_NT-10.0-18362 <scrubbed> 3.1.4-340.x86_64 2020-05-22 08:28 UTC x86_64 Msys&#xA;</scrubbed>

    &#xA;&#xA;

    The issue that I'm facing is that despite passing the options --enable-static and --disable-shared, the executables generated still require libchromaprint.dll, libfdk-aac-2.dll, and libgme.dll to run. What I expected was that FFmpeg would execute independently of any DLL files since I passed those options to ./configure.

    &#xA;

  • How to extract motion vectors from h264 without a full decode on the CPU

    25 septembre 2020, par Adrian May

    I'm trying to use my nose as a pointing device. The plan is to encode the video stream from a webcam pointed at my face as h264 or the like, get the motion vectors, cook the numbers a bit and chuck them into /dev/uinput to make the mouse pointer move about. The uinput bit was easy.

    &#xA;

    This has to work with zero discernable latency. This, for instance :

    &#xA;

    #!/bin/bash&#xA;[ -p pipe.mkv ] || mkfifo pipe.mkv&#xA;ffmpeg -y -rtbufsize 1M -s 640x360 -vcodec mjpeg -i /dev/video0 -c h264_nvenc pipe.mkv &amp;&#xA;ffplay -flags2 &#x2B;export_mvs -vf codecview=mv=pf&#x2B;bf&#x2B;bb pipe.mkv&#xA;

    &#xA;

    shows that the vectors are there but with a latency of several seconds which is unusable in a mouse. I know that the first ffmpeg step is working very fast by using the GPU, so either the pipe or the h264 decode in the second step is introducing the latency.

    &#xA;

    I tried MV Tractus (same as mpegflow I think) in a similar pipe arrangement and it was also very slow. They do a full h264 decode on the CPU and I think that's the problem cos I can see them imposing a lot of load on one CPU. If the pipe had caused the delay by buffering badly then the CPU wouldn't have been loaded. I guess ffplay also did the decoding on the CPU and I couldn't persuade it not to, but it only wants to draw arrows which are no use to me.

    &#xA;

    I think there are several approaches, and I'd like advice on which would be best, or if there's something even better I don't know about. I could :

    &#xA;

      &#xA;
    • Decode in hardware and get the motion vectors. So far this has failed. I tried combining ffmpeg's extract_mvs.c and hw_decode.c samples but no motion vectors turn up. vdpau is the only decoder I got working on my linux box. I have a nvidia gpu.
    • &#xA;

    • Do a minimal parse of the h264 to fish out the motion vectors only, ignoring all the other data. I think this would mean putting some kind of "motion only" option in libav's parser, but I'm not at all familiar with that code.
    • &#xA;

    • Find some other h264 parsing library that has said option and also unpacks the container.
    • &#xA;

    • Forget about hardware accelerated encoding and use a stripped down encoder to make only the motion vectors on either CPU or GPU. I suspect this would be slow cos I think calculating the motion vectors is the hardest part of the algorithm.
    • &#xA;

    &#xA;

    I'm tending towards the second option but I need some help figuring out where in the libav code to do it.

    &#xA;