Recherche avancée

Médias (1)

Mot : - Tags -/ticket

Autres articles (99)

  • MediaSPIP 0.1 Beta version

    25 avril 2011, par

    MediaSPIP 0.1 beta is the first version of MediaSPIP proclaimed as "usable".
    The zip file provided here only contains the sources of MediaSPIP in its standalone version.
    To get a working installation, you must manually install all-software dependencies on the server.
    If you want to use this archive for an installation in "farm mode", you will also need to proceed to other manual (...)

  • HTML5 audio and video support

    13 avril 2011, par

    MediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
    The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
    For older browsers the Flowplayer flash fallback is used.
    MediaSPIP allows for media playback on major mobile platforms with the above (...)

  • ANNEXE : Les plugins utilisés spécifiquement pour la ferme

    5 mars 2010, par

    Le site central/maître de la ferme a besoin d’utiliser plusieurs plugins supplémentaires vis à vis des canaux pour son bon fonctionnement. le plugin Gestion de la mutualisation ; le plugin inscription3 pour gérer les inscriptions et les demandes de création d’instance de mutualisation dès l’inscription des utilisateurs ; le plugin verifier qui fournit une API de vérification des champs (utilisé par inscription3) ; le plugin champs extras v2 nécessité par inscription3 (...)

Sur d’autres sites (7302)

  • Why is ffmpeg's hstack so much slower than overlay and pad ?

    27 janvier 2021, par cgenco

    I'm using ffmpeg to stitch together two videos of people chatting into a video with each of them side-by-side, like this :

    


    left.mp4 + right.mp4 = out.mp4

    


    Here's the command I'm currently using to get this done, which runs at 2.5x on my 13" M1 MacBook Pro :

    


    ffmpeg -y -i left.mp4 -i right.mp4 -filter_complex "
  [0:v] crop=w=in_w/2 [croppedLeft];
  [1:v][1:v] overlay=x=overlay_w/4 [shiftedRight];
  [shiftedRight][croppedLeft] overlay [vout];
  [0:a][1:a] amix [aout]
" -map "[vout]" -map "[aout]" -ac 2 out.mp4


    


    This command crops the left video to half of its original width (cropping so the video is centered), then shifts the right video a quarter of its width to the right, then overlays the left video on the left half of the output merged with the shifted right video.

    


    One day on my weekly fun-time read-through the FFmpeg filters documentation I stumbled on a filter named hstack, which is described as being "faster than using overlay and pad filter to create same output."

    


    My ex wife can affirm that there are few higher priorities in my life than going faster, so I altered my ffmpeg script to use hstack instead of two overlays :

    


    ffmpeg -y -i left.mp4 -i right.mp4 -filter_complex "
  [0:v] crop=w=in_w/2 [croppedLeft];
  [1:v] crop=w=in_w/2 [croppedRight];
  [croppedLeft][croppedRight] vstack [vout];
  [0:a][1:a] amix [aout]
" -map "[vout]" -map "[aout]" -ac 2 out.mp4


    


    ...but that command runs painfully slowly, like 0.1x. It takes multiple minutes to render a single second.

    


    So uhhh what's going on here ? Why is hstack taking so long when it's supposed to be faster ?

    


    I've tried this on both the M1 native build from OSXExperts (version N-99816-g3da35b7) and the standard ffmpeg from brew and hstack is just as slow on each.

    


  • Inconsistent behavior when placing adjacent video overlays

    24 juillet 2023, par abingham

    This script uses ffmpeg to overlay a red [1] and blue [2] stream atop a yellow [0] stream. The intent is that the red stream will be exactly adjacent to the blue stream ; at frame 1731 the output video should transition immediately from pure red to pure blue :

    


    # These values result in a blank frame where the videos join at around 58s
LENGTH=1800
SPLIT_AT=1731

# These values work just fine.
# LENGTH=1800
# SPLIT_AT=2

ffmpeg \
-f lavfi -r 30 -i "color=c=yellow:s=1920x1080" \
-f lavfi -r 30 -i "color=c=990000:s=1920x1080" \
-f lavfi -r 30 -i "color=c=000099:s=1920x1080" \
-filter_complex \
"[1]concat=n=1[red_concat];"\
"[red_concat]tpad=start=0[red_pad];"\
"[0][red_pad]overlay=enable=between(n\,0\,$SPLIT_AT):eof_action=repeat[red_overlay];"\
"[2]concat=n=1[blue_concat];"\
"[blue_concat]tpad=start=$SPLIT_AT[blue_pad];"\
"[red_overlay][blue_pad]overlay=enable=between(n\,$SPLIT_AT\,$LENGTH):eof_action=repeat[blue_overlay];"\
"[blue_overlay]trim=end_frame=$LENGTH[full];" \
 -map "[full]" video.mp4 -y


    


    However, what I see is that there is a single black frame between the pure red and pure blue. This black seems to be coming from the frames that tpad adds.

    


    What I don't understand is why I'm seeing that black frame. As far as I can tell, my uses of tpad and overlay/enable-between should produce a seamless transition from red to blue. And indeed, if I change the SPLIT_AT value to almost anything else I get exactly that ; the value 1731 is "special" in some way.

    


    Can anyone explain what's going on ? Maybe more importantly, can anyone reproduce this problem ?

    


    My first thought was that I've got some form of fencepost error in how I'm using tpad or enable-between. However, that doesn't explain why the script works for most values but not 1731. This difference in results feels like there's a rounding error somewhere, but since I'm dealing purely in terms of integer frame numbers, I don't see how I could be introducing them.

    


    A few notes :

    


      

    • This is a drastically reduced example from a larger video compositing system. Things like the use of concat may appear unnecessary, but they're representative of what the larger system is doing.

      


    • 


    • I'm using ffmpeg-6.0 installed via homebrew on an M1 macbook

      


    • 


    


  • How to convert png images to a HEVC(h265) video with alpha

    20 décembre 2022, par ImWH

    I want to convert a sequence of png images to a HEVC video with alpha, I would like to use ffmpeg to do this work and find the following command(images name are p01.png, p02.png...) :

    


    ffmpeg -framerate 10 -i p%02d.png -c:v hevc_videotoolbox -require_sw 1 -allow_sw 1 -alpha_quality 0.5 -vtag hvc1 hevc.mov


    


    the print is here :

    


    ffmpeg version 5.1.2 Copyright (c) 2000-2022 the FFmpeg developers
  built with Apple clang version 14.0.0 (clang-1400.0.29.202)
  configuration: --prefix=/opt/homebrew/Cellar/ffmpeg/5.1.2_1 --enable-shared --enable-pthreads --enable-version3 --cc=clang --host-cflags= --host-ldflags= --enable-ffplay --enable-gnutls --enable-gpl --enable-libaom --enable-libbluray --enable-libdav1d --enable-libmp3lame --enable-libopus --enable-librav1e --enable-librist --enable-librubberband --enable-libsnappy --enable-libsrt --enable-libtesseract --enable-libtheora --enable-libvidstab --enable-libvmaf --enable-libvorbis --enable-libvpx --enable-libwebp --enable-libx264 --enable-libx265 --enable-libxml2 --enable-libxvid --enable-lzma --enable-libfontconfig --enable-libfreetype --enable-frei0r --enable-libass --enable-libopencore-amrnb --enable-libopencore-amrwb --enable-libopenjpeg --enable-libspeex --enable-libsoxr --enable-libzmq --enable-libzimg --disable-libjack --disable-indev=jack --enable-videotoolbox --enable-neon
  libavutil      57. 28.100 / 57. 28.100
  libavcodec     59. 37.100 / 59. 37.100
  libavformat    59. 27.100 / 59. 27.100
  libavdevice    59.  7.100 / 59.  7.100
  libavfilter     8. 44.100 /  8. 44.100
  libswscale      6.  7.100 /  6.  7.100
  libswresample   4.  7.100 /  4.  7.100
  libpostproc    56.  6.100 / 56.  6.100
Input #0, image2, from 'p%02d.png':
  Duration: 00:00:09.60, start: 0.000000, bitrate: N/A
  Stream #0:0: Video: png, rgba(pc), 375x812 [SAR 3779:3779 DAR 375:812], 10 fps, 10 tbr, 10 tbn
Stream mapping:
  Stream #0:0 -> #0:0 (png (native) -> hevc (hevc_videotoolbox))
Press [q] to stop, [?] for help
[1]    16348 trace trap  ffmpeg -framerate 10 -i p%02d.png -c:v hevc_videotoolbox -require_sw 1  1  0.


    


    It seems like there is no error and the video has been generated but the size of video is always 0kb, I try to change the parameter for command but nothing changed, so is there something working with the command ? or Is there is a simple way to create the HEVC video using png image ? Thanks

    


    My Environment :
Apple silicon(MacBook with M1 Pro clip) Mac OS 13.0
ffmpeg : 5.1.2