Recherche avancée

Médias (91)

Autres articles (74)

  • La file d’attente de SPIPmotion

    28 novembre 2010, par

    Une file d’attente stockée dans la base de donnée
    Lors de son installation, SPIPmotion crée une nouvelle table dans la base de donnée intitulée spip_spipmotion_attentes.
    Cette nouvelle table est constituée des champs suivants : id_spipmotion_attente, l’identifiant numérique unique de la tâche à traiter ; id_document, l’identifiant numérique du document original à encoder ; id_objet l’identifiant unique de l’objet auquel le document encodé devra être attaché automatiquement ; objet, le type d’objet auquel (...)

  • Personnaliser en ajoutant son logo, sa bannière ou son image de fond

    5 septembre 2013, par

    Certains thèmes prennent en compte trois éléments de personnalisation : l’ajout d’un logo ; l’ajout d’une bannière l’ajout d’une image de fond ;

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

Sur d’autres sites (5389)

  • FFmpeg - Generating waveform image from webm file not working due to overlay filter

    18 juillet 2018, par jfizz

    I am attempting to generate a waveform image from a webm file. I am using the overlay filter to overlay the waveform atop a background. The waveform does not show whenever the overlay filter is used. The failing command :

    ffmpeg -i big_buck_bunny.webm -filter_complex "showwavespic=s=640x120[fg];color=s=640x120:color=#ff0000[bg];[bg][fg]overlay=format=auto" -frames:v 1 output.jpg

    The result :

    enter image description here

    I have also tried using an image as the background with the same result :

    ffmpeg -i big_buck_bunny.webm -i bg.jpg -filter_complex "showwavespic=s=640x120[fg];[1:v][fg]overlay=format=auto" -frames:v 1 output.jpg

    Generating a waveform without a background works as expected :

    ffmpeg -i big_buck_bunny.webm -filter_complex "showwavespic=s=640x120" -frames:v 1 output.jpg

    enter image description here

    Here is the video I am using.

    I have tried various webm files with the same result. I also tested some .mp4,.mov,.ogv files which worked as expected. I am using the latest ffmpeg version, below is my banner.

    ffmpeg version 4.0.1 Copyright (c) 2000-2018 the FFmpeg developers
    built with Apple LLVM version 9.1.0 (clang-902.0.39.2)
    configuration: --prefix=/usr/local/Cellar/ffmpeg/4.0.1 --enable-shared --
    enable-pthreads --enable-version3 --enable-hardcoded-tables --enable-avresample --cc=clang --host-cflags= --host-ldflags= --enable-gpl --enable-libmp3lame --enable-libx264 --enable-libxvid --enable-opencl --enable-videotoolbox --disable-lzma
    libavutil      56. 14.100 / 56. 14.100
    libavcodec     58. 18.100 / 58. 18.100
    libavformat    58. 12.100 / 58. 12.100
    libavdevice    58.  3.100 / 58.  3.100
    libavfilter     7. 16.100 /  7. 16.100
    libavresample   4.  0.  0 /  4.  0.  0
    libswscale      5.  1.100 /  5.  1.100
    libswresample   3.  1.100 /  3.  1.100
    libpostproc    55.  1.100 / 55.  1.100
  • Split 2 video FFMpeg

    19 juillet 2018, par Евгений Бреус

    can you help me split 2 videos
    enter image description here

    ffmpeg -i input1.mp4 -i input2.mp4 -filter_complex "[0:v][1:v]vstack[v];[0:a][1:a]amerge=inputs=2[a]" -map "[v]" -map "[a]" -ac 2 output.mp4

    Don’t work

    My complete log

       C:\Users\forumrulez\Desktop\POST>ffmpeg -i input1.mp4 -i input2.mp4 -filter_complex "[0:v][1:v]vstack[v];[0:a][1:a]amerge=inputs=2[a]" -map "[v]" -map "[a]" -ac 2 output.mp4
    ffmpeg version N-91447-g0bd48ab2d9 Copyright (c) 2000-2018 the FFmpeg developers
     built with gcc 7.3.1 (GCC) 20180709
     configuration: --enable-gpl --enable-version3 --enable-sdl2 --enable-bzlib --enable-fontconfig --enable-gnutls --enable-iconv --enable-libass --enable-libbluray --enable-libfreetype --enable-libmp3lame --enable-libopencore-amrnb --enable-libopencore-amrwb --enable-libopenjpeg --enable-libopus --enable-libshine --enable-libsnappy --enable-libsoxr --enable-libtheora --enable-libtwolame --enable-libvpx --enable-libwavpack --enable-libwebp --enable-libx264 --enable-libx265 --enable-libxml2 --enable-libzimg --enable-lzma --enable-zlib --enable-gmp --enable-libvidstab --enable-libvorbis --enable-libvo-amrwbenc --enable-libmysofa --enable-libspeex --enable-libxvid --enable-libaom --enable-libmfx --enable-amf --enable-ffnvcodec --enable-cuvid --enable-d3d11va --enable-nvenc --enable-nvdec --enable-dxva2 --enable-avisynth
     libavutil      56. 18.102 / 56. 18.102
     libavcodec     58. 21.104 / 58. 21.104
     libavformat    58. 17.101 / 58. 17.101
     libavdevice    58.  4.101 / 58.  4.101
     libavfilter     7. 25.100 /  7. 25.100
     libswscale      5.  2.100 /  5.  2.100
     libswresample   3.  2.100 /  3.  2.100
     libpostproc    55.  2.100 / 55.  2.100
    Input #0, mov,mp4,m4a,3gp,3g2,mj2, from 'input1.mp4':
     Metadata:
       major_brand     : isom
       minor_version   : 512
       compatible_brands: isomiso2avc1mp41
       encoder         : Lavf58.17.101
     Duration: 00:01:00.08, start: 0.000000, bitrate: 586 kb/s
       Stream #0:0(und): Video: h264 (High) (avc1 / 0x31637661), yuv420p, 260x480 [SAR 128:39 DAR 16:9], 428 kb/s, 24 fps, 24 tbr, 12288 tbn, 48 tbc (default)
       Metadata:
         handler_name    : VideoHandler
       Stream #0:1(und): Audio: aac (LC) (mp4a / 0x6134706D), 44100 Hz, stereo, fltp, 152 kb/s (default)
       Metadata:
         handler_name    : SoundHandler
    Input #1, mov,mp4,m4a,3gp,3g2,mj2, from 'input2.mp4':
     Metadata:
       major_brand     : isom
       minor_version   : 512
       compatible_brands: isomiso2avc1mp41
       encoder         : Lavf58.17.101
     Duration: 00:01:00.08, start: 0.000000, bitrate: 703 kb/s
       Stream #1:0(und): Video: h264 (High) (avc1 / 0x31637661), yuv420p, 360x480 [SAR 64:27 DAR 16:9], 545 kb/s, 24 fps, 24 tbr, 12288 tbn, 48 tbc (default)
       Metadata:
         handler_name    : VideoHandler
       Stream #1:1(und): Audio: aac (LC) (mp4a / 0x6134706D), 44100 Hz, stereo, fltp, 152 kb/s (default)
       Metadata:
         handler_name    : SoundHandler
    File 'output.mp4' already exists. Overwrite ? [y/N] y
    Stream mapping:
     Stream #0:0 (h264) -> vstack:input0
     Stream #0:1 (aac) -> amerge:in0
     Stream #1:0 (h264) -> vstack:input1
     Stream #1:1 (aac) -> amerge:in1
     vstack -> Stream #0:0 (libx264)
     amerge -> Stream #0:1 (aac)
    Press [q] to stop, [?] for help
    [Parsed_amerge_1 @ 00000165b1e6af00] No channel layout for input 1
    [Parsed_amerge_1 @ 00000165b1e6af00] Input channel layouts overlap: output layout will be determined by the number of distinct input channels
    [Parsed_vstack_0 @ 00000165b29f1040] Input 1 width 360 does not match input 0 width 260.
    [Parsed_vstack_0 @ 00000165b29f1040] Failed to configure output pad on Parsed_vstack_0
    Error reinitializing filters!
    Failed to inject frame into filter network: Invalid argument
    Error while processing the decoded data for stream #1:0
    Conversion failed!
  • Ubuntu 16.04 + OpenCV + FFMPEG + CUDA don't open default camera

    20 juillet 2018, par Newton Pasqualini Filho

    I am trying to work with accelerated OpenCV using my own build of FFMPEG with CUDA support for GPU video processing using Python.

    My application must support any kind of video devices, including USB and IP cameras.

    I have a config file parsed with ConfigParser that loads the camera device path, as described in OpenCV documentation we can access usb camera by index number and for IP camera we use RTSP protocol so the path could be sometimes 0 or rtsp://192.168.1.2/0.

    To force the use of FFMPEG behind OpenCV I am trying to open the video capture stream like bellow :

    video_capture = cv.VideoCapture()
    stream_opened = video_capture.open(camera_path, cv.CAP_FFMPEG)

    The problem is when I try to open the default USB camera camera_path = 0 the video_capture.open method always returns false, but when it is a RTSP like camera_path = "rtsp://192.168.1.2/1" it open the video capture stream ok.

    When I let OpenCV decide what backend will be used it calls to GStreammer and it opens the capture stream like this :

    video_capture = cv.VideoCapture(camera_path)

    I suspect that FFMPEG does not have any support for capture video from a USB camera, does it ? Is there some missing feature to my ffmpeg build ?

    ffmpeg version N-91487-g1809f1c Copyright (c) 2000-2018 the FFmpeg developers
    built with gcc 5.4.0 (Ubuntu 5.4.0-6ubuntu1~16.04.10) 20160609
    configuration: --prefix=/usr --pkg-config-flags=--static --enable-cuda-sdk --enable-cuvid --enable-libnpp --extra-cflags=-I/usr/local/cuda/include/ --extra-ldflags=-L/usr/local/cuda/lib64/ --nvccflags='-gencode arch=compute_61,code=sm_61 -O2' --enable-gpl --enable-libass --enable-libfdk-aac --enable-libx264 --extra-libs=-lpthread --enable-libx265 --enable-nvenc --enable-nonfree
    libavutil      56. 18.102 / 56. 18.102
    libavcodec     58. 21.105 / 58. 21.105
    libavformat    58. 17.101 / 58. 17.101
    libavdevice    58.  4.101 / 58.  4.101
    libavfilter     7. 26.100 /  7. 26.100
    libswscale      5.  2.100 /  5.  2.100
    libswresample   3.  2.100 /  3.  2.100
    libpostproc    55.  2.100 / 55.  2.100
    Hyper fast Audio and Video encoder
    usage: ffmpeg [options] [[infile options] -i infile]... {[outfile options] outfile}...