Recherche avancée

Médias (1)

Mot : - Tags -/copyleft

Autres articles (35)

  • Les autorisations surchargées par les plugins

    27 avril 2010, par

    Mediaspip core
    autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs

  • Contribute to a better visual interface

    13 avril 2011

    MediaSPIP is based on a system of themes and templates. Templates define the placement of information on the page, and can be adapted to a wide range of uses. Themes define the overall graphic appearance of the site.
    Anyone can submit a new graphic theme or template and make it available to the MediaSPIP community.

  • HTML5 audio and video support

    13 avril 2011, par

    MediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
    The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
    For older browsers the Flowplayer flash fallback is used.
    MediaSPIP allows for media playback on major mobile platforms with the above (...)

Sur d’autres sites (5420)

  • ffmpeg converting video to images while video file is being written

    15 octobre 2015, par user3398227

    Hopefully an easy question for an ffmpeg expert !

    I’m currently converting large (+6GB) mpeg video into an image sequence - which is working well using the below ffmpeg command :

    ffmpeg -i "input.mpeg" -vf - fps=fps=2 -f image2 -qscale 1 -s 1026x768 "output%6d.jpg"

    however i have to wait for the file to finish being written to disk before i kick off ffmpeg - but this takes a good hour or so to finish writing, but what i’ve noticed is that ffmpeg can start reading the file while its being written to disk - the only snag here is it gets to the end of the file and stops before the file has finished being written...

    Question is, is there a way that ffmpeg can convert to an image sequence at the same pace the video is being written (and not exit out ?)... or know to wait for the next frame to be written from the source. (unfortunately the input doesn’t support streaming, I only get a network drive and file to work off.. ) I thought i read somewhere that ffmpeg can process at the video frame rate but cant seem to find this command for love or money in the doco !!

    Thanks !

  • Demuxing .AVI Video Muxed using mpegts on 3 Video Streams

    11 juin 2022, par nwf1115

    I have output from a Python program that comes out as a single .AVI that contains 3 video streams, however, it seems that they are all combined. The videos are not concatenated, but they are interlaced. The code that produces it is pretty abstract and the only thing I can tell is that they used Gstreamer to combine the videos. I can pick out H.264 encoding and mpegts and output as an .AVI. Using ffprobe on the file I get the following output :

    


    Input #0, mpegts, from 'output.avi':
  Duration: 00:05:00.07, start: 12045.450000, bitrate: 33695 kb/s
  Program 1
    Stream #0:0[0xdd]: Video: h264 (Main) ([27][0][0][0] / 0x001B), yuv420p(progressive), >1920x1080 [SAR 1:1 DAR 16:9], 15 fps, 15 tbr, 90k tbn, 30 tbc


    


    I'm needing to separate the videos contained in this .AVI file, but I'm having trouble doing so using ffmpeg since there are not separate streams for each video.

    


    Does anyone have any tips on how to do this using ffmpeg or gstreamer ?

    


    Update : Upon further experimentation, it does not seem that the frames are interlaced, rather the stream of data was in some way, combined in some manner.

    


  • FFMPEG : How to combine video and image (video template)

    22 février 2024, par clo5ure

    Goal

    


    I have a video and an image (a template for the video) that I'm trying to combine into one output video (1080w x 1920h - 9:16 aspect ratio).

    


      

    • Input video - 1920x1080
    • 


    • Input image - 1080x1920
    • 


    • Output video - 1080x1920
    • 


    


    This image shows what I'm trying to accomplish. The two orange sections are the input image - it's a single .png with a transparent section in the middle for the video.

    


    As mentioned in the title, I'm trying to accomplish this using FFMPEG. More specifically, I'm using the fluent-ffmpeg npm package.

    


    Current Status

    


    I can read in both inputs just fine but I have issues getting the two to play nicely with one another.

    


    If I get the overlay working then my output video is 1920x1080 instead of 1080x1920.

    


    If I get the output video dimensions right, then the video is either stretched or I get errors adding my overlay.

    


    Code

    


    Here's what I have at the moment. I'm happy to answer any questions. Thank you in advance for taking a look :)

    


    var ffmpeg = require('fluent-ffmpeg');
var command = ffmpeg();
var timemark = null;

command
  .on('end', onEnd )
  .on('progress', onProgress)
  .on('error', onError)
  .input('./input-video.mp4')
  .input('./template.png')
  .complexFilter([
    {
      filter: 'scale',
      options: { width: 1080, height: 1920 }
    },
    // {
    //   filter: 'overlay',
    //   options: { x: 100, y: 100 }
    // },
  ])
  .outputFps(30)
  .output('./output-video.mp4')
  .run();

/* Misc */

function onProgress(progress){
  if (progress.timemark != timemark) {
    timemark = progress.timemark;
    console.log('Time mark: ' + timemark + "...");
  }
}

function onError(err, stdout, stderr) {
  console.log('Cannot process video: ' + err.message);
}

function onEnd() {
  console.log('Finished processing');
}