Recherche avancée

Médias (1)

Mot : - Tags -/publishing

Autres articles (51)

  • Les autorisations surchargées par les plugins

    27 avril 2010, par

    Mediaspip core
    autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

  • Des sites réalisés avec MediaSPIP

    2 mai 2011, par

    Cette page présente quelques-uns des sites fonctionnant sous MediaSPIP.
    Vous pouvez bien entendu ajouter le votre grâce au formulaire en bas de page.

Sur d’autres sites (7662)

  • ffmpeg converting video to images while video file is being written

    15 octobre 2015, par user3398227

    Hopefully an easy question for an ffmpeg expert !

    I’m currently converting large (+6GB) mpeg video into an image sequence - which is working well using the below ffmpeg command :

    ffmpeg -i "input.mpeg" -vf - fps=fps=2 -f image2 -qscale 1 -s 1026x768 "output%6d.jpg"

    however i have to wait for the file to finish being written to disk before i kick off ffmpeg - but this takes a good hour or so to finish writing, but what i’ve noticed is that ffmpeg can start reading the file while its being written to disk - the only snag here is it gets to the end of the file and stops before the file has finished being written...

    Question is, is there a way that ffmpeg can convert to an image sequence at the same pace the video is being written (and not exit out ?)... or know to wait for the next frame to be written from the source. (unfortunately the input doesn’t support streaming, I only get a network drive and file to work off.. ) I thought i read somewhere that ffmpeg can process at the video frame rate but cant seem to find this command for love or money in the doco !!

    Thanks !

  • Demuxing .AVI Video Muxed using mpegts on 3 Video Streams

    11 juin 2022, par nwf1115

    I have output from a Python program that comes out as a single .AVI that contains 3 video streams, however, it seems that they are all combined. The videos are not concatenated, but they are interlaced. The code that produces it is pretty abstract and the only thing I can tell is that they used Gstreamer to combine the videos. I can pick out H.264 encoding and mpegts and output as an .AVI. Using ffprobe on the file I get the following output :

    


    Input #0, mpegts, from 'output.avi':
  Duration: 00:05:00.07, start: 12045.450000, bitrate: 33695 kb/s
  Program 1
    Stream #0:0[0xdd]: Video: h264 (Main) ([27][0][0][0] / 0x001B), yuv420p(progressive), >1920x1080 [SAR 1:1 DAR 16:9], 15 fps, 15 tbr, 90k tbn, 30 tbc


    


    I'm needing to separate the videos contained in this .AVI file, but I'm having trouble doing so using ffmpeg since there are not separate streams for each video.

    


    Does anyone have any tips on how to do this using ffmpeg or gstreamer ?

    


    Update : Upon further experimentation, it does not seem that the frames are interlaced, rather the stream of data was in some way, combined in some manner.

    


  • FFMPEG : How to combine video and image (video template)

    22 février 2024, par clo5ure

    Goal

    


    I have a video and an image (a template for the video) that I'm trying to combine into one output video (1080w x 1920h - 9:16 aspect ratio).

    


      

    • Input video - 1920x1080
    • 


    • Input image - 1080x1920
    • 


    • Output video - 1080x1920
    • 


    


    This image shows what I'm trying to accomplish. The two orange sections are the input image - it's a single .png with a transparent section in the middle for the video.

    


    As mentioned in the title, I'm trying to accomplish this using FFMPEG. More specifically, I'm using the fluent-ffmpeg npm package.

    


    Current Status

    


    I can read in both inputs just fine but I have issues getting the two to play nicely with one another.

    


    If I get the overlay working then my output video is 1920x1080 instead of 1080x1920.

    


    If I get the output video dimensions right, then the video is either stretched or I get errors adding my overlay.

    


    Code

    


    Here's what I have at the moment. I'm happy to answer any questions. Thank you in advance for taking a look :)

    


    var ffmpeg = require('fluent-ffmpeg');
var command = ffmpeg();
var timemark = null;

command
  .on('end', onEnd )
  .on('progress', onProgress)
  .on('error', onError)
  .input('./input-video.mp4')
  .input('./template.png')
  .complexFilter([
    {
      filter: 'scale',
      options: { width: 1080, height: 1920 }
    },
    // {
    //   filter: 'overlay',
    //   options: { x: 100, y: 100 }
    // },
  ])
  .outputFps(30)
  .output('./output-video.mp4')
  .run();

/* Misc */

function onProgress(progress){
  if (progress.timemark != timemark) {
    timemark = progress.timemark;
    console.log('Time mark: ' + timemark + "...");
  }
}

function onError(err, stdout, stderr) {
  console.log('Cannot process video: ' + err.message);
}

function onEnd() {
  console.log('Finished processing');
}