Recherche avancée

Médias (1)

Mot : - Tags -/bug

Autres articles (45)

  • La file d’attente de SPIPmotion

    28 novembre 2010, par

    Une file d’attente stockée dans la base de donnée
    Lors de son installation, SPIPmotion crée une nouvelle table dans la base de donnée intitulée spip_spipmotion_attentes.
    Cette nouvelle table est constituée des champs suivants : id_spipmotion_attente, l’identifiant numérique unique de la tâche à traiter ; id_document, l’identifiant numérique du document original à encoder ; id_objet l’identifiant unique de l’objet auquel le document encodé devra être attaché automatiquement ; objet, le type d’objet auquel (...)

  • Contribute to documentation

    13 avril 2011

    Documentation is vital to the development of improved technical capabilities.
    MediaSPIP welcomes documentation by users as well as developers - including : critique of existing features and functions articles contributed by developers, administrators, content producers and editors screenshots to illustrate the above translations of existing documentation into other languages
    To contribute, register to the project users’ mailing (...)

  • Use, discuss, criticize

    13 avril 2011, par

    Talk to people directly involved in MediaSPIP’s development, or to people around you who could use MediaSPIP to share, enhance or develop their creative projects.
    The bigger the community, the more MediaSPIP’s potential will be explored and the faster the software will evolve.
    A discussion list is available for all exchanges between users.

Sur d’autres sites (7847)

  • How to read remote video on Amazon S3 using ffmpeg

    19 septembre 2012, par virtualize

    I need to create poster frames from videos hosted on Amazon S3 via ffmpeg.

    So is there a way to use the remote video file directly in ffmpeg command line like this :
    ffmpeg -i "http://bucket.s3.amazonaws.com/video.mp4" -ss 00:00:10 -vframes 1 -f image2 "image%03d.jpg"

    ffmpeg just returns :

    http://bucket.s3.amazonaws.com/video.mp4: I/O error occurred<br />
    Usually that means that input file is truncated and/or corrupted.

    I also tried forcing ffmpeg to use the videos mp4 container for reading :
    ffmpeg -f mp4 -i "http://bucket.s3.amazonaws.com/video.mp4" ...
    But no luck.

    Wget this video from S3 and processing it locally works fine of course,
    as well as reading the file remotely from other 'standard' http servers.
    So I know that ffmpeg supports remote file reading, but why not on S3 ?

  • Amazon Alexa Audio Encoding- Few audios are not playing [closed]

    3 mars 2023, par Vijayanath Viswanathan

    I am encoding audio for Alexa audio using ffmpeg like below,

    &#xA;&#xA;

    ffmpeg -i  -ac 2 -codec:a libmp3lame -b:a 48k -ar 16000 &#xA;

    &#xA;&#xA;

    The problem is few of the audios are playing properly but few are not. I am using same Project Rate and Quality (Project Rate 16000 and Quality to 48 kbps) for all audios which needs to be converted. Anybody knows is there any basic quality for source.mp3 to encode to Project Rate 16000 and Quality to 48 kbps ?

    &#xA;&#xA;

    The response I am getting from alexa for faulty file is, "There is a problem with skill response".

    &#xA;

  • Best practices for developing scalable video transcoding server on Amazon Web Services ?

    6 septembre 2016, par undefined

    What do people think are the most important issues when developing an application that is going to allow users to upload video and images to a server and have them transcoded by FFMPEG and stored in amazon S3 ? I have a couple of options ;

    1) install FFMPEG on the same server that handles file uploads, when a video is uploaded and stored on EC2 instance, call FFMPEG to convert it then when done, write the file to S3 bucket and dispose of the original.

    How scalable is this ? What happens when many users upload at the same time ? How do I manage multiple processes at once ? How do I know when to start another instance and load balance this configuration ?

    2) Have one server for processing uploads (updating database, renaming files etc) and one server for doing transcoding. Again what is the best way to manage multiple processes ? should I be looking at Amazon SQS for this ? Can I tell the transcoding server to get the file from the upload server or should I copy the file to the transcoding server ? Should I just store all files on S3 and SQS can read from there. I am trying to have as little traffic as possible.

    I am running a linux box as the upload server and have FFMPEG running on this.

    Any advice on best practices for setting up such a configuration would be appreciated. Many thanks