Recherche avancée

Médias (1)

Mot : - Tags -/bug

Autres articles (24)

  • La file d’attente de SPIPmotion

    28 novembre 2010, par

    Une file d’attente stockée dans la base de donnée
    Lors de son installation, SPIPmotion crée une nouvelle table dans la base de donnée intitulée spip_spipmotion_attentes.
    Cette nouvelle table est constituée des champs suivants : id_spipmotion_attente, l’identifiant numérique unique de la tâche à traiter ; id_document, l’identifiant numérique du document original à encoder ; id_objet l’identifiant unique de l’objet auquel le document encodé devra être attaché automatiquement ; objet, le type d’objet auquel (...)

  • Contribute to documentation

    13 avril 2011

    Documentation is vital to the development of improved technical capabilities.
    MediaSPIP welcomes documentation by users as well as developers - including : critique of existing features and functions articles contributed by developers, administrators, content producers and editors screenshots to illustrate the above translations of existing documentation into other languages
    To contribute, register to the project users’ mailing (...)

  • Supporting all media types

    13 avril 2011, par

    Unlike most software and media-sharing platforms, MediaSPIP aims to manage as many different media types as possible. The following are just a few examples from an ever-expanding list of supported formats : images : png, gif, jpg, bmp and more audio : MP3, Ogg, Wav and more video : AVI, MP4, OGV, mpg, mov, wmv and more text, code and other data : OpenOffice, Microsoft Office (Word, PowerPoint, Excel), web (html, CSS), LaTeX, Google Earth and (...)

Sur d’autres sites (5425)

  • rtsp to rtmp using ffmpeg or any tool wrapper

    27 août 2017, par Chakri

    I have a requirement where I need to restream the RTSP stream from camera source to RTMP server. I know this may sound a repeated question but my exact scenario is I cannot do it manually over command line with ffmpeg command. I need a wrapper where I receive the rtsp and rtmp url from external source say through REST invocation. Then the code can trigger the ffmpeg restream.

    Basically flow is like this :

    1. Camera source application sends RTSP read event(could be basic HTTP(REST) request with RTSP url, metadata about camera info, serial no etc) to my streamer app

    Ex : /usr/bin/ffmpeg -i rtsp ://10.144.11.22:554/stream1 -f flv rtmp ://10.13.11.121:1935/stream1

    1. Streamer app computes the RTMP server url for corresponding camera and triggers a ffmpeg command to stream RTSP to RTMP

    2. Streamer app triggers above(2) in separate thread and keeps reading the logs for monitoring purpose. Also identifies the end of RTSP stream and sends an update(Example : RTSP END) event to UI

    Now at point(2) I need a suggestion. Here I need a stable wrapper/api which can help. I tried this through some Java wrappers but the process hangs or fails to read the output from ffmpeg. Also I need to handle streams from many cameras where spawning thread for each one could be exhaustive.

    So I am looking for some similar api/wrapper in C++ or Go Lang which might have more closer interaction in handling ffmpeg command.

    Please point if similar issue is addressed elsewhere

  • How to decouple between ffmpeg and rtp server ?

    2 février 2023, par dror

    I have a ffmpeg-based worker that handles video-generation jobs at very high throughput.
Long videos need to be streamed while being generated.

    


    For that purpose, I have introduced a WebRTC server named Janus-Gateway with its streaming plugin, and set the application's output to an rtp:// endpoint at that server (ffmpeg can stream a single stream using the ​RTP protocol).

    


    In order to avoid buffering problems on the other hand, the streaming is done through the ffmpeg's -re option, which means that the stream will be streamed in real-time, i.e. it slows it down to simulate live streaming.

    


    [ffmpeg-based app] (#1)--> [rtp://janus:port @ webrtc server] (#2)--> [webrtc subscribers]


    


    How can I continue processing video jobs at high throughput while streaming the results at real-time speed ? I need somehow to decouple ffmpeg output (stage #1) so that consumers at stage #2 get streams at natural playback speed.

    


  • Capturing two video sources into a single MKV with Java

    7 avril 2014, par Luke

    Problem

    I need my application to let the user record its screen, web cam and audio simultaneously. I have been able to accomplish this using a mix of Sarxos, Robot and Xuggler. But the various output files have different timing and are out of sync when played back synchronously.

    This is not hard to imagine given that the capturing is started asynchronously.


    Possible solution

    I'm thinking that capturing the various sources into an MKV file with multiple video tracks might be a better solution. Some timer could poll the sources in parallel at regular intervals. Or at the very least blank audio/video-data could be inserted before any late capture start, to keep the video's and audio in sync.

    JMF would allow for very easy capturing, but doesn't seem to support MKV and is supposedly very outdated. I question whether it will work properly on current PCs and Macs. The FMJ Project might be an alternative, but MKV-support seems to be restricted to FFMPEG. This would defeat the purpose, as any kind of muxing or re-encoding afterwards would still leave the videos and audio out of sync. I could use Xuggler, as I do now, but it's impossible to find any relevant documentation or hints.

    Still, recording into a single file seems to make sense in order to get the timing straight.


    My question

    Is anyone aware of a Java solution that would allow for encoding two video sources and an audio source into a single MKV container ?