Recherche avancée

Médias (1)

Mot : - Tags -/MediaSPIP

Autres articles (35)

  • Des sites réalisés avec MediaSPIP

    2 mai 2011, par

    Cette page présente quelques-uns des sites fonctionnant sous MediaSPIP.
    Vous pouvez bien entendu ajouter le votre grâce au formulaire en bas de page.

  • Support audio et vidéo HTML5

    10 avril 2011

    MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
    Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
    Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
    Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...)

  • HTML5 audio and video support

    13 avril 2011, par

    MediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
    The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
    For older browsers the Flowplayer flash fallback is used.
    MediaSPIP allows for media playback on major mobile platforms with the above (...)

Sur d’autres sites (4012)

  • FFMPEG How to apply subtitles [ass] on a live video stream

    19 août 2014, par Lewisou

    I am going to develop a camera recording application.

    One of the feature is to apply subtitles while recording. Users can key in subtitles on the camera preview screen when the camera is recording.

    I know I can apply the subtitles after the record is finished by running :

    ffmpeg -i ./file.mp4 -c:v libx264 -c:a copy -vf "ass=t.ass" out.mp4

    but which takes a very long time. So it will be better to do it on the fly (while recording a video).

    Notice : the subtitle itself is also a live-stream.

    Any one help ?

  • ffmpeg with multiple live inputs [closed]

    9 juin 2013, par maddanio

    we are trying to get ffmpeg to combine multiple live streams into one. In principle we got it to work by mosaicing the streams like so :

    ffmpeg -rtsp_transport tcp -r 25 -i rtsp ://root:password@192.168.178.91:554/axis-media/media.amp \
    -rtsp_transport tcp -r 25 -i rtsp ://root:password@192.168.178.92:554/axis-media/media.amp \
    -filter_complex "[0:0]pad=iw*2:ih[a] ;[a][1:0]overlay=w" ...

    the problem though is that ffmpeg seems to start the input streams at significantly different times, so that there is about a 1 second shift between the streams, that remains throughout. Now we do have correct time stamps (pts) on both streams. Can we somehow tell ffmpeg to respect those and use them to align the streams correctly ?

  • Is it possible to use FFMPEG for h264 live decoding ?

    9 avril 2018, par elpha01

    I’m receiving in my C#/.NET application a H264 stream as byte arrays.
    Firstly I receive the SPS and PPS and then all NAL_units.

    I haven’t find good tools to make h264 decoding so I’m currently using FFMPEG for testing with this article as help : A Simple c# Wrapper for ffMpeg

    For the moment I write the sps and the pps in a provisory .264 file with the first NAL unit then I decode it with FFMPEG. It works for the first frame but I don’t know how I can decode the following frame. I tried to rewrite the SPS and PPS in a second file with the second NAL unit without the first one but it doesn’t work.

    How can I use FFMPEG to decode a live h264 stream when receiving NAL units.

    Thank you