Recherche avancée

Médias (1)

Mot : - Tags -/framasoft

Autres articles (69)

  • Multilang : améliorer l’interface pour les blocs multilingues

    18 février 2011, par

    Multilang est un plugin supplémentaire qui n’est pas activé par défaut lors de l’initialisation de MediaSPIP.
    Après son activation, une préconfiguration est mise en place automatiquement par MediaSPIP init permettant à la nouvelle fonctionnalité d’être automatiquement opérationnelle. Il n’est donc pas obligatoire de passer par une étape de configuration pour cela.

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

  • Automated installation script of MediaSPIP

    25 avril 2011, par

    To overcome the difficulties mainly due to the installation of server side software dependencies, an "all-in-one" installation script written in bash was created to facilitate this step on a server with a compatible Linux distribution.
    You must have access to your server via SSH and a root account to use it, which will install the dependencies. Contact your provider if you do not have that.
    The documentation of the use of this installation script is available here.
    The code of this (...)

Sur d’autres sites (7132)

  • ffmpeg with multiple live inputs [closed]

    9 juin 2013, par maddanio

    we are trying to get ffmpeg to combine multiple live streams into one. In principle we got it to work by mosaicing the streams like so :

    ffmpeg -rtsp_transport tcp -r 25 -i rtsp ://root:password@192.168.178.91:554/axis-media/media.amp \
    -rtsp_transport tcp -r 25 -i rtsp ://root:password@192.168.178.92:554/axis-media/media.amp \
    -filter_complex "[0:0]pad=iw*2:ih[a] ;[a][1:0]overlay=w" ...

    the problem though is that ffmpeg seems to start the input streams at significantly different times, so that there is about a 1 second shift between the streams, that remains throughout. Now we do have correct time stamps (pts) on both streams. Can we somehow tell ffmpeg to respect those and use them to align the streams correctly ?

  • Is it possible to use FFMPEG for h264 live decoding ?

    9 avril 2018, par elpha01

    I’m receiving in my C#/.NET application a H264 stream as byte arrays.
    Firstly I receive the SPS and PPS and then all NAL_units.

    I haven’t find good tools to make h264 decoding so I’m currently using FFMPEG for testing with this article as help : A Simple c# Wrapper for ffMpeg

    For the moment I write the sps and the pps in a provisory .264 file with the first NAL unit then I decode it with FFMPEG. It works for the first frame but I don’t know how I can decode the following frame. I tried to rewrite the SPS and PPS in a second file with the second NAL unit without the first one but it doesn’t work.

    How can I use FFMPEG to decode a live h264 stream when receiving NAL units.

    Thank you

  • Live streaming of processed frames to AWS

    22 avril 2021, par MinasCham

    I'm working on a project where i need to capture live video feed from an RTSP camera source, process the video frame-by-frame and stream the result to an AWS Service.

    


    So far, my solution :

    


      

    • Captures frames from the RTSP camera source using OpenCV and performs some processing.
    • 


    • Feeds the processed frames to an ffmpeg pipe that packages the content for online streaming (HTTP Live Streaming - hls) and saves it locally.
    • 


    • Transfers the media content to an Amazon Kinesis Video Stream using a Gstreamer pipeline element with kvssink as a sink element.
    • 


    


    My questions are :

    


      

    • Currently I'm saving the content both locally and on an Amazon Kinesis Video Stream. Is this efficient ?
    • 


    • Is it possible to directly stream the frames to the Amazon kinesis Video Stream (perhaps by connecting the ffmpeg output with the gstreamer pipeline element) ?
    • 


    • Is the file format suitable for this implementation or would it better to encode the media differently ?
    •