Recherche avancée

Médias (2)

Mot : - Tags -/documentation

Autres articles (112)

  • MediaSPIP version 0.1 Beta

    16 avril 2011, par

    MediaSPIP 0.1 beta est la première version de MediaSPIP décrétée comme "utilisable".
    Le fichier zip ici présent contient uniquement les sources de MediaSPIP en version standalone.
    Pour avoir une installation fonctionnelle, il est nécessaire d’installer manuellement l’ensemble des dépendances logicielles sur le serveur.
    Si vous souhaitez utiliser cette archive pour une installation en mode ferme, il vous faudra également procéder à d’autres modifications (...)

  • MediaSPIP 0.1 Beta version

    25 avril 2011, par

    MediaSPIP 0.1 beta is the first version of MediaSPIP proclaimed as "usable".
    The zip file provided here only contains the sources of MediaSPIP in its standalone version.
    To get a working installation, you must manually install all-software dependencies on the server.
    If you want to use this archive for an installation in "farm mode", you will also need to proceed to other manual (...)

  • Websites made ​​with MediaSPIP

    2 mai 2011, par

    This page lists some websites based on MediaSPIP.

Sur d’autres sites (15080)

  • Anomalie #4704 : Par sinum tri le nombre par ordre alpha

    26 mars 2021

    Pour que par num mette en dernier les sans numéros, il faut qu’il fasse automatiquement le "par sinum" du coup... (en ajoutant sinum dans le select...)
    Mais du coup, comment tu le désactiverais si tu veux l’ancien comportement ?

  • Live AAC and H264 data into live stream

    10 mai 2024, par tzuleger

    I have a remote camera that captures H264 encoded video data and AAC encoded audio data, places the data into a custom ring buffer, which then is sent to a Node.js socket server, where the packet of information is detected as audio or video and then handled accordingly. That data should turn into a live stream, the protocol doesn't matter, but the delay has to be around 4 seconds and can be played on iOS and Android devices.

    


    After reading hundreds of pages of documentation, questions, or solutions on the internet, I can't seem to find anything about handling two separate streams of AAC and H264 data to create a live stream.

    


    Despite attempting many different ways of achieving this goal, even having a working implementation of HLS, I want to revisit ALL options of live streaming, and I am hoping someone out there can give me advice or guidance to specific documentation on how to achieve this goal.

    


    To be specific, this is our goal :

    


      

    • Stream AAC and H264 data from remote cellular camera to a server which will do some work on that data to live stream to one user (possibly more users in the future) on a mobile iOS or Android device
    • 


    • Delay of the live stream should be a maximum of 4 seconds, if the user has bad signal, then a longer delay is okay, as we obviously cannot do anything about that.
    • 


    • We should not have to re-encode our data. We've explored WebRTC, but that requires OPUS audio packets and thus requires us to re-encode the data, which would be expensive for our server to run.
    • 


    


    Any and all help, ranging from re-visiting an old approach we took to exploring new ones, is appreciated.

    


    I can provide code snippets as well for our current implementation of LLHLS if it helps, but I figured this post is already long enough.

    


    I've tried FFmpeg with named pipes, I expected it to just work, but FFmpeg kept blocking on the first named pipe input. I thought of just writing the data out to two files and then using FFmpeg, but it's continuous data and I don't have enough knowledge on FFmpeg on how I could use that type of implementation to create one live stream.

    


    I've tried implementing our own RTSP server on the camera using Gstreamer (our camera had its RTSP server stripped out, wasn't my call) but the camera's flash storage cannot handle having GStreamer on it, so that wasn't an option.

    


    My latest attempt was using a derivation of hls-parser to create an HLS manifest and mux.js to create MP4 containers for .m4s fragmented mp4 files and do an HLS live stream. This was my most successful attempt, where we successfully had a live stream going, but the delay was up to 16 seconds, as one would expect with HLS live streaming. We could drop the target duration down to 2 seconds and get about 6-8 seconds delay, but this could be unreliable, as these cameras could have no signal making it relatively expensive to send so many IDR frames with such low bandwidth.

    


    With the delay being the only factor left, I attempted to upgrade the implementation to support Apple's Low Latency HLS. It seems to work, as the right partial segments are getting requested and everything that makes LLHLS is working as intended, but the delay isn't going down when played on iOS' native AVPlayer, as a matter of fact, it looks like it worsened.

    


    I would also like to disclaim, my knowledge on media streaming is fairly limited. I've learned most of what I speak of in this post over the past 3 months by reading RFCs, documentation, and stackoverflow/reddit questions and answers. If anything appears to be confusing, it might be just my lack of understanding of it.

    


  • Creating a transparent color mask that covers half of every frame in a video with opencv/moviepy/ffmpeg

    24 octobre 2019, par Nyktofob

    I’m doing a pose estimation in rats separated into two compartments of a cage. To make my tracking better I want to put a transparent color mask that covers half of every frame so that one of the rats will appear to the neural net as being of different color, thus making it easier for the net to differentiate those rats. My question is : Is there any easy way to do this using python (moviepy/opencv/ffmpeg) ?

    Bare in mind, I’m completely new to python, so I’d appreciate any recommendation of a tutorial or detailed answer.

    I tried using blender to do this, but failed miserably.

    I expect to be able to modify the video, so that there will be a rectangular transparent mask in some color that covers certain part of every frame (Let’s say video is 1920x1080 pixels, so from pixel 960 on x axis everything to the right is transparently green)