Recherche avancée

Médias (1)

Mot : - Tags -/musée

Autres articles (53)

  • Keeping control of your media in your hands

    13 avril 2011, par

    The vocabulary used on this site and around MediaSPIP in general, aims to avoid reference to Web 2.0 and the companies that profit from media-sharing.
    While using MediaSPIP, you are invited to avoid using words like "Brand", "Cloud" and "Market".
    MediaSPIP is designed to facilitate the sharing of creative media online, while allowing authors to retain complete control of their work.
    MediaSPIP aims to be accessible to as many people as possible and development is based on expanding the (...)

  • Les images

    15 mai 2013
  • Mediabox : ouvrir les images dans l’espace maximal pour l’utilisateur

    8 février 2011, par

    La visualisation des images est restreinte par la largeur accordée par le design du site (dépendant du thème utilisé). Elles sont donc visibles sous un format réduit. Afin de profiter de l’ensemble de la place disponible sur l’écran de l’utilisateur, il est possible d’ajouter une fonctionnalité d’affichage de l’image dans une boite multimedia apparaissant au dessus du reste du contenu.
    Pour ce faire il est nécessaire d’installer le plugin "Mediabox".
    Configuration de la boite multimédia
    Dès (...)

Sur d’autres sites (7695)

  • combine 4 video with delay side by side using ffmpeg

    9 janvier 2021, par MeTe-30

    I have at least 4 video from video conference crated by meetecho/janus-gateway
    
Janus create two mjr video and audio file for each user, first i merged them into one webm file, then convert all to 500*500 videos.
    
Now i'm trying to combine these videos like mosaic and found this code :

    


    ffmpeg -i 1.webm -i 2.webm -i 3.webm -i 4.webm \
-speed 8 -deadline realtime -filter_complex "[0]pad=2*iw:2*ih[l]; \
[1]setpts=PTS-STARTPTS+428/TB[1v]; [l][1v]overlay=x=W/2[a]; \
[2]setpts=PTS-STARTPTS+439/TB[2v]; [a][2v]overlay=y=H/2[b]; \
[3]setpts=PTS-STARTPTS+514/TB[3v]; [b][3v]overlay=y=H/2:x=W/2[v]; \
[1]adelay=428372|428372[1a]; \
[2]adelay=439999|439999[2a]; \
[3]adelay=514589|514589[3a]; \
[0][1a][2a][3a]amix=inputs=4[a]" \
-map "[v]" -map "[a]" merged.webm


    


    I calculated the delayed times from created dateTime of each file, related to the first video.

    


    my problmes :

    


      

    1. This code is not working ! after minutes of console freezing, it shows this line :

      


      Killed 29 fps=0.1 q=0.0 size= 1kB time=00:04:30.07 bitrate= 0.0kbits/s speed=0.896x

      


    2. 


    3. I didn't find out the meaning of letters before and after overlay, [1v][2v][3v][l][a][b][v]...

      


    4. 


    


  • Looking for a free alternative RTSP server for Node.js [closed]

    2 juin 2020, par Maoration

    I'm looking at running my node.js server as an RTSP streaming server, as well as an http server. the ability to get some ffmpeg video output to stream as rtsp to 'localhost' (which will be the server listening), and for multiple clients to request a stream from the server via the rtsp ://... protocol

    



    The most common online implementation is :
https://www.npmjs.com/package/rtsp-streaming-server

    



    However, this is licensed under GPL-3.0, meaning my product would have to be open-source, or I'll be violating the terms of use. I'm afraid thats not possible..

    



    Other common results when searching for a solution are :

    



    https://www.npmjs.com/package/rtsp-server
which just seems to wrap to lower level rtsp protocol messages.

    



    https://www.npmjs.com/package/node-media-server
which provides solutions to many use cases, but I couldnt figure out how to configure it as an RTSP server, or if this would even be possible.

    



    So, any alternatives ? other suggestions ?

    


  • Mixing audio stream into video stream using ffmpeg while retaining original audio from the video stream as background [duplicate]

    18 mars 2020, par Core7s

    I have a live video stream and a live audio stream. The audio stream has to get mixed into the video stream and the resulting stream is pushed to YouTube. I have this working ffmpeg -i videostream -i audiostream -map 0:1 -map 1:0 -shortest -c:a copy -c:v copy -f flv streamout-youtube;
    However I am losing the original audio in the videostream. I want to have the original audio playing at maybe 20% of the original volume.
    I looked around a lot and found maybe I can modify the above command to use -filter_complex to get ffmpeg -i videostream -i audiostream -filter_complex "volume=0.2" -map 0:1 -map 1:0 -shortest -c:a copy -c:v copy -f flv streamout-youtube;

    However, when I add the -filter_complex option, the stream goes dead meaning it doesn’t get pushed to youtube anymore. I have verified the stream keys are correct. This seems to be a rather simple thing but I can’t put my finger on what’s wrong.

    I am using rtmp module of nginx to receive, mix and push out streams.