Recherche avancée

Médias (1)

Mot : - Tags -/pirate bay

Autres articles (53)

  • Personnaliser en ajoutant son logo, sa bannière ou son image de fond

    5 septembre 2013, par

    Certains thèmes prennent en compte trois éléments de personnalisation : l’ajout d’un logo ; l’ajout d’une bannière l’ajout d’une image de fond ;

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

  • Emballe médias : à quoi cela sert ?

    4 février 2011, par

    Ce plugin vise à gérer des sites de mise en ligne de documents de tous types.
    Il crée des "médias", à savoir : un "média" est un article au sens SPIP créé automatiquement lors du téléversement d’un document qu’il soit audio, vidéo, image ou textuel ; un seul document ne peut être lié à un article dit "média" ;

Sur d’autres sites (5759)

  • Merging input Streams with nodejs/ffmpeg

    14 septembre 2020, par jAndy

    I'm creating a very basic and rudimentary Video-Web-Chat. On the client side, I'm going to use a simple getUserMedia API call to capture the webcam data and send video-data as data-blob to my server.

    


    From There, I'm planning to either use the fluent-ffmpeg library or just spawn ffmpeg myself and pipe that raw data to ffmpeg, which in turn, does some magic and pushes that out as HLS stream to an Amazon AWS Service (for instance), which then gets actually displayed on a Web Browser for all participating people in the video chat.

    


    So far, I think all of this should be fairly easy to implement, but I keep my head spinning around the question, how I can create a "combined" or "merged" frame and stream, so the output HLS data from my server to the distributing cloud service has only to be one combined data stream to receive.

    


    If there are 3 people in that video chat, my server receives 3 data streams from those clients and combines these data streams (from the individual web-cam data sources) into one output stream.

    


    How could that be accomplished ?
Can I "create" a new frame with ffmpeg, so to speak ? I would be very thankful if anybody could give me a heads up here, maybe I'm thinking in a complete wrong direction.

    


    Another question which arises to me is, if I really can just "dump" any data, which I'm receiving from a binary blob created from getUserMedia or MultiStreamRecorder to ffmpeg or if I have to specify somewhere and somehow the exact codecs being used etc.?

    


  • avcodec/mediacodecdec : add workaround for buggy amlogic mpeg2 decoder

    26 avril 2018, par Aman Gupta
    avcodec/mediacodecdec : add workaround for buggy amlogic mpeg2 decoder
    

    I tested the previous mediacodec changes on seven different Android
    TV devices, with both mpeg2 and h264 content. All except one worked
    as expected. The exception was the MiBox3 running Android 6.0.1,
    where playback would freeze on a frame every few seconds. I tested
    two other AMLogic devices with newer Android versions that did not
    show the same problem. H264 decoding on the MiBox3 was also not affected,
    so this workaround applies only to OMX.amlogic.mpeg2.decoder.awesome
    on Android API22.

    There is a rumor that Xiaomi is planning to release Android Oreo for
    the MiBox3, so I will revisit in a few months to confirm whether this
    is specific to os/driver version or the chipset used in that device.

    Signed-off-by : Aman Gupta <aman@tmm1.net>
    Signed-off-by : Matthieu Bouron <matthieu.bouron@gmail.com>

    • [DH] libavcodec/mediacodecdec.c
  • creating illusion of live streaming (internet radio) using ffmpeg

    29 décembre 2014, par user259060

    I am working on a project that involves live streaming but without seeking (just like internet radio). I am using ffmpeg and ffserver.

    • I could just send the song to ffserver feed using ffmpeg but the problem is that the whole song / file is getting dumped. I don’t want that to happen.

    • First I segmented the song using ffmpeg -threads 1 -i INPUT.mp3 -ar 24000 -acodec libmp3lame -ac 1 -aq 1 -ab 64k -map 0:0 -f segment -segment_time 2 -segment_list /PATH/TO/LIST/outputlist.m3u8 -segment_format mpegts /OUTPUT/PATH/output%05d.mp3 (this is just an example) .

    • As you can see that the segment time is 2 seconds. What I’m actually planning to achieve is that I want to send first segmented file say output00001.mp3 to ffserver feed and wait for 1 second then send the the second segmented files say output00002.mp3 to ffserver feed and so on till the end of the song. This creates an illusion of radio live streaming. I was able to implement this without a problem using python.

    PROBLEM :

    The problem I faced while listening to song (htttp ://foo:port/test1.mp3) is that after 2nd second I hear a few millisecond pause and then the song continues to play which is very irritating. This happens after every segmented song completes playing. Is there any solution to eradicate the pause ? Is there a technique to make song play live (that means if my server shuts down, the song should stop immediately) ?