Recherche avancée

Médias (2)

Mot : - Tags -/plugins

Autres articles (111)

  • Des sites réalisés avec MediaSPIP

    2 mai 2011, par

    Cette page présente quelques-uns des sites fonctionnant sous MediaSPIP.
    Vous pouvez bien entendu ajouter le votre grâce au formulaire en bas de page.

  • Support audio et vidéo HTML5

    10 avril 2011

    MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
    Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
    Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
    Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...)

  • HTML5 audio and video support

    13 avril 2011, par

    MediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
    The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
    For older browsers the Flowplayer flash fallback is used.
    MediaSPIP allows for media playback on major mobile platforms with the above (...)

Sur d’autres sites (8225)

  • stream a video file using srt protocol [closed]

    13 novembre 2022, par programmer

    I want to stream a file over the network from a server. I will need to
send the AVFormatContext, and individual AVPackets over the network.For this reason i use libav library(ffmpeg) and srt protocol to send to specific port. But i can not find a function that send avpackets in srt. I find srt_sendmsg() function but it is not usable in this case.Does it have any solution ?
thanks in advance.

    


    As i say i want to design a srt server that streams specific file. it should read the video file and then send it on specific port that clients can connect to that and receive live stream.I want to use c++ programming , libav libraray and srt protocol

    


  • Is it possible to pipe an ffmpeg output with multiple files (HLS or DASH)

    29 août 2020, par New Dev

    I'm using FFmpeg to generate fragmented MP4s in a dash and HLS format.

    


    ffmpeg -i input.mov -f dash -seg_duration 6 -hls_playlist true output.mbd


    


    The above (simplified) command outputs multiple files, in addition to output.mdb (e.g. actual segments, master.m3u8, etc...)

    


    Is there a way to get each produced file into their individual and separate output streams ?

    



    


    Broader context :

    


    I'm trying to build a transcoder in Node.js running in Google Cloud, with the idea being that it writes directly to a Google Storage through a writable stream. I can only create a stream per file, but since the number of files is dynamic, I'm not sure how to obtain a stream from each file.

    


  • How to detect silence in a pyAV AudioFrame ?

    29 janvier 2024, par Sachin Dole

    I want to process streaming audio (coming in from a person speaking on the peer of a webRTC peer connection) to detect when the person is done talking. I have got the audio track and access to individual frames. I see that each frame can be converted to an nd_array using Frame.to_ndarray. I can also see values in the ndarray changing depending on what the person is speaking, what pitch, what volume etc. Now, I want to detect silence on the stream. My question is what is in the ndarray and how can I make sense of the data ?

    


            while True:
            try:
                frame:AudioFrame = await track.recv()
                frame_nd_array = frame.to_ndarray() 


    


    Where can I learn what is in the frame_nd_array ?