Recherche avancée

Médias (3)

Mot : - Tags -/Valkaama

Autres articles (111)

  • Submit bugs and patches

    13 avril 2011

    Unfortunately a software is never perfect.
    If you think you have found a bug, report it using our ticket system. Please to help us to fix it by providing the following information : the browser you are using, including the exact version as precise an explanation as possible of the problem if possible, the steps taken resulting in the problem a link to the site / page in question
    If you think you have solved the bug, fill in a ticket and attach to it a corrective patch.
    You may also (...)

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

  • Script d’installation automatique de MediaSPIP

    25 avril 2011, par

    Afin de palier aux difficultés d’installation dues principalement aux dépendances logicielles coté serveur, un script d’installation "tout en un" en bash a été créé afin de faciliter cette étape sur un serveur doté d’une distribution Linux compatible.
    Vous devez bénéficier d’un accès SSH à votre serveur et d’un compte "root" afin de l’utiliser, ce qui permettra d’installer les dépendances. Contactez votre hébergeur si vous ne disposez pas de cela.
    La documentation de l’utilisation du script d’installation (...)

Sur d’autres sites (9988)

  • How to detect silence in a pyAV AudioFrame ?

    29 janvier 2024, par Sachin Dole

    I want to process streaming audio (coming in from a person speaking on the peer of a webRTC peer connection) to detect when the person is done talking. I have got the audio track and access to individual frames. I see that each frame can be converted to an nd_array using Frame.to_ndarray. I can also see values in the ndarray changing depending on what the person is speaking, what pitch, what volume etc. Now, I want to detect silence on the stream. My question is what is in the ndarray and how can I make sense of the data ?

    


            while True:
            try:
                frame:AudioFrame = await track.recv()
                frame_nd_array = frame.to_ndarray() 


    


    Where can I learn what is in the frame_nd_array ?

    


  • How to batch auto-generate videos according to audio length

    16 septembre 2020, par yociba

    I have hundreds of folders of Spanish short stories, with each folder having dozens of individual .MP3 files for each dialogue spoken. Along with the .MP3 files, lies in the folder, a JSON file of the format ["folder1-dialogue-1", "folder1-dialogue-2",...].

    


    I want to generate images, each containing the text values of the JSON file, like folder1-dialogue-1 and generate videos combining that image with its corresponding audio (folder1-dialogue-1.mp3), and then merge all the videos in the folder into one - folder1.mp4

    


    How do I go about doing this ?

    


    P.S I'm not straight up asking for code, but just advice as to how I go about doing this.

    


  • vulkan : fix sw_frame_deps counter

    20 novembre 2024, par Lynne
    vulkan : fix sw_frame_deps counter
    

    The code used as a basis was the buffer dependency code, where the
    counter was incremented after each buffer, but for the sw_frame dep
    API, we only support adding individual frames at a time.

    • [DH] libavutil/vulkan.c