
Recherche avancée
Médias (2)
-
Exemple de boutons d’action pour une collection collaborative
27 février 2013, par
Mis à jour : Mars 2013
Langue : français
Type : Image
-
Exemple de boutons d’action pour une collection personnelle
27 février 2013, par
Mis à jour : Février 2013
Langue : English
Type : Image
Autres articles (67)
-
Gestion des droits de création et d’édition des objets
8 février 2011, parPar défaut, beaucoup de fonctionnalités sont limitées aux administrateurs mais restent configurables indépendamment pour modifier leur statut minimal d’utilisation notamment : la rédaction de contenus sur le site modifiables dans la gestion des templates de formulaires ; l’ajout de notes aux articles ; l’ajout de légendes et d’annotations sur les images ;
-
Keeping control of your media in your hands
13 avril 2011, parThe vocabulary used on this site and around MediaSPIP in general, aims to avoid reference to Web 2.0 and the companies that profit from media-sharing.
While using MediaSPIP, you are invited to avoid using words like "Brand", "Cloud" and "Market".
MediaSPIP is designed to facilitate the sharing of creative media online, while allowing authors to retain complete control of their work.
MediaSPIP aims to be accessible to as many people as possible and development is based on expanding the (...) -
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir
Sur d’autres sites (10284)
-
Is it possible to pipe an ffmpeg output with multiple files (HLS or DASH)
29 août 2020, par New DevI'm using FFmpeg to generate fragmented MP4s in a dash and HLS format.


ffmpeg -i input.mov -f dash -seg_duration 6 -hls_playlist true output.mbd



The above (simplified) command outputs multiple files, in addition to
output.mdb
(e.g. actual segments,master.m3u8
, etc...)

Is there a way to get each produced file into their individual and separate output streams ?



Broader context :


I'm trying to build a transcoder in Node.js running in Google Cloud, with the idea being that it writes directly to a Google Storage through a writable stream. I can only create a stream per file, but since the number of files is dynamic, I'm not sure how to obtain a stream from each file.


-
How to detect silence in a pyAV AudioFrame ?
29 janvier 2024, par Sachin DoleI want to process streaming audio (coming in from a person speaking on the peer of a webRTC peer connection) to detect when the person is done talking. I have got the audio track and access to individual frames. I see that each frame can be converted to an nd_array using Frame.to_ndarray. I can also see values in the ndarray changing depending on what the person is speaking, what pitch, what volume etc. Now, I want to detect silence on the stream. My question is what is in the ndarray and how can I make sense of the data ?


while True:
 try:
 frame:AudioFrame = await track.recv()
 frame_nd_array = frame.to_ndarray() 



Where can I learn what is in the frame_nd_array ?


-
How to batch auto-generate videos according to audio length
16 septembre 2020, par yocibaI have hundreds of folders of Spanish short stories, with each folder having dozens of individual .MP3 files for each dialogue spoken. Along with the .MP3 files, lies in the folder, a JSON file of the format
["folder1-dialogue-1", "folder1-dialogue-2",...]
.

I want to generate images, each containing the text values of the JSON file, like
folder1-dialogue-1
and generate videos combining that image with its corresponding audio(folder1-dialogue-1.mp3)
, and then merge all the videos in the folder into one -folder1.mp4


How do I go about doing this ?


P.S I'm not straight up asking for code, but just advice as to how I go about doing this.