
Recherche avancée
Médias (91)
-
Géodiversité
9 septembre 2011, par ,
Mis à jour : Août 2018
Langue : français
Type : Texte
-
USGS Real-time Earthquakes
8 septembre 2011, par
Mis à jour : Septembre 2011
Langue : français
Type : Texte
-
SWFUpload Process
6 septembre 2011, par
Mis à jour : Septembre 2011
Langue : français
Type : Texte
-
La conservation du net art au musée. Les stratégies à l’œuvre
26 mai 2011
Mis à jour : Juillet 2013
Langue : français
Type : Texte
-
Podcasting Legal guide
16 mai 2011, par
Mis à jour : Mai 2011
Langue : English
Type : Texte
-
Creativecommons informational flyer
16 mai 2011, par
Mis à jour : Juillet 2013
Langue : English
Type : Texte
Autres articles (36)
-
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir -
Librairies et binaires spécifiques au traitement vidéo et sonore
31 janvier 2010, parLes logiciels et librairies suivantes sont utilisées par SPIPmotion d’une manière ou d’une autre.
Binaires obligatoires FFMpeg : encodeur principal, permet de transcoder presque tous les types de fichiers vidéo et sonores dans les formats lisibles sur Internet. CF ce tutoriel pour son installation ; Oggz-tools : outils d’inspection de fichiers ogg ; Mediainfo : récupération d’informations depuis la plupart des formats vidéos et sonores ;
Binaires complémentaires et facultatifs flvtool2 : (...) -
Support audio et vidéo HTML5
10 avril 2011MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...)
Sur d’autres sites (3560)
-
Can FFMPEG be used to change the transparency of a single PNG ?
19 juillet 2020, par ChameleonI know that FFMPEG can manipulate transparency using fade in/fade out over a series of frames. I just haven't found a way to generate a png with a specified transparency setting.


I am in the process of creating a lot (20 in first set) of procedurally-generated videos. Each video requires several starts and stops, where with each stop, a png overlay describing what is seen will be displayed for several (actual number TBD, and may vary) frames. I'd like each overlay to fade in over x frames, display for y frames at full opacity and then fade out over y frames. The documentation on FFMPEG really sucks at explaining fade effects, and no one I've found showing usage actually explains what all of the parameters do.


The original sources for the videos are CGI png frames with a transparent background. I will be making the descriptive overlays with transparent backgrounds (same resolution as the CGI frames) - I'm really hoping to not have to manually save each overlay (8 overlays per video at a minimum) with 4 to 6 (or more) transparency settings - The company is being real indecisive as to the length of the fade in/out or the "hang time" of the descriptive overlays.


I have already created a Python script that uses FFMPEG to place overlays on specific frames, then uses FFMPEG to stitch all frames into a single video, adds a background to the video and then places the company's watermark on the video. It already manages the key frames and what images should be overlaid (and for how many frames). It is working well, but the company doesn't like the lack of fade. I already have the framework in place to manage the fade in/out duration (I'm just missing the answer to this question). I created the script because I have no doubt that I'll have to generate the final output on the first couple of videos a number of times to appease the stakeholders.


I'm really hoping to find a fairly easy way to hand FFMPEG a png with a transparent background and have FFMPEG hand me a copy of the image whose non-transparent part is now "x percent" transparent (or "y percent" opaque). I know from failed attempts that FFMPEG doesn't choke trying to make a region more than 100% transparent.


It's a real pain to get approval to install new software on the workstation, so I'm not actually interested in any suggestion that doesn't use FFMPEG or a pretty vanilla Python installation. - It's not that the other software might not be useful, it just means that if other software is needed, I'll have to manually create the frames.


-
Writing Live-Multimedia-Application using OpenGL & Co. saving output to disc [closed]
21 janvier 2013, par user1997286I want to write an application that does the following thing :
- Getting Commands via ArtNET (DMX over Ethernet, a Control Protocol) for each object (called Layer)
- each Layer could be one of the following : Live Camera Stream, Movie, Image
- each layer could be translated, rotated or stretched
- on each layer I can set filters (Like a Kaleidoscope Effect, Blur, Color Correction, etc.)
- the rsulting video-stream is in the 3d-space
- I want to display each part of the image on one Projector (in total up to 3 ones) using a TripleHead2GO (3 Projectors display a different region of my DVI-Output). Each Projecector-Image should have own Soft-Edge and Keystone parameters.
- the resulting image will also be shown on a Preview-Screen with some Information overlay.
I think all that should be possible with opengl and openal (for the movie audio)
I think I'll use C++, OpenGL for Graphics, OpenAL for Audio, if needed ffmpeg for Video conversion, Ubuntu/Debian as OS.
The software is used to do Multimedia-Shows on Concerts including Cameras & Co.
All that should happen Live (On a FullHD output), Having i7 3770, GLX 670 and 16GB of Ram for at least 8 Layers. (4 Live-Images at once + Some Overlays like the Actors Name and some Logos)
But now comes the question.
Is it also Posible to do the following with that setting :
- Writing the output Image with all the 3d translations to a Movie File (To Master a DVD later) with Audio
- Mixing Audio from different Inputs & Files (Ambience Mics, Signal from the Sound Mixer, Playbacks from my own application) to more than one Mix (eg. one Mix for the Recording, one Mix for Live)
- Stream that Output Complete or in Parts (e.g. the left Part of the Image) over the Network (For example, Projector 1 is near the Server, so I connect it using DVI, Projector 2+3 is connected to a Computer that receives the streams for that two projectors (with soft edge on each stream) and Screen 4 is outside the Concert Hall and shows the complete Live-Stream.
- What GUI-Framework should I use for that ?
- is it perhaps event performant enough to use Java for that ?
- is it posible to use that mechanism for just rendering (eg. I have stored the cut points on Disc and saved every single camera stream to change some errors later or cut out some parts)
-
What is Behavioural Segmentation and Why is it Important ?
28 septembre 2023, par Erin — Analytics Tips