
Recherche avancée
Médias (2)
-
Rennes Emotion Map 2010-11
19 octobre 2011, par
Mis à jour : Juillet 2013
Langue : français
Type : Texte
-
Carte de Schillerkiez
13 mai 2011, par
Mis à jour : Septembre 2011
Langue : English
Type : Texte
Autres articles (43)
-
HTML5 audio and video support
13 avril 2011, parMediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
For older browsers the Flowplayer flash fallback is used.
MediaSPIP allows for media playback on major mobile platforms with the above (...) -
De l’upload à la vidéo finale [version standalone]
31 janvier 2010, parLe chemin d’un document audio ou vidéo dans SPIPMotion est divisé en trois étapes distinctes.
Upload et récupération d’informations de la vidéo source
Dans un premier temps, il est nécessaire de créer un article SPIP et de lui joindre le document vidéo "source".
Au moment où ce document est joint à l’article, deux actions supplémentaires au comportement normal sont exécutées : La récupération des informations techniques des flux audio et video du fichier ; La génération d’une vignette : extraction d’une (...) -
Librairies et binaires spécifiques au traitement vidéo et sonore
31 janvier 2010, parLes logiciels et librairies suivantes sont utilisées par SPIPmotion d’une manière ou d’une autre.
Binaires obligatoires FFMpeg : encodeur principal, permet de transcoder presque tous les types de fichiers vidéo et sonores dans les formats lisibles sur Internet. CF ce tutoriel pour son installation ; Oggz-tools : outils d’inspection de fichiers ogg ; Mediainfo : récupération d’informations depuis la plupart des formats vidéos et sonores ;
Binaires complémentaires et facultatifs flvtool2 : (...)
Sur d’autres sites (6364)
-
[FFmpeg on Windows] : help me converting a set of .png files to a video file
31 octobre 2016, par Stefano FedeleHi everybody I am on windows 10, I installed FFMpeg and I would like to use it to convert 100 .png image files into a video file.
I am using it via the prompt, which is the Windows console, and I am struggling with the syntax (which probability comes from Linux). I set the folder in which ffmpeg.exe is contained as a windows system path as suggested on many websites and I moved the prompt to the folders in which the image files are contained which is reported here :
cd C:\video\pnts + ag 40 nm- during-tmpyp 10-12- laser power 7mw- kinetic100- exposure time 1s- gain
I would like to convert the image files contained in this folder into a video file, the names of those image files are reported here :
pnts + ag 40 nm- during-tmpyp 10-12- laser power 7mw- kinetic100- exposure time 1s- gain 20 001.png
pnts + ag 40 nm- during-tmpyp 10-12- laser power 7mw- kinetic100- exposure time 1s- gain 20 002.png
...
...
pnts + ag 40 nm- during-tmpyp 10-12- laser power 7mw- kinetic100- exposure time 1s- gain 20 100.png
The number of frames per second whould be 2.
I read on internet that I should have to type something like this
ffmpeg -i image-%03d.png video.webm
But it looks like there are a lot of variables to consider and all the times I try to set them and run FFmpeg I get an error from the prompt.
Is there anyone who could suggest me the proper syntax, please ? -
running ECSTask in IOS
26 janvier 2017, par DecKnoI am new to IOS platform. I have a code snippet written using JAVA API.
AmazonECS amazonECS = new AmazonECSClient(credentials).withRegion(usWest1);
String command="bash /opt/run-task.sh "+"/mnt/s3/inputFile1::/mnt/s3/inputFile2"+" "+ "outputFile";
ContainerOverride containerOverrides = new ContainerOverride().withCommand(command).withName("<<container>>");
TaskOverride overrides = new TaskOverride().withContainerOverrides(containerOverrides);
RunTaskRequest runTaskRequest = new RunTaskRequest().withCluster("<<cluster>>").withTaskDefinition("<<task definition="definition" arn="arn">>")
.withOverrides(overrides).withGeneralProgressListener(new ProgressListener() {
@Override
public void progressChanged(ProgressEvent progressEvent) {
System.out.println(progressEvent.getBytesTransferred());
System.out.println(progressEvent.getEventType());
}
});
Task task = new Task().withTaskDefinitionArn("<<task definition="definition" arn="arn">>")
.withOverrides(overrides);
RunTaskResult runTaskResult = amazonECS.runTask(runTaskRequest).withTasks(task);
List<failure> failures = runTaskResult.withTasks(task).getFailures();
</failure></task></task></cluster></container>I am using ffmpeg to merge few video files as single file. I need to know if there is equivalent functionality available in IOS.
-
Live streaming of processed frames to AWS
22 avril 2021, par MinasChamI'm working on a project where i need to capture live video feed from an RTSP camera source, process the video frame-by-frame and stream the result to an AWS Service.


So far, my solution :


- 

- Captures frames from the RTSP camera source using
OpenCV
and performs some processing. - Feeds the processed frames to an
ffmpeg
pipe that packages the content for online streaming (HTTP Live Streaming - hls
) and saves it locally. - Transfers the media content to an Amazon Kinesis Video Stream using a
Gstreamer
pipeline element withkvssink
as a sink element.








My questions are :


- 

- Currently I'm saving the content both locally and on an Amazon Kinesis Video Stream. Is this efficient ?
- Is it possible to directly stream the frames to the Amazon kinesis Video Stream (perhaps by connecting the
ffmpeg
output with thegstreamer
pipeline element) ? - Is the file format suitable for this implementation or would it better to encode the media differently ?








- Captures frames from the RTSP camera source using