
Recherche avancée
Médias (1)
-
The pirate bay depuis la Belgique
1er avril 2013, par
Mis à jour : Avril 2013
Langue : français
Type : Image
Autres articles (39)
-
ANNEXE : Les plugins utilisés spécifiquement pour la ferme
5 mars 2010, parLe site central/maître de la ferme a besoin d’utiliser plusieurs plugins supplémentaires vis à vis des canaux pour son bon fonctionnement. le plugin Gestion de la mutualisation ; le plugin inscription3 pour gérer les inscriptions et les demandes de création d’instance de mutualisation dès l’inscription des utilisateurs ; le plugin verifier qui fournit une API de vérification des champs (utilisé par inscription3) ; le plugin champs extras v2 nécessité par inscription3 (...)
-
Ajouter notes et légendes aux images
7 février 2011, parPour pouvoir ajouter notes et légendes aux images, la première étape est d’installer le plugin "Légendes".
Une fois le plugin activé, vous pouvez le configurer dans l’espace de configuration afin de modifier les droits de création / modification et de suppression des notes. Par défaut seuls les administrateurs du site peuvent ajouter des notes aux images.
Modification lors de l’ajout d’un média
Lors de l’ajout d’un média de type "image" un nouveau bouton apparait au dessus de la prévisualisation (...) -
HTML5 audio and video support
13 avril 2011, parMediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
For older browsers the Flowplayer flash fallback is used.
MediaSPIP allows for media playback on major mobile platforms with the above (...)
Sur d’autres sites (9772)
-
Audiopulse record multiple xfbv screens with python
7 octobre 2020, par Eric LagardaI’m trying to do something complicated and I need some help.


I’m able to record a session (video and audio) with python using xvfb and ffmpeg.


I just want to visit a page using selenium and record the video and audio. It’s working but when I try to run multiple instances, the audio is mixing between records.


How can achieve that ? I’m using pulseaudio. I know that with also you can set like a interface per instance, but I don’t know how to do it.


Thanks for you support and comments.


-
FFMPEG not “cutting” as expected in Android
30 décembre 2020, par Anees UI referred FFMPEG not "cutting" as expected to split video in chunks in below format :


00:00:00 - 00:00:01


00:00:01 - 00:00:02


00:00:02 - 00:00:03


00:00:03 - 00:00:04


00:00:04 - 00:00:05


Here is the command i used :


String[] cmd1 = new String []{"-ss", "00:00:00.000", "-i", inputVideoUrl, "-t", "00:00:01.000", "-c:v", "libx264", "-strict", "-2", outputPath};

String[] cmd2 = new String []{"-ss", "00:00:01.000", "-i", inputVideoUrl, "-t", "00:00:02.000", "-c:v", "libx264", "-strict", "-2", outputPath};

String[] cmd3 = new String []{"-ss", "00:00:02.000", "-i", inputVideoUrl, "-t", "00:00:03.000", "-c:v", "libx264", "-strict", "-2", outputPath};

String[] cmd4 = new String []{"-ss", "00:00:03.000", "-i", inputVideoUrl, "-t", "00:00:04.000", "-c:v", "libx264", "-strict", "-2", outputPath};

String[] cmd5 = new String []{"-ss", "00:00:04.000", "-i", inputVideoUrl, "-t", "00:00:05.000", "-c:v", "libx264", "-strict", "-2", outputPath};



And i use this library :
https://github.com/teanersener/mobile-ffmpeg


implementation 'com.arthenica:mobile-ffmpeg-full-gpl:4.2.2.LTS'



When i execute i get the duration starting from beginning of the video all the time. Please suggest what am i doing wrong, i searched and tried changing commands but nothing works.


00:00:00 - 00:00:01


00:00:00 - 00:00:02


00:00:00 - 00:00:03


00:00:00 - 00:00:04


00:00:00 - 00:00:05


Let me share another example to explain my requirement :
inputVideoUrl contains a video of any length(ex : 5min)
I want to cut a user selected portion of it(ex first one minute) in the form of 5 split videos.


split 1 : 00 sec - 12.0sec


split 2 : 12.1 sec - 24.0sec


split 3 : 24.1 sec - 36.0sec


split 4 : 36.1 sec - 48.0sec


split 5 : 48.1 sec - 60.0sec


And later i will provide these split videos to a player playlist.


-
Is it possible to send a temporary slate (image or video) into a running Azure Live Event RTMP-stream ?
15 novembre 2020, par Brian FrischI'm currently building a video streaming app which leverages Azure Media Services Live Events.


It consists of :


- 

- a mobile app that can stream live video and.
- a web client that plays the live event video.
- a producer screen with controls to start and stop the web client access to the video.
- a server that handles various operations around the entire system










It's working very well, but I would like to add a feature that would enable the producer to add some elegance to the experience. Therefore I'm trying to get my head around how I can enable the producer be able to switch the incoming source of the stream to a pre-recorded video or event a still image at any point during the recording, and also to switch back to live-video. A kill-switch of some kind, that would cover waiting-time if there's technical difficulties on the set, and it could also be used for pre-/post-roll branding slates when introing and outroing a video event. I would like this source switch to be embedded in the video stream (also so it would be possible to get this into the final video-product if I need it in an archive for later playback)


I'm trying to do it in a way where the producer can set a timestamp for when the video override should come in, and when it should stop. The I want to have my server respond to these timestamps and send the instructions over RTMP to the Azure Live Event. Is it possible to send such an instruction ("Hey, play this video-bit/show this image in the stream for x-seconds") in the RTMP-protocol ? I've tried to figure it out, and I've read about SCTE-35 markers and such, but I have not been able to find any examples on how to do it, so I'm a bit stuck.


My plan-B is to make it possible to stream an image from the mobile application that already handles the live video-stream, but I'm initially targeting an architecture where the mobile app is unaware of anything else than live streaming, and this override switch should preferably be handled by the server, which is a firebase functions setup.


If you are able to see other ways of doing it, I'm all ears.


I've already tried to build a ffmpeg method that listens to updates to the producer-set state, and then streams an image to the same RTMP-url that the video goes to from the mobile app. But it only works when the live video isn't already streaming - it seems like I cannot take over a RTMP-stream when it's already running.