
Recherche avancée
Médias (1)
-
Richard Stallman et le logiciel libre
19 octobre 2011, par
Mis à jour : Mai 2013
Langue : français
Type : Texte
Autres articles (69)
-
Multilang : améliorer l’interface pour les blocs multilingues
18 février 2011, parMultilang est un plugin supplémentaire qui n’est pas activé par défaut lors de l’initialisation de MediaSPIP.
Après son activation, une préconfiguration est mise en place automatiquement par MediaSPIP init permettant à la nouvelle fonctionnalité d’être automatiquement opérationnelle. Il n’est donc pas obligatoire de passer par une étape de configuration pour cela. -
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir -
Automated installation script of MediaSPIP
25 avril 2011, parTo overcome the difficulties mainly due to the installation of server side software dependencies, an "all-in-one" installation script written in bash was created to facilitate this step on a server with a compatible Linux distribution.
You must have access to your server via SSH and a root account to use it, which will install the dependencies. Contact your provider if you do not have that.
The documentation of the use of this installation script is available here.
The code of this (...)
Sur d’autres sites (7132)
-
ffmpeg with multiple live inputs [closed]
9 juin 2013, par maddaniowe are trying to get ffmpeg to combine multiple live streams into one. In principle we got it to work by mosaicing the streams like so :
ffmpeg -rtsp_transport tcp -r 25 -i rtsp ://root:password@192.168.178.91:554/axis-media/media.amp \
-rtsp_transport tcp -r 25 -i rtsp ://root:password@192.168.178.92:554/axis-media/media.amp \
-filter_complex "[0:0]pad=iw*2:ih[a] ;[a][1:0]overlay=w" ...the problem though is that ffmpeg seems to start the input streams at significantly different times, so that there is about a 1 second shift between the streams, that remains throughout. Now we do have correct time stamps (pts) on both streams. Can we somehow tell ffmpeg to respect those and use them to align the streams correctly ?
-
Is it possible to use FFMPEG for h264 live decoding ?
9 avril 2018, par elpha01I’m receiving in my C#/.NET application a H264 stream as byte arrays.
Firstly I receive the SPS and PPS and then all NAL_units.I haven’t find good tools to make h264 decoding so I’m currently using FFMPEG for testing with this article as help : A Simple c# Wrapper for ffMpeg
For the moment I write the sps and the pps in a provisory .264 file with the first NAL unit then I decode it with FFMPEG. It works for the first frame but I don’t know how I can decode the following frame. I tried to rewrite the SPS and PPS in a second file with the second NAL unit without the first one but it doesn’t work.
How can I use FFMPEG to decode a live h264 stream when receiving NAL units.
Thank you
-
Live streaming of processed frames to AWS
22 avril 2021, par MinasChamI'm working on a project where i need to capture live video feed from an RTSP camera source, process the video frame-by-frame and stream the result to an AWS Service.


So far, my solution :


- 

- Captures frames from the RTSP camera source using
OpenCV
and performs some processing. - Feeds the processed frames to an
ffmpeg
pipe that packages the content for online streaming (HTTP Live Streaming - hls
) and saves it locally. - Transfers the media content to an Amazon Kinesis Video Stream using a
Gstreamer
pipeline element withkvssink
as a sink element.








My questions are :


- 

- Currently I'm saving the content both locally and on an Amazon Kinesis Video Stream. Is this efficient ?
- Is it possible to directly stream the frames to the Amazon kinesis Video Stream (perhaps by connecting the
ffmpeg
output with thegstreamer
pipeline element) ? - Is the file format suitable for this implementation or would it better to encode the media differently ?








- Captures frames from the RTSP camera source using