
Recherche avancée
Médias (1)
-
Rennes Emotion Map 2010-11
19 octobre 2011, par
Mis à jour : Juillet 2013
Langue : français
Type : Texte
Autres articles (44)
-
Les autorisations surchargées par les plugins
27 avril 2010, parMediaspip core
autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs -
Contribute to a better visual interface
13 avril 2011MediaSPIP is based on a system of themes and templates. Templates define the placement of information on the page, and can be adapted to a wide range of uses. Themes define the overall graphic appearance of the site.
Anyone can submit a new graphic theme or template and make it available to the MediaSPIP community. -
ANNEXE : Les plugins utilisés spécifiquement pour la ferme
5 mars 2010, parLe site central/maître de la ferme a besoin d’utiliser plusieurs plugins supplémentaires vis à vis des canaux pour son bon fonctionnement. le plugin Gestion de la mutualisation ; le plugin inscription3 pour gérer les inscriptions et les demandes de création d’instance de mutualisation dès l’inscription des utilisateurs ; le plugin verifier qui fournit une API de vérification des champs (utilisé par inscription3) ; le plugin champs extras v2 nécessité par inscription3 (...)
Sur d’autres sites (8056)
-
FFMpeg - Segment audio to chunks
8 mars 2015, par fr3shI have tried this example in order to segment a given video file using ffmpeg into an m3u8 file and smaller chunks (.ts files). This actually worked great. Is it possible to do practically the same thing with audio input ?
This was my most promising approach so far (capturing live audio on Windows OS) :ffmpeg -f dshow -i audio="<name of="of" input="input" device="device">" -acodec libmp3lame -ab 64000 | segmenter - 10 stream stream.m3u8 http:///stream/stream/ 5 1
</name>But this returns this error :
At least one output file must be specified.
Could not open input file, make sure it is an mpegts file: -1I really would not know how to convert the live audio stream to an mpegts file.
Could anyone please give me a hint ?Thanks a lot
-
iOS - How can I stream Encoded Video Frames(from AVFoundation and VideoToolBox) from device to server via RTP
2 octobre 2015, par ASP PeekI am trying to stream live video from my iPhone device to server using RTP.
Using AVFoundation’s AVCaptureVideoDataOutput, I was able to get CMSampleBuffer for video. I then feed these frames as they arrive into VideoToolBox’s VTCompressionSessionEncodeFrame() and is able to get Encoded CMSampleBuffer.Now to send these encoded Frames via RTP, I came across FFMPEG and found its built library for iOS device. (https://github.com/kewlbear/FFmpeg-iOS-build-script)
However I am not able to find any iOS example or sample code or any documentation that explains the process of sending the encoded frames via RTP for iOS apps.
Is there any existing example or documentation that can explain me how can I send the encoded CMSampleBuffers to server via RTP using FFMPEG.
Thanks in Advance :)
-
How can I stream video in swift using FFmpeg in real time ?
13 novembre 2019, par rInI want to make a real time video streaming app but I’m new to Swift and live streaming..
Video input captured by AVCaptureSession is from iPhone and I want to encode the input from MPEG-4 to MPEG-2.
I will use FFmpeg library to encode the video input in swift.Here’s my questions.
-
I don’t know how to deliver video data to FFmpeg functions. Actually I’m not certain that I have to use FFmpeg library to encode video data from MPEG-4 to MPEG-2 transport stream. Is there any API that can encode video data in Swift by Apple ?
-
How can I deal with video data from AVCaptureSession ? Are there frames or h.264 video in CMSampleBuffer ? I want to know what type of data is in CMSampleBuffer.
I’m struggling to solve this problem. Please let me know anything if you have some experiences about this kind of project. Any ideas or advice for a better approach are welcome.
-