
Recherche avancée
Médias (1)
-
Rennes Emotion Map 2010-11
19 octobre 2011, par
Mis à jour : Juillet 2013
Langue : français
Type : Texte
Autres articles (66)
-
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir -
HTML5 audio and video support
13 avril 2011, parMediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
For older browsers the Flowplayer flash fallback is used.
MediaSPIP allows for media playback on major mobile platforms with the above (...) -
Support audio et vidéo HTML5
10 avril 2011MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...)
Sur d’autres sites (9321)
-
RTSP streaming on Android client using FFMpeg
10 août 2013, par rurtleI am working on a hobby project the goal for which is to develop an Android application capable of streaming live feeds captured through web cams in a LAN setting using FFMpeg as the underlying engine. So far, I did the following -
A. Compiling and generating FFMpeg related libraries for the following releases -
FFMpeg version : 2.0
NDK version : r8e & r9
Android Platform version : android-16 & android-18thisthisthisthis
Toolchain version : 4.6 & 4.8
Platform built on : Fedora 18 (x86_64)B. Creating the files Android.mk & Application.mk in appropriate path.
However, when it came to writing the native code for accessing appropriate functionality of FFMpeg from the application layer using Java, I'm stuck with following questions -
a) Which all of FFMpeg's features I need to make available from native to app layer for streaming real-time feeds ?
b) In order to compile FFMpeg for Android, I followed this link. Whether the compilation options are sufficient for handling *.sdp streams or do I need to modify it ?
c) Do I need to make use of live555 ?I am totally new to FFMpeg and Android application development and this is going to be my first serious project for Android platform. I have been searching for relevant tutorials dealing with RTSP streaming using FFMpeg for a while now without much success. Moreover, I tried the latest development build of VLC player and found it to be great for streaming real-time feeds. However, it's a complex beast and the goal for my project is of quite limited nature, mostly learning - in a short time span.
Could you suggest some pointers (e.g. links, documents or sample code) on how can I write the native code for utilizing FFMpeg library and subsequently use those functionality from the app layer for streaming real-time feeds ? Moreover, will really appreciate if you could let me know the kind of background knowledge necessary for this project from a functional standpoint (in a language agnostic sense).
-
How to stream from dynamically created Selenium Moon browser pods to RTMP using ffmpeg ?
5 avril, par Ahmed Seddik BouchibaI'm running Selenium Moon (from Aerokube) in my minikube cluster for automated browser testing, and I need to stream the browser sessions to an RTMP endpoint. Here's my current setup :


$ kubectl get all -n my-namespace
NAME READY STATUS RESTARTS AGE
pod/minio-5b85cc6cd6-nznwl 1/1 Running 1 (48m ago) 2d8h
pod/moon-7db8b9c76f-8jzzm 4/4 Running 0 34m
pod/moon-7db8b9c76f-kp78j 4/4 Running 0 34m

NAME TYPE CLUSTER-IP EXTERNAL-IP PORT(S) AGE
service/minio NodePort 10.99.231.25 <none> 9000:31147/TCP,9001:31125/TCP 2d8h
service/moon NodePort 10.100.244.255 <none> 4444:30683/TCP,9090:32243/TCP,8080:32571/TCP 34m

NAME READY UP-TO-DATE AVAILABLE AGE
deployment.apps/minio 1/1 1 1 2d8h
deployment.apps/moon 2/2 2 2 34m

NAME DESIRED CURRENT READY AGE
replicaset.apps/minio-5b85cc6cd6 1 1 1 2d8h
replicaset.apps/moon-7db8b9c76f 2 2 2 34m
</none></none>


The challenge :


- 

- Moon dynamically creates browser pods when tests run
- I need to stream the browser activity to an RTMP endpoint using ffmpeg
- Moon was installed with Helm








What I've tried to understand :


- 

- How to integrate ffmpeg with browser pods that are created on-demand
- How to modify the Moon Helm chart to inject ffmpeg capabilities
- How to capture display from the browser pods
- How to stream to RTMP endpoints










Has anyone implemented something similar ? I'm looking for a solution that works with the dynamic nature of Moon's browser pod creation. Any examples, Helm chart modifications, or approaches would be greatly appreciated.


Thanks in advance !


-
Recommendations for real-time pixel-level analysis of television (TV) video
6 décembre 2011, par Randall Cook[Note : This is a rewrite of an earlier question that was considered inappropriate and closed.]
I need to do some pixel-level analysis of television (TV) video. The exact nature of this analysis is not pertinent, but it basically involves looking at every pixel of every frame of TV video, starting from an MPEG-2 transport stream. The host platform will be server-class, multiprocessor 64-bit Linux machines.
I need a library that can handle the decoding of the transport stream and present me with the image data in real-time. OpenCV and ffmpeg are two libraries that I am considering for this work. OpenCV is appealing because I have heard it has easy to use APIs and rich image analysis support, but I have no experience using it. I have used ffmpeg in the past for extracting video frame data from files for analysis, but it lacks image analysis support (though Intel's IPP can supplement).
In addition to general recommendations for approaches to this problem (excluding the actual image analysis), I have some more specific questions that would help me get started :
- Are ffmpeg or OpenCV commonly used in industry as a foundation for real-time
video analysis, or is there something else I should be looking at ? - Can OpenCV decode video frames in real time, and still leave enough
CPU left over to do nontrivial image analysis, also in real-time ? - Is sufficient to use ffpmeg for MPEG-2 transport stream decoding, or
is it preferable to just use an MPEG-2 decoding library directly (and if so, which one) ? - Are there particular pixel formats for the output frames that ffmpeg
or OpenCV is particularly efficient at producing (like RGB, YUV, or YUV422, etc) ?
- Are ffmpeg or OpenCV commonly used in industry as a foundation for real-time