
Recherche avancée
Médias (1)
-
Rennes Emotion Map 2010-11
19 octobre 2011, par
Mis à jour : Juillet 2013
Langue : français
Type : Texte
Autres articles (31)
-
HTML5 audio and video support
13 avril 2011, parMediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
For older browsers the Flowplayer flash fallback is used.
MediaSPIP allows for media playback on major mobile platforms with the above (...) -
Support audio et vidéo HTML5
10 avril 2011MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...) -
De l’upload à la vidéo finale [version standalone]
31 janvier 2010, parLe chemin d’un document audio ou vidéo dans SPIPMotion est divisé en trois étapes distinctes.
Upload et récupération d’informations de la vidéo source
Dans un premier temps, il est nécessaire de créer un article SPIP et de lui joindre le document vidéo "source".
Au moment où ce document est joint à l’article, deux actions supplémentaires au comportement normal sont exécutées : La récupération des informations techniques des flux audio et video du fichier ; La génération d’une vignette : extraction d’une (...)
Sur d’autres sites (5744)
-
Connect external cameras to iOS and decompress to a usable form
27 septembre 2017, par Ping ChenI want to create a 2 camera setup which can send 1 of the camera views out as an RTMP stream depending on the motion intensity detected. The chosen camera view can change if motion intensity on the views changes.
I imagine that I could use an iPhone/iPad as encoding/streaming hub as well as 1 of the cameras. And connect a WiFi camera to the iPad/iPhone to feed the 2nd camera view.
My goals for the iOS side are :
- Connect with a WiFi camera on the local network
- Decode the data and run motion intensity detection on the WiFi camera feed AND the iPhone/iPad’s own camera feed with Brad Larson’s GPUImage framework https://github.com/BradLarson/GPUImage
- Stream out the chosen camera view. depending on motion detected
Larson’s GPUImage framework works with an AVCaptureSession subclass. I’m only familiar with AVFoundation objects, but am a complete noob with it comes to VideoToolbox and some of the lower level iOS video stuff. Through googling, I kind of know that VTDecompressionSession is what I’d get from the WiFi camera. I have no clue how I can manipulate that to a usable form for my purposes.
I’ve dug through stackoverflow answers such as : https://stackoverflow.com/a/29525001/7097455
Very informative, but maybe I don’t even know to ask the correct questions
-
Reading geolocation data from a video file using FFMpeg/Xuggler
23 décembre 2015, par agent154Using the MediaInfo application, I am able to see that a file taken with an iPhone 5 contains geolocation metadata, tagged both with ©xyz and com.apple.quicktime.location.ISO6709. I am not able to find any way to get this data using xuggler, however.
Format : MPEG-4
Format profile : QuickTime
Codec ID : qt 0000.00 (qt )
File size : 7.50 MiB
Duration : 3s 537ms
Overall bit rate : 17.8 Mbps
Recorded date : 2015-12-17T14:32:23-0330
Encoded date : UTC 2015-12-17 18:02:23
Tagged date : UTC 2015-12-17 18:02:27
Writing application : 8.4.1
Writing library : Apple QuickTime
Model : iPhone 5
©xyz : +47.5184-052.8046+133.390/
Make : Apple
com.apple.quicktime.make : Apple
com.apple.quicktime.creationdate : 2015-12-17T14:32:23-0330
com.apple.quicktime.location.ISO6709 : +47.5184-052.8046+133.390/
com.apple.quicktime.software : 8.4.1
com.apple.quicktime.model : iPhone 5As an aside, there seems to be a lot of metadata on this file that I can’t immediately find while debugging via xuggler.
The question at Reading Geolocation from Quicktime Movies with Java (Xuggler) ? is asking the exact same question, but has no answers or comments at all, and is over 4 years old.
Is anybody aware of a way to be able to get this data using xuggler as it is, or how I can modify the source and re-compile to make it work ? I am required to get this data for a work project. Thanks.
-
time progress bar does not scale by sec in FFMPEG chopped mp4 in iOS and Android devices
10 juillet 2017, par Xi XiaoI have used the below command to cut out a 44sec length video. The result video can be found here.
ffmpeg -i friends.mkv -ss 00:15:30.0 -t 44 -vf scale=640:360 video.mp4
One odd issue appears from the generated
video.mp4
.when playing it in iphone 7 and huawei honor8 android, the time progress bar does not scale by second.
What I mean is, for example, the video can start to play either from 0 second or 6 second. But, if I drag the progress bar to time between 0 - 6, such as 4, the video player will jump to 0 and play, instead of playing from 4 second.
When I use Javascript to
setTime
for 2 second for html5 video player, it does not work either, i.e. the player will play from 0 second instead of 2. Code is as below :var vid = document.getElementById("video1");
vid.currentTime = 2;
vid.play();screenshot made in iphone 7
I stop the video, try set the progress bar between 0 and 6 as in screenshot. When I click
play
, the player will play from 0.In what devices have I tested it ?
- This issue appears in iphone 7, huawei honor8 android phone.
- This issue does not appear in PC browser, nor in ipad.
At last, I suspect that the issue roots from video codec, but had no success to locate the keyword to explain my issue and google the solution.