
Recherche avancée
Médias (1)
-
The pirate bay depuis la Belgique
1er avril 2013, par
Mis à jour : Avril 2013
Langue : français
Type : Image
Autres articles (55)
-
Les autorisations surchargées par les plugins
27 avril 2010, parMediaspip core
autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs -
HTML5 audio and video support
13 avril 2011, parMediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
For older browsers the Flowplayer flash fallback is used.
MediaSPIP allows for media playback on major mobile platforms with the above (...) -
Support audio et vidéo HTML5
10 avril 2011MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...)
Sur d’autres sites (10233)
-
images->video->web canvas : RGB/YUV issues
5 février 2016, par nrobWe’ve written an web app which :
- takes 3D, time dependent weather data
- tiles each 3D time point to make a 2D frame (written out as a png image)
- stitches these frames together into a video (using ffmpeg/avconv)
- streams this video into a web app
- polls the canvas for frames
- sends the frames to the GPU where they are converted back to 3D and ray traced
You can see the app here, code here and you can see the data video here
Currently the pngs are written as RGB images, the video codec is in YUV and getting frames from the canvas returns RGB. As such there is a significant loss of information due to the conversion between image spaces.
Does anyone have suggestions what is the best way round this ?
I’ve tried a bunch of RGB video codecs, but I can’t get any to work, and I don’t know if the web browser will support it anyway. Can anyone suggest a good RGB codec (both lossy and lossless would be great)
Also, is it possible to write to YUV images/read them from a video canvas in HTML5 ?
Ultimately, I don’t even want anything to do with images/videos, I’m just hacking the codecs to stream/compress large animated 3D data volumes
-
FFMPEG HLS streaming and transcodding on the fly to HTML player - video duration changes while transcoding
20 août 2019, par Thomas ThoI am trying to make a video streaming server and watch videos directly from web browser. The idea is to make the server to stream video from remote server, transcode with different audio format in local server, and then instantly stream to the client (this is specific way I need it to function).
This is the FFMPEG code im currently using :ffmpeg -i "url" -c:v copy -c:a aac -ac 2 -f hls -hls_time 60 -hls_playlist_type event -hls_flags independent_segments out.m3u8
The HLS stream is attached to the HTML player with hls.js and it works. However, the video duration is constantly changing while video is being transcoded. I have tried to change video duration with JS like
$('video').duration = 120;
with no luck.How do i make the player to display the video file duration instead of stream current transcoded time ?
I am also planning to implement video seeking but i am clueless. The current idea is to send seeking time to the server, terminate ffmpeg, and start from specific time. However, i think the player might get stuck on loading and will not start playing without reloading.
-
Store a videostream from webcam but as 1s long files that can later be merged together [duplicate]
17 février 2018, par Jennifer RThis question is an exact duplicate of :
I’m trying to store a videostream from a webcam but need it to be 1s long files. I have never worked with storing videofiles so i’m not sure if and how i should use opencv and how those videofiles need to be contained that i can merge them together. I tried the old windows cmd code
copy /b *.avi combined.avi
but since avi is sort of a container type i only got as far as the first file. I’m guessing since there’s some container data limiting the first file and i can’t watch it.
I need help in both questions. How can i store a video stream and how do i merge those videos ? Do i use ffmpeg for the latter and is it possible to edit the container data to the correct length ?