
Recherche avancée
Médias (1)
-
SPIP - plugins - embed code - Exemple
2 septembre 2013, par
Mis à jour : Septembre 2013
Langue : français
Type : Image
Autres articles (50)
-
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir -
Support audio et vidéo HTML5
10 avril 2011MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...) -
HTML5 audio and video support
13 avril 2011, parMediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
For older browsers the Flowplayer flash fallback is used.
MediaSPIP allows for media playback on major mobile platforms with the above (...)
Sur d’autres sites (8112)
-
Video from VFR Image Sequence
12 octobre 2014, par Stryker33I am using
ffmpeg
to create video from an image sequence that is taken from the AndroidCamera
’sPreviewCallback
methodonPreviewFrame
...The images are written to a pipe that is connected to
ffmpeg
’sstdin
using the command :ffmpeg -f image2pipe -vcodec mjpeg -i - -f flv -vcodec libx264
The problem that’s arising is that the output video is very short as compared to the actual recording time and all of the frames are shown very rapidly...
But when the frame size is set to the lowest supported preview size, the video appears to be in sync with the actual recording time...
As far as I reckon this seems to be an issue related to the frame rate of the input image sequence and that of the output video...
But the main problem is that the frames that are generated from
onPreviewFrame
are of variable rates...Is there any way to construct a smooth video from an image sequence having variable frame rate...?
Also, the image sequence is muxed with audio from the microphone which also appears to be out of sync with the video...
Could the video generated using the above process and audio from the microphone be muxed in perfect synchronization...?
-
How to draw a waveform from an RTSP audio using ffmpeg and Python
9 mars 2023, par S AndrewI have a Hikvision camera. Using ffmpeg, I can extract the audio from it and save it in
wav
file using below code :

import os
os.system("ffmpeg -i rtsp://admin:password@192.168.0.27:554/Streaming/Channels/101/ -q:a 0 -map a -t 10 file.wav")



It creates
file.wav
file and when played I can hear the audio recorded from camera. Now I am planning to draw the waveform of these audio's. For this I have below code :

os.system("ffmpeg -i rtsp://admin:password@192.168.0.27:554/Streaming/Channels/101/ -filter_complex showwavespic -frames:v 1 output.png")



and below is the output I get after pressing q


[q] command received. Exiting.

Finishing stream 0:0 without any data written to it.
Output #0, image2, to 'output.png':
 Metadata:
 title : Media Presentation
 encoder : Lavf59.26.100
 Stream #0:0: Video: png, rgba, 600x240 [SAR 1:1 DAR 5:2], q=2-31, 200 kb/s, 1 fps, 1 tbn
 Metadata:
 encoder : Lavc59.36.100 png
frame= 0 fps=0.0 q=0.0 Lsize=N/A time=00:00:00.00 bitrate=N/A speed= 0x 
video:0kB audio:0kB subtitle:0kB other streams:0kB global headers:0kB muxing overhead: unknown
Output file is empty, nothing was encoded (check -ss / -t / -frames parameters if used)



and there is no file generated. I tried the above code with and mp3 file and it generated the output.png with waveform. How can I resolve the issue ?


-
YUV Raw frames to video stream
30 septembre 2014, par Ahmed NassarIm trying to stream raw YUV frames in an array generated in a C++ program to video using FFPEG. Can anyone point me to the right direction ?