
Recherche avancée
Autres articles (98)
-
Mise à jour de la version 0.1 vers 0.2
24 juin 2013, parExplications des différents changements notables lors du passage de la version 0.1 de MediaSPIP à la version 0.3. Quelles sont les nouveautés
Au niveau des dépendances logicielles Utilisation des dernières versions de FFMpeg (>= v1.2.1) ; Installation des dépendances pour Smush ; Installation de MediaInfo et FFprobe pour la récupération des métadonnées ; On n’utilise plus ffmpeg2theora ; On n’installe plus flvtool2 au profit de flvtool++ ; On n’installe plus ffmpeg-php qui n’est plus maintenu au (...) -
Personnaliser en ajoutant son logo, sa bannière ou son image de fond
5 septembre 2013, parCertains thèmes prennent en compte trois éléments de personnalisation : l’ajout d’un logo ; l’ajout d’une bannière l’ajout d’une image de fond ;
-
Multilang : améliorer l’interface pour les blocs multilingues
18 février 2011, parMultilang est un plugin supplémentaire qui n’est pas activé par défaut lors de l’initialisation de MediaSPIP.
Après son activation, une préconfiguration est mise en place automatiquement par MediaSPIP init permettant à la nouvelle fonctionnalité d’être automatiquement opérationnelle. Il n’est donc pas obligatoire de passer par une étape de configuration pour cela.
Sur d’autres sites (9245)
-
Extracting Metadata from a video file (.mkv) using FFmpeg
14 avril 2023, par Ashutosh SinglaI have a video file that contains 4 streams, 3 videos streams and one steam for metadata.


Stream Info :


Input #0, matroska,webm, from 'output_master.mkv':
Metadata:
title : Azure Kinect
encoder : libmatroska-1.4.9
creation_time : 2021-05-20T12:11:15.000000Z
K4A_DEPTH_DELAY_NS: 0
K4A_WIRED_SYNC_MODE: MASTER
K4A_COLOR_FIRMWARE_VERSION: 1.6.110
K4A_DEPTH_FIRMWARE_VERSION: 1.6.79
K4A_DEVICE_SERIAL_NUMBER: 000123102712
K4A_START_OFFSET_NS: 298800000
Duration: 00:00:40.03, start: 0.000000, bitrate: 480934 kb/s

Stream #0:0(eng): Video: mjpeg (Baseline) (MJPG / 0x47504A4D), yuvj422p(pc, bt470bg/unknown/unknown), 2048x1536, SAR 1:1 DAR 4:3, 30 fps, 30 tbr, 1000k tbn (default)
Metadata:
 title : COLOR
 K4A_COLOR_TRACK : 14499183330009048
 K4A_COLOR_MODE : MJPG_1536P
Stream #0:1(eng): Video: rawvideo (b16g / 0x67363162), gray16be, 640x576, SAR 1:1 DAR 10:9, 30 fps, 30 tbr, 1000k tbn (default)
Metadata:
 title : DEPTH
 K4A_DEPTH_TRACK : 429408169412322196
 K4A_DEPTH_MODE : NFOV_UNBINNED
Stream #0:2(eng): Video: rawvideo (b16g / 0x67363162), gray16be, 640x576, SAR 1:1 DAR 10:9, 30 fps, 30 tbr, 1000k tbn (default)
Metadata:
 title : IR
 K4A_IR_TRACK : 194324406376800992
 K4A_IR_MODE : ACTIVE
Stream #0:3: Attachment: none
Metadata:
 filename : calibration.json
 mimetype : application/octet-stream
 K4A_CALIBRATION_FILE: calibration.json



I am using the below command to extract
Stream #0:0
,Stream #0:1
, andStream #0:2
by changingmap 0:X
.

ffmpeg -i output_master.mkv -c copy -allow_raw_vfw 1 -map 0:0 temp_0.mkv 



To extract the metadata from all the streams and store them in metadata.txt, I am using the command below :


ffprobe -v quiet -show_format -show_streams -print_format json output_master.mkv > metadata.txt



What should be the command to extract
Stream #0:3
? Any help would be appreciated.

-
How to stream live h.264 (IP camera) video to browser ? (bonus : low bandwidth and latency)
4 octobre 2018, par Ryan GriggsI need to stream live h.264-encoded video from an IP camera to the browser, while supporting all common browsers and mobile devices (i.e. Android, Firefox, Chrome, IE, Safari (Mac OS and iOS)), and while keeping bandwidth requirements and latency to a minimum.
MPEG-DASH requires browser support for Media Source Extensions, which are NOT supported by iOS. So that’s out.
HLS is only supported by Safari and Edge.
Also DASH seems to impose a latency of several seconds, which is not preferable.
I would like to be able to chunk the incoming h.264 data (i.e. fragmented MP4), pass the chunked data to the browser via Websockets, then dump the chunks into some sort of player as they arrive.
Broadway and its forks are a javascript h.264 decoder, and there is a Broadway-stream project that supports streams instead of files, but the docs are poor and I can only find examples of streaming when the source is not live.
The most pressing question is : how do I hand the "chunked data" to a player or Video HTML element as it arrives at the browser ?
I think the ideal setup would be to
- Use ffmpeg to transcode the original video to a chunked format (fMP4)
- Pipe the chunked output to a Node JS app which emits each chunk out through a Websocket to all connected viewers
- Viewers’ browsers dump each incoming chunk into some sort of decoder which renders the video.
I’m clear up to the point of handing the received chunks to a video decoder. How can that be done without depending on Media Source Extensions, and allowing viewers to join the stream at random times ?
-
RTMP_ReadPacket, failed to read RTMP packet header
14 juin 2018, par Dilivio CatI am trying to livestream a 24/7 looping stream from my Ubuntu server to youtube.
I have a gif that is looping and my music gets shuffled from a music list.
for i in ./*.mp3 ; do echo "file '$i'" ; done | shuf >mp3list.txt ffmpeg -f
concat -i mp3list.txt -fflags +genpts -ignore_loop 0 -i giffilename.gif -f
flv rtmp://a.rtmp.youtube.com/live2/(mykeyhere)But I am getting this error when I run my script :
RTMP_ReadPacket, failed to read RTMP packet header : Unknown error
occutredm/live2/(mykeyhere)I did some research, but couldn’t find anything about this specific error. I opened port 1935 and it did not change anything.
The OS is Ubuntu 16.04 on a VPS.
Thanks for your help,Dilie