
Recherche avancée
Médias (3)
-
Exemple de boutons d’action pour une collection collaborative
27 février 2013, par
Mis à jour : Mars 2013
Langue : français
Type : Image
-
Exemple de boutons d’action pour une collection personnelle
27 février 2013, par
Mis à jour : Février 2013
Langue : English
Type : Image
-
Collections - Formulaire de création rapide
19 février 2013, par
Mis à jour : Février 2013
Langue : français
Type : Image
Autres articles (59)
-
Personnaliser en ajoutant son logo, sa bannière ou son image de fond
5 septembre 2013, parCertains thèmes prennent en compte trois éléments de personnalisation : l’ajout d’un logo ; l’ajout d’une bannière l’ajout d’une image de fond ;
-
Ecrire une actualité
21 juin 2013, parPrésentez les changements dans votre MédiaSPIP ou les actualités de vos projets sur votre MédiaSPIP grâce à la rubrique actualités.
Dans le thème par défaut spipeo de MédiaSPIP, les actualités sont affichées en bas de la page principale sous les éditoriaux.
Vous pouvez personnaliser le formulaire de création d’une actualité.
Formulaire de création d’une actualité Dans le cas d’un document de type actualité, les champs proposés par défaut sont : Date de publication ( personnaliser la date de publication ) (...) -
Gestion générale des documents
13 mai 2011, parMédiaSPIP ne modifie jamais le document original mis en ligne.
Pour chaque document mis en ligne il effectue deux opérations successives : la création d’une version supplémentaire qui peut être facilement consultée en ligne tout en laissant l’original téléchargeable dans le cas où le document original ne peut être lu dans un navigateur Internet ; la récupération des métadonnées du document original pour illustrer textuellement le fichier ;
Les tableaux ci-dessous expliquent ce que peut faire MédiaSPIP (...)
Sur d’autres sites (7765)
-
How do I use FFMPEG to live stream a video to Vimeo ?
9 janvier 2021, par KwixsterI'm attempting to live stream to Vimeo a pre-recorded video using a FFMPEG CmdLine. The following CmdLine works just fine if the video is 1280 x 720 but if the video is 1920 x 1080 it is jerky and unwatchable. I have the Vimeo account that allows live streaming. I know very little about FFMPEG ; the line I'm using is copied from another forum and modified (path to video file, URL, key) for my purposes. I know that my PC and internet speed (24 mbps upload) is more than capable of handling this. It works fine if a stream from Wirecast.


My ultimate goal is to write a program that streams videos to Vimeo and onto my website at a scheduled time without me having to be present to push a button. That part I've got handled. Its the FFMPEG part that I'm having trouble with.


Here's the CmdLine and the FFMPEG report :


**C:\Users\kirby>ffmpeg -re -nostdin -i "D:\Dropbox\ACJ\ACJ Videos\Kirchen-Ox Blood_1080.mp4" -vcodec libx264 -preset:v ultrafast -acodec aac -f flv rtmp://rtmp-global.cloud.vimeo.com/live/149679ba-d76c-4bdf-895a-84fddccxxxxx**

ffmpeg version N-100554-g89c9c42c5b Copyright (c) 2000-2021 the FFmpeg developers
 built with gcc 9.3-win32 (GCC) 20200320
 configuration: --prefix=/ffbuild/prefix --pkg-config-flags=--static --pkg-config=pkg-config --cross-prefix=x86_64-w64-mingw32- --arch=x86_64 --target-os=mingw32 --enable-gpl --enable-version3 --disable-debug --enable-shared --disable-static --disable-debug --disable-w32threads --enable-pthreads --enable-iconv --enable-zlib --enable-libxml2 --enable-libfreetype --enable-libfribidi --enable-gmp --enable-lzma --enable-fontconfig --enable-opencl --enable-libvmaf --enable-vulkan --enable-libvorbis --enable-amf --enable-libaom --enable-avisynth --enable-libdav1d --enable-libdavs2 --enable-ffnvcodec --enable-cuda-llvm --enable-libglslang --enable-libass --enable-libbluray --enable-libmp3lame --enable-libopus --enable-libtheora --enable-libvpx --enable-libwebp --enable-libmfx --enable-libopencore-amrnb --enable-libopencore-amrwb --enable-libopenjpeg --enable-librav1e --enable-librubberband --enable-schannel --enable-sdl2 --enable-libsoxr --enable-libsrt --enable-libsvtav1 --enable-libtwolame --enable-libuavs3d --enable-libvidstab --enable-libx264 --enable-libx265 --enable-libxavs2 --enable-libxvid --enable-libzimg --extra-cflags=-DLIBTWOLAME_STATIC --extra-cxxflags= --extra-ldflags=-pthread --extra-libs=-lgomp
 libavutil 56. 63.100 / 56. 63.100
 libavcodec 58.115.102 / 58.115.102
 libavformat 58. 65.101 / 58. 65.101
 libavdevice 58. 11.103 / 58. 11.103
 libavfilter 7. 95.100 / 7. 95.100
 libswscale 5. 8.100 / 5. 8.100
 libswresample 3. 8.100 / 3. 8.100
 libpostproc 55. 8.100 / 55. 8.100
Input #0, mov,mp4,m4a,3gp,3g2,mj2, from 'D:\Dropbox\ACJ\ACJ Videos\Kirchen-Ox Blood_1080.mp4':
 Metadata:
 major_brand : mp42
 minor_version : 0
 compatible_brands: mp42mp41
 creation_time : 2021-01-09T17:24:45.000000Z
 Duration: 00:05:20.70, start: 0.000000, bitrate: 10135 kb/s
 Stream #0:0(eng): Video: h264 (Main) (avc1 / 0x31637661), yuv420p, 1920x1080 [SAR 1:1 DAR 16:9], 9815 kb/s, 29.97 fps, 29.97 tbr, 30k tbn, 59.94 tbc (default)
 Metadata:
 creation_time : 2021-01-09T17:24:45.000000Z
 handler_name : ?Mainconcept Video Media Handler
 vendor_id : [0][0][0][0]
 encoder : AVC Coding
 Stream #0:1(eng): Audio: aac (LC) (mp4a / 0x6134706D), 48000 Hz, stereo, fltp, 317 kb/s (default)
 Metadata:
 creation_time : 2021-01-09T17:24:45.000000Z
 handler_name : #Mainconcept MP4 Sound Media Handler
 vendor_id : [0][0][0][0]
Stream mapping:
 Stream #0:0 -> #0:0 (h264 (native) -> h264 (libx264))
 Stream #0:1 -> #0:1 (aac (native) -> aac (native))
[libx264 @ 0000024360133040] using SAR=1/1
[libx264 @ 0000024360133040] using cpu capabilities: MMX2 SSE2Fast SSSE3 SSE4.2 AVX FMA3 BMI2 AVX2
[libx264 @ 0000024360133040] profile Constrained Baseline, level 4.0, 4:2:0, 8-bit
[libx264 @ 0000024360133040] 264 - core 161 - H.264/MPEG-4 AVC codec - Copyleft 2003-2020 - http://www.videolan.org/x264.html - options: cabac=0 ref=1 deblock=0:0:0 analyse=0:0 me=dia subme=0 psy=1 psy_rd=1.00:0.00 mixed_ref=0 me_range=16 chroma_me=1 trellis=0 8x8dct=0 cqm=0 deadzone=21,11 fast_pskip=1 chroma_qp_offset=0 threads=12 lookahead_threads=2 sliced_threads=0 nr=0 decimate=1 interlaced=0 bluray_compat=0 constrained_intra=0 bframes=0 weightp=0 keyint=250 keyint_min=25 scenecut=0 intra_refresh=0 rc=crf mbtree=0 crf=23.0 qcomp=0.60 qpmin=0 qpmax=69 qpstep=4 ip_ratio=1.40 aq=0
Output #0, flv, to 'rtmp://rtmp-global.cloud.vimeo.com/live/149679ba-d76c-4bdf-895a-84fddccxxxxx':
 Metadata:
 major_brand : mp42
 minor_version : 0
 compatible_brands: mp42mp41
 encoder : Lavf58.65.101
 Stream #0:0(eng): Video: h264 ([7][0][0][0] / 0x0007), yuv420p(progressive), 1920x1080 [SAR 1:1 DAR 16:9], q=2-31, 29.97 fps, 1k tbn (default)
 Metadata:
 creation_time : 2021-01-09T17:24:45.000000Z
 handler_name : ?Mainconcept Video Media Handler
 vendor_id : [0][0][0][0]
 encoder : Lavc58.115.102 libx264
 Side data:
 cpb: bitrate max/min/avg: 0/0/0 buffer size: 0 vbv_delay: N/A
 Stream #0:1(eng): Audio: aac (LC) ([10][0][0][0] / 0x000A), 48000 Hz, stereo, fltp, 128 kb/s (default)
 Metadata:
 creation_time : 2021-01-09T17:24:45.000000Z
 handler_name : #Mainconcept MP4 Sound Media Handler
 vendor_id : [0][0][0][0]
 encoder : Lavc58.115.102 aac
frame= 531 fps= 21 q=23.0 size= 22503kB time=00:00:17.62 bitrate=10461.5kbits/s speed=0.686x


 



-
Create and insert new FFMPEG packet to live stream
25 septembre 2019, par TUrwinI’m currently receiving an FFMPEG encoded livestream and saving it directly to a .ts file.
The data has 2 encoded streams : One for video, and one for KLV metadata.
I have another dataset streaming in which is raw KLV data.
What I want to achieve is to encode the raw KLV packet data into FFMPEG packets and mux them into the same .ts file as the other data streams without having to decode and re-encode the existing FFMPEG packets.
Is there a way to achieve this ?
-
Queuing multiple video files as a single live video stream with nodejs and ffmpeg
24 avril 2016, par bomtempii would like to stream a video with node nodejs using multiple little video sequences that are created on the fly.
It means that the server queues multiple files, but the client must see it as only one single stream.Why would i need such feature ?
I have an application that creates images on the fly on the server, and i need to send them as a single video to the client, as if it was on a live source.
I cannot use ffmpeg to concatenate these files and convert them to a single video because i don’t have the complete sequence.I can do the trick with images sequences, using "multipart/x-mixed-replace" content-type, in a mjpeg format. Here is how i do :
http.createServer(function(req, stream) {
stream.writeHead(200, {
'Content-Type': 'multipart/x-mixed-replace;boundary=myboundary',
'Cache-Control': 'no-cache',
'Expires': '-1',
'Connection': 'close',
'Pragma': 'no-cache'
});
setInterval(function() {
var data = ... // here launch the function to get the binary data from the latest image created on the server
stream.write("--myboundary\r\n");
stream.write("Content-Type: image/jpeg\r\n");
stream.write("Content-Length: " + data.length + "\r\n");
stream.write("\r\n");
stream.write(data);
stream.write("\r\n");
}, 300);
}).listen(port, function() {
});The problem with this format is that it has no sound.
I tried to serve the sound from a different endpoint, and use ffmpeg to assemble them in live and send them back to the server where it will be forwarded to a new endpoint containing the sound and video, but after just a few seconds the video becomes more and more laggy and freezes.So that is why i thought that i should start by generating small videos on the server, by taking the 24 or 30 latest images created on the fly, concatenate them with ffmpeg and queue them on nodejs. That would just add approximately 1 second delay to the client.
Thanks for your help !