
Recherche avancée
Médias (2)
-
Rennes Emotion Map 2010-11
19 octobre 2011, par
Mis à jour : Juillet 2013
Langue : français
Type : Texte
-
Carte de Schillerkiez
13 mai 2011, par
Mis à jour : Septembre 2011
Langue : English
Type : Texte
Autres articles (35)
-
Support de tous types de médias
10 avril 2011Contrairement à beaucoup de logiciels et autres plate-formes modernes de partage de documents, MediaSPIP a l’ambition de gérer un maximum de formats de documents différents qu’ils soient de type : images (png, gif, jpg, bmp et autres...) ; audio (MP3, Ogg, Wav et autres...) ; vidéo (Avi, MP4, Ogv, mpg, mov, wmv et autres...) ; contenu textuel, code ou autres (open office, microsoft office (tableur, présentation), web (html, css), LaTeX, Google Earth) (...)
-
Supporting all media types
13 avril 2011, parUnlike most software and media-sharing platforms, MediaSPIP aims to manage as many different media types as possible. The following are just a few examples from an ever-expanding list of supported formats : images : png, gif, jpg, bmp and more audio : MP3, Ogg, Wav and more video : AVI, MP4, OGV, mpg, mov, wmv and more text, code and other data : OpenOffice, Microsoft Office (Word, PowerPoint, Excel), web (html, CSS), LaTeX, Google Earth and (...)
-
Création définitive du canal
12 mars 2010, parLorsque votre demande est validée, vous pouvez alors procéder à la création proprement dite du canal. Chaque canal est un site à part entière placé sous votre responsabilité. Les administrateurs de la plateforme n’y ont aucun accès.
A la validation, vous recevez un email vous invitant donc à créer votre canal.
Pour ce faire il vous suffit de vous rendre à son adresse, dans notre exemple "http://votre_sous_domaine.mediaspip.net".
A ce moment là un mot de passe vous est demandé, il vous suffit d’y (...)
Sur d’autres sites (7384)
-
ffmpeg continuous live streaming of dynamic files from directory
4 septembre 2022, par user72261So I have been using ffmpeg to create a live stream from files stored in a folder using concat list.


This works perfectly for local, static content but I have a security system in my warehouse that uploads .mp4 files to my server in the office.


This system doesn't have any live stream functionality so I would like to set up a live stream myself using ffmpeg to continually read the latest files from the folder.


I tried setting up a script to read all files in the directory and sort them based on the file name that is a UNIX timestamp of the date/time the video was taken. Then had the script select the latest segments after that last used segment and create a new concat list such as below


file '00000.mpd'
file '00001.mp4'
file '00002.mp4'
file '00003.mp4'
file '00004.mp4'



After doing some reading I found out to make it loop that file you could reference its self in the list like so


file '00000.mpd'
file '00001.mp4'
file '00002.mp4'
file '00003.mp4'
file '00004.mp4'
file 'concat.txt'



This seemed to work ok until I started getting the error impossible to open concat.txt. This file doesn't exist, which I can only guess is because ffmpeg is trying to open the file as its being updated by the script.


Has anyone come across this before or know another way to do it ?


-
Subtitles in ffmpeg/libavfilter
15 juin 2021, par Captain JackI have a C program to read video/audio with libav/ffmpeg libraries and decode it.



I am playing with some filters and most work just fine. I can draw text, overlay logos, flip and invert video colours. However, I am having big issues overlaying subtitles.



My filter is very simple.



const char *vfilter_descr = "[in]subtitles=subs.srt[out]";




On the console I get this :



[Parsed_subtitles_0 @ 0x7fe76c703240] Shaper: FriBidi 0.19.7 (SIMPLE) HarfBuzz-ng 2.4.0 (COMPLEX)
[Parsed_subtitles_0 @ 0x7fe76c703240] Using font provider coretext
[Parsed_subtitles_0 @ 0x7fe76c703240] fontselect: (Arial, 400, 0) -> /Library/Fonts/Microsoft/Arial.ttf, -1, ArialMT
[Parsed_subtitles_0 @ 0x7fe76c703240] fontselect: (Arial, 400, 100) -> /Library/Fonts/Microsoft/Arial Italic.ttf, -1, Arial-ItalicMT




...which somewhat confirms that subtitles are loading, though I am not sure why there are two fonts being loaded ?



However, they are not showing at all - almost as if they never loaded. I tried several different files, including ASS ones but no luck.



ffmpeg version is the latest one.



$ ffmpeg -v
ffmpeg version 4.1.3 Copyright (c) 2000-2019 the FFmpeg developers
 built with Apple LLVM version 9.0.0 (clang-900.0.39.2)
 configuration: --prefix=/usr/local/Cellar/ffmpeg/4.1.3_1 --enable-shared --enable-pthreads --enable-version3 --enable-hardcoded-tables --enable-avresample --cc=clang --host-cflags='-I/Library/Java/JavaVirtualMachines/adoptopenjdk-11.0.2.jdk/Contents/Home/include -I/Library/Java/JavaVirtualMachines/adoptopenjdk-11.0.2.jdk/Contents/Home/include/darwin' --host-ldflags= --enable-ffplay --enable-gnutls --enable-gpl --enable-libaom --enable-libbluray --enable-libmp3lame --enable-libopus --enable-librubberband --enable-libsnappy --enable-libtesseract --enable-libtheora --enable-libvorbis --enable-libvpx --enable-libx264 --enable-libx265 --enable-libxvid --enable-lzma --enable-libfontconfig --enable-libfreetype --enable-frei0r --enable-libass --enable-libopencore-amrnb --enable-libopencore-amrwb --enable-libopenjpeg --enable-librtmp --enable-libspeex --enable-videotoolbox --disable-libjack --disable-indev=jack --enable-libaom --enable-libsoxr
 libavutil 56. 22.100 / 56. 22.100
 libavcodec 58. 35.100 / 58. 35.100
 libavformat 58. 20.100 / 58. 20.100
 libavdevice 58. 5.100 / 58. 5.100
 libavfilter 7. 40.101 / 7. 40.101
 libavresample 4. 0. 0 / 4. 0. 0
 libswscale 5. 3.100 / 5. 3.100
 libswresample 3. 3.100 / 3. 3.100
 libpostproc 55. 3.100 / 55. 3.100




Any ideas ?


-
Mixed Reality WebRTC without Signalling Server
25 mai 2021, par SilverLifeI am trying to find a way, which allows me to use Mixed Reality WebRTC (link to git-repo) without a signalling server.
In detail, I want to create a sdp-file from my ffmpeg Video sender and use this sdp-description in my unity-Project to bypass the signaling process and receive the ffmpeg video stream.
Is there a way of doing so with Mixed Reality WebRTC ? I was already searching for the line of code, where the sdp-file is created within MR WebRTC but I didn´t find it.


I am relatively new to this topic and I am not sure if this works at all but since ffmpeg is not directly compatible with webrtc I was thinking that this might be the most promising approach.