Recherche avancée

Médias (2)

Mot : - Tags -/map

Autres articles (35)

  • Support de tous types de médias

    10 avril 2011

    Contrairement à beaucoup de logiciels et autres plate-formes modernes de partage de documents, MediaSPIP a l’ambition de gérer un maximum de formats de documents différents qu’ils soient de type : images (png, gif, jpg, bmp et autres...) ; audio (MP3, Ogg, Wav et autres...) ; vidéo (Avi, MP4, Ogv, mpg, mov, wmv et autres...) ; contenu textuel, code ou autres (open office, microsoft office (tableur, présentation), web (html, css), LaTeX, Google Earth) (...)

  • Supporting all media types

    13 avril 2011, par

    Unlike most software and media-sharing platforms, MediaSPIP aims to manage as many different media types as possible. The following are just a few examples from an ever-expanding list of supported formats : images : png, gif, jpg, bmp and more audio : MP3, Ogg, Wav and more video : AVI, MP4, OGV, mpg, mov, wmv and more text, code and other data : OpenOffice, Microsoft Office (Word, PowerPoint, Excel), web (html, CSS), LaTeX, Google Earth and (...)

  • Création définitive du canal

    12 mars 2010, par

    Lorsque votre demande est validée, vous pouvez alors procéder à la création proprement dite du canal. Chaque canal est un site à part entière placé sous votre responsabilité. Les administrateurs de la plateforme n’y ont aucun accès.
    A la validation, vous recevez un email vous invitant donc à créer votre canal.
    Pour ce faire il vous suffit de vous rendre à son adresse, dans notre exemple "http://votre_sous_domaine.mediaspip.net".
    A ce moment là un mot de passe vous est demandé, il vous suffit d’y (...)

Sur d’autres sites (7384)

  • ffmpeg continuous live streaming of dynamic files from directory

    4 septembre 2022, par user72261

    So I have been using ffmpeg to create a live stream from files stored in a folder using concat list.

    


    This works perfectly for local, static content but I have a security system in my warehouse that uploads .mp4 files to my server in the office.

    


    This system doesn't have any live stream functionality so I would like to set up a live stream myself using ffmpeg to continually read the latest files from the folder.

    


    I tried setting up a script to read all files in the directory and sort them based on the file name that is a UNIX timestamp of the date/time the video was taken. Then had the script select the latest segments after that last used segment and create a new concat list such as below

    


    file '00000.mpd'
file '00001.mp4'
file '00002.mp4'
file '00003.mp4'
file '00004.mp4'


    


    After doing some reading I found out to make it loop that file you could reference its self in the list like so

    


    file '00000.mpd'
file '00001.mp4'
file '00002.mp4'
file '00003.mp4'
file '00004.mp4'
file 'concat.txt'


    


    This seemed to work ok until I started getting the error impossible to open concat.txt. This file doesn't exist, which I can only guess is because ffmpeg is trying to open the file as its being updated by the script.

    


    Has anyone come across this before or know another way to do it ?

    


  • Subtitles in ffmpeg/libavfilter

    15 juin 2021, par Captain Jack

    I have a C program to read video/audio with libav/ffmpeg libraries and decode it.

    



    I am playing with some filters and most work just fine. I can draw text, overlay logos, flip and invert video colours. However, I am having big issues overlaying subtitles.

    



    My filter is very simple.

    



    const char *vfilter_descr = "[in]subtitles=subs.srt[out]";


    



    On the console I get this :

    



    [Parsed_subtitles_0 @ 0x7fe76c703240] Shaper: FriBidi 0.19.7 (SIMPLE) HarfBuzz-ng 2.4.0 (COMPLEX)
[Parsed_subtitles_0 @ 0x7fe76c703240] Using font provider coretext
[Parsed_subtitles_0 @ 0x7fe76c703240] fontselect: (Arial, 400, 0) -> /Library/Fonts/Microsoft/Arial.ttf, -1, ArialMT
[Parsed_subtitles_0 @ 0x7fe76c703240] fontselect: (Arial, 400, 100) -> /Library/Fonts/Microsoft/Arial Italic.ttf, -1, Arial-ItalicMT


    



    ...which somewhat confirms that subtitles are loading, though I am not sure why there are two fonts being loaded ?

    



    However, they are not showing at all - almost as if they never loaded. I tried several different files, including ASS ones but no luck.

    



    ffmpeg version is the latest one.

    



    $ ffmpeg -v
ffmpeg version 4.1.3 Copyright (c) 2000-2019 the FFmpeg developers
  built with Apple LLVM version 9.0.0 (clang-900.0.39.2)
  configuration: --prefix=/usr/local/Cellar/ffmpeg/4.1.3_1 --enable-shared --enable-pthreads --enable-version3 --enable-hardcoded-tables --enable-avresample --cc=clang --host-cflags='-I/Library/Java/JavaVirtualMachines/adoptopenjdk-11.0.2.jdk/Contents/Home/include -I/Library/Java/JavaVirtualMachines/adoptopenjdk-11.0.2.jdk/Contents/Home/include/darwin' --host-ldflags= --enable-ffplay --enable-gnutls --enable-gpl --enable-libaom --enable-libbluray --enable-libmp3lame --enable-libopus --enable-librubberband --enable-libsnappy --enable-libtesseract --enable-libtheora --enable-libvorbis --enable-libvpx --enable-libx264 --enable-libx265 --enable-libxvid --enable-lzma --enable-libfontconfig --enable-libfreetype --enable-frei0r --enable-libass --enable-libopencore-amrnb --enable-libopencore-amrwb --enable-libopenjpeg --enable-librtmp --enable-libspeex --enable-videotoolbox --disable-libjack --disable-indev=jack --enable-libaom --enable-libsoxr
  libavutil      56. 22.100 / 56. 22.100
  libavcodec     58. 35.100 / 58. 35.100
  libavformat    58. 20.100 / 58. 20.100
  libavdevice    58.  5.100 / 58.  5.100
  libavfilter     7. 40.101 /  7. 40.101
  libavresample   4.  0.  0 /  4.  0.  0
  libswscale      5.  3.100 /  5.  3.100
  libswresample   3.  3.100 /  3.  3.100
  libpostproc    55.  3.100 / 55.  3.100


    



    Any ideas ?

    


  • Mixed Reality WebRTC without Signalling Server

    25 mai 2021, par SilverLife

    I am trying to find a way, which allows me to use Mixed Reality WebRTC (link to git-repo) without a signalling server.
In detail, I want to create a sdp-file from my ffmpeg Video sender and use this sdp-description in my unity-Project to bypass the signaling process and receive the ffmpeg video stream.
Is there a way of doing so with Mixed Reality WebRTC ? I was already searching for the line of code, where the sdp-file is created within MR WebRTC but I didn´t find it.

    


    I am relatively new to this topic and I am not sure if this works at all but since ffmpeg is not directly compatible with webrtc I was thinking that this might be the most promising approach.