Recherche avancée

Médias (29)

Mot : - Tags -/Musique

Autres articles (86)

  • Keeping control of your media in your hands

    13 avril 2011, par

    The vocabulary used on this site and around MediaSPIP in general, aims to avoid reference to Web 2.0 and the companies that profit from media-sharing.
    While using MediaSPIP, you are invited to avoid using words like "Brand", "Cloud" and "Market".
    MediaSPIP is designed to facilitate the sharing of creative media online, while allowing authors to retain complete control of their work.
    MediaSPIP aims to be accessible to as many people as possible and development is based on expanding the (...)

  • Participer à sa traduction

    10 avril 2011

    Vous pouvez nous aider à améliorer les locutions utilisées dans le logiciel ou à traduire celui-ci dans n’importe qu’elle nouvelle langue permettant sa diffusion à de nouvelles communautés linguistiques.
    Pour ce faire, on utilise l’interface de traduction de SPIP où l’ensemble des modules de langue de MediaSPIP sont à disposition. ll vous suffit de vous inscrire sur la liste de discussion des traducteurs pour demander plus d’informations.
    Actuellement MediaSPIP n’est disponible qu’en français et (...)

  • MediaSPIP v0.2

    21 juin 2013, par

    MediaSPIP 0.2 est la première version de MediaSPIP stable.
    Sa date de sortie officielle est le 21 juin 2013 et est annoncée ici.
    Le fichier zip ici présent contient uniquement les sources de MediaSPIP en version standalone.
    Comme pour la version précédente, il est nécessaire d’installer manuellement l’ensemble des dépendances logicielles sur le serveur.
    Si vous souhaitez utiliser cette archive pour une installation en mode ferme, il vous faudra également procéder à d’autres modifications (...)

Sur d’autres sites (17301)

  • ffmpeg : Infinite length output when overlaying subtitles onto black image

    18 juillet 2020, par rosuav

    I'm trying to do some analysis of image-based subtitles by outputting them as a sequence of PNGs to a pipe. My command line looks like this :

    


    ffmpeg -y -i $INPUTFILE -f lavfi -i color=c=black:s=1920x1080 -filter_complex "[1:v][0:s:5]overlay[v]" -shortest -map "[v]" -c:v png -f image2pipe - | pike subspng.pike


    


    In theory, -shortest should mean that the stream stops at the shortest input, which would be roughly seven minutes of input file. Instead, my script receives an infinite sequence of black frames after the last frame of subtitles, until I send FFMPEG a SIGINT. Placing -shortest before -filter_complex has the same effect.

    


    Is there a different way to force the filtering to stop at the end of the input file ?

    


    EDIT : Using the shortest=1 flag on the overlay filter also doesn't help, even in combination with -shortest.

    


  • Is there any way in ffmpeg to cut a video with embedded subtitles and multiple audio tracks accordingly ? [closed]

    21 février 2024, par Harsh

    I have a video which had multiple audio tracks and embedded subtitles. I tried using ffmpeg to cut it, but I could not get the subtitles with it, nor the audio tracks. Here's the command I tried :

    


    ffmpeg -i input.mkv -ss 00:01:16 -to 00:02:46 -map "[0:0]:v" -map "[0:3]:a" -map "[0:4]:s" -c copy output.mkv 


    


    I tried using this command -

    


    ffprobe -v info input.mkv


    


    to collect information about the streams and substituted the streams, but for some reason it gave an error which read something along the lines of - Output with label 0:0 does not exist. Is there an alternative to achieve the result I wish ?

    


    Edit - I tried this command :

    


    ffmpeg -i input.mkv -ss 00:01:16 -to 00:02:46 -map 0 -c:a copy -c:s copy output.mkv


    


    It does the task BUT I have trouble seeing the video on my pc, it just shows a black screen (VLC) but on my mobile when I watch it using PLAYit, it works fine but I can't see the first 8 seconds for some reason. Anyways fix for this ?

    


  • Live streaming webvtt subtitles with HLS protocol

    5 mai, par Victor Ruiz

    I need a tool to generate HLS subtitles in live mode. Specifically, I want to create an HTTP server that serves .m3u8 and .webvtt files which are continuously updated over time.
Such HLS stream could be consumed via HTTP requests by HLS JS/ffplay multimedia players.

    


    The .webvtt files will be generated by an automatic transcriber, so the program must update the .m3u8 playlist accordingly whenever a new subtitle is produced.

    


    I only want to stream subtitle channels—no audio. The video channel can simply display a black chroma background.

    


    I attempted to use FFmpeg with a Linux pipe as input for the streaming .webvtt subtitles, along with a video file for the video stream. The output .webvtt and .m3u8 files were written to a folder and served via an NGINX server. However, FFmpeg fails after it reads the initial content of the .webvtt input from the pipe. If I inject more content afterward, it gets skipped.

    


    How can I achieve HLS subtitle streaming in live mode ? Can FFmpeg be used for this purpose, or do I need a different tool ?