Recherche avancée

Médias (0)

Mot : - Tags -/metadatas

Aucun média correspondant à vos critères n’est disponible sur le site.

Autres articles (52)

  • Mise à jour de la version 0.1 vers 0.2

    24 juin 2013, par

    Explications des différents changements notables lors du passage de la version 0.1 de MediaSPIP à la version 0.3. Quelles sont les nouveautés
    Au niveau des dépendances logicielles Utilisation des dernières versions de FFMpeg (>= v1.2.1) ; Installation des dépendances pour Smush ; Installation de MediaInfo et FFprobe pour la récupération des métadonnées ; On n’utilise plus ffmpeg2theora ; On n’installe plus flvtool2 au profit de flvtool++ ; On n’installe plus ffmpeg-php qui n’est plus maintenu au (...)

  • Personnaliser en ajoutant son logo, sa bannière ou son image de fond

    5 septembre 2013, par

    Certains thèmes prennent en compte trois éléments de personnalisation : l’ajout d’un logo ; l’ajout d’une bannière l’ajout d’une image de fond ;

  • Ecrire une actualité

    21 juin 2013, par

    Présentez les changements dans votre MédiaSPIP ou les actualités de vos projets sur votre MédiaSPIP grâce à la rubrique actualités.
    Dans le thème par défaut spipeo de MédiaSPIP, les actualités sont affichées en bas de la page principale sous les éditoriaux.
    Vous pouvez personnaliser le formulaire de création d’une actualité.
    Formulaire de création d’une actualité Dans le cas d’un document de type actualité, les champs proposés par défaut sont : Date de publication ( personnaliser la date de publication ) (...)

Sur d’autres sites (6869)

  • Hello, I have some problems with ffmpeg.concat, there are logs

    12 mai 2022, par Matvey

    Log :

    


    [swscaler @ 0x27d3c10] [swscaler @ 0x2886fe0] deprecated pixel format used, make sure you did set range correctly
[swscaler @ 0x27d3c10] [swscaler @ 0x283be10] deprecated pixel format used, make sure you did set range correctly
Output #0, avi, to '/home/pi/Desktop/VIDEOREG/2022:05:12_09:51:57:856_v1.avi':
  Metadata:
    software        : Lavf58.45.100
    ISFT            : Lavf59.20.101
  Stream #0:0: Video: mpeg4 (FMP4 / 0x34504D46), yuv420p(tv, bt470bg/unknown/unknown, progressive), 1080x720, q=2-31, 200 kb/s, 25 fps, 25>
    Metadata:
      encoder         : Lavc59.25.100 mpeg4
    Side data:
      cpb: bitrate max/min/avg: 0/0/200000 buffer size: 0 vbv_delay: N/A
Output #0, avi, to '/home/pi/Desktop/VIDEOREG/2022:05:12_09:51:57:911_v3.avi':
  Metadata:
    software        : Lavf58.45.100
    ISFT            : Lavf59.20.101
  Stream #0:0: Video: mpeg4 (FMP4 / 0x34504D46), yuv420p(tv, bt470bg/unknown/unknown, progressive), 1080x720, q=2-31, 200 kb/s, 25 fps, 25>
    Metadata:
      encoder         : Lavc59.25.100 mpeg4
    Side data:
      cpb: bitrate max/min/avg: 0/0/200000 buffer size: 0 vbv_delay: N/A
Output #0, avi, to '/home/pi/Desktop/VIDEOREG/2022:05:12_09:51:57:890_v2.avi':
  Metadata:
    software        : Lavf58.45.100
    ISFT            : Lavf59.20.101
  Stream #0:0: Video: mpeg4 (FMP4 / 0x34504D46), yuv420p(tv, bt470bg/unknown/unknown, progressive), 1080x720, q=2-31, 200 kb/s, 25 fps, 25>
    Metadata:
      encoder         : Lavc59.25.100 mpeg4
    Side data:
      cpb: bitrate max/min/avg: 0/0/200000 buffer size: 0 vbv_delay: N/A
[mjpeg @ 0x27aaa10] mjpeg_decode_dc: bad vlc: 0:0 (0x27b7c20)trate=8122.4kbits/s speed=1.33e+04x
[mjpeg @ 0x27aaa10] error dc
[mjpeg @ 0x27aaa10] error y=18 x=37
[mjpeg @ 0x27aaa10] EOI missing, emulating
[mjpeg @ 0x29551c0] error count: 65
[mjpeg @ 0x29551c0] error y=2 x=17
###############################################


    


    My part of code for concatenation audio and video :

    


    def video_with_audio(number, filename, video):
        ms = datetime.today().strftime('%f') + '0000'
        ms = ms[:3]
        name = datetime.today().strftime('%Y:%m:%d_%H:%M:%S:')
        title_video = f'/home/pi/Desktop/VIDEOREG/{name}{ms}{number}.avi'
        print(title_video)
        os.system("ffmpeg -i " + video + " -vf setpts=PTS*2 " + title_video)
        ####################################################
        ms_1 = datetime.today().strftime('%f') + '0000'
        ms_1 = ms_1[:3]
        name_1 = datetime.today().strftime('%Y:%m:%d_%H:%M:%S:')
        title_video = f'/home/pi/Desktop/VIDEOREG/{name}{ms}{number}.avi'
        title_video_1 = f'/home/pi/Desktop/VIDEOREG/{name_1}{ms_1}{number}.avi'
        input_audio = ffmpeg.input(filename)
        input_video = ffmpeg.input(title_video)
        print(title_video_1)
        ffmpeg.concat(input_video, input_audio, v=1, a=1).output(title_video_1).run()


    


    There are slight delays during video recording, because I print information from sensors on video, can this affect the concatenation of video with audio ? (Saved video is accelerated ? that's why at the beginning I used slowing video down.)

    


  • FFMPEG video editing application. Need time and date stamp burned into video

    11 mai 2022, par Jacob

    I am developing an application for video editing. The main component of this application is to produce a single video file from several video files captured from a camcorder with the time and date stamp displayed on the final rendered video, much like the final product from a security camera. I have figured out, by using FFMPEG, how to burn the date and time into the video with a .SRT file as well as with DrawText like the following :

    


    ffmpeg -y -i video.mp4 -vf “drawtext=fontfile=roboto.ttf:fontsize=12:fontcolor=yellow:text='%{pts\:localtime\:1575526882\:%A, %d, %B %Y %I\\\:%M\\\:%S %p}'" -preset ultrafast -f mp4 output_new.mp4    


    


    I would rather use the DrawText method so the user does not have to wait longer while creating the .SRT files. I am new to FFMPEG and I find their documentation very confusing. I guess I am hoping there is someone out there who has experience with it.

    


    Everything seems to work when I pass in the date created meta data from the video file and drawtext just does its thing. The problem is my application allows for editing of the video. I do this, for lack of better solution, by allowing the user to select beginning and ending frames they do not want, from the UI and then the code simply deletes the frames from the directory where they were split and saved. I then use FFMPEG to iterate through the directory and combine the remaining frames to make a video file.

    


    This approach starts the time and date from the date created metadata ; however, cutting the frames out of the video will make the DT stamp inaccurate, due to the missing frames.

    


    Is there any way to tell FFMPEG to burn in the date and time from date/time retrieved from each individual frame ? I appreciate any advice that you may have.

    


  • How to use ffmpeg to overlay waveforms on xstack mosaics and specify specific audio for playback

    1er mai 2022, par kellib

    I would like to make a mosaic of multiple titled streams, 1) specifying which of the audio streams to play and 2) overlay waveforms at the bottom of each of the video tiles for the audio that they belong to.

    


    I'm successfully able to create the titled mosaic of streams with the code below.

    


    However :

    


      

    1. I'm having a hard time figuring out how to specify just one of the specific audio sources. I found amix, but I don't really want to mix them, I just want to specify audio [a0], or [a1], or [a2], etc.
    2. 


    


    and

    


      

    1. I'm having a hard time figuring out how to overlay the wave forms at the bottom of the video for each of the tiles. I struggled trying to figure out putting showwaves into the mix. Is it possible ?
    2. 


    


    I want each tile to look like this, but since these are rtmp streams, they need to play-out the matching waveforms dynamically with each stream. https://dragonquest64.blogspot.com/2020/01/ffmpeg-audio-waveform.html

    


    If someone could point me in the right direction, that would be great. I'm getting close, but I'm pretty new to all of this, and have already spent way more time than I should have, so would love a little help.

    


    ffmpeg \
-i rtmp://my.cdn.com/srcEncoders/STREAM-1 \
-i rtmp://my.cdn.com/srcEncoders/STREAM-2 \
-i rtmp://my.cdn.com/srcEncoders/STREAM-3 \
-i rtmp://my.cdn.com/srcEncoders/STREAM-4 \
  -filter_complex " \
      [0:v] setpts=PTS-STARTPTS, scale=qvga \
    , drawtext=text=STREAM-1:fontsize=20:x=10:y=10:fontcolor=white:box=1:boxcolor=black@0.5:boxborderw=5 [a0]; \
      [1:v] setpts=PTS-STARTPTS, scale=qvga \
    , drawtext=text=STREAM-2:fontsize=20:x=10:y=10:fontcolor=white:box=1:boxcolor=black@0.5:boxborderw=5 [a1]; \
      [2:v] setpts=PTS-STARTPTS, scale=qvga \
    , drawtext=text=STREAM-3:fontsize=20:x=10:y=10:fontcolor=white:box=1:boxcolor=black@0.5:boxborderw=5 [a2]; \
      [3:v] setpts=PTS-STARTPTS, scale=qvga \ 
    , drawtext=text=STREAM-4:fontsize=20:x=10:y=10:fontcolor=white:box=1:boxcolor=black@0.5:boxborderw=5 [a3]; \
      [a0][a1][a2][a3]xstack=inputs=4:layout=0_0|0_h0|w0_0|w0_h0[out]; \
    amix=inputs=1
      " \
  -map "[out]" \
 -c:v libx264 -b:v 1000k -g 30 -keyint_min 120 -profile:v baseline -preset veryfast -f mpegts "udp://127.0.0.1:1234?pkt_size=1316"