Recherche avancée

Médias (1)

Mot : - Tags -/framasoft

Autres articles (85)

  • Gestion des droits de création et d’édition des objets

    8 février 2011, par

    Par défaut, beaucoup de fonctionnalités sont limitées aux administrateurs mais restent configurables indépendamment pour modifier leur statut minimal d’utilisation notamment : la rédaction de contenus sur le site modifiables dans la gestion des templates de formulaires ; l’ajout de notes aux articles ; l’ajout de légendes et d’annotations sur les images ;

  • (Dés)Activation de fonctionnalités (plugins)

    18 février 2011, par

    Pour gérer l’ajout et la suppression de fonctionnalités supplémentaires (ou plugins), MediaSPIP utilise à partir de la version 0.2 SVP.
    SVP permet l’activation facile de plugins depuis l’espace de configuration de MediaSPIP.
    Pour y accéder, il suffit de se rendre dans l’espace de configuration puis de se rendre sur la page "Gestion des plugins".
    MediaSPIP est fourni par défaut avec l’ensemble des plugins dits "compatibles", ils ont été testés et intégrés afin de fonctionner parfaitement avec chaque (...)

  • Activation de l’inscription des visiteurs

    12 avril 2011, par

    Il est également possible d’activer l’inscription des visiteurs ce qui permettra à tout un chacun d’ouvrir soit même un compte sur le canal en question dans le cadre de projets ouverts par exemple.
    Pour ce faire, il suffit d’aller dans l’espace de configuration du site en choisissant le sous menus "Gestion des utilisateurs". Le premier formulaire visible correspond à cette fonctionnalité.
    Par défaut, MediaSPIP a créé lors de son initialisation un élément de menu dans le menu du haut de la page menant (...)

Sur d’autres sites (9690)

  • lavu : add an API function to return the Libav version string

    2 juillet 2015, par wm4
    lavu : add an API function to return the Libav version string
    

    This returns something like "v12_dev0-1332-g333a27c". This is much more
    useful than the individual library versions, of which there are too
    many, and which are very hard to map back to releases or git commits.

    Signed-off-by : Janne Grunau <janne-libav@jannau.net>

    • [DH] .gitignore
    • [DH] Makefile
    • [DH] cmdutils.c
    • [DH] doc/APIchanges
    • [DH] libavutil/avutil.h
    • [DH] libavutil/utils.c
  • Produce waveform video from audio using FFMPEG

    30 novembre 2020, par RhythmicDevil

    I am trying to create a waveform video from audio. My goal is to produce a video that looks something like this

    &#xA;

    enter image description here

    &#xA;

    For my test I have an mp3 that plays a short clipped sound. There are 4 bars of 1/4 notes and 4 bars of 1/8 notes played at 120bpm. I am having some trouble coming up with the right combination of preprocessing and filtering to produce a video that looks like the image. The colors dont have to be exact, I am more concerned with the shape of the beats. I tried a couple of different approaches using showwaves and showspectrum. I cant quite wrap my head around why when using showwaves the beats go past so quickly, but using showspectrum produces a video where I can see each individual beat.

    &#xA;

    ShowWaves

    &#xA;

    ffmpeg -i beat_test.mp3 -filter_complex "[0:a]showwaves=s=1280x100:mode=cline:rate=25:scale=sqrt,format=yuv420p[v]" -map "[v]" -map 0:a output_wav.mp4&#xA;

    &#xA;

    This link will download the output of that command.

    &#xA;

    ShowSpectrum

    &#xA;

    ffmpeg -i beat_test.mp3 -filter_complex "[0:a]showspectrum=s=1280x100:mode=combined:color=intensity:saturation=5:slide=1:scale=cbrt,format=yuv420p[v]" -map "[v]" -an -map 0:a output_spec.mp4&#xA;

    &#xA;

    This link will download the output of that command.

    &#xA;

    I posted the simple examples because I didn't want to confuse the issue by adding all the variations I have tried.

    &#xA;

    In practice I suppose I can get away with the output from showspectrum but I'd like to understand where/how I am thinking about this incorrectly. Thanks for any advice.

    &#xA;

    Here is a link to the source audio file.

    &#xA;

  • ffmpeg png to png quality loss

    13 avril 2018, par kilo

    I did a python script that managed to unshuffle a shuffled (png) image according to a specific pattern, that python script uses ffmpeg and does 12 encodes to unshuffling it (by cropping a specific part and pasting it over the existing picture).
    As such the same file is re-encoded into a new file each time, which shouldn’t be a problem since i am doing png conversion (lossless, right ?), but i still lose quality on it.

    Here are the pictures :

    Notice the loss of quality on the "ONE PUNCH MAN" text. The rest of the picture is, seemingly, literally identical. So the problem seems to be with the colors.

    Here are the ffmpeg commands i ran to get to the output :

    ffmpeg -loglevel panic -y -i "output/001.png" -i "001.png" -qscale:v 2 -filter_complex "[0:v]crop=200:280:200:0[t];[0:v][t]overlay=0:280" "output/001.png"
    ffmpeg -loglevel panic -y -i "output/001.png" -i "001.png" -qscale:v 2 -filter_complex "[0:v]crop=200:280:400:0[t];[0:v][t]overlay=0:560" "output/001.png"
    ffmpeg -loglevel panic -y -i "output/001.png" -i "001.png" -qscale:v 2 -filter_complex "[0:v]crop=200:280:600:0[t];[0:v][t]overlay=0:840" "output/001.png"
    ffmpeg -loglevel panic -y -i "output/001.png" -i "001.png" -qscale:v 2 -filter_complex "[1:v]crop=200:280:0:280[t];[0:v][t]overlay=200:0" "output/001.png"
    ffmpeg -loglevel panic -y -i "output/001.png" -i "001.png" -qscale:v 2 -filter_complex "[0:v]crop=200:280:400:280[t];[0:v][t]overlay=200:560" "output/001.png"
    ffmpeg -loglevel panic -y -i "output/001.png" -i "001.png" -qscale:v 2 -filter_complex "[0:v]crop=200:280:600:280[t];[0:v][t]overlay=200:840" "output/001.png"
    ffmpeg -loglevel panic -y -i "output/001.png" -i "001.png" -qscale:v 2 -filter_complex "[1:v]crop=200:280:0:560[t];[0:v][t]overlay=400:0" "output/001.png"
    ffmpeg -loglevel panic -y -i "output/001.png" -i "001.png" -qscale:v 2 -filter_complex "[1:v]crop=200:280:200:560[t];[0:v][t]overlay=400:280" "output/001.png"
    ffmpeg -loglevel panic -y -i "output/001.png" -i "001.png" -qscale:v 2 -filter_complex "[0:v]crop=200:280:600:560[t];[0:v][t]overlay=400:840" "output/001.png"
    ffmpeg -loglevel panic -y -i "output/001.png" -i "001.png" -qscale:v 2 -filter_complex "[1:v]crop=200:280:0:840[t];[0:v][t]overlay=600:0" "output/001.png"
    ffmpeg -loglevel panic -y -i "output/001.png" -i "001.png" -qscale:v 2 -filter_complex "[1:v]crop=200:280:200:840[t];[0:v][t]overlay=600:280" "output/001.png"
    ffmpeg -loglevel panic -y -i "output/001.png" -i "001.png" -qscale:v 2 -filter_complex "[1:v]crop=200:280:400:840[t];[0:v][t]overlay=600:560" "output/001.png"

    Anyone got any idea why is there this quality loss ?
    Strangely enough, there is no quality loss when i do it in an entirely different way (crop each square into an individual file, then each of them are put into a 1x2 vstack with the next one, then each of the resulting 1x2 files are vstacked with a second file to make a 1x4 file, then each of those are hstacked to make a 2x4 file, and finally we hstack the two resulting file for the resulting 4x4 output), even though there is more than double the amount of encodes.