Recherche avancée

Médias (91)

Autres articles (63)

  • Les autorisations surchargées par les plugins

    27 avril 2010, par

    Mediaspip core
    autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs

  • Creating farms of unique websites

    13 avril 2011, par

    MediaSPIP platforms can be installed as a farm, with a single "core" hosted on a dedicated server and used by multiple websites.
    This allows (among other things) : implementation costs to be shared between several different projects / individuals rapid deployment of multiple unique sites creation of groups of like-minded sites, making it possible to browse media in a more controlled and selective environment than the major "open" (...)

  • Ajouter notes et légendes aux images

    7 février 2011, par

    Pour pouvoir ajouter notes et légendes aux images, la première étape est d’installer le plugin "Légendes".
    Une fois le plugin activé, vous pouvez le configurer dans l’espace de configuration afin de modifier les droits de création / modification et de suppression des notes. Par défaut seuls les administrateurs du site peuvent ajouter des notes aux images.
    Modification lors de l’ajout d’un média
    Lors de l’ajout d’un média de type "image" un nouveau bouton apparait au dessus de la prévisualisation (...)

Sur d’autres sites (9889)

  • Use ffmpeg to compile RGBA image sequence into two separate videos (RGB + Alpha)

    27 novembre 2018, par MirceaKitsune

    I plan on using Blender to render animated sequences for a game. I need the videos to have transparent backgrounds in the format expected by its engine. This format involves the transparency being defined as a separate grayscale video of equal FPS and duration. Since splitting the RGB and Alpha channels in Blender is more complicated, I’d prefer doing this directly from ffmpeg.

    The input directory contains an image sequence of png files in RGBA format (eg : 0000.png to 0100.png). What I need ffmpeg to do is compile them into two separate videos : One containing only the RGB channels (leave transparency black) and another containing only the grayscale alpha channel. In the end I’d have something like my_video.mp4 + my_video_mask.mp4.

    I’m familiar with compiling non-transparent image sequences into video using :

    ffmpeg -f image2 -i /path/to/sequence/%04d.png output.mp4

    But I don’t know how to extract the alpha channel and make it a separate video file. What is the simplest ffmpeg command to achieve this result ?

  • How can I render frames decoded by FFmpeg using hardware decoding with D3D11 ?

    14 juin 2024, par mercuric taylor

    I have completed the process of decoding a video frame using FFmpeg. The format of the decoded frame is AV_PIX_FMT_NV12. Now, I want to render this frame to the screen using D3D11. My questions are :

    


      

    1. What is the equivalent concept in D3D11 for a decoded frame ? Is it a texture ?
    2. 


    3. I have seen many solutions that convert NV12 data to RGB, but it seems that DX11 does not require this conversion anymore.
    4. 


    5. I just want to display this frame, and since my frame is on the GPU, is there a more convenient way to render directly on the GPU without copying ?
    6. 


    


    Please forgive my not-so-good English. Can anyone provide a reference example ?

    


    I have already referenced this open-source project. texthttps://github.com/balapradeepswork/D3D11NV12Rendering/tree/master/D3D11NV12Rendering
But I don't understand it very well.
Since I don't use dx11 to make game, just for show video, I am looking forward for a more easy solution. It has worried me for weeks, can anyone give me some advice(even good tutorial, some I can found is too old). Thanks sincerely !

    


  • avcodec/sanm : implement STOR/FTCH for ANIMv1

    14 mars, par Manuel Lauss
    avcodec/sanm : implement STOR/FTCH for ANIMv1
    

    Handle STOR/FTCH the same way the RA1 game engine does :
    On STOR, save the next following FOBJ (not the decoded image)
    in a buffer ; decode it on FTCH.
    The RA1 codecs and the fobj handler now take an explicit
    GetByteContext in order to be able to replay stored data.

    Used extensively by Rebel Assault 1 for e.g. backgrounds and
    the cockpit overlay.

    For ANIMv2 keep the current system to store the decoded image, as
    replaying a FOBJ would not work with codecs37/47/48 due to sequence
    violations.

    Signed-off-by : Manuel Lauss <manuel.lauss@gmail.com>

    • [DH] libavcodec/sanm.c