Recherche avancée

Médias (1)

Mot : - Tags -/musée

Autres articles (105)

  • Support audio et vidéo HTML5

    10 avril 2011

    MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
    Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
    Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
    Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...)

  • HTML5 audio and video support

    13 avril 2011, par

    MediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
    The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
    For older browsers the Flowplayer flash fallback is used.
    MediaSPIP allows for media playback on major mobile platforms with the above (...)

  • De l’upload à la vidéo finale [version standalone]

    31 janvier 2010, par

    Le chemin d’un document audio ou vidéo dans SPIPMotion est divisé en trois étapes distinctes.
    Upload et récupération d’informations de la vidéo source
    Dans un premier temps, il est nécessaire de créer un article SPIP et de lui joindre le document vidéo "source".
    Au moment où ce document est joint à l’article, deux actions supplémentaires au comportement normal sont exécutées : La récupération des informations techniques des flux audio et video du fichier ; La génération d’une vignette : extraction d’une (...)

Sur d’autres sites (11128)

  • How to port signal() to sigaction() ?

    28 septembre 2016, par Shark

    due to recent problems discovered with NDK12 and NDK13b2, i’m thinking of ’porting’ libx264’s use of signal() (and missing bsd_signal() in ndk12) to use sigaction() instead.

    The problem is, I’m not quite sure what’s the simple&fastest way to replace signal() calls with sigaction() ones.

    For all i see, it’s mainly used in x264-snapshot/common/cpu.c in the following manner :

    using the following signal handler :

    static void sigill_handler( int sig )
    {
       if( !canjump )
       {
           signal( sig, SIG_DFL );
           raise( sig );
       }

       canjump = 0;
       siglongjmp( jmpbuf, 1 );
    }

    This is the problematic x264_cpu_detect function... currently, i’m guessing i only need to tackle the ARM version, but i’ ; ; still have to replace all occurances of signal() with sigaction() so i might just cover both of them to get the thing building...

    FYI - the NDK13 beta2 still has "unstable" libc and the build doesn’t fail on this part, but rather the first invocation of the rand() function somewhere else... So i’m out of luck and replacing the signal() calls might be better than just waiting for the official NDK13 release. I’m doing this to get rid of text-relocations so i can run the library (and doubango) on API 24 (Android N)

    the problematic part of function that invokes signal() :

    #elif SYS_LINUX

    uint32_t x264_cpu_detect( void )
    {
       static void (*oldsig)( int );

       oldsig = signal( SIGILL, sigill_handler );
       if( sigsetjmp( jmpbuf, 1 ) )
       {
           signal( SIGILL, oldsig );
           return 0;
       }

       canjump = 1;
       asm volatile( "mtspr 256, %0\n\t"
                     "vand 0, 0, 0\n\t"
                     :
                     : "r"(-1) );
       canjump = 0;

       signal( SIGILL, oldsig );

       return X264_CPU_ALTIVEC;
    }
    #endif

    #elif ARCH_ARM

    void x264_cpu_neon_test( void );
    int x264_cpu_fast_neon_mrc_test( void );

    uint32_t x264_cpu_detect( void )
    {
       int flags = 0;
    #if HAVE_ARMV6
       flags |= X264_CPU_ARMV6;

       // don't do this hack if compiled with -mfpu=neon
    #if !HAVE_NEON
       static void (* oldsig)( int );
       oldsig = signal( SIGILL, sigill_handler );
       if( sigsetjmp( jmpbuf, 1 ) )
       {
           signal( SIGILL, oldsig );
           return flags;
       }

       canjump = 1;
       x264_cpu_neon_test();
       canjump = 0;
       signal( SIGILL, oldsig );
    #endif

       flags |= X264_CPU_NEON;

       // fast neon -> arm (Cortex-A9) detection relies on user access to the
       // cycle counter; this assumes ARMv7 performance counters.
       // NEON requires at least ARMv7, ARMv8 may require changes here, but
       // hopefully this hacky detection method will have been replaced by then.
       // Note that there is potential for a race condition if another program or
       // x264 instance disables or reinits the counters while x264 is using them,
       // which may result in incorrect detection and the counters stuck enabled.
       // right now Apple does not seem to support performance counters for this test
    #ifndef __MACH__
       flags |= x264_cpu_fast_neon_mrc_test() ? X264_CPU_FAST_NEON_MRC : 0;
    #endif
       // TODO: write dual issue test? currently it's A8 (dual issue) vs. A9 (fast      mrc)
    #endif
       return flags;
    }

    #else

    uint32_t x264_cpu_detect( void )
    {
       return 0;
    }

    So the question is really this : what would be the quickest/easiest//fastest way to replace the signal() calls with sigaction() ones while preserving the current functionality ?

    EDIT :
    The reason i’m trying to get rid of signal() are these build errors :

    /home/devshark/SCRATCH/doubango/thirdparties/android/armv5te/lib/dist/libx264.a(cpu.o):cpu.c:function sigill_handler: error: undefined reference to 'bsd_signal'

    /home/devshark/SCRATCH/doubango/thirdparties/android/armv5te/lib/dist/libx264.a(cpu.o):cpu.c:function x264_cpu_detect: error: undefined reference to 'bsd_signal'
    /home/devshark/SCRATCH/doubango/thirdparties/android/armv5te/lib/dist/libx264.a(cpu.o):cpu.c:function x264_cpu_detect: error: undefined reference to 'bsd_signal'

    /home/devshark/SCRATCH/doubango/thirdparties/android/armv5te/lib/dist/libx264.a(cpu.o):cpu.c:function x264_cpu_detect: error: undefined reference to 'bsd_signal'

    I already know that this is a known NDK12 problem, that might be solved by bringing bsd_signal back to the libc in NDK13. However, in it’ beta state with it’s unstable libc - it’s currently missing the rand() function and simply waiting for it might not do the trick. But in the worst-case scenario, i guess i’ll just have to wait for it and retry after it’s release.

    But as it currently is, the prebuilt version of the library i want to use has text-relocations and is being rejected by phones running newer API / version of the android OS.

    EDIT2 :
    I also know that signal() usually works by using sigaction() under the hood, but maybe i won’t get bsd_signal related build-errors... since i’m suspecting that this one isn’t using it. It’s obviously using bsd_signal, which may or may not be the same underlying thing :/

  • FFmpeg : Create a Video Slideshow from PNG Images for MS PowerPoint

    20 septembre 2016, par Klaidonis

    I am using FFmpeg on Windows 7 to create a video from PNG image sequence for Microsoft PowerPoint.

    The best results I have achieved so far, is by using the following command :

    ffmpeg -framerate 10 -start_number 3 -i .\folder\name_%d.png -q:v 0 -c:v libx264 -pix_fmt yuv420p output-video.avi

    It seems perfect, 50 images of total size 30 MB are converted into 200 KB video with no loss in the quality. Placing it in PowerPoint also seems right, but there is a slight color shift (yellow appears darker and possibly more orangish). By using some other conversion options, I obtained a video in PowerPoint where the first image of the video (like album cover art) is exactly as the original but the rest of the video plays with the mentioned color shift.

    When I play this file in VLC, it’s good. Although, if in the settings "Use hardware YUV->RGB conversions" is enabled, colors appear a bit washed out, and white color is a bit gray.

    I also tried to convert the images to a GIF file and at first it seems good but outer edges and numbers from the left and top side are blurred, and the background has turned from white to a bit gray color, although, white segments not on the background are white. The output size is 18 MB. I ended up with a way better GIF 600 KB by converting from the first video file ; just it is slightly more dotted, and the background again is grayish.

    ffmpeg -i output-video.avi output-video.gif

    Could someone help ?

  • Shadow on Central Image via FFMPEG [closed]

    9 octobre 2023, par Ashish Chaturvedi

    enter image description here

    


    To create an image for use in portrait orientations and for YouTube thumbnails in both 16 by 9 and 9 by 16 aspect ratios,

    


    I want to add a shadow effect to the central cover art image.

    


    The desired effect involves placing the main image in the center and surrounding it with a blurred background.

    


    The primary focus is to achieve a shadowing effect on the central image via the FFMPEG command as show in the below image.

    


    Any help/idea will be more helpful.