Recherche avancée

Médias (1)

Mot : - Tags -/3GS

Autres articles (104)

  • Support audio et vidéo HTML5

    10 avril 2011

    MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
    Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
    Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
    Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...)

  • HTML5 audio and video support

    13 avril 2011, par

    MediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
    The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
    For older browsers the Flowplayer flash fallback is used.
    MediaSPIP allows for media playback on major mobile platforms with the above (...)

  • De l’upload à la vidéo finale [version standalone]

    31 janvier 2010, par

    Le chemin d’un document audio ou vidéo dans SPIPMotion est divisé en trois étapes distinctes.
    Upload et récupération d’informations de la vidéo source
    Dans un premier temps, il est nécessaire de créer un article SPIP et de lui joindre le document vidéo "source".
    Au moment où ce document est joint à l’article, deux actions supplémentaires au comportement normal sont exécutées : La récupération des informations techniques des flux audio et video du fichier ; La génération d’une vignette : extraction d’une (...)

Sur d’autres sites (9937)

  • Revision f167433d9c : fix the mv_ref_idx issue The following issue was reported : https://code.google

    20 août 2013, par Jim Bankoski

    Changed Paths :
     Modify /vp9/common/vp9_mvref_common.c



    fix the mv_ref_idx issue

    The following issue was reported :
    https://code.google.com/p/webm/issues/detail?id=601&q=jimbankoski&sort=-id&colsp
    ec=ID%20Pri%20mstone%20ReleaseBlock%20Type%20Component%20Status%20Owner%20Summar
    y

    This code makes the choice and code cleaner and removes any question
    about whether the border needs to be checked.

    Change-Id : Ia7aecfb3168e340618805bd318499176c2989597

  • how to us google ML kit Selfie segmentation in flutter to overlay segmented user on top of chosen image or video ? [closed]

    4 avril 2024, par Amr

    I'm working on implementing a virtual background feature using the Google ML Kit selfie segmentation Flutter plugin. The goal is to separate the user from the background live from the camera feed, overlay the user on a chosen image or video, and then create a new video combining these elements.

    


    Currently, I'm successfully obtaining the mask using the plugin. However, I'm concerned about performance issues if I directly copy pixels to achieve the overlay effect, especially in real-time scenarios.

    


    I'm looking for alternative approaches or best practices to efficiently achieve this functionality without sacrificing performance. Any insights or suggestions on how to approach this would be greatly appreciated.

    


    I thought about using ffmpeg but not sure how to take the user pixels from the camera feed, I am not a flutter developer so not sure how to do so.

    


  • google function : TypeError : Cannot read property 'name' of undefined

    9 décembre 2019, par Juggernaught

    I’m trying to use Google cloud function to transcode videos from a storage bucket and outputs into another bucket using code posted online but with a few adjustments.

    but I get the following error :

    TypeError : Cannot read property ’name’ of undefined at transcodeVideo
    (/srv/index.js:17:56) at /worker/worker.js:825:24 at at
    process._tickDomainCallback (internal/process/next_tick.js:229:7)

    Index.js

    const {Storage} = require('@google-cloud/storage');
    const projectId = 'cc18-223318';
    const storage = new Storage({
       projectId: projectId,
    });
    const ffmpegPath = require('@ffmpeg-installer/ffmpeg').path;
    const ffmpeg = require('fluent-ffmpeg');

    const transcodedBucket = storage.bucket('2400p');
    const uploadBucket = storage.bucket('inpuut');
    ffmpeg.setFfmpegPath(ffmpegPath);

    exports.transcodeVideo = function transcodeVideo(event, callback) {
     const file = event.data;
     // Ensure that you only proceed if the file is newly created, and exists.
     if (file.metageneration !== '1' || file.resourceState !== 'exists') {
       callback();
       return;
     }

     // Open write stream to new bucket, modify the filename as needed.
     const remoteWriteStream = transcodedBucket.file(file.name.replace('.webm', '.mp4'))
       .createWriteStream({
         metadata: {
           metadata: file.metadata, // You may not need this, my uploads have associated metadata
           contentType: 'video/mp4', // This could be whatever else you are transcoding to
         },
       });

     // Open read stream to our uploaded file
     const remoteReadStream = uploadBucket.file(file.name).createReadStream();

     // Transcode
     ffmpeg()
       .input(remoteReadStream)
       .outputOptions('-c:v libx264') // Change these options to whatever suits your needs
       .outputOptions('-c:a copy')
       .outputOptions('-vf "scale=4800:2400"')
       .outputOptions('-b:a 160k')
       .outputOptions('-b:a 160k')
       .outputOptions('-x264-params mvrange=511')
       .outputOptions('-f mp4')
       .outputOptions('-preset slow')
       .outputOptions('-movflags frag_keyframe+empty_moov')
       .outputOptions('-crf 18')
       // https://github.com/fluent-ffmpeg/node-fluent-ffmpeg/issues/346#issuecomment-67299526
       .on('start', (cmdLine) => {
         console.log('Started ffmpeg with command:', cmdLine);
       })
       .on('end', () => {
         console.log('Successfully re-encoded video.');
         callback();
       })
       .on('error', (err, stdout, stderr) => {
         console.error('An error occured during encoding', err.message);
         console.error('stdout:', stdout);
         console.error('stderr:', stderr);
         callback(err);
       })
       .pipe(remoteWriteStream, { end: true }); // end: true, emit end event when readable stream ends
    };