Recherche avancée

Médias (2)

Mot : - Tags -/plugins

Autres articles (96)

  • MediaSPIP 0.1 Beta version

    25 avril 2011, par

    MediaSPIP 0.1 beta is the first version of MediaSPIP proclaimed as "usable".
    The zip file provided here only contains the sources of MediaSPIP in its standalone version.
    To get a working installation, you must manually install all-software dependencies on the server.
    If you want to use this archive for an installation in "farm mode", you will also need to proceed to other manual (...)

  • HTML5 audio and video support

    13 avril 2011, par

    MediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
    The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
    For older browsers the Flowplayer flash fallback is used.
    MediaSPIP allows for media playback on major mobile platforms with the above (...)

  • ANNEXE : Les plugins utilisés spécifiquement pour la ferme

    5 mars 2010, par

    Le site central/maître de la ferme a besoin d’utiliser plusieurs plugins supplémentaires vis à vis des canaux pour son bon fonctionnement. le plugin Gestion de la mutualisation ; le plugin inscription3 pour gérer les inscriptions et les demandes de création d’instance de mutualisation dès l’inscription des utilisateurs ; le plugin verifier qui fournit une API de vérification des champs (utilisé par inscription3) ; le plugin champs extras v2 nécessité par inscription3 (...)

Sur d’autres sites (8898)

  • Merging input Streams with nodejs/ffmpeg

    14 septembre 2020, par jAndy

    I'm creating a very basic and rudimentary Video-Web-Chat. On the client side, I'm going to use a simple getUserMedia API call to capture the webcam data and send video-data as data-blob to my server.

    


    From There, I'm planning to either use the fluent-ffmpeg library or just spawn ffmpeg myself and pipe that raw data to ffmpeg, which in turn, does some magic and pushes that out as HLS stream to an Amazon AWS Service (for instance), which then gets actually displayed on a Web Browser for all participating people in the video chat.

    


    So far, I think all of this should be fairly easy to implement, but I keep my head spinning around the question, how I can create a "combined" or "merged" frame and stream, so the output HLS data from my server to the distributing cloud service has only to be one combined data stream to receive.

    


    If there are 3 people in that video chat, my server receives 3 data streams from those clients and combines these data streams (from the individual web-cam data sources) into one output stream.

    


    How could that be accomplished ?
Can I "create" a new frame with ffmpeg, so to speak ? I would be very thankful if anybody could give me a heads up here, maybe I'm thinking in a complete wrong direction.

    


    Another question which arises to me is, if I really can just "dump" any data, which I'm receiving from a binary blob created from getUserMedia or MultiStreamRecorder to ffmpeg or if I have to specify somewhere and somehow the exact codecs being used etc.?

    


  • swscale/aarch64 : add vscale specializations

    13 août 2022, par Swinney, Jonathan
    swscale/aarch64 : add vscale specializations
    

    This commit adds new code paths for vscale when filterSize is 2, 4, or
    8. By using specialized code with unrolling to match the filterSize we
    can improve performance.

    On AWS c7g (Graviton 3, Neoverse V1) instances :
    before after
    yuv2yuvX_2_0_512_accurate_neon : 558.8 268.9
    yuv2yuvX_4_0_512_accurate_neon : 637.5 434.9
    yuv2yuvX_8_0_512_accurate_neon : 1144.8 806.2
    yuv2yuvX_16_0_512_accurate_neon : 2080.5 1853.7

    Signed-off-by : Jonathan Swinney <jswinney@amazon.com>
    Signed-off-by : Martin Storsjö <martin@martin.st>

    • [DH] libswscale/aarch64/output.S
    • [DH] libswscale/aarch64/swscale.c
  • Error : ffmpeg exited with code 1 on AWS Lambda

    16 juin 2022, par Hassnain Alvi

    I am using fluent-ffmpeg nodejs package to run ffmpeg for audio conversion on AWS Lambda. I am using this FFmpeg layer for lambda.&#xA;Here is my code

    &#xA;

      const bitrate64 = ffmpeg("file.mp3").audioBitrate(&#x27;64k&#x27;);&#xA;    bitrate64.outputOptions([&#xA;        &#x27;-preset slow&#x27;,&#xA;    &#x27;-g 48&#x27;,&#xA;    "-map", "0:0",&#xA;        &#x27;-hls_time 6&#x27;,&#xA;        &#x27;-master_pl_name master.m3u8&#x27;,&#xA;        &#x27;-hls_segment_filename 64k/fileSequence%d.ts&#x27;&#xA;    ])&#xA;    .output(&#x27;./64k/prog_index.m3u8&#x27;)&#xA;    .on(&#x27;progress&#x27;, function(progress) {&#xA;        console.log(&#x27;Processing 64k bitrate: &#x27; &#x2B; progress.percent &#x2B; &#x27;% done&#x27;)&#xA;    })  &#xA;    .on(&#x27;end&#x27;, function(err, stdout, stderr) {&#xA;        console.log(&#x27;Finished processing 64k bitrate!&#x27;)&#xA;    })&#xA;    .run() &#xA;

    &#xA;

    after running it via AWS lambda I get following error message

    &#xA;

    ERROR   Uncaught Exception  &#xA;{&#xA;    "errorType": "Error",&#xA;    "errorMessage": "ffmpeg exited with code 1: Conversion failed!\n",&#xA;    "stack": [&#xA;        "Error: ffmpeg exited with code 1: Conversion failed!",&#xA;        "",&#xA;        "    at ChildProcess.<anonymous> (/var/task/node_modules/fluent-ffmpeg/lib/processor.js:182:22)",&#xA;        "    at ChildProcess.emit (events.js:198:13)",&#xA;        "    at ChildProcess.EventEmitter.emit (domain.js:448:20)",&#xA;        "    at Process.ChildProcess._handle.onexit (internal/child_process.js:248:12)"&#xA;    ]&#xA;}&#xA;</anonymous>

    &#xA;

    I don't get any more info so I am not sure what's going on. Can anyone tell me what's wrong here and how can I enable more detailed logs ?

    &#xA;