Recherche avancée

Médias (91)

Autres articles (66)

  • Installation en mode ferme

    4 février 2011, par

    Le mode ferme permet d’héberger plusieurs sites de type MediaSPIP en n’installant qu’une seule fois son noyau fonctionnel.
    C’est la méthode que nous utilisons sur cette même plateforme.
    L’utilisation en mode ferme nécessite de connaïtre un peu le mécanisme de SPIP contrairement à la version standalone qui ne nécessite pas réellement de connaissances spécifique puisque l’espace privé habituel de SPIP n’est plus utilisé.
    Dans un premier temps, vous devez avoir installé les mêmes fichiers que l’installation (...)

  • Multilang : améliorer l’interface pour les blocs multilingues

    18 février 2011, par

    Multilang est un plugin supplémentaire qui n’est pas activé par défaut lors de l’initialisation de MediaSPIP.
    Après son activation, une préconfiguration est mise en place automatiquement par MediaSPIP init permettant à la nouvelle fonctionnalité d’être automatiquement opérationnelle. Il n’est donc pas obligatoire de passer par une étape de configuration pour cela.

  • List of compatible distributions

    26 avril 2011, par

    The table below is the list of Linux distributions compatible with the automated installation script of MediaSPIP. Distribution nameVersion nameVersion number Debian Squeeze 6.x.x Debian Weezy 7.x.x Debian Jessie 8.x.x Ubuntu The Precise Pangolin 12.04 LTS Ubuntu The Trusty Tahr 14.04
    If you want to help us improve this list, you can provide us access to a machine whose distribution is not mentioned above or send the necessary fixes to add (...)

Sur d’autres sites (11086)

  • google speech to text errors out (grpc invalid deadline NaN)

    15 décembre 2019, par jamescharlesworth

    I have a ffmpeg script that cuts an audio file into a short 5 second clip, however after I cut the file, calling the google speech recognize command errors out.

    Creating a clip - full code link :

    const uri = 'http://traffic.libsyn.com/joeroganexp/p1400.mp3?dest-id=19997';
    const command = ffmpeg(got.stream(uri));
    command
     .seek(0)
     .duration(5)
     .audioBitrate(128)
     .format('mp3')
    ...

    which works fine and creates ./clip2.mp3.

    I then take that file and upload it to speech to text api and it times out (script here. When I put timeout and maxRetries argument I can get the actual error :

    Error: 2 UNKNOWN: Getting metadata from plugin failed with error: Deadline is too far in the future
       at Object.callErrorFromStatus (/Users/jamescharlesworth/Downloads/demo/node_modules/@grpc/grpc-js/build/src/call.js:30:26)
       at Http2CallStream.<anonymous> (/Users/jamescharlesworth/Downloads/demo/node_modules/@grpc/grpc-js/build/src/client.js:96:33)
       at Http2CallStream.emit (events.js:215:7)
       at /Users/jamescharlesworth/Downloads/demo/node_modules/@grpc/grpc-js/build/src/call-stream.js:98:22
       at processTicksAndRejections (internal/process/task_queues.js:75:11) {
     code: 2,
     details: 'Getting metadata from plugin failed with error: Deadline is too far in the future',
     metadata: Metadata { internalRepr: Map {}, options: {} },
     note: 'Exception occurred in retry method that was not classified as transient'
    }
    </anonymous>

    Stepping through the grpc code i see that the deadline is an invalid date.
    enter image description here
    This seems to be causing the issue but i assume it may be from incorrect params passed into the speech client.recognize() method.

    A few other things to note :

    • The script works for some audio files, not all
    • I can upload the broken my clip mp3 clip2.mp3 to the demo app here and it works fine.
    • If I change the seek command of my ffmpeg script to start at 0.01 speech recognize command will work (however it breaks other audio clips as its not the correct starting point). I notice that when i do this the png of the mp3 gets stripped out and is a much smaller file size
  • How to Add PulseAudio Server to quay.io/browser/google-chrome-stable Docker Image for Audio Support with Screen Recording ?

    17 avril, par Ahmed Seddik Bouchiba

    I’m trying to set up an environment for recording the screen of a Chrome browser running in a Docker container, and I need to enable audio support. I’m using the quay.io/browser/google-chrome-stable:133.0.6943.98-6 image for the browser and quay.io/aerokube/xvfb:21.1 for the virtual framebuffer to capture the screen.

    &#xA;

    However, I’m facing an issue where audio is not supported in the Chrome Docker image, which I need for recording. The setup involves using FFmpeg in a separate container to stream the recorded video, but without audio from the browser, this setup isn’t complete.

    &#xA;

    I’m looking for guidance on how to add a PulseAudio server to the Chrome image to enable audio support. Specifically :

    &#xA;

    How can I configure the Docker image quay.io/browser/google-chrome-stable:133.0.6943.98-6 to support PulseAudio?&#xA;&#xA;Are there any considerations or best practices when adding PulseAudio to a headless browser Docker container?&#xA;&#xA;Is it possible to run the PulseAudio server in a separate container and link it to the Chrome container, or should it be included directly in the Chrome container?&#xA;

    &#xA;

    Any help on adding PulseAudio support to this Chrome Docker image would be greatly appreciated !

    &#xA;

    Additional Context :

    &#xA;

    The goal is to run a headless Chrome browser with audio support to record the browser’s activities (both video and audio) and stream it using FFmpeg.&#xA;&#xA;I’m using Docker Compose to orchestrate the containers but haven’t figured out how to integrate PulseAudio into the setup effectively.&#xA;

    &#xA;

    Thanks in advance !

    &#xA;

  • avformat/hlsenc : Handle NULL input in IO open and close utility functions

    15 décembre 2017, par Karthick J
    avformat/hlsenc : Handle NULL input in IO open and close utility functions
    

    Signed-off-by : Steven Liu <lq@chinaffmpeg.org>

    • [DH] libavformat/hlsenc.c