Recherche avancée

Médias (39)

Mot : - Tags -/audio

Autres articles (35)

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

  • Encoding and processing into web-friendly formats

    13 avril 2011, par

    MediaSPIP automatically converts uploaded files to internet-compatible formats.
    Video files are encoded in MP4, Ogv and WebM (supported by HTML5) and MP4 (supported by Flash).
    Audio files are encoded in MP3 and Ogg (supported by HTML5) and MP3 (supported by Flash).
    Where possible, text is analyzed in order to retrieve the data needed for search engine detection, and then exported as a series of image files.
    All uploaded files are stored online in their original format, so you can (...)

  • Les formats acceptés

    28 janvier 2010, par

    Les commandes suivantes permettent d’avoir des informations sur les formats et codecs gérés par l’installation local de ffmpeg :
    ffmpeg -codecs ffmpeg -formats
    Les format videos acceptés en entrée
    Cette liste est non exhaustive, elle met en exergue les principaux formats utilisés : h264 : H.264 / AVC / MPEG-4 AVC / MPEG-4 part 10 m4v : raw MPEG-4 video format flv : Flash Video (FLV) / Sorenson Spark / Sorenson H.263 Theora wmv :
    Les formats vidéos de sortie possibles
    Dans un premier temps on (...)

Sur d’autres sites (6890)

  • Need help understanding this script which uses ffmpeg to send rtmp input to node.js script

    4 juin 2022, par Arpit Shukla

    I was trying to understand this shell script which uses ffmpeg to take an rtmp input stream and send it to a node.js script. But I am having trouble understanding the syntax. Can someone please explain what is going on here ?

    


    The script :

    


    while :
do
  echo "Loop start"

  feed_time=$(ffprobe -v error -show_entries format=start_time -of default=noprint_wrappers=1:nokey=1 $RTMP_INPUT)
  printf "feed_time value: ${feed_time}"

  if [ ! -z "${feed_time}" ]
  then
  ffmpeg -i $RTMP_INPUT -tune zerolatency -muxdelay 0 -af "afftdn=nf=-20, highpass=f=200, lowpass=f=3000" -vn -sn -dn -f wav -ar 16000 -ac 1 - 2>/dev/null | node src/transcribe.js $feed_time

  else
  echo "FFprobe returned null as a feed time."
  
  fi

  echo "Loop finish"
  sleep 3
done


    


      

    • What is feed_time here ? What does it represent ?
    • 


    • What is this portion doing - 2>/dev/null | node src/transcribe.js $feed_time ?
    • 


    • What is the use of sleep 3 ? Does this mean that we are sending audio stream to node.js in chuncks of 3 seconds ?
    • 


    


  • How to record (and process ?) a video that is streamable from Android

    13 mai 2016, par afollestad

    My company’s app relies heavily on video recording and playback of web-based videos. I use the MediaRecorder API to record videos, through this library designed by me : https://github.com/afollestad/material-camera.

    For playback, I use this library which is basically a wrapper around Google’s ExoPlayer library : https://github.com/brianwernick/ExoMedia.

    It works fine for the most part with small videos, especially if I decrease bit rates for audio and video. However, larger and higher quality videos have many issues. Sometimes they seem to buffer forever, sometimes playback doesn’t even start successfully, etc. Again, these videos are being streamed over HTTP from Amazon S3.


    I’ve read a little bit about FFMPEG, and how it can process MP4’s for "faststart", splitting the files into chunks for DASH, etc. However, FFMPEG solutions for Android seem a bit complex, so...

    Is there anyway to record MP4’s from Android, with MediaRecorder, MediaCodec, or some other API which results in a video file that is fast to stream ? It amazes me how well Snapchat has figured this out.

  • Carrierwave video not being processsed before uploading to S3

    12 décembre 2013, par Cramps

    I'm using Carrierwave, Carrierwave-video and Carrierwave-video-thumbnailer to process videos and make a thumbnail when they're uploaded. This was all working nicely while I was saving the files on my file system. However, now that I've added uploading to Amazon S3 using the carrierwave-aws gem, the videos are being uploaded to S3 without being processed first. It's as if the process encode_video and version :thumb are being skipped by the uploader.

    Here's what was working for me at first (before adding S3) :

    class VideoUploader < CarrierWave::Uploader::Base

    include CarrierWave::Video
    include CarrierWave::Video::Thumbnailer

    storage :file

    def store_dir
       "upload/path/"
    end

    process encode_video: [{ bunch of video options}]

    version :thumb do
       process thumbnail: [{ bunch of thumbnailer options }]
       def full_filename for_file
           png_name for_file, version_name
       end  
    end

    def png_name for_file, version_name
       %Q{#{version_name}_#{for_file.chomp(File.extname(for_file))}.png}
    end

    Now it's really just the same, except it's using storage :aws instead.