Advanced search

Medias (91)

Other articles (34)

  • MediaSPIP version 0.1 Beta

    16 April 2011, by

    MediaSPIP 0.1 beta est la première version de MediaSPIP décrétée comme "utilisable".
    Le fichier zip ici présent contient uniquement les sources de MediaSPIP en version standalone.
    Pour avoir une installation fonctionnelle, il est nécessaire d’installer manuellement l’ensemble des dépendances logicielles sur le serveur.
    Si vous souhaitez utiliser cette archive pour une installation en mode ferme, il vous faudra également procéder à d’autres modifications (...)

  • Librairies et binaires spécifiques au traitement vidéo et sonore

    31 January 2010, by

    Les logiciels et librairies suivantes sont utilisées par SPIPmotion d’une manière ou d’une autre.
    Binaires obligatoires FFMpeg : encodeur principal, permet de transcoder presque tous les types de fichiers vidéo et sonores dans les formats lisibles sur Internet. CF ce tutoriel pour son installation; Oggz-tools : outils d’inspection de fichiers ogg; Mediainfo : récupération d’informations depuis la plupart des formats vidéos et sonores;
    Binaires complémentaires et facultatifs flvtool2 : extraction / (...)

  • Support audio et vidéo HTML5

    10 April 2011

    MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
    Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
    Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
    Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...)

On other websites (4082)

  • iOS Objective C Opus audio stream kAudioFormatOpus — specification / conversion query

    16 April 2022, by deffodeffo

    I'm working on a React Native voice app wanting to record straight to an Opus stream.

    


    On the iOS side at device level, I'm working in Objective C and I'm using an AVAudioSession recorder with formatID set to kAudioFormatOpus. The recorder captures audio data in the specified format and passes packets for upstreaming to the React Native app.

    


    This is all working well and emitting a stream of audio data when I run my code on an iOS Simulator or real device.

    


    My problem is when I receive the audio stream in the backend, I find I'm unable to reliably decode the Opus stream using ffmpeg. (I wish to decode to PCM.)

    


    Whilst I could post up specific code to show what I'm doing in more detail, my question at this stage is more generic in nature:

    


    Is anyone familiar with the format of Opus audio stream that is generated by iOS AudioSession recorder under audio format kAudioFormatOpus ? Is anyone able to suggest proven conversion techniques? (e.g. some ffmpeg commands), or else hook me up with some links to format specs so I can figure out what is going on here ?

    


    The Apple Developer documentation contains zero useful information: https://developer.apple.com/documentation/coreaudiotypes/1572096-audio_data_format_identifiers/kaudioformatopus?changes=la__2&language=objc

    


    I have looked all over the internet, but I'm unable to find any useful spec info that corresponds to the stream that AVAudioSession recorder is outputting. I have seen a few posts which say Apple have not fully complied with the Opus spec, but I don't know enough about the proper structure of Opus to ascertain this for myself.

    


    Any help would be very much appreciated.

    


    Thanks all

    


  • Live streaming and simultaneous local/server video saving with Insta360/Theta 360 camera [closed]

    13 August 2023, by Fornow

    I'm currently working on a project that involves live streaming video from a 360 camera, specifically the Insta360 and Theta models, while also saving the streamed video either locally or on a remote server. I'm relatively new to both live streaming and working with 360 cameras, so I'm seeking guidance on the best approach to achieve this.

    


    My primary goals are as follows:

    


      

    1. Live Streaming: I want to be able to stream the real-time video captured by the 360 camera to a web platform or application, allowing users to experience the immersive 360 content as it happens.

      


    2. 


    3. Simultaneous Video Saving: In addition to live streaming, I also need to save the streamed video. This can either be saved locally on the device running the streaming process or on a remote server. The saved video should ideally retain its 360 nature and high-quality resolution.

      


    4. 


    


    I've been researching various technologies and frameworks like WebRTC for live streaming, but I'm unsure about the compatibility and best practices when dealing specifically with 360 cameras like Insta360 and Theta. Additionally, I'm uncertain about the most efficient way to save the streamed video while maintaining its immersive properties.

    


    If anyone has experience with live streaming from 360 cameras and simultaneously saving the content, could you please provide insights into the following:

    


      

    • Recommended libraries, SDKs, or frameworks for live streaming 360 video from Insta360 or Theta cameras.
    • 


    • Tips for ensuring the streamed video retains its 360 attributes and high quality.
    • 


    • Best practices for saving the streamed video either locally or on a remote server while the live stream is ongoing.
    • 


    


    Any code examples, tutorials, or step-by-step guides would be greatly appreciated. Thank you in advance for your help!

    


  • How to convert any video format to mp4 format using FFmpeg API

    27 April 2017, by krish

    Using FFmpeg API :ConvertLiveMediaTask

    ConvertLiveMedia(Stream inputStream, string inputFormat, Stream outputStream, string outputFormat, ConvertSettings settings);

    Here is my code, Requirement : want to convert video file from any format to mp4 format and conversion should happen on fly using ConvertLiveMedia API

    Please help me for this problem, I googled so many times i didnt get a solution for this.

    //Declarations part

    const int megabyte = 1024 * 1024;

           static string theFilename = @"F:\BackUp\Short video clip-nature.mp4";

           static void Main(string[] args)
           {
               ChunkedData(theFilename, 0);
           }

           private static void ChunkedData(string theFilename, long whereToStartReading = 0)
           {

               FileStream fileStram = new FileStream(theFilename, FileMode.Open, FileAccess.Read);
               using (fileStram)
               {
                   byte[] buffer = new byte[megabyte];
                   fileStram.Seek(whereToStartReading, SeekOrigin.Begin);
                   int bytesRead = fileStram.Read(buffer, 0, megabyte);
                   while (bytesRead > 0)
                   {
                       StreamData(buffer, bytesRead);
                       bytesRead = fileStram.Read(buffer, 0, megabyte);
                   }
               }
           }

           private static void StreamData(byte[] buffer, int bytesRead)
           {
               // Have no idea what to give for ConvertSettings, i simply pass a object
               ConvertSettings cs = new ConvertSettings();

               var ffmpeg = new FFMpegConverter();

               Stream inputStream = new MemoryStream(buffer);

               Stream outputStream = new MemoryStream();

               try
               {
                   // Here im getting outputStream capacity = 0, length = 0, position =0.
                   ffmpeg.ConvertLiveMedia(inputStream, Format.mp4, outputStream, Format.avi, cs);

               }
               catch (Exception ex)
               {
                   Console.WriteLine(ex.ToString());
               }


           }