Recherche avancée

Médias (3)

Mot : - Tags -/spip

Autres articles (79)

  • Des sites réalisés avec MediaSPIP

    2 mai 2011, par

    Cette page présente quelques-uns des sites fonctionnant sous MediaSPIP.
    Vous pouvez bien entendu ajouter le votre grâce au formulaire en bas de page.

  • HTML5 audio and video support

    13 avril 2011, par

    MediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
    The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
    For older browsers the Flowplayer flash fallback is used.
    MediaSPIP allows for media playback on major mobile platforms with the above (...)

  • De l’upload à la vidéo finale [version standalone]

    31 janvier 2010, par

    Le chemin d’un document audio ou vidéo dans SPIPMotion est divisé en trois étapes distinctes.
    Upload et récupération d’informations de la vidéo source
    Dans un premier temps, il est nécessaire de créer un article SPIP et de lui joindre le document vidéo "source".
    Au moment où ce document est joint à l’article, deux actions supplémentaires au comportement normal sont exécutées : La récupération des informations techniques des flux audio et video du fichier ; La génération d’une vignette : extraction d’une (...)

Sur d’autres sites (8822)

  • FFMPEG - Streaming GoPro4 to RSTP

    20 mai 2017, par Winged

    Greeting !
    I’m planning to capture a realtime streaming video from GoPro4 and load it into TouchDesigner in order to use this footage in a projection mapping project.

    So far I was able to play GoPro4 stream using ffplay doing as follows :
    1. Host GoPro4 WiFi hotspot.
    2. Connect a laptop WiFi to GoPro4’s hotspot.
    3. Restart GoPro stream service by opening ’http://10.5.5.9/gp/gpControl/execute?p1=gpStream&c1=restart’ in s browser.
    4. Run ffplay command in order to open ffplay stream video window : ’ffplay -an -fflags nobuffer -f:v mpegts -probesize 8192 rtp ://10.5.5.9:8554’

    The next step I opt to make it to re-stream this signal into RTSP protocol so TouchDesigner could receive this signal.

    And here is my question - how to re-stream this input to RTSP ?

  • Error using javacv library for live streaming camera to server

    22 juin 2015, par Prathyush Kumar

    I am new to live-streaming i’m following this link
    here for implementing camera streaming to server with ffserver listening on port 8090.But my code breaks at this line

    public void startRecording() {
       try {
           recorder.start();    //breaks here
           startTime = System.currentTimeMillis();
           recording = true;
           audioThread.start();
       } catch (FFmpegFrameRecorder.Exception e) {
           e.printStackTrace();
       }
    }

    I debugged a little bit to find out what is happening and i found out in this line in FFmpegFrameRecorder.class

    if((this.oformat.flags() & 1) == 0) {
       AVIOContext pb1 = new AVIOContext((Pointer)null);
       if((ret = avformat.avio_open(pb1, this.filename, 2)) < 0) {//throwing exception here
           this.release();
           throw new Exception("avio_open error() error " + ret + ": Could not open \'" + this.filename + "\'");
       }

       this.oc.pb(pb1);
    }

    ret is -5 which is less than 0 and so it cannot open the file(rtmp ://live:live@192.168.0.115:8090/live/test.flv).

    I am not able to understand whats happening in avformat.class Please help

    Is this any configuration problem for ffserver or is it anything else ?I’m not able to figure this out please any help would be greatly appreciated.

  • (IOS) Streaming raw h.264 video from and iphone to a RTSP

    10 octobre 2015, par Eric Jovenitti

    Ok so i’m new to video encoding and decoding and RTSP, but I need to take a third party component video stream i.e. (A quad copter drone) and output that stream to a RTSP. The stream is in raw h.264. The stream comes from the drones sdk i.e.,

    -(void) camera:(DJICamera*)camera didReceivedVideoData:(uint8_t*)videoBuffer length:(int)length{
    uint8_t* pBuffer = (uint8_t*)malloc(length);
    memcpy(pBuffer, videoBuffer, length);
    [[[VideoPreviewer instance] dataQueue] push:pBuffer length:length];

    }

    As you can see the raw video comes in from camera in the uint8_t.
    To my understanding this a decoded raw h.264 stream with Nalu’s an a IDR frame.

    So I want to stream this video to a RTSP so it can be watch on VLC .

    Does anyone know of tutorials or demo code to stream this video via a RTSP ?

    Additionally I found this demo : http://www.gdcl.co.uk/2013/02/20/iOS-Video-Encoding.html

    This sends the iphone camera’s video out to a RTSP and it work well.
    But i can’t figure out how to inject the drones streaming video into this demo RTSP app.

    This demo uses
    AVCaptureDevice to capture the cameras streaming data.

    So my questions from this demo are :
    Is there anyway to make my drones stream an AVCAptureDevice ?
    If not, can i just inject the drones raw data into the where the iphones camera data would be and just use the same properties that the iphome camera already configured ?

    if i can’t do those options can I just use the demos RTSP part ?
    If that’s possible, i can that the iphone camera data is being encoded. Does this mean i would need to encode the drones stream ? if so, how would i go about doing this ? The drones SKD uses FFmpeg to output the video so it can be seen in real time on the iphones screen.

    Thanks so much for the help !