Recherche avancée

Médias (0)

Mot : - Tags -/masques

Aucun média correspondant à vos critères n’est disponible sur le site.

Autres articles (38)

  • Le plugin : Podcasts.

    14 juillet 2010, par

    Le problème du podcasting est à nouveau un problème révélateur de la normalisation des transports de données sur Internet.
    Deux formats intéressants existent : Celui développé par Apple, très axé sur l’utilisation d’iTunes dont la SPEC est ici ; Le format "Media RSS Module" qui est plus "libre" notamment soutenu par Yahoo et le logiciel Miro ;
    Types de fichiers supportés dans les flux
    Le format d’Apple n’autorise que les formats suivants dans ses flux : .mp3 audio/mpeg .m4a audio/x-m4a .mp4 (...)

  • MediaSPIP v0.2

    21 juin 2013, par

    MediaSPIP 0.2 is the first MediaSPIP stable release.
    Its official release date is June 21, 2013 and is announced here.
    The zip file provided here only contains the sources of MediaSPIP in its standalone version.
    To get a working installation, you must manually install all-software dependencies on the server.
    If you want to use this archive for an installation in "farm mode", you will also need to proceed to other manual (...)

  • Personnaliser en ajoutant son logo, sa bannière ou son image de fond

    5 septembre 2013, par

    Certains thèmes prennent en compte trois éléments de personnalisation : l’ajout d’un logo ; l’ajout d’une bannière l’ajout d’une image de fond ;

Sur d’autres sites (6407)

  • (IOS) Streaming raw h.264 video from and iphone to a RTSP

    10 octobre 2015, par Eric Jovenitti

    Ok so i’m new to video encoding and decoding and RTSP, but I need to take a third party component video stream i.e. (A quad copter drone) and output that stream to a RTSP. The stream is in raw h.264. The stream comes from the drones sdk i.e.,

    -(void) camera:(DJICamera*)camera didReceivedVideoData:(uint8_t*)videoBuffer length:(int)length{
    uint8_t* pBuffer = (uint8_t*)malloc(length);
    memcpy(pBuffer, videoBuffer, length);
    [[[VideoPreviewer instance] dataQueue] push:pBuffer length:length];

    }

    As you can see the raw video comes in from camera in the uint8_t.
    To my understanding this a decoded raw h.264 stream with Nalu’s an a IDR frame.

    So I want to stream this video to a RTSP so it can be watch on VLC .

    Does anyone know of tutorials or demo code to stream this video via a RTSP ?

    Additionally I found this demo : http://www.gdcl.co.uk/2013/02/20/iOS-Video-Encoding.html

    This sends the iphone camera’s video out to a RTSP and it work well.
    But i can’t figure out how to inject the drones streaming video into this demo RTSP app.

    This demo uses
    AVCaptureDevice to capture the cameras streaming data.

    So my questions from this demo are :
    Is there anyway to make my drones stream an AVCAptureDevice ?
    If not, can i just inject the drones raw data into the where the iphones camera data would be and just use the same properties that the iphome camera already configured ?

    if i can’t do those options can I just use the demos RTSP part ?
    If that’s possible, i can that the iphone camera data is being encoded. Does this mean i would need to encode the drones stream ? if so, how would i go about doing this ? The drones SKD uses FFmpeg to output the video so it can be seen in real time on the iphones screen.

    Thanks so much for the help !

  • Error using javacv library for live streaming camera to server

    22 juin 2015, par Prathyush Kumar

    I am new to live-streaming i’m following this link
    here for implementing camera streaming to server with ffserver listening on port 8090.But my code breaks at this line

    public void startRecording() {
       try {
           recorder.start();    //breaks here
           startTime = System.currentTimeMillis();
           recording = true;
           audioThread.start();
       } catch (FFmpegFrameRecorder.Exception e) {
           e.printStackTrace();
       }
    }

    I debugged a little bit to find out what is happening and i found out in this line in FFmpegFrameRecorder.class

    if((this.oformat.flags() & 1) == 0) {
       AVIOContext pb1 = new AVIOContext((Pointer)null);
       if((ret = avformat.avio_open(pb1, this.filename, 2)) < 0) {//throwing exception here
           this.release();
           throw new Exception("avio_open error() error " + ret + ": Could not open \'" + this.filename + "\'");
       }

       this.oc.pb(pb1);
    }

    ret is -5 which is less than 0 and so it cannot open the file(rtmp ://live:live@192.168.0.115:8090/live/test.flv).

    I am not able to understand whats happening in avformat.class Please help

    Is this any configuration problem for ffserver or is it anything else ?I’m not able to figure this out please any help would be greatly appreciated.

  • FFMPEG - Streaming GoPro4 to RSTP

    20 mai 2017, par Winged

    Greeting !
    I’m planning to capture a realtime streaming video from GoPro4 and load it into TouchDesigner in order to use this footage in a projection mapping project.

    So far I was able to play GoPro4 stream using ffplay doing as follows :
    1. Host GoPro4 WiFi hotspot.
    2. Connect a laptop WiFi to GoPro4’s hotspot.
    3. Restart GoPro stream service by opening ’http://10.5.5.9/gp/gpControl/execute?p1=gpStream&c1=restart’ in s browser.
    4. Run ffplay command in order to open ffplay stream video window : ’ffplay -an -fflags nobuffer -f:v mpegts -probesize 8192 rtp ://10.5.5.9:8554’

    The next step I opt to make it to re-stream this signal into RTSP protocol so TouchDesigner could receive this signal.

    And here is my question - how to re-stream this input to RTSP ?