Recherche avancée

Médias (91)

Autres articles (62)

  • Participer à sa traduction

    10 avril 2011

    Vous pouvez nous aider à améliorer les locutions utilisées dans le logiciel ou à traduire celui-ci dans n’importe qu’elle nouvelle langue permettant sa diffusion à de nouvelles communautés linguistiques.
    Pour ce faire, on utilise l’interface de traduction de SPIP où l’ensemble des modules de langue de MediaSPIP sont à disposition. ll vous suffit de vous inscrire sur la liste de discussion des traducteurs pour demander plus d’informations.
    Actuellement MediaSPIP n’est disponible qu’en français et (...)

  • Les autorisations surchargées par les plugins

    27 avril 2010, par

    Mediaspip core
    autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs

  • L’agrémenter visuellement

    10 avril 2011

    MediaSPIP est basé sur un système de thèmes et de squelettes. Les squelettes définissent le placement des informations dans la page, définissant un usage spécifique de la plateforme, et les thèmes l’habillage graphique général.
    Chacun peut proposer un nouveau thème graphique ou un squelette et le mettre à disposition de la communauté.

Sur d’autres sites (13318)

  • Video Conferencing in HTML5 : WebRTC via Web Sockets

    1er janvier 2014, par silvia

    A bit over a week ago I gave a presentation at Web Directions Code 2012 in Melbourne. Maxine and John asked me to speak about something related to HTML5 video, so I went for the new shiny : WebRTC – real-time communication in the browser.

    Presentation slides

    I only had 20 min, so I had to make it tight. I wanted to show off video conferencing without special plugins in Google Chrome in just a few lines of code, as is the promise of WebRTC. To a large extent, I achieved this. But I made some interesting discoveries along the way. Demos are in the slide deck.

    UPDATE : Opera 12 has been released with WebRTC support.

    Housekeeping : if you want to replicate what I have done, you need to install a Google Chrome Web Browser 19+. Then make sure you go to chrome ://flags and activate the MediaStream and PeerConnection experiment(s). Restart your browser and now you can experiment with this feature. Big warning up-front : it’s not production-ready, since there are still changes happening to the spec and there is no compatible implementation by another browser yet.

    Here is a brief summary of the steps involved to set up video conferencing in your browser :

    1. Set up a video element each for the local and the remote video stream.
    2. Grab the local camera and stream it to the first video element.
    3. (*) Establish a connection to another person running the same Web page.
    4. Send the local camera stream on that peer connection.
    5. Accept the remote camera stream into the second video element.

    Now, the most difficult part of all of this – believe it or not – is the signalling part that is required to build the peer connection (marked with (*)). Initially I wanted to run completely without a server and just enter the remote’s IP address to establish the connection. This is, however, not a functionality that the PeerConnection object provides [might this be something to add to the spec ?].

    So, you need a server known to both parties that can provide for the handshake to set up the connection. All the examples that I have seen, such as https://apprtc.appspot.com/, use a channel management server on Google’s appengine. I wanted it all working with HTML5 technology, so I decided to use a Web Socket server instead.

    I implemented my Web Socket server using node.js (code of websocket server). The video conferencing demo is in the slide deck in an iframe – you can also use the stand-alone html page. Works like a treat.

    While it is still using Google’s STUN server to get through NAT, the messaging for setting up the connection is running completely through the Web Socket server. The messages that get exchanged are plain SDP message packets with a session ID. There are OFFER, ANSWER, and OK packets exchanged for each streaming direction. You can see some of it in the below image :

    WebRTC demo

    I’m not running a public WebSocket server, so you won’t be able to see this part of the presentation working. But the local loopback video should work.

    At the conference, it all went without a hitch (while the wireless played along). I believe you have to host the WebSocket server on the same machine as the Web page, otherwise it won’t work for security reasons.

    A whole new world of opportunities lies out there when we get the ability to set up video conferencing on every Web page – scary and exciting at the same time !

  • Revert "mpegts : do not set pts for missing dts in video streams"

    2 mars 2014, par Michael Niedermayer
    Revert "mpegts : do not set pts for missing dts in video streams"
    

    This reverts commit f65afef1df49f53e14c8d4173ff960fff8d44ecb.

    If only pts is coded then dts must be equal pts
    See Rec. ITU-T H.222.0 (06/2012) / ISO/IEC 13818-1:2013 (E) 2.7.5

    Found-by : Baptiste Coudurier

    • [DH] libavformat/mpegts.c
  • FFMpeg integration with iOS 7

    2 janvier 2014, par metsburg

    I am trying to integrate FFMpeg with iOS 7.

    My objective is to decode a video stream which is coming in through H.264 video codec.

    I've built the FFMpeg library using methodology described here : Trying to compile the FFMPEG libraries for iPhoneOS platform with armv6 and arv7 architecture

    When I compile using :

    ./configure
    --prefix=armv7
    --disable-ffmpeg
    --disable-ffplay
    --disable-ffprobe
    --disable-ffserver
    --enable-avresample
    --enable-cross-compile
    --sysroot="/Applications/Xcode.app/Contents/Developer/Platforms/iPhoneOS.platform/Developer/SDKs/iPhoneOS7.0.sdk"
    --target-os=darwin
    --cc="/usr/llvm-gcc-4.2/bin/llvm-gcc-4.2"
    --extra-cflags="-arch armv7 -mfpu=neon -miphoneos-version-min=6.1"
    --extra-ldflags="-arch armv7 -isysroot /Applications/Xcode.app/Contents/Developer/Platforms/iPhoneOS.platform/Developer/SDKs/iPhoneOS7.0.sdk -miphoneos-version-min=7.0"
    --arch=arm --cpu=cortex-a9 --enable-pic

    all the *.a files do not get generated.

    I'm trying to create a sample FFMpeg project, as described in http://sol3.typepad.com/exotic_particles/2012/10/building-a-simple-frame-recorder-part-1.html?cid=6a00d8343a141353ef019b01f49c8e970d#comment-6a00d8343a141353ef019b01f49c8e970d

    However, I keep getting the following error while trying to build the project :

    Undefined symbols for architecture i386:

     "_av_close_input_file", referenced from:

         -[FrameExtractor dealloc] in FFmpegDecoder(FrameExtractor.o)

     "_av_find_stream_info", referenced from:

         -[FrameExtractor initWithVideo:] in FFmpegDecoder(FrameExtractor.o)

     "_av_free", referenced from:

         -[FrameExtractor dealloc] in FFmpegDecoder(FrameExtractor.o)

     "_av_free_packet", referenced from:

         -[FrameExtractor stepFrame] in FFmpegDecoder(FrameExtractor.o)

     "_av_open_input_file", referenced from:

         -[FrameExtractor initWithVideo:] in FFmpegDecoder(FrameExtractor.o)

     "_av_read_frame", referenced from:

         -[FrameExtractor stepFrame] in FFmpegDecoder(FrameExtractor.o)

     "_av_register_all", referenced from:

         -[FrameExtractor initWithVideo:] in FFmpegDecoder(FrameExtractor.o)

     "_avcodec_alloc_frame", referenced from:

         -[FrameExtractor initWithVideo:] in FFmpegDecoder(FrameExtractor.o)

     "_avcodec_close", referenced from:

         -[FrameExtractor dealloc] in FFmpegDecoder(FrameExtractor.o)

     "_avcodec_decode_video2", referenced from:

         -[FrameExtractor stepFrame] in FFmpegDecoder(FrameExtractor.o)

     "_avcodec_find_decoder", referenced from:

         -[FrameExtractor initWithVideo:] in FFmpegDecoder(FrameExtractor.o)

     "_avcodec_flush_buffers", referenced from:

         -[FrameExtractor seekTime:] in FFmpegDecoder(FrameExtractor.o)

     "_avcodec_open", referenced from:

         -[FrameExtractor initWithVideo:] in FFmpegDecoder(FrameExtractor.o)

     "_avcodec_register_all", referenced from:

         -[FrameExtractor initWithVideo:] in FFmpegDecoder(FrameExtractor.o)

     "_avformat_seek_file", referenced from:

         -[FrameExtractor seekTime:] in FFmpegDecoder(FrameExtractor.o)

     "_avpicture_alloc", referenced from:

         -[FrameExtractor setupScaler] in FFmpegDecoder(FrameExtractor.o)

     "_avpicture_free", referenced from:

         -[FrameExtractor setupScaler] in FFmpegDecoder(FrameExtractor.o)

         -[FrameExtractor dealloc] in FFmpegDecoder(FrameExtractor.o)

     "_sws_freeContext", referenced from:

         -[FrameExtractor setupScaler] in FFmpegDecoder(FrameExtractor.o)

         -[FrameExtractor dealloc] in FFmpegDecoder(FrameExtractor.o)

     "_sws_getContext", referenced from:

         -[FrameExtractor setupScaler] in FFmpegDecoder(FrameExtractor.o)

     "_sws_scale", referenced from:

         -[FrameExtractor convertFrameToRGB] in FFmpegDecoder(FrameExtractor.o)

    ld: symbol(s) not found for architecture i386

    clang: error: linker command failed with exit code 1 (use -v to see invocation)

    Please help.