Recherche avancée

Médias (1)

Mot : - Tags -/ticket

Autres articles (111)

  • Multilang : améliorer l’interface pour les blocs multilingues

    18 février 2011, par

    Multilang est un plugin supplémentaire qui n’est pas activé par défaut lors de l’initialisation de MediaSPIP.
    Après son activation, une préconfiguration est mise en place automatiquement par MediaSPIP init permettant à la nouvelle fonctionnalité d’être automatiquement opérationnelle. Il n’est donc pas obligatoire de passer par une étape de configuration pour cela.

  • Script d’installation automatique de MediaSPIP

    25 avril 2011, par

    Afin de palier aux difficultés d’installation dues principalement aux dépendances logicielles coté serveur, un script d’installation "tout en un" en bash a été créé afin de faciliter cette étape sur un serveur doté d’une distribution Linux compatible.
    Vous devez bénéficier d’un accès SSH à votre serveur et d’un compte "root" afin de l’utiliser, ce qui permettra d’installer les dépendances. Contactez votre hébergeur si vous ne disposez pas de cela.
    La documentation de l’utilisation du script d’installation (...)

  • Support de tous types de médias

    10 avril 2011

    Contrairement à beaucoup de logiciels et autres plate-formes modernes de partage de documents, MediaSPIP a l’ambition de gérer un maximum de formats de documents différents qu’ils soient de type : images (png, gif, jpg, bmp et autres...) ; audio (MP3, Ogg, Wav et autres...) ; vidéo (Avi, MP4, Ogv, mpg, mov, wmv et autres...) ; contenu textuel, code ou autres (open office, microsoft office (tableur, présentation), web (html, css), LaTeX, Google Earth) (...)

Sur d’autres sites (11307)

  • s is not recognized as an internal or external command in ffmpeg when streaming to Vk.com

    4 décembre 2018, par Anirudha Gupta

    I am trying to streaming using ffmpeg windows binary. It’s work on youtube and other site but not on vk.com

    ffmpeg -re -i file:"C:\Users\{uername}\Desktop\test.mp4" -acodec libmp3lame  -ar 44100 -b:a 128k -pix_fmt yuv420p -profile:v baseline -bufsize 6000k -framerate 24 -vb 64k -maxrate 1500k -deinterlace -vcodec libx264 -preset veryfast -g 30 -r 30 -crf 23 -c:a copy -f flv rtmp://stream2.vkuserlive.com:443/live?srv=629725&s=aWQ9QnE3dk44Uk5jdGcmc2lnbj1lR3pCWmdSdTdxTVpCTDdkMUVGR0lRPT0=/Bq7vN8RNctg

    It’s keep saying

    Conversion failed!
    's' is not recognized as an internal or external command,
    operable program or batch file.

    This happen on new version. If I test it on old version, it’s telling me that handshake failed with server. Anyone have idea on what is it, I am not removing my key and url so you can test. All you need is mp4 file in your pc and you can run this thing in your PC.

  • How to work with data received from streaming services in my Java application ?

    24 novembre 2020, par gabriel garcia

    I'm currently trying to develop an "streaming client" as a way to organize multiple stream services (twitch, yt, mitele...) in a single desktop application written in Java.

    


    It basically relies on streamlink (which relies in ffmpeg) thanks to all it's features so my project could be defined as a frontend for streamlink.

    


    Straight to the point, one of the features I'd like to add it is the option to programatically record streams in the background and showing this video stream to the user when it's requested. Since there's also the possibility that the user wants to watch the stream without recording it, I'm forced to work with all that byte-like data sent from those streaming sources.

    


    So, the problem is basically that I do not know much about video coding/decoding/muxing/demuxing nor video theory like container structure, video formats and such.

    


    But the idea is to work with all the data sent from the stream source (let's say twitch, for example), read this bytes (I'm not sure what kind of information is sent to the client nor format) from the java.lang.Process's stdout and then present it to the client.

    


    Here's another problem : I don't know how to play video streams in JavaFX and I don't think it's even supported right now. So I would have to extract each frame and sound associated from the stdout and show them to the user each time a new frame is received (oups, another problem since I don't know when does each frame starts/ends since I'm reading each stdout's line).

    


    As a summary :

    


      

    • What kind of data am I receiving from the streaming source ?
    • 


    • How can I know when does each frame starts/stops ?
    • 


    • How can I extract the image and sound from each frame ?
    • 


    


    I hope I'm not asking too much and that you could shed some light upon my darkness.

    


  • How to setup HLS Live Video Streaming from iOS Device

    9 septembre 2017, par Sean

    Good day everyone !

    So, as the title suggests, i am developing an app with similar functionality to that off Periscope and Facebook Live video streaming. Here is what the end goal is :

    • A Broadcasting device [user]
    • EC2 Instance [Hosting an ffmpeg transcoder]
    • Cloudfront Distrubution [CDN]
    • 1 to n viewers of the live feed

    I’ve been doing a lot of googling and what I cant seem to figure out is :

    As you send chunks of video to the server from the Broadcaster, how do
    you create an
    .m3u8 playlist when you don’t have all the chunks of video yet (e.g. the
    device sends its first 5second chunk of video) ?

    It seems a .m3u8 file is created from a .mp4 file that is already complete, then broken down into chunks... But i’m sending chunks of the video to the server, how can it generate the .m3u8 file when more chunks are still coming from the Broadcaster, so the watchers / clients can continuously stitch together the video chunks ?

    I’ll be happy to clarify this question further. Thanks !