Recherche avancée

Médias (1)

Mot : - Tags -/bug

Autres articles (18)

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

  • Configuration spécifique d’Apache

    4 février 2011, par

    Modules spécifiques
    Pour la configuration d’Apache, il est conseillé d’activer certains modules non spécifiques à MediaSPIP, mais permettant d’améliorer les performances : mod_deflate et mod_headers pour compresser automatiquement via Apache les pages. Cf ce tutoriel ; mode_expires pour gérer correctement l’expiration des hits. Cf ce tutoriel ;
    Il est également conseillé d’ajouter la prise en charge par apache du mime-type pour les fichiers WebM comme indiqué dans ce tutoriel.
    Création d’un (...)

  • Participer à sa documentation

    10 avril 2011

    La documentation est un des travaux les plus importants et les plus contraignants lors de la réalisation d’un outil technique.
    Tout apport extérieur à ce sujet est primordial : la critique de l’existant ; la participation à la rédaction d’articles orientés : utilisateur (administrateur de MediaSPIP ou simplement producteur de contenu) ; développeur ; la création de screencasts d’explication ; la traduction de la documentation dans une nouvelle langue ;
    Pour ce faire, vous pouvez vous inscrire sur (...)

Sur d’autres sites (5774)

  • Play Raspberry Pi h264 stream in C# app

    5 septembre 2019, par CoreMeltdown

    I have a Raspberry Pi board with dedicated camera that records video only in h264. I am looking for the best method to stream and play recorded video in real-time (as in, less than 1 sec delay) in c# windows forms app. The additional requirement is that such stream can be easily processed before displaying, for example for searching for objects on the image.

    Stuff I tried :

    - VLC server on raspi and VLC control in c# forms app <- simple solution, with RTSP, but has a serious flaw, which is a 3sec delay in image displayed. I couldn’t fix it with buffor size/options etc.

    - creating a socket on raspi with nc, receiving raw h264 data in c# and passing it to mplayer frontend <- If I simply start raspivid | nc and on the laptop nc | mplayer, i get exactly the results i want, the video i get is pretty much realtime, but the problem arises when i try to create mplayer frontend in c# and simulate the nc.exe. Maybe I’m passing the h264 data wrong (simply write them to stdin) or maybe something else.

    - using https://github.com/cisco/openh264 <- I compiled everything, but i can’t even get to decode sample vid.h264 i recorded on raspi with h264dec.exe, not to mention using it in c#.

    h264dec.exe vid.h264 out.yuv

    This produces 0bytes out.yuv file, while :

    h264dec.exe  vid.h264

    Gives me error message : "No input file specified in configuration file."

    - ffmpeg <- I implemented ffplay.exe playback in c# app but the lack of easy method to take screencaps etc. discouraged me to further investigate and develop.

    I’m not even sure whether I’m properly approaching the subject, so I’d be really thankful for every piece of advice I can get.

    EDIT
    Here is my ’working’ solution I am trying to implement in c#

    raspivid --width 400 --height 300 -t 9999999 --framerate 25 --output - | nc -l 5884

    nc ip_addr 5884 | mplayer -nosound -fps 100 -demuxer +h264es -cache 1024 -

    The key here is FPS 100, becuase then mplayer skips lag and plays the video it immediately receives with normal speed.
    The issue here is that I don’t know how to pass video data from socket into mplayer via c#, because I guess it is not done via stdin (already tried that).

  • Play Raspberry Pi h264 stream in C# app

    29 octobre 2016, par CoreMeltdown

    I have a Raspberry Pi board with dedicated camera that records video only in h264. I am looking for the best method to stream and play recorded video in real-time (as in, less than 1 sec delay) in c# windows forms app. The additional requirement is that such stream can be easily processed before displaying, for example for searching for objects on the image.

    Stuff I tried :

    - VLC server on raspi and VLC control in c# forms app <- simple solution, with RTSP, but has a serious flaw, which is a 3sec delay in image displayed. I couldn’t fix it with buffor size/options etc.

    - creating a socket on raspi with nc, receiving raw h264 data in c# and passing it to mplayer frontend <- If I simply start raspivid | nc and on the laptop nc | mplayer, i get exactly the results i want, the video i get is pretty much realtime, but the problem arises when i try to create mplayer frontend in c# and simulate the nc.exe. Maybe I’m passing the h264 data wrong (simply write them to stdin) or maybe something else.

    - using https://github.com/cisco/openh264 <- I compiled everything, but i can’t even get to decode sample vid.h264 i recorded on raspi with h264dec.exe, not to mention using it in c#.

    h264dec.exe vid.h264 out.yuv

    This produces 0bytes out.yuv file, while :

    h264dec.exe  vid.h264

    Gives me error message : "No input file specified in configuration file."

    - ffmpeg <- I implemented ffplay.exe playback in c# app but the lack of easy method to take screencaps etc. discouraged me to further investigate and develop.

    I’m not even sure whether I’m properly approaching the subject, so I’d be really thankful for every piece of advice I can get.

    EDIT
    Here is my ’working’ solution I am trying to implement in c#

    raspivid --width 400 --height 300 -t 9999999 --framerate 25 --output - | nc -l 5884

    nc ip_addr 5884 | mplayer -nosound -fps 100 -demuxer +h264es -cache 1024 -

    The key here is FPS 100, becuase then mplayer skips lag and plays the video it immediately receives with normal speed.
    The issue here is that I don’t know how to pass video data from socket into mplayer via c#, because I guess it is not done via stdin (already tried that).

  • Fallback input for ffmpeg

    22 septembre 2018, par Daniel Cantarin

    I’m doing some transcoding from a third-party remote input stream that I do not control.

    This input stream has errors from time to time, that I would like to mitigate before sending the stream to my transcoding pipeline, avoiding this way some possible problems in the output.

    I have several ideas regarding different problems. But the most basic scenario I would like to set up is as follows : when the stream is down, or it somehow loses some frames, I want to fill that video gap with a secondary input (like a blank screen, for example).

    For this simple task, I would like to use ffmpeg. I know it can mix, let’s say, an input stream with a fullscreen black square static image. However, I have to deal with this other condition : ffmpeg would run in the same infraestructure for the actual transcoding pipeline. That infraestructure must use its computing power for rendering the output. So, whatever ffmpeg command I end up using should use the minimum possible computing power.

    My actual problem : if I use -vcodec copy, in order to use minimum CPU, I can’t alter the original stream. But if I alter the original stream (by mixing it with some other stream), the operation uses CPU.

    My question : Is there a way to use -vcodec copy, but with a fallback input (instead of a mixed one) for when there are video gaps in the primary stream ?

    Thanks in advance.