Recherche avancée

Médias (91)

Autres articles (37)

  • HTML5 audio and video support

    13 avril 2011, par

    MediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
    The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
    For older browsers the Flowplayer flash fallback is used.
    MediaSPIP allows for media playback on major mobile platforms with the above (...)

  • De l’upload à la vidéo finale [version standalone]

    31 janvier 2010, par

    Le chemin d’un document audio ou vidéo dans SPIPMotion est divisé en trois étapes distinctes.
    Upload et récupération d’informations de la vidéo source
    Dans un premier temps, il est nécessaire de créer un article SPIP et de lui joindre le document vidéo "source".
    Au moment où ce document est joint à l’article, deux actions supplémentaires au comportement normal sont exécutées : La récupération des informations techniques des flux audio et video du fichier ; La génération d’une vignette : extraction d’une (...)

  • Support audio et vidéo HTML5

    10 avril 2011

    MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
    Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
    Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
    Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...)

Sur d’autres sites (5999)

  • Is there any library in python or can we use FFMPEG to detect the appearance of the lines pattern in video that is shown in the image attached below ? [closed]

    21 septembre 2022, par Shreyas R

    Video defect appearing frame

    


    Hi All,

    


    I am trying to find a way to detect the appearance of this lines pattern embedded in the link below.
If I get any match of this pattern in a video, I need to record that particular timestamp of the video where I got the pattern match.

    


    My idea is to take the image and compare it against the video to find for approximate match percentage. If I get it, then I record that particular timestamp within a .txt file.

    


    Is there any better way to achieve this ?
I think we can use Python or FFMPEG video filters to get a solution for this.

    


    It would really help if anyone can support with a solution. Thanks

    


    Update :
I got a reference where they have proposed a solution similar to my concern.

    


    https://aws.amazon.com/blogs/media/metfc-automatically-compare-two-videos-to-find-common-content/

    


    So we can try to achieve some output from this solution.

    


  • ffmpeg udp/tcp stream receive frame not same as sent

    21 novembre 2013, par vivienlwt

    I am streaming a video on raspberrypi using command :

    ffmpeg -re -threads 2 -i sample_video.m2v -f mpegts - | \
    ffmpeg -f mpegts -i - -c copy -f mpegts udp://192.168.1.100:12345

    The remote PC with 192.168.1.100 uses ffmpeg library to listen to the input stream. For example :

    informat = ffmpeg::av_find_input_format("mpegts");
    avformat_open_input(&pFormatCtx, "udp://192.168.1.100:12345", informat, options);

    However, when I compute the hash value of each decoded frame on two sides (i.e. raspberrypi and PC), they DON'T MATCH at all. A weird thing is, among 2000 frames, there are in total 10 frames whose hash value are the same on the sender and receiver side. The match result look like this :

    00000....00011000...00011110000...000

    where 0 indicates non-match and 1 indicates match. The matched frame appeared 2 6 in sequence and appeared rarely while most of the other frames has different hash value.

    The hash is computed on the frame data buffer extracted using avpicture_layout(). On the Pi side, I just stream the video to a local port and there's a local process using the same code to decode and hash the frames :

    ffmpeg -re -threads 2 -i sample_video.m2v -f mpegts - | \
    ffmpeg -f mpegts -i - -c copy -f mpegts udp://localhost:12345
    ...

    The streaming source raspberry pi, is connected directly to the PC using cable. I don't think it is a packet loss problem. Because, first, I rerun the same process several times and the hash value of the received frames are the same (otherwise the result should be different because packet loss is probabilistic). Secondly, I even try to stream on tcp ://192.168.1.100:12345 (and "tcp ://192.168.1.100:12345 ?listen" on PC), and the received frame hash are still the same - different than the hash result on the Pi.

    So, does anyone know why the streaming to a remote address will yield different decoded frames ? Maybe I am missing some details.

    Thanks in advance !!

  • Frame-by-frame video decoding and processing on Android

    22 octobre 2016, par Jason M

    I am working on a project to decode a recorded video on Android into yuv420sp frames with API17 for processing. I have implemented it on Win and iOS with ffmpeg and native api for decoding, respectively. Now I want to reuse the same c++ processing code on Android API17, but found difficulty. I saw straight API to do that in API21, but sadly I have to keep that better compatibility.

    At first, I tried to make the best use of APIs of hardware decoding (following MediaCodec get all frames from video and some others) but found that the data in the outputbuffer does not match that described by the format. For example, I have 1920x1080 frames decoded into yuv420sp but found that y channel actually span 1928x1080. Since I had similar headaches working with still image capturing that was never worked out(YUV image taken from takePicture not match defined format), I am not sure if I may ever solve this.

    Later, I tried to go back to ffmpeg but found most building guides either out of date(for ffmpeg 2.x) or a little hard to understand(such as http://writingminds.github.io/ffmpeg-android/, where they do provide nice pre-built binaries but no instructions on using it. Maybe I should download the sources and use the headers there ?).

    Any suggestions ?