Recherche avancée

Médias (91)

Autres articles (106)

  • Soumettre bugs et patchs

    10 avril 2011

    Un logiciel n’est malheureusement jamais parfait...
    Si vous pensez avoir mis la main sur un bug, reportez le dans notre système de tickets en prenant bien soin de nous remonter certaines informations pertinentes : le type de navigateur et sa version exacte avec lequel vous avez l’anomalie ; une explication la plus précise possible du problème rencontré ; si possibles les étapes pour reproduire le problème ; un lien vers le site / la page en question ;
    Si vous pensez avoir résolu vous même le bug (...)

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

  • Encoding and processing into web-friendly formats

    13 avril 2011, par

    MediaSPIP automatically converts uploaded files to internet-compatible formats.
    Video files are encoded in MP4, Ogv and WebM (supported by HTML5) and MP4 (supported by Flash).
    Audio files are encoded in MP3 and Ogg (supported by HTML5) and MP3 (supported by Flash).
    Where possible, text is analyzed in order to retrieve the data needed for search engine detection, and then exported as a series of image files.
    All uploaded files are stored online in their original format, so you can (...)

Sur d’autres sites (8971)

  • avformat/smacker : Improve timestamps

    24 juin 2020, par Andreas Rheinhardt
    avformat/smacker : Improve timestamps
    

    A Smacker file can contain up to seven audio tracks. Up until now,
    the pts for the i. audio packet contained in a Smacker frame was
    simply the end timestamp of the last i. audio packet contained in
    an earlier Smacker frame.

    The problem with this is that a Smacker stream need not contain data in
    every Smacker frame and so the current i. audio packet present may come
    from a different underlying stream than the last i. audio packet
    contained in an earlier frame.

    The sample hypnotix.smk* exhibits this. It has three audio tracks and
    the first of the three has a longer first packet, so that the audio for
    the first track is contained in only 235 packets contained in the first
    235 Smacker frames ; the end timestamp of this track is 166696 (about 7.56s
    at a timebase of 1/22050) ; the other two audio tracks both have 253 packets
    contained in the first 253 Smacker frames. Up until now, the 236th
    packet of the second track being the first audio packet in the 236th
    Smacker frame would get the end timestamp of the last first audio packet
    from the last Smacker frame containing a first audio packet and said
    last audio packet is the first audio packet from the 235th Smacker frame
    from the first audio track, so that the timestamp is 166696. In contrast,
    the 236th packet from the third track (whose packets contain the same number
    of samples as the packets from the second track) has a timestamp of
    156116 (because its timestamp is derived from the end timestamp of the
    235th packet of the second audio track). In the end, the second track
    ended up being 177360/22050 s = 8.044s long ; in contrast, the third
    track was 166780/22050 s = 7.56s long which also coincided with the
    video.

    This commit fixes this by not using timestamps from other tracks for
    a packet's pts.

    * : https://samples.ffmpeg.org/game-formats/smacker/wetlands/hypnotix.smk

    Reviewed-by : Timotej Lazar <timotej.lazar@araneo.si>
    Signed-off-by : Andreas Rheinhardt <andreas.rheinhardt@gmail.com>

    • [DH] libavformat/smacker.c
  • Display FFMPEG decoded frame in a GLFW window

    17 juin 2020, par Infecto

    I am implementing the client program of a game where the server sends encoded frames of the game to the client (via UDP), while the client decodes them (via FFMPEG) and displays them in a GLFW window. &#xA;My program has two threads :

    &#xA;&#xA;

      &#xA;
    1. Thread 1 : renders the content of the uint8_t* variable dataToRender
    2. &#xA;

    3. Thread 2 : keeps obtaining frames from the server, decodes them and updates dataToRender accordingly
    4. &#xA;

    &#xA;&#xA;

    Thread 1 does the typical rendering of a GLFW window in a while-loop. I have already tried to display some dummy frame data (a completely red frame) and it worked :

    &#xA;&#xA;

    while (!glfwWindowShouldClose(window)) {&#xA;    glClear(GL_COLOR_BUFFER_BIT | GL_DEPTH_BUFFER_BIT);&#xA;    ...&#xA;&#xA;    glBindTexture(GL_TEXTURE_2D, tex_handle);&#xA;    glTexImage2D(GL_TEXTURE_2D, 0, GL_RGB, window_width, window_height, 0, GL_RGB, GL_UNSIGNED_BYTE, dataToRender);&#xA;    ...&#xA;    glfwSwapBuffers(window);&#xA;}&#xA;

    &#xA;&#xA;

    Thread 2 is where I am having trouble. I am unable to properly store the decoded frame into my dataToRender variable. On top if it, the frame data is originally in YUV format and needs to be converted to RGB. I use FFMPEG's sws_scale for that, which also gives me a bad dst image pointers error output in the console. Here's the code snippet responsible for that part :

    &#xA;&#xA;

            size_t data_size = frameBuffer.size();  // frameBuffer is a std::vector where I accumulate the frame data chunks&#xA;        uint8_t* data = frameBuffer.data();  // convert the vector to a pointer&#xA;        picture->format = AV_PIX_FMT_RGB24;&#xA;        av_frame_get_buffer(picture, 1);&#xA;        while (data_size > 0) {&#xA;            int ret = av_parser_parse2(parser, c, &amp;pkt->data, &amp;pkt->size,&#xA;                data, data_size, AV_NOPTS_VALUE, AV_NOPTS_VALUE, 0);&#xA;            if (ret &lt; 0) {&#xA;                fprintf(stderr, "Error while parsing\n");&#xA;                exit(1);&#xA;            }&#xA;            data &#x2B;= ret;&#xA;            data_size -= ret;&#xA;&#xA;            if (pkt->size) {&#xA;                swsContext = sws_getContext(&#xA;                    c->width, c->height,&#xA;                    AV_PIX_FMT_YUV420P, c->width, c->height,&#xA;                    AV_PIX_FMT_RGB24, SWS_BILINEAR, NULL, NULL, NULL&#xA;                );&#xA;                uint8_t* rgb24[1] = { data };&#xA;                int rgb24_stride[1] = { 3 * c->width };&#xA;                sws_scale(swsContext, rgb24, rgb24_stride, 0, c->height, picture->data, picture->linesize);&#xA;&#xA;                decode(c, picture, pkt, outname);&#xA;                // TODO: copy content of picture->data[0] to "dataToRender" maybe?&#xA;            }&#xA;        }&#xA;

    &#xA;&#xA;

    I have already tried doing another sws_scale to copy the content to dataToRender and I cannot get rid of the bad dst image pointers error. Any advice or solution to the problem would be greatly appreciated as I have been stuck for days on this.

    &#xA;

  • Set "start" field to 0 in mp3 with ffmpeg

    11 juin 2020, par TomatoCo

    I'm trying to change the bitrate and sample rate of an MP3 to match another to try and stop a small audio glitch from occurring when some game tries to play it. I've got the sample rate and bitrate right where I want them, but I can't get the "start" portion of

    &#xA;&#xA;

      Duration: 00:03:33.81, start: 0.025057, bitrate: 196 kb/s&#xA;    Stream #0:0: Audio: mp3, 44100 Hz, stereo, fltp, 196 kb/s&#xA;

    &#xA;&#xA;

    to go to 0, like the mp3 I'm trying to replace. The target looks like :

    &#xA;&#xA;

      Duration: 00:06:47.59, start: 0.000000, bitrate: 196 kb/s&#xA;    Stream #0:0: Audio: mp3, 44100 Hz, stereo, fltp, 196 kb/s&#xA;

    &#xA;&#xA;

    I've tried a variety of silenceremove filters and -ss flags to try and trim it, but I can't get rid of that "start" field. Google is failing me. What args am I looking for ?

    &#xA;