Recherche avancée

Médias (2)

Mot : - Tags -/kml

Autres articles (60)

  • La file d’attente de SPIPmotion

    28 novembre 2010, par

    Une file d’attente stockée dans la base de donnée
    Lors de son installation, SPIPmotion crée une nouvelle table dans la base de donnée intitulée spip_spipmotion_attentes.
    Cette nouvelle table est constituée des champs suivants : id_spipmotion_attente, l’identifiant numérique unique de la tâche à traiter ; id_document, l’identifiant numérique du document original à encoder ; id_objet l’identifiant unique de l’objet auquel le document encodé devra être attaché automatiquement ; objet, le type d’objet auquel (...)

  • Des sites réalisés avec MediaSPIP

    2 mai 2011, par

    Cette page présente quelques-uns des sites fonctionnant sous MediaSPIP.
    Vous pouvez bien entendu ajouter le votre grâce au formulaire en bas de page.

  • Gestion des droits de création et d’édition des objets

    8 février 2011, par

    Par défaut, beaucoup de fonctionnalités sont limitées aux administrateurs mais restent configurables indépendamment pour modifier leur statut minimal d’utilisation notamment : la rédaction de contenus sur le site modifiables dans la gestion des templates de formulaires ; l’ajout de notes aux articles ; l’ajout de légendes et d’annotations sur les images ;

Sur d’autres sites (5945)

  • af_hdcd : allow all HDCD sample rates

    7 septembre 2016, par Burt P
    af_hdcd : allow all HDCD sample rates
    

    The PM Model Two could output HDCD-encoded audio in CD and all
    DVD-Audio sample rates. (44100, 48000, 88200, 96000, 176400, and
    192000 Hz)

    Signed-off-by : Burt P <pburt0@gmail.com>

    • [DH] libavfilter/af_hdcd.c
  • Sending per frame metadata with H264 encoded frames

    21 septembre 2013, par user2459280

    We're looking for a way to send per frame metadata (for example an ID) with H264 encoded frames from a server to a client.

    We're currently developing a remote rendering application, where both client and server side are actively involved.
    The server renders a high quality image with all effects, lighting etc.
    The client also has model-informations and renders a diffuse image that is used when the bandwidth is too low or the images have to be warped in order to avoid stuttering .

    So far we're encoding the frames on the server side with ffmpeg and streaming them with live555 to the client, who receives an rtsp-stream and decodes the frames again using ffmpeg.

    For our application, we now need to send per frame metadata.
    We want the client to tell the server where the camera is right now.
    Ideally we'd be able to send the client's view matrix to the server, render the corresponding frame and send it back to the client together with its view matrix. So when the client receives a frame, we need to know exactly at what camera position the frame was rendered.

    Alternatively we could also tag each view matrix with an ID, send it to the server, render the frame and tag it with the same ID and send it back. In this case we'd have to assign the right matrix to the frame again on the client side.

    After several attempts to realize the above intent with ffmpeg we came to the conclusion that ffmpeg does not provide the required functionality. ffmpeg only provides a fix, predefined set of fields for metadata, that either cannot store a matrix or can only be set for every key frame, which is not frequently enough for our purpose.

    Now we're considering using live555. So far we have an on demand Server, witch gets a VideoSubsession with a H264VideoStreamDiscreteFramer to contain our own FramedSource class. In this class we load the encoded AVPacket (from ffmpeg) and send its data-buffer over the network. Now we need a way to send some kind of metadata with every frame to the client.

    Do you have any ideas how to solve this metadata problem with live555 oder another library ?

    Thanks for your help !

  • ffmpeg to opengl texture with glTexImage2D

    12 juin 2013, par 2kPi

    I'm trying to make a player in OpenGL. I use ffmpeg to decode the video.

    I searched several tutorial (StackOverflow, ....) how to convert my pFrameRGB-> data [0] texture in OpenGL, but it still shows me a white square. I converted my video to 512x256 to comply with 2 ^ n.

    I know my pFrameRGB is correct because I can create frameX.ppm with "SaveFrame (...)" function.

    I used the following source ( https://github.com/arashafiei/dranger-ffmpeg-tuto/blob/master/tutorial02.c ) code as a model by adapting for my player.

    And this is my source code :

    http://www.sourcepod.com/uagkel22-19078

    Does one of you have a solution to my problem ?

    Edit 1 :

    I delete :

    glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_MIN_FILTER, GL_LINEAR_MIPMAP_NEAREST);
    glTexParameterf(GL_TEXTURE_2D, GL_TEXTURE_MAG_FILTER, GL_LINEAR);

    glTexParameterf( GL_TEXTURE_2D, GL_TEXTURE_WRAP_S, 1 ? GL_REPEAT : GL_CLAMP );
    glTexParameterf( GL_TEXTURE_2D, GL_TEXTURE_WRAP_T, 1 ? GL_REPEAT : GL_CLAMP );

    And replaced :

    glTexImage2D( GL_TEXTURE_2D, 0, GL_RGB, pCodecCtx->width, pCodecCtx->height, 0, GL_RGB, GL_UNSIGNED_BYTE, pFrameRGB->data[0]);

    by :

    if(firstRendering){
    glBindTexture(GL_TEXTURE_2D, texture);
    glTexImage2D( GL_TEXTURE_2D, 0, GL_RGB, pCodecCtx->width, pCodecCtx->height, 0, GL_RGB, GL_UNSIGNED_BYTE, pFrameRGB->data[0]);

    firstRendering = 0;

    }else{

    glActiveTexture(texture);
    glBindTexture(GL_TEXTURE_2D, texture);
    glTexSubImage2D( GL_TEXTURE_2D, 0, 0, 0, pCodecCtx->width, pCodecCtx->height, GL_RGB, GL_UNSIGNED_BYTE, pFrameRGB->data[0]);
    }

    And i try to run but stiil nothing...