Recherche avancée

Médias (0)

Mot : - Tags -/auteurs

Aucun média correspondant à vos critères n’est disponible sur le site.

Autres articles (21)

  • La sauvegarde automatique de canaux SPIP

    1er avril 2010, par

    Dans le cadre de la mise en place d’une plateforme ouverte, il est important pour les hébergeurs de pouvoir disposer de sauvegardes assez régulières pour parer à tout problème éventuel.
    Pour réaliser cette tâche on se base sur deux plugins SPIP : Saveauto qui permet une sauvegarde régulière de la base de donnée sous la forme d’un dump mysql (utilisable dans phpmyadmin) mes_fichiers_2 qui permet de réaliser une archive au format zip des données importantes du site (les documents, les éléments (...)

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

  • Encodage et transformation en formats lisibles sur Internet

    10 avril 2011

    MediaSPIP transforme et ré-encode les documents mis en ligne afin de les rendre lisibles sur Internet et automatiquement utilisables sans intervention du créateur de contenu.
    Les vidéos sont automatiquement encodées dans les formats supportés par HTML5 : MP4, Ogv et WebM. La version "MP4" est également utilisée pour le lecteur flash de secours nécessaire aux anciens navigateurs.
    Les documents audios sont également ré-encodés dans les deux formats utilisables par HTML5 :MP3 et Ogg. La version "MP3" (...)

Sur d’autres sites (5047)

  • ffmpeg FPS information mismatch with the video

    8 décembre 2017, par Adorn

    I have a bunch of videos with some statistics of what is happening inside a video. One such piece of information is given in terms of time of the video in seconds up to one decimal point.

    To get the FPS of a video, I am using ffmpeg -i

    But when I manually compute one particular frame’s time using given FPS, it does not match.

    For example, ffmpeg outputs FPS = 30.

    I look at the video statistics, the frame at the 156.8 = 2.368 has to be
    4704'th frame. I open the video using ’skvideo’, read all the frames, and view the 4704'th frame. It is some frame around time 2.12 ?. I checked multiple such instances in multiple videos and this is a common behavior.

    I do not understand why this is so and how can I get around the problem ?

    As such I am not bounded by ffmpeg. Skvideo is being used to read the videos. I tried opencv, as of now it does not work with VideoCapture, and reinstalling it is costly for me time wise. But I guess ’opencv/skvideo’ should not matter, one can count the frames manually as well.

    So, in the solution, I am looking out for -

    1. Given timestamps of inside of a video, how can I find a frame of that particular time location ?

    2. In case someone might have already worked on this, this is related to THUMOS dataset. I am on Ubuntu 16.04

    EDIT_1

    Actually I can be more specific as it is a publicly available data. The time bounds are of an important activity. For example, in a video, when does basketball dunk occurs ? It is given in pairs - [start end]. Some videos have multiple activities, some have only one.

    Here is a sample video, and following are the activity times.

    [[  16.5, 20.8],
    [  26.6, 32.2],
    [  34.8, 42.1],
    [  47.8, 50.0],
    [  58.1, 62.9],
    [  65.6, 67.2],
    [  68.5, 74.0],
    [  76.4, 78.3],
    [  78.7, 79.8],
    [  80.8, 82.1],
    [  85.0, 87.3],
    [  90.1, 91.4],
    [  98.5, 100.3]]

    I also tried checking manually, 32.87 FPS "almost" works for few videos but not for all. and almost means it is off by 10 frames. This is a huge difference for my task, and I need exact frame.

    Also, there has to be some way, because it can be visually observed with multiple video players that times in the dataset are correct.

  • Typesetting

    9 juin 2010, par Mikko Koppanen — Imagick, PHP stuff

    Ever had the situation where you have a piece of string which you need to overlay on an image ? Maybe a situation where the area reserved for the string is known in pixels but you need to know the font size to fill most of the area ? Think no more !

    Here is a small example of how to fit a certain piece of a string on to an area of which you know the width and the height or only the width. The magic happens through the ImageMagick CAPTION : format. You can see from the example images how the parameters actually affect the image.

    1. < ?php
    2.  
    3. /* How wide is our image */
    4. $image_width = 200 ;
    5.  
    6. /* Give zero for autocalculating the height */
    7. $image_height = 200 ;
    8.  
    9. /* Specify the text */
    10. $text = "Lorem ipsum dolor sit amet, consectetuer adipiscing elit.
    11.     Mauris lectus mi, mattis non, euismod vel, sagittis nec, ipsum." ;
    12.  
    13. /* Instanciate imagick */
    14. $im = new Imagick() ;
    15.  
    16. /* Create new image using caption : pseudo format */
    17. $im->newPseudoImage( $image_width, $image_height, "caption :" . $text ) ;
    18.  
    19. /* Put 1px border around the image */
    20. $im->borderImage( ’black’, 1, 1 ) ;
    21.  
    22. /* PNG format */
    23. $im->setImageFormat( "png")  ;
    24.  
    25. /* Output */
    26. header( "Content-Type : image/png" ) ;
    27. echo $im ;
    28.  
    29.  ?>

    Here is image with width 100 and height 0 :

    width_100_height_0.png

    Width 100 Height 50 :

    width_100_height_50.png

    Width 200 Height 200 (as you can see the font size is now larger) :

    width_200_height_200.png

  • QOpenGLWidget video rendering perfomance in multiple processes

    24 novembre 2018, par MasterAler

    My problem may seem vague without code, but it actually isn’t.

    So, there I’ve got an almost properly-working widget, which renders video frames.

    Qt 5.10 and QOpenGLWidget subclassing worked fine, I didn’t make any sophisticated optimizations — there are two textures and a couple of shaders, converting YUV pixel format to RGB — glTexImage2D() + shaders, no buffers.

    Video frames are obtained from FFMPEG, it shows great performance due to hardware acceleration... when there is only one video window.

    The piece of software is a "video wall" — multiple independent video windows on the same screen. Of course, multi-threading would be the preferred solution, but legacy holds for now, I can’t change it.

    So, 1 window with Full HD video consumes 2% CPU & 8-10% GPU regardless of the size of the window. But 7-10 similar windows, launched from the same executable at the same time consume almost all the CPU. My math says that 2 x 8 != 100...

    My best guesses are :

    • This is a ffmpeg decoder issue, hardware acceleration still is not magic, some hardware pipeline stalls
    • 7-8-9 independent OpenGL contexts cost a lot more than 1 cost x N
    • I’m not using PUBO or some other complex techniques to improve OpenGL rendering. It still explains nothing, but at least it is a guess

    The behavior is the same on Ubuntu, where decoding uses different codec (I mean that using GPU accelerated or CPU accelerated codecs makes no difference !), so, it makes more probable that I’m missing something about OpenGL... or not, because launching 6-7 Qt examples with dynamic textures shows normal growth of CPU usages — it is approximately a sum for the number of windows.

    Anyway, it becomes quite tricky for me to profile the case, so I hope someone could have been solving the similar problem before and could share his experience with me. I’d be appreciated for any ideas, how to deal with the described riddle.

    I can add any pieces of code if that helps.