
Recherche avancée
Médias (10)
-
Demon Seed
26 septembre 2011, par
Mis à jour : Septembre 2011
Langue : English
Type : Audio
-
Demon seed (wav version)
26 septembre 2011, par
Mis à jour : Avril 2013
Langue : English
Type : Audio
-
The four of us are dying (wav version)
26 septembre 2011, par
Mis à jour : Avril 2013
Langue : English
Type : Audio
-
Corona radiata (wav version)
26 septembre 2011, par
Mis à jour : Avril 2013
Langue : English
Type : Audio
-
Lights in the sky (wav version)
26 septembre 2011, par
Mis à jour : Avril 2013
Langue : English
Type : Audio
-
Head down (wav version)
26 septembre 2011, par
Mis à jour : Avril 2013
Langue : English
Type : Audio
Autres articles (21)
-
La sauvegarde automatique de canaux SPIP
1er avril 2010, parDans le cadre de la mise en place d’une plateforme ouverte, il est important pour les hébergeurs de pouvoir disposer de sauvegardes assez régulières pour parer à tout problème éventuel.
Pour réaliser cette tâche on se base sur deux plugins SPIP : Saveauto qui permet une sauvegarde régulière de la base de donnée sous la forme d’un dump mysql (utilisable dans phpmyadmin) mes_fichiers_2 qui permet de réaliser une archive au format zip des données importantes du site (les documents, les éléments (...) -
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir -
Encodage et transformation en formats lisibles sur Internet
10 avril 2011MediaSPIP transforme et ré-encode les documents mis en ligne afin de les rendre lisibles sur Internet et automatiquement utilisables sans intervention du créateur de contenu.
Les vidéos sont automatiquement encodées dans les formats supportés par HTML5 : MP4, Ogv et WebM. La version "MP4" est également utilisée pour le lecteur flash de secours nécessaire aux anciens navigateurs.
Les documents audios sont également ré-encodés dans les deux formats utilisables par HTML5 :MP3 et Ogg. La version "MP3" (...)
Sur d’autres sites (5047)
-
ffmpeg FPS information mismatch with the video
8 décembre 2017, par AdornI have a bunch of videos with some statistics of what is happening inside a video. One such piece of information is given in terms of time of the video in seconds up to one decimal point.
To get the FPS of a video, I am using
ffmpeg -i
But when I manually compute one particular frame’s time using given FPS, it does not match.
For example, ffmpeg outputs FPS = 30.
I look at the video statistics, the frame at the 156.8 = 2.368 has to be
4704'th
frame. I open the video using ’skvideo’, read all the frames, and view the4704'th
frame. It is some frame around time 2.12 ?. I checked multiple such instances in multiple videos and this is a common behavior.I do not understand why this is so and how can I get around the problem ?
As such I am not bounded by ffmpeg. Skvideo is being used to read the videos. I tried
opencv
, as of now it does not work withVideoCapture
, and reinstalling it is costly for me time wise. But I guess ’opencv/skvideo’ should not matter, one can count the frames manually as well.So, in the solution, I am looking out for -
-
Given timestamps of inside of a video, how can I find a frame of that particular time location ?
-
In case someone might have already worked on this, this is related to
THUMOS
dataset. I am on Ubuntu 16.04
EDIT_1
Actually I can be more specific as it is a publicly available data. The time bounds are of an important activity. For example, in a video, when does basketball dunk occurs ? It is given in pairs - [start end]. Some videos have multiple activities, some have only one.
Here is a sample video, and following are the activity times.
[[ 16.5, 20.8],
[ 26.6, 32.2],
[ 34.8, 42.1],
[ 47.8, 50.0],
[ 58.1, 62.9],
[ 65.6, 67.2],
[ 68.5, 74.0],
[ 76.4, 78.3],
[ 78.7, 79.8],
[ 80.8, 82.1],
[ 85.0, 87.3],
[ 90.1, 91.4],
[ 98.5, 100.3]]I also tried checking manually, 32.87 FPS "almost" works for few videos but not for all. and almost means it is off by 10 frames. This is a huge difference for my task, and I need exact frame.
Also, there has to be some way, because it can be visually observed with multiple video players that times in the dataset are correct.
-
-
Typesetting
Ever had the situation where you have a piece of string which you need to overlay on an image ? Maybe a situation where the area reserved for the string is known in pixels but you need to know the font size to fill most of the area ? Think no more !
Here is a small example of how to fit a certain piece of a string on to an area of which you know the width and the height or only the width. The magic happens through the ImageMagick CAPTION : format. You can see from the example images how the parameters actually affect the image.
-
< ?php
-
-
/* How wide is our image */
-
$image_width = 200 ;
-
-
/* Give zero for autocalculating the height */
-
$image_height = 200 ;
-
-
/* Specify the text */
-
$text = "Lorem ipsum dolor sit amet, consectetuer adipiscing elit.
-
Mauris lectus mi, mattis non, euismod vel, sagittis nec, ipsum." ;
-
-
/* Instanciate imagick */
-
$im = new Imagick() ;
-
-
/* Create new image using caption : pseudo format */
-
$im->newPseudoImage( $image_width, $image_height, "caption :" . $text ) ;
-
-
/* Put 1px border around the image */
-
$im->borderImage( ’black’, 1, 1 ) ;
-
-
/* PNG format */
-
$im->setImageFormat( "png") ;
-
-
/* Output */
-
header( "Content-Type : image/png" ) ;
-
echo $im ;
-
-
?>
Here is image with width 100 and height 0 :
Width 100 Height 50 :
Width 200 Height 200 (as you can see the font size is now larger) :
-
-
QOpenGLWidget video rendering perfomance in multiple processes
24 novembre 2018, par MasterAlerMy problem may seem vague without code, but it actually isn’t.
So, there I’ve got an almost properly-working widget, which renders video frames.
Qt 5.10 and QOpenGLWidget subclassing worked fine, I didn’t make any sophisticated optimizations — there are two textures and a couple of shaders, converting YUV pixel format to RGB —
glTexImage2D()
+ shaders, no buffers.Video frames are obtained from FFMPEG, it shows great performance due to hardware acceleration... when there is only one video window.
The piece of software is a "video wall" — multiple independent video windows on the same screen. Of course, multi-threading would be the preferred solution, but legacy holds for now, I can’t change it.
So, 1 window with Full HD video consumes 2% CPU & 8-10% GPU regardless of the size of the window. But 7-10 similar windows, launched from the same executable at the same time consume almost all the CPU. My math says that 2 x 8 != 100...
My best guesses are :
- This is a ffmpeg decoder issue, hardware acceleration still is not magic, some hardware pipeline stalls
- 7-8-9 independent OpenGL contexts cost a lot more than 1 cost x N
- I’m not using PUBO or some other complex techniques to improve OpenGL rendering. It still explains nothing, but at least it is a guess
The behavior is the same on Ubuntu, where decoding uses different codec (I mean that using GPU accelerated or CPU accelerated codecs makes no difference !), so, it makes more probable that I’m missing something about OpenGL... or not, because launching 6-7 Qt examples with dynamic textures shows normal growth of CPU usages — it is approximately a sum for the number of windows.
Anyway, it becomes quite tricky for me to profile the case, so I hope someone could have been solving the similar problem before and could share his experience with me. I’d be appreciated for any ideas, how to deal with the described riddle.
I can add any pieces of code if that helps.