
Recherche avancée
Autres articles (111)
-
Emballe médias : à quoi cela sert ?
4 février 2011, parCe plugin vise à gérer des sites de mise en ligne de documents de tous types.
Il crée des "médias", à savoir : un "média" est un article au sens SPIP créé automatiquement lors du téléversement d’un document qu’il soit audio, vidéo, image ou textuel ; un seul document ne peut être lié à un article dit "média" ; -
Supporting all media types
13 avril 2011, parUnlike most software and media-sharing platforms, MediaSPIP aims to manage as many different media types as possible. The following are just a few examples from an ever-expanding list of supported formats : images : png, gif, jpg, bmp and more audio : MP3, Ogg, Wav and more video : AVI, MP4, OGV, mpg, mov, wmv and more text, code and other data : OpenOffice, Microsoft Office (Word, PowerPoint, Excel), web (html, CSS), LaTeX, Google Earth and (...)
-
Script d’installation automatique de MediaSPIP
25 avril 2011, parAfin de palier aux difficultés d’installation dues principalement aux dépendances logicielles coté serveur, un script d’installation "tout en un" en bash a été créé afin de faciliter cette étape sur un serveur doté d’une distribution Linux compatible.
Vous devez bénéficier d’un accès SSH à votre serveur et d’un compte "root" afin de l’utiliser, ce qui permettra d’installer les dépendances. Contactez votre hébergeur si vous ne disposez pas de cela.
La documentation de l’utilisation du script d’installation (...)
Sur d’autres sites (8078)
-
How to build opencv_ffmpeg.so for Linux/MacOS
27 octobre 2016, par BlackHow to build opencv_ffmpeg.so like opencv_ffmpeg.dll in Windows for OpenCV use in java app.
I followed build guide of FFmpeg.org but when build completed, it just generated many lib*.so.
How to build opencv_ffmpeg.so for shipping in my java application and load it dynamically when app runs ? -
OBS alternative for server : Creating a continuous video streaming to RTMP server and beign able to manipulate it using NodeJS
27 juillet 2016, par futufuWhat I want is to be able to create a livestream from a Ubuntu v14.04 server to a RTMP Server (like Twitch) and to be able to use NodeJS to control visual aspects (adding layers, text, images) and add different sources (video files, others livestreams, etc). Like having OBS running on a server.
What I’ve done/researched so far :
FFmpeg
With ffmpeg I can can create video files streams like that :
ffmpeg -re -i video.mp4 -c:v libx264 -preset fast -c:a aac -ab 128k -ar 44100 -f flv rtmp://example.com
Also using the
filter_complex
I can create something near to a layer like this tutorial explains :
https://trac.ffmpeg.org/wiki/Create%20a%20mosaic%20out%20of%20several%20input%20videosBut I found the following problems :
- The streams that I create with ffmpeg only last until the video file is over, if I wanted to stream multiple video files (dynamic playlist) it would interrupt the stream between each file ;
- The manipulation is very limited as far as I am concerned, I can’t edit
filter_complex
once ffmpeg is executing ; - Can’t display text and create animated overlays, like sliding text.
I tried to search for any cli/nodejs package that is able to create a continuos video stream and manipulate it to use as input source for ffmpeg which streams to the RTMP server.
Can someone give me more information about what I am trying to do ?
I’m playing with
github.com/fluent-ffmpeg/node-fluent-ffmpeg
to see if I have a different outcome. -
how to fill AvFrame pixel by pixel
18 mai 2017, par masoud khanloi try to encode opengl output to a mp4 file using ffmpeg but when i want to simply fill AvFrame with pixel colors with this code snippet :
for (y = 0; y < c->height; y++) {
for (x = 0; x < c->width; x++) {
int offset = 3 * (x + y * c->width);
if (x > 2* c->width / 3 && x width )
{
frame->data[0][offset + 0] = 0; // B
frame->data[0][offset + 1] = 0; // G
frame->data[0][offset + 2] = 255; // R
}
else if(x < 2 * c->width / 3 && x > c->width / 3) {
frame->data[0][offset + 0] = 0; // B
frame->data[0][offset + 1] = 255; // G
frame->data[0][offset + 2] = 0; // R
}
else {
frame->data[0][offset + 0] = 255; // B
frame->data[0][offset + 1] = 0; // G
frame->data[0][offset + 2] = 0; // R
}
}then i have this video output :
and i had this options for ffmpeg :
codec : libx264rgb
& pix_fmt : AV_PIX_FMT_BGR24in fact i expect something like this video output :
i’m also new in ffmpeg and i realy dont know advanced stuffs about video encoding ,so anybody knows what is problem ?