
Recherche avancée
Autres articles (34)
-
L’utiliser, en parler, le critiquer
10 avril 2011La première attitude à adopter est d’en parler, soit directement avec les personnes impliquées dans son développement, soit autour de vous pour convaincre de nouvelles personnes à l’utiliser.
Plus la communauté sera nombreuse et plus les évolutions seront rapides ...
Une liste de discussion est disponible pour tout échange entre utilisateurs. -
Contribute to documentation
13 avril 2011Documentation is vital to the development of improved technical capabilities.
MediaSPIP welcomes documentation by users as well as developers - including : critique of existing features and functions articles contributed by developers, administrators, content producers and editors screenshots to illustrate the above translations of existing documentation into other languages
To contribute, register to the project users’ mailing (...) -
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir
Sur d’autres sites (5374)
-
Unsplitting two video files with ffmpeg [duplicate]
2 septembre 2016, par exebookThis question is an exact duplicate of :
I cannot find a better word for this, because I searched for "combine", "join", "merge", "concat" and each time I get the information on how to get two videos into one so that the resulting file will play first video and then continues to the second one. I, on the other hand, need them to play simultaneously in the resulting video, so that first file goes to the left, and the second one to the right. It is a video chat recording of two person web cameras that need to get into one file to play at once. How to do that with ffmpeg ?
-
Tips for encoding a live stream to IIS Media Services (or Azure Live Media Services) with FFMPEG ?
6 mars 2014, par user3104748We're trying to encode assets, either live or static, in a live stream to IIS Media Services using ffmpeg. Can anyone provide pointers for exactly what kinds of parameters we should be using and setting ?
As part of our test, just to see if we can get things to work, we have a standard plain-old MP4 video static asset that we're trying to stream to the server. It seems to work on the client side, but when we try to view the video on the receiving end, we get nothing.
Here's an example of the command we're using, where gg.mp4 is the static MP4 video (obviously (hostname) is the name of our host and not the actual word in parenthesis :)...
ffmpeg -y -re -i gg.mp4 -movflags isml+frag_keyframe -f ismv -threads 0 -c:a libvo_aacenc -ac 2 -b:a 64k -c:v libx264 -preset fast -profile:v baseline -g 48 -keyint_min 48 -map 0:v -b:v:0 477k -s:v:0 368x152 -map 0:v -b:v:1 331k -s:v:1 288x120 -map 0:v -b:v:2 230k -s:v:2 224x92 -map 0:a:0 http://(hostname)/ingest.isml/Streams(video)
-
How to use ffmpeg with h.265 (or h.264) coded ?
8 mai 2015, par randomuser1Normally I start recording the camera image with the command :
ffmpeg -y -f vfwcap -r 25 -i 0 OUT.mp4
but I’m not sure which coded do I use in here (I’m just beginning my adventure with ffmpeg), however I found on this webpage this command :
ffmpeg -i INPUT -c:a copy -x265-params crf=25 OUT.mov
But when I run it - I get the following error :
INPUT: No such file or directory"
I changed the INPUT word above also to
0
, so the command isffmpeg -i 0 -c:a copy -x265-params crf=25 OUT.mov
,
but error stays similar (0: no such file or directory
).
How can I grab the camera image and save it to a file while using the H.265 ?