
Recherche avancée
Médias (1)
-
Carte de Schillerkiez
13 mai 2011, par
Mis à jour : Septembre 2011
Langue : English
Type : Texte
Autres articles (93)
-
MediaSPIP 0.1 Beta version
25 avril 2011, parMediaSPIP 0.1 beta is the first version of MediaSPIP proclaimed as "usable".
The zip file provided here only contains the sources of MediaSPIP in its standalone version.
To get a working installation, you must manually install all-software dependencies on the server.
If you want to use this archive for an installation in "farm mode", you will also need to proceed to other manual (...) -
Multilang : améliorer l’interface pour les blocs multilingues
18 février 2011, parMultilang est un plugin supplémentaire qui n’est pas activé par défaut lors de l’initialisation de MediaSPIP.
Après son activation, une préconfiguration est mise en place automatiquement par MediaSPIP init permettant à la nouvelle fonctionnalité d’être automatiquement opérationnelle. Il n’est donc pas obligatoire de passer par une étape de configuration pour cela. -
Les notifications de la ferme
1er décembre 2010, parAfin d’assurer une gestion correcte de la ferme, il est nécessaire de notifier plusieurs choses lors d’actions spécifiques à la fois à l’utilisateur mais également à l’ensemble des administrateurs de la ferme.
Les notifications de changement de statut
Lors d’un changement de statut d’une instance, l’ensemble des administrateurs de la ferme doivent être notifiés de cette modification ainsi que l’utilisateur administrateur de l’instance.
À la demande d’un canal
Passage au statut "publie"
Passage au (...)
Sur d’autres sites (6375)
-
Video duration not detected by moviepy
27 août 2023, par KyvI have downloaded a video online, and wanted to add a background sound to it.
I followed this link :
Combining an audio file with video file in python
But, while reading the file


my_clip = mpe.VideoFileClip('file.mp4')



I get the error [mov,mp4,m4a,3gp,3g2,mj2 @ 00000152769e78c0] moov atom not found : path/to/file.mp4 Invalid data found when processing input from line 320 in the github resource https://github.com/victorcampos7/tensorflow-ffmpeg/blob/master/ffmpeg_reader.py
How to overcome this issue ?

The video I am testing on : https://simgool.com/salon/SHY4xAqxVZ

-
What is the proper input format for drawtext's timecode option in fluent-ffmpeg (node.js) ?
24 janvier 2023, par Gil DekelI am trying to generate short video clips from 3 images with a simple millisecond counter at the top left per user request. So far I was able to generate the clips using the
videoshow npm lib
, no issues here.


However, when I try to add the timer via
fluent-ffmpeg
'sdrawtext
filter to an already generated video, I run into the following issue.


Here's what I have so far :



ffmpeg('video.mp4')
 .on('error', function(err, stdout, stderr) {
 console.log("ffmpeg stdout:\n" + stdout);
 console.log("ffmpeg stderr:\n" + stderr);
 })
 .videoCodec('libx264')
 .format('mp4')
 .fps(fps)
 .complexFilter([
 {
 filter: 'drawtext',
 options: {
 fontsize: 15,
 timecode: '00\:00\:00\:00', //<--- Issue is most likely here
 fontsize: 32,
 fontcolor: 'white',
 boxcolor: 'black',
 box: 1
 }
 }
 ]).save('out.mp4');




From what I found online I am quite convinced that my problem is with the
timecode
input format. Here's my error log :


ffmpeg stderr:
ffmpeg version 3.4.2 Copyright (c) 2000-2018 the FFmpeg developers
 built with Apple LLVM version 9.0.0 (clang-900.0.39.2)
 configuration: --prefix=/usr/local/Cellar/ffmpeg/3.4.2 --enable-shared --enable-pthreads --enable-version3 --enable-hardcoded-tables --enable-avresample --cc=clang --host-cflags= --host-ldflags= --disable-jack --enable-gpl --enable-libass --enable-libfdk-aac --enable-libfontconfig --enable-libfreetype --enable-libmp3lame --enable-libopus --enable-libvorbis --enable-libvpx --enable-libx264 --enable-libx265 --enable-libxvid --enable-opencl --enable-videotoolbox --disable-lzma --enable-nonfree
 libavutil 55. 78.100 / 55. 78.100
 libavcodec 57.107.100 / 57.107.100
 libavformat 57. 83.100 / 57. 83.100
 libavdevice 57. 10.100 / 57. 10.100
 libavfilter 6.107.100 / 6.107.100
 libavresample 3. 7. 0 / 3. 7. 0
 libswscale 4. 8.100 / 4. 8.100
 libswresample 2. 9.100 / 2. 9.100
 libpostproc 54. 7.100 / 54. 7.100
Input #0, mov,mp4,m4a,3gp,3g2,mj2, from 'video.mp4':
 Metadata:
 major_brand : isom
 minor_version : 512
 compatible_brands: isomiso2avc1mp41
 encoder : Lavf57.83.100
 Duration: 00:00:01.03, start: 0.000000, bitrate: 585 kb/s
 Stream #0:0(und): Video: h264 (High) (avc1 / 0x31637661), yuv420p, 640x1136 [SAR 923:924 DAR 130:231], 573 kb/s, 60 fps, 60 tbr, 15360 tbn, 120 tbc (default)
 Metadata:
 handler_name : VideoHandler
[Parsed_drawtext_0 @ 0x7f91aa700700] Both text and text file provided. Please provide only one
[AVFilterGraph @ 0x7f91aa700440] Error initializing filter 'drawtext' with args 'fontsize=32:timecode=00:00:00:00:fontcolor=white:boxcolor=black:box=1'
Error initializing complex filters.
Invalid argument




Specifically this :



[Parsed_drawtext_0 @ 0x7f91aa700700] Both text and text file provided. Please provide only one




When I run what should be the equivalent to my code in shell, it works perfectly :



ffmpeg -i video.mp4 \
 -vf drawtext="fontsize=15:timecode='00\:00\:00\:00':rate=60:fontsize=32:fontcolor='white':\
 boxcolor=0x000000AA:box=1" \
 -f mp4 out.mp4




I already spent a whole day trying to find any previous solutions to this problem, or documentation online regarding the proper format of
drawtext
'stimecode
option within node.js with no luck. It seems most of the information out there is for bash/shell use. I need to do this programmatically on a server per request, and I just can't find the right example to solve this.


Any help to resolve this would be appreciated.


-
Video Encode different parts of video using different bitrates
8 février 2015, par khateebIf I’ve a video like : video-lectures
And I need to encode the slides with high bitrate (as it’s the most important to the user) and encode the rest of the view with low bitrate.
-
Select the part to encode with high bitratre.
-
Separately Encoding the different parts with different bitrates.
- Combine the different parts into one video.
The purpose is to reduce the total file-size to be published online, hence reduce the bandwidth needed.
I need to do that using ffmpeg.
I’ll use that approach in encoding video-lectures recorded with still camera (180 degree view like the images above), since I know which regions are more useful to students (board and slides) than other regions. -