
Recherche avancée
Médias (1)
-
The Slip - Artworks
26 septembre 2011, par
Mis à jour : Septembre 2011
Langue : English
Type : Texte
Autres articles (78)
-
Creating farms of unique websites
13 avril 2011, parMediaSPIP platforms can be installed as a farm, with a single "core" hosted on a dedicated server and used by multiple websites.
This allows (among other things) : implementation costs to be shared between several different projects / individuals rapid deployment of multiple unique sites creation of groups of like-minded sites, making it possible to browse media in a more controlled and selective environment than the major "open" (...) -
HTML5 audio and video support
13 avril 2011, parMediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
For older browsers the Flowplayer flash fallback is used.
MediaSPIP allows for media playback on major mobile platforms with the above (...) -
Support audio et vidéo HTML5
10 avril 2011MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...)
Sur d’autres sites (9589)
-
Is there an efficient way to use ffmpeg to perform a large quantity of cuts from a single file ?
16 mars 2024, par Giuliano OliveriI'm trying to cut video files into smaller chunks. (each one being one word said in the video, so they're not all of equal size)


I've tried a lot of different approaches to try to be as efficient as possible, but I can't get the runtime to be under 2/3rd of the original video length. That's an issue because I'm trying to process 400+ hours of video.


Is there a more efficient way to do this ? Or am I doomed to run this for weeks ?


Here is the command for my best attempt so far


ffmpeg -hwaccel cuda -hwaccel_output_format cuda -ss start_timestamp -t to_timestamp -i file_name -vf "fps=30,scale_cuda=1280:720" -c:v h264_nvenc -y output_file



Note that the machine running the code has a 4090
This command is then executed via python, which gives it the right timestamps and file paths for each smaller clip in a for loop


I think it's wasting a lot of time calling a new process each time, however I haven't been able to get better results with a split filter ; but here's the ffmpeg-python code for that attempt :


Creation of the stream :


inp = (
 ffmpeg
 .input(file_name, hwaccel="cuda", hwaccel_output_format="cuda")
 .filter("fps",fps=30)
 .filter('scale_cuda', '1280','720')
 .filter_multi_output('split')
)



Which then gets called in a for loop


(
 ffmpeg
 .filter(inp, 'trim', start=row[1]['start'], end=row[1]['end'])
 .filter('setpts', 'PTS-STARTPTS')
 .output(output_file,vcodec='h264_nvenc')
 .run()
)



-
Splitting audio tracks with incorrect length - FFMPEG
26 mars 2018, par channaeVersion : com.writingminds:FFmpegAndroid:0.3.2
I have an audio file with length 43 seconds. And I wrote an algorithm to split at each 10 seconds mark where a word ends (For this I used IBM Watson to get ending timestamp). So cropping duration is always around 10 seconds to 11 seconds. Of course except the 5th one. I have printed my commands so that you will understand my use-case better.
System.out: Split Command: -y -i /storage/emulated/0/AudioClipsForSpeakerRecognition/merge.wav -ss 00:00:00.000 -codec copy -t 00:00:10.010 /storage/emulated/0/AudioClipsForSpeakerRecognition/segment_1.wav
System.out: Split Command: -y -i /storage/emulated/0/AudioClipsForSpeakerRecognition/merge.wav -ss 00:00:10.010 -codec copy -t 00:00:21.090 /storage/emulated/0/AudioClipsForSpeakerRecognition/segment_2.wav
System.out: Split Command: -y -i /storage/emulated/0/AudioClipsForSpeakerRecognition/merge.wav -ss 00:00:21.090 -codec copy -t 00:00:30.480 /storage/emulated/0/AudioClipsForSpeakerRecognition/segment_3.wav
System.out: Split Command: -y -i /storage/emulated/0/AudioClipsForSpeakerRecognition/merge.wav -ss 00:00:30.480 -codec copy -t 00:00:40.120 /storage/emulated/0/AudioClipsForSpeakerRecognition/segment_4.wav
System.out: Split Command: -y -i /storage/emulated/0/AudioClipsForSpeakerRecognition/merge.wav -ss 00:00:40.120 -codec copy -t 00:00:43.000 /storage/emulated/0/AudioClipsForSpeakerRecognition/segment_5.wavHowever when playing all cropped audio files I noticed segment_1 is about 10 seconds and segment_2 is about 20 seconds etc. Therefore some of the audio parts belong to segment_1 also available in segment 2 etc etc. Why is this happening ?
Appreciate your response.
-
Gujarati text properly not show on output video using FFmpeg Library in Android
15 mars 2024, par sanjay dangarGujarati text overly on video using FFmpeg Library on Android


below my code in Gujarati text twin word proper does not show


I am Enter this text ="વડાપ્રધાનશ્રી નરેન્દ્રભાઇ મોદી." but output video in proper not show


fun addFrame2VideoEditFun(
 videoPath: String,
 fontPath: String,
 outputPath: String
 ): Array<string> {
 val inputs: ArrayList<string> = ArrayList()
 var textFile = "વડાપ્રધાનશ્રી નરેન્દ્રભાઇ મોદી."
 inputs.apply {
 add("-i")
 add(videoPath)
 add("-vf")
 add("drawtext=fontfile=$fontPath:text=$textFile:fontsize=24:fontcolor=white:x=(w-text_w)/2:y=h-line_h")
 add("-c:a")
 add("copy")
 add(outputPath)
 }
 return inputs.toArray(arrayOfNulls(inputs.size))
 }
</string></string>


my output video in below text show




Gujarati text properly does not show on output video using FFmpeg Library.