Recherche avancée

Médias (0)

Mot : - Tags -/navigation

Aucun média correspondant à vos critères n’est disponible sur le site.

Autres articles (29)

  • HTML5 audio and video support

    13 avril 2011, par

    MediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
    The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
    For older browsers the Flowplayer flash fallback is used.
    MediaSPIP allows for media playback on major mobile platforms with the above (...)

  • De l’upload à la vidéo finale [version standalone]

    31 janvier 2010, par

    Le chemin d’un document audio ou vidéo dans SPIPMotion est divisé en trois étapes distinctes.
    Upload et récupération d’informations de la vidéo source
    Dans un premier temps, il est nécessaire de créer un article SPIP et de lui joindre le document vidéo "source".
    Au moment où ce document est joint à l’article, deux actions supplémentaires au comportement normal sont exécutées : La récupération des informations techniques des flux audio et video du fichier ; La génération d’une vignette : extraction d’une (...)

  • Support audio et vidéo HTML5

    10 avril 2011

    MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
    Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
    Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
    Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...)

Sur d’autres sites (3870)

  • Splitting audio tracks with incorrect length - FFMPEG

    26 mars 2018, par channae

    Version : com.writingminds:FFmpegAndroid:0.3.2

    I have an audio file with length 43 seconds. And I wrote an algorithm to split at each 10 seconds mark where a word ends (For this I used IBM Watson to get ending timestamp). So cropping duration is always around 10 seconds to 11 seconds. Of course except the 5th one. I have printed my commands so that you will understand my use-case better.

    System.out: Split Command: -y -i /storage/emulated/0/AudioClipsForSpeakerRecognition/merge.wav -ss 00:00:00.000 -codec copy -t 00:00:10.010 /storage/emulated/0/AudioClipsForSpeakerRecognition/segment_1.wav

    System.out: Split Command: -y -i /storage/emulated/0/AudioClipsForSpeakerRecognition/merge.wav -ss 00:00:10.010 -codec copy -t 00:00:21.090 /storage/emulated/0/AudioClipsForSpeakerRecognition/segment_2.wav

    System.out: Split Command: -y -i /storage/emulated/0/AudioClipsForSpeakerRecognition/merge.wav -ss 00:00:21.090 -codec copy -t 00:00:30.480 /storage/emulated/0/AudioClipsForSpeakerRecognition/segment_3.wav

    System.out: Split Command: -y -i /storage/emulated/0/AudioClipsForSpeakerRecognition/merge.wav -ss 00:00:30.480 -codec copy -t 00:00:40.120 /storage/emulated/0/AudioClipsForSpeakerRecognition/segment_4.wav

    System.out: Split Command: -y -i /storage/emulated/0/AudioClipsForSpeakerRecognition/merge.wav -ss 00:00:40.120 -codec copy -t 00:00:43.000 /storage/emulated/0/AudioClipsForSpeakerRecognition/segment_5.wav

    However when playing all cropped audio files I noticed segment_1 is about 10 seconds and segment_2 is about 20 seconds etc. Therefore some of the audio parts belong to segment_1 also available in segment 2 etc etc. Why is this happening ?

    Appreciate your response.

  • Frame works/Packages for video editing in IOS ?

    16 juillet 2014, par 2vision2

    I’ve just started working on video editing application for IOS.

    What i am trying to do :

    Trying to create a typical Video Editing application with the following features.

    1. Frames split-up(thumbnails)
    2. Adding/merging videos
    3. Adding audio to the video.
    4. Audio/Video Fadeouts.
    5. Zoom in and Zoom out functionalities
    6. Adding subtitles/Titles
    7. Audio Filtering.

    What have i done so far :

    Put a detail pre development study and found that FFMPEG can do the job. I have taken
    the latest version of FFMPEG and built for IOS as .a files.

    However, i feel extremely hard to create a sample programs as i am beginner in ios.

    Also i’ve found that the IOS SDK frameworks like CoreVideo, CoreAnimation, CALayers may help.

    Now reading about OpenGLES.

    What’s my problem :

    1. Any help regarding finding the framework/package (from apple or third party) for the video editing application ?

    2. Is FFMPEG the only option ? If yes, can you guys give me a link or sample programs to start with FFMPEG ?

  • Why sliced thread affect so much on realtime encoding using ffmpeg x264 ?

    13 avril 2016, par CurtisGuo

    I’m using ffmpeg libx264 to encode a 720p screen captured from x11 in realtime with a fps of 30.
    when I use -tune zerolatency paramenter, the average encode time per-frame can be as large as 12ms with profile baseline.

    After a study of the ffmpeg x264 source code, I found that the key parameter leading to such long encode time is sliced-threads which enabled by -tune zerolatency. After disabled using -x264-params sliced-threads=0 the encode time can be as low as 2ms

    And with sliced-threads disabled, the CPU usage will be 40%, while only 20% when enabled.

    Can someone explain the details about this sliced-thread ? Especially in realtime encoding(assume no frame is buffered to be encoded. only encode when a frame is captured).