Recherche avancée

Médias (1)

Mot : - Tags -/graphisme

Autres articles (49)

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

  • HTML5 audio and video support

    13 avril 2011, par

    MediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
    The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
    For older browsers the Flowplayer flash fallback is used.
    MediaSPIP allows for media playback on major mobile platforms with the above (...)

  • Support audio et vidéo HTML5

    10 avril 2011

    MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
    Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
    Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
    Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...)

Sur d’autres sites (8131)

  • Terminal command to upgrade FFMPEG

    14 avril 2023, par Savvy Turtle

    Everything is still installed but once I upgraded my server from CentOS 7.6 to Cloud Linux FFMPEG does not seem to work anymore.

    


    When I go and debug it in the admin section where you upload a test video and hit debug it shows a toast to upload a video. and on the front end of https://turtle.tube/ users upload a video and it gets stuck at 100% upload and processed.

    


    So I'm wondering whether I need to ugrade FFMPEG to work with Cloud Linux update or If I need to restart it.

    


    Any help would be great thanks.

    


    Savvy

    


    Checked Version, looked for terminal command to upgrade FFMPEG to latest stable release of : FFmpeg 5.1.3 "Riemann" or FFmpeg 6.0 "Von Neumann" but have not figured out the command to upgrade yet.

    


    nor the command to restart FFMPEG if that even exist.

    


    Current version of FFMPEG shows :

    


    ffmpeg -version

    


    ffmpeg version 3.4.12 Copyright (c) 2000-2022 the FFmpeg developers
built with gcc 4.8.5 (GCC) 20150623 (Red Hat 4.8.5-44)
configuration : —prefix=/usr —bindir=/usr/bin —datadir=/usr/share/ffmpeg —docdir=/usr/share/doc/ffmpeg —incdir=/usr/include/ffmpeg —libdir=/usr/lib64 —mandir=/usr/share/man —arch=x86_64 —optflags='-O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector-strong —param=ssp-buffer-size=4 -grecord-gcc-switches -m64 -mtune=generic' —extra-ldflags='-Wl,-z,relro ' —extra-cflags=' ' —enable-libopencore-amrnb —enable-libopencore-amrwb —enable-libvo-amrwbenc —enable-version3 —enable-bzlib —disable-crystalhd —enable-fontconfig —enable-gcrypt —enable-gnutls —enable-ladspa —enable-libass —enable-libbluray —enable-libcdio —enable-libdrm —enable-indev=jack —enable-libfreetype —enable-libfribidi —enable-libgsm —enable-libmp3lame —enable-nvenc —enable-openal —enable-opencl —enable-opengl —enable-libopenjpeg —enable-libopus —disable-encoder=libopus —enable-libpulse —enable-librsvg —enable-libsoxr —enable-libspeex —enable-libtheora —enable-libvorbis —enable-libv4l2 —enable-libvidstab —enable-libvpx —enable-libx264 —enable-libx265 —enable-libxvid —enable-libzvbi —enable-avfilter —enable-avresample —enable-libmodplug —enable-postproc —enable-pthreads —disable-static —enable-shared —enable-gpl —disable-debug —disable-stripping —shlibdir=/usr/lib64 —enable-libmfx —enable-runtime-cpudetect
libavutil 55. 78.100 / 55. 78.100
libavcodec 57.107.100 / 57.107.100
libavformat 57. 83.100 / 57. 83.100
libavdevice 57. 10.100 / 57. 10.100
libavfilter 6.107.100 / 6.107.100
libavresample 3. 7. 0 / 3. 7. 0
libswscale 4. 8.100 / 4. 8.100
libswresample 2. 9.100 / 2. 9.100
libpostproc 54. 7.100 / 54. 7.100

    


  • ffmpeg convert RTMP stream to HLS ? RED5

    23 février 2017, par Lewis Day

    Before posting this question ; I followed this post ;
    https://ffmpeg.org/pipermail/ffmpeg-user/2013-November/018558.html I also tried to find solutions to the problem I’m having.

    ffmpeg -v verbose -i "rtmp://domain.com:5080/stream1" -c:v libx264 -c:a aac -ac 1 -strict -2 -crf 18 -profile:v baseline -maxrate 400k -bufsize 1835k -pix_fmt yuv420p -flags -global_header -hls_time 10 -hls_list_size 6 -hls_wrap 10 -start_number 1 '/home/domain/public_html/hts-streams/stream1.m3u8'

    Okay above is the Ffmpeg command I’m trying to execute to convert my RTMP stream into a .m3u8.

    However the command is failing and I’m being returned this ;

    I have changed my domain to domain.com

     configuration: --prefix=/usr --bindir=/usr/bin --datadir=/usr/share/ffmpeg --incdir=/usr/include/ffmpeg --libdir=/usr/lib64 --mandir=/usr/share/man --arch=x86_64 --extra-cflags='-O2 -g -pipe -Wall -Wp,-D_FORTIFY_SOURCE=2 -fexceptions -fstack-protector --param=ssp-buffer-size=4 -m64 -mtune=generic' --enable-bzlib --disable-crystalhd --enable-gnutls --enable-libass --enable-libcdio --enable-libcelt --enable-libdc1394 --disable-indev=jack --enable-libfreetype --enable-libgsm --enable-libmp3lame --enable-openal --enable-libopenjpeg --enable-libpulse --enable-librtmp --enable-libschroedinger --enable-libspeex --enable-libtheora --enable-libvorbis --enable-libv4l2 --enable-libx264 --enable-libxvid --enable-x11grab --enable-avfilter --enable-postproc --enable-pthreads --disable-static --enable-shared --enable-gpl --disable-debug --disable-stripping --shlibdir=/usr/lib64 --enable-runtime-cpudetect
     libavutil      51. 35.100 / 51. 35.100
     libavcodec     53. 61.100 / 53. 61.100
     libavformat    53. 32.100 / 53. 32.100
     libavdevice    53.  4.100 / 53.  4.100
     libavfilter     2. 61.100 /  2. 61.100
     libswscale      2.  1.100 /  2.  1.100
     libswresample   0.  6.100 /  0.  6.100
     libpostproc    52.  0.100 / 52.  0.100
    Parsing...
    Parsed protocol: 0
    Parsed host    : domain.com
    Parsed app     : stream1
    RTMP_Connect1, ... connected, handshaking
    HandShake: Type Answer   : 48
    HandShake: Type mismatch: client sent 3, server answered 72
    RTMP_Connect1, handshake failed.
    rtmp://domain.com:5080/stream1: Operation not permitted

    The command then fails to convert the stream due to a failed handshake. What is causing this ?

  • Bootstrapping an AI UGC system — video generation is expensive, APIs are limiting, and I need help navigating it all [closed]

    24 juin, par Barack _ Ouma

    I’m building a solo AI-powered UGC (User-Generated Content) platform — something that automates the creation of short-form content using AI avatars, voices, visuals, and scripts. But I’ve hit a wall with video generation and API limitations.

    


    So far, I’ve integrated TTS and voice cloning (using ElevenLabs), and I’ve gotten image generation working. But video generation (especially talking avatars) has been a nightmare — both financially and technically.

    


    🛠️ Features I’m trying to build :

    


    AI avatars (face + lip-syncing)
Script generation (LLM-driven)
Image generation
Video composition

    


    I’m trying to build an AI faceless content creation automtion platform alternative to Makeugc.com or Reelfarm.org or postbridge.com — just trying to create a working pipeline for automated content.

    


    ❌ Challenges so far :

    


    Services like D-ID, Synthesia, Magic Hour, and Luma are either paywalled, have no trials, or are very expensive.

    


    D-ID does support avatar creation, but you need to pay upfront to even access those features. There's no easy/free entry point.

    


    Tools like Google Veo 3 are powerful but clearly not accessible for indie builders.
I’ve looked into open-source models like WAN 2.1, CogVideo, etc., but I have no clue how to run them or what infra is needed.

    


    Now I’m torn between buying my own GPU or renting compute power to self-host these models.

    


    💸 Cost is a huge blocker

    


    I’ve been looking through Replicate’s pricing, and while some models (especially image gen) are manageable, video models get expensive fast. Even GPU rental rates stack up quickly, especially if you’re testing often or experimenting with pipelines. Plus, idle time billing doesn’t help.

    


    💭 What I could really use help with :

    


    Has anyone successfully stitched together APIs (voice, avatar, video) into a working UGC pipeline ?

    


    Should I use separate services (e.g. ElevenLabs + Synthesia + WAN) or try to host my own end-to-end system ?

    


    Is it cheaper (long term) to buy a used GPU like a 4090 and run things locally ? Or better to rent compute short-term ?

    


    Any open-source solutions that are beginner-friendly or have minimal setup ?
Any existing frameworks or wrappers for UGC media pipelines that make all this easier ?

    


    I’ve spent weeks researching, testing APIs, and hitting walls — and while I’ve learned a lot, I’d really appreciate any guidance from folks who’ve been here before.
Thanks in advance 🙏

    


    And good luck to everyone else trying to build with AI on a budget — this stuff isn’t as plug-and-play as it looks on launch videos 💀