
Recherche avancée
Médias (1)
-
Bug de détection d’ogg
22 mars 2013, par
Mis à jour : Avril 2013
Langue : français
Type : Video
Autres articles (81)
-
Les autorisations surchargées par les plugins
27 avril 2010, parMediaspip core
autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs -
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir -
HTML5 audio and video support
13 avril 2011, parMediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
For older browsers the Flowplayer flash fallback is used.
MediaSPIP allows for media playback on major mobile platforms with the above (...)
Sur d’autres sites (9649)
-
errors while building ffmpeg with ndk16b standalone toolchain
11 octobre 2018, par soni sahuI am trying to build ffmpeg with android ndk18b toolchain. My configuration looks like this.
TOOLCHAIN_PREFIX=/home/git/ndks/stand_alone_toolchain_ndk16b
NDK_SYSROOT=$TOOLCHAIN_PREFIX/sysroot
NDK_ARCH='arm'
NDK_ABIARCH='arm-linux-androideabi'
target_host=arm-linux-androideabi
export CPP="${CROSS_PREFIX}clang++"
export PATH=${TOOLCHAIN_PREFIX}/bin:$PATH
export CROSS_PREFIX=${TOOLCHAIN_PREFIX}/bin/${NDK_ABIARCH}-
export CC="${CROSS_PREFIX}clang"
export CXX="${CROSS_PREFIX}clang++"
export AS="${CROSS_PREFIX}clang"
export AR="${CROSS_PREFIX}ar"
export LD="${CROSS_PREFIX}ld"
export RANLIB="${CROSS_PREFIX}ranlib"
export STRIP="${CROSS_PREFIX}strip"
export OBJDUMP="${CROSS_PREFIX}objdump"
export CPP="${CROSS_PREFIX}cpp"
export GCONV="${CROSS_PREFIX}gconv"
export NM="${CROSS_PREFIX}nm"
export SIZE="${CROSS_PREFIX}size"
# Tell configure what flags Android requires.
export CFLAGS="-fPIE -fPIC"
export LDFLAGS="-pie"
./configure \
--cpu="armv7-a" \
--enable-pic \
--disable-runtime-cpudetect \
--enable-pthreads \
--enable-hardcoded-tables \
--prefix=$PREFIX \
--disable-doc \
--disable-ffplay \
--disable-ffprobe \
--disable-ffserver \
--disable-doc \
--disable-network \
--enable-libmp3lame \
--enable-libx264 \
--enable-gpl \
--extra-ldflags="-latomic -L${NDK_SYSROOT}/usr/lib -L$TOOLCHAIN_PREFIX/arm-linux-androideabi/lib -L$TOOLCHAIN_PREFIX/lib -Llibmp3lame/lib -Lx264/android/armeabi-v7a/lib $LDFLAGS -v -lc -lm -ldl -llog -march=armv7-a"\
--extra-cflags="-ffast-math -funroll-loops -mfloat-abi=softfp -mfpu=vfpv3-d16 -Ilibmp3lame/include -Ix264/android/armeabi-v7a/include $CFLAGS -march=armv7-a -marm -mfloat-abi=softfp -mfpu=neon -mtune=cortex-a8 -mthumb -D__thumb__ -fno-exceptions -fno-rtti -march=armv7-a -Wl,--fix-cortex-a8" \
--arch=arm\
--target-os=linux \
--enable-cross-compile \
--cross-prefix=$TOOLCHAIN_PREFIX/bin/arm-linux-androideabi- \
--nm=${NM} \
--cc=${CC} \
--cxx=${CXX} \
--ld=${LD} \
--ar=${AR} \
--as=${AS} \
--strip=${STRIP}I have build libmp3lame using this http://developer.samsung.com/android/technical-docs/Porting-and-using-LAME-MP3-on-Android-with-JNI and placed inside libmp3lame/lib folder.
I am getting these undefined errors in config.log.
/home/git/ndks/stand_alone_toolchain_ndk18b/bin/arm-linux-androideabi-clang
-D_ISOC99_SOURCE -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE -Dstrtod=avpriv_strtod -DPIC -fPIE -fPIC -ffast-math -funroll-loops -mfloat-abi=softfp -mfpu=vfpv3-d16 -Ilibmp3lame/include -Ix264/android/armeabi-v7a/include -fPIE -fPIC -march=armv7-a -marm -mfloat-abi=softfp -mfpu=neon -mtune=cortex-a8 -mthumb -D__thumb__ -fno-exceptions -fno-rtti -march=armv7-a -Wl,—fix-cortex-a8 -march=armv7-a -std=c11 -fomit-frame-pointer -fPIC -mthumb -c -o /tmp/ffconf.PT1BY4UI.o /tmp/ffconf.5z69vrhC.c clang70 : warning :
-Wl,—fix-cortex-a8 : ’linker’ input unused [-Wunused-command-line-argument]
/home/git/ndks/stand_alone_toolchain_ndk18b/bin/arm-linux-androideabi-ld
-pie -latomic -L/home/git/ndks/stand_alone_toolchain_ndk18b/sysroot/usr/lib -L/home/git/ndks/stand_alone_toolchain_ndk18b/arm-linux-androideabi/lib
-L/home/git/ndks/stand_alone_toolchain_ndk18b/lib -Llibmp3lame/lib -Lx264/android/armeabi-v7a/lib -pie -v -lc -lm -ldl -llog -march=armv7-a -o /tmp/ffconf.o9IsP7bS /tmp/ffconf.PT1BY4UI.o -lmp3lame -lm -lz libmp3lame/lib/libmp3lame.a(set_get.o)(.ARM.exidx.text.lame_set_num_samples+0x0) :
error : undefined reference to ’__aeabi_unwind_cpp_pr0’
libmp3lame/lib/libmp3lame.a(set_get.o)(.ARM.exidx.text.lame_set_ogg+0x0) :
error : undefined reference to ’__aeabi_unwind_cpp_pr0’
libmp3lame/lib/libmp3lame.a(set_get.o)(.ARM.exidx.text.lame_get_ogg+0x0) :
error : undefined reference to ’__aeabi_unwind_cpp_pr0’
libmp3lame/lib/libmp3lame.a(set_get.o)(.ARM.exidx.text.lame_set_quality+0x0) :
error : undefined reference to ’__aeabi_unwind_cpp_pr0’
D :/work_dir/battefield3/chimpoon/proj.android/app/jni/../../../../lame/lame-3.100/./libmp3lame/set_get.c:2161 :
error : undefined reference to ’__aeabi_uidiv’
D :/work_dir/battefield3/chimpoon/proj.android/app/jni/../../../../lame/lame-3.100/./libmp3lame/set_get.c:2165 :
error : undefined reference to ’__aeabi_uidivmod’
D :/work_dir/battefield3/chimpoon/proj.android/app/jni/../../../../lame/lame-3.100/./libmp3lame/set_get.c:2170 :
error : undefined reference to ’__aeabi_uidiv’Please help me to resolve this issue. Thanks.
-
Writing Live-Multimedia-Application using OpenGL & Co. saving output to disc [closed]
21 janvier 2013, par user1997286I want to write an application that does the following thing :
- Getting Commands via ArtNET (DMX over Ethernet, a Control Protocol) for each object (called Layer)
- each Layer could be one of the following : Live Camera Stream, Movie, Image
- each layer could be translated, rotated or stretched
- on each layer I can set filters (Like a Kaleidoscope Effect, Blur, Color Correction, etc.)
- the rsulting video-stream is in the 3d-space
- I want to display each part of the image on one Projector (in total up to 3 ones) using a TripleHead2GO (3 Projectors display a different region of my DVI-Output). Each Projecector-Image should have own Soft-Edge and Keystone parameters.
- the resulting image will also be shown on a Preview-Screen with some Information overlay.
I think all that should be possible with opengl and openal (for the movie audio)
I think I'll use C++, OpenGL for Graphics, OpenAL for Audio, if needed ffmpeg for Video conversion, Ubuntu/Debian as OS.
The software is used to do Multimedia-Shows on Concerts including Cameras & Co.
All that should happen Live (On a FullHD output), Having i7 3770, GLX 670 and 16GB of Ram for at least 8 Layers. (4 Live-Images at once + Some Overlays like the Actors Name and some Logos)
But now comes the question.
Is it also Posible to do the following with that setting :
- Writing the output Image with all the 3d translations to a Movie File (To Master a DVD later) with Audio
- Mixing Audio from different Inputs & Files (Ambience Mics, Signal from the Sound Mixer, Playbacks from my own application) to more than one Mix (eg. one Mix for the Recording, one Mix for Live)
- Stream that Output Complete or in Parts (e.g. the left Part of the Image) over the Network (For example, Projector 1 is near the Server, so I connect it using DVI, Projector 2+3 is connected to a Computer that receives the streams for that two projectors (with soft edge on each stream) and Screen 4 is outside the Concert Hall and shows the complete Live-Stream.
- What GUI-Framework should I use for that ?
- is it perhaps event performant enough to use Java for that ?
- is it posible to use that mechanism for just rendering (eg. I have stored the cut points on Disc and saved every single camera stream to change some errors later or cut out some parts)
-
Can FFMPEG be used to change the transparency of a single PNG ?
19 juillet 2020, par ChameleonI know that FFMPEG can manipulate transparency using fade in/fade out over a series of frames. I just haven't found a way to generate a png with a specified transparency setting.


I am in the process of creating a lot (20 in first set) of procedurally-generated videos. Each video requires several starts and stops, where with each stop, a png overlay describing what is seen will be displayed for several (actual number TBD, and may vary) frames. I'd like each overlay to fade in over x frames, display for y frames at full opacity and then fade out over y frames. The documentation on FFMPEG really sucks at explaining fade effects, and no one I've found showing usage actually explains what all of the parameters do.


The original sources for the videos are CGI png frames with a transparent background. I will be making the descriptive overlays with transparent backgrounds (same resolution as the CGI frames) - I'm really hoping to not have to manually save each overlay (8 overlays per video at a minimum) with 4 to 6 (or more) transparency settings - The company is being real indecisive as to the length of the fade in/out or the "hang time" of the descriptive overlays.


I have already created a Python script that uses FFMPEG to place overlays on specific frames, then uses FFMPEG to stitch all frames into a single video, adds a background to the video and then places the company's watermark on the video. It already manages the key frames and what images should be overlaid (and for how many frames). It is working well, but the company doesn't like the lack of fade. I already have the framework in place to manage the fade in/out duration (I'm just missing the answer to this question). I created the script because I have no doubt that I'll have to generate the final output on the first couple of videos a number of times to appease the stakeholders.


I'm really hoping to find a fairly easy way to hand FFMPEG a png with a transparent background and have FFMPEG hand me a copy of the image whose non-transparent part is now "x percent" transparent (or "y percent" opaque). I know from failed attempts that FFMPEG doesn't choke trying to make a region more than 100% transparent.


It's a real pain to get approval to install new software on the workstation, so I'm not actually interested in any suggestion that doesn't use FFMPEG or a pretty vanilla Python installation. - It's not that the other software might not be useful, it just means that if other software is needed, I'll have to manually create the frames.