Recherche avancée

Médias (10)

Mot : - Tags -/wav

Autres articles (75)

  • La file d’attente de SPIPmotion

    28 novembre 2010, par

    Une file d’attente stockée dans la base de donnée
    Lors de son installation, SPIPmotion crée une nouvelle table dans la base de donnée intitulée spip_spipmotion_attentes.
    Cette nouvelle table est constituée des champs suivants : id_spipmotion_attente, l’identifiant numérique unique de la tâche à traiter ; id_document, l’identifiant numérique du document original à encoder ; id_objet l’identifiant unique de l’objet auquel le document encodé devra être attaché automatiquement ; objet, le type d’objet auquel (...)

  • Websites made ​​with MediaSPIP

    2 mai 2011, par

    This page lists some websites based on MediaSPIP.

  • Participer à sa traduction

    10 avril 2011

    Vous pouvez nous aider à améliorer les locutions utilisées dans le logiciel ou à traduire celui-ci dans n’importe qu’elle nouvelle langue permettant sa diffusion à de nouvelles communautés linguistiques.
    Pour ce faire, on utilise l’interface de traduction de SPIP où l’ensemble des modules de langue de MediaSPIP sont à disposition. ll vous suffit de vous inscrire sur la liste de discussion des traducteurs pour demander plus d’informations.
    Actuellement MediaSPIP n’est disponible qu’en français et (...)

Sur d’autres sites (9063)

  • How can combine two separate scripts being piped together to make one script instead of two ?

    27 mars 2016, par user556068

    For the past couple hours I’ve been banging my head against the wall trying to figure out something I thought would be simple. Maybe it is but it’s beyond me at the moment. So I have now two scripts. Originallly they were part of the same but I could never make it work how it should. So the first part uses curl to download a file from a site. Then using grep and sedto filter out the text I need which is then put into a plain text file as a long list of website urls ; one per line. The last part of the 1st script calls on youtube -dl to read the batch file in order to obtain the web addresses where the actual content is located. I hope that makes sense.

    youtube-dl reads the batch file and outputs a new list urls into the terminal. This second list is not saved to file because it doesn’t need to be. These urls change from day to day or hour to hour. Using the read command, these urls are then passed to ffmpeg using a predetermined set of arguments for the input and output. Ffmpeg is executed on every url it receives and runs quietly in the background.

    The first paragraph describes script1.sh and paragraph 2 obviously describes script2.sh. When I pipe them together like script1.sh | script2.sh it works better than I ever thought possible. Maybe i’m nitpicking at this point but the idea is to have 1 unified script. For the moment I have simplified it by adding an alias to my .bash_profile.

    Here are the last two commands of script1.

    sed 's/\"\,/\//g' > "$HOME/file2.txt";
    cat $HOME/file2.txt | youtube-dl --ignore-config -iga -

    The trailing - allows youtube-dl to read from stdin.

    The second part of the script ; what I’m calling script2 at this point begins with

    while read -r input
    do
    ffmpeg [arg] [input] [arg2] [output]

    What am i not seeing that is causing the script to hang when the two halves are combined yet work perfectly if one is piped into the other ?

  • How can I run command line FFMPEG and accept multiple pipes (video and audio) without blocking on the first input ?

    18 février 2016, par Version135b

    I’m trying to mux h264 and aac created with MediaCodec using FFMPEG, and also use FFMPEG’s RTMP support to send to youtube. I’ve created two pipes, and am writing from java (android) through WriteableByteChannels. I can send to one pipe just fine (accepting null audio) like this :

    ./ffmpeg -f lavfi -i aevalsrc=0 -i "files/camera-test.h264" -acodec aac -vcodec copy -bufsize 512k -f flv "rtmp://a.rtmp.youtube.com/live2/XXXX"

    YouTube streaming works perfectly (but I have no audio). Using two pipes this is my command :

    ./ffmpeg \
    -i "files/camera-test.h264" \
    -i "files/audio-test.aac" \
    -vcodec copy \
    -acodec copy \
    -map 0:v:0 -map 1:a:0 \
    -f flv "rtmp://a.rtmp.youtube.com/live2/XXXX""

    The pipes are created with mkfifo , and opened from java like this :

    pipeWriterVideo = Channels.newChannel(new FileOutputStream(outputFileVideo.toString()));

    The order of execution (for now in my test phase) is creation of the files, starting ffmpeg (through adb shell) and then starting recording which opens the channels. ffmpeg will immediately open the h264 stream and then wait, since it is reading from the pipe the first channel open (for video) will successfully run. When it comes to trying to open the audio the same way, it fails because ffmpeg has not actually started reading from the pipe. I can open a second terminal window and cat the audio file and my app spits out what i hope is encoded aac, but ffmpeg fails, usually just sitting there waiting. Here is the verbose output :

    ffmpeg version N-78385-g855d9d2 Copyright (c) 2000-2016 the FFmpeg
    developers
     built with gcc 4.8 (GCC)
     configuration: --prefix=/home/dev/svn/android-ffmpeg-with-rtmp/src/ffmpeg/android/arm
       --enable-shared --disable-static --disable-doc --disable-ffplay
       --disable-ffprobe --disable-ffserver --disable-symver
       --cross-prefix=/home/dev/dev/android-ndk-r10e/toolchains/arm-linux-androideabi-4.8/prebuilt/linux-x86_64/bin/arm-linux-androideabi-
       --target-os=linux --arch=arm --enable-cross-compile
       --enable-librtmp --enable-pic --enable-decoder=h264
       --sysroot=/home/dev/dev/android-ndk-r10e/platforms/android-19/arch-arm
       --extra-cflags='-Os -fpic -marm'
       --extra-ldflags='-L/home/dev/svn/android-ffmpeg-with-rtmp/src/openssl-android/libs/armeabi '
       --extra-ldexeflags=-pie --pkg-config=/usr/bin/pkg-config
     libavutil      55. 17.100 / 55. 17.100
     libavcodec     57. 24.102 / 57. 24.102
     libavformat    57. 25.100 / 57. 25.100
     libavdevice    57.  0.101 / 57.  0.101
     libavfilter     6. 31.100 /  6. 31.100
     libswscale      4.  0.100 /  4.  0.100
     libswresample   2.  0.101 /  2.  0.101
    matched as AVOption 'debug' with argument 'verbose'.
    Trailing options were found on the commandline.
    Finished splitting the commandline.
    Parsing a group of options: global .
    Applying option async (audio sync method) with argument 1.
    Successfully parsed a group of options.
    Parsing a group of options: input file files/camera-test.h264.
    Successfully parsed a group of options.
    Opening an input file: files/camera-test.h264.
    [file @ 0xb503b100] Setting default whitelist 'file'

    I think if I could just get ffmpeg to start listening to both pipes, the rest would work out !

    Thanks for your time.

    EDIT :
    I’ve made progress by decoupling the audio pipe connection and encoding, but now as soon as the video stream has been passed it errors on audio. I started a separate thread to create the WriteableByteChannel for audio and it never gets passed the FileOutputStream creation.

    matched as AVOption 'debug' with argument 'verbose'.
    Trailing options were found on the commandline.
    Finished splitting the commandline.
    Parsing a group of options: global .
    Successfully parsed a group of options.
    Parsing a group of options: input file files/camera-test.h264.
    Successfully parsed a group of options.
    Opening an input file: files/camera-test.h264.
    [file @ 0xb503b100] Setting default whitelist 'file'
    [h264 @ 0xb503c400] Format h264 probed with size=2048 and score=51
    [h264 @ 0xb503c400] Before avformat_find_stream_info() pos: 0 bytes read:15719 seeks:0
    [h264 @ 0xb5027400] Current profile doesn't provide more RBSP data in PPS, skipping
    [h264 @ 0xb503c400] max_analyze_duration 5000000 reached at 5000000 microseconds st:0
    [h264 @ 0xb503c400] After avformat_find_stream_info() pos: 545242 bytes read:546928 seeks:0 frames:127
    Input #0, h264, from 'files/camera-test.h264':
     Duration: N/A, bitrate: N/A
       Stream #0:0, 127, 1/1200000: Video: h264 (Baseline), 1 reference frame, yuv420p(left), 854x480 (864x480), 1/50, 25 fps, 25 tbr, 1200k tbn, 50 tbc
    Successfully opened the file.
    Parsing a group of options: input file files/audio-test.aac.
    Applying option vcodec (force video codec ('copy' to copy stream)) with argument copy.
    Successfully parsed a group of options.
    Opening an input file: files/audio-test.aac.
    Unknown decoder 'copy'
    [AVIOContext @ 0xb5054020] Statistics: 546928 bytes read, 0 seeks

    Here is where I attempt to open the audio pipe.

    new Thread(){
        public void run(){
             Log.d("Audio", "pre thread");
             FileOutputStream fs = null;
             try {
                  fs = new FileOutputStream("/data/data/android.com.android.grafika/files/audio-test.aac");
             } catch (FileNotFoundException e) {
                  e.printStackTrace();
             }
             Log.d("Audio", "made fileoutputstream");  //never hits here
             mVideoEncoder.pipeWriterAudio = Channels.newChannel(fs);
             Log.d("Audio", "made it past opening audio pipe");
        }
    }.start();

    Thanks.

  • Compiling FFmpeg : libx264 not found

    17 mars 2016, par SaidTagnit

    I hope some one helps me the solve this problems. I was trying to compile FFmpeg 2.2.3 library under ubuntu 12.04LTS for android using android ndk r10e by following this tutorials :

    Tutorial1

    Tutorial2

    here is my build_android.sh file :

    #!/bin/bash
    NDK=/home/rango/Desktop/android-ndk-r10e
    SYSROOT=$NDK/platforms/android-19/arch-arm/
    TOOLCHAIN=$NDK/toolchains/arm-linux-androideabi-4.8/prebuilt/linux-x86

    #ADDI_LDFLAGS="-L /usr/local/lib"
    #ADDI_CFLAGS="-I /usr/include"

    function build_one
    {
    ./configure \
    --prefix=$PREFIX \
    --enable-shared \
    --enable-static \
    --disable-doc \
    --disable-ffmpeg \
    --disable-ffplay \
    --disable-ffprobe \
    --disable-ffserver \
    --disable-doc \
    --disable-symver \
    --enable-protocol=concat \
    --enable-protocol=file \
    --enable-muxer=mp4 \
    --enable-demuxer=mpegts \
    --enable-memalign-hack \
    --enable-gpl \
    --enable-libx264 \
    --cross-prefix=$TOOLCHAIN/bin/arm-linux-androideabi- \
    --target-os=linux \
    --arch=arm \
    --enable-cross-compile \
    --sysroot=$SYSROOT \
    --extra-cflags="-Os -fpic -marm $ADDI_CFLAGS" \
    --extra-ldflags="$ADDI_LDFLAGS" \

    make clean
    make -j4
    make install
    }

    CPU=arm
    PREFIX=$(pwd)/android/$CPU

    build_one

    when i execute buid_android.sh script without —enable-libx264 \ line everything is going well and i can get .a files on android/arm/ folder. but with this line it fails and show the following error in the console :

    ERROR : libx264 not found

    If you think configure made a mistake, make sure you are using the
    latest version from Git. If the latest version fails, report the
    problem to the ffmpeg-user@ffmpeg.org mailing list or IRC #ffmpeg on
    irc.freenode.net. Include the log file "config.log" produced by
    configure as this will help

    it seems like it can’t find where the libraries are, as i get a lot of the following errors in config.log file :

    "LIBNAME".h : No such file or directory
    newlib.h : No such file or
    directory mingw.h : No such file or directory
    x264.h : No such file or
    directory

    Here is the tail of config.log file :

    check_mathfunc truncf 1
    check_ld cc
    check_cc
    BEGIN /tmp/ffconf.zGKqGin6.c
       1   #include
       2   float foo(float f, float g) { return truncf(f); }
       3   int main(void){ return (int) foo; }
    END /tmp/ffconf.zGKqGin6.c
    /home/rango/Desktop/android-ndk-r10e/toolchains/arm-linux-androideabi-4.8/prebuilt/linux-x86/bin/arm-linux-androideabi-gcc --sysroot=/home/rango/Desktop/android-ndk-r10e/platforms/android-19/arch-arm/ -isysroot /home/rango/Desktop/android-ndk-r10e/platforms/android-19/arch-arm/ -D_ISOC99_SOURCE -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE -Dstrtod=avpriv_strtod -DPIC -Os -fpic -marm -march=armv5te -std=c99 -fomit-frame-pointer -fPIC -marm -pthread -c -o /tmp/ffconf.8Q9ke3aO.o /tmp/ffconf.zGKqGin6.c
    /home/rango/Desktop/android-ndk-r10e/toolchains/arm-linux-androideabi-4.8/prebuilt/linux-x86/bin/arm-linux-androideabi-gcc --sysroot=/home/rango/Desktop/android-ndk-r10e/platforms/android-19/arch-arm/ -isysroot /home/rango/Desktop/android-ndk-r10e/platforms/android-19/arch-arm/ -Wl,--as-needed -o /tmp/ffconf.3sjTkc5z /tmp/ffconf.8Q9ke3aO.o -lm -lz -pthread
    check_lib x264.h x264_encoder_encode -lx264
    check_header x264.h
    check_cpp
    BEGIN /tmp/ffconf.zGKqGin6.c
       1   #include
       2   int x;
    END /tmp/ffconf.zGKqGin6.c
    /home/rango/Desktop/android-ndk-r10e/toolchains/arm-linux-androideabi-4.8/prebuilt/linux-x86/bin/arm-linux-androideabi-gcc --sysroot=/home/rango/Desktop/android-ndk-r10e/platforms/android-19/arch-arm/ -isysroot /home/rango/Desktop/android-ndk-r10e/platforms/android-19/arch-arm/ -D_ISOC99_SOURCE -D_FILE_OFFSET_BITS=64 -D_LARGEFILE_SOURCE -Dstrtod=avpriv_strtod -DPIC -Os -fpic -marm -march=armv5te -std=c99 -fomit-frame-pointer -fPIC -marm -pthread -E -o /tmp/ffconf.8Q9ke3aO.o /tmp/ffconf.zGKqGin6.c
    /tmp/ffconf.zGKqGin6.c:1:18: fatal error: x264.h: No such file or directory
    #include
                     ^
    compilation terminated.
    ERROR: libx264 not found