Recherche avancée

Médias (1)

Mot : - Tags -/sintel

Autres articles (100)

  • Multilang : améliorer l’interface pour les blocs multilingues

    18 février 2011, par

    Multilang est un plugin supplémentaire qui n’est pas activé par défaut lors de l’initialisation de MediaSPIP.
    Après son activation, une préconfiguration est mise en place automatiquement par MediaSPIP init permettant à la nouvelle fonctionnalité d’être automatiquement opérationnelle. Il n’est donc pas obligatoire de passer par une étape de configuration pour cela.

  • MediaSPIP 0.1 Beta version

    25 avril 2011, par

    MediaSPIP 0.1 beta is the first version of MediaSPIP proclaimed as "usable".
    The zip file provided here only contains the sources of MediaSPIP in its standalone version.
    To get a working installation, you must manually install all-software dependencies on the server.
    If you want to use this archive for an installation in "farm mode", you will also need to proceed to other manual (...)

  • Publier sur MédiaSpip

    13 juin 2013

    Puis-je poster des contenus à partir d’une tablette Ipad ?
    Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir

Sur d’autres sites (13308)

  • Build OpenCv using static ffmpeg library

    24 juillet 2020, par gaurav agarwal

    I am trying to build opencv for one of the commercial product. That commercial product already contains a static build of ffmpeg executable. While building OpenCv I am seeing bunch of failure most probably due to linking issues with ffmpeg.

    


    Is it possible to build OpenCv using static ffmpeg library ? In OpenCv installation doc I can see mentioned prerequisite is :

    


    


    ffmpeg or libav development packages : libavcodec-dev, libavformat-dev,
libswscale-dev

    


    


    I guess this means OpenCv needs these lib to be present as shared libs. Any expert opinion will be helpful !!!

    


  • IOException when trying to run ffmpeg command on Android

    20 avril 2020, par Rashid Faheem

    I am using this library through gradle with Android Studio 3.6.

    



    implementation 'com.github.hiteshsondhi88.libffmpeg:FFmpegAndroid:0.2.5'


    



    This is path of input and output file.

    



    String VideoIn = Environment.getExternalStorageDirectory().getAbsolutePath() + "/Download/Ab.mp4";
String VideoOut = Environment.getExternalStorageDirectory().getAbsolutePath() + "/Download/Ab1.avi";


    



    This is code where I load ffmpeg and try to execute command to convert an mp4 into avi. (Same command i use on windows to do this task). But it say IOException (No such file or directory) although input file is present in Download folder.

    



    private void convert(String VideoIn, String VideoOut) {
    ffmpeg = FFmpeg.getInstance(this);
    try {
        ffmpeg.loadBinary(new LoadBinaryResponseHandler() {
            @Override
            public void onFailure() {
                super.onFailure();
            }

            @Override
            public void onSuccess() {
                super.onSuccess();
            }

            @Override
            public void onStart() {
                super.onStart();
            }

            @Override
            public void onFinish() {
                super.onFinish();
            }
        });
    } catch (FFmpegNotSupportedException e) {
        Log.e(TAG, e.getMessage());
    }

    try {
        ffmpeg.execute("-i " + VideoIn + " " + VideoOut, new ExecuteBinaryResponseHandler() {
            @Override
            public void onSuccess(String message) {
                Log.e(TAG, message);
            }

            @Override
            public void onProgress(String message) {
                Log.e(TAG, message);
            }

            @Override
            public void onFailure(String message) {
                Log.e(TAG, message);
            }

            @Override
            public void onStart() {
                Log.e(TAG, "onStart called");
            }

            @Override
            public void onFinish() {
                Log.e(TAG, "onFinish called");
            }
        });
    } catch (FFmpegCommandAlreadyRunningException ex) {
        Log.e(TAG, ex.getMessage());
    }
}


    



    Here is log data for your to understand. I am not that expert in that so I couldn't figure out what is wrong with my code. Path is correct and file is there. Maybe I am calling this command wrong or what ?

    



    I searched google and stack overflow much but nothing helped me solve it. Any help would be appreciated.

    



    04-20 12:08:26.050 12696-12696/com.example.myapplication E/Video to Mp3: Input File Path: 
/storage/emulated/0/Download/Ab.mp4
04-20 12:08:26.050 12696-12696/com.example.myapplication E/Video to Mp3: Output File Path: 
/storage/emulated/0/Download/Ab1.avi
04-20 12:08:26.050 12696-12696/com.example.myapplication E/Video to Mp3: Device not supported
04-20 12:08:26.050 12696-12696/com.example.myapplication E/Video to Mp3: onStart called
04-20 12:08:26.084 12696-13027/com.example.myapplication E/FFmpeg: Exception while trying to run: 
/data/user/0/com.example.myapplication/files/ffmpeg -i /storage/emulated/0/Download/Ab.mp4 
/storage/emulated/0/Download/Ab1.avi
java.io.IOException: Error running exec(). Command: [/data/user/0/com.example.myapplication/files/ffmpeg, -i, /storage/emulated/0/Download/Ab.mp4, /storage/emulated/0/Download/Ab1.avi] Working Directory: null Environment: null
    at java.lang.ProcessManager.exec(ProcessManager.java:211)
    at java.lang.Runtime.exec(Runtime.java:186)
    at java.lang.Runtime.exec(Runtime.java:259)
    at java.lang.Runtime.exec(Runtime.java:202)
    at com.github.hiteshsondhi88.libffmpeg.ShellCommand.run(ShellCommand.java:10)
    at com.github.hiteshsondhi88.libffmpeg.FFmpegExecuteAsyncTask.doInBackground(FFmpegExecuteAsyncTask.java:38)
    at com.github.hiteshsondhi88.libffmpeg.FFmpegExecuteAsyncTask.doInBackground(FFmpegExecuteAsyncTask.java:10)
    at android.os.AsyncTask$2.call(AsyncTask.java:307)
    at java.util.concurrent.FutureTask.run(FutureTask.java:237)
    at android.os.AsyncTask$SerialExecutor$1.run(AsyncTask.java:246)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1113)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:588)
    at java.lang.Thread.run(Thread.java:833)
 Caused by: java.io.IOException: No such file or directory
    at java.lang.ProcessManager.exec(Native Method)
    at java.lang.ProcessManager.exec(ProcessManager.java:209)
    at java.lang.Runtime.exec(Runtime.java:186) 
    at java.lang.Runtime.exec(Runtime.java:259) 
    at java.lang.Runtime.exec(Runtime.java:202) 
    at com.github.hiteshsondhi88.libffmpeg.ShellCommand.run(ShellCommand.java:10) 
    at com.github.hiteshsondhi88.libffmpeg.FFmpegExecuteAsyncTask.doInBackground(FFmpegExecuteAsyncTask.java:38) 
    at com.github.hiteshsondhi88.libffmpeg.FFmpegExecuteAsyncTask.doInBackground(FFmpegExecuteAsyncTask.java:10) 
    at android.os.AsyncTask$2.call(AsyncTask.java:307) 
    at java.util.concurrent.FutureTask.run(FutureTask.java:237) 
    at android.os.AsyncTask$SerialExecutor$1.run(AsyncTask.java:246) 
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1113) 
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:588) 
    at java.lang.Thread.run(Thread.java:833) 
04-20 12:08:26.084 12696-12696/com.example.myapplication E/Video to Mp3: onFinish called


    


  • Using ffmpeg with an audio input from System.IO.Stream

    17 juillet 2020, par kevingoos

    So I am creating a project where I need to convert audio comming from azure text to speech.
    
I wanted to use ffmpeg for this, but I am not an expert with all these parameters.
So far I alread got it working where I read from a $.wav file and convert it to an output System.IO.Stream comming from ffmpeg.

    


    But I want to change it that I don't always save a wav file to disk. Instead now I get a System.IO.Stream from azure text to speech, but I have no clue for the params and the documentation is not really clear for me...

    


    var azureAudio = await new AzureSpeechService().Speak(text);

var psi = new ProcessStartInfo
{
    FileName = @"C:\ffmpeg-20200715-a54b367-win64-static\bin\ffmpeg.exe",
    Arguments = "-i pipe:0 -ac 2 -f s16le -ar 48000 pipe:1",
    RedirectStandardOutput = true,
    UseShellExecute = false
};
var ffmpeg = Process.Start(psi);

var inputStream = ffmpeg.StandardInput.BaseStream;

azureAudio.CopyTo(inputStream);

inputStream.Flush();
inputStream.Close();

var outputStream = ffmpeg.StandardOutput.BaseStream;
await SendAudioAsync(client, outputStream);