Recherche avancée

Médias (0)

Mot : - Tags -/api

Aucun média correspondant à vos critères n’est disponible sur le site.

Autres articles (74)

  • MediaSPIP version 0.1 Beta

    16 avril 2011, par

    MediaSPIP 0.1 beta est la première version de MediaSPIP décrétée comme "utilisable".
    Le fichier zip ici présent contient uniquement les sources de MediaSPIP en version standalone.
    Pour avoir une installation fonctionnelle, il est nécessaire d’installer manuellement l’ensemble des dépendances logicielles sur le serveur.
    Si vous souhaitez utiliser cette archive pour une installation en mode ferme, il vous faudra également procéder à d’autres modifications (...)

  • Amélioration de la version de base

    13 septembre 2013

    Jolie sélection multiple
    Le plugin Chosen permet d’améliorer l’ergonomie des champs de sélection multiple. Voir les deux images suivantes pour comparer.
    Il suffit pour cela d’activer le plugin Chosen (Configuration générale du site > Gestion des plugins), puis de configurer le plugin (Les squelettes > Chosen) en activant l’utilisation de Chosen dans le site public et en spécifiant les éléments de formulaires à améliorer, par exemple select[multiple] pour les listes à sélection multiple (...)

  • Mise à jour de la version 0.1 vers 0.2

    24 juin 2013, par

    Explications des différents changements notables lors du passage de la version 0.1 de MediaSPIP à la version 0.3. Quelles sont les nouveautés
    Au niveau des dépendances logicielles Utilisation des dernières versions de FFMpeg (>= v1.2.1) ; Installation des dépendances pour Smush ; Installation de MediaInfo et FFprobe pour la récupération des métadonnées ; On n’utilise plus ffmpeg2theora ; On n’installe plus flvtool2 au profit de flvtool++ ; On n’installe plus ffmpeg-php qui n’est plus maintenu au (...)

Sur d’autres sites (12498)

  • How to Output Mjpeg from Kokorin Jaffree FFmpeg via UDP to a Localhost Port ?

    14 octobre 2022, par roba

    I have a Java program which displays dual webcams and records them to file in FHD 30fps H264/H265. It uses Sarxos Webcam for the initial setup and display but when recording, it switches to Jaffree FFmpeg. During recording Sarxos Webcam must release its webcam access and cannot display while recording continues.

    


    I have tried recording with Xuggler/Sarxos but Sarxos seems to only access raw video from the webcams which creates limitations in the frame rate and resolution which can be achieved. At 1920x1080 the cameras can only deliver 5 fps raw video.

    


    I am trying to direct mjpeg streams from Jaffree to localports for display purposes during recording but I cannot figure out how to do it.

    


    Simultaneous recording plus sending to a port can be done from the terminal with the following :

    


    ffmpeg -f  dshow  -video_size 1920x1080 -rtbufsize 944640k -framerate 25 -vcodec mjpeg  -i video="Logitech Webcam C930e" -pix_fmt yuv420p -c:v libx264 outFHDx25.mp4 -f mpegts udp://localhost:1234?pkt_size=188&buffer_size=65535


    


    and viewed from the port in a different terminal like this :

    


    ffplay -i udp://localhost:1234


    


    The video which displays is a little blocky compared with the video recorded to file. Any suggestions on how to improve this would be appreciated.

    


    Note that FFPlay is not included in Jaffree FFMpeg.

    


    I would like to send the mjpeg to a port and then read it into the Sarxos Webcam viewer to display while recording is in progress.

    


    The Jaffree Java code for recording the output of one webcam to file follows. It takes the mjpeg/yuv422p output from the webcam and normally encodes it to file as H264/yuv420p :

    


    public static FFmpeg createTestFFmpeg() {
      String camera1Ref = "video=" + cam1Vid + ":audio=" + cam1Aud;
          return FFmpeg.atPath()
              .addArguments("-f", "dshow")  //selects dshow for Windows
              .addArguments("-video_size", resString)  //video resolution  eg 1920x1080          
              .addArguments("-rtbufsize", rtBufResultString) 
              .addArguments("-thread_queue_size", threadQ)
              .addArguments("-framerate", fpsString)   // capture frame rate  eg 30fps         
              .addArguments(codec, vidString)  //set capture encode mode from camera
              .addArgument(audio) //on or off
              .addArguments("-i", camera1Ref)   // name of camera to capture
              .addArguments("-pix_fmt", pixFmt)
              .addArguments("-c:v", enc2)  //eg enc2 = "libx264", "h264_nvenc"
              .addArguments(enc3, enc4)  //enc3 = "-crf", enc4 = "20"
              .addArguments(enc5, enc6)  //enc5 = "-gpu:v", enc6 = "0"
              .addArguments(enc7, enc8)  //enc7 = "-cq:v", enc8 = "20"
              .addArguments(enc9, enc10)  //enc9 = "-rc:v", enc10 = "vbr"
              .addArguments(enc11, enc12)  //enc11 = "-tune:v", enc12 = "ll"
              .addArguments(enc13, enc14)  //enc13 = "-preset:v", enc14 = "p1" 
              .addArguments(enc15,enc16)  //enc15 = "-b:v", enc16 = "0"
              .addArguments(enc17, enc18)  //enc17 = "-maxrate:v", enc18 = "5000k"
              .addArguments(enc19, enc20)  //enc19 = "-bufsize:v", enc20 = "5000k"
              .addArguments(enc21, enc22)  //enc21 = "-profile:v", enc22 = "main"
              .addArgument(noFFStats) //"-nostats"{, stops logging progress/statistics
              .addArguments("-loglevel", ffLogLevel)  //error logging
              .addArgument(bannerResultString)  // "-hide_banner"
              .addArguments("-rtbufsize", rtBufResultString) 
              .setOverwriteOutput(true)   // overwrite filename if it exists  Boolean = overwriteFile
              .addOutput(
                  UrlOutput
                      .toUrl(filePathL))                    
              .setProgressListener(new ProgressListener(){
                  @Override
                  public void onProgress(FFmpegProgress progress){
                     if(ffProgress){ 
                          System.out.println(progress);
                          
                     } 
                    }
            } );
            
   }


    


    How and where do I add the code to output mjpeg via UDP to a localport while simultaneously writing H264 to a file, and what is the syntax ? I am sure it must be simple but I seem to have tried all of the permutations without success. I can write to a file OR I can output to a port but I cannot do both.

    


  • avcodec/mediacodecdec : don't break out if both input and output port return try again

    19 octobre 2022, par Zhao Zhili
    avcodec/mediacodecdec : don't break out if both input and output port return try again
    

    At the beginning of decoding, if we feed mediacodec too fast, the
    input port will return try again. It takes some time for mediacodec
    to consume bitstream and output frame. So the output port also return
    try again. It possible that mediacodec_receive_frame doesn't consume
    any AVPacket and no AVFrame is output. Then both avcodec_send_packet()
    and avcodec_receive_frame() return EAGAIN, which shouldn't happen.

    This bug can be produced with decoding benchmark on Pixel 3.

    Signed-off-by : Zhao Zhili <zhilizhao@tencent.com>

    • [DH] libavcodec/mediacodecdec.c
  • which port ffmpeg is using for streaming to a rtmp(or rtmps) server running in a cloud location

    12 janvier 2023, par elsa

    I am working with ffmpeg and rtmp camera streaming in a constrained environment.&#xA;I have rtmp server running on a cloud (say eg. 11.23.123.13 in port 1935 for rtmp & 4936 for rtmps).&#xA;However is there any port on sending end , i should request team to open to avoid blocking the stream sending from device end to rtmp server end(camera streaming end). Only port 443 is open in device where camera connected and ffmpeg commands are running.

    &#xA;

    It would be helpful if somebody give insights on how network is connecting from ffmpeg encoder to rtmp server

    &#xA;

      &#xA;
    1. Getting the details how network is connecting from ffmpeg encoder to rtmp server ?
    2. &#xA;

    3. Which ports to be open in device end(where camera connecting) to start ffmpeg streaming to rtmp server running on cloud ?
    4. &#xA;

    &#xA;