Recherche avancée

Médias (0)

Mot : - Tags -/latitude

Aucun média correspondant à vos critères n’est disponible sur le site.

Autres articles (54)

  • Des sites réalisés avec MediaSPIP

    2 mai 2011, par

    Cette page présente quelques-uns des sites fonctionnant sous MediaSPIP.
    Vous pouvez bien entendu ajouter le votre grâce au formulaire en bas de page.

  • Les autorisations surchargées par les plugins

    27 avril 2010, par

    Mediaspip core
    autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs

  • Possibilité de déploiement en ferme

    12 avril 2011, par

    MediaSPIP peut être installé comme une ferme, avec un seul "noyau" hébergé sur un serveur dédié et utilisé par une multitude de sites différents.
    Cela permet, par exemple : de pouvoir partager les frais de mise en œuvre entre plusieurs projets / individus ; de pouvoir déployer rapidement une multitude de sites uniques ; d’éviter d’avoir à mettre l’ensemble des créations dans un fourre-tout numérique comme c’est le cas pour les grandes plate-formes tout public disséminées sur le (...)

Sur d’autres sites (5325)

  • How to make video from images using Java + x264 ; cross platform solution required

    19 octobre 2014, par Shashank Tulsyan

    I have made a software which records my entire day into a video.
    Example video : https://www.youtube.com/watch?v=ITZYMMcubdw (Note : >16hrs compressed in 2mins, video speed too high, might cause epilepsy :P )

    The approach that I use right now is, Avisynth + x264 + Java.
    This is very very efficient. The video for entire day is created in 3-4mins, and reduced to a size of 40-50MB. This is perfect, the only issue is that this solution is not cross platform.
    Does anyone have a better idea ?

    I tried using java based x246 libraries but

    1. They are slow as hell
    2. The video output size is too big
    3. The video quality is not satisfactory.

    Some website suggest a command such as :

    x264.exe --crf 18 --fps 24 --input-res 1920x1080 --input-csp rgb -o "T:\crf18.mkv" "T:\___BBB\big_buck_bunny_%05d.png"

    There are 2 problems with this approach.

    1. As far as I know, x264 does accept image sequence as input, ffmpeg does
    2. The input images are not named in sequence such as image01.png , image02.png etc. They are named as timestamp_as_longinteger.png . So inorder to allow x264 to accept these images as input, I have to rename all of them ( i make a symbolic link for all images in a new folder ). This approach is again unsatisfactory, because I need more flexibility in selecting/unselecting files which would be converted to a video. Right now my approach is a hack.

    The best solution is x264. But not sure how I can send it an image sequence from Java. That too, images which are not named in sequential fashion.


    BTW The purpose of making video is going back in time, and finding out how time was spend/wasted.
    The software is aware of what the user is doing. So using this I can find out (visually) how a class evolved with time. How much time I spend on a particular class/package/module/project/customer. The granuality right now is upto the class level, I wish to take it to the function level. The software is called jitendriya.

    Here is a sample graph


    Here is 1 solution
    How does one encode a series of images into H264 using the x264 C API ?

    But this is for C. If I have to do the same in java, and in a cross plaform fashion, I will have to resort to JNA/JNI. JNA might have a significant performance hit. JNI would be more work.
    FFMpeg also looks like a nice alternative, but I am still not satisfied by any of these solutions looking at the pros and cons.


    Solution Adapted.

    package weeklyvideomaker;

    import java.awt.AWTException;
    import java.awt.Rectangle;
    import java.awt.Robot;
    import java.awt.Toolkit;
    import java.awt.image.BufferedImage;
    import java.io.File;
    import java.io.IOException;
    import java.io.OutputStream;
    import java.util.Calendar;
    import java.util.LinkedList;
    import java.util.logging.Level;
    import java.util.logging.Logger;
    import neembuu.release1.util.StreamGobbler;
    import org.shashaank.activitymonitor.ScreenCaptureHandler;
    import org.shashaank.jitendriya.JitendriyaParams;

    /**
    *
    * @author Shashank
    */
    public class DirectVideoScreenHandler implements ScreenCaptureHandler {
       private final JitendriyaParams  jp;

       private String extension="264";
       private boolean lossless=false;
       private String fps="24/1";

       private Process p = null;
       private Rectangle r1;
       private Robot r;

       private int currentDay;

       private static final String[]weeks={"sun","mon","tue","wed","thu","fri","sat"};

       public DirectVideoScreenHandler(JitendriyaParams jp) {
           this.jp = jp;
       }

       public String getExtension() {
           return extension;
       }

       public void setExtension(String extension) {
           this.extension = extension;
       }

       public boolean isLossless() {
           return lossless;
       }

       public void setLossless(boolean lossless) {
           this.lossless = lossless;
       }

       public String getFps() {
           return fps;
       }

       public void setFps(String fps) {
           this.fps = fps;
       }

       private static int getday(){
           return Calendar.getInstance().get(Calendar.DAY_OF_WEEK) - 1;
       }

       public void make()throws IOException,AWTException{
           currentDay = getday();
           File week = jp.getWeekFolder();

           String destinationFile = week+"\\videos\\"+weeks[currentDay]+"_"+System.currentTimeMillis()+"_direct."+extension;

           r = new Robot();
           r1 = getScreenSize();

           ProcessBuilder pb = makeProcess(destinationFile, 0, r1.width, r1.height);

           p = pb.start();
           StreamGobbler out = new StreamGobbler(p.getInputStream(), "out");
           StreamGobbler err = new StreamGobbler(p.getErrorStream(), "err");
           out.start();err.start();
       }

       private static Rectangle getScreenSize(){
           return new Rectangle(Toolkit.getDefaultToolkit().getScreenSize());
       }

       private void screenShot(OutputStream os)throws IOException{        
           BufferedImage bi = r.createScreenCapture(r1);
           int[]intRawData = ((java.awt.image.DataBufferInt)
                   bi.getRaster().getDataBuffer()).getData();
           byte[]rawData = new byte[intRawData.length*3];
           for (int i = 0; i < intRawData.length; i++) {
               int rgb = intRawData[i];
               rawData[ i*3 + 0 ] = (byte) (rgb >> 16);
               rawData[ i*3 + 1 ] = (byte) (rgb >> 8);
               rawData[ i*3 + 2 ] = (byte) (rgb);
           }
           os.write(rawData);
       }

       private ProcessBuilder makeProcess(String destinationFile, int numberOfFrames,
               int width, int height){
           LinkedList<string> commands = new LinkedList&lt;>();
           commands.add("\""+encoderPath()+"\"");
           if(true){
               commands.add("-");
               if(lossless){
                   commands.add("--qp");
                   commands.add("0");
               }
               commands.add("--keyint");
               commands.add("240");
               commands.add("--sar");
               commands.add("1:1");
               commands.add("--output");
               commands.add("\""+destinationFile+"\"");
               if(numberOfFrames>0){
                   commands.add("--frames");
                   commands.add(String.valueOf(numberOfFrames));
               }else{
                   commands.add("--stitchable");
               }
               commands.add("--fps");
               commands.add(fps);
               commands.add("--input-res");
               commands.add(width+"x"+height);
               commands.add("--input-csp");
               commands.add("rgb");//i420
           }
           return new ProcessBuilder(commands);
       }

       private String encoderPath(){
           return jp.getToolsPath()+File.separatorChar+"x264_64.exe";
       }

       @Override public void run() {
           try {
               if(p==null){
                   make();
               }
               if(currentDay!=getday()){// day changed
                   destroy();
                   return;
               }
               if(!r1.equals(getScreenSize())){// screensize changed
                   destroy();
                   return;
               }
               screenShot(p.getOutputStream());
           } catch (Exception ex) {
               Logger.getLogger(DirectVideoScreenHandler.class.getName()).log(Level.SEVERE, null, ex);
           }
       }

       private void destroy()throws Exception{
           p.getOutputStream().flush();
           p.getOutputStream().close();
           p.destroy();
           p = null;
       }

    }
    </string>

    package weeklyvideomaker;

    import org.shashaank.jitendriya.JitendriyaParams;

    /**
    *
    * @author Shashank
    */
    public class DirectVideoScreenHandlerTest {
       public static void main(String[] args)throws Exception {
           JitendriyaParams  jp = new JitendriyaParams.Builder()
                   .setToolsPath("F:\\GeneralProjects\\JReminder\\development_environment\\tools")
                   .setOsDependentDataFolderPath("J:\\jt_data")
                   .build();
           DirectVideoScreenHandler w = new DirectVideoScreenHandler(jp);
           w.setExtension("264");
           w.setFps("24/1");
           w.setLossless(false);
           w.make();

           for (int i = 0; ; i++) {
               w.run();
               Thread.sleep(1000);
           }
       }
    }
  • Why does FFMPEG adds up extra bit rate to video ? [migrated]

    5 juin 2012, par user735647

    I use FFMPEG (command line Input) to convert my videos to a specific output format. The problem I am facing is when I try to pass a constant bit rate(700 kbps) to FFMPEG, the result is an output video with a different bit rate(say 1000 kbps). This phenomenon occurs invariably for all videos.Why is this happening ? I need to maintain a constant bit rate. Can anyone help me out.

    My FFMPEG version is 0.5

    The command line parameter which I am passing to FFMPEG is,

    -i {inputfile}
    -b 700k -ab 64k
    -vcodec libx264
    -acodec libfaac -ac 2 -ar 44100
    -y -s 320x240
    {outputfile}

    EDIT :

    I was able to force CBR with a fluctuation of 3% when I used the following parameters.

    ffmpeg -i myfile.avi
    -b 4000k -minrate 4000k
    -maxrate 4000k -bufsize 1835k   out.m2v

    But when I used -maxrate and - minrate along with my parameter set I was not able to force CBR. My parameter set is as follows,

    -i {inputfile}
    -b 1200k -minrate 1200k
    -maxrate 1200k -bufsize 1200k
    -ab 64k -vcodec libx264
    -acodec libfaac -ac 2 -ar 44100
    -y -s 320x240
    {outputfile}

    Why is this happening ?

  • Not able to encode image with ffmpeg

    29 juin 2012, par user1310596

    I am trying to encode one image with the help of ffmpeg library.. Actually i want to encode live video but for now i am starting with encoding an image..
    Here is my code for that..

       av_register_all();
       avcodec_init();
       avcodec_register_all();
       avformat_alloc_context();

       AVCodec *codec;
       AVCodecContext *ctx= NULL;
       int out_size, size, outbuf_size;
       AVFrame *picture;
       uint8_t *outbuf;
       unsigned char *flvdata = malloc(sizeof(unsigned char) * 30);


       outbuf_size = 100000;
       outbuf = malloc(outbuf_size);


       printf("Video encoding\n");

       codec = avcodec_find_encoder(CODEC_ID_FLV1);
       if (!codec) {
               fprintf(stderr, "codec not found\n");
               exit(1);
       }

       ctx= avcodec_alloc_context();
       picture= avcodec_alloc_frame();


       ctx->width = 320;
       ctx->height = 240;
       ctx -> sample_rate = 11025;
       ctx -> time_base.den = 1000;
       ctx -> time_base.num = 23976;
       ctx -> codec_id = CODEC_ID_FLV1;
       ctx -> codec_type = CODEC_TYPE_VIDEO;
       ctx->pix_fmt = PIX_FMT_YUV420P;

       if (avcodec_open(ctx, codec) &lt; 0) {
               fprintf(stderr, "could not open codec\n");
               exit(1);
       }

       outbuf_size = 100000;
       outbuf = malloc(outbuf_size);
       size = ctx->width * ctx->height;

       AVFrame* outpic = avcodec_alloc_frame();
       int nbytes = avpicture_get_size(PIX_FMT_YUV420P, ctx->width, ctx->height);

       uint8_t* outbuffer = (uint8_t*)av_malloc(nbytes);

       fflush(stdout);

       int numBytes = avpicture_get_size(PIX_FMT_YUV420P, ctx->width, ctx->height);

       UIImage *image = [UIImage imageNamed:[NSString stringWithFormat:@"0.jpg"]];
       CGImageRef newCgImage = [image CGImage];

       CGDataProviderRef dataProvider = CGImageGetDataProvider(newCgImage);
       CFDataRef bitmapData = CGDataProviderCopyData(dataProvider);
       long dataLength = CFDataGetLength(bitmapData);

       uint8_t *buffer = (uint8_t *)av_malloc(dataLength);
       buffer = (uint8_t *)CFDataGetBytePtr(bitmapData);

       for(int i = 0; i &lt; dataLength; i++)
       {
               if((i + 1) % 16 == 1 &amp;&amp; i != 1)
                       printf("\n");
               printf("%X\t",buffer[i]); // getting something different than the     actual hex value of the image
       }



       outpic -> pts = 0;        

       avpicture_fill((AVPicture*)picture, buffer, PIX_FMT_RGB8, ctx->width, ctx->height);

       avpicture_fill((AVPicture*)outpic, outbuffer, PIX_FMT_YUV420P, ctx->width, ctx->height);

       struct SwsContext* fooContext = sws_getContext(ctx->width, ctx->height,
                                                              PIX_FMT_RGB8,
                                                              ctx->width, ctx->height,
                                                              PIX_FMT_YUV420P,
                                                              SWS_FAST_BILINEAR, NULL, NULL, NULL);

       sws_scale(fooContext, picture->data, picture->linesize, 0, ctx->height, outpic->data, outpic->linesize);

       printf("abcdefghijklmnop");
       out_size = avcodec_encode_video(ctx, outbuf, outbuf_size, outpic);
       printf("\n\n out_size %d   outbuf_size %d",out_size,outbuf_size);

    so the size of image is 29 kb and the avcodec_encode_video is returning 20143.. So is it correct ? I mean i am encoding so its size should also decrease... And i have opened that image in hex mode and i am getting different data than that is in buffer (as shown in the code). so i think that buffer is not getting correct data. Right ?
    Can anyone please help me with my code ?
    Thank you in advance...