Recherche avancée

Médias (1)

Mot : - Tags -/iphone

Autres articles (97)

  • MediaSPIP 0.1 Beta version

    25 avril 2011, par

    MediaSPIP 0.1 beta is the first version of MediaSPIP proclaimed as "usable".
    The zip file provided here only contains the sources of MediaSPIP in its standalone version.
    To get a working installation, you must manually install all-software dependencies on the server.
    If you want to use this archive for an installation in "farm mode", you will also need to proceed to other manual (...)

  • HTML5 audio and video support

    13 avril 2011, par

    MediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
    The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
    For older browsers the Flowplayer flash fallback is used.
    MediaSPIP allows for media playback on major mobile platforms with the above (...)

  • ANNEXE : Les plugins utilisés spécifiquement pour la ferme

    5 mars 2010, par

    Le site central/maître de la ferme a besoin d’utiliser plusieurs plugins supplémentaires vis à vis des canaux pour son bon fonctionnement. le plugin Gestion de la mutualisation ; le plugin inscription3 pour gérer les inscriptions et les demandes de création d’instance de mutualisation dès l’inscription des utilisateurs ; le plugin verifier qui fournit une API de vérification des champs (utilisé par inscription3) ; le plugin champs extras v2 nécessité par inscription3 (...)

Sur d’autres sites (7965)

  • mp4 video file with gps location

    4 mars 2017, par Charles Chay

    I had a understanding that there is no official support to insert gps location metadata into .mp4 files. However, I see a gps location metadata inserted into the video file while I was using my Sony Z2 with geotagging on. ffmpeg was able to detect the metadata after exporting the same file to my laptop. Anyone knows if there a new interface method or api declared for such purpose in andriod ndk ?

  • YUY2 image ==>>sws_scale ==>>x264_encoder_encode doesn't work in Windows

    8 décembre 2011, par shiju sasi

    I have a multi media app in Windows using x264 built using MSYS-MingW and ffmpeg Windows binaries. This works for most of the cameras which capture data in RGB24 and RGB32 formats in most of the OSes. But when I tested the app on a Windows 7 (64 bit OS) Sony Vaio Laptop which has an integrated webcam capturing in YUY2 format, the x264_enoder_encode crashes. The sws_scale to convert the YUY2 data to YUV420 Planar any way works fine here too and returns proper stride values. Please check the relevant code fragments that I have attached below.

        x264_param_apply_profile(&m_param, "baseline");
          m_pEncoder = x264_encoder_open(&m_param);
          x264_encoder_parameters(m_pEncoder,&m_param);

          m_encoderConvertCtx = sws_getContext(g_iWidth, g_iHeight, PIX_FMT_YUYV422, SCALE_WIDTH, SCALE_HEIGHT, PIX_FMT_YUV420P, SWS_BICUBIC, NULL, NULL, NULL);

          x264_picture_t m_pic_in, m_pic_out;   //X264 picture variables to get the X264 encoded picture out.
          x264_picture_init(&m_pic_in);
    m_pic_in.i_type = X264_CSP_I420;

    x264_nal_t*  m_nals;


    srcstride = g_iWidth * 2;  //For YUYV422 Packed

    AVFrame* pictIn;
    AVFrame* pictOut;

    int iInBytes = avpicture_get_size(PIX_FMT_YUV420P, SCALE_WIDTH, SCALE_HEIGHT);
    uint8_t* outbuffer = (uint8_t*)av_malloc(iInBytes);
    pictOut = avcodec_alloc_frame();

    avpicture_fill((AVPicture*)pictOut, outbuffer, PIX_FMT_YUV420P, SCALE_WIDTH, SCALE_HEIGHT);

    sws_scale(m_encoderConvertCtx, &in_buf, &srcstride, 0, g_iHeight, pictOut->data, pictOut->linesize); //Scale from YUYV422 Packed to YUV420 Plane


    ///Code after Scale begins

    memcpy(m_pic_in.img.plane[0],pictOut->data[0],SCALE_WIDTH * SCALE_HEIGHT);
    memcpy(m_pic_in.img.plane[1],pictOut->data[1],SCALE_WIDTH * SCALE_HEIGHT/4);
    memcpy(m_pic_in.img.plane[2],pictOut->data[2],SCALE_WIDTH * SCALE_HEIGHT/4);
    m_pic_in.img.plane[3] = 0;

    for(int iPlane = 0; iPlane < 3; iPlane++)
    {
    m_pic_in.img.i_stride[iPlane] = pictOut->linesize[iPlane];
    }
    m_pic_in.img.i_stride[3] = 0;

    int frame_size = x264_encoder_encode(m_pEncoder, &m_nals, &i_nals, &m_pic_in, &m_pic_out);

    Please help if possible, as this has been consuming a lot of time at my end. But I am not able to dig in to the library side for debugging. Any experienced hands are requested to assist.

  • MPEG-TS Segments HTTP Live Streaming

    5 juin 2013, par user1069624

    I'm trying to interleave MPEG-TS segments but failing. One set of segments was actually captured using the built in camera in the laptop, then encoded using FFMPEG with the following command :

    ffmpeg -er 4 -y -f video4linux2 -s 640x480 -r 30 -i %s -isync -f mpegts -acodec libmp3lame -ar 48000 -ab 64k -s 640x480 -vcodec libx264 -fflags +genpts -b 386k -coder 0 -me_range 16 -keyint_min 25 -i_qfactor 0.71 -bt 386k -maxrate 386k -bufsize 386k -qcomp 0.6 -qmin 10 -qmax 51 -qdiff 4 -aspect 640:480

    And the other one is an avi file that was encoded using the following command :

    fmpeg -er 4 -y -f avi -s 640x480 -r 30 -i ./DSCF2021.AVI -vbsf dump_extra -f mpegts -acodec libmp3lame -ar 48000 -ab 64k -s 640x480 -vcodec libx264 -fflags +genpts -b 386k -coder 0 -me_range 16 -keyint_min 25 -i_qfactor 0.71 -bt 386k -maxrate 386k -bufsize 386k -qcomp 0.6 -qmin 10 -qmax 51 -qdiff 4 -aspect 640:480

    Then the output is segmented into ts segments using an open source segmenter.

    If both come from the same source (both from the camera) they work fine. However in this case, the second set of segments freeze. Time passes, but the video does not move..
    So i think it's an encoding problem. So my question is, how should i change the ffmpeg command for this to work ?

    By interleave I mean, having a playlist with the first set of segments, and another playlist with the other set of segments, and having the client call one then the other (HTTP Live Streaming)

    The ffprobe output of one of the first set of segments :

    Input #0, mpegts, from 'live1.ts':
     Duration: 00:00:09.76, start: 1.400000, bitrate: 281 kb/s
     Program 1 Service01
       Metadata:
         name            : Service01
         provider_name   : FFmpeg
       Stream #0.0[0x100]: Video: h264, yuv420p, 640x480 [PAR 1:1 DAR 4:3], 29.92 fps, 29.92 tbr, 90k tbn, 59.83 tbc
       Stream #0.1[0x101]: Audio: aac, 48000 Hz, stereo, s16, 111 kb/s

    The ffprobe output of one of the second set of segments :

    Input #0, mpegts, from 'ad1.ts':
     Duration: 00:00:09.64, start: 1.400000, bitrate: 578 kb/s
     Program 1 Service01
       Metadata:
         name            : Service01
         provider_name   : FFmpeg
       Stream #0.0[0x100]: Video: h264, yuv420p, 640x480 [PAR 1:1 DAR 4:3], 25 fps, 25 tbr, 90k tbn, 50 tbc
       Stream #0.1[0x101]: Audio: aac, 48000 Hz, stereo, s16, 22 kb/s

    Thank you,