Recherche avancée

Médias (91)

Autres articles (39)

  • (Dés)Activation de fonctionnalités (plugins)

    18 février 2011, par

    Pour gérer l’ajout et la suppression de fonctionnalités supplémentaires (ou plugins), MediaSPIP utilise à partir de la version 0.2 SVP.
    SVP permet l’activation facile de plugins depuis l’espace de configuration de MediaSPIP.
    Pour y accéder, il suffit de se rendre dans l’espace de configuration puis de se rendre sur la page "Gestion des plugins".
    MediaSPIP est fourni par défaut avec l’ensemble des plugins dits "compatibles", ils ont été testés et intégrés afin de fonctionner parfaitement avec chaque (...)

  • Le plugin : Podcasts.

    14 juillet 2010, par

    Le problème du podcasting est à nouveau un problème révélateur de la normalisation des transports de données sur Internet.
    Deux formats intéressants existent : Celui développé par Apple, très axé sur l’utilisation d’iTunes dont la SPEC est ici ; Le format "Media RSS Module" qui est plus "libre" notamment soutenu par Yahoo et le logiciel Miro ;
    Types de fichiers supportés dans les flux
    Le format d’Apple n’autorise que les formats suivants dans ses flux : .mp3 audio/mpeg .m4a audio/x-m4a .mp4 (...)

  • Les autorisations surchargées par les plugins

    27 avril 2010, par

    Mediaspip core
    autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs

Sur d’autres sites (6431)

  • Read a Bytes image from Amazon Kinesis output in python

    14 février 2020, par Varun_Rathinam

    I used imageio.get_reader(BytesIO(a),  'ffmpeg') to load a bytes image and save it as normal image.

    But the below error throws when I read the image using imageio.get_reader(BytesIO(a),  'ffmpeg')

    Traceback (most recent call last):
     File "<stdin>", line 1, in <module>
     File "/home/tango/anaconda3/lib/python3.6/site-packages/imageio/core/functions.py", line 186, in get_reader
       return format.get_reader(request)
     File "/home/tango/anaconda3/lib/python3.6/site-packages/imageio/core/format.py", line 164, in get_reader
       return self.Reader(self, request)
     File "/home/tango/anaconda3/lib/python3.6/site-packages/imageio/core/format.py", line 214, in __init__
       self._open(**self.request.kwargs.copy())
     File "/home/tango/anaconda3/lib/python3.6/site-packages/imageio/plugins/ffmpeg.py", line 323, in _open
       self._initialize()
     File "/home/tango/anaconda3/lib/python3.6/site-packages/imageio/plugins/ffmpeg.py", line 466, in _initialize
       self._meta.update(self._read_gen.__next__())
     File "/home/tango/anaconda3/lib/python3.6/site-packages/imageio_ffmpeg/_io.py", line 150, in read_frames
       raise IOError(fmt.format(err2))
    OSError: Could not load meta information
    === stderr ===

    ffmpeg version 4.2 Copyright (c) 2000-2019 the FFmpeg developers
     built with gcc 7.3.0 (crosstool-NG 1.23.0.449-a04d0)
     configuration: --prefix=/home/tango/anaconda3 --cc=/home/conda/feedstock_root/build_artifacts/ffmpeg_1566210161358/_build_env/bin/x86_64-conda_cos6-linux-gnu-cc --disable-doc --disable-openssl --enable-avresample --enable-gnutls --enable-gpl --enable-hardcoded-tables --enable-libfreetype --enable-libopenh264 --enable-libx264 --enable-pic --enable-pthreads --enable-shared --enable-static --enable-version3 --enable-zlib --enable-libmp3lame
     libavutil      56. 31.100 / 56. 31.100
     libavcodec     58. 54.100 / 58. 54.100
     libavformat    58. 29.100 / 58. 29.100
     libavdevice    58.  8.100 / 58.  8.100
     libavfilter     7. 57.100 /  7. 57.100
     libavresample   4.  0.  0 /  4.  0.  0
     libswscale      5.  5.100 /  5.  5.100
     libswresample   3.  5.100 /  3.  5.100
     libpostproc    55.  5.100 / 55.  5.100
    [matroska,webm @ 0x5619b9da3cc0] File ended prematurely
    [matroska,webm @ 0x5619b9da3cc0] Could not find codec parameters for stream 0 (Video: h264, none, 1280x720): unspecified pixel format
    Consider increasing the value for the 'analyzeduration' and 'probesize' options
    Input #0, matroska,webm, from '/tmp/imageio_zm6hhpgr':
     Metadata:
       title           : Kinesis Video SDK
       encoder         : Kinesis Video SDK 1.0.0
       AWS_KINESISVIDEO_FRAGMENT_NUMBER: 91343852333183888465720004820715065721442989478
       AWS_KINESISVIDEO_SERVER_TIMESTAMP: 1580791384.096
       AWS_KINESISVIDEO_PRODUCER_TIMESTAMP: 1580791377.843
     Duration: N/A, bitrate: N/A
       Stream #0:0(eng): Video: h264, none, 1280x720, SAR 1:1 DAR 16:9, 1k tbr, 1k tbn, 2k tbc (default)
       Metadata:
         title           : kinesis_video
    Stream mapping:
     Stream #0:0 -> #0:0 (h264 (native) -> rawvideo (native))
    Press [q] to stop, [?] for help
    Cannot determine format of input stream 0:0 after EOF
    Error marking filters as finished
    Conversion failed!
    </module></stdin>

    The above approach to read a MKV bytes file was done based on this thread

    Or is there is any approach to parse and read the MKV bytes file.

  • tiff series to mp4 using ffmpeg

    11 avril 2018, par Learning from masters

    I have a big tiff file which contains several (1000) images inside. How can I convert this file into a mp4 video ?

    I tried : ffmpeg -i test.tiff test.mp4 but it gives me :

    ffmpeg version N-82966-g6993bb4 Copyright (c) 2000-2016 the FFmpeg
    developers built with gcc 5.4.0 (GCC) configuration : —enable-gpl
    —enable-version3 —enable-dxva2 —enable-libmfx -
    -enable-nvenc —enable-avisynth —enable-bzlib —enable-fontconfig —enable-frei 0r —enable-gnutls —enable-iconv —enable-libass —enable-libbluray —enable-li bbs2b —enable-libcaca —enable-libfreetype —enable-libgme —enable-libgsm —en able-libilbc —enable-libmodplug —enable-libmp3lame —enable-libopencore-amrnb
    —enable-libopencore-amrwb —enable-libopenh264 —enable-libopenjpeg —enable-li bopus —enable-librtmp —enable-libsnappy —enable-libsoxr —enable-libspeex —e nable-libtheora —enable-libtwolame —enable-libvidstab —enable-libvo-amrwbenc
    —enable-libvorbis —enable-libvpx —enable-libwavpack —enable-libwebp —enable
    -libx264 —enable-libx265 —enable-libxavs —enable-libxvid —enable-libzimg —e nable-lzma —enable-decklink —enable-zlib libavutil 55. 43.100 / 55. 43.100 libavcodec 57. 70.100 /
    57. 70.100 libavformat 57. 61.100 / 57. 61.100 libavdevice 57. 2.100 / 57. 2.100 libavfilter 6. 68.100 / 6. 68.100 libswscale 4. 3.101 / 4. 3.101 libswresample 2. 4.100 /
    2. 4.100 libpostproc 54. 2.100 / 54. 2.100 [tiff_pipe @ 000000000042a6e0] Stream #0 : not enough frames to estimate rate ; co
    nsider increasing probesize [tiff_pipe @ 000000000042a6e0] decoding
    for stream 0 failed [tiff_pipe @ 000000000042a6e0] Could not find
    codec parameters for stream 0 (Vid eo : tiff, none) : unspecified size
    Consider increasing the value for the ’analyzeduration’ and
    ’probesize’ options Input #0, tiff_pipe, from ’test.tiff’ : Duration :
    N/A, bitrate : N/A
    Stream #0:0 : Video : tiff, none, 25 tbr, 25 tbn, 25 tbc [buffer @ 0000000000454460] Unable to parse option value "0x0" as image size
    [buffer @ 0000000000454460] Unable to parse option value "-1" as pixel
    format [buffer @ 0000000000454460] Unable to parse option value "0x0"
    as image size [buffer @ 0000000000454460] Error setting option
    video_size to value 0x0. [graph 0 input from stream 0:0 @
    0000000000454340] Error applying options to the filter. Error opening
    filters !

  • Why doesn't this FFmpeg code create a video from a series of images ?

    20 octobre 2011, par user551117

    I have successfully compiled the FFmpeg library for use in an iOS application. I would like to use it for encoding a video from a series of images, but I can't seem to make it work.

    The following is the code that I am using to encode this video :

    AVCodec *codec;
    AVCodecContext *c= NULL;
    int i, out_size, size, outbuf_size;
    FILE *f;
    AVFrame *picture;
    uint8_t *outbuf;

    printf("Video encoding\n");

    /// find the mpeg video encoder
    codec=avcodec_find_encoder(CODEC_ID_MPEG4);
    //codec = avcodec_find_encoder(CODEC_ID_MPEG4);
    if (!codec) {
       fprintf(stderr, "codec not found\n");
       exit(1);
    }

    c= avcodec_alloc_context();
    picture= avcodec_alloc_frame();

    // put sample parameters
    c->bit_rate = 400000;
    /// resolution must be a multiple of two
    c->width = 320;
    c->height = 480;
    //frames per second
    c->time_base= (AVRational){1,25};
    c->gop_size = 10; /// emit one intra frame every ten frames
    c->max_b_frames=1;
    c->pix_fmt = PIX_FMT_YUV420P;

    //open it
    if (avcodec_open(c, codec) &lt; 0) {
       fprintf(stderr, "could not open codec\n");
       exit(1);
    }

    f = fopen([[NSTemporaryDirectory() stringByAppendingPathComponent:filename] UTF8String], "w");
    if (!f) {
       fprintf(stderr, "could not open %s\n",[filename UTF8String]);
       exit(1);
    }

    // alloc image and output buffer
    outbuf_size = 100000;
    outbuf = malloc(outbuf_size);
    size = c->width * c->height;

    #pragma mark -
    AVFrame* outpic = avcodec_alloc_frame();
    int nbytes = avpicture_get_size(PIX_FMT_YUV420P, c->width, c->height);

    //create buffer for the output image
    uint8_t* outbuffer = (uint8_t*)av_malloc(nbytes);

    #pragma mark -  
    for(i=1;i&lt;48;i++) {
       fflush(stdout);

       int numBytes = avpicture_get_size(PIX_FMT_YUV420P, c->width, c->height);
       uint8_t *buffer = (uint8_t *)av_malloc(numBytes*sizeof(uint8_t));

       UIImage *image = [UIImage imageNamed:[NSString stringWithFormat:@"%d.png", i]];
       CGImageRef newCgImage = [image CGImage];

       CGDataProviderRef dataProvider = CGImageGetDataProvider(newCgImage);
       CFDataRef bitmapData = CGDataProviderCopyData(dataProvider);
       buffer = (uint8_t *)CFDataGetBytePtr(bitmapData);  

       avpicture_fill((AVPicture*)picture, buffer, PIX_FMT_RGB24, c->width, c->height);
       avpicture_fill((AVPicture*)outpic, outbuffer, PIX_FMT_YUV420P, c->width, c->height);

       struct SwsContext* fooContext = sws_getContext(c->width, c->height,
                                                      PIX_FMT_RGB24,
                                                      c->width, c->height,
                                                      PIX_FMT_YUV420P,
                                                      SWS_FAST_BILINEAR, NULL, NULL, NULL);

       //perform the conversion
       sws_scale(fooContext, picture->data, picture->linesize, 0, c->height, outpic->data, outpic->linesize);
       // Here is where I try to convert to YUV

       // encode the image

       out_size = avcodec_encode_video(c, outbuf, outbuf_size, outpic);
       printf("encoding frame %3d (size=%5d)\n", i, out_size);
       fwrite(outbuf, 1, out_size, f);

       free(buffer);
       buffer = NULL;      

    }

    // get the delayed frames
    for(; out_size; i++) {
       fflush(stdout);

       out_size = avcodec_encode_video(c, outbuf, outbuf_size, NULL);
       printf("write frame %3d (size=%5d)\n", i, out_size);
       fwrite(outbuf, 1, outbuf_size, f);      
    }

    // add sequence end code to have a real mpeg file
    outbuf[0] = 0x00;
    outbuf[1] = 0x00;
    outbuf[2] = 0x01;
    outbuf[3] = 0xb7;
    fwrite(outbuf, 1, 4, f);
    fclose(f);
    free(outbuf);

    avcodec_close(c);
    av_free(c);
    av_free(picture);
    printf("\n");

    What could be wrong with this code ?