Recherche avancée

Médias (1)

Mot : - Tags -/artwork

Autres articles (65)

  • Websites made ​​with MediaSPIP

    2 mai 2011, par

    This page lists some websites based on MediaSPIP.

  • Creating farms of unique websites

    13 avril 2011, par

    MediaSPIP platforms can be installed as a farm, with a single "core" hosted on a dedicated server and used by multiple websites.
    This allows (among other things) : implementation costs to be shared between several different projects / individuals rapid deployment of multiple unique sites creation of groups of like-minded sites, making it possible to browse media in a more controlled and selective environment than the major "open" (...)

  • Les autorisations surchargées par les plugins

    27 avril 2010, par

    Mediaspip core
    autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs

Sur d’autres sites (12539)

  • CoreAudio : how to retrieve actual sampling rate of raw data ?

    13 août 2014, par jyavenard

    When attempting to play AAC-HE content in an mp4 container, the reported sampling rate found in the mp4 container appears to be half of the actual sampling rate.

    E.g it appears as 24kHz instead of 48kHz.

    Using the FFmpeg AAC decoder, retrieving the actual sampling rate can be done by simply decoding an audio packet using

    avcodec_decode_audio4

    And looking at AVCodecContext::sample_rate which will be updated appropriately. From that it’s easy to adapt the output.

    With CoreAudio decoder, I would use a AudioConverterRef set the input and output AudioStreamBasicDescription
    and call AudioConverterFillComplexBuffer

    As the converter performs all the required internal conversion including resampling it’s fine. But it plays the content after resampling it to 24kHz (as that’s what the input AudioStreamBasicDescription contains.

    Would there be a way to retrieve the actual sampling rate as found be the decoder (rather than the demuxer) in a similar fashion as one can with FFmpeg ?

    Would prefer to avoid losing audio quality if at all possible, and not downmix data

    Thanks

  • how to put image files from one folder to make a video in python

    23 septembre 2015, par Gaurav16

    I have written a code to collect all image files from one folder, so that I can make a video out of it.

    But while compiling I getting an error saying "Error creating movie, return code: 4 Try running with --verbose-debug"

     #import numpy as np
    import matplotlib.pyplot as plt
    import matplotlib.animation as animation
    import glob, os
    from PIL import Image
    fig = plt.figure()
    #size = 128, 128
    ims=[]
    #print(len(ims))
    #im=Image.open(ims[5])
    #im.show()
    for infile in glob.glob("*.png"):
       file, ext = os.path.splitext(infile)
       im = Image.open(infile)
       im1=plt.imshow(im)
       ims.append([im1])
    print(len(ims))    
    ani = animation.ArtistAnimation(fig, ims)
    ani.save('MovWave.mpeg', writer="ffmpeg")
    plt.show()
  • Feeding a series of images to ffmpeg as each image is created [closed]

    5 février 2013, par Mark Schneider

    I'm trying to use ffmpeg to build a 1280x720 slide-show from a sequence of pictures and videos, but I have concerns about potential disk I/O bottleneck.

    I expect a typical slide-show to have about 50 pictures and 2-3 videos (10-15 seconds each at 30 fps). I would like to show each picture for 3-4 seconds (possibly with a
    Ken Burns effect) with a smooth 2 second crossfade between each set of pictures (or for pictures adjacent to videos - between the picture and the first/last frame of the video).

    Given about 50 pictures, the crossfades alone would amount to about 3,000 images (50 transitions x 2 secs/transition x 30 fps). And I suppose if I implement a Ken Burns effect during each picture's 3-4 second showing, I'd have to provide ffmpeg with individual images for each of those frames. (I'm writing a script in Ruby that will pull a list of images from a database and in turn call ImageMagick to create the individual images for each frame. As I understand it, the RMagick library interfaces with ImageMagick such that the output images come back as in-memory objects without needing to write to disk. FWIW, I'm developing in Windows 8 and will deploy to Heroku.)

    All of the slideshow examples I've found online feed ffmpeg a set of images which have already been created. However, in an effort to avoid waiting on considerable disk I/O, I'd like to feed each image to ffmpeg as the image is created rather than create them all in advance.

    Is there a way to send each image file to ffmpeg on the fly as the file is created in memory ?