
Recherche avancée
Médias (1)
-
The Slip - Artworks
26 septembre 2011, par
Mis à jour : Septembre 2011
Langue : English
Type : Texte
Autres articles (65)
-
Websites made with MediaSPIP
2 mai 2011, parThis page lists some websites based on MediaSPIP.
-
Creating farms of unique websites
13 avril 2011, parMediaSPIP platforms can be installed as a farm, with a single "core" hosted on a dedicated server and used by multiple websites.
This allows (among other things) : implementation costs to be shared between several different projects / individuals rapid deployment of multiple unique sites creation of groups of like-minded sites, making it possible to browse media in a more controlled and selective environment than the major "open" (...) -
Les autorisations surchargées par les plugins
27 avril 2010, parMediaspip core
autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs
Sur d’autres sites (12539)
-
CoreAudio : how to retrieve actual sampling rate of raw data ?
13 août 2014, par jyavenardWhen attempting to play AAC-HE content in an mp4 container, the reported sampling rate found in the mp4 container appears to be half of the actual sampling rate.
E.g it appears as 24kHz instead of 48kHz.
Using the FFmpeg AAC decoder, retrieving the actual sampling rate can be done by simply decoding an audio packet using
avcodec_decode_audio4
And looking at
AVCodecContext::sample_rate
which will be updated appropriately. From that it’s easy to adapt the output.With CoreAudio decoder, I would use a
AudioConverterRef
set the input and outputAudioStreamBasicDescription
and callAudioConverterFillComplexBuffer
As the converter performs all the required internal conversion including resampling it’s fine. But it plays the content after resampling it to 24kHz (as that’s what the input
AudioStreamBasicDescription
contains.Would there be a way to retrieve the actual sampling rate as found be the decoder (rather than the demuxer) in a similar fashion as one can with FFmpeg ?
Would prefer to avoid losing audio quality if at all possible, and not downmix data
Thanks
-
how to put image files from one folder to make a video in python
23 septembre 2015, par Gaurav16I have written a code to collect all image files from one folder, so that I can make a video out of it.
But while compiling I getting an error saying
"Error creating movie, return code: 4 Try running with --verbose-debug"
#import numpy as np
import matplotlib.pyplot as plt
import matplotlib.animation as animation
import glob, os
from PIL import Image
fig = plt.figure()
#size = 128, 128
ims=[]
#print(len(ims))
#im=Image.open(ims[5])
#im.show()
for infile in glob.glob("*.png"):
file, ext = os.path.splitext(infile)
im = Image.open(infile)
im1=plt.imshow(im)
ims.append([im1])
print(len(ims))
ani = animation.ArtistAnimation(fig, ims)
ani.save('MovWave.mpeg', writer="ffmpeg")
plt.show() -
Feeding a series of images to ffmpeg as each image is created [closed]
5 février 2013, par Mark SchneiderI'm trying to use ffmpeg to build a 1280x720 slide-show from a sequence of pictures and videos, but I have concerns about potential disk I/O bottleneck.
I expect a typical slide-show to have about 50 pictures and 2-3 videos (10-15 seconds each at 30 fps). I would like to show each picture for 3-4 seconds (possibly with a
Ken Burns effect) with a smooth 2 second crossfade between each set of pictures (or for pictures adjacent to videos - between the picture and the first/last frame of the video).Given about 50 pictures, the crossfades alone would amount to about 3,000 images (50 transitions x 2 secs/transition x 30 fps). And I suppose if I implement a Ken Burns effect during each picture's 3-4 second showing, I'd have to provide ffmpeg with individual images for each of those frames. (I'm writing a script in Ruby that will pull a list of images from a database and in turn call ImageMagick to create the individual images for each frame. As I understand it, the RMagick library interfaces with ImageMagick such that the output images come back as in-memory objects without needing to write to disk. FWIW, I'm developing in Windows 8 and will deploy to Heroku.)
All of the slideshow examples I've found online feed ffmpeg a set of images which have already been created. However, in an effort to avoid waiting on considerable disk I/O, I'd like to feed each image to ffmpeg as the image is created rather than create them all in advance.
Is there a way to send each image file to ffmpeg on the fly as the file is created in memory ?