
Recherche avancée
Médias (91)
-
Géodiversité
9 septembre 2011, par ,
Mis à jour : Août 2018
Langue : français
Type : Texte
-
USGS Real-time Earthquakes
8 septembre 2011, par
Mis à jour : Septembre 2011
Langue : français
Type : Texte
-
SWFUpload Process
6 septembre 2011, par
Mis à jour : Septembre 2011
Langue : français
Type : Texte
-
La conservation du net art au musée. Les stratégies à l’œuvre
26 mai 2011
Mis à jour : Juillet 2013
Langue : français
Type : Texte
-
Podcasting Legal guide
16 mai 2011, par
Mis à jour : Mai 2011
Langue : English
Type : Texte
-
Creativecommons informational flyer
16 mai 2011, par
Mis à jour : Juillet 2013
Langue : English
Type : Texte
Autres articles (40)
-
Use, discuss, criticize
13 avril 2011, parTalk to people directly involved in MediaSPIP’s development, or to people around you who could use MediaSPIP to share, enhance or develop their creative projects.
The bigger the community, the more MediaSPIP’s potential will be explored and the faster the software will evolve.
A discussion list is available for all exchanges between users. -
HTML5 audio and video support
13 avril 2011, parMediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
For older browsers the Flowplayer flash fallback is used.
MediaSPIP allows for media playback on major mobile platforms with the above (...) -
De l’upload à la vidéo finale [version standalone]
31 janvier 2010, parLe chemin d’un document audio ou vidéo dans SPIPMotion est divisé en trois étapes distinctes.
Upload et récupération d’informations de la vidéo source
Dans un premier temps, il est nécessaire de créer un article SPIP et de lui joindre le document vidéo "source".
Au moment où ce document est joint à l’article, deux actions supplémentaires au comportement normal sont exécutées : La récupération des informations techniques des flux audio et video du fichier ; La génération d’une vignette : extraction d’une (...)
Sur d’autres sites (5597)
-
How to combine/concatenate videos stored in AWS S3 bucket based on title of the file name
9 juin 2020, par orangecubeI am using a service that allows me to record videos that get automatically pushed to a folder (submissions) in an S3 bucket. There are multiple videos however they need to be grouped together and concatenated so the output is one video per group.



So, basically, any tips on how I can take videos based on the title and stitch them together ?



Example :



Submissions folder will have :



a-100-2.mp4
a-200-6.mp4
b-123-5.mp4




Expected output in processed folder :



a.mp4 - (both 'a' videos get stitched together)
b.mp4 - (only 'b' gets sent over since there is only one video.)




Thanks in advance !



Edit : Some additional and detailed information below if it helps.



The files will be labeled with :
name-location-video_token-stream_token.mp4



Need help creating a script or process that will concatenate the videos using the procedure outlined below :



Processing rules (back end) :



- 

-
Check if videos have same video_token in ‘submissions folder’. If so, keep the newest one and delete old ones.
-
Take all videos in ‘submissions folder’ with same name and location in title and concatenate the videos. Save output video to a new folder in the bucket labeled as the location for the folder name. 
Output file name : 
name-location-year.mp4.







EXAMPLE :



Submissions folder :
joey-toronto-001-354.mp4



joey-toronto-001-241.mp4 - this will be deleted



joey-toronto-103-452.mp4



alex-montreal-352-232.mp4



alex-montreal-452-223.mp4



Resulting output :



Toronto folder :



Joey-toronto-2020.mp4



Montreal folder :



Alex-montreal-2020.mp4


-
-
Why PyAudio doesn't read 'mp3' ?
22 octobre 2020, par freshITmeatI tried to read file that I give with absolute path.
When I run my code first that I see is this message :


D:\prog\datascience\anaconda\lib\site-packages\pydub\utils.py:170: RuntimeWarning: Couldn't find ffmpeg or avconv - defaulting to ffmpeg, but may not work
 warn("Couldn't find ffmpeg or avconv - defaulting to ffmpeg, but may not work", RuntimeWarning)



I tried this :


PATH_TO_FFMPEG = 'D:\\prog\\ffmpeg-win-2.2.2\\ffmpeg.exe'
pydub.AudioSegment.converter = r'D:\\prog\\ffmpeg-win-2.2.2\\ffmpeg.exe'



And I separately installed
ffmpeg
withpip
. But it didn't help.
When I try this :

raw_sound = pydub.AudioSegment.from_mp3(file=track_path)



where
track_path
is correct absolute path generated automatically.
So I got this this error :

Traceback (most recent call last):
 File "D:\prog\PyCharm Community Edition 2020.2.2\plugins\python-ce\helpers\pydev\pydevd.py", line 1448, in _exec
 pydev_imports.execfile(file, globals, locals) # execute the script
 File "D:\prog\PyCharm Community Edition 2020.2.2\plugins\python-ce\helpers\pydev\_pydev_imps\_pydev_execfile.py", line 18, in execfile
 exec(compile(contents+"\n", file, 'exec'), glob, loc)
 File "D:/testtask2/test_task/testtask/get_mffc.py", line 165, in <module>
 slice_all_in_a_dir('May 27 2020 LNC/Hydrophone 1/raw_records')
 File "D:/testtask2/test_task/testtask/get_mffc.py", line 70, in slice_all_in_a_dir
 slice_samples(track_path= [file],
 File "D:/testtask2/test_task/testtask/get_mffc.py", line 48, in slice_samples
 raw_sound = pydub.AudioSegment.from_mp3(file=track_path)
 File "D:\prog\datascience\anaconda\lib\site-packages\pydub\audio_segment.py", line 738, in from_mp3
 return cls.from_file(file, 'mp3', parameters=parameters)
 File "D:\prog\datascience\anaconda\lib\site-packages\pydub\audio_segment.py", line 680, in from_file
 stdin_data = file.read()
AttributeError: 'list' object has no attribute 'read'
python-BaseException
</module>


-
Can't overlay a WebM transpartent video on top of a MP4 using ffmpeg [closed]
27 octobre 2020, par HervéI have been struggling for hours trying to overlay a WebM transpartent video on top of a MP4 using ffmpeg. The two videos have the same duration. It should therefore be easy but I am getting desperate.


"c:\Program Files\ffmpeg\bin\ffmpeg.exe" ^
-i drone.mp4 ^
-i trans.webm -pix_fmt yuv444p ^
-filter_complex "[1:v]format=rgba,colorchannelmixer=aa=0.5[trans];[0:v][trans]overlay=10:10" ^
out.mp4



I tried MANY MANY different options, formats, but kept obtaining the same outcome : ffmpeg simply ignores the alpha channel of the trans.webm file and considers its background as black. I used the colorchannelmixer=aa=0.5 and the 10:10 offset to actually see the problem. My goal is really to keep the original subtle alpha channel of trans.webm, NOT to set the black color to transparent.


Some information about the two files :


drone.mp4


Metadata:
 creation_time : 2020-10-27T10:18:38.000000Z
 handler_name : VideoHandler
 encoder : h264
 Stream #0:1(eng): Audio: aac (LC) (mp4a / 0x6134706D), 44100 Hz, stereo, fltp, 191 kb/s (default)
 Metadata:
 creation_time : 2020-10-27T10:18:38.000000Z
 handler_name : SoundHandler



trans.webm


Input #0, matroska,webm, from 'trans.webm':
 Metadata:
 ENCODER : Lavf58.51.100
 Duration: 00:00:27.00, start: 0.000000, bitrate: 28 kb/s
 Stream #0:0: Video: vp9 (Profile 0), yuv420p(tv), 1920x1080, SAR 1:1 DAR 16:9, 25 fps, 25 tbr, 1k tbn, 1k tbc (default)
 Metadata:
 alpha_mode : 1
 ENCODER : Lavc58.100.100 libvpx-vp9
 DURATION : 00:00:27.000000000