
Recherche avancée
Médias (2)
-
Exemple de boutons d’action pour une collection collaborative
27 février 2013, par
Mis à jour : Mars 2013
Langue : français
Type : Image
-
Exemple de boutons d’action pour une collection personnelle
27 février 2013, par
Mis à jour : Février 2013
Langue : English
Type : Image
Autres articles (82)
-
Encoding and processing into web-friendly formats
13 avril 2011, parMediaSPIP automatically converts uploaded files to internet-compatible formats.
Video files are encoded in MP4, Ogv and WebM (supported by HTML5) and MP4 (supported by Flash).
Audio files are encoded in MP3 and Ogg (supported by HTML5) and MP3 (supported by Flash).
Where possible, text is analyzed in order to retrieve the data needed for search engine detection, and then exported as a series of image files.
All uploaded files are stored online in their original format, so you can (...) -
MediaSPIP Player : problèmes potentiels
22 février 2011, parLe lecteur ne fonctionne pas sur Internet Explorer
Sur Internet Explorer (8 et 7 au moins), le plugin utilise le lecteur Flash flowplayer pour lire vidéos et son. Si le lecteur ne semble pas fonctionner, cela peut venir de la configuration du mod_deflate d’Apache.
Si dans la configuration de ce module Apache vous avez une ligne qui ressemble à la suivante, essayez de la supprimer ou de la commenter pour voir si le lecteur fonctionne correctement : /** * GeSHi (C) 2004 - 2007 Nigel McNie, (...) -
Supporting all media types
13 avril 2011, parUnlike most software and media-sharing platforms, MediaSPIP aims to manage as many different media types as possible. The following are just a few examples from an ever-expanding list of supported formats : images : png, gif, jpg, bmp and more audio : MP3, Ogg, Wav and more video : AVI, MP4, OGV, mpg, mov, wmv and more text, code and other data : OpenOffice, Microsoft Office (Word, PowerPoint, Excel), web (html, CSS), LaTeX, Google Earth and (...)
Sur d’autres sites (5882)
-
aacenc_pred : rework the way prediction is done
29 août 2015, par Rostislav Pehlivanovaacenc_pred : rework the way prediction is done
This commit completely alters the algorithm of prediction.
The original commit which introduced prediction was completely
incorrect to even remotely care about what the actual coefficients
contain or whether any options were enabled. Not my actual fault.This commit treats prediction the way the decoder does and expects
to do : like lossy encryption. Everything related to prediction now
happens at the very end but just before quantization and encoding
of coefficients. On the decoder side, prediction happens before
anything has had a chance to even access the coefficients.Also the original implementation had problems because it actually
touched the band_type of special bands which already had their
scalefactor indices marked and it’s a wonder the asserion wasn’t
triggered when transmitting those.Overall, this now drastically increases audio quality and you should
think about enabling it if you don’t plan on playing anything encoded
on really old low power ultra-embedded devices since they might not
support decoding of prediction or AAC-Main. Though the specifications
were written ages ago and as times change so do the FLOPS.Signed-off-by : Rostislav Pehlivanov <atomnuker@gmail.com>
-
ffmpeg replacing audio track with track from another video but different length (sync audio)
27 août 2019, par MattBackground
I have digitized some old Canon Video8 tapes. I used a SONY Digital8 camera (which is backwards compatible with Video8) to output DV (using it’s inbuilt ADC). The conversion process worked well for the video but the audio came through jumpy/distorted in places. This left me with a problem, was it the Camera or the tape ? So I bought another Samsung Video8 camera (just analog) and using the SONY’s passthrough feature output from the Samsung (Composite & mono audio) into the SONY which output the DV. Much to my delight it worked ! The audio was clear.
Result
DV01.avi - Good Video / Crap Audio
DV02.avi - Crap Video / Good AudioOk so obviously what I would like to do is take the video track from DV01 and the audio track from DV02 and join/mux ? them WITHOUT re-encoding.
Problem 1 : They have different start times so just copying over the audio track will result it not being in sync.
After some googling I found you can use ffmpeg to take care of this :
Firstly here is the Video info using :
ffmpeg -i DV01.avi
Input #0, avi, from 'DV01.avi':
Duration: 02:53:06.68, start: 0.000000, bitrate: 28878 kb/s
Stream #0:0: Video: dvvideo, yuv420p, 720x576 [SAR 16:15 DAR 4:3], 25000 kb/s, 25 fps, 25 tbr, 25 tbn, 25 tbc
Stream #0:1: Audio: pcm_s16le, 32000 Hz, stereo, s16, 1024 kb/s
Stream #0:2: Audio: pcm_s16le, 32000 Hz, stereo, s16, 1024 kb/sMuxing :
ffmpeg -itsoffset 4 -i DV01.avi -i DV02.avi -map 0:v -map 1:a -c copy output.avi
In the example above I am delaying the start of the video 4 seconds (I think ?) This is just an example I haven’t actually tried it as the file sizes are 40Gb !
So my QUESTION is :
Given the background above what would be the best way to join/mux the two streams together (without re-encoding) with the audio being in sync. Given that syncing the audio to the video may need millisecond tweaking I don’t believe trial and error is a good idea (I don’t want to tweak it by 10ms then rinse and repeat for a 40Gb file) ?
I just had a thought, could I say create a 10 second clip (from the start) of each video and use them to find the reference/sync start point then use that when muxing the 40Gb versions ?
Anyway, you get the idea. Looking for ways to solve this problem. Thanks !
-
FFMPEG - local video to UDP streaming to OpenCV - video quality degraded
6 juin 2021, par user3925023my goal is to re-stream local video content / desktop screencasting, to an UDP flow that I need to process on a Python script.



Here is the FFMPEG script that I'm using :



ffmpeg -re -i C:\Users\test\Downloads\out.ts -strict -2 -c:v copy -an -preset slower -tune stillimage -b 11200k -f rawvideo udp://127.0.0.1:5000




And here is the simple Python script supposed to read the stream flow :



import cv2

cap = cv2.VideoCapture('udp://127.0.0.1:5000',cv2.CAP_FFMPEG)
if not cap.isOpened():
 print('VideoCapture not opened')
 exit(-1)
width = cap.get(cv2.CAP_PROP_FRAME_WIDTH) # float
height = cap.get(cv2.CAP_PROP_FRAME_HEIGHT) # float
print(str(width))
print(str(height))
while True:
 ret, frame = cap.read()
 imgray = frame[int(round((height/100)*70,0)):int(round((width/100)*42,0)), int(round((height/100)*74,0)):int(round((width/100)*54,0))]
 if not ret:
 print('frame empty')
 break
 cv2.imshow('image', imgray)
 if cv2.waitKey(1)&0XFF == ord('q'):
 break
cap.release()




I'm able to visualize portion of the stream video as expect, but I'm facing lot of issue in video quality degradation, specially video artifact probably due missing packet processing :






Also these are error log I'm geting from script :



[h264 @ 0000026eb272f280] error while decoding MB 105 66, bytestream -21
[h264 @ 0000026eb2fcb740] error while decoding MB 100 53, bytestream -11
[h264 @ 0000026eb272f280] error while decoding MB 32 22, bytestream -11
[h264 @ 0000026ead9ee300] error while decoding MB 60 20, bytestream -25
[h264 @ 0000026eb27f00c0] error while decoding MB 9 62, bytestream -5
[h264 @ 0000026ead9ee780] error while decoding MB 85 44, bytestream -5
[h264 @ 0000026eb27f0800] error while decoding MB 64 25, bytestream -15
[h264 @ 0000026eb272f280] error while decoding MB 112 23, bytestream -17
[h264 @ 0000026eb2735200] error while decoding MB 30 21, bytestream -7




Actually I don't care about video fluidity,I can also reduce the FPS, important thing is the video quality. Not sure if I'm doing wrong on the scripting python part or if I'm using wrong FFMPEG command.



Many Thanks