
Recherche avancée
Médias (1)
-
Bug de détection d’ogg
22 mars 2013, par
Mis à jour : Avril 2013
Langue : français
Type : Video
Autres articles (112)
-
MediaSPIP 0.1 Beta version
25 avril 2011, parMediaSPIP 0.1 beta is the first version of MediaSPIP proclaimed as "usable".
The zip file provided here only contains the sources of MediaSPIP in its standalone version.
To get a working installation, you must manually install all-software dependencies on the server.
If you want to use this archive for an installation in "farm mode", you will also need to proceed to other manual (...) -
Websites made with MediaSPIP
2 mai 2011, parThis page lists some websites based on MediaSPIP.
-
HTML5 audio and video support
13 avril 2011, parMediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
For older browsers the Flowplayer flash fallback is used.
MediaSPIP allows for media playback on major mobile platforms with the above (...)
Sur d’autres sites (14564)
-
displaying a baseline h264 frames stream in browsers
6 août 2021, par Thabet SabhaSo, I have a server that receives a live rtsp stream then generates baseline h264 frames using ffmpeg, which then are sent via an rtcDataChannel to browser, and while the frames arrive as intended, I can't figure out a way to display them on my html5 videoElement,
here is a simplified version of my current approach :


const remoteStream = new MediaSource();
myVideoElement.src = window.URL.createObjectURL(remoteStream);

// called when remoteStream.readyState === "open"
let sourceBuffer = remoteStream.addSourceBuffer('video/mp4; codecs="avc1.4d002a"');

// this gets called when ever a new frame is received from the webrtc data channel.
function onFrame(frame) {
 sourceBuffer.appendBuffer(new Uint8Array(frame));

 /*
 console.log(frame) ==> <buffer 00="00" 01="01" 41="41" 9b="9b" a0="a0" 22="22" 80="80" a5="a5" d7="d7" 42="42" ea="ea" 34="34" 14="14" 85="85" ba="ba" bc="bc" 1b="1b" f2="f2" 71="71" 0d="0d" 8b="8b" e1="e1" 3c="3c" 52="52" d5="d5" 8c="8c" ef="ef" c1="c1" 89="89" 10="10" c5="c5" 05="05" 78="78" ee="ee" 1d="1d" 03="03" 8d="8d" 2896="2896" more="more" bytes="bytes">
 */
}
</buffer>


ffmpeg options :


[
 "-rtsp_transport", "tcp",
 "-i", `${rtspCamURL}`, 
 "-framerate", "15",
 "-c:v", "libx264",
 "-vprofile", "baseline",
 "-b:v", "600k",
 "-bufsize", "600k",
 "-pix_fmt", "yuv420p",
 '-tune', 'zerolatency',
 "-preset", "ultrafast",
 "-f", "rawvideo",
 '-'
]; 



ffmpeg stream is then split using NAL delimiter (to generate individual frames) then each frame is sent via the data channel like so :

Buffer.concat([nalDelimiter, frame])
.

I am not sure if i'm missing something as i'm not getting any helpful errors due to the remoteSource closing as soon as the first frame arrives for some reason.


or does the media source just not support raw h264 frames, and if so is there a workaround to solve this issue ? (even if it has to do with changing the ffmpeg params.


-
Bash : sort find results using part of a filename
13 novembre 2015, par utt50I have 3 webcams set up in a building, uploading still images to a webserver. I’m using ffmpeg to encode the jpgs to mp4 video.
The directories are set up like this :
Cam1/201504
Cam1/201505
Cam2/201504
Cam2/201505
Cam3/201504
Cam3/201505I’m using the following bash loop/ffmpeg parameters to make one video per camera, per year. This works well so far (well... except that my SSD is rapidly degrading in performance - too many simultaneous read/write operations ?) :
find Cam2/2013* -name "*.jpg" -print0 | xargs -0 cat | ffmpeg -f image2pipe -framerate 30 -vcodec mjpeg -i - -vcodec libx264 -profile:v baseline -level 3.0 -movflags +faststart -crf 19 -pix_fmt yuv420p -r 30 "Cam2-2013-30fps-19crf.mp4"
The individual files are named like this (confusing ffmpeg’s built-in file sequencer) :
Cam1_2015052413543201.jpg
Cam1_2015052413544601.jpg
Cam2_2015052413032601.jpg
Cam2_2015052413544901.jpgI now need to create one video for an entire year across all 3 cameras, ordered by timestamp. To accomplish this, I need to sort the find results by the segment of the filename after the underscore.
What do I pipe the find output to to accomplish this ? For example, the files above would be ordered like this :
Cam2_2015052413032601.jpg
Cam1_2015052413543201.jpg
Cam1_2015052413544601.jpg
Cam2_2015052413544901.jpgAny help is very much appreciated !
-
ffmpeg is not executing clip.run() when compiled with pyinstaller
24 septembre 2018, par ThriskelWhen executing the .exe file in another machine that doens’t have it’s requeriments installed, running the programs gives the error :
Exception in thread Thread-1:
Traceback (most recent call last):
File "threading.py", line 916, in _bootstrap_inner
File "threading.py", line 864, in run
File "y2m.py", line 80, in workit
File "site-packages\ffmpeg\_run.py", line 202, in run
File "subprocess.py", line 707, in __init__
File "subprocess.py", line 992, in _execute_child
FileNotFoundError: [WinError 2] The system cannot find the file specifiedI’m compiling the exe file using the command :
pyinstaller y2m.py
I’ve been reading every other post in this and others forums refered to this kind of problem but I don’t seem to find or understand the way of fixing this.
The source code is in here
and the warny2m.txt file is in here
Things that I have tried :
- using —paths=pathToLibFolder
- using -w
- using an older pyinstaller version
- adding python3 to path
- using -p DIR to add C :\Python3\Lib location then sub locations, also individual scripts. (those 3 ways to see if it would work)
EDIT :
When using ffmpeg_run.py the code fails. line is
clip.run()
where clip presents the input and output file.