
Recherche avancée
Médias (1)
-
Bug de détection d’ogg
22 mars 2013, par
Mis à jour : Avril 2013
Langue : français
Type : Video
Autres articles (45)
-
Qu’est ce qu’un éditorial
21 juin 2013, parEcrivez votre de point de vue dans un article. Celui-ci sera rangé dans une rubrique prévue à cet effet.
Un éditorial est un article de type texte uniquement. Il a pour objectif de ranger les points de vue dans une rubrique dédiée. Un seul éditorial est placé à la une en page d’accueil. Pour consulter les précédents, consultez la rubrique dédiée.
Vous pouvez personnaliser le formulaire de création d’un éditorial.
Formulaire de création d’un éditorial Dans le cas d’un document de type éditorial, les (...) -
Contribute to translation
13 avril 2011You can help us to improve the language used in the software interface to make MediaSPIP more accessible and user-friendly. You can also translate the interface into any language that allows it to spread to new linguistic communities.
To do this, we use the translation interface of SPIP where the all the language modules of MediaSPIP are available. Just subscribe to the mailing list and request further informantion on translation.
MediaSPIP is currently available in French and English (...) -
Les formats acceptés
28 janvier 2010, parLes commandes suivantes permettent d’avoir des informations sur les formats et codecs gérés par l’installation local de ffmpeg :
ffmpeg -codecs ffmpeg -formats
Les format videos acceptés en entrée
Cette liste est non exhaustive, elle met en exergue les principaux formats utilisés : h264 : H.264 / AVC / MPEG-4 AVC / MPEG-4 part 10 m4v : raw MPEG-4 video format flv : Flash Video (FLV) / Sorenson Spark / Sorenson H.263 Theora wmv :
Les formats vidéos de sortie possibles
Dans un premier temps on (...)
Sur d’autres sites (7155)
-
Background blur on concat videos make everything more blurry
7 août 2020, par AlphaDjangoHi I have a bunch of temp video files that I want to concat together, some are different aspect ratios and resolutions so I want the end video to be a 1920x1080p with the empty space on the sides to be a box blur.


I have it working how I want but the video clips in the middle become more blurry than the original temp files.
Could someone let me know if I can change the code to keep original quality ? It will also show it's dropping frames a few thousand in a 20 minute video which isn't terrible but I would prefer it to not drop frames.


"ffmpeg -i "concat :" + python script adds all file locations + "" -s 1920x1080 -filter_complex [0:v]scale=ih16/9 :-1,boxblur=luma_radius=min(h,w)/20:luma_power=2:chroma_radius=min(cw,ch)/20:chroma_power=1[bg] ;[bg][0:v]overlay=(W-w)/2 :(H-h)/2,crop=h=iw9/16 -af aselect=concatdec_select,aresample=async=1 Rendered_Videos/output.mp4"


-
Can't Stream Ogg From ffmpeg Through stdout
16 janvier 2012, par dave mankoffTo get the the crux of it, why does the first command work, but the second command does not. They produce slightly differently sized files with different contents :
ffmpeg -i test.wav -f ogg -acodec libvorbis test.a.ogg
ffmpeg -i test.wav -f ogg -acodec libvorbis - > test.b.oggtest.a.ogg will play properly and has no problems. test.b.ogg starts in the middle of the source audio and has stops and gaps in the audio. It also does not report the length of the track.
I want to transcode source files on the fly into ogg for a program I am writing and I am trying to pipe the stdout from ffmpeg into my program. Putting the results into an intermediary file will kill performance since the transcoding is supposed to happen on demand.
-
Connect a remote Ip camera as a Webrtc client
5 avril 2017, par idoshI have 2 cameras :
- An internal webcam embedded in my laptop.
- A remote IP camera that is connected to my laptop through Wifi (transmits TCP, raw H264 data - no container). I’m getting the stream using node.js.
My goal is to create a Webrtc network and connect the remote camera as another client.
I’m trying to figure out possible solutions :
- My naive thinking was that I would stream the remote camera payload to the browser. But as I came to understand the browser can’t handle the stream without a container. Fair enough. But I don’t understand why it does handle the video stream that arrives from my internal camera (from the navigator.getUserMedia() function). what’s the difference between the two streams ? why can’t I mimic the stream from the remote camera as the input ?
- To bypass this problem I thought about creating a virtual camera using Manycam (or Manycam like app). To accomplish that I need to convert my TCP stream into an RTP stream (in order to feed Manycam). Though I did saw some info in ffmpeg command line, I couldn’t find info in their node.js api package "fluent-ffmpeg". Is it possible to do it using fluent-ffmpeg ? Or only using the command line tool ? Would it require another rtp server in the middle such as this one ?.
- Third option I read about is using node.js as a client in Webrtc. I saw it was implemented in "simple-peer". I tried it out using their co-work with socket.io (socket.io-p2p). unfortunately I couldn’t get it to work / : When i’m trying to create a socket/peer in the server - it throws errors, as it expect options that are only available on the client-side (like window, location, etc.). Am I doing something wrong ? maybe there is more suitable framework for this matter ?
- Forth option is to use a streaming server in the middle such as Kurnto. From my understanding it receives rtp as an input and transmits it as a webrtc client. I feel it’s the most excessive option, but maybe it’s not so bad (I have to admit that I haven’t investigate this option yet).
any thoughts ?
thanks !