
Recherche avancée
Médias (1)
-
Richard Stallman et le logiciel libre
19 octobre 2011, par
Mis à jour : Mai 2013
Langue : français
Type : Texte
Autres articles (37)
-
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir -
Keeping control of your media in your hands
13 avril 2011, parThe vocabulary used on this site and around MediaSPIP in general, aims to avoid reference to Web 2.0 and the companies that profit from media-sharing.
While using MediaSPIP, you are invited to avoid using words like "Brand", "Cloud" and "Market".
MediaSPIP is designed to facilitate the sharing of creative media online, while allowing authors to retain complete control of their work.
MediaSPIP aims to be accessible to as many people as possible and development is based on expanding the (...) -
Librairies et binaires spécifiques au traitement vidéo et sonore
31 janvier 2010, parLes logiciels et librairies suivantes sont utilisées par SPIPmotion d’une manière ou d’une autre.
Binaires obligatoires FFMpeg : encodeur principal, permet de transcoder presque tous les types de fichiers vidéo et sonores dans les formats lisibles sur Internet. CF ce tutoriel pour son installation ; Oggz-tools : outils d’inspection de fichiers ogg ; Mediainfo : récupération d’informations depuis la plupart des formats vidéos et sonores ;
Binaires complémentaires et facultatifs flvtool2 : (...)
Sur d’autres sites (3927)
-
react cant find ffmpegwasm
18 septembre 2024, par MartinI am trying to create a working example for ffmpeg wasm with react js in my browser.


I have been following this very simple example :
https://www.youtube.com/watch?v=-OTc0Ki7Sv0&ab_channel=Fireship


installed ffmpeg locally inside my react repo node_modules as seen here :



And followed to tutorial video to edit the App.jsx so it looks like this :


import React, { useState, useEffect } from 'react';
import './App.css';

import { createFFmpeg, fetchFile } from '@ffmpeg/ffmpeg';
const ffmpeg = createFFmpeg({
 log: true,
});
function App() {
 
 const [ready, setReady] = useState(false);

 const load = async () => {
 console.log('load()')
 await ffmpeg.load();
 setReady(true);
 }

 useEffect(()=>{
 load();
 }, [])

 return (
 <div classname="App">
 content
 </div>
 );
}

export default App;




But this leads to error messages in my win10 command prompt terminal saying it cant find the ffmpeg files :


[16:07:47] [snowpack] [404] Not Found (/node_modules/@ffmpeg/core/dist/ffmpeg-core.js)
[16:07:47] [snowpack] [404] Not Found (/node_modules/@ffmpeg/core/dist/ffmpeg-core.wasm)
[16:07:47] [snowpack] [404] Not Found (/node_modules/@ffmpeg/core/dist/ffmpeg-core.worker.js)



I've even tried moving the ffmpeg files to my public folder and editing the code to find them like so :


const ffmpeg = createFFmpeg({
 log: true,
 corePath: '../public/@ffmpeg/core/dist/ffmpeg-core.js',
});



But the same error occured. Why doesn't my react App.jsx file correctly find the ffmpeg files in my node_modules folder ?


-
Adding Subtitles to a VAAPI/QSV 10bit Accelerated Transcode
17 juin 2023, par EnverexI've been converting some of my BluRays to watch on a streaming machine elsewhere in the house but I'm having trouble when it comes to burning subtitles, or more specifically, I'm having trouble figuring out what FFMPEG wants from me to make the process actually work.


It's easy enough to do this in software, but I'm using hardware decoding and encoding and that's where the complexity seems to come from (VAAPI/QSV for decoding, QSV AV1 for encoding).


I can use the following when dealing with SDR content :


-vf "scale_vaapi=w='min(1920,iw)':h=-8:mode=nl_anamorphic:format=p010le:extra_hw_frames=120,hwmap=derive_device=qsv,format=qsv"



And the following when dealing with HDR content :


-vf "scale_vaapi=w='min(1920,iw)':h=-8:mode=nl_anamorphic:format=p010le,tonemap_vaapi=format=p010le:t=bt709:m=bt709:p=bt709:extra_hw_frames=120,hwmap=derive_device=qsv,format=qsv"



But I cannot, in the hundreds of permutations I've tried now, find a single way to shoehorn subtitle baking into the process. I'm using SRT files for simplicity, so I just need to add
subtitles=blah.srt
somewhere to get it to work, but the crux of the issue is knowing where in the chain it needs to go and more importantly, what supporting arguments it needs with it (e.g. hwupload, hwdownload, and their associated switches, etc).

Pretty much every single attempt just results in :




Impossible to convert between the formats supported by the filter 'graph 0 input from stream 0:0' and the filter 'auto_scale_0'
Error reinitializing filters !
Failed to inject frame into filter network : Function not implemented




So, what am I missing ?


-
FFMPEG and FFPlay can access rtsp stream from one ip, but from other ip, it can't
28 mai 2022, par CrearThe situation is kind of complex. I was archiving several CCTV camera feeds (rtsp, h264, no audio) through OpenCV, which worked but the CPU utilization was too high and started to lose some frames time by time.


To reduce the CPU utilization, I started to use FFMPEG to skip the decoding and encoding processes, which worked perfectly on my home machine. However, when I connected to my university VPN and tried to deploy it on our lab server, FFmpeg couldn't read any frame, ffplay couldn't get anything either. However, OpenCV, VLC Player and IINA Player could still read and display the feed.


In Summary,


1 FFMPEG/ffplay


1.1 can only read the feed from my home network(Wi-Fi, optimum)


1.2 from other two networks, the error message says : "Could not find codec parameters for stream 0 (Video : h264, none) : unspecified size
Consider increasing the value for the 'analyzeduration' (0) and 'probesize' (5000000) options"


2 IINA/VLC Player, OpenCV
These tools can get the video all the time.


I'm wondering whether it's related to some specific port access, that the ffmpeg required but the others don't. I'd appreciate it if anyone can provide any suggestions.


As references, the tested ffplay command is simple :


ffplay 'the rtsp address'



Thanks



Update


More tests have been performed.


By specifying rtsp_transport as TCP, ffplay can play the video, but FFmpeg can't access the video. (In the beginning, when both FFmpeg and ffplay worked through my home network, it was UDP)
The FFmpeg command is as follows :


ffmpeg -i rtsp://the_ip_address/axis-media/media.amp -hide_banner -c:v copy -s 1920x1080 -segment_time 00:30:00 -f segment -strftime 1 -reset_timestamps 1 -rtsp_transport tcp "%Y-%m-%d-%H-%M-%S_Test.mp4"



Please help...