
Recherche avancée
Médias (1)
-
Video d’abeille en portrait
14 mai 2011, par
Mis à jour : Février 2012
Langue : français
Type : Video
Autres articles (104)
-
Support audio et vidéo HTML5
10 avril 2011MediaSPIP utilise les balises HTML5 video et audio pour la lecture de documents multimedia en profitant des dernières innovations du W3C supportées par les navigateurs modernes.
Pour les navigateurs plus anciens, le lecteur flash Flowplayer est utilisé.
Le lecteur HTML5 utilisé a été spécifiquement créé pour MediaSPIP : il est complètement modifiable graphiquement pour correspondre à un thème choisi.
Ces technologies permettent de distribuer vidéo et son à la fois sur des ordinateurs conventionnels (...) -
HTML5 audio and video support
13 avril 2011, parMediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
For older browsers the Flowplayer flash fallback is used.
MediaSPIP allows for media playback on major mobile platforms with the above (...) -
De l’upload à la vidéo finale [version standalone]
31 janvier 2010, parLe chemin d’un document audio ou vidéo dans SPIPMotion est divisé en trois étapes distinctes.
Upload et récupération d’informations de la vidéo source
Dans un premier temps, il est nécessaire de créer un article SPIP et de lui joindre le document vidéo "source".
Au moment où ce document est joint à l’article, deux actions supplémentaires au comportement normal sont exécutées : La récupération des informations techniques des flux audio et video du fichier ; La génération d’une vignette : extraction d’une (...)
Sur d’autres sites (9937)
-
Revision f167433d9c : fix the mv_ref_idx issue The following issue was reported : https://code.google
20 août 2013, par Jim BankoskiChanged Paths :
Modify /vp9/common/vp9_mvref_common.c
fix the mv_ref_idx issueThe following issue was reported :
https://code.google.com/p/webm/issues/detail?id=601&q=jimbankoski&sort=-id&colsp
ec=ID%20Pri%20mstone%20ReleaseBlock%20Type%20Component%20Status%20Owner%20Summar
yThis code makes the choice and code cleaner and removes any question
about whether the border needs to be checked.Change-Id : Ia7aecfb3168e340618805bd318499176c2989597
-
how to us google ML kit Selfie segmentation in flutter to overlay segmented user on top of chosen image or video ? [closed]
4 avril 2024, par AmrI'm working on implementing a virtual background feature using the Google ML Kit selfie segmentation Flutter plugin. The goal is to separate the user from the background live from the camera feed, overlay the user on a chosen image or video, and then create a new video combining these elements.


Currently, I'm successfully obtaining the mask using the plugin. However, I'm concerned about performance issues if I directly copy pixels to achieve the overlay effect, especially in real-time scenarios.


I'm looking for alternative approaches or best practices to efficiently achieve this functionality without sacrificing performance. Any insights or suggestions on how to approach this would be greatly appreciated.


I thought about using ffmpeg but not sure how to take the user pixels from the camera feed, I am not a flutter developer so not sure how to do so.


-
google function : TypeError : Cannot read property 'name' of undefined
9 décembre 2019, par JuggernaughtI’m trying to use Google cloud function to transcode videos from a storage bucket and outputs into another bucket using code posted online but with a few adjustments.
but I get the following error :
TypeError : Cannot read property ’name’ of undefined at transcodeVideo
(/srv/index.js:17:56) at /worker/worker.js:825:24 at at
process._tickDomainCallback (internal/process/next_tick.js:229:7)Index.js
const {Storage} = require('@google-cloud/storage');
const projectId = 'cc18-223318';
const storage = new Storage({
projectId: projectId,
});
const ffmpegPath = require('@ffmpeg-installer/ffmpeg').path;
const ffmpeg = require('fluent-ffmpeg');
const transcodedBucket = storage.bucket('2400p');
const uploadBucket = storage.bucket('inpuut');
ffmpeg.setFfmpegPath(ffmpegPath);
exports.transcodeVideo = function transcodeVideo(event, callback) {
const file = event.data;
// Ensure that you only proceed if the file is newly created, and exists.
if (file.metageneration !== '1' || file.resourceState !== 'exists') {
callback();
return;
}
// Open write stream to new bucket, modify the filename as needed.
const remoteWriteStream = transcodedBucket.file(file.name.replace('.webm', '.mp4'))
.createWriteStream({
metadata: {
metadata: file.metadata, // You may not need this, my uploads have associated metadata
contentType: 'video/mp4', // This could be whatever else you are transcoding to
},
});
// Open read stream to our uploaded file
const remoteReadStream = uploadBucket.file(file.name).createReadStream();
// Transcode
ffmpeg()
.input(remoteReadStream)
.outputOptions('-c:v libx264') // Change these options to whatever suits your needs
.outputOptions('-c:a copy')
.outputOptions('-vf "scale=4800:2400"')
.outputOptions('-b:a 160k')
.outputOptions('-b:a 160k')
.outputOptions('-x264-params mvrange=511')
.outputOptions('-f mp4')
.outputOptions('-preset slow')
.outputOptions('-movflags frag_keyframe+empty_moov')
.outputOptions('-crf 18')
// https://github.com/fluent-ffmpeg/node-fluent-ffmpeg/issues/346#issuecomment-67299526
.on('start', (cmdLine) => {
console.log('Started ffmpeg with command:', cmdLine);
})
.on('end', () => {
console.log('Successfully re-encoded video.');
callback();
})
.on('error', (err, stdout, stderr) => {
console.error('An error occured during encoding', err.message);
console.error('stdout:', stdout);
console.error('stderr:', stderr);
callback(err);
})
.pipe(remoteWriteStream, { end: true }); // end: true, emit end event when readable stream ends
};