
Recherche avancée
Autres articles (44)
-
Les autorisations surchargées par les plugins
27 avril 2010, parMediaspip core
autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs -
Contribute to documentation
13 avril 2011Documentation is vital to the development of improved technical capabilities.
MediaSPIP welcomes documentation by users as well as developers - including : critique of existing features and functions articles contributed by developers, administrators, content producers and editors screenshots to illustrate the above translations of existing documentation into other languages
To contribute, register to the project users’ mailing (...) -
Prérequis à l’installation
31 janvier 2010, parPréambule
Cet article n’a pas pour but de détailler les installations de ces logiciels mais plutôt de donner des informations sur leur configuration spécifique.
Avant toute chose SPIPMotion tout comme MediaSPIP est fait pour tourner sur des distributions Linux de type Debian ou dérivées (Ubuntu...). Les documentations de ce site se réfèrent donc à ces distributions. Il est également possible de l’utiliser sur d’autres distributions Linux mais aucune garantie de bon fonctionnement n’est possible.
Il (...)
Sur d’autres sites (4050)
-
how to add text to the video in android ?
18 septembre 2015, par Sajad NorouziIn my app I record a video and I wanna overlay the url of my company on the right bottom of that video. I record video with MediaRecorder and there is no problem wiht this part but for adding text to the video, I wasn’t successful. I searched and found some library like ffmpeg and javaCV which let user overlay a text on the video but I found using them in the android hard. after that found OnPreviewFrame which returns each frame of the camerapreivew. I want to get all frames after user touch record button and draw text using canvas to each frame and then save all those frame together until recording video stops, and videos duration are fix and 15 seconds. Is it a good solution ? if it is, how should I implement OnPreviewFrame function ? I know how to draw canvas but I don’t know how to convert data[] to the bitmap.
-
Latency issue with CMU Sphinx 4
22 septembre 2015, par vijaymI have written the speech recognition application using CMU sphinx 4 and followed the details from this link. I have defined the Acoustic,Dictionary and Language Model as below
configuration.setAcousticModelPath("resource:/edu/cmu/sphinx/models/en-us/en-us");
configuration.setDictionaryPath("resource:/edu/cmu/sphinx/models/en-us/cmudict-en-us.dict");
configuration.setLanguageModelPath("resource:/edu/cmu/sphinx/models/en-us/en-us.lm.bin");With the above configuration the 20 minutes of wav file takes almost close to 20 minutes to do the transcription.Hence than I tried to pass the user defined config.xml. I did n’t find the configuration manager option to pass the user defined config.xml with the current version of Sphinx4.Then I had written own recognizer by extending the
AbstractSpeechRecognizer.java
class(It may be useless) and changed few parameters of config.xml and I tried it but still no improvement.I have downloaded video and audio across multiple source and converted into WAV file using FFMPEG
The command is as below
ffmpeg -i input.mp3 -acodec pcm_s16le -ac 1 -ar 16000 output.wav
Environment Details :
Java 8
Ubuntu 14.04
RAM 4GB
I5 Processor
What I would like to know is, what I am missing here and how to improve the performance ?
-
Chrome "stalling" when streaming mp3 file from nodejs windows only
8 octobre 2015, par Alan HollisWe’ve got a really annoying bug when trying to send mp3 data. We’ve got the following set up.
Web cam producing aac -> ffmpeg convert to adts -> send to nodejs server -> ffmpeg on server converts adts to mp3 -> mp3 then streamed to browser.
This works *perfectly" on Linux ( chrome with HTML5 and flash, firefox flash only )
However on windows the sound just "stalls", no matter what combination we use ( browser/html5/flash ). If however we shutdown the server the sound then immediately starts to play as we expect.
For some reason on windows based machines it’s as if the sound is being buffered "waiting" for something but we don’t know what that is.
Any help would be greatly appreciated.
Relevant code in node
res.setHeader('Connection', 'Transfer-Encoding');
res.setHeader('Content-Type', 'audio/mpeg');
res.setHeader('Transfer-Encoding', 'chunked');
res.writeHeader('206');
that.eventEmitter.on('soundData', function (data) {
debug("Got sound data" + data.cameraId + " " + req.params.camera_id);
if (req.params.camera_id == data.cameraId) {
debug("Sending data direct to browser");
res.write(data.sound);
}
});Code on browser
soundManager.setup({
url: 'http://dashboard.agricamera.co.uk/themes/agricamv2/swf/soundmanager2.swf',
useHTML5Audio: false,
onready: function () {
that.log("Sound manager is now ready")
var mySound = soundManager.createSound({
url: src,
autoLoad: true,
autoPlay: true,
stream: true,
});
}
});