
Recherche avancée
Médias (29)
-
#7 Ambience
16 octobre 2011, par
Mis à jour : Juin 2015
Langue : English
Type : Audio
-
#6 Teaser Music
16 octobre 2011, par
Mis à jour : Février 2013
Langue : English
Type : Audio
-
#5 End Title
16 octobre 2011, par
Mis à jour : Février 2013
Langue : English
Type : Audio
-
#3 The Safest Place
16 octobre 2011, par
Mis à jour : Février 2013
Langue : English
Type : Audio
-
#4 Emo Creates
15 octobre 2011, par
Mis à jour : Février 2013
Langue : English
Type : Audio
-
#2 Typewriter Dance
15 octobre 2011, par
Mis à jour : Février 2013
Langue : English
Type : Audio
Autres articles (74)
-
Des sites réalisés avec MediaSPIP
2 mai 2011, parCette page présente quelques-uns des sites fonctionnant sous MediaSPIP.
Vous pouvez bien entendu ajouter le votre grâce au formulaire en bas de page. -
Support de tous types de médias
10 avril 2011Contrairement à beaucoup de logiciels et autres plate-formes modernes de partage de documents, MediaSPIP a l’ambition de gérer un maximum de formats de documents différents qu’ils soient de type : images (png, gif, jpg, bmp et autres...) ; audio (MP3, Ogg, Wav et autres...) ; vidéo (Avi, MP4, Ogv, mpg, mov, wmv et autres...) ; contenu textuel, code ou autres (open office, microsoft office (tableur, présentation), web (html, css), LaTeX, Google Earth) (...)
-
Changer son thème graphique
22 février 2011, parLe thème graphique ne touche pas à la disposition à proprement dite des éléments dans la page. Il ne fait que modifier l’apparence des éléments.
Le placement peut être modifié effectivement, mais cette modification n’est que visuelle et non pas au niveau de la représentation sémantique de la page.
Modifier le thème graphique utilisé
Pour modifier le thème graphique utilisé, il est nécessaire que le plugin zen-garden soit activé sur le site.
Il suffit ensuite de se rendre dans l’espace de configuration du (...)
Sur d’autres sites (8232)
-
TCL script unable to work with multiple camera on windows 7 platform
27 novembre 2017, par M. D. PMy TCL script works properly for single camera. the TCL script is as follow :
proc liveStreaming {} {
#open the config file.
set f [open "C:/main/video_config.txt" r]
#To retrive the values from the config file.
while {![eof $f]} {
set part [split [gets $f] "="]
set props([string trimright [lindex $part 0]]) [string trimleft [lindex $part 1]]
}
close $f
#camera selection to live streaming.
set camera "video"
append cctv $camera "=" $props(cctv)
#ffmpeg command to capture live streaming in background
exec ffmpeg -f dshow -s 1280x720 -i $cctv c:/test/sample.avi >& c:/test/temp.txt &
}
liveStreaming
//above code is to capture video using ffmpeg and tcl.My text file "video_config" is as follow :
cctv=Logitech HD Webcam C525
as, i want it to run with multiple camera : the text file should be as follow :
cctv=Integrated Webcam,Logitech HD Webcam C525
The problem is "," (comma), my TCL script is unable to predict the coma (,) . can any one provide me the proper TCL script which can predict the coma, so that my TCL script can work with multiple camera.
Error Report :
ffmpeg version N-89127-g8f4702a93f Copyright (c) 2000-2017 the FFmpeg developers
built with gcc 7.2.0 (GCC)
configuration: --enable-gpl --enable-version3 --enable-sdl2 --enable-bzlib --enable-fontconfig --enable-gnutls --enable-iconv --enable-libass --enable-libbluray --enable-libfreetype --enable-libmp3lame --enable-libopenjpeg --enable-libopus --enable-libshine --enable-libsnappy --enable-libsoxr --enable-libtheora --enable-libtwolame --enable-libvpx --enable-libwavpack --enable-libwebp --enable-libx264 --enable-libx265 --enable-libxml2 --enable-libzimg --enable-lzma --enable-zlib --enable-gmp --enable-libvidstab --enable-libvorbis --enable-cuda --enable-cuvid --enable-d3d11va --enable-nvenc --enable-dxva2 --enable-avisynth --enable-libmfx
libavutil 56. 0.100 / 56. 0.100
libavcodec 58. 3.103 / 58. 3.103
libavformat 58. 2.100 / 58. 2.100
libavdevice 58. 0.100 / 58. 0.100
libavfilter 7. 2.100 / 7. 2.100
libswscale 5. 0.101 / 5. 0.101
libswresample 3. 0.101 / 3. 0.101
libpostproc 55. 0.100 / 55. 0.100
[dshow @ 05944ac0] Could not find video device with name [Integrated Webcam,Logitech HD Webcam C525] among source devices of type video.
video=Integrated Webcam,Logitech HD Webcam C525: I/O error -
How to stream live video from DJI Professional 3 camera ?
29 avril 2017, par raullalvesI have to get the live stream video from DJI Phantom 3 camera in my C++ application, in order to do a Computer Vision processing in OpenCV.
First I tried sending the H264 raw data through an UDP socket, inside this callback :
mReceivedVideoDataCallBack = new CameraReceivedVideoDataCallback() {
@Override
public void onResult(byte[] videoBuffer, int size) {
//Here, I call a method from a class I created, that sends the buffer through UDP
if (gravar_trigger) controleVideo.enviarFrame(videoBuffer, size);
if (mCodecManager != null) mCodecManager.sendDataToDecoder(videoBuffer, size);
}
};That communication above works well. However, I haven’t been able to decode that UDP H264 data in my C++ desktop application. I have tested with FFmpeg lib, but couldn’t get to alocate an
AVPacket
with my UDP data, in order to decode usingavcodec_send_packet
andavcodec_receive_frame
. I also had problems withAVCodecContext
, since my UDP communication wasn’t a stream like RTSP, where it could get information about its source. Therefore, I had to change how I was trying to solve the problem.Then, I found libstreaming, in which can be associate to stream the android video camera to a Wowza Server, creating something like a RTSP stream connection, where the data could be obtained in my final C++ application easily using OpenCV
videoCapture
. However, libstreaming uses its ownsurfaceView
. In other words, I would have to link the libstreamingsurfaceView
with the DJI Drone’svideoSurface
. I’m really new to Android, so don’t have any clue of how to do that.To sum up, is that the correct approach ? Someone has a better idea ? Thanks in advance
-
How to stream live video from DJI Professional 3 camera ?
5 juin 2017, par raullalvesI have to get the live stream video from DJI Phantom 3 camera in my C++ application, in order to do a Computer Vision processing in OpenCV.
First I tried sending the H264 raw data through an UDP socket, inside this callback :
mReceivedVideoDataCallBack = new CameraReceivedVideoDataCallback() {
@Override
public void onResult(byte[] videoBuffer, int size) {
//Here, I call a method from a class I created, that sends the buffer through UDP
if (gravar_trigger) controleVideo.enviarFrame(videoBuffer, size);
if (mCodecManager != null) mCodecManager.sendDataToDecoder(videoBuffer, size);
}
};That communication above works well. However, I haven’t been able to decode that UDP H264 data in my C++ desktop application. I have tested with FFmpeg lib, but couldn’t get to alocate an
AVPacket
with my UDP data, in order to decode usingavcodec_send_packet
andavcodec_receive_frame
. I also had problems withAVCodecContext
, since my UDP communication wasn’t a stream like RTSP, where it could get information about its source. Therefore, I had to change how I was trying to solve the problem.Then, I found libstreaming, in which can be associate to stream the android video camera to a Wowza Server, creating something like a RTSP stream connection, where the data could be obtained in my final C++ application easily using OpenCV
videoCapture
. However, libstreaming uses its ownsurfaceView
. In other words, I would have to link the libstreamingsurfaceView
with the DJI Drone’svideoSurface
. I’m really new to Android, so don’t have any clue of how to do that.To sum up, is that the correct approach ? Someone has a better idea ? Thanks in advance