
Recherche avancée
Autres articles (96)
-
MediaSPIP Core : La Configuration
9 novembre 2010, parMediaSPIP Core fournit par défaut trois pages différentes de configuration (ces pages utilisent le plugin de configuration CFG pour fonctionner) : une page spécifique à la configuration générale du squelettes ; une page spécifique à la configuration de la page d’accueil du site ; une page spécifique à la configuration des secteurs ;
Il fournit également une page supplémentaire qui n’apparait que lorsque certains plugins sont activés permettant de contrôler l’affichage et les fonctionnalités spécifiques (...) -
MediaSPIP Player : problèmes potentiels
22 février 2011, parLe lecteur ne fonctionne pas sur Internet Explorer
Sur Internet Explorer (8 et 7 au moins), le plugin utilise le lecteur Flash flowplayer pour lire vidéos et son. Si le lecteur ne semble pas fonctionner, cela peut venir de la configuration du mod_deflate d’Apache.
Si dans la configuration de ce module Apache vous avez une ligne qui ressemble à la suivante, essayez de la supprimer ou de la commenter pour voir si le lecteur fonctionne correctement : /** * GeSHi (C) 2004 - 2007 Nigel McNie, (...) -
Encoding and processing into web-friendly formats
13 avril 2011, parMediaSPIP automatically converts uploaded files to internet-compatible formats.
Video files are encoded in MP4, Ogv and WebM (supported by HTML5) and MP4 (supported by Flash).
Audio files are encoded in MP3 and Ogg (supported by HTML5) and MP3 (supported by Flash).
Where possible, text is analyzed in order to retrieve the data needed for search engine detection, and then exported as a series of image files.
All uploaded files are stored online in their original format, so you can (...)
Sur d’autres sites (10661)
-
Read H264 streaming from Elp H264 with OpenCV + Pyhton
15 janvier 2017, par Francesco PaissanI’m trying to read an udp streaming of a H264 encoded image. The software structure is like follows :
On a BeagleBoneBlack (Ubuntu 16.04) I’ve an Elp H264 cam (see here : https://www.amazon.com/ELP-Support-Android-Windows-Surveillance/dp/B00VDSBH9G ). I stream frames with ffmpeg on a Unicast UDP Stream.
I want to read this images from python and opencv to be able to process them.
I tried with this simple code to see if the cap is opened or not :
import cv2
try:
cap = cv2.VideoCapture("udp://localhost:1234/")
cap.set(CV_CAP_PROP_FOURCC, CV_FOURCC('H', '2', '6', '4'));
except Exception, e:
print str(e)But when I run this script python says :
DtsGetHWFeatures : Create File Failed DtsGetHWFeatures : Create File
Failed Running DIL (3.22.0) Version DtsDeviceOpen : Opening HW in mode
0 DtsDeviceOpen : Create File Failed libva info : VA-API version 0.38.1
libva info : va_getDriverName() returns -1 libva error :
va_getDriverName() failed with unknown libva error,driver_name=(null)
libva info : VA-API version 0.38.1 libva info : va_getDriverName()
returns -1 libva error : va_getDriverName() failed with unknown libva
error,driver_name=(null) libva info : VA-API version 0.38.1 libva info :
va_getDriverName() returns -1 libva error : va_getDriverName() failed
with unknown libva error,driver_name=(null) libva info : VA-API version
0.38.1 libva info : va_getDriverName() returns -1 libva error : va_getDriverName() failed with unknown libva error,driver_name=(null)
GStreamer Plugin : Embedded video playback halted ; module vaapidecode
reported : Could not initialize supporting library. OpenCV
Error : Unspecified error (GStreamer : unable to start pipeline ) in
cvCaptureFromCAM_GStreamer, file /builddir/build/BUILD/opencv-
2.4.12.3/modules/highgui/src/cap_gstreamer.cpp, line 816 /builddir/build/BUILD/opencv-2.4.12.3/modules/highgui/src/cap_gstreamer.cpp:816 :
error : (-2) GStreamer : unable to start pipeline in function
cvCaptureFromCAM_GStreamerCan anybody help me ?
Thanks,
Francesco.
-
PipedInputStream / PipedOutputStream, ImageIO and ffmpeg
19 avril 2015, par jdevelopI have the following code in Scala :
val pos = new PipedOutputStream()
val pis = new PipedInputStream(pos)
Future {
LOG.trace("Start rendering")
generateFrames(videoRenderParams.length) {
img ⇒ ImageIO.write(img, "PNG", pos)
}
pos.flush()
IOUtils.closeQuietly(pos)
LOG.trace("Finished rendering")
} onComplete {
case Success(_) ⇒
LOG.trace("Complete successfully")
case Failure(err) ⇒
LOG.error("Can't render stuff", err)
IOUtils.closeQuietly(pis)
IOUtils.closeQuietly(pos)
}
val prc = (ffmpegCli #< pis).!(logger)the Future simply writes the generated images one by one to the OutputStream. Now the ffmpeg process reads the input images from stdin and converts them to MP4 file.
That works pretty well, but for some reason sometimes I’m getting the following stacktraces :
I/O error Pipe closed for process: <input stream="stream" />
java.io.IOException: Pipe closed
at java.io.PipedInputStream.checkStateForReceive(PipedInputStream.java:260)
at java.io.PipedInputStream.receive(PipedInputStream.java:226)
at java.io.PipedOutputStream.write(PipedOutputStream.java:149)
at scala.sys.process.BasicIO$.loop$1(BasicIO.scala:236)
at scala.sys.process.BasicIO$.transferFullyImpl(BasicIO.scala:242)
at scala.sys.process.BasicIO$.transferFully(BasicIO.scala:223)
at scala.sys.process.ProcessImpl$PipeThread.runloop(ProcessImpl.scala:159)
at scala.sys.process.ProcessImpl$PipeSource.run(ProcessImpl.scala:179)At the same time I’m getting the following error from another stream :
javax.imageio.IIOException: I/O error writing PNG file!
at com.sun.imageio.plugins.png.PNGImageWriter.write(PNGImageWriter.java:1168)
at javax.imageio.ImageWriter.write(ImageWriter.java:615)
at javax.imageio.ImageIO.doWrite(ImageIO.java:1612)
at javax.imageio.ImageIO.write(ImageIO.java:1578)
atSo it seems that the streams were broken somewhere in between, so ffmpeg can not read the data, and ImageIO can not write the data.
What is even more interesting - the problem is reproducible only on certain Linux server (Amazon). It works flawlessly on other Linux boxes. So I wonder if somebody could point me out to the possible causes of this error.
What I’ve tried so far :
- use Oracle JDK 8 and OpenJDK
- use different versions of FFMPEG
Nothing worked by the moment.
-
AWS : Best way to generate a thumbnail for every frame of a s3 uploaded video
4 janvier 2018, par danielfrancaI need to process a video file, transcode it and generate a thumbnail for every frame.
It should happen every time there’s a new video on a specific AWS bucket.
I found out that AWS Lambda should be the best service for that
However, it is not working as expected and I’ll explain why
I’ve created a simple Python2.7 file using FFVideo
It seems that this library doesn’t support Python3.It is a nice abstraction on top of ffmpeg
To deploy the package I had run
lld
on the FFVideo shared object, and then copied everything to my project directory, as described in their documentation.
Zipped it and upload to AWS LambdaYet it doesn’t work, I keep getting errors as if the /usr/lib64/libstdc++ is missing, even after copied it to the projecct dir, also tried /usr/lib64 and /lib64
Then as a second thought I wonder if just running
ffmpeg
wouldn’t be easier...
So I just copied ffmpeg to the project dir and did a simple Python script to call it.Missing shared objects, ok,
lld
again and copied everything to the directory.Then AWS Lambda seems to be completely broken, I can’t save it anymore and it just says "Fix errors before saving"
But no error message, nothingI even have attempted to write inline a simple code, but now AWS Lambda don’t even open the online editor.
I also tried to remove all the shared objects I have added, returning to the original state, but still same generic error.
Same thing if I just create a new lambda function with same old code.Doesn’t matter what I do it never even enable the Save button anymore.
I thought it might be just some AWS unstability, but it been a while.I’ve looked to a similar project using Node
and it doesn’t seem to include anything except ffmpegMy other idea is to use SQS to trigger a python script somewhere else to create the thumbnails
Any idea how is the best approach for that ?