
Recherche avancée
Médias (91)
-
MediaSPIP Simple : futur thème graphique par défaut ?
26 septembre 2013, par
Mis à jour : Octobre 2013
Langue : français
Type : Video
-
avec chosen
13 septembre 2013, par
Mis à jour : Septembre 2013
Langue : français
Type : Image
-
sans chosen
13 septembre 2013, par
Mis à jour : Septembre 2013
Langue : français
Type : Image
-
config chosen
13 septembre 2013, par
Mis à jour : Septembre 2013
Langue : français
Type : Image
-
SPIP - plugins - embed code - Exemple
2 septembre 2013, par
Mis à jour : Septembre 2013
Langue : français
Type : Image
-
GetID3 - Bloc informations de fichiers
9 avril 2013, par
Mis à jour : Mai 2013
Langue : français
Type : Image
Autres articles (64)
-
Submit bugs and patches
13 avril 2011Unfortunately a software is never perfect.
If you think you have found a bug, report it using our ticket system. Please to help us to fix it by providing the following information : the browser you are using, including the exact version as precise an explanation as possible of the problem if possible, the steps taken resulting in the problem a link to the site / page in question
If you think you have solved the bug, fill in a ticket and attach to it a corrective patch.
You may also (...) -
Publier sur MédiaSpip
13 juin 2013Puis-je poster des contenus à partir d’une tablette Ipad ?
Oui, si votre Médiaspip installé est à la version 0.2 ou supérieure. Contacter au besoin l’administrateur de votre MédiaSpip pour le savoir -
HTML5 audio and video support
13 avril 2011, parMediaSPIP uses HTML5 video and audio tags to play multimedia files, taking advantage of the latest W3C innovations supported by modern browsers.
The MediaSPIP player used has been created specifically for MediaSPIP and can be easily adapted to fit in with a specific theme.
For older browsers the Flowplayer flash fallback is used.
MediaSPIP allows for media playback on major mobile platforms with the above (...)
Sur d’autres sites (11604)
-
FFMPEG API — How much do stream parameters change frame-to-frame ?
8 septembre 2015, par kaceyI’m trying to extract raw streams from devices and files using ffmpeg. I notice the crucial frame information (Video : width, height, pixel format, color space, Audio : sample format) is stored both in the AVCodecContext and in the AVFrame. This means I can access it prior to the stream playing and I can access it for every frame.
How much do I need to account for these values changing frame-to-frame ? I found https://ffmpeg.org/doxygen/trunk/demuxing__decoding_8c_source.html#l00081 which indicates that at least width, height, and pixel format may change frame to frame.
- Will the color space and sample format also change frame to frame ?
- Will these changes be temporary (a single frame) or lasting (a significant block of frames) and is there any way to predict for this stream which behavior will occur ?
- Is there a way to find the most descriptive attributes that this stream is possible of producing, such that I can scale all the lower-quality frames up, but not offer a result that is mindlessly higher-quality than the source, even if this is a device or a network stream where I cannot play all the frames in advance ?
The fundamental question is : how do I resolve the flexibility of this API with the restriction that raw streams (my output) do not have any way of specifying a change of stream attributes mid-stream. I imagine I will need to either predict the most descriptive attributes to give the stream, or offer a new stream when the attributes change. Which choice to make depends on whether these values will change rapidly or stay relatively stable.
-
embed video stream with custom meta data
15 mai 2022, par Sergey KolesnikI have an optical system that provides a UDP video stream.


From device specification FAQ :




Both single metadata (KLV) stream and compressed video (H.264) with metadata (KLV) are available on Ethernet link. Compressed video and metadata are coupled in the same stream compliant with STANAG 4609 standard. Each encoded video stream is encapsulated with the associated metadata within an MPEG-TS single program stream over Ethernet UDP/IP/ The video and metadata are synchronized through the use of timestamps.




Also there are other devices that provide data about the state of an aircraft (velocity, coords, etc). This data should be displayed on a client GUI display alongside with video. Of course it has to be synchronized with the current video frame.


One of the approaches I though of is to embed this data into the video stream. But I am not sure if it is possible or should I use another (than UDP) protocol for this purpose.


Is it possible/reasonable to use such approach ? Is ffmpeg library suitable in this case ?
If not, what are the other ways to synchronize data with a video frame.
Latency is crucial. Although bandwidth is limited to 2-5 Mbps.



It seems to be possible using ffmpeg :
AVPacket
can be provided with additional data using functionav_packet_add_side_data
which takes a preallocated buffer, size and a typeAVPacketSideDataType
.
However, I am not sure for now, which enum value ofAVPacketSideDataType
can be used for custom user-provided binary data.

Something similar that might be used for my needs :


How do I encode KLV packets to an H.264 video using libav*


-
How to use prebuilt FFmpeg in Android Studio
26 mai 2016, par vxh.vietI’m sure this is a very basic question but since this is the my first time messing around with the NDK, a lot of thing is still very unclear to me.
Use case :
-
I’m trying to develop a video scrubbing feature so fast and accurate frame seeking is crucial. I’ve tried most of the available players out there but the performance is still not up to my demand. That’s why I’m going down the FFmpeg route.
-
Basically, what I’m looking for is FFmpeg input seeking. I’ve tried WrtingMinds’ ffmpeg-android-java. However it is a file based implementation which means the
out.jpg
need to be written to external memory and read back which has a big hit on performance (roughly 1000 milliseconds for 1 seek). -
That’s why I’m trying to built my own FFmpeg player to do the input seeking in JNI and push back the
byte[]
to be displayed in Java.
Question : After a lot of struggling with the NDK, I’ve managed to set it up and successfully calling the JNI method from my Java code. The structure is as below :
MyApp
-app
-MyFFmpegPlayer
-build
-libs
-src
-main
-java
-com.example.myffmpegplayer
+HelloJNI.java
-jni
+MyFFmpegPlayer.cAfter some fail attempt to build FFmpeg on Windows, I’ve decided to use WritingMinds prebuilt FFmpeg. However, after extraction they just come up as plain
ffmpeg
files (not.so
file) so I don’t really know how to use these.It would be a great gratitude, if someone can just chime in and give me a good starting point for my next step.
Thank you so much for your time.
-