Recherche avancée

Médias (91)

Autres articles (97)

  • MediaSPIP 0.1 Beta version

    25 avril 2011, par

    MediaSPIP 0.1 beta is the first version of MediaSPIP proclaimed as "usable".
    The zip file provided here only contains the sources of MediaSPIP in its standalone version.
    To get a working installation, you must manually install all-software dependencies on the server.
    If you want to use this archive for an installation in "farm mode", you will also need to proceed to other manual (...)

  • Multilang : améliorer l’interface pour les blocs multilingues

    18 février 2011, par

    Multilang est un plugin supplémentaire qui n’est pas activé par défaut lors de l’initialisation de MediaSPIP.
    Après son activation, une préconfiguration est mise en place automatiquement par MediaSPIP init permettant à la nouvelle fonctionnalité d’être automatiquement opérationnelle. Il n’est donc pas obligatoire de passer par une étape de configuration pour cela.

  • ANNEXE : Les plugins utilisés spécifiquement pour la ferme

    5 mars 2010, par

    Le site central/maître de la ferme a besoin d’utiliser plusieurs plugins supplémentaires vis à vis des canaux pour son bon fonctionnement. le plugin Gestion de la mutualisation ; le plugin inscription3 pour gérer les inscriptions et les demandes de création d’instance de mutualisation dès l’inscription des utilisateurs ; le plugin verifier qui fournit une API de vérification des champs (utilisé par inscription3) ; le plugin champs extras v2 nécessité par inscription3 (...)

Sur d’autres sites (7134)

  • lavu/opencl : replace va_ext.h with standard name

    23 novembre 2018, par Ruiling Song
    lavu/opencl : replace va_ext.h with standard name
    

    Khronos OpenCL header (https://github.com/KhronosGroup/OpenCL-Headers)
    uses cl_va_api_media_sharing_intel.h. And Intel's official OpenCL driver
    for Intel GPU (https://github.com/intel/compute-runtime) was compiled
    against Khronos OpenCL header. So it's better to align with Khronos.

    Signed-off-by : Ruiling Song <ruiling.song@intel.com>

    • [DH] configure
    • [DH] libavutil/hwcontext_opencl.c
  • How to pack the pyav.packet and distribute to another computer

    31 juillet 2024, par lambertk

    I'm currently working on projects which needs to read frames from RTSP server on single entry of computer, do some preprocessing and distribute these frames with preprocessed metadata to different backend for different purpose.

    &#xA;

    And after googling, I found that PyAV could be the solution which can retrieve the video from RTSP source and make it packets, which could possibly be sent to another computer.

    &#xA;

    Considering the network bandwidth, transmit the packets instead of the decoded frames could be better choice.

    &#xA;

    But now comes the problem, socket/MQ, usually only allows to send bytes or string.
    &#xA;Encode the PyAV.packet.Packet object into byte is easy by bytes(packet), but I couldn't find out the way to decode it back to PyAV.packet.Packet object.

    &#xA;

    I've tried to use pickle to serialize the packet, but this method is not implemented in PyAV, and was rejected by the official team.

    &#xA;

    I've also tried to use another package called msgpack, which also failed to serialize the packet.

    &#xA;

    I've tried the following code after reading the source code of PyAV

    &#xA;

    packet_bytes = bytes(packet)&#xA;pt = av.packet.Packet(len(packet_bytes))&#xA;pt.update(packet_bytes)&#xA;

    &#xA;

    the update function seems did not update anything

    &#xA;

    Is there anyway to decode the bytes back to packet object ?

    &#xA;

    Or, can someone give out a way to encode the frame packet and the preprocessed metadata (which is differ frame by frame) together (like H264 SEI Message, which I tried, but could not be inserted when using Python) then send to backend ?

    &#xA;

  • ffmpeg simple 5fps webcam capturing but results in poor performance

    16 septembre 2022, par irous

    I'm using Windows 10, intel core i5, 8GB ram. and download ffmpeg from official site : https://www.gyan.dev/ffmpeg/builds/

    &#xA;

    I'm trying to capture 5fps video from webcam but the output fps is only 3.4 and speed is 0.668x. (I also tried with libx264 but the speed is same 0.5).
    &#xA;This is my command :

    &#xA;

    ffmpeg ^&#xA;  -f dshow -r 5 -s 640x480 -i video="Sony Visual Communication Camera" ^&#xA;  -f rawvideo output.raw&#xA;

    &#xA;

    screenshot of console output :&#xA;enter image description here

    &#xA;

    The cpu usage for above ffmpeg command is about 0.1% and also ram/disk usage is relatively low.
    &#xA;If I change the fps to 30 (-r 30), the output is : fps 7.9, speed 0.261x and CPU usage is still very low 0.1%.

    &#xA;

    But I can easily record 30fps video with the same format yuyv422 in Directshow GraphEdit :&#xA;enter image description here

    &#xA;

    So my question is why does ffmpeg show such poor performance even with 5fps video capturing ? If it can only produce 3.4 fps in the first case why it can output up to 7.9 fps in the latter case (when I set desired fps to 30) ? Did I mis-config ffmpeg ?
    &#xA;Any help is appreciated.

    &#xA;