Recherche avancée

Médias (0)

Mot : - Tags -/xml-rpc

Aucun média correspondant à vos critères n’est disponible sur le site.

Autres articles (59)

  • Des sites réalisés avec MediaSPIP

    2 mai 2011, par

    Cette page présente quelques-uns des sites fonctionnant sous MediaSPIP.
    Vous pouvez bien entendu ajouter le votre grâce au formulaire en bas de page.

  • Les autorisations surchargées par les plugins

    27 avril 2010, par

    Mediaspip core
    autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs

  • Personnaliser les catégories

    21 juin 2013, par

    Formulaire de création d’une catégorie
    Pour ceux qui connaissent bien SPIP, une catégorie peut être assimilée à une rubrique.
    Dans le cas d’un document de type catégorie, les champs proposés par défaut sont : Texte
    On peut modifier ce formulaire dans la partie :
    Administration > Configuration des masques de formulaire.
    Dans le cas d’un document de type média, les champs non affichés par défaut sont : Descriptif rapide
    Par ailleurs, c’est dans cette partie configuration qu’on peut indiquer le (...)

Sur d’autres sites (7028)

  • Anomalie #2985 (Nouveau) : Les docs attachés à un auteur ne sont pas visibles pour les rédacs dans...

    26 avril 2013, par Suske -

    Voir http://forum.spip.net/fr_251736.html

    Je confirme.

    Par contre ils s’afficheraient dans le public, ce que je veux bien croire mais pas testé.

    Une question d’autorisation j’imagine.

  • runtime error when linking ffmpeg libraries in qt creator

    6 juillet 2012, par dxthegreat

    I'm quite new around here but i hear that if you want a question answered, stackoverflow is the place to ask it >.<. So i hope that my question isn't too trivial that everyone will get annoyed at my lack of research (I've tried googling for two days already D= no progress !)
    I've also asked this question in the Qt forums, but i figured i'd ask here too.

    so...

    For the last few days I’ve been fiddling around with opengl and the like, trying to write a video player.

    However, when i try to import the ffmpeg libraries (avcodec, avformat, avutils etc), an error occurs on runtime (the program compiles fine). When compiled and run in debug mode, the error message gives me only a memory address and error code 135 (DLL not found).

    This error occurs when i include a function from those libraries in my code (e.g. av_register_all()) and it occurs regardless of whether the function is actually called.

    So i’m thinking that I’m doing something wrong when linking these libraries.
    I’m currently using :
    Windows vista (32bit),
    Qt creator 2.4.1 based on Qt 4.7.4 (32bit),
    Zeranoe’s FFmpeg build git-3233ad4 (2012-06-30)

    My .pro file consists of :

    QT       += core gui opengl

    TARGET = test
    TEMPLATE = app


    SOURCES += main.cpp\
           mainwindow.cpp \
       glwidget.cpp

    HEADERS += mainwindow.h \
       glwidget.h \

    FORMS    += mainwindow.ui


    LIBS += -L"$$_PRO_FILE_PWD_/libraries/ffmpeg/libs/" -lavcodec -lavformat -lavutil
    INCLUDEPATH += libraries/ffmpeg/includes

    I’ve tried many variations to the LIBS += line and checked my filepath many times. However, the DLL not found error occurs in all of these variations =(.

    Is there something I’m forgetting when doing these includes ?

    Thanks in advance >.<,
    (young and naive) aspiring dev

  • How to build and link FFMPEG to iOS ?

    30 juin 2015, par Alexander Tkachenko

    all !

    I know, there are a lot of questions here about FFMPEG on iOS, but no one answer is appropriate for my case :(
    Something strange happens each case when I am trying to link FFMPEG in my project, so please, help me !

    My task is to write video-chat application for iOS, that uses RTMP-protocol for publishing and reading video-stream to/from custom Flash Media Server.

    I decided to use rtmplib, free open-source library for streaming FLV video over RTMP, as it is the only appropriate library.

    Many problem appeared when I began research of it, but later I understood how it should work.

    Now I can read live stream of FLV video(from url) and send it back to channel, with the help of my application.

    My trouble now is in sending video FROM Camera.
    Basic operations sequence, as I understood, should be the following :

    1. Using AVFoundation, with the help of sequence (Device-AVCaptureSession-AVVideoDataOutput-> AVAssetWriter) I write this to a file(If you need, I can describe this flow more detailed, but in the context of question it is not important). This flow is necessary to make hardware-accelerated conversion of live video from the camera into H.264 codec. But it is in MOV container format. (This is completed step)

    2. I read this temporary file with each sample written, and obtain the stream of bytes of video-data, (H.264 encoded, in QuickTime container). (this is allready completed step)

    3. I need to convert videodata from QuickTime container format to FLV. And it all in real-time.(packet - by - packet)

    4. If i will have the packets of video-data, contained in FLV container format, I will be able to send packets over RTMP using rtmplib.

    Now, the most complicated part for me, is step 3.

    I think, I need to use ffmpeg lib to this conversion (libavformat). I even found the source code, showing how to decode h.264 data packets from MOV file (looking in libavformat, i found that it is possible to extract this packets even from byte stream, which is more appropriate for me). And having this completed, I will need to encode packets into FLV(using ffmpeg or manually, in a way of adding FLV-headers to h.264 packets, it is not problem and is easy, if I am correct).

    FFMPEG has great documentation and is very powerfull library, and I think, there won’t be a problem to use it. BUT the problem here is that I can not got it working in iOS project.

    I have spend 3 days reading documentation, stackoverflow and googling the answer on the question "How to build FFMPEG for iOS" and I think, my PM is gonna fire me if I will spend one more week on trying to compile this library :))

    I tried to use many different build-scripts and configure files, but when I build FFMPEG, i Got libavformat, libavcodec, etc. for x86 architecture (even when I specify armv6 arch in build-script). (I use "lipo -info libavcodec.a" to show architectures)

    So I cannot build this sources, and decided to find prebuilt FFMPEG, that is build for architecture armv7, armv6, i386.

    I have downloaded iOS Comm Lib from MidnightCoders from github, and it contains example of usage FFMPEG, it contains prebuilt .a files of avcodec,avformat and another FFMPEG libraries.

    I check their architecture :

    iMac-2:MediaLibiOS root# lipo -info libavformat.a
    Architectures in the fat file: libavformat.a are: armv6 armv7 i386

    And I found that it is appropriate for me !
    When I tried to add this libraries and headers to xCode project, It compiles fine(and I even have no warnings like "Library is compiled for another architecture"), and I can use structures from headers, but when I am trying to call C-function from libavformat (av_register_all()), the compiler show me error message "Symbol(s) not found for architecture armv7 : av_register_all".

    I thought, that maybe there are no symbols in lib, and tried to show them :

    root# nm -arch armv6 libavformat.a | grep av_register_all
    00000000 T _av_register_all

    Now I am stuck here, I don’t understand, why xCode can not see this symbols, and can not move forward.

    Please, correct me if I am wrong in the understanding of flow for publishing RTMP-stream from iOS, and help me in building and linking FFMPEG for iOS.

    I have iPhone 5.1. SDK and xCode 4.2.