Recherche avancée

Médias (1)

Mot : - Tags -/MediaSPIP

Autres articles (47)

  • Personnaliser en ajoutant son logo, sa bannière ou son image de fond

    5 septembre 2013, par

    Certains thèmes prennent en compte trois éléments de personnalisation : l’ajout d’un logo ; l’ajout d’une bannière l’ajout d’une image de fond ;

  • MediaSPIP v0.2

    21 juin 2013, par

    MediaSPIP 0.2 est la première version de MediaSPIP stable.
    Sa date de sortie officielle est le 21 juin 2013 et est annoncée ici.
    Le fichier zip ici présent contient uniquement les sources de MediaSPIP en version standalone.
    Comme pour la version précédente, il est nécessaire d’installer manuellement l’ensemble des dépendances logicielles sur le serveur.
    Si vous souhaitez utiliser cette archive pour une installation en mode ferme, il vous faudra également procéder à d’autres modifications (...)

  • Mise à disposition des fichiers

    14 avril 2011, par

    Par défaut, lors de son initialisation, MediaSPIP ne permet pas aux visiteurs de télécharger les fichiers qu’ils soient originaux ou le résultat de leur transformation ou encodage. Il permet uniquement de les visualiser.
    Cependant, il est possible et facile d’autoriser les visiteurs à avoir accès à ces documents et ce sous différentes formes.
    Tout cela se passe dans la page de configuration du squelette. Il vous faut aller dans l’espace d’administration du canal, et choisir dans la navigation (...)

Sur d’autres sites (5487)

  • Truly live streaming to Android/iPhone

    4 juillet 2012, par Tsaukpaetra

    I have spent quite a while (past week) trying this to little avail. However, what I want seems completely unheard of. So far, I have reviewed recommendations available through google, which include encoding a static file into multiple static files in different formats, creating a playlist that hosts static files in an m3u8 file (files which get added to the playlist as streaming continues).
    I have also seen ideas involving rtmp, rtsp etc which are completely out of the question because of their incompatibility.
    Ideally, I would have one webpage that would link to the stream (http://server/video.mp4) and/or show it in a webpage (via the video tag). With that in mind, the most likely format would be h264+aac in mp4 container.

    Unfortunately, (and probably because the file has no duration metadata) it does not work. I can use a desktop player (such as VLC) to open the stream and play it, but my iPhone and Android both give their respective "Can't be played" messages.

    I don't think the problem is caused by the devices' ability to stream, for I have made a streaming shoutcast server work just fine (mp3 only).

    Currently, the closest I have become is using the following setup on my win32 machine :

    FFMPEG Command: : ffmpeg -f dshow -i video="Logitech Webcam 200":audio="Microphone (Webcam 200)" -b:v 180k -bt 240k -vcodec libx264 -tune zerolatency -profile:v baseline -preset ultrafast -r 10 -strict -2 -acodec aac -ac 2 -ar 48000 -ab 32k -f flv "udp ://127.0.0.1:1234"

    VLC: : Stream from udp ://127.0.0.1:1234 to http:// :8080/video.mp4 (No Transcoding), basically just to convert the UDP stream into an http-accessible stream.

    Any hints or suggestions would be warmly welcomed !

  • ffmpeg audio and the iphone

    17 juin 2019, par michelle

    has anyone been able to make ffmpeg work with audio queues, I get an error when I try to create the queue.

    ret = avcodec_open(enc, codec);

    if (ret < 0) {
       NSLog(@"Error: Could not open video decoder: %d", ret);
       av_close_input_file(avfContext);
       return;
    }


    if (audio_index >= 0) {
       AudioStreamBasicDescription
       audioFormat;

       audioFormat.mFormatID = -1;

       audioFormat.mSampleRate =
       avfContext->streams[audio_index]->codec->sample_rate;

       audioFormat.mFormatFlags = 0;

       switch (avfContext->streams[audio_index]->codec->codec_id)
       {
       case CODEC_ID_MP3:
           audioFormat.mFormatID = kAudioFormatMPEGLayer3;
           break;

       case CODEC_ID_AAC:
           audioFormat.mFormatID = kAudioFormatMPEG4AAC;
           audioFormat.mFormatFlags = kMPEG4Object_AAC_Main;
           break;

       case CODEC_ID_AC3:
           audioFormat.mFormatID = kAudioFormatAC3;
           break;
       default:
           break;
       }

       if (audioFormat.mFormatID != -1) {
           audioFormat.mBytesPerPacket = 0;
           audioFormat.mFramesPerPacket =
           avfContext->streams[audio_index]->codec->frame_size;

           audioFormat.mBytesPerFrame = 0;

           audioFormat.mChannelsPerFrame = avfContext->streams[audio_index]->codec->channels;

           audioFormat.mBitsPerChannel = 0;

           if (ret = AudioQueueNewOutput(&audioFormat, audioQueueOutputCallback, self, NULL, NULL, 0, &audioQueue)) {

               NSLog(@"Error creating audio output queue: %d", ret);
           }

    The issues only with the audio,

    Video is perfect if only I can figure out how to get audio queues to work.

    http://web.me.com/cannonwc/Site/Photos_6.html

    I though of remoteio but there is’nt much doc on that.

    I will share the code for the complete class with anyone that helps me get it to work.

    The idea is to have a single view controller that plays any streaming video passed to it, similar to ffplay on the iphone but without the sdl overhead.

  • Sharing AVPacket data in shared memory

    13 décembre 2023, par Salvo Passaro

    I'm redesigning an application (heavily based on FFmpeg) moving some of the video muxing to a separate process from where the demuxing happens.

    


    Until now I had a pool of threads doing demuxing work, producing AVPackets to be inserted in a queue to be later consumed by another pool of threads doing muxing work.

    


    Now, I'd like to maintain a shared memory approach for the AVPacket queueing between processes but I'm not sure how to properly allocate the internal AVPacket buffer in shared memory.

    


    IPC's probably gonna be enabled by something like this