Recherche avancée

Médias (9)

Mot : - Tags -/soundtrack

Autres articles (54)

  • MediaSPIP v0.2

    21 juin 2013, par

    MediaSPIP 0.2 est la première version de MediaSPIP stable.
    Sa date de sortie officielle est le 21 juin 2013 et est annoncée ici.
    Le fichier zip ici présent contient uniquement les sources de MediaSPIP en version standalone.
    Comme pour la version précédente, il est nécessaire d’installer manuellement l’ensemble des dépendances logicielles sur le serveur.
    Si vous souhaitez utiliser cette archive pour une installation en mode ferme, il vous faudra également procéder à d’autres modifications (...)

  • MediaSPIP version 0.1 Beta

    16 avril 2011, par

    MediaSPIP 0.1 beta est la première version de MediaSPIP décrétée comme "utilisable".
    Le fichier zip ici présent contient uniquement les sources de MediaSPIP en version standalone.
    Pour avoir une installation fonctionnelle, il est nécessaire d’installer manuellement l’ensemble des dépendances logicielles sur le serveur.
    Si vous souhaitez utiliser cette archive pour une installation en mode ferme, il vous faudra également procéder à d’autres modifications (...)

  • Les autorisations surchargées par les plugins

    27 avril 2010, par

    Mediaspip core
    autoriser_auteur_modifier() afin que les visiteurs soient capables de modifier leurs informations sur la page d’auteurs

Sur d’autres sites (6074)

  • ffmpeg m3u8 segment file and relative path for file names

    15 mars 2013, par Luuk D. Jansen

    This must be a simple issue, but I have a difficulty with the M3U8 file created by FFMPEG as it contains the absolute path to the files, and not just the file names which I need. I am trying to be able to view the stream while it is encoded, and this seems the last hurdle.

    My FFMPEG command is (ignore the comma's) :

    ffmpeg, -i, /Users/User/Documents/Java/server/data/media/1/1/Promo-2013.mov, -dn, -y, -f, mpegts, -c:a, libfdk_aac, -strict, -2, -ar, 48000, -b:a, 128000, -s, 640x360, -c:v, libx264, -b:v, 700000, -aspect, 640:360, -r, 25, -level, 3.0, -vprofile, baseline, -flags, +loop, -cmp, chroma, -coder, 0, -me_range, 16, -keyint_min, 25, -sc_threshold, 40, -i_qfactor, 0.71, -bt, 200k, -maxrate, 700000, -bufsize, 700000, -rc_eq, 'blurCplx^(1-qComp)', -qcomp, 0.6, -qmin, 10, -qmax, 51, -qdiff, 4, -level, 30, -g, 30, -async, 2, -partitions, +parti4x4+partp8x8+partb8x8, -subq, 5, -trellis, 1, -refs, 1, -vbsf, h264_mp4toannexb, -flags, -global_header, -map, 0, -f, segment, -segment_time, 10, -segment_list, /Users/User/Documents/Java/server/data/media/1/3/prog_index.m3u8, -segment_list_flags, +live, -segment_format, mpegts, /Users/User/Documents/Java/server/data/media/1/3/segment_%05d.ts

    The resulting file is :

    #EXTM3U
    #EXT-X-VERSION:3
    #EXT-X-MEDIA-SEQUENCE:0
    #EXT-X-ALLOW-CACHE:YES
    #EXT-X-TARGETDURATION:12
    #EXTINF:10.005333,
    /Users/User/Documents/Java/server/data/media/1/10/segment_00000.ts
    #EXTINF:11.952000,
    /Users/User/Documents/Java/server/data/media/1/10/segment_00001.ts
    #EXTINF:10.002667,
    /Users/User/Documents/Java/server/data/media/1/10/segment_00002.ts
    #EXTINF:10.008000,
    /Users/User/Documents/Java/server/data/media/1/10/segment_00003.ts
    #EXTINF:0.266667,
    /Users/User/Documents/Java/server/data/media/1/10/segment_00004.ts
    #EXT-X-ENDLIST

    It all works perfect, but I only want e.g. 'segment_00000.ts, and not the complete absolute path, as that is of no use for live streaming. As this is a very ordinary application, I must do something wrong.

  • fate : Add a —target-samples path parameter

    20 mai 2013, par Martin Storsjö
    fate : Add a —target-samples path parameter
    

    This allows having the samples accessible via different paths
    on the target and on the host.

    Signed-off-by : Martin Storsjö <martin@martin.st>

    • [DBH] configure
    • [DBH] tests/fate.sh
    • [DBH] tests/fate/aac.mak
    • [DBH] tests/fate/ac3.mak
    • [DBH] tests/fate/adpcm.mak
    • [DBH] tests/fate/alac.mak
    • [DBH] tests/fate/als.mak
    • [DBH] tests/fate/amrnb.mak
    • [DBH] tests/fate/amrwb.mak
    • [DBH] tests/fate/atrac.mak
    • [DBH] tests/fate/audio.mak
    • [DBH] tests/fate/bmp.mak
    • [DBH] tests/fate/cdxl.mak
    • [DBH] tests/fate/cover-art.mak
    • [DBH] tests/fate/demux.mak
    • [DBH] tests/fate/dfa.mak
    • [DBH] tests/fate/dpcm.mak
    • [DBH] tests/fate/ea.mak
    • [DBH] tests/fate/filter-audio.mak
    • [DBH] tests/fate/filter-video.mak
    • [DBH] tests/fate/flac.mak
    • [DBH] tests/fate/h264.mak
    • [DBH] tests/fate/image.mak
    • [DBH] tests/fate/indeo.mak
    • [DBH] tests/fate/lossless-audio.mak
    • [DBH] tests/fate/lossless-video.mak
    • [DBH] tests/fate/microsoft.mak
    • [DBH] tests/fate/monkeysaudio.mak
    • [DBH] tests/fate/mp3.mak
    • [DBH] tests/fate/mpc.mak
    • [DBH] tests/fate/pcm.mak
    • [DBH] tests/fate/probe.mak
    • [DBH] tests/fate/prores.mak
    • [DBH] tests/fate/qt.mak
    • [DBH] tests/fate/qtrle.mak
    • [DBH] tests/fate/real.mak
    • [DBH] tests/fate/screen.mak
    • [DBH] tests/fate/utvideo.mak
    • [DBH] tests/fate/video.mak
    • [DBH] tests/fate/voice.mak
    • [DBH] tests/fate/vorbis.mak
    • [DBH] tests/fate/vpx.mak
    • [DBH] tests/fate/vqf.mak
    • [DBH] tests/fate/wavpack.mak
    • [DBH] tests/fate/wma.mak
  • How to using every 5 sec generate video output File Path to Encode with RTMP Formate write data in ios ? [on hold]

    16 juillet 2015, par Sandeep Joshi
    (void) segmentRecording:(NSTimer*)timer {
    if (!shouldBeRecording) {
       [timer invalidate];
    }
    AVAssetWriter *tempAssetWriter = self.assetWriter;
    AVAssetWriterInput *tempAudioEncoder = self.audioEncoder;
    AVAssetWriterInput *tempVideoEncoder = self.videoEncoder;
    self.assetWriter = queuedAssetWriter;
    self.audioEncoder = queuedAudioEncoder;
    self.videoEncoder = queuedVideoEncoder;
    NSLog(@"Switching encoders");

    dispatch_async(segmentingQueue, ^{
       if (tempAssetWriter.status == AVAssetWriterStatusWriting) {
           @try {
               [tempAudioEncoder markAsFinished];
               [tempVideoEncoder markAsFinished];
               [tempAssetWriter finishWritingWithCompletionHandler:^{
                   if (tempAssetWriter.status == AVAssetWriterStatusFailed) {
                       [self showError:tempAssetWriter.error];
                   } else {
                       [self uploadLocalURL:tempAssetWriter.outputURL];
                   }
               }];
           }
           @catch (NSException *exception) {
               NSLog(@"Caught exception: %@", [exception description]);
               //[BugSenseController logException:exception withExtraData:nil];
           }
       }
       self.segmentCount++;
       if (self.readyToRecordAudio &amp;&amp; self.readyToRecordVideo) {
           NSError *error = nil;
           self.queuedAssetWriter = [[AVAssetWriter alloc] initWithURL:[OWUtilities urlForRecordingSegmentCount:segmentCount basePath:self.basePath] fileType:(NSString *)kUTTypeMPEG4 error:&amp;error];
           if (error) {
               [self showError:error];
           }
           self.queuedVideoEncoder = [self setupVideoEncoderWithAssetWriter:self.queuedAssetWriter formatDescription:videoFormatDescription bitsPerSecond:videoBPS];
           self.queuedAudioEncoder = [self setupAudioEncoderWithAssetWriter:self.queuedAssetWriter formatDescription:audioFormatDescription bitsPerSecond:audioBPS];
           //NSLog(@"Encoder switch finished");
       }
    });}



    (void) uploadLocalURL:(NSURL*)url {
    NSLog(@"upload local url: %@", url);
    NSString *inputPath = [url path];
    NSString *outputPath = [inputPath stringByReplacingOccurrencesOfString:@".mp4" withString:@".ts"];
    NSString *outputFileName = [outputPath lastPathComponent];
    NSDictionary *options = @{kFFmpegOutputFormatKey: @"mpegts"};
    NSLog(@"%@ conversion...", outputFileName);
    [ffmpegWrapper convertInputPath:[url path] outputPath:outputPath options:options progressBlock:nil completionBlock:^(BOOL success, NSError *error) {
       if (success) {
           if (!isRtmpConnected) {
               isRtmpConnected = [rtmp openWithURL:HOST_URL enableWrite:YES];
           }
           isRtmpConnected = [rtmp isConnected];

           if (isRtmpConnected) {

               NSData *video = [NSData dataWithContentsOfURL:[NSURL URLWithString:outputPath]];
               NSUInteger length = [video length];
               NSUInteger chunkSize = 1024 * 5;;
               NSUInteger offset = 0;
               NSLog(@"original video length: %lu \n chunkSize : %lu", length,chunkSize);
             // Let's split video to small chunks to publish to media server
               do {
                   NSUInteger thisChunkSize = length - offset > chunkSize ? chunkSize : length - offset;
                   NSData* chunk = [NSData dataWithBytesNoCopy:(char *)[video bytes] + offset
                                                        length:thisChunkSize
                                                  freeWhenDone:NO];
                   offset += thisChunkSize;

                   // Write new chunk to rtmp server
                   NSLog(@"%lu", (unsigned long)[rtmp write:chunk]);
                   sleep(1);
               } while (offset &lt; length);
           }else{
               [rtmp close];
           }


       } else {
           NSLog(@"conversion error: %@", error.userInfo);
       }
    }];}

    This code use for live streaming for send data using RTMP Wrapper.
    Not write in Socket properly because every 5 second to generate different file output file.

    This is proper way ?

    I have no idea how to get NSData in proper way.

    Please help me .