Recherche avancée

Médias (1)

Mot : - Tags -/artwork

Autres articles (95)

  • Les sons

    15 mai 2013, par
  • Automated installation script of MediaSPIP

    25 avril 2011, par

    To overcome the difficulties mainly due to the installation of server side software dependencies, an "all-in-one" installation script written in bash was created to facilitate this step on a server with a compatible Linux distribution.
    You must have access to your server via SSH and a root account to use it, which will install the dependencies. Contact your provider if you do not have that.
    The documentation of the use of this installation script is available here.
    The code of this (...)

  • MediaSPIP en mode privé (Intranet)

    17 septembre 2013, par

    À partir de la version 0.3, un canal de MediaSPIP peut devenir privé, bloqué à toute personne non identifiée grâce au plugin "Intranet/extranet".
    Le plugin Intranet/extranet, lorsqu’il est activé, permet de bloquer l’accès au canal à tout visiteur non identifié, l’empêchant d’accéder au contenu en le redirigeant systématiquement vers le formulaire d’identification.
    Ce système peut être particulièrement utile pour certaines utilisations comme : Atelier de travail avec des enfants dont le contenu ne doit pas (...)

Sur d’autres sites (11784)

  • How do I run the qt ffmpeg wrapper example ? (error : cannot find -lavutil)

    29 août 2013, par James

    I am using QT 4.8 on the raspberry pi trying to save a binary stream to a video file like mp4. I was able to create and save static images by converting it into a QImage, but QT does not seem to support actually saving that stream (or the series of QImages) to a video file.

    The only option I may have found is qt ffmpeg Wrapper
    https://code.google.com/p/qtffmpegwrapper/

    but the example won't run out of the box. I am having the same exact problem as this guy :

    https://code.google.com/p/qtffmpegwrapper/issues/detail?id=14&colspec=ID%20Type%20Status%20Priority%20Milestone%20Owner%20Summary%20Opened%20Closed%20Modified#makechanges

    All I did was download it and try to run the included example, but no go.
    Am I supposed to download ffmpeg files prior to using this ? apt-get install ffmpeg ?

    Thanks !

  • Concat video/audio files downloaded from mpd dash manifest

    15 mai 2022, par Bloworlf Mathurin

    I have this manifest I fetched from an url :

    


    &lt;?xml version="1.0"?>&#xA;<mpd type="dynamic" xmlns="urn:mpeg:dash:schema:mpd:2011" profiles="urn:mpeg:dash:profile:isoff-live:2011" availabilitystarttime="2022-05-14T21:18:50-07:00" availabilityendtime="2022-05-14T22:35:36-07:00" timeshiftbufferdepth="PT20S" suggestedpresentationdelay="PT2S" minbuffertime="PT1S" publishtime="2022-05-14T21:18:50-07:00" minimumupdateperiod="PT1S" validationerrors="" currentservertimems="0" firstavtimems="1652588331877" lastvideoframets="0" loapstreamid="17944405669984029" publishframetime="939">&#xA;    <period start="PT0S">&#xA;        <adaptationset segmentalignment="true" maxwidth="432" maxheight="766" maxframerate="30">&#xA;            <representation mimetype="video/mp4" codecs="avc1.4d401e" width="432" height="766" framerate="30" startwithsap="1" bandwidth="38894" maxbandwidth="46772" playbackresolutionmos="432:82.84,720:65.43,216:70.39" qualityclass="sd" qualitylabel="432p">&#xA;                <segmenttemplate presentationtimeoffset="0" timescale="1000" initialization="some/url/some_id-init.m4v?ms=m_C&amp;amp;ccb=2-4" media="some/url/some_id-$Time$.m4v?ms=m_C&amp;amp;ccb=2-4">&#xA;                    <segmenttimeline>&#xA;                        <s t="4092613" d="2000"></s>&#xA;                        <s t="4094613" d="2000"></s>&#xA;                        <s t="4096613" d="2000"></s>&#xA;                        <s t="4098613" d="2000"></s>&#xA;                        <s t="4100613" d="2000"></s>&#xA;                        <s t="4102613" d="2000"></s>&#xA;                        <s t="4104613" d="2000"></s>&#xA;                        <s t="4106613" d="2000"></s>&#xA;                        <s t="4108613" d="2000"></s>&#xA;                        <s t="4110613" d="2000"></s>&#xA;                    </segmenttimeline>&#xA;                </segmenttemplate>&#xA;            </representation>&#xA;        </adaptationset>&#xA;        <adaptationset segmentalignment="true">&#xA;            <representation mimetype="audio/mp4" codecs="mp4a.40.2" audiosamplingrate="44100" startwithsap="1" bandwidth="48979" maxbandwidth="57732">&#xA;                <segmenttemplate presentationtimeoffset="0" timescale="1000" initialization="some/url/some_id-init.m4a?ms=m_C&amp;amp;ccb=2-4" media="some/url/some_id-$Time$.m4a?ms=m_C&amp;amp;ccb=2-4">&#xA;                    <segmenttimeline>&#xA;                        <s t="4092613" d="2000"></s>&#xA;                        <s t="4094613" d="2000"></s>&#xA;                        <s t="4096613" d="2000"></s>&#xA;                        <s t="4098613" d="2000"></s>&#xA;                        <s t="4100613" d="2000"></s>&#xA;                        <s t="4102613" d="2000"></s>&#xA;                        <s t="4104613" d="2000"></s>&#xA;                        <s t="4106613" d="2000"></s>&#xA;                        <s t="4108613" d="2000"></s>&#xA;                        <s t="4110613" d="2000"></s>&#xA;                    </segmenttimeline>&#xA;                </segmenttemplate>&#xA;            </representation>&#xA;        </adaptationset>&#xA;    </period>&#xA;</mpd>&#xA;

    &#xA;

    I manage to download all the files (init file + 10 segment files) and put them in an array

    &#xA;

    File[] files = downloadSegments();

    &#xA;

    So I have :&#xA;[0] -> file-0.m4v (which is the init file)&#xA;[1] -> file-1.m4v (1st segment)&#xA;... and so on.

    &#xA;

    My question is how can I concat/append all these files into another file (final_segment.m4v) ?

    &#xA;

    I looked around a lot and am now using 'com.arthenica:ffmpeg-kit-video:4.5.1-1'&#xA;I've last tried :

    &#xA;

    FFmpegSession session = FFmpegKit.execute("-i file-0.m4v -i file-1.m4v ... -c copy final_segment.m4v");&#xA;

    &#xA;

    Is there something I'm missing ?

    &#xA;

    Also I assume that I'll have to do the same for the audio segments. So I will have 2 files (final_segment.m4v and final_segment.m4a) that I will have to merge/mix together.

    &#xA;

    If you could help me with some piece of code, that would be great.

    &#xA;

  • How to correctly use h264_amf on Windows ?

    27 octobre 2022, par Harshiv

    I'm trying to use hardware encoding with FFMPEG on Windows 10 with AMD Radeon r5 M330 DGPU on laptop : HP-AC026-TX. FFMPEG version is 5.1.2-full_build-www.gyan.dev from gyan.dev. I'm using this :

    &#xA;

    fmpeg ^&#xA;    -hide_banner ^&#xA;    -v verbose ^&#xA;    -i "%1" ^&#xA;    -c:v h264_amf -acodec copy -y out.mp4&#xA;

    &#xA;

    but it fails with :

    &#xA;

    [h264_amf @ 000001fb84d80b80] AMF initialisation succeeded via D3D11.&#xA;[h264_amf @ 000001fb847ab3c0] CreateComponent(AMFVideoEncoderVCE_AVC) failed with error 36&#xA;Error initializing output stream 0:0 -- Error while opening encoder for output stream #0:0 - maybe incorrect parameters such as bit_rate, rate, width or height&#xA;

    &#xA;

    The arguments seem as generic as I could hope, FFMPEG should use sensible defaults, so I can't Understand why it is failing.&#xA;Entire FFMPEG output is :

    &#xA;

    [h264 @ 000001fb84706c40] Reinit context to 1920x1088, pix_fmt: yuv420p&#xA;Input #0, mov,mp4,m4a,3gp,3g2,mj2, from &#x27;1366_pngl_20220816_161558.mp4&#x27;:&#xA;  Metadata:&#xA;    major_brand     : mp42&#xA;    minor_version   : 0&#xA;    compatible_brands: isommp42&#xA;    creation_time   : 2022-08-16T10:51:42.000000Z&#xA;    com.android.version: 11&#xA;    com.android.capture.fps: 30.000000&#xA;  Duration: 00:03:41.33, start: 0.000000, bitrate: 7517 kb/s&#xA;  Stream #0:0[0x1](eng): Video: h264 (High), 1 reference frame (avc1 / 0x31637661), yuv420p(tv, bt709, progressive, left), 1920x1080 (1920x1088), 7259 kb/s, SAR 1:1 DAR 16:9, 29.96 fps, 29.99 tbr, 90k tbn (default)&#xA;    Metadata:&#xA;      creation_time   : 2022-08-16T10:51:42.000000Z&#xA;      handler_name    : VideoHandle&#xA;      vendor_id       : [0][0][0][0]&#xA;  Stream #0:1[0x2](eng): Audio: aac (LC) (mp4a / 0x6134706D), 48000 Hz, stereo, fltp, 256 kb/s (default)&#xA;    Metadata:&#xA;      creation_time   : 2022-08-16T10:51:42.000000Z&#xA;      handler_name    : SoundHandle&#xA;      vendor_id       : [0][0][0][0]&#xA;Stream mapping:&#xA;  Stream #0:0 -> #0:0 (h264 (native) -> h264 (h264_amf))&#xA;  Stream #0:1 -> #0:1 (copy)&#xA;Press [q] to stop, [?] for help&#xA;[h264 @ 000001fb84787400] Reinit context to 1920x1088, pix_fmt: yuv420p&#xA;[graph 0 input from stream 0:0 @ 000001fb84cfd780] w:1920 h:1080 pixfmt:yuv420p tb:1/90000 fr:90000/3001 sar:1/1&#xA;[h264_amf @ 000001fb84d80b80] AMF initialisation succeeded via D3D11.&#xA;[h264_amf @ 000001fb847ab3c0] CreateComponent(AMFVideoEncoderVCE_AVC) failed with error 36&#xA;Error initializing output stream 0:0 -- Error while opening encoder for output stream #0:0 - maybe incorrect parameters such as bit_rate, rate, width or height&#xA;[AVIOContext @ 000001fb846f89c0] Statistics: 0 bytes written, 0 seeks, 0 writeouts&#xA;[AVIOContext @ 000001fb846f3f40] Statistics: 371902 bytes read, 3 seeks&#xA;Conversion failed!&#xA;

    &#xA;

    What should I do to make hardware encoding work ?

    &#xA;