Recherche avancée

Médias (91)

Autres articles (11)

  • Supporting all media types

    13 avril 2011, par

    Unlike most software and media-sharing platforms, MediaSPIP aims to manage as many different media types as possible. The following are just a few examples from an ever-expanding list of supported formats : images : png, gif, jpg, bmp and more audio : MP3, Ogg, Wav and more video : AVI, MP4, OGV, mpg, mov, wmv and more text, code and other data : OpenOffice, Microsoft Office (Word, PowerPoint, Excel), web (html, CSS), LaTeX, Google Earth and (...)

  • Le plugin : Podcasts.

    14 juillet 2010, par

    Le problème du podcasting est à nouveau un problème révélateur de la normalisation des transports de données sur Internet.
    Deux formats intéressants existent : Celui développé par Apple, très axé sur l’utilisation d’iTunes dont la SPEC est ici ; Le format "Media RSS Module" qui est plus "libre" notamment soutenu par Yahoo et le logiciel Miro ;
    Types de fichiers supportés dans les flux
    Le format d’Apple n’autorise que les formats suivants dans ses flux : .mp3 audio/mpeg .m4a audio/x-m4a .mp4 (...)

  • Encoding and processing into web-friendly formats

    13 avril 2011, par

    MediaSPIP automatically converts uploaded files to internet-compatible formats.
    Video files are encoded in MP4, Ogv and WebM (supported by HTML5) and MP4 (supported by Flash).
    Audio files are encoded in MP3 and Ogg (supported by HTML5) and MP3 (supported by Flash).
    Where possible, text is analyzed in order to retrieve the data needed for search engine detection, and then exported as a series of image files.
    All uploaded files are stored online in their original format, so you can (...)

Sur d’autres sites (3057)

  • FFMPEG and FFPlay can access rtsp stream from one ip, but from other ip, it can't

    28 mai 2022, par Crear

    The situation is kind of complex. I was archiving several CCTV camera feeds (rtsp, h264, no audio) through OpenCV, which worked but the CPU utilization was too high and started to lose some frames time by time.

    


    To reduce the CPU utilization, I started to use FFMPEG to skip the decoding and encoding processes, which worked perfectly on my home machine. However, when I connected to my university VPN and tried to deploy it on our lab server, FFmpeg couldn't read any frame, ffplay couldn't get anything either. However, OpenCV, VLC Player and IINA Player could still read and display the feed.

    


    In Summary,

    


    1 FFMPEG/ffplay

    


    1.1 can only read the feed from my home network(Wi-Fi, optimum)

    


    1.2 from other two networks, the error message says : "Could not find codec parameters for stream 0 (Video : h264, none) : unspecified size
Consider increasing the value for the 'analyzeduration' (0) and 'probesize' (5000000) options"

    


    2 IINA/VLC Player, OpenCV
These tools can get the video all the time.

    


    I'm wondering whether it's related to some specific port access, that the ffmpeg required but the others don't. I'd appreciate it if anyone can provide any suggestions.

    


    As references, the tested ffplay command is simple :

    


    ffplay 'the rtsp address'


    


    Thanks

    



    


    Update

    


    More tests have been performed.

    


    By specifying rtsp_transport as TCP, ffplay can play the video, but FFmpeg can't access the video. (In the beginning, when both FFmpeg and ffplay worked through my home network, it was UDP)
The FFmpeg command is as follows :

    


    ffmpeg -i rtsp://the_ip_address/axis-media/media.amp -hide_banner -c:v copy -s 1920x1080 -segment_time 00:30:00 -f segment -strftime 1 -reset_timestamps 1 -rtsp_transport tcp "%Y-%m-%d-%H-%M-%S_Test.mp4"


    


    Please help...

    


  • Pims.open Throws "UnkownFormat Error" with "Invalid Argument" On One Machine But Not on Another

    12 août 2021, par KaceBellamy

    I'm trying to do some video processing for a physics experiment, and I want to do it on my much more powerful Windows desktop computer as opposed to my Mac laptop.

    


    The following code works like a dream when run as a Jupyter notebook on my Mac :

    


    import matplotlib as mpl
from mpl_toolkits import mplot3d
import pims
import trackpy as tp 

@pims.pipeline
def gray(image):
    return image[:, :, 1]  # Take just the green channel

frames = gray(pims.open('output.mp4'))


    


    but on my Windows machine I get this error :

    


    ---------------------------------------------------------------------------&#xA;UnknownFormatError                        Traceback (most recent call last)&#xA;~\AppData\Local\Temp/ipykernel_12152/704954007.py in <module>&#xA;      1 #Actually converts the video. Might be VERY processor intensive... or not?&#xA;----> 2 frames = gray(pims.open(&#x27;output.mp4&#x27;)) #Make the File Name whatever file you like!&#xA;&#xA;~\miniconda3\lib\site-packages\pims\api.py in open(sequence, **kwargs)&#xA;    207             warn(message)&#xA;    208             exceptions &#x2B;= message &#x2B; &#x27;\n&#x27;&#xA;--> 209     raise UnknownFormatError("All handlers returned exceptions:\n" &#x2B; exceptions)&#xA;    210 &#xA;    211 &#xA;&#xA;UnknownFormatError: All handlers returned exceptions:&#xA;<class> errored: [Errno 22] Invalid argument: &#x27;output.mp4&#x27;&#xA;<class> errored: [Errno 22] Invalid argument: &#x27;output.mp4&#x27;&#xA;<class> errored: Could not load meta information&#xA;=== stderr ===&#xA;&#xA;ffmpeg version 4.2.2 Copyright (c) 2000-2019 the FFmpeg developers&#xA;  built with gcc 9.2.1 (GCC) 20200122&#xA;  configuration: --enable-gpl --enable-version3 --enable-sdl2 --enable-fontconfig --enable-gnutls --enable-iconv --enable-libass --enable-libdav1d --enable-libbluray --enable-libfreetype --enable-libmp3lame --enable-libopencore-amrnb --enable-libopencore-amrwb --enable-libopenjpeg --enable-libopus --enable-libshine --enable-libsnappy --enable-libsoxr --enable-libtheora --enable-libtwolame --enable-libvpx --enable-libwavpack --enable-libwebp --enable-libx264 --enable-libx265 --enable-libxml2 --enable-libzimg --enable-lzma --enable-zlib --enable-gmp --enable-libvidstab --enable-libvorbis --enable-libvo-amrwbenc --enable-libmysofa --enable-libspeex --enable-libxvid --enable-libaom --enable-libmfx --enable-amf --enable-ffnvcodec --enable-cuvid --enable-d3d11va --enable-nvenc --enable-nvdec --enable-dxva2 --enable-avisynth --enable-libopenmpt&#xA;  libavutil      56. 31.100 / 56. 31.100&#xA;  libavcodec     58. 54.100 / 58. 54.100&#xA;  libavformat    58. 29.100 / 58. 29.100&#xA;  libavdevice    58.  8.100 / 58.  8.100&#xA;  libavfilter     7. 57.100 /  7. 57.100&#xA;  libswscale      5.  5.100 /  5.  5.100&#xA;  libswresample   3.  5.100 /  3.  5.100&#xA;  libpostproc    55.  5.100 / 55.  5.100&#xA;C:\Users\Callum\OneDrive - The University of Chicago\output.mp4: Invalid argument&#xA;<class> errored: MoviePy error: failed to read the duration of file output.mp4.&#xA;Here are the file infos returned by ffmpeg:&#xA;&#xA;ffmpeg version 4.2.2 Copyright (c) 2000-2019 the FFmpeg developers&#xA;  built with gcc 9.2.1 (GCC) 20200122&#xA;  configuration: --enable-gpl --enable-version3 --enable-sdl2 --enable-fontconfig --enable-gnutls --enable-iconv --enable-libass --enable-libdav1d --enable-libbluray --enable-libfreetype --enable-libmp3lame --enable-libopencore-amrnb --enable-libopencore-amrwb --enable-libopenjpeg --enable-libopus --enable-libshine --enable-libsnappy --enable-libsoxr --enable-libtheora --enable-libtwolame --enable-libvpx --enable-libwavpack --enable-libwebp --enable-libx264 --enable-libx265 --enable-libxml2 --enable-libzimg --enable-lzma --enable-zlib --enable-gmp --enable-libvidstab --enable-libvorbis --enable-libvo-amrwbenc --enable-libmysofa --enable-libspeex --enable-libxvid --enable-libaom --enable-libmfx --enable-amf --enable-ffnvcodec --enable-cuvid --enable-d3d11va --enable-nvenc --enable-nvdec --enable-dxva2 --enable-avisynth --enable-libopenmpt&#xA;  libavutil      56. 31.100 / 56. 31.100&#xA;  libavcodec     58. 54.100 / 58. 54.100&#xA;  libavformat    58. 29.100 / 58. 29.100&#xA;  libavdevice    58.  8.100 / 58.  8.100&#xA;  libavfilter     7. 57.100 /  7. 57.100&#xA;  libswscale      5.  5.100 /  5.  5.100&#xA;  libswresample   3.  5.100 /  3.  5.100&#xA;  libpostproc    55.  5.100 / 55.  5.100&#xA;output.mp4: Invalid argument&#xA;</class></class></class></class></module>

    &#xA;

    output.mp4 is just a run of the mill video file taken on a CCD camera ; I've tried converting it to a .MOV, I've tried other video files taken on different cameras, and I've tried running the file through FFmpeg to impose a 30 fps framerate ; everything I've tried works fine on my Mac and throws the above error on my Windows machine.

    &#xA;

    For reference, I installed necessary packages for this code on both machines this morning, so it should all be up to date and the same on both.

    &#xA;

    Any ideas as to what's up ? Thanks !

    &#xA;

  • How to batch process a series of video files with powershell and other-transcode/ffmpeg

    7 juin 2022, par DarkDiamond

    TL ;DR

    &#xA;

    What did I do wrong in the following PowerShell-Script ? It does not work as expected.

    &#xA;


    &#xA;

    I am recording some of my lectures in my university with a photo camera. This works pretty well although I have to split the single lecture into three to four parts because the camera can only record 29 minutes of video in one take. I know that this is a common issue related to some licensensing problem that most photo cameras simply don't have the right license to record longer videos. But it confronts me with the problem that I later have to edit the files together after I did some post processing on them.

    &#xA;

    With the camera I produce up to four video files with sizes around 3.5 GB which is way to big in order to be of any use because our IT department understandably doesn't want to host so much data, as I produce around 22 GB of video material each week.

    &#xA;

    Some time ago I came across a very useful tool called "other-video-transcoding" by Don Melton over on GitHub, written in ruby, that allows me to compress the files to a reasonable file size without any visual loss. In addition I crop the videos to remove the part of each frame that is neither the board nor a place where my professor stands in order to decrease the filesize even further and do some privacy protection by cutting out most of the students.

    &#xA;

    As the tools are accessable via the command line, it is relatively easy to configure and does not cost additional computational power to render a nice gui, so I can edit one of the 29 minute clips in less than 10 minutes.

    &#xA;

    Now I wanted to optimize my workflow by writing a PowerShell script that only takes the parameters what to crop and which files to work on and then does the rest on its own so I can just start the script and then do something else while my laptop renders the new files.

    &#xA;

    So far I have the following :

    &#xA;

    $video_path = Get-ChildItem ..\ -Directory | findstr "SoSe"&#xA;&#xA;Get-ChildItem $video_path -name | findstr ".MP4" | Out-File temp.txt -Append &#xA;Get-Content temp.txt | ForEach-Object {"file " &#x2B; $_} >> .\files.txt&#xA;&#xA;Get-ChildItem $video_path |&#xA;Foreach-Object {&#xA;other-transcode --hevc --mp4 --target 3000 --crop 1920:780:0:0 $_.FullName&#xA;}&#xA;&#xA;#other-transcode --hevc --mp4 --crop 1920:720:60:0 ..\SoSe22_Theo1_videos_v14_RAW\&#xA;ffmpeg -f concat -i files.txt -c copy merged.mp4&#xA;Remove-Item .\temp.txt&#xA;

    &#xA;

    but it does not quite do what I it expect to do.&#xA;This is my file system :

    &#xA;

    sciebo/&#xA;└── SoSe22_Theo1_videos/&#xA;    ├── SoSe22_Theo1_videos_v16/&#xA;    │   ├── SoSe22_Theo1_videos_v16_KOMPR/&#xA;    │   │   ├── C0001.mp4&#xA;    │   │   ├── C0002.mp4&#xA;    │   │   ├── C0003.mp4&#xA;    │   │   ├── C0004.mp4&#xA;    │   │   ├── temp.txt&#xA;    │   │   ├── files.txt&#xA;    │   │   └── merged.mp4&#xA;    │   └── SoSe22_Theo1_videos_v16_RAW/&#xA;    │       ├── C0001.mp4&#xA;    │       ├── C0002.mp4&#xA;    │       ├── C0003.mp4&#xA;    │       └── C0004.mp4&#xA;    └── SoSe22_Theo1_videos_v17/&#xA;        ├── SoSe22_Theo1_videos_v17_KOMPR&#xA;        └── SoSe22_Theo1_videos_v17_RAW/&#xA;            ├── C0006.mp4&#xA;            ├── C0007.mp4&#xA;            ├── C0008.mp4&#xA;            └── C0009.mp4&#xA;

    &#xA;

    where the 16th lecture is already processed and the 17th is not. I always have the raw video data in the folders ending on RAW and the edited/compressed output files in the one ending on KOMPR. Note that the video files in the KOMPR folder are the output files of the other-transcode tool.

    &#xA;

    The real work happens in the line where it says

    &#xA;

    other-transcode --hevc --mp4 --target 3000 --crop 1920:780:0:0 $_.FullName&#xA;

    &#xA;

    and in the line

    &#xA;

    ffmpeg -f concat -i files.txt -c copy merged.mp4&#xA;

    &#xA;

    where I concat the output files into the final version I can upload to our online learning platform.&#xA;What is wrong with my script ? In the end I'd like to pass the --crop parameter just to my script, but that is not the primary problem.

    &#xA;


    &#xA;

    A little information on the transcoding script so you don't have to look into the documentation :
    &#xA;As the last argument the tool takes the location of the video files to work on, be it relative or absolute file paths. The output is placed in the folder the script is called in, so if I cd into one of the KOMPR directories and then call

    &#xA;

    other-transcode --mp4 ../SoSe22_Theo1_videos_v16_RAW/C0001.mp4&#xA;

    &#xA;

    a new file C0001.mp4 is created in the KOMPR directory and the transcoded video and old audio are written to that new video file.

    &#xA;