
Recherche avancée
Autres articles (92)
-
Websites made with MediaSPIP
2 mai 2011, parThis page lists some websites based on MediaSPIP.
-
List of compatible distributions
26 avril 2011, parThe table below is the list of Linux distributions compatible with the automated installation script of MediaSPIP. Distribution nameVersion nameVersion number Debian Squeeze 6.x.x Debian Weezy 7.x.x Debian Jessie 8.x.x Ubuntu The Precise Pangolin 12.04 LTS Ubuntu The Trusty Tahr 14.04
If you want to help us improve this list, you can provide us access to a machine whose distribution is not mentioned above or send the necessary fixes to add (...) -
Creating farms of unique websites
13 avril 2011, parMediaSPIP platforms can be installed as a farm, with a single "core" hosted on a dedicated server and used by multiple websites.
This allows (among other things) : implementation costs to be shared between several different projects / individuals rapid deployment of multiple unique sites creation of groups of like-minded sites, making it possible to browse media in a more controlled and selective environment than the major "open" (...)
Sur d’autres sites (10841)
-
Write Live Photo metadata to video using FFMPEG
13 août 2021, par Luke BurnsAccording to Apple Live Photo file format, there are three pieces of metadata that need to be written for a JPEG and MOV to be accepted as a live photo. I can use exiftool and ffmpeg to write the necessary content identifier metadata.


For the JPEG :


exiftool -TagsFromFile reference.jpeg -makernotes -ContentIdentifier image.jpeg
exiftool -ContentIdentifier="$id" image.jpeg



Similarly, ffmpeg can be used to write the top-level Quicktime metadata with matching id.


However I'm having trouble with the timed metadata :
["com.apple.quicktime.still-image-time" : 0xFF]
.

I can't even manage to produce a copy of an existing live photo MOV file using ffmpeg that preserves the necessary timed metadata.


ffmpeg -i original.mov -map 0 -c copy -movflags use_metadata_tags copy.mov



copies the global metadata (i.e. com.apple.quicktime.content.identifier), but loses the necessary still-image-time which can be confirmed using exiftool :


> exiftool -G -U -ee original.mov | grep 'Still Image Time'
[QuickTime] Still Image Time : -1
> exiftool -G -U -ee copy.mov | grep 'Still Image Time'
> 



How can I write the timed metadata using ffmpeg—specifically the still-image-time data ?


Edit : it looks like this may be happening because ffmpeg does not know how to handle the
mebx
tag on data streams :

[mov @ 0x7fb232091400] Unknown hldr_type for mebx, writing dummy values



And ffmpeg doesn't seem to have a way to copy unknown streams. This appears to also be a problem for dealing with the
fdsc
tag in GoPro metadata streams (e.g. see https://gist.github.com/radimklaska/8974637522a751adb49db0de3be482c9#file-gopro_hevc_to_dnxhd-sh-L125), so it's often copied over asgpmd
data, which ensures it's not overwritten with dummy values, but this trick doesn't work in the case of live photosmebx
metadata.

-
linking libavcodec os x ffmpeg ApplicationServices.framework
30 décembre 2015, par Grady PlayerHere is my error : (newer version of ffmpeg)
dyld: Library not loaded: /System/Library/Frameworks/ApplicationServices.framework/Versions/A/ApplicationServices
Referenced from: /Users/me/path/to/lib/libavcodec.dylib
Reason: Incompatible library version: libavcodec.dylib requires version 64.0.0 or later, but ApplicationServices provides version 1.0.0I have read the questions on SO that say you need to link ApplicationServices.framework... which I have done, it makes no difference...
I have searched for other versions of that framework :
$ mdfind ApplicationServices.framework
/System/Library/Frameworks/ApplicationServices.frameworkjust one...
so then inspecting frameworks...
$ otool -L /System/Library/Frameworks/ApplicationServices.framework/Versions/A/ApplicationServices
/System/Library/Frameworks/ApplicationServices.framework/Versions/A/ApplicationServices:
/System/Library/Frameworks/ApplicationServices.framework/Versions/A/ApplicationServices (compatibility version 1.0.0, current version 48.0.0)
/System/Library/Frameworks/CoreGraphics.framework/Versions/A/CoreGraphics (compatibility version 64.0.0, current version 600.0.0)
...and libavcodec
otool -L ../lib/libavcodec.dylib
../lib/libavcodec.dylib:
libavcodec.dylib (compatibility version 56.0.0, current version 56.60.100)
@loader_path/libswresample.dylib (compatibility version 1.0.0, current version ...
...
/System/Library/Frameworks/ApplicationServices.framework/Versions/A/ApplicationServices (compatibility version 64.0.0, current version 600.0.0)I am compiling everything with
-mmacosx-version-min=10.7
so maybe a bug in the toolchain ?
compiler :
clang -v
Apple LLVM version 7.0.2 (clang-700.1.81)
Target: x86_64-apple-darwin15.2.0
Thread model: posixlinker :
ld -v
@(#)PROGRAM:ld PROJECT:ld64-253.9
configured to support archs: armv6 armv7 armv7s arm64 i386 x86_64 x86_64h armv6m armv7k armv7m armv7em (tvOS)
LTO support using: Apple LLVM 7.0.2 (clang-700.1.81) -
How to simplify the script for ffmpeg ?
27 mars 2018, par Pravesh KumarI want this script to be simplified
I am having this given simple script ; I would like to simplify this script can anyone help me to add text with the proper time interval and fade in and out effect.
ffmpeg -y -i video.mp4 -filter_complex \
"[0]split[base][text];[text] \
drawtext=fontfile=gvr.otf:text='Which of these is not an event listener adapter defined in the java.awt.event package?': fontcolor=white: fontsize=40: x=100:y=200, \
format=yuva444p,fade=t=in:st=1:d=1:alpha=1,fade=t=out:st=8:d=1:alpha=1[subtitles]; \
[base][subtitles]overlay" test_out.mp4
ffmpeg -y -i test_out.mp4 -filter_complex \
"[0]split[base][text];[text] \
drawtext=fontfile=gvr.otf:text='a) public void apple(String s, int i) {}': fontcolor=white: fontsize=40: x=100:y=(200) + 50 * 2, \
format=yuva444p,fade=t=in:st=3:d=1:alpha=1,fade=t=out:st=8:d=1:alpha=1[subtitles]; \
[base][subtitles]overlay" test_out1.mp4
ffmpeg -y -i test_out1.mp4 -filter_complex \
"[0]split[base][text];[text] \
drawtext=fontfile=gvr.otf:text='b) public int apple(int i, String s) {}': fontcolor=white: fontsize=40: x=100:y=(200) + 50 * 3, \
format=yuva444p,fade=t=in:st=4:d=1:alpha=1,fade=t=out:st=8:d=1:alpha=1[subtitles]; \
[base][subtitles]overlay" test_out2.mp4
ffmpeg -y -i test_out2.mp4 -filter_complex \
"[0]split[base][text];[text] \
drawtext=fontfile=gvr.otf:text='c) public void apple(int i, String mystring) {}': fontcolor=white: fontsize=40: x=100:y=(200) + 50 * 4, \
format=yuva444p,fade=t=in:st=5:d=1:alpha=1,fade=t=out:st=8:d=1:alpha=1[subtitles]; \
[base][subtitles]overlay" test_out3.mp4
ffmpeg -y -i test_out3.mp4 -filter_complex \
"[0]split[base][text];[text] \
drawtext=fontfile=gvr.otf:text='d) public void Apple(int i, String s) {}': fontcolor=white: fontsize=40: x=100:y=(200) + 50 * 5, \
format=yuva444p,fade=t=in:st=6:d=1:alpha=1,fade=t=out:st=8:d=1:alpha=1[subtitles]; \
[base][subtitles]overlay" test_out4.mp4