23:17
How do I losslessly split videos by file size? For example, I have a 1.92GB video and I want to split it into 200MB videos.
I tried using ffmpeg and putting in -fs, -c copy, and 200000000 for 200MB, but I only got one file that's not even the correct file size. How do I properly use it?
15:24
I am building a YouTube video trimmer, and currently, it takes the YouTube video ID, downloads the whole video from the server (Even it is like 10 hours long), and then it trims it according to user's inputted timeframe using ffmpeg.
Now the problem with this is that it takes so much time to download the whole video even if we want a small piece of it hence it is highly impractical.
I was thinking to implement something like an HTML5 video player does when you seek the video forward. It just jumps to the part where you seek-ed to without downloading the part you skipped (...)
07:41
everyone,It is sorry to bother you,but i need some help.
I'm working on an embedded deployment project,doing object detection work to real-time video stream. So I have to port my c++ inference prog to RKNN1808 platform. I compile this program with CMake tool,but I cant finish my work because opencv lib cant be compiled rightly.
To FFmpeg,my configure commend is following:
./configure --enable-cross-compile --cross-prefix=/home/midsummer/Tool/gcc-linaro-6.3.1-2017.05-x86_64_aarch64-linux-gnu/bin/aarch64-linux-gnu- --target-os=linux --arch=aarch64 (...)
06:03
I want to extract single frames (as thumbs) from videos, without having to download the full video file. I would like to just read the mp4 header (the info should be in the moov atom) and then download the required byte ranges. Afaik thats what browsers (HTML5 video) do when you skip to an unbuffered part
I already looked at: How can HTML5 video's byte-range requests (pseudo-streaming) work? but i can't figure out how to use it with ffmpeg for example.
Thanks a (...)
03:19
There is a USB camera that sends H264 frames with a frequency of 30 fps in a resolution of 1920x1080, GOP size 30 or 60 (1 or 2 I frame per second depending on the camera), which requires throughput of about 6 mbit/s. It is necessary to reduce this to no more than 2 mbit/s. All this is done on a weak ARMv7, so the option of transcoding is extremely resource intensive, especially considering that I did not find how to do it as a GPU Mali.
Given that the basic functionality is performed on the framework ffmpeg, is it possible to reduce the frame rate, ideally to 10 fps, WITHOUT loss of (...)
03:05
I'm testing the ddagrab filter with this command line:
ffmpeg -f lavfi -i ddagrab -c:v h264_nvenc -cq 18 -v 40 -t 10 -y out.mp4
FFmpeg exits with this error message:
OpenEncodeSessionEx failed: no encode device (1): (no details)
My laptop has Nvidia Geforce GTX 1050 Ti, Driver version 527.37. Windows 10 Home, Version 22H2, OS build 19045.4291. DirectX version 12.
This command works without problems:
ffmpeg -f gdigrab -i desktop -c:v h264_nvenc -t 10 -y out.mp4
I test these command on my other laptop with Nvidia Geforce RTX 2060, Driver version (...)