Title of the post fails to describe what this fully does and you didn't seem to have read the repo. This generates a live interaction with the website, with scrolls, timings, clicks, route navigations, etc. A mixture of AI generated cues that you can use via skills + scripts to be feed with those cues that will process the video automatically via headless chromium and ffmpeg. Seems really interesting
You could do this with ffmpeg on a screenshot PNG.
ffmpeg -i screenshot.png -vf \ "crop=iw:1080:0:n*(20000-1080)/600,format=yuv420p" \ -t 10 -r 60 output.mp4
Title of the post fails to describe what this fully does and you didn't seem to have read the repo. This generates a live interaction with the website, with scrolls, timings, clicks, route navigations, etc. A mixture of AI generated cues that you can use via skills + scripts to be feed with those cues that will process the video automatically via headless chromium and ffmpeg. Seems really interesting
What problem does this solve?
Many websites have looping or scroll-triggered animations, or play gifs / videos that a static image wouldn't capture.