using generative ai image-to-video ML to hallucinate extended video sequences from a single source image.
you will need an API key from replicate.com exported as an environment variable.
tested on debian linux
this requires ffmpeg
this costs money to run *see replicate.com pricing
put your images in the ./images folder. mp4 output is in the output folder.
sudo apt-get install ffmpeg
git clone https://github.com/m-onz/artifice
npm i
export REPLICATE_API_TOKEN=r8_BRU**********************************
node hallucinate.js
wait a long time!
You can add images to the images
folder from any source. I used this model
With the prompt:
shocking abstract 3D art in the style of andy warhol and francis bacon for a gallery that shocks the viewer exploring digital, glitch and modern culture, distorted abstract wireframe mesh forms
You can use this model to generate a .gif from an .mp4 video.
You can see an example video of this output here
- Updated to use /minimax/video-01 for 6 second video image-to-video generation
- You can use a text-to-image model to generate initial images.. or use any image you like