Skip to content
/ artifice Public

using generative ai image-to-video ML to hallucinate extended video sequences from an initial source image.

License

Notifications You must be signed in to change notification settings

m-onz/artifice

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

27 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

artifice

about

using generative ai image-to-video ML to hallucinate extended video sequences from a single source image.

you will need an API key from replicate.com exported as an environment variable.

dependencies

tested on debian linux

this requires ffmpeg

this costs money to run *see replicate.com pricing

put your images in the ./images folder. mp4 output is in the output folder.

sudo apt-get install ffmpeg
git clone https://github.com/m-onz/artifice
npm i
export REPLICATE_API_TOKEN=r8_BRU**********************************
node hallucinate.js

wait a long time!

generating initial images

You can add images to the images folder from any source. I used this model

With the prompt:

shocking abstract 3D art in the style of andy warhol and francis bacon for a gallery that shocks the viewer exploring digital, glitch and modern culture, distorted abstract wireframe mesh forms

generate gifs

You can use this model to generate a .gif from an .mp4 video.

an example video

You can see an example video of this output here

models

  • Updated to use /minimax/video-01 for 6 second video image-to-video generation
  • You can use a text-to-image model to generate initial images.. or use any image you like

About

using generative ai image-to-video ML to hallucinate extended video sequences from an initial source image.

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published