Initial Release
https://github.com/Jony12rb/Hackupc24_inter/assets/99959182/cfde0de4-f683-4a11-bebf-c219fd4ef7b1
The following photo shows the schema of the project. Our model is a text-to-video project, although we do not use a generative video model. Instead, we use the user’s photos together with ai music generation to create a video. The user can upload photos, and the model will generate a video with the photos and music. The music is generated using an ai model, and the video is created using the photo and the music. The user can download the video and share it with friends.
Moreover, we offer the functionality of selecting the photos and music using natural language. The user can write a sentence, and the model will select the photos and music based on the sentence, which makes it more user-friendly. Then, it will create the music according to the sentence and the photos.
After cloning the repo, stand on the main folder and do the following:
docker run -d --name iris-comm -p 1972:1972 -p 52773:52773 -e IRIS_PASSWORD=demo -e IRIS_USERNAME=demo intersystemsdc/iris-community:latest
:information_source: After running the above command, you can access the System Management Portal via http://localhost:52773/csp/sys/UtilHome.csp. Please note you may need to configure your web server separately when using another product edition.
conda create --name iris-vector-search python=3.10
conda activate iris-vector-search
pip install -r requirements.txt
conda install -c conda-forge ffmpeg
Create a .env file in the main folder and write the following:
OPENAI_API_KEY=your_openai_api_key
REPLICATE_API_KEY=your_replicate_api_key
streamlit run ./demo/GalleryVideoclipGenerator.py