Extending Text2Video-Zero for Multi-ControlNet

dc.contributor.advisorShi, Humphrey
dc.contributor.advisorMossberg, Barbara
dc.contributor.authorBacken, Ben
dc.date.accessioned2023-08-18T15:48:20Z
dc.date.available2023-08-18T15:48:20Z
dc.date.issued2023
dc.description15 pagesen_US
dc.description.abstractThis research paper presents an extension to the Text2Video-Zero (T2V0) generative model, augmenting the synthesis of video from textual and video inputs. The project focuses on enhancing the functionality and accessibility of T2V0 by integrating Stable Diffusion’s (SD) support for multiple ControlNets, implementing frame-wise masking for selective ControlNet application, and introducing memory optimizations to enable running the model on consumer-grade hardware. The paper also provides a high-level overview of SD, explores experimental features, and offers practical tips for generating videos using these tools. Additionally, we include a demonstration video showcasing T2V0 with Multi-ControlNet. The video highlights the early potential of text-to-video models for storytelling. Ultimately, the study strives to expand the capabilities and accessibility of T2V0, increasing users' control over their generated outputs while upholding the democratic principles of open-source AI.en_US
dc.identifier.orcid0009-0005-0548-7369
dc.identifier.urihttps://hdl.handle.net/1794/28647
dc.language.isoen_US
dc.publisherUniversity of Oregon
dc.rightsCC BY-NC-ND 4.0
dc.subjecttext-to-videoen_US
dc.subjectStable Diffusionen_US
dc.subjectControlNeten_US
dc.subjectmachine learningen_US
dc.subjectgenerative modelsen_US
dc.titleExtending Text2Video-Zero for Multi-ControlNet
dc.typeThesis/Dissertation

Files

Original bundle
Now showing 1 - 1 of 1
Loading...
Thumbnail Image
Name:
Backen_Ben_Thesis_CHC.pdf
Size:
470.33 KB
Format:
Adobe Portable Document Format
Description:
License bundle
Now showing 1 - 1 of 1
Name:
license.txt
Size:
2.12 KB
Format:
Plain Text
Description: