Skip to content

Video Generation via Tokens #5

@ClashLuke

Description

@ClashLuke

If we tokenise frames of a video with a VQGAN, we can autoregressively predict the next token using our current language model. More specifically, using our current context of 2 million tokens, we could fit 2048 frames (~34 minutes at 1 FPS) with current state-of-the-art image quantisation models.
This issue is about implementing such a model end-to-end and having a working demo.

Metadata

Metadata

Assignees

No one assigned

    Labels

    MLRequires machine-learning knowledge (can be built up on the fly)downstreamChanges code wrapping the core modelresearchCreative project that might fail but could give high returns

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions