Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Int4 inference #45

Open
dmenig opened this issue Oct 30, 2024 · 2 comments
Open

Int4 inference #45

dmenig opened this issue Oct 30, 2024 · 2 comments

Comments

@dmenig
Copy link

dmenig commented Oct 30, 2024

Is there any plan to release an int4 version ? Specifically, I'm interested in the video understanding part.

@xffxff
Copy link
Collaborator

xffxff commented Nov 12, 2024

Yes! You can check out https://github.com/mobiusml/hqq/blob/master/examples/hf/aria_multimodal.py, a HQQ 4-bit version. This implementation is about 4-6x faster and takes 3x less VRAM!

@AIWarper
Copy link

AIWarper commented Dec 5, 2024

Yes! You can check out https://github.com/mobiusml/hqq/blob/master/examples/hf/aria_multimodal.py, a HQQ 4-bit version. This implementation is about 4-6x faster and takes 3x less VRAM!

I was looking into this as I need a local Video Vision solution - is it correct to assume the model was 12 shards at ~4gb a shard?

Also, you mentioned 3x less VRAM - what are the upper limits of the video vision you can achieve on 24gb or less?

EDIT: I tested locally with a 4090. Single image inference was fine but any video OOMed me no matter the length or resolution

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants