Did any one succeed on getting OPT-13B-Erebus to work on 3080ti? #94
Replies: 2 comments 3 replies
-
The memory usage of OPT-13B-Erebus with I am not sure if it is possible to mix offloading with 8bit (I think not), so the best currently available option would be to use deepspeed: https://github.com/oobabooga/text-generation-webui/wiki/DeepSpeed or simply --auto-devices with --gpu-memory 10 or similar. |
Beta Was this translation helpful? Give feedback.
-
Anyone trying this, I got it working on a system with 3080ti with 64GB ram using the following command |
Beta Was this translation helpful? Give feedback.
-
I'm using
python server.py --auto-devices --cai-chat --load-in-8bit --listen --listen-port=8888
But I'm getting
I have 12GB of VRAM and 64GB or RAM, and I think those should load the model correctly with
--auto-devices
Is there something wrong that I'm doing or is it impossible to get it to run with this hardware?
Beta Was this translation helpful? Give feedback.
All reactions