OOM on 3090
#60
by
TheBigBlockPC
- opened
I tried running this LLM on my dual 3090 PC but it runs on of memory in a single or even both 3090s. Quantizing uding botsandbytes doesn't work. I use transformers. Can someone help fixing that
TheBigBlockPC
changed discussion status to
closed