Spaces:
Sleeping
Sleeping
kwabs22
commited on
Commit
·
a7e3cb8
1
Parent(s):
1c670bf
CUDA location is probably zero issue
Browse files
app.py
CHANGED
|
@@ -11,7 +11,7 @@ model = None
|
|
| 11 |
def loadmodel():
|
| 12 |
global tokenizer, model
|
| 13 |
tokenizer = AutoTokenizer.from_pretrained("ISTA-DASLab/Meta-Llama-3.1-70B-AQLM-PV-2Bit-1x16")
|
| 14 |
-
model = AutoModelForCausalLM.from_pretrained("ISTA-DASLab/Meta-Llama-3.1-70B-AQLM-PV-2Bit-1x16", torch_dtype=
|
| 15 |
#model = model.to('cuda') # Move the model to GPU if available
|
| 16 |
pass
|
| 17 |
|
|
|
|
| 11 |
def loadmodel():
|
| 12 |
global tokenizer, model
|
| 13 |
tokenizer = AutoTokenizer.from_pretrained("ISTA-DASLab/Meta-Llama-3.1-70B-AQLM-PV-2Bit-1x16")
|
| 14 |
+
model = AutoModelForCausalLM.from_pretrained("ISTA-DASLab/Meta-Llama-3.1-70B-AQLM-PV-2Bit-1x16", torch_dtype='auto', device_map= 'auto') #torch_dtype=torch.float16
|
| 15 |
#model = model.to('cuda') # Move the model to GPU if available
|
| 16 |
pass
|
| 17 |
|