Skip to content

SafetensorError: Error while deserializing header: HeaderTooLarge #10

Open
@Nicolas-dds

Description

@Nicolas-dds

SafetensorError Traceback (most recent call last)
Cell In[2], line 11
7 # MODEL_PATH = '/home/quang/working/LLMs/oobabooga_linux/text-generation-webui/models/TheBloke_Wizard-Vicuna-7B-Uncensored-GPTQ'
8 # CHECKPOINT_PATH = '/home/quang/working/LLMs/oobabooga_linux/text-generation-webui/models/TheBloke_Wizard-Vicuna-7B-Uncensored-GPTQ/Wizard-Vicuna-7B-Uncensored-GPTQ-4bit-128g.no-act-order.safetensors'
10 DEVICE = torch.device('cuda:0')
---> 11 model, tokenizer = load_model_main(MODEL_PATH, CHECKPOINT_PATH, DEVICE)

File /mnt/c/Users/nlave/Maestria/Tesis/generativeAgent_LLM/server/model.py:56, in load_model_main(model_para, checkpoint_para, device)
55 def load_model_main(model_para, checkpoint_para, device):
---> 56 model = load_quant(model_para, checkpoint_para, 4, 128)
57 model.to(device)
58 tokenizer = AutoTokenizer.from_pretrained(model_para)

File /mnt/c/Users/nlave/Maestria/Tesis/generativeAgent_LLM/server/model.py:47, in load_quant(model, checkpoint, wbits, groupsize)
45 if checkpoint.endswith('.safetensors'):
46 from safetensors.torch import load_file as safe_load
---> 47 model.load_state_dict(safe_load(checkpoint))
48 else:
49 model.load_state_dict(torch.load(checkpoint))

File /mnt/c/Users/nlave/Maestria/Tesis/generativeAgent_LLM/env/lib/python3.10/site-packages/safetensors/torch.py:308, in load_file(filename, device)
285 """
286 Loads a safetensors file into torch format.
287
(...)
305 ```
306 """
307 result = {}
--> 308 with safe_open(filename, framework="pt", device=device) as f:
309 for k in f.keys():
310 result[k] = f.get_tensor(k)

SafetensorError: Error while deserializing header: HeaderTooLarge

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions