Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Running convert fails with BadZipFile (Bad CRC-32) #4365

Closed
4 tasks done
itsJoKr opened this issue Dec 7, 2023 · 14 comments
Closed
4 tasks done

Running convert fails with BadZipFile (Bad CRC-32) #4365

itsJoKr opened this issue Dec 7, 2023 · 14 comments

Comments

@itsJoKr
Copy link

itsJoKr commented Dec 7, 2023

Prerequisites

Please answer the following questions for yourself before submitting an issue.

  • I am running the latest code. Development is very rapid so there are no tagged versions as of now.
  • I carefully followed the README.md.
  • I searched using keywords relevant to my issue to make sure that I am creating a new issue that is not already open (or closed).
  • I reviewed the Discussions, and have a new bug or useful enhancement to share.

Expected Behavior

The convert runs successfully.

Current Behavior

The convert fails with error like

  File "/Library/Frameworks/Python.framework/Versions/3.12/lib/python3.12/zipfile/__init__.py", line 996, in _update_crc
    raise BadZipFile("Bad CRC-32 for file %r" % self.name)

Environment and Context

Please provide detailed information about your computer setup. This is important in case the issue is not reproducible except for under certain specific conditions.

MacBook Pro, M1 Pro, MacOS Sonoma
Python: 3.12.0
Make: GNU Make 3.81

I tried to run for both OpenLlama-3B and Llama-7B-chat, same error.

Steps to Reproduce

Please provide detailed steps for reproducing the issue. We are not sitting in front of your screen, so the more detail the better.

  1. Clone llama and llama.cpp
  2. Make llama.cpp
  3. Download llama 7B-chat model
  4. Run convert

Failure Logs

llama.cpp$ git log | head 
commit 5a7d3125e7c24f223659b7f0b7aa7736986e92c0

llama.cpp$ python3 --version
Python 3.12.0

llama.cpp$ pip list | egrep "torch|numpy|sentencepiece"
numpy              1.26.2
sentencepiece      0.1.99

@skidd-level-100
Copy link

same issue for all of my models, worked before git pull + recomp

@LIMPIX31
Copy link

LIMPIX31 commented Jan 9, 2024

This seems to have been fixed on the last commit.

@skidd-level-100
Copy link

This seems to have been fixed on the last commit.

not for me still 'convert.py path/to/model' and 'convert.py path/to/model.bin' fail with the last to lines being:

raise BadZipFile("Bad CRC-32 for file %r" % self.name)

zipfile.BadZipFile: Bad CRC-32 for file 'archive/data/13'

@aspanner
Copy link

I've got the same issue. Here's my test run after downloading the model from hf:

python3.12 llama.cpp/convert.py aiopsmodel-hfv2 --outfile llama-2-7b-aiopsfinetunedv2-q8_0-gguf --outtype q8_0
Loading model file aiopsmodel-hfv2/pytorch_model-00001-of-00002.bin
Loading model file aiopsmodel-hfv2/pytorch_model-00001-of-00002.bin
Loading model file aiopsmodel-hfv2/pytorch_model-00002-of-00002.bin
params = Params(n_vocab=32001, n_embd=4096, n_layer=32, n_ctx=4096, n_ff=11008, n_head=32, n_head_kv=32, n_experts=None, n_experts_used=None, f_norm_eps=1e-05, rope_scaling_type=None, f_rope_freq_base=None, f_rope_scale=None, n_orig_ctx=None, rope_finetuned=None, ftype=<GGMLFileType.MostlyQ8_0: 7>, path_model=PosixPath('aiopsmodel-hfv2'))
Found vocab files: {'tokenizer.model': PosixPath('aiopsmodel-hfv2/tokenizer.model'), 'vocab.json': None, 'tokenizer.json': PosixPath('aiopsmodel-hfv2/tokenizer.json')}
Loading vocab file 'aiopsmodel-hfv2/tokenizer.model', type 'spm'
Vocab info: <SentencePieceVocab with 32000 base tokens and 1 added tokens>
Special vocab info: <SpecialVocab with 0 merges, special tokens {'bos': 1, 'eos': 2, 'unk': 0, 'pad': 0}, add special tokens unset>
Permuting layer 0
Permuting layer 1
Permuting layer 2
Permuting layer 3
Permuting layer 4
Permuting layer 5
Permuting layer 6
Permuting layer 7
Permuting layer 8
Permuting layer 9
Permuting layer 10
Permuting layer 11
Permuting layer 12
Permuting layer 13
Permuting layer 14
Permuting layer 15
Permuting layer 16
Permuting layer 17
Permuting layer 18
Permuting layer 19
Permuting layer 20
Permuting layer 21
Permuting layer 22
Permuting layer 23
Permuting layer 24
Permuting layer 25
Permuting layer 26
Permuting layer 27
Permuting layer 28
Permuting layer 29
Permuting layer 30
Permuting layer 31
model.embed_tokens.weight -> token_embd.weight | F16 | [32000, 4096]
model.layers.0.self_attn.q_proj.weight -> blk.0.attn_q.weight | F16 | [4096, 4096]
model.layers.0.self_attn.k_proj.weight -> blk.0.attn_k.weight | F16 | [4096, 4096]
model.layers.0.self_attn.v_proj.weight -> blk.0.attn_v.weight | F16 | [4096, 4096]
model.layers.0.self_attn.o_proj.weight -> blk.0.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.0.attn_rot_embd
model.layers.0.mlp.gate_proj.weight -> blk.0.ffn_gate.weight | F16 | [11008, 4096]
model.layers.0.mlp.up_proj.weight -> blk.0.ffn_up.weight | F16 | [11008, 4096]
model.layers.0.mlp.down_proj.weight -> blk.0.ffn_down.weight | F16 | [4096, 11008]
model.layers.0.input_layernorm.weight -> blk.0.attn_norm.weight | F16 | [4096]
model.layers.0.post_attention_layernorm.weight -> blk.0.ffn_norm.weight | F16 | [4096]
model.layers.1.self_attn.q_proj.weight -> blk.1.attn_q.weight | F16 | [4096, 4096]
model.layers.1.self_attn.k_proj.weight -> blk.1.attn_k.weight | F16 | [4096, 4096]
model.layers.1.self_attn.v_proj.weight -> blk.1.attn_v.weight | F16 | [4096, 4096]
model.layers.1.self_attn.o_proj.weight -> blk.1.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.1.attn_rot_embd
model.layers.1.mlp.gate_proj.weight -> blk.1.ffn_gate.weight | F16 | [11008, 4096]
model.layers.1.mlp.up_proj.weight -> blk.1.ffn_up.weight | F16 | [11008, 4096]
model.layers.1.mlp.down_proj.weight -> blk.1.ffn_down.weight | F16 | [4096, 11008]
model.layers.1.input_layernorm.weight -> blk.1.attn_norm.weight | F16 | [4096]
model.layers.1.post_attention_layernorm.weight -> blk.1.ffn_norm.weight | F16 | [4096]
model.layers.2.self_attn.q_proj.weight -> blk.2.attn_q.weight | F16 | [4096, 4096]
model.layers.2.self_attn.k_proj.weight -> blk.2.attn_k.weight | F16 | [4096, 4096]
model.layers.2.self_attn.v_proj.weight -> blk.2.attn_v.weight | F16 | [4096, 4096]
model.layers.2.self_attn.o_proj.weight -> blk.2.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.2.attn_rot_embd
model.layers.2.mlp.gate_proj.weight -> blk.2.ffn_gate.weight | F16 | [11008, 4096]
model.layers.2.mlp.up_proj.weight -> blk.2.ffn_up.weight | F16 | [11008, 4096]
model.layers.2.mlp.down_proj.weight -> blk.2.ffn_down.weight | F16 | [4096, 11008]
model.layers.2.input_layernorm.weight -> blk.2.attn_norm.weight | F16 | [4096]
model.layers.2.post_attention_layernorm.weight -> blk.2.ffn_norm.weight | F16 | [4096]
model.layers.3.self_attn.q_proj.weight -> blk.3.attn_q.weight | F16 | [4096, 4096]
model.layers.3.self_attn.k_proj.weight -> blk.3.attn_k.weight | F16 | [4096, 4096]
model.layers.3.self_attn.v_proj.weight -> blk.3.attn_v.weight | F16 | [4096, 4096]
model.layers.3.self_attn.o_proj.weight -> blk.3.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.3.attn_rot_embd
model.layers.3.mlp.gate_proj.weight -> blk.3.ffn_gate.weight | F16 | [11008, 4096]
model.layers.3.mlp.up_proj.weight -> blk.3.ffn_up.weight | F16 | [11008, 4096]
model.layers.3.mlp.down_proj.weight -> blk.3.ffn_down.weight | F16 | [4096, 11008]
model.layers.3.input_layernorm.weight -> blk.3.attn_norm.weight | F16 | [4096]
model.layers.3.post_attention_layernorm.weight -> blk.3.ffn_norm.weight | F16 | [4096]
model.layers.4.self_attn.q_proj.weight -> blk.4.attn_q.weight | F16 | [4096, 4096]
model.layers.4.self_attn.k_proj.weight -> blk.4.attn_k.weight | F16 | [4096, 4096]
model.layers.4.self_attn.v_proj.weight -> blk.4.attn_v.weight | F16 | [4096, 4096]
model.layers.4.self_attn.o_proj.weight -> blk.4.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.4.attn_rot_embd
model.layers.4.mlp.gate_proj.weight -> blk.4.ffn_gate.weight | F16 | [11008, 4096]
model.layers.4.mlp.up_proj.weight -> blk.4.ffn_up.weight | F16 | [11008, 4096]
model.layers.4.mlp.down_proj.weight -> blk.4.ffn_down.weight | F16 | [4096, 11008]
model.layers.4.input_layernorm.weight -> blk.4.attn_norm.weight | F16 | [4096]
model.layers.4.post_attention_layernorm.weight -> blk.4.ffn_norm.weight | F16 | [4096]
model.layers.5.self_attn.q_proj.weight -> blk.5.attn_q.weight | F16 | [4096, 4096]
model.layers.5.self_attn.k_proj.weight -> blk.5.attn_k.weight | F16 | [4096, 4096]
model.layers.5.self_attn.v_proj.weight -> blk.5.attn_v.weight | F16 | [4096, 4096]
model.layers.5.self_attn.o_proj.weight -> blk.5.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.5.attn_rot_embd
model.layers.5.mlp.gate_proj.weight -> blk.5.ffn_gate.weight | F16 | [11008, 4096]
model.layers.5.mlp.up_proj.weight -> blk.5.ffn_up.weight | F16 | [11008, 4096]
model.layers.5.mlp.down_proj.weight -> blk.5.ffn_down.weight | F16 | [4096, 11008]
model.layers.5.input_layernorm.weight -> blk.5.attn_norm.weight | F16 | [4096]
model.layers.5.post_attention_layernorm.weight -> blk.5.ffn_norm.weight | F16 | [4096]
model.layers.6.self_attn.q_proj.weight -> blk.6.attn_q.weight | F16 | [4096, 4096]
model.layers.6.self_attn.k_proj.weight -> blk.6.attn_k.weight | F16 | [4096, 4096]
model.layers.6.self_attn.v_proj.weight -> blk.6.attn_v.weight | F16 | [4096, 4096]
model.layers.6.self_attn.o_proj.weight -> blk.6.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.6.attn_rot_embd
model.layers.6.mlp.gate_proj.weight -> blk.6.ffn_gate.weight | F16 | [11008, 4096]
model.layers.6.mlp.up_proj.weight -> blk.6.ffn_up.weight | F16 | [11008, 4096]
model.layers.6.mlp.down_proj.weight -> blk.6.ffn_down.weight | F16 | [4096, 11008]
model.layers.6.input_layernorm.weight -> blk.6.attn_norm.weight | F16 | [4096]
model.layers.6.post_attention_layernorm.weight -> blk.6.ffn_norm.weight | F16 | [4096]
model.layers.7.self_attn.q_proj.weight -> blk.7.attn_q.weight | F16 | [4096, 4096]
model.layers.7.self_attn.k_proj.weight -> blk.7.attn_k.weight | F16 | [4096, 4096]
model.layers.7.self_attn.v_proj.weight -> blk.7.attn_v.weight | F16 | [4096, 4096]
model.layers.7.self_attn.o_proj.weight -> blk.7.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.7.attn_rot_embd
model.layers.7.mlp.gate_proj.weight -> blk.7.ffn_gate.weight | F16 | [11008, 4096]
model.layers.7.mlp.up_proj.weight -> blk.7.ffn_up.weight | F16 | [11008, 4096]
model.layers.7.mlp.down_proj.weight -> blk.7.ffn_down.weight | F16 | [4096, 11008]
model.layers.7.input_layernorm.weight -> blk.7.attn_norm.weight | F16 | [4096]
model.layers.7.post_attention_layernorm.weight -> blk.7.ffn_norm.weight | F16 | [4096]
model.layers.8.self_attn.q_proj.weight -> blk.8.attn_q.weight | F16 | [4096, 4096]
model.layers.8.self_attn.k_proj.weight -> blk.8.attn_k.weight | F16 | [4096, 4096]
model.layers.8.self_attn.v_proj.weight -> blk.8.attn_v.weight | F16 | [4096, 4096]
model.layers.8.self_attn.o_proj.weight -> blk.8.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.8.attn_rot_embd
model.layers.8.mlp.gate_proj.weight -> blk.8.ffn_gate.weight | F16 | [11008, 4096]
model.layers.8.mlp.up_proj.weight -> blk.8.ffn_up.weight | F16 | [11008, 4096]
model.layers.8.mlp.down_proj.weight -> blk.8.ffn_down.weight | F16 | [4096, 11008]
model.layers.8.input_layernorm.weight -> blk.8.attn_norm.weight | F16 | [4096]
model.layers.8.post_attention_layernorm.weight -> blk.8.ffn_norm.weight | F16 | [4096]
model.layers.9.self_attn.q_proj.weight -> blk.9.attn_q.weight | F16 | [4096, 4096]
model.layers.9.self_attn.k_proj.weight -> blk.9.attn_k.weight | F16 | [4096, 4096]
model.layers.9.self_attn.v_proj.weight -> blk.9.attn_v.weight | F16 | [4096, 4096]
model.layers.9.self_attn.o_proj.weight -> blk.9.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.9.attn_rot_embd
model.layers.9.mlp.gate_proj.weight -> blk.9.ffn_gate.weight | F16 | [11008, 4096]
model.layers.9.mlp.up_proj.weight -> blk.9.ffn_up.weight | F16 | [11008, 4096]
model.layers.9.mlp.down_proj.weight -> blk.9.ffn_down.weight | F16 | [4096, 11008]
model.layers.9.input_layernorm.weight -> blk.9.attn_norm.weight | F16 | [4096]
model.layers.9.post_attention_layernorm.weight -> blk.9.ffn_norm.weight | F16 | [4096]
model.layers.10.self_attn.q_proj.weight -> blk.10.attn_q.weight | F16 | [4096, 4096]
model.layers.10.self_attn.k_proj.weight -> blk.10.attn_k.weight | F16 | [4096, 4096]
model.layers.10.self_attn.v_proj.weight -> blk.10.attn_v.weight | F16 | [4096, 4096]
model.layers.10.self_attn.o_proj.weight -> blk.10.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.10.attn_rot_embd
model.layers.10.mlp.gate_proj.weight -> blk.10.ffn_gate.weight | F16 | [11008, 4096]
model.layers.10.mlp.up_proj.weight -> blk.10.ffn_up.weight | F16 | [11008, 4096]
model.layers.10.mlp.down_proj.weight -> blk.10.ffn_down.weight | F16 | [4096, 11008]
model.layers.10.input_layernorm.weight -> blk.10.attn_norm.weight | F16 | [4096]
model.layers.10.post_attention_layernorm.weight -> blk.10.ffn_norm.weight | F16 | [4096]
model.layers.11.self_attn.q_proj.weight -> blk.11.attn_q.weight | F16 | [4096, 4096]
model.layers.11.self_attn.k_proj.weight -> blk.11.attn_k.weight | F16 | [4096, 4096]
model.layers.11.self_attn.v_proj.weight -> blk.11.attn_v.weight | F16 | [4096, 4096]
model.layers.11.self_attn.o_proj.weight -> blk.11.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.11.attn_rot_embd
model.layers.11.mlp.gate_proj.weight -> blk.11.ffn_gate.weight | F16 | [11008, 4096]
model.layers.11.mlp.up_proj.weight -> blk.11.ffn_up.weight | F16 | [11008, 4096]
model.layers.11.mlp.down_proj.weight -> blk.11.ffn_down.weight | F16 | [4096, 11008]
model.layers.11.input_layernorm.weight -> blk.11.attn_norm.weight | F16 | [4096]
model.layers.11.post_attention_layernorm.weight -> blk.11.ffn_norm.weight | F16 | [4096]
model.layers.12.self_attn.q_proj.weight -> blk.12.attn_q.weight | F16 | [4096, 4096]
model.layers.12.self_attn.k_proj.weight -> blk.12.attn_k.weight | F16 | [4096, 4096]
model.layers.12.self_attn.v_proj.weight -> blk.12.attn_v.weight | F16 | [4096, 4096]
model.layers.12.self_attn.o_proj.weight -> blk.12.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.12.attn_rot_embd
model.layers.12.mlp.gate_proj.weight -> blk.12.ffn_gate.weight | F16 | [11008, 4096]
model.layers.12.mlp.up_proj.weight -> blk.12.ffn_up.weight | F16 | [11008, 4096]
model.layers.12.mlp.down_proj.weight -> blk.12.ffn_down.weight | F16 | [4096, 11008]
model.layers.12.input_layernorm.weight -> blk.12.attn_norm.weight | F16 | [4096]
model.layers.12.post_attention_layernorm.weight -> blk.12.ffn_norm.weight | F16 | [4096]
model.layers.13.self_attn.q_proj.weight -> blk.13.attn_q.weight | F16 | [4096, 4096]
model.layers.13.self_attn.k_proj.weight -> blk.13.attn_k.weight | F16 | [4096, 4096]
model.layers.13.self_attn.v_proj.weight -> blk.13.attn_v.weight | F16 | [4096, 4096]
model.layers.13.self_attn.o_proj.weight -> blk.13.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.13.attn_rot_embd
model.layers.13.mlp.gate_proj.weight -> blk.13.ffn_gate.weight | F16 | [11008, 4096]
model.layers.13.mlp.up_proj.weight -> blk.13.ffn_up.weight | F16 | [11008, 4096]
model.layers.13.mlp.down_proj.weight -> blk.13.ffn_down.weight | F16 | [4096, 11008]
model.layers.13.input_layernorm.weight -> blk.13.attn_norm.weight | F16 | [4096]
model.layers.13.post_attention_layernorm.weight -> blk.13.ffn_norm.weight | F16 | [4096]
model.layers.14.self_attn.q_proj.weight -> blk.14.attn_q.weight | F16 | [4096, 4096]
model.layers.14.self_attn.k_proj.weight -> blk.14.attn_k.weight | F16 | [4096, 4096]
model.layers.14.self_attn.v_proj.weight -> blk.14.attn_v.weight | F16 | [4096, 4096]
model.layers.14.self_attn.o_proj.weight -> blk.14.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.14.attn_rot_embd
model.layers.14.mlp.gate_proj.weight -> blk.14.ffn_gate.weight | F16 | [11008, 4096]
model.layers.14.mlp.up_proj.weight -> blk.14.ffn_up.weight | F16 | [11008, 4096]
model.layers.14.mlp.down_proj.weight -> blk.14.ffn_down.weight | F16 | [4096, 11008]
model.layers.14.input_layernorm.weight -> blk.14.attn_norm.weight | F16 | [4096]
model.layers.14.post_attention_layernorm.weight -> blk.14.ffn_norm.weight | F16 | [4096]
model.layers.15.self_attn.q_proj.weight -> blk.15.attn_q.weight | F16 | [4096, 4096]
model.layers.15.self_attn.k_proj.weight -> blk.15.attn_k.weight | F16 | [4096, 4096]
model.layers.15.self_attn.v_proj.weight -> blk.15.attn_v.weight | F16 | [4096, 4096]
model.layers.15.self_attn.o_proj.weight -> blk.15.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.15.attn_rot_embd
model.layers.15.mlp.gate_proj.weight -> blk.15.ffn_gate.weight | F16 | [11008, 4096]
model.layers.15.mlp.up_proj.weight -> blk.15.ffn_up.weight | F16 | [11008, 4096]
model.layers.15.mlp.down_proj.weight -> blk.15.ffn_down.weight | F16 | [4096, 11008]
model.layers.15.input_layernorm.weight -> blk.15.attn_norm.weight | F16 | [4096]
model.layers.15.post_attention_layernorm.weight -> blk.15.ffn_norm.weight | F16 | [4096]
model.layers.16.self_attn.q_proj.weight -> blk.16.attn_q.weight | F16 | [4096, 4096]
model.layers.16.self_attn.k_proj.weight -> blk.16.attn_k.weight | F16 | [4096, 4096]
model.layers.16.self_attn.v_proj.weight -> blk.16.attn_v.weight | F16 | [4096, 4096]
model.layers.16.self_attn.o_proj.weight -> blk.16.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.16.attn_rot_embd
model.layers.16.mlp.gate_proj.weight -> blk.16.ffn_gate.weight | F16 | [11008, 4096]
model.layers.16.mlp.up_proj.weight -> blk.16.ffn_up.weight | F16 | [11008, 4096]
model.layers.16.mlp.down_proj.weight -> blk.16.ffn_down.weight | F16 | [4096, 11008]
model.layers.16.input_layernorm.weight -> blk.16.attn_norm.weight | F16 | [4096]
model.layers.16.post_attention_layernorm.weight -> blk.16.ffn_norm.weight | F16 | [4096]
model.layers.17.self_attn.q_proj.weight -> blk.17.attn_q.weight | F16 | [4096, 4096]
model.layers.17.self_attn.k_proj.weight -> blk.17.attn_k.weight | F16 | [4096, 4096]
model.layers.17.self_attn.v_proj.weight -> blk.17.attn_v.weight | F16 | [4096, 4096]
model.layers.17.self_attn.o_proj.weight -> blk.17.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.17.attn_rot_embd
model.layers.17.mlp.gate_proj.weight -> blk.17.ffn_gate.weight | F16 | [11008, 4096]
model.layers.17.mlp.up_proj.weight -> blk.17.ffn_up.weight | F16 | [11008, 4096]
model.layers.17.mlp.down_proj.weight -> blk.17.ffn_down.weight | F16 | [4096, 11008]
model.layers.17.input_layernorm.weight -> blk.17.attn_norm.weight | F16 | [4096]
model.layers.17.post_attention_layernorm.weight -> blk.17.ffn_norm.weight | F16 | [4096]
model.layers.18.self_attn.q_proj.weight -> blk.18.attn_q.weight | F16 | [4096, 4096]
model.layers.18.self_attn.k_proj.weight -> blk.18.attn_k.weight | F16 | [4096, 4096]
model.layers.18.self_attn.v_proj.weight -> blk.18.attn_v.weight | F16 | [4096, 4096]
model.layers.18.self_attn.o_proj.weight -> blk.18.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.18.attn_rot_embd
model.layers.18.mlp.gate_proj.weight -> blk.18.ffn_gate.weight | F16 | [11008, 4096]
model.layers.18.mlp.up_proj.weight -> blk.18.ffn_up.weight | F16 | [11008, 4096]
model.layers.18.mlp.down_proj.weight -> blk.18.ffn_down.weight | F16 | [4096, 11008]
model.layers.18.input_layernorm.weight -> blk.18.attn_norm.weight | F16 | [4096]
model.layers.18.post_attention_layernorm.weight -> blk.18.ffn_norm.weight | F16 | [4096]
model.layers.19.self_attn.q_proj.weight -> blk.19.attn_q.weight | F16 | [4096, 4096]
model.layers.19.self_attn.k_proj.weight -> blk.19.attn_k.weight | F16 | [4096, 4096]
model.layers.19.self_attn.v_proj.weight -> blk.19.attn_v.weight | F16 | [4096, 4096]
model.layers.19.self_attn.o_proj.weight -> blk.19.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.19.attn_rot_embd
model.layers.19.mlp.gate_proj.weight -> blk.19.ffn_gate.weight | F16 | [11008, 4096]
model.layers.19.mlp.up_proj.weight -> blk.19.ffn_up.weight | F16 | [11008, 4096]
model.layers.19.mlp.down_proj.weight -> blk.19.ffn_down.weight | F16 | [4096, 11008]
model.layers.19.input_layernorm.weight -> blk.19.attn_norm.weight | F16 | [4096]
model.layers.19.post_attention_layernorm.weight -> blk.19.ffn_norm.weight | F16 | [4096]
model.layers.20.self_attn.q_proj.weight -> blk.20.attn_q.weight | F16 | [4096, 4096]
model.layers.20.self_attn.k_proj.weight -> blk.20.attn_k.weight | F16 | [4096, 4096]
model.layers.20.self_attn.v_proj.weight -> blk.20.attn_v.weight | F16 | [4096, 4096]
model.layers.20.self_attn.o_proj.weight -> blk.20.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.20.attn_rot_embd
model.layers.20.mlp.gate_proj.weight -> blk.20.ffn_gate.weight | F16 | [11008, 4096]
model.layers.20.mlp.up_proj.weight -> blk.20.ffn_up.weight | F16 | [11008, 4096]
model.layers.20.mlp.down_proj.weight -> blk.20.ffn_down.weight | F16 | [4096, 11008]
model.layers.20.input_layernorm.weight -> blk.20.attn_norm.weight | F16 | [4096]
model.layers.20.post_attention_layernorm.weight -> blk.20.ffn_norm.weight | F16 | [4096]
model.layers.21.self_attn.q_proj.weight -> blk.21.attn_q.weight | F16 | [4096, 4096]
model.layers.21.self_attn.k_proj.weight -> blk.21.attn_k.weight | F16 | [4096, 4096]
model.layers.21.self_attn.v_proj.weight -> blk.21.attn_v.weight | F16 | [4096, 4096]
model.layers.21.self_attn.o_proj.weight -> blk.21.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.21.attn_rot_embd
model.layers.21.mlp.gate_proj.weight -> blk.21.ffn_gate.weight | F16 | [11008, 4096]
model.layers.21.mlp.up_proj.weight -> blk.21.ffn_up.weight | F16 | [11008, 4096]
model.layers.21.mlp.down_proj.weight -> blk.21.ffn_down.weight | F16 | [4096, 11008]
model.layers.21.input_layernorm.weight -> blk.21.attn_norm.weight | F16 | [4096]
model.layers.21.post_attention_layernorm.weight -> blk.21.ffn_norm.weight | F16 | [4096]
model.layers.22.self_attn.q_proj.weight -> blk.22.attn_q.weight | F16 | [4096, 4096]
model.layers.22.self_attn.k_proj.weight -> blk.22.attn_k.weight | F16 | [4096, 4096]
model.layers.22.self_attn.v_proj.weight -> blk.22.attn_v.weight | F16 | [4096, 4096]
model.layers.22.self_attn.o_proj.weight -> blk.22.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.22.attn_rot_embd
model.layers.22.mlp.gate_proj.weight -> blk.22.ffn_gate.weight | F16 | [11008, 4096]
model.layers.22.mlp.up_proj.weight -> blk.22.ffn_up.weight | F16 | [11008, 4096]
model.layers.22.mlp.down_proj.weight -> blk.22.ffn_down.weight | F16 | [4096, 11008]
model.layers.22.input_layernorm.weight -> blk.22.attn_norm.weight | F16 | [4096]
model.layers.22.post_attention_layernorm.weight -> blk.22.ffn_norm.weight | F16 | [4096]
model.layers.23.self_attn.q_proj.weight -> blk.23.attn_q.weight | F16 | [4096, 4096]
model.layers.23.self_attn.k_proj.weight -> blk.23.attn_k.weight | F16 | [4096, 4096]
model.layers.23.self_attn.v_proj.weight -> blk.23.attn_v.weight | F16 | [4096, 4096]
model.layers.23.self_attn.o_proj.weight -> blk.23.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.23.attn_rot_embd
model.layers.23.mlp.gate_proj.weight -> blk.23.ffn_gate.weight | F16 | [11008, 4096]
model.layers.23.mlp.up_proj.weight -> blk.23.ffn_up.weight | F16 | [11008, 4096]
model.layers.23.mlp.down_proj.weight -> blk.23.ffn_down.weight | F16 | [4096, 11008]
model.layers.23.input_layernorm.weight -> blk.23.attn_norm.weight | F16 | [4096]
model.layers.23.post_attention_layernorm.weight -> blk.23.ffn_norm.weight | F16 | [4096]
model.layers.24.self_attn.q_proj.weight -> blk.24.attn_q.weight | F16 | [4096, 4096]
model.layers.24.self_attn.k_proj.weight -> blk.24.attn_k.weight | F16 | [4096, 4096]
model.layers.24.self_attn.v_proj.weight -> blk.24.attn_v.weight | F16 | [4096, 4096]
model.layers.24.self_attn.o_proj.weight -> blk.24.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.24.attn_rot_embd
model.layers.24.mlp.gate_proj.weight -> blk.24.ffn_gate.weight | F16 | [11008, 4096]
model.layers.24.mlp.up_proj.weight -> blk.24.ffn_up.weight | F16 | [11008, 4096]
model.layers.24.mlp.down_proj.weight -> blk.24.ffn_down.weight | F16 | [4096, 11008]
model.layers.24.input_layernorm.weight -> blk.24.attn_norm.weight | F16 | [4096]
model.layers.24.post_attention_layernorm.weight -> blk.24.ffn_norm.weight | F16 | [4096]
model.layers.25.self_attn.q_proj.weight -> blk.25.attn_q.weight | F16 | [4096, 4096]
model.layers.25.self_attn.k_proj.weight -> blk.25.attn_k.weight | F16 | [4096, 4096]
model.layers.25.self_attn.v_proj.weight -> blk.25.attn_v.weight | F16 | [4096, 4096]
model.layers.25.self_attn.o_proj.weight -> blk.25.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.25.attn_rot_embd
model.layers.25.mlp.gate_proj.weight -> blk.25.ffn_gate.weight | F16 | [11008, 4096]
model.layers.25.mlp.up_proj.weight -> blk.25.ffn_up.weight | F16 | [11008, 4096]
model.layers.25.mlp.down_proj.weight -> blk.25.ffn_down.weight | F16 | [4096, 11008]
model.layers.25.input_layernorm.weight -> blk.25.attn_norm.weight | F16 | [4096]
model.layers.25.post_attention_layernorm.weight -> blk.25.ffn_norm.weight | F16 | [4096]
model.layers.26.self_attn.q_proj.weight -> blk.26.attn_q.weight | F16 | [4096, 4096]
model.layers.26.self_attn.k_proj.weight -> blk.26.attn_k.weight | F16 | [4096, 4096]
model.layers.26.self_attn.v_proj.weight -> blk.26.attn_v.weight | F16 | [4096, 4096]
model.layers.26.self_attn.o_proj.weight -> blk.26.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.26.attn_rot_embd
model.layers.26.mlp.gate_proj.weight -> blk.26.ffn_gate.weight | F16 | [11008, 4096]
model.layers.26.mlp.up_proj.weight -> blk.26.ffn_up.weight | F16 | [11008, 4096]
model.layers.26.mlp.down_proj.weight -> blk.26.ffn_down.weight | F16 | [4096, 11008]
model.layers.26.input_layernorm.weight -> blk.26.attn_norm.weight | F16 | [4096]
model.layers.26.post_attention_layernorm.weight -> blk.26.ffn_norm.weight | F16 | [4096]
model.layers.27.self_attn.q_proj.weight -> blk.27.attn_q.weight | F16 | [4096, 4096]
model.layers.27.self_attn.k_proj.weight -> blk.27.attn_k.weight | F16 | [4096, 4096]
model.layers.27.self_attn.v_proj.weight -> blk.27.attn_v.weight | F16 | [4096, 4096]
model.layers.27.self_attn.o_proj.weight -> blk.27.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.27.attn_rot_embd
model.layers.27.mlp.gate_proj.weight -> blk.27.ffn_gate.weight | F16 | [11008, 4096]
model.layers.27.mlp.up_proj.weight -> blk.27.ffn_up.weight | F16 | [11008, 4096]
model.layers.27.mlp.down_proj.weight -> blk.27.ffn_down.weight | F16 | [4096, 11008]
model.layers.27.input_layernorm.weight -> blk.27.attn_norm.weight | F16 | [4096]
model.layers.27.post_attention_layernorm.weight -> blk.27.ffn_norm.weight | F16 | [4096]
model.layers.28.self_attn.q_proj.weight -> blk.28.attn_q.weight | F16 | [4096, 4096]
model.layers.28.self_attn.k_proj.weight -> blk.28.attn_k.weight | F16 | [4096, 4096]
model.layers.28.self_attn.v_proj.weight -> blk.28.attn_v.weight | F16 | [4096, 4096]
model.layers.28.self_attn.o_proj.weight -> blk.28.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.28.attn_rot_embd
model.layers.28.mlp.gate_proj.weight -> blk.28.ffn_gate.weight | F16 | [11008, 4096]
model.layers.28.mlp.up_proj.weight -> blk.28.ffn_up.weight | F16 | [11008, 4096]
model.layers.28.mlp.down_proj.weight -> blk.28.ffn_down.weight | F16 | [4096, 11008]
model.layers.28.input_layernorm.weight -> blk.28.attn_norm.weight | F16 | [4096]
model.layers.28.post_attention_layernorm.weight -> blk.28.ffn_norm.weight | F16 | [4096]
model.layers.29.self_attn.q_proj.weight -> blk.29.attn_q.weight | F16 | [4096, 4096]
model.layers.29.self_attn.k_proj.weight -> blk.29.attn_k.weight | F16 | [4096, 4096]
model.layers.29.self_attn.v_proj.weight -> blk.29.attn_v.weight | F16 | [4096, 4096]
model.layers.29.self_attn.o_proj.weight -> blk.29.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.29.attn_rot_embd
model.layers.29.mlp.gate_proj.weight -> blk.29.ffn_gate.weight | F16 | [11008, 4096]
model.layers.29.mlp.up_proj.weight -> blk.29.ffn_up.weight | F16 | [11008, 4096]
model.layers.29.mlp.down_proj.weight -> blk.29.ffn_down.weight | F16 | [4096, 11008]
model.layers.29.input_layernorm.weight -> blk.29.attn_norm.weight | F16 | [4096]
model.layers.29.post_attention_layernorm.weight -> blk.29.ffn_norm.weight | F16 | [4096]
model.layers.30.self_attn.q_proj.weight -> blk.30.attn_q.weight | F16 | [4096, 4096]
model.layers.30.self_attn.k_proj.weight -> blk.30.attn_k.weight | F16 | [4096, 4096]
model.layers.30.self_attn.v_proj.weight -> blk.30.attn_v.weight | F16 | [4096, 4096]
model.layers.30.self_attn.o_proj.weight -> blk.30.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.30.attn_rot_embd
model.layers.30.mlp.gate_proj.weight -> blk.30.ffn_gate.weight | F16 | [11008, 4096]
model.layers.30.mlp.up_proj.weight -> blk.30.ffn_up.weight | F16 | [11008, 4096]
model.layers.30.mlp.down_proj.weight -> blk.30.ffn_down.weight | F16 | [4096, 11008]
model.layers.30.input_layernorm.weight -> blk.30.attn_norm.weight | F16 | [4096]
model.layers.30.post_attention_layernorm.weight -> blk.30.ffn_norm.weight | F16 | [4096]
model.layers.31.self_attn.q_proj.weight -> blk.31.attn_q.weight | F16 | [4096, 4096]
model.layers.31.self_attn.k_proj.weight -> blk.31.attn_k.weight | F16 | [4096, 4096]
model.layers.31.self_attn.v_proj.weight -> blk.31.attn_v.weight | F16 | [4096, 4096]
model.layers.31.self_attn.o_proj.weight -> blk.31.attn_output.weight | F16 | [4096, 4096]
skipping tensor blk.31.attn_rot_embd
model.layers.31.mlp.gate_proj.weight -> blk.31.ffn_gate.weight | F16 | [11008, 4096]
model.layers.31.mlp.up_proj.weight -> blk.31.ffn_up.weight | F16 | [11008, 4096]
model.layers.31.mlp.down_proj.weight -> blk.31.ffn_down.weight | F16 | [4096, 11008]
model.layers.31.input_layernorm.weight -> blk.31.attn_norm.weight | F16 | [4096]
model.layers.31.post_attention_layernorm.weight -> blk.31.ffn_norm.weight | F16 | [4096]
model.norm.weight -> output_norm.weight | F16 | [4096]
lm_head.weight -> output.weight | F16 | [32000, 4096]
Writing llama-2-7b-aiopsfinetunedv2-q8_0-gguf, format 7
Ignoring added_tokens.json since model matches vocab size without it.
gguf: This GGUF file is for Little Endian only
gguf: Setting special token type bos to 1
gguf: Setting special token type eos to 2
gguf: Setting special token type unk to 0
gguf: Setting special token type pad to 0
Traceback (most recent call last):
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 1483, in
main()
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 1477, in main
OutputFile.write_all(outfile, ftype, params, model, vocab, special_vocab,
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 1144, in write_all
for i, ((name, lazy_tensor), ndarray) in enumerate(zip(model.items(), ndarrays)):
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 918, in bounded_parallel_map
futures.append(executor.submit(func, next(iterable)))
^^^^^^^^^^^^^^
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 924, in bounded_parallel_map
result = futures.pop(0).result()
^^^^^^^^^^^^^^^^^^^^^^^
File "/usr/lib64/python3.12/concurrent/futures/_base.py", line 449, in result
return self.__get_result()
^^^^^^^^^^^^^^^^^^^
File "/usr/lib64/python3.12/concurrent/futures/_base.py", line 401, in __get_result
raise self._exception
File "/usr/lib64/python3.12/concurrent/futures/thread.py", line 58, in run
result = self.fn(*self.args, **self.kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 1101, in do_item
tensor = lazy_tensor.load().to_ggml()
^^^^^^^^^^^^^^^^^^
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 649, in load
ret = self._load()
^^^^^^^^^^^^
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 659, in load
return self.load().astype(data_type)
^^^^^^^^^^^
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 649, in load
ret = self._load()
^^^^^^^^^^^^
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 737, in load
return lazy_tensor.load().permute(n_head, n_head_kv)
^^^^^^^^^^^^^^^^^^
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 649, in load
ret = self._load()
^^^^^^^^^^^^
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 809, in load
return UnquantizedTensor(storage.load(storage_offset, elm_count).reshape(size))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "/home/user1/git/aiops/model_operations/llama.cpp/convert.py", line 796, in load
data = fp.read(size)
^^^^^^^^^^^^^
File "/usr/lib64/python3.12/zipfile/init.py", line 979, in read
data = self._read1(n)
^^^^^^^^^^^^^^
File "/usr/lib64/python3.12/zipfile/init.py", line 1069, in _read1
self._update_crc(data)
File "/usr/lib64/python3.12/zipfile/init.py", line 997, in _update_crc
raise BadZipFile("Bad CRC-32 for file %r" % self.name)
zipfile.BadZipFile: Bad CRC-32 for file 'pytorch_model-00001-of-00002/data/2'

It did start with the requirments.txt not installing. I had to replace the ~= with >= to make it install.
I am using a clean virtual python environement.

@JaCraig
Copy link

JaCraig commented Feb 22, 2024

It did start with the requirments.txt not installing. I had to replace the ~= with >= to make it install. I am using a clean virtual python environement.

This is due to ImpImporter being removed recently from pip (the Python thread where ImpImporter is deprecated). The zipfile issue also seems to be 3.12 specific. Switching to 3.10 in my environment fixed the issue for me.

@wdavo
Copy link

wdavo commented Mar 4, 2024

Thanks @JaCraig - Can confirm that switching from 3.12 to 3.10 fixed the issue for me as well.

@harshil21
Copy link

Facing the same issue. I don't want to downgrade my python just for this. I think llama.cpp should support Python 3.12. The latest pytorch version (2.2.1) now also supports Python 3.12 (pytorch/pytorch#110436 (comment)) so it shouldn't be a problem to support it.

@MarcoLv412
Copy link

Same issue here, on windows 10 PC:
Traceback (most recent call last):
File "D:\eosusa\cpp\llama.cpp\convert.py", line 1486, in
main()
File "D:\eosusa\cpp\llama.cpp\convert.py", line 1480, in main
OutputFile.write_all(outfile, ftype, params, model, vocab, special_vocab,
File "D:\eosusa\cpp\llama.cpp\convert.py", line 1162, in write_all
of.write_tensor_data(ftype, model, concurrency)
File "D:\eosusa\cpp\llama.cpp\convert.py", line 1093, in write_tensor_data
for i, ((name, lazy_tensor), ndarray) in enumerate(zip(model.items(), ndarrays)):
File "D:\eosusa\cpp\llama.cpp\convert.py", line 941, in bounded_parallel_map
result = futures.pop(0).result()
^^^^^^^^^^^^^^^^^^^^^^^
File "C:\ProgramData\miniconda3\Lib\concurrent\futures_base.py", line 449, in result
return self.__get_result()
^^^^^^^^^^^^^^^^^^^
File "C:\ProgramData\miniconda3\Lib\concurrent\futures_base.py", line 401, in __get_result
raise self._exception
File "C:\ProgramData\miniconda3\Lib\concurrent\futures\thread.py", line 58, in run
result = self.fn(*self.args, **self.kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\eosusa\cpp\llama.cpp\convert.py", line 1126, in do_item
tensor = lazy_tensor.load().to_ggml()
^^^^^^^^^^^^^^^^^^
File "D:\eosusa\cpp\llama.cpp\convert.py", line 666, in load
ret = self._load()
^^^^^^^^^^^^
File "D:\eosusa\cpp\llama.cpp\convert.py", line 676, in load
return self.load().astype(data_type)
^^^^^^^^^^^
File "D:\eosusa\cpp\llama.cpp\convert.py", line 666, in load
ret = self.load()
^^^^^^^^^^^^
File "D:\eosusa\cpp\llama.cpp\convert.py", line 826, in load
return UnquantizedTensor(storage.load(storage_offset, elm_count).reshape(size))
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "D:\eosusa\cpp\llama.cpp\convert.py", line 810, in load
fp = self.zip_file.open(info)
^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\ProgramData\miniconda3\Lib\zipfile_init
.py", line 1643, in open
raise BadZipFile(f"Overlapped entries: {zinfo.orig_filename!r} (possible zip bomb)")
zipfile.BadZipFile: Overlapped entries: 'consolidated.00/data/73' (possible zip bomb)

python --version
Python 3.12.2

All requriements.txt installed.

@github-actions github-actions bot added the stale label Apr 25, 2024
@ifoxhz
Copy link

ifoxhz commented Apr 28, 2024

the same issue on macos:
data = self._read1(n)
File "/usr/local/Cellar/python@3.9/3.9.13_2/Frameworks/Python.framework/Versions/3.9/lib/python3.9/zipfile.py", line 1014, in _read1
self._update_crc(data)
File "/usr/local/Cellar/python@3.9/3.9.13_2/Frameworks/Python.framework/Versions/3.9/lib/python3.9/zipfile.py", line 942, in _update_crc
raise BadZipFile("Bad CRC-32 for file %r" % self.name)
zipfile.BadZipFile: Bad CRC-32 for file 'pytorch_model-00001-of-00002/data/144'

llmama.cpp version:
commit 4e96a81 (HEAD -> master, origin/master, origin/HEAD)
Author: Anas Ahouzi 112881240+aahouzi@users.noreply.github.com
Date: Tue Apr 23 02:53:18 2024 +0200

python3.10

@github-actions github-actions bot removed the stale label Apr 29, 2024
@Leonidas-Li
Copy link

It did start with the requirments.txt not installing. I had to replace the ~= with >= to make it install. I am using a clean virtual python environement.

This is due to ImpImporter being removed recently from pip (the Python thread where ImpImporter is deprecated). The zipfile issue also seems to be 3.12 specific. Switching to 3.10 in my environment fixed the issue for me.

Thanks @JaCraig Solved my problem.

@QingtaoLi1
Copy link

QingtaoLi1 commented May 20, 2024

Same issue here, on windows 10 PC: Traceback (most recent call last): File "D:\eosusa\cpp\llama.cpp\convert.py", line 1486, in main() File "D:\eosusa\cpp\llama.cpp\convert.py", line 1480, in main OutputFile.write_all(outfile, ftype, params, model, vocab, special_vocab, File "D:\eosusa\cpp\llama.cpp\convert.py", line 1162, in write_all of.write_tensor_data(ftype, model, concurrency) File "D:\eosusa\cpp\llama.cpp\convert.py", line 1093, in write_tensor_data for i, ((name, lazy_tensor), ndarray) in enumerate(zip(model.items(), ndarrays)): File "D:\eosusa\cpp\llama.cpp\convert.py", line 941, in bounded_parallel_map result = futures.pop(0).result() ^^^^^^^^^^^^^^^^^^^^^^^ File "C:\ProgramData\miniconda3\Lib\concurrent\futures_base.py", line 449, in result return self.__get_result() ^^^^^^^^^^^^^^^^^^^ File "C:\ProgramData\miniconda3\Lib\concurrent\futures_base.py", line 401, in __get_result raise self._exception File "C:\ProgramData\miniconda3\Lib\concurrent\futures\thread.py", line 58, in run result = self.fn(*self.args, **self.kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 1126, in do_item tensor = lazy_tensor.load().to_ggml() ^^^^^^^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 666, in load ret = self.load() ^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 676, in load return self.load().astype(data_type) ^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 666, in load ret = self.load() ^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 826, in load return UnquantizedTensor(storage.load(storage_offset, elm_count).reshape(size)) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "D:\eosusa\cpp\llama.cpp\convert.py", line 810, in load fp = self.zip_file.open(info) ^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\ProgramData\miniconda3\Lib\zipfile__init.py", line 1643, in open raise BadZipFile(f"Overlapped entries: {zinfo.orig_filename!r} (possible zip bomb)") zipfile.BadZipFile: Overlapped entries: 'consolidated.00/data/73' (possible zip bomb)

python --version Python 3.12.2

All requriements.txt installed.

@MarcoLv412 Same error message here on Windows 11, python=3.8. Try adding "--concurrency=1" which works for me. Seems some problems in multi-thread zipfile reading.

@teleprint-me
Copy link
Contributor

teleprint-me commented May 20, 2024

@mofosyne This is a bug. It is python specific though. Add to docs for now?

Using safetensors is a valid workaround.

@github-actions github-actions bot added the stale label Jun 21, 2024
Copy link
Contributor

github-actions bot commented Jul 6, 2024

This issue was closed because it has been inactive for 14 days since being marked as stale.

@github-actions github-actions bot closed this as completed Jul 6, 2024
@LianaT
Copy link

LianaT commented Sep 1, 2024

Is this still an issue or has it been fixed because it's marked as closed? I'd rather stick with Python 3.12 which I already have on macos, instead of having to switch to python 3.10

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests