X-iZhang commited on
Commit
4717b60
·
verified ·
1 Parent(s): 01736c5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +16 -16
app.py CHANGED
@@ -61,24 +61,24 @@ def safe_load_pretrained_model(model_path, model_base=None, model_name=None, **k
61
  model_path, model_base, model_name, **kwargs
62
  )
63
 
64
- # # Fix tokenizer pad_token_id if it's None (common issue with Llama 3 models)
65
- # if tokenizer.pad_token_id is None:
66
- # if tokenizer.eos_token_id is not None:
67
- # # tokenizer.pad_token_id = tokenizer.eos_token_id
68
- # tokenizer.pad_token_id = 1
69
- # print(f'[INFO] Set pad_token_id to eos_token_id: {tokenizer.eos_token_id}')
70
- # else:
71
- # tokenizer.pad_token_id = 0
72
- # print('[INFO] Set pad_token_id to 0 (default)')
73
-
74
- # Also ensure pad_token is set (Llama 3 specific) - CRITICAL for output
75
- if tokenizer.pad_token is None:
76
- if tokenizer.eos_token is not None:
77
  tokenizer.pad_token_id = 128001
78
- print(f'[INFO] Set pad_token to eos_token: {tokenizer.pad_token_id}')
79
  else:
80
- tokenizer.add_special_tokens({'pad_token': '[PAD]'})
81
- print('[INFO] Added [PAD] token to tokenizer')
 
 
 
 
 
 
 
 
 
82
 
83
  # Set padding_side to left for Llama 3 (prevents empty generation)
84
  if hasattr(tokenizer, 'padding_side'):
 
61
  model_path, model_base, model_name, **kwargs
62
  )
63
 
64
+ # Fix tokenizer pad_token_id if it's None (common issue with Llama 3 models)
65
+ if tokenizer.pad_token_id is None:
66
+ if tokenizer.eos_token_id is not None:
67
+ # tokenizer.pad_token_id = tokenizer.eos_token_id
 
 
 
 
 
 
 
 
 
68
  tokenizer.pad_token_id = 128001
69
+ print(f'[INFO] Set pad_token_id to eos_token_id: {tokenizer.pad_token_id}')
70
  else:
71
+ tokenizer.pad_token_id = 0
72
+ print('[INFO] Set pad_token_id to 0 (default)')
73
+
74
+ # # Also ensure pad_token is set (Llama 3 specific) - CRITICAL for output
75
+ # if tokenizer.pad_token is None:
76
+ # if tokenizer.eos_token is not None:
77
+ # tokenizer.pad_token_id = tokenizer.eos_token
78
+ # print(f'[INFO] Set pad_token to eos_token: {tokenizer.pad_token_id}')
79
+ # else:
80
+ # tokenizer.add_special_tokens({'pad_token': '[PAD]'})
81
+ # print('[INFO] Added [PAD] token to tokenizer')
82
 
83
  # Set padding_side to left for Llama 3 (prevents empty generation)
84
  if hasattr(tokenizer, 'padding_side'):