|
27 | 27 | from QEfficient.utils.run_utils import ApiRunnerInternVL, ApiRunnerVlm
|
28 | 28 | from QEfficient.utils.test_utils import InternProcessor
|
29 | 29 |
|
30 |
| -HF_TOKEN = "" |
31 | 30 | NEW_GENERATION_TOKENS = 10
|
32 | 31 | test_models_config = [
|
33 | 32 | # CONFIG PARAMS NEEDED FOR A MODEL TO BE TESTED
|
|
104 | 103 | def load_image_text_to_text_model(model_config):
|
105 | 104 | model_path = hf_download(
|
106 | 105 | repo_id=model_config._name_or_path,
|
107 |
| - hf_token=HF_TOKEN, |
108 | 106 | ignore_patterns=["*.onnx", "*.ot", "*.md", "*.tflite", "*.pdf", "*.h5", "*.msgpack"],
|
109 | 107 | )
|
110 | 108 | try:
|
111 | 109 | model_hf = AutoModelForImageTextToText.from_pretrained(
|
112 | 110 | model_path,
|
113 | 111 | low_cpu_mem_usage=False,
|
114 |
| - token=HF_TOKEN, |
115 | 112 | config=model_config,
|
116 | 113 | )
|
117 | 114 | except ValueError:
|
118 | 115 | model_hf = AutoModelForCausalLM.from_pretrained(
|
119 | 116 | model_path,
|
120 | 117 | low_cpu_mem_usage=False,
|
121 |
| - token=HF_TOKEN, |
122 | 118 | trust_remote_code=True,
|
123 | 119 | config=model_config,
|
124 | 120 | )
|
@@ -160,9 +156,7 @@ def check_image_text_to_text_pytorch_vs_kv_vs_ort_vs_ai100(
|
160 | 156 | ):
|
161 | 157 | model_config = {"model_name": model_name}
|
162 | 158 | model_config["img_size"] = img_size
|
163 |
| - config = AutoConfig.from_pretrained( |
164 |
| - model_config["model_name"], token=HF_TOKEN, trust_remote_code=True, padding=True |
165 |
| - ) |
| 159 | + config = AutoConfig.from_pretrained(model_config["model_name"], trust_remote_code=True, padding=True) |
166 | 160 | config = set_num_layers(config, n_layer=n_layer)
|
167 | 161 | model_hf, _ = load_image_text_to_text_model(config)
|
168 | 162 | processor = AutoProcessor.from_pretrained(model_name, trust_remote_code=True, padding=True)
|
@@ -199,7 +193,6 @@ def check_image_text_to_text_pytorch_vs_kv_vs_ort_vs_ai100(
|
199 | 193 | model_config["model_name"],
|
200 | 194 | kv_offload=kv_offload,
|
201 | 195 | config=config,
|
202 |
| - token=HF_TOKEN, |
203 | 196 | )
|
204 | 197 |
|
205 | 198 | # pytorch_kv_tokens = api_runner.run_vlm_kv_model_on_pytorch(qeff_model.model)
|
@@ -284,7 +277,6 @@ def check_intern_image_text_to_text_pytorch_vs_kv_vs_ort_vs_ai100(
|
284 | 277 | model_config["model_name"],
|
285 | 278 | kv_offload=kv_offload,
|
286 | 279 | config=config,
|
287 |
| - token=HF_TOKEN, |
288 | 280 | )
|
289 | 281 | # pytorch_kv_tokens = api_runner.run_vlm_kv_model_on_pytorch(qeff_model.model)
|
290 | 282 | # assert (pytorch_hf_tokens == pytorch_kv_tokens).all(), (
|
|
0 commit comments