In the following cell,
bigger_finetuned_output = bigger_finetuned_model(test_question)
bigger_finetuned_model = BasicModelRunner(model_name_to_id["bigger_model_name"])
bigger_finetuned_output = bigger_finetuned_model(test_question)
print("Bigger (2.8B) finetuned model (test): ", bigger_finetuned_output)
I’m getting the following error. And I checked that model_name_to_id
is {'bigger_model_name': '06ad41e68cd839fb475a0c1a4ee7a3ad398228df01c9396a97788295d5a0f8bb'}
status code: 400
---------------------------------------------------------------------------
HTTPError Traceback (most recent call last)
File /usr/local/lib/python3.9/site-packages/llama/engine/lamini.py:253, in Lamini.make_web_request(self, url, http_method, json)
252 try:
--> 253 resp.raise_for_status()
254 except requests.exceptions.HTTPError as e:
File /usr/local/lib/python3.9/site-packages/requests/models.py:1021, in Response.raise_for_status(self)
1020 if http_error_msg:
-> 1021 raise HTTPError(http_error_msg, response=self)
HTTPError: 400 Client Error: Bad Request for url: http://internal-aws-prod-internal-revproxy-alb-11660607.us-west-1.elb.amazonaws.com/rev-proxy/lamini/v2/lamini/completions
During handling of the above exception, another exception occurred:
UserError Traceback (most recent call last)
Cell In[24], line 2
1 bigger_finetuned_model = BasicModelRunner(model_name_to_id["bigger_model_name"])
----> 2 bigger_finetuned_output = bigger_finetuned_model(test_question)
3 print("Bigger (2.8B) finetuned model (test): ", bigger_finetuned_output)
File /usr/local/lib/python3.9/site-packages/llama/runners/basic_model_runner.py:52, in BasicModelRunner.__call__(self, inputs)
49 else:
50 # Singleton
51 input_objects = Input(input=inputs)
---> 52 output_objects = self.llm(
53 input=input_objects,
54 output_type=Output,
55 model_name=self.model_name,
56 enable_peft=self.enable_peft,
57 )
58 if isinstance(output_objects, list):
59 outputs = [o.output for o in output_objects]
File /usr/local/lib/python3.9/site-packages/llama/engine/typed_lamini.py:13, in TypedLamini.__call__(self, *args, **kwargs)
12 def __call__(self, *args, **kwargs):
---> 13 result = super().__call__(*args, **kwargs)
14 if isinstance(result, list):
15 if "output_type" in kwargs:
File /usr/local/lib/python3.9/site-packages/llama/engine/lamini.py:81, in Lamini.__call__(self, input, output_type, stop_tokens, model_name, enable_peft, random, max_tokens)
69 req_data = self.make_llm_req_map(
70 self.id,
71 model_name or self.model_name,
(...)
78 max_tokens,
79 )
80 url = self.api_prefix + "completions"
---> 81 return self.make_web_request(url, "post", req_data)
File /usr/local/lib/python3.9/site-packages/llama/engine/lamini.py:281, in Lamini.make_web_request(self, url, http_method, json)
279 except Exception:
280 json_response = {}
--> 281 raise UserError(json_response.get("detail", "UserError"))
282 if resp.status_code == 503:
283 try:
UserError: Currently this user has support for base models: ['hf-internal-testing/tiny-random-gpt2', 'EleutherAI/pythia-70m', 'EleutherAI/pythia-70m-deduped', 'EleutherAI/pythia-70m-v0', 'EleutherAI/pythia-70m-deduped-v0', 'EleutherAI/neox-ckpt-pythia-70m-deduped-v0', 'EleutherAI/neox-ckpt-pythia-70m-v1', 'EleutherAI/neox-ckpt-pythia-70m-deduped-v1', 'EleutherAI/gpt-neo-125m', 'EleutherAI/pythia-160m', 'EleutherAI/pythia-160m-deduped', 'EleutherAI/pythia-160m-deduped-v0', 'EleutherAI/neox-ckpt-pythia-70m', 'EleutherAI/neox-ckpt-pythia-160m', 'EleutherAI/neox-ckpt-pythia-160m-deduped-v1', 'EleutherAI/pythia-410m', 'EleutherAI/pythia-410m-v0', 'EleutherAI/pythia-410m-deduped', 'EleutherAI/pythia-410m-deduped-v0', 'EleutherAI/neox-ckpt-pythia-410m', 'EleutherAI/neox-ckpt-pythia-410m-deduped-v1', 'cerebras/Cerebras-GPT-111M', 'cerebras/Cerebras-GPT-256M', 'meta-llama/Llama-2-7b-hf', 'meta-llama/Llama-2-7b-chat-hf', 'meta-llama/Llama-2-13b-chat-hf', 'meta-llama/Llama-2-70b-chat-hf'