05_Training_lab_student status code 400

In the following cell,

bigger_finetuned_output = bigger_finetuned_model(test_question)

bigger_finetuned_model = BasicModelRunner(model_name_to_id["bigger_model_name"])
bigger_finetuned_output = bigger_finetuned_model(test_question)
print("Bigger (2.8B) finetuned model (test): ", bigger_finetuned_output)

I’m getting the following error. And I checked that model_name_to_id is {'bigger_model_name': '06ad41e68cd839fb475a0c1a4ee7a3ad398228df01c9396a97788295d5a0f8bb'}

status code: 400
---------------------------------------------------------------------------
HTTPError                                 Traceback (most recent call last)
File /usr/local/lib/python3.9/site-packages/llama/engine/lamini.py:253, in Lamini.make_web_request(self, url, http_method, json)
    252 try:
--> 253     resp.raise_for_status()
    254 except requests.exceptions.HTTPError as e:

File /usr/local/lib/python3.9/site-packages/requests/models.py:1021, in Response.raise_for_status(self)
   1020 if http_error_msg:
-> 1021     raise HTTPError(http_error_msg, response=self)

HTTPError: 400 Client Error: Bad Request for url: http://internal-aws-prod-internal-revproxy-alb-11660607.us-west-1.elb.amazonaws.com/rev-proxy/lamini/v2/lamini/completions

During handling of the above exception, another exception occurred:

UserError                                 Traceback (most recent call last)
Cell In[24], line 2
      1 bigger_finetuned_model = BasicModelRunner(model_name_to_id["bigger_model_name"])
----> 2 bigger_finetuned_output = bigger_finetuned_model(test_question)
      3 print("Bigger (2.8B) finetuned model (test): ", bigger_finetuned_output)

File /usr/local/lib/python3.9/site-packages/llama/runners/basic_model_runner.py:52, in BasicModelRunner.__call__(self, inputs)
     49 else:
     50     # Singleton
     51     input_objects = Input(input=inputs)
---> 52 output_objects = self.llm(
     53     input=input_objects,
     54     output_type=Output,
     55     model_name=self.model_name,
     56     enable_peft=self.enable_peft,
     57 )
     58 if isinstance(output_objects, list):
     59     outputs = [o.output for o in output_objects]

File /usr/local/lib/python3.9/site-packages/llama/engine/typed_lamini.py:13, in TypedLamini.__call__(self, *args, **kwargs)
     12 def __call__(self, *args, **kwargs):
---> 13     result = super().__call__(*args, **kwargs)
     14     if isinstance(result, list):
     15         if "output_type" in kwargs:

File /usr/local/lib/python3.9/site-packages/llama/engine/lamini.py:81, in Lamini.__call__(self, input, output_type, stop_tokens, model_name, enable_peft, random, max_tokens)
     69 req_data = self.make_llm_req_map(
     70     self.id,
     71     model_name or self.model_name,
   (...)
     78     max_tokens,
     79 )
     80 url = self.api_prefix + "completions"
---> 81 return self.make_web_request(url, "post", req_data)

File /usr/local/lib/python3.9/site-packages/llama/engine/lamini.py:281, in Lamini.make_web_request(self, url, http_method, json)
    279     except Exception:
    280         json_response = {}
--> 281     raise UserError(json_response.get("detail", "UserError"))
    282 if resp.status_code == 503:
    283     try:

UserError: Currently this user has support for base models: ['hf-internal-testing/tiny-random-gpt2', 'EleutherAI/pythia-70m', 'EleutherAI/pythia-70m-deduped', 'EleutherAI/pythia-70m-v0', 'EleutherAI/pythia-70m-deduped-v0', 'EleutherAI/neox-ckpt-pythia-70m-deduped-v0', 'EleutherAI/neox-ckpt-pythia-70m-v1', 'EleutherAI/neox-ckpt-pythia-70m-deduped-v1', 'EleutherAI/gpt-neo-125m', 'EleutherAI/pythia-160m', 'EleutherAI/pythia-160m-deduped', 'EleutherAI/pythia-160m-deduped-v0', 'EleutherAI/neox-ckpt-pythia-70m', 'EleutherAI/neox-ckpt-pythia-160m', 'EleutherAI/neox-ckpt-pythia-160m-deduped-v1', 'EleutherAI/pythia-410m', 'EleutherAI/pythia-410m-v0', 'EleutherAI/pythia-410m-deduped', 'EleutherAI/pythia-410m-deduped-v0', 'EleutherAI/neox-ckpt-pythia-410m', 'EleutherAI/neox-ckpt-pythia-410m-deduped-v1', 'cerebras/Cerebras-GPT-111M', 'cerebras/Cerebras-GPT-256M', 'meta-llama/Llama-2-7b-hf', 'meta-llama/Llama-2-7b-chat-hf', 'meta-llama/Llama-2-13b-chat-hf', 'meta-llama/Llama-2-70b-chat-hf'

Hi @Jiaquan_He,

Thanks for letting us know. I have forwarded this along.

Best,
Mubsi

Hi @Jiaquan_He,

This should be fixed now.

Best,
Mubsi