Thanks to the team for actively reviewing and responding to issues raised in this forum.
I am currently facing multiple issues while working on Assignment #5 and this is blocking my progress, requesting support to review and fix them soon
Issue #1 – Unable to access the model (Blocking)
Description:
When invoking the function generate_with_single_input, the model fails to load.
Error Message:
InvalidRequestError: Error code: 400 - {
"message": "Unable to access non-serverless model meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo.
Please visit https://api.together.ai/models/meta-llama/Meta-Llama-3.1-8B-Instruct-Turbo
to create and start a new dedicated endpoint for the model.",
"type_": "invalid_request_error",
"code": "model_not_available"
}
Steps to Reproduce:
-
Run the function
generate_with_single_input -
Model invocation fails with the error above.
Current Status: Unresolved – blocking assignment progress for the last 2 days.
Issue #2 – Possible model response format mismatch (Need confirmation)
I am not sure if this is an actual issue and would appreciate confirmation.
It appears that the model being used internally is Qwen/Qwen3.5-9B, which seems to return its response in the reasoning field rather than the content field.
However, in the function:
check_if_faq_or_product(query, simplified=False)
the validation logic checks the content field (expecting values like FAQ or Product). Since content is blank when using Qwen/Qwen3.5-9B, the validation fails.
**Current Status:**Unresolved and possibly related to the model access issue above.
Issue #3 – Unable to load get_phoenix_endpoint during library import
Description: While importing the required libraries, the function get_phoenix_endpoint could not be loaded.
Workaround Applied: I temporarily resolved this by adding get_phoenix_endpoint inside utils.py.
Current Status: Temporarily resolved.

