Este contenido no está disponible en el idioma seleccionado.
Chapter 1. Troubleshooting OpenShift Lightspeed
Review solutions and workarounds for common installation, configuration, and operational issues encountered with OpenShift Lightspeed.
1.1. 502 Bad Gateway errors in the interface Copiar enlaceEnlace copiado en el portapapeles!
Wait for service pods to finish starting to avoid 502 Bad Gateway errors.
Wait a few minutes after deploying OpenShift Lightspeed and OpenShift Container Platform before trying the interface again.
1.2. Operator missing from the OperatorHub list Copiar enlaceEnlace copiado en el portapapeles!
The OperatorHub displays the OpenShift Lightspeed Operator only for supported architectures. Filtering prevents the Operator from appearing on anything other than the x86_64 architecture.
1.3. Reasoning model generates delineator prompt Copiar enlaceEnlace copiado en el portapapeles!
Reasoning models use tags such as THOUGHT or reasoning to separate inner logic from the final answer.
OpenShift Lightspeed does not control these tags or add them to the output. This feature is part of the model itself.
Procedure
You can turn off these tags using one of the following methods:
-
Add a keyword to your prompt if the model supports it, such as
/nothink. Check your model documentation for the specific keyword. - Disable the delineator feature in the inference server configuration settings. For more information, see the documentation for the inference server or for the model you are using.
1.4. Troubleshoot API authentication failures Copiar enlaceEnlace copiado en el portapapeles!
Use the status codes and error details to identify and resolve common authentication failures when connecting to the OpenShift Lightspeed API.
| Status code | Description | Example detail |
|---|---|---|
| 401 Unauthorized |
The |
|
| 403 Forbidden |
The token is invalid, expired, or the user lacks RBAC permissions for the |
|
| 500 Internal Error |
An unexpected error occurred through the Kubernetes |
|
1.5. Resolving prompt is too long errors Copiar enlaceEnlace copiado en el portapapeles!
To resolve the Prompt is too long error, adjust the model parameters or reduce the input length for the query to fit within the supported context window.
This error occurs when the total number of tokens (the input query, RAG context, and expected response) exceeds the model context window.
Procedure
- Verify that you have set the context window value correctly for your specific model and provider.
- Set a lower value for the maximum response tokens parameter to allow more space for the input query and context.
- Shorten the query or reduce the size of any attached files.
1.6. Resolving truncated responses Copiar enlaceEnlace copiado en el portapapeles!
To resolve truncated or incomplete model responses, increase the response token limit or use follow-up prompts to retrieve the remaining content.
This issue occurs when the model reaches its pre-configured response token limit.
Procedure
- Verify that the model supports a higher response token limit.
- Increase the token limit value in the OpenShift Lightspeed configuration.
-
If the response is still cut off, type
continueas a follow-up query to prompt the model to provide the remaining text.
Set the response token value in reasonable proportion to the context window value. Setting this value too high reserves tokens and might limit the size of your input query.
1.7. Resolving issues with conversation history Copiar enlaceEnlace copiado en el portapapeles!
To maintain conversation history, optimize the balance between the context window and response tokens. This prevents earlier dialogue from being truncated when the model reaches its context limit.
Procedure
- Verify the context window is correctly set for your specific model and provider.
- Lower the max response tokens value to increase the remaining space available for conversation history.