Video upload date:  · Duration: PT1H46M27S  · Language: EN

After changing the response-generating LLM databricks video

generative-ai video for after changing the response-generating LLM in a RAG pipeline from GPT-4 to a model with a shorter context length that the company

This is a dedicated watch page for a single video.

Full Certification Question

After changing the response-generating LLM in a RAG pipeline from GPT-4 to a model with a shorter context length that the company self-hosts, the Generative AI Engineer is getting the following error: {"error_code": "BAD_REQUEST", "message": "Bad request: rpc error: code = InvalidArgument desc = prompt token count (4595) cannot exceed 4096..."} What TWO solutions should the Generative AI Engineer implement without changing the response-generating model? (Choose two.)