Spring Sale Special Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: xmas50

Databricks Databricks-Generative-AI-Engineer-Associate - Databricks Certified Generative AI Engineer Associate

After changing the response generating LLM in a RAG pipeline from GPT-4 to a model with a shorter context length that the company self-hosts, the Generative AI Engineer is getting the following error:

What TWO solutions should the Generative AI Engineer implement without changing the response generating model? (Choose two.)

A.

Use a smaller embedding model to generate

B.

Reduce the maximum output tokens of the new model

C.

Decrease the chunk size of embedded documents

D.

Reduce the number of records retrieved from the vector database

E.

Retrain the response generating model using ALiBi

Which indicator should be considered to evaluate the safety of the LLM outputs when qualitatively assessing LLM responses for a translation use case?

A.

The ability to generate responses in code

B.

The similarity to the previous language

C.

The latency of the response and the length of text generated

D.

The accuracy and relevance of the responses

A Generative AI Engineer I using the code below to test setting up a vector store:

Assuming they intend to use Databricks managed embeddings with the default embedding model, what should be the next logical function call?

A.

vsc.get_index()

B.

vsc.create_delta_sync_index()

C.

vsc.create_direct_access_index()

D.

vsc.similarity_search()

When developing an LLM application, it’s crucial to ensure that the data used for training the model complies with licensing requirements to avoid legal risks.

Which action is NOT appropriate to avoid legal risks?

A.

Reach out to the data curators directly before you have started using the trained model to let them know.

B.

Use any available data you personally created which is completely original and you can decide what license to use.

C.

Only use data explicitly labeled with an open license and ensure the license terms are followed.

D.

Reach out to the data curators directly after you have started using the trained model to let them know.

A Generative AI Engineer is experimenting with using parameters to configure an agent in Mosaic Agent Framework. However, they are struggling to get the agent to respond with relevant information with this configuration:

config = {"prompt_template": "You are a trivia bot. Generate a question based on the user's input: {user_input}", "input_vars": ["user_input"], "parameters": {"temperature": 0.01, "max_tokens": 500}}

Which error is causing the problem?

A.

The prompt does not parse the user's input vars

B.

The prompt does not set the retriever schema

C.

The prompt does not list available agents for the LLM to call

D.

The prompt is not wrapped in ChatModel

A Generative Al Engineer interfaces with an LLM with prompt/response behavior that has been trained on customer calls inquiring about product availability. The LLM is designed to output “In Stock” if the product is available or only the term “Out of Stock” if not.

Which prompt will work to allow the engineer to respond to call classification labels correctly?

A.

Respond with “In Stock” if the customer asks for a product.

B.

You will be given a customer call transcript where the customer asks about product availability. The outputs are either “In Stock” or “Out of Stock”. Format the output in JSON, for example: {“call_id”: “123”, “label”: “In Stock”}.

C.

Respond with “Out of Stock” if the customer asks for a product.

D.

You will be given a customer call transcript where the customer inquires about product availability. Respond with “In Stock” if the product is available or “Out of Stock” if not.

A Generative Al Engineer is building a system which will answer questions on latest stock news articles.

Which will NOT help with ensuring the outputs are relevant to financial news?

A.

Implement a comprehensive guardrail framework that includes policies for content filters tailored to the finance sector.

B.

Increase the compute to improve processing speed of questions to allow greater relevancy analysis

C Implement a profanity filter to screen out offensive language

C.

Incorporate manual reviews to correct any problematic outputs prior to sending to the users

A Generative Al Engineer wants their (inetuned LLMs in their prod Databncks workspace available for testing in their dev workspace as well. All of their workspaces are Unity Catalog enabled and they are currently logging their models into the Model Registry in MLflow.

What is the most cost-effective and secure option for the Generative Al Engineer to accomplish their gAi?

A.

Use an external model registry which can be accessed from all workspaces

B.

Setup a script to export the model from prod and import it to dev.

C.

Setup a duplicate training pipeline in dev, so that an identical model is available in dev.

D.

Use MLflow to log the model directly into Unity Catalog, and enable READ access in the dev workspace to the model.

A Generative AI Engineer is designing an LLM-powered live sports commentary platform. The platform provides real-time updates and LLM-generated analyses for any users who would like to have live summaries, rather than reading a series of potentially outdated news articles.

Which tool below will give the platform access to real-time data for generating game analyses based on the latest game scores?

A.

DatabrickslQ

B.

Foundation Model APIs

C.

Feature Serving

D.

AutoML

A Generative AI Engineer has a provisioned throughput model serving endpoint as part of a RAG application and would like to monitor the serving endpoint’s incoming requests and outgoing responses. The current approach is to include a micro-service in between the endpoint and the user interface to write logs to a remote server.

Which Databricks feature should they use instead which will perform the same task?

A.

Vector Search

B.

Lakeview

C.

DBSQL

D.

Inference Tables