0% found this document useful (0 votes)
42 views5 pages

OCI GEN AI Test 2

The document contains a series of questions and answers related to the OCI Generative AI service, focusing on techniques like Retrieval-Augmented Generation, T-Few fine-tuning, and various parameters in model generation. It covers topics such as vector normalization, model storage, and the importance of certain parameters in the generation process. The document also highlights correct and incorrect responses to the questions, indicating areas of knowledge and misunderstanding.

Uploaded by

jimmyy.jain2
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
42 views5 pages

OCI GEN AI Test 2

The document contains a series of questions and answers related to the OCI Generative AI service, focusing on techniques like Retrieval-Augmented Generation, T-Few fine-tuning, and various parameters in model generation. It covers topics such as vector normalization, model storage, and the importance of certain parameters in the generation process. The document also highlights correct and incorrect responses to the questions, indicating areas of knowledge and misunderstanding.

Uploaded by

jimmyy.jain2
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 5

Question 1Incorrect

☁️How does the Retrieval-Augmented Generation (RAG) Token technique differ from RAG Sequence
when generating a model’s response?

RAG Token retrieves relevant documents for each part of the response and constructs the answer
incrementally.

RAG Token does not use document retrieval but generates responses based on pre-existing
knowledge only.

RAG Token retrieves documents only at the beginning of the response generation and uses those for
the entire content.

Unlike RAG Sequence, RAG Token generates the entire response at once without considering
individual parts.

Question 2Incorrect

☁️Which is a key advantage of using T-Few over Vanilla fine-tuning in the OCI Generative AI service?

Faster training time and lower cost

Enhanced generalization to unseen data

Reduced model complexity

Increased model interpretability

Question 3Incorrect

☁️How do Dot Product and Cosine Distance differ in their application to comparing text embeddings
in natural language processing?

Dot Product is used for semantic analysis, whereas Cosine Distance is used for syntactic comparisons.

Dot Product assesses the overall similarity in content, whereas Cosine Distance measures topical
relevance.

Dot Product calculates the literal overlap of words, whereas Cosine Distance evaluates the stylistic
similarity.

Dot Product measures the magnitude and direction of vectors, whereas Cosine Distance focuses on
the orientation regardless of magnitude.

Question 4Correct

☁️Why is normalization of vectors important before indexing in a hybrid search system?

It standardizes vector lengths for meaningful comparison using metrics such as Cosine Similarity.

It converts all sparse vectors to dense vectors.

It ensures that all vectors represent keywords only.

It significantly reduces the size of the database.


Question 5Correct

☁️Which is NOT a category of pretrained foundational models available in the OCI Generative AI
service?

Generation models

Translation models

Embedding models

Summarization models

Question 6Correct

☁️Which is a key characteristic of the annotation process used in T-Few fine-tuning?

T-Few fine-tuning uses annotated data to adjust a fraction of model weights.

T-Few fine-tuning requires manual annotation of input-output pairs.

T-Few fine-tuning involves updating the weights of all layers in the model.

T-Few fine-tuning relies on unsupervised learning techniques for annotation.

Question 7Correct

☁️Which statement is true about the "Top p" parameter of the OCI Generative AI Generation
models?

"Top p" assigns penalties to frequently occurring tokens.

"Top p" determines the maximum number of tokens per response.

"Top p" limits token selection based on the sum of their probabilities.

"Top p" selects tokens from the "Top k" tokens sorted by probability.

Question 8Correct

☁️How are fine-tuned customer models stored to enable strong data privacy and security in the OCI
Generative AI service?

Stored in an unencrypted form in Object Storage

Stored in Key Management service

Stored in Object Storage encrypted by default

Shared among multiple customers for efficiency


Question 9Correct

☁️What is the purpose of the "stop sequence" parameter in the OCI Generative AI Generation
models?

It assigns a penalty to frequently occurring tokens to reduce repetitive text.

It specifies a string that tells the model to stop generating more content.

It determines the maximum number of tokens the model can generate per response.

It controls the randomness of the model’s output, affecting its creativity.

Question 10Correct

☁️What does a higher number assigned to a token signify in the "Show Likelihoods" feature of the
language model token generation?

The token is less likely to follow the current token.

The token will be the only one considered in the next generation step.

The token is more likely to follow the current token.

The token is unrelated to the current token and will not be used.

Question 11Correct

☁️Which Oracle Accelerated Data Science (ADS) class can be used to deploy a Large Language Model
(LLM) application to OCI Data Science model deployment?

GenerativeAI

TextLoader

RetrievalQA

ChainDeployment

Question 12Correct

☁️Which role does a "model endpoint" serve in the inference workflow of the OCI Generative AI
service?

Hosts the training data for fine-tuning custom models

Updates the weights of the base model during the fine-tuning process

Evaluates the performance metrics of the custom models

Serves as a designated point for user requests and model responses


Question 13Correct

☁️What issue might arise from using small data sets with the Vanilla fine-tuning method in the OCI
Generative AI service?

Data Leakage

Overfitting

Underfitting

Model Drift

Question 14Incorrect

☁️Given the following code:

prompt - PromptTomplate (Input_variables=( "human input ", "city"], template=template)

Which statement is true about PromtTemplate in relation to input_variables?

PromptTemplate can support only a single variable at a time.

PromptTemplate is unable to use any variables.

PromptTemplate supports any number of variables, including the possibility of having none.

PromptTemplate requires a minimum of two variables to function properly.

Question 15Incorrect

☁️In LangChain, which retriever search type is used to balance between relevancy and diversity?

mmr

similarity_score_threshold

similarity

top k

Question 16Correct

☁️How does the architecture of dedicated AI clusters contribute to minimizing GPU memory
overhead for T-Few fine-tuned model inference?

By optimizing GPU memory utilization for each model’s unique parameters

By loading the entire model into GPU memory for efficient processing

By allocating separate GPUs for each model instance

By sharing base model weights across multiple fine-tuned models on the same group of GPUs

Question 17Correct
☁️Given the following code:

chain - prompt | 11m

Which statement is true about LangChain Expression Language (LCEL)?

LCEL is a declarative and preferred way to compose chains together.

LCEL is an older Python library for building Large Language Models.

LCEL is a programming language used to write documentation for LangChain.

LCEL is a legacy method for creating chains in LangChain.

Question 18Correct

☁️What is the purpose of the "stop sequence" parameter in the OCI Generative AI Generation
models?

It determines the maximum number of tokens the model can generate per response.

It specifies a string that tells the model to stop generating more content.

It assigns a penalty to frequently occurring tokens to reduce repetitive text.

It controls the randomness of the model’s output, affecting its creativity.

Question 19Correct

☁️Which is NOT a built-in memory type in LangChain?

ConvorsationImageMemory

ConversationTokenBufferMemory

ConvorsationBufferMemory

ConversationSummaryMemory

Question 20Correct

☁️Which technique involves prompting the Large Language Model (LLM) to emit intermediate
reasoning steps as part of its response?

Step-Back Prompting

In-context Learning

Least-to-most Prompting

Chain-of-Thought

You might also like