
Answer-first summary for fast verification
Answer: Provide labeled data with the prompt field and the completion field.
## Detailed Explanation Fine-tuning a foundation model in Amazon Bedrock requires providing **labeled training data** where each example consists of a **prompt** (the input to the model) and a **completion** (the desired output). This structured approach allows the model to learn specific patterns and behaviors tailored to the company's proprietary data and use case. ### Why Option A is Correct: - **Labeled Data Structure**: Fine-tuning fundamentally requires supervised learning with labeled examples. The prompt-completion format is the standard structure for fine-tuning large language models, enabling the model to adjust its parameters based on the relationship between inputs and desired outputs. - **Amazon Bedrock Requirements**: When fine-tuning models through Amazon Bedrock, the training dataset must be formatted with prompt and completion fields to align with the service's fine-tuning capabilities. - **Domain Adaptation**: By providing company-specific labeled data, the model learns to generate more accurate responses for the AI search tool's particular domain and requirements. ### Why Other Options Are Incorrect: **Option B**: Preparing a .txt file with multiple lines in .csv format is not the appropriate data format for fine-tuning in Amazon Bedrock. While CSV files can be used for data storage, the specific prompt-completion structure (typically in JSONL format) is required for effective fine-tuning. **Option C**: Purchasing Provisioned Throughput is not necessary for the fine-tuning process itself. Provisioned Throughput is required for **inference** with custom models after fine-tuning, but fine-tuning can be performed without it. This is a common point of confusion where deployment requirements are mistaken for training requirements. **Option D**: Training the model on journals and textbooks represents pre-training or continued pre-training, not fine-tuning. Fine-tuning specifically refers to adapting a pre-trained foundation model with targeted, labeled data for a specific task or domain, not general domain expansion through unstructured text. ### Best Practices Consideration: For successful fine-tuning in Amazon Bedrock: 1. Ensure high-quality, representative labeled data that reflects real-world use cases 2. Maintain proper data formatting with clear prompt-completion pairs 3. Consider data volume requirements based on the specific foundation model being fine-tuned 4. Validate data quality to avoid reinforcing biases or incorrect patterns This approach aligns with AWS best practices for customizing foundation models to achieve higher accuracy for specific business applications.
Ultimate access to all questions.
No comments yet.
Author: LeetQuiz Editorial Team
A company is using a foundation model from Amazon Bedrock for an AI search tool and wants to improve its accuracy by fine-tuning it with the company's proprietary data. Which approach will enable the successful fine-tuning of the model?
A
Provide labeled data with the prompt field and the completion field.
B
Prepare the training dataset by creating a .txt file that contains multiple lines in .csv format.
C
Purchase Provisioned Throughput for Amazon Bedrock.
D
Train the model on journals and textbooks.