
Ultimate access to all questions.
Question: 3
You are preparing a large legal document to be used in a generative AI model for text summarization. The document has many chapters, and each chapter contains multiple sections with varying lengths. The model you're using has a token limit of 2048 tokens for processing. Which of the following chunking strategies would best ensure efficient processing of the document without exceeding the token limit?
Explanation:
When preparing a large legal document for use in a generative AI model that has a token limit of 2048, the most efficient way to chunk the text is to split the document into smaller, manageable sections that respect natural language boundaries, such as sentences. This ensures that each chunk is coherent and meaningful without exceeding the model's token limit.