Generated text reflects the underlying dataset's biases, which can include historical and systemic bias; often lacks representation from diverse perspectives; and may not reflect the full reality of possible legal outcomes.
Examples:
It is usually unclear how GenAI tools determine which cases or other sources to reference, summarize, or otherwise bring into a generated response—this is the "black box" of AI.
Examples:
These tools also typically do not account for contextual information, which means that their output reflects bias due to the narrow information it is drawing on.
Example:
Many legal research tasks do not have one right answer; a series of possible answers or angles to the problem may exist. Even a relatively simple task like summarizing a case reflects a certain bias—both in terms of what aspects of the case to prioritize in the summary, and in the language used to summarize the facts.
Examples:
The analysis layer may also be biased based on the language model itself, which may have learned certain assumptions about words and phrases based on its dataset that do not reflect their legal interpretation.
Example:
A researcher's personal bias (worldview, implicit bias, personal experiences, etc.) will be reflected in the prompt they provide, which will in turn affect what the system provides as output.
Example:
Automation bias: A researcher may be more likely to trust the output of a system because it comes from a machine.
Example: