Day Four
DEMYSTIFYING: Setting and understanding reasonable levels of AI risk tolerance
Despite the common misconception that AI tools should perform flawlessly - like a calculator does - they are not infallible. Like an eager articling student, AI is ready and well equipped to assist, but may also misunderstand the nature of a request, fail to find the best source, or simply make a mistake. While it can significantly enhance efficiency and decision-making processes, AI has its limitations.
RISK: Compromising the quality of your work
The main risk lies in the assumption that AI outputs are always correct, which can lead to reliance on inaccurate results. Conversely, completely dismissing AI tools due to their potential for error may prevent individuals and organizations from benefiting from the efficiencies they offer. Ultimately, failing to balance trust in AI with a critical understanding of its limitations can compromise the quality of your work, especially in legal practice where accuracy is paramount.
MITIGATION
Mitigating these risks means making sure that AI technology complements rather than substitutes human judgment; it is critical that the lawyer be an active partner with AI. Ensure that all team members are informed and vigilant, and remember that lawyers are responsible for reviewing and validating AI-generated work before it is presented to clients or used in court.
POLICY CHECKLIST FOR YOUR ORGANIZATION
Remember that your role as lawyer doesn't change.
did you know…?
In 2019, IBM's Project Debater, an AI system designed to engage in persuasive argumentation, debated against two human debaters in San Francisco. The debate focused on whether preschools should be subsidized, with Project Debater arguing in favour of the motion and the human debaters arguing against it. The event showcased the AI's ability to formulate coherent arguments, respond to counterarguments, and engage in persuasive speech. While Project Debater's arguments were based on data and information it had been trained on rather than personal experience or beliefs, the debate demonstrated how AI can be used to engage in complex and nuanced discussions with humans on a wide range of topics. No winner was named.
HAVE YOU TRIED THIS FOR FUN…?
Exploring AI risk tolerance and error handling can be made engaging through interactive AI simulations or games. Imagine a text-based adventure game powered by AI, where players make decisions based on AI advice that may not always be perfect. Such games can simulate scenarios where AI limitations become apparent, teaching players the importance of critical assessment and human oversight in AI applications. This approach not only educates but also provides a hands-on experience with AI's intricacies, making the learning process enjoyable and impactful.