By Madison Johnson, Esq. | Manager, Large Markets Big Law is coming off a highly profitable year in 2024, and one implication of this financial success is the generation of cash flow that can be invested...
By Madison Johnson, Esq. | Manager, Large Markets The new year has introduced some terminology into the vocabulary of everyone who is trying to keep up with the pace of innovation in artificial intelligence...
In-house legal professionals are entering 2025 with an expectation that the year will bring challenges they can anticipate, as well as a host of headaches that are unexpected right now. They are considering...
By Madison Johnson, Esq | Manager, Large Markets The return of President Trump to the White House for another four-year term has large law firms moving to adapt their service offerings and practice area...
The past year was filled with exciting breakthroughs of products built with new generative artificial intelligence (Gen AI) technology, but it also ushered in a peculiar new regulatory skirmish that in...
By Geoffrey D. Ivnik, Esq. | Director of Large Markets, LexisNexis
The emergence of generative artificial intelligence (Gen AI) tools is an exciting breakthrough in our industry, but it is appropriate to also approach this new technology with caution. We are all aware by now of the risks of relying on Gen AI outputs without proper oversight by a lawyer, just as a lawyer would oversee the work product of a summer associate or a paralegal.
One way to mitigate these risks is to make sure that you are using a Legal AI tool — a Gen AI solution trained for the legal profession — that is grounded in authoritative legal content and developed with “Responsible AI” principles. This is a critical foundation that ensures the AI system producing outputs in response to your prompts has been developed, deployed and governed to comply with all relevant ethics and laws.
“Responsible AI is an approach to developing and deploying artificial intelligence from both an ethical and legal standpoint,” explains the International Organization for Standardization. “The goal is to employ AI in a safe, trustworthy and ethical way. Using AI responsibly should increase transparency while helping to reduce issues such as AI bias.”
Responsible AI is about more than doing the right thing; it is an important risk management guardrail to increase confidence that the Legal AI tool you are using will avoid potential reputational and financial damage to your law firm in the future.
Here are five keys to the responsible development of a Legal AI tool, drawn from the Responsible AI Principles at RELX, the parent company of LexisNexis:
The AI tool should have an appropriate level of transparency for each application and use case to ensure that different users can understand and trust the output.
Mathematical accuracy doesn’t guarantee freedom from bias, so the AI should be developed with procedures, extensive review and documentation processes, and the use of automated bias detection tools.
Humans must have ownership and accountability over the development, use and outcomes of AI systems. This requires an appropriate level of human oversight throughout the lifecycle of development and deployment, including ongoing quality assurance of machine outputs to pre-empt unintended use.
The AI tool should be developed with robust data management and security policies and procedures, ensuring that personal information is handled in accordance with all applicable privacy laws and regulations — as well as privacy principles that require the developer to always act as responsible data stewards. For example, LexisNexis has made data security and privacy for customers a priority by opting out of certain Microsoft AI monitoring features to ensure OpenAI cannot access or retain confidential customer data.
This technology should be built only after a reflection on: the sphere of influence of a new product; a map of stakeholders beyond direct users; and the domain to which the tool applies, including any potential impact related to an individual’s health, career or rights.
LexisNexis has been an industry leader in the development of responsible and trustworthy AI tools for several years. This began with our use of extractive AI models that used various machine learning techniques and has now advanced to the transformational technology of Gen AI with the breakthrough Lexis+ AI platform.
Lexis+ AI supports legal professionals with the ethical and responsible adoption of Gen AI tools. The platform is being built according to a framework of pre-defined principles, ethics and rules that guide everything we do. All answers are grounded in the world’s largest repository of accurate and exclusive legal content from LexisNexis, including case law, statutes, treatises and more. In fact, Lexis+ AI is the only Legal AI solution that provides linked citations in responses.
This innovative platform enables conversational search, insightful summarization, intelligent legal drafting, and document upload and analysis capabilities — all in a seamless user experience. It incorporates multiple large language models to match the best model for each research task.
We’re now taking the industry to the next level with the launch of our second-generation legal generative AI assistant on Lexis+ AI. The new version of our AI Assistant on Lexis+ AI delivers an even more personalized experience that will support legal professionals in making informed decisions faster, generating outstanding work, and freeing up time for them to focus on other efforts that drive value. All existing Lexis+ AI customers have access to the enhanced AI Assistant.
If you want to learn more about how Lexis+ AI can help legal professionals achieve better outcomes, or to sign up for the Lexis+ AI Insider program that provides the latest in Legal AI educational content, visit www.lexisnexis.com/ai.