Trialing and Evaluating AI Software? Five Hallmarks of a Successful Comparative Trial
Everyone wants to use AI and reap the rewards with increased productivity, more automation, cost savings and streamlined workflows. But how does a law firm or corporate law department compare different AI tools and offerings?
November 25, 2019 at 07:00 AM
5 minute read
To some, artificial intelligence (AI) tools are still the new shiny object in legal. For others, they are proven solutions helping legal teams gain efficiencies and a competitive edge, especially in the contract analysis arena. Regardless of legal's AI journey, everyone wants to use AI and reap the rewards with increased productivity, more automation, cost savings and streamlined workflows.
The big challenge: How does a law firm or corporate law department compare different AI tools and offerings? Without the necessary subject matter expertise? How do you ask the right questions? How do you do this … cost-effectively? comprehensively? methodically? and collaboratively?
Over an eighteen-month period, I conducted a comprehensive comparative trial of nine well-known machine-learning and AI contract analysis software solutions. The universe of tested tools has since increased to 12. The main objective of the trials was to thoroughly evaluate the capabilities of each of these technologies. The end goal was to help a consortium of six global legal departments make informed buying decisions based on each providers performance. Since the original trial, law firms (including those that are already using one of the tools) and other players (including large corporates interested in acquiring the technologies) have been looking for guidance from the trial.
The results were in some cases surprising. Overall, there were far fewer solutions than we expected who had developed to the point that they could learn to recognise new clause types in real time. For the members of our consortium, they benefited from basing their decisions on the outcome of standardized tests rather than just opting for a tool that seems 'safe.' One of the best solutions even managed to identify clauses in our test set that we, the researchers, initially overlooked!
Setting up a comparative trial of this magnitude was no easy feat. Individual law firms or legal departments might easily spend upwards of 200 hours in running a comparable trial spanning document search, RFI development and vendor selection, solution setup and training, trialing solutions, compiling results and final decision making. What I ended up with was a long list of 'lessons learned' and 'best practices,' which I have managed to whittle down to a top five:
Put thought into your RFI: An RFI should be crafted to solicit interest from vendors who are qualified to participate—think through your RFI from the vendor's perspective before you hit send. Approach vendors in a collaborative and collegial fashion. Avoid 'gotcha' type questions or asking them to commit resources beyond what is really necessary. Focus on questions that are specifically going to yield the information you actually need. Make sure vendors clearly understand what you are asking in order to ensure accurate RFI responses.
Build bridges, not barriers: Based on my experience, RFIs crafted by teams of lawyers are often far more complex than they need to be to achieve the desired result. Provide a concise and simple RFI that encourages collaboration with vendors and transparency with regard to trial costs and resource commitments.
More is not always more: Since the trial process is often cumbersome and time consuming to begin with, it's tempting to cram in as many testing criteria as possible. Testing too many software applications for too many features and functions over complicates the process and unnecessarily extends the bake-off period. Instead, avoid scope creep by focusing on a shorter list of key outcomes that are most important to your firm.
Test and test again: In our comparative trial, we devised a four-part testing protocol for each of the dozen machine learning tools. This included understanding the technology; going 'under the hood' with the respective technology stacks; reconciling our testing with what we deemed vital as part of the RFI process; and lastly, running actual, comprehensive performance and training tests on each of the AI/ML tools and scoring them against standardized criteria. While testing protocols will vary, it is critical that testing criteria are uniform and consistent in order to in the end produce true 'apple v. apple' comparisons.
Report Results … then act: It is paramount that trial results are properly reported, shared with the appropriate constituents, and recommendations are implemented. Above all, setting up a comprehensive comparative trial with the above lessons learned in mind allows firms to base their decisions on actual vendor performance instead of following the crowd or acting on some educated guesses.
Friedrich is an entrepreneur, business builder and legal industry leader, with expertise in improving the competitiveness and financial performance of legal service providers. Over the past 20 years, Friedrich has consulted with and worked for law firms, legal departments and alternative legal service providers in 17 countries spanning four continents. He can be reached at [email protected].
This content has been archived. It is available through our partners, LexisNexis® and Bloomberg Law.
To view this content, please continue to their sites.
Not a Lexis Subscriber?
Subscribe Now
Not a Bloomberg Law Subscriber?
Subscribe Now
NOT FOR REPRINT
© 2025 ALM Global, LLC, All Rights Reserved. Request academic re-use from www.copyright.com. All other uses, submit a request to [email protected]. For more information visit Asset & Logo Licensing.
You Might Like
View AllTrending Stories
- 1SDNY US Attorney Damian Williams Lands at Paul Weiss
- 2Litigators of the Week: A Knockout Blow to Latest FCC Net Neutrality Rules After ‘Loper Bright’
- 3Litigator of the Week Runners-Up and Shout-Outs
- 4Norton Rose Sues South Africa Government Over Ethnicity Score System
- 5KMPG Wants to Provide Legal Services in the US. Now All Eyes Are on Their Big Four Peers
Who Got The Work
Michael G. Bongiorno, Andrew Scott Dulberg and Elizabeth E. Driscoll from Wilmer Cutler Pickering Hale and Dorr have stepped in to represent Symbotic Inc., an A.I.-enabled technology platform that focuses on increasing supply chain efficiency, and other defendants in a pending shareholder derivative lawsuit. The case, filed Oct. 2 in Massachusetts District Court by the Brown Law Firm on behalf of Stephen Austen, accuses certain officers and directors of misleading investors in regard to Symbotic's potential for margin growth by failing to disclose that the company was not equipped to timely deploy its systems or manage expenses through project delays. The case, assigned to U.S. District Judge Nathaniel M. Gorton, is 1:24-cv-12522, Austen v. Cohen et al.
Who Got The Work
Edmund Polubinski and Marie Killmond of Davis Polk & Wardwell have entered appearances for data platform software development company MongoDB and other defendants in a pending shareholder derivative lawsuit. The action, filed Oct. 7 in New York Southern District Court by the Brown Law Firm, accuses the company's directors and/or officers of falsely expressing confidence in the company’s restructuring of its sales incentive plan and downplaying the severity of decreases in its upfront commitments. The case is 1:24-cv-07594, Roy v. Ittycheria et al.
Who Got The Work
Amy O. Bruchs and Kurt F. Ellison of Michael Best & Friedrich have entered appearances for Epic Systems Corp. in a pending employment discrimination lawsuit. The suit was filed Sept. 7 in Wisconsin Western District Court by Levine Eisberner LLC and Siri & Glimstad on behalf of a project manager who claims that he was wrongfully terminated after applying for a religious exemption to the defendant's COVID-19 vaccine mandate. The case, assigned to U.S. Magistrate Judge Anita Marie Boor, is 3:24-cv-00630, Secker, Nathan v. Epic Systems Corporation.
Who Got The Work
David X. Sullivan, Thomas J. Finn and Gregory A. Hall from McCarter & English have entered appearances for Sunrun Installation Services in a pending civil rights lawsuit. The complaint was filed Sept. 4 in Connecticut District Court by attorney Robert M. Berke on behalf of former employee George Edward Steins, who was arrested and charged with employing an unregistered home improvement salesperson. The complaint alleges that had Sunrun informed the Connecticut Department of Consumer Protection that the plaintiff's employment had ended in 2017 and that he no longer held Sunrun's home improvement contractor license, he would not have been hit with charges, which were dismissed in May 2024. The case, assigned to U.S. District Judge Jeffrey A. Meyer, is 3:24-cv-01423, Steins v. Sunrun, Inc. et al.
Who Got The Work
Greenberg Traurig shareholder Joshua L. Raskin has entered an appearance for boohoo.com UK Ltd. in a pending patent infringement lawsuit. The suit, filed Sept. 3 in Texas Eastern District Court by Rozier Hardt McDonough on behalf of Alto Dynamics, asserts five patents related to an online shopping platform. The case, assigned to U.S. District Judge Rodney Gilstrap, is 2:24-cv-00719, Alto Dynamics, LLC v. boohoo.com UK Limited.
Featured Firms
Law Offices of Gary Martin Hays & Associates, P.C.
(470) 294-1674
Law Offices of Mark E. Salomone
(857) 444-6468
Smith & Hassler
(713) 739-1250