One of the more subtly cruel administrative experiences one can have is handling a loved one’s estate after they pass away. The court system was not created with bereaved people in mind, and there are forms to file, deadlines to meet, and procedures that vary based on what the deceased owned and how they owned it. The Alaska Virtual Assistant, a generative AI chatbot that has been in development for almost a year and a half and is intended to assist regular people in navigating the probate process without the need for legal representation, was created by Alaska’s court system to address precisely this issue. It was a good idea. It has been much more difficult to execute.
When the Alaska Court System team realized in real time how challenging it is to develop an AI tool that is both helpful and consistently accurate in a situation where inaccurate information causes real harm, what was initially planned to be a three-month project turned into more than fifteen months. “We need to be 100% accurate, and that’s really difficult with this technology,” one of the project’s leaders and administrative director for the court system, Stacey Marz, stated in an interview with NBC News. That could be the epitome of a thousand government AI projects that started out hopeful but ultimately failed.
| Category | Details |
|---|---|
| Project Name | Alaska Virtual Assistant (AVA) |
| Developed By | Alaska Court System (ACS) + LawDroid (Tom Martin) |
| Support Organization | National Center for State Courts (NCSC) |
| Purpose | Help self-represented residents navigate probate — the process of transferring a deceased person’s property |
| Originally Planned Duration | 3 months |
| Actual Duration (as of Jan 2026) | 15+ months |
| Key Personnel | Stacey Marz (ACS Administrative Director), Jeannie Sato (Director, Access to Justice Services), Aubrie Souza (NCSC Consultant) |
| Technology | Retrieval-augmented generative AI, built on OpenAI’s GPT family |
| Technology Partner | LawDroid (founded by Tom Martin, lawyer and law professor) |
| Test Questions Used | Reduced from 91 to 16 |
| Estimated Cost Per 20 Queries | ~$0.11 (model inference only) |
| Planned Launch | Late January 2026 |
| Notable Hallucination | Directed users to a law school alumni network — Alaska has no law school |
| Deloitte Stat | Less than 6% of local government practitioners prioritizing AI for service delivery |

The hallucination issue started early and continued. In one of the more notable instances, the chatbot recommended contacting a network of law school alumni when a user inquired about where to find legal assistance in Alaska. Alaska does not have a law school. The system confidently, helpfully, and incorrectly pulled information from sources other than its carefully selected knowledge base of court probate documents, even though it was not supposed to. Tom Martin, a lawyer and law professor who created AVA through his business LawDroid, has put a lot of effort into limiting the chatbot’s access to pertinent court documents. However, limiting what a model will say is a truly challenging engineering problem, not a setting you can change, as anyone who has worked with large language models is aware.
The testing regimen provides a unique account of the project’s origins. In order to assess the accuracy and usefulness of AVA, the team first created a set of 91 questions that covered a variety of probate topics, such as which forms to use in particular circumstances. Given the stakes and the requirement for human review at every stage, the 91-question test proved to be too time-consuming to administer and assess. Eventually, it was narrowed down to 16 questions, some of which addressed common questions, some of which covered subjects on which AVA had previously provided incorrect answers, and some of which covered more complicated situations. Methodological details that have not been fully disclosed will determine whether that reduction is a retreat from rigor or a sensible prioritization.
The particular human texture that this project was attempting to serve is difficult to ignore. AVA’s initial iterations were said to be overly sympathetic, beginning with condolences that user testing showed were unwelcome. People didn’t want an AI chatbot to show sympathy when they were trying to process a bank account or transfer a car title after losing a loved one. They sought explanations. The difference between creating AI in theory and using it to help people in real-world challenging situations is captured by that little detail. The condolences were taken down. However, editing away the deeper issue—getting the answers correct—proved to be much more difficult.
Here, the larger context is important. According to a recent Deloitte report, less than 6% of local government practitioners currently view AI as a priority tool for service delivery. This can be explained by the experience of the Alaskan court system. It takes time, resources, and human oversight to even partially close the gap between using AI in a consumer product and using it in a public legal context, where the cost of an incorrect response is not a slightly irritated user but rather a harmed litigant who might miss a filing deadline or submit the wrong form. “It was just so very labor-intensive to do this,” Marz said to NBC News, “despite all the buzz about generative AI and everybody saying this is going to revolutionize self-help and democratize access to the courts.”
Although government technology projects typically have aspirational launch dates, AVA was supposed to launch in late January 2026. According to Marz, the team’s objectives had changed to something more constrained and verifiable rather than attempting to completely replicate what a human facilitator could accomplish. That recalibration isn’t exactly a failure. It is what responsible organizations do when the technology doesn’t perform as advertised. The goal was justified. The challenge was genuine.
Disclaimer
Nothing published on Creative Learning Guild — including news articles, legal news, lawsuit summaries, settlement guides, legal analysis, financial commentary, expert opinion, educational content, or any other material — constitutes legal advice, financial advice, investment advice, or professional counsel of any kind. All content on this website is provided strictly for informational, educational, and news reporting purposes only. Consult your legal or financial advisor before taking any step.
