Evaluating GenAI solutions: science or art?

alt text

GenAI (Generative Artificial Intelligence) has enabled several new applications through the generation of targeted content.

Developing a new product involving GenAI, how do we know it effectively addresses its intended use?

Being a new field that is getting a lot of traction, there are several practices, theories, and perspectives. You might be developing, using, or benefitting from GenAI-related products. Either way, you will need to prioritize the most relevant aspects to you. In this article, I will share my perspective on the most relevant principles that helped me shape my thinking in designing solutions involving GenAI.

Is it science or art?

Unlike more traditional AI, this novel field involves generating, or creating, if you prefer, content rather than providing answers. Being a field that involves “creativity”, there is some degree of subjectivity, especially in judging the quality of the generated content. While scientific rigour is still relevant, there are also a lot of areas that require creative thinking and management of perspectives, hence the need for a well-rounded approach. For example, how did I determine which images to include in this article? All of the images were generated using GenAI, and I relied on my own judgment to evaluate their quality and relevance. And how about the previous sentence? That was also rephrased through GenAI and I considered it “better” than the initial sentence that I wrote. How about the rest of the article? A good portion of this article was written directly, or “handwritten” in the sense that I typed every character and used GenAI as a reviewer rather than a content creator.

In this article, I will focus on the “why” and “what”, touching on the “how” very briefly.

In other words, I will share why I believe that some principles are crucial and what I believe that the solution needs to address. About the “how”, I have seen these challenges addressed in different ways and I would strongly recommend designing approaches that fit the purpose, accounting for both the use case and the broader environment. The prioritized aspects are new challenges, back-testing strategy, deployment & operations. Each section covers an aspect in more detail.

New challenges

alt text

What are the key aspects of evaluating a GenAI solution?

To determine the effectiveness of a solution involving GenAI, the key aspects are not necessarily the same as those for more traditional AI.

Let’s start with a concrete example: imagine that you want to grow the customer base of a real estate app providing potential buyers with recommendations. To be more effective towards your goal, you would like to improve the customer experience with a new chatbot that provides valuable details such as energy class, potential upgrades, and suggested locations.

Your solution involves an OpenAI chatbot that includes a RAG (Retrieval Augmented Generation) component. Will it provide the users with the answers they are expecting?

Evaluating GenAI is more challenging than evaluating traditional AI due to its complex, qualitative outputs as opposed to the straightforward numerical predictions from traditional AI.

Traditional AI, which produces numerical outcomes like forecasts, can be easily evaluated by comparing the generated numbers with known or expected values. In contrast, GenAI creates new content, such as text or images. The quality of GenAI outputs, being more qualitative and subjective, is harder to assess. It requires understanding the content’s purpose, target audience, and broader context, making it a more complex process that goes beyond numerical comparisons and involves human judgement and feedback.

Back-testing strategy

alt text

Backtesting involves evaluating the solution within a controlled, simulated environment with pre-defined scenarios and expected outcomes.

To test the effectiveness of a solution, the assumption is that the future behaviour will be somehow consistent with a simulated environment. For instance, in our real estate app use case, a simulated environment involves sample queries from end users, either real, generated or simulated.

The final outcome of a GenAI solution has some degree of subjectivity.

The final outcome is text, and its effectiveness depends on the satisfaction of the end user. There are different angles to measure it, and it’s arguably impossible to have an objective evaluation in a simulated environment due to its complexity. Analytical metrics partially answer the problem and are not necessarily related to the effectiveness of the solution.

A part of the solution can be assessed more quantitatively. Each aspect likely requires separate backtesting.

It is possible to break down the solution into smaller steps and objectively measure some aspects. A key aspect in a RAG solution is the capability to retrieve relevant information. In terms of backtesting, the key aspects are:

This is just an example, related to basic RAG solutions. Depending on the design of the solution, where GenAI might not be the only component, there might be different aspects to test.

Another key aspect is ethical compliance.

Mitigating potential ethical risks is arguably a priority for most companies. One option is to perform automated checks in a simulated environment. For instance, solution back-testing could check for discrimination, demeaning behaviour, and other aspects of Responsible AI. In this case, there could, for example, be checks for consistent experiences across requests from different geographies, as well as detection of misleading answers that could have negative consequences.

Meta-GenAI can be a powerful tool to generate scenarios and evaluate them.

In all the steps requiring automated generation of content, GenAI can help. For instance, it can generate sample questions and evaluate different aspects of the answer. I have seen backtesting solutions implemented with the help of GenAI. This is arguably a strong point in favour of it being an art, where you can get creative in the use of GenAI, without compromising on scientific rigour. A crucial risk to consider is that GenAI’s behavior may not be entirely predictable, making it all the more important to exercise additional caution.

Deployment & operations

alt text

A GenAI solution is dependent on data, which might be subject to change. This change needs to be managed to mitigate future risks.

GenAI solutions based on RAG patterns make use of a knowledge base, which might be subject to change. Another moving part could be the core GenAI model, which might be subject to re-training, especially now that the applications are becoming more mature.

Each aspect can be handled through LMOps.

LMOps (Language Models Machine Learning Operations) is a new term, based on the more traditional MLOps (Machine Learning Operations), contextualized to LLM. There are different practices and, in my opinion, the commonality is the need to manage changes in both technology, through more traditional techniques such as CI/CD pipelines, and data. Changes in the data are often managed as part of a more traditional approach, triggering the release of a newer version of the solution if more effective, and providing a domain expert with the option to keep a previous version if appropriate.

A new aspect to manage is the knowledge base.

In traditional machine learning, a change in the data would trigger model re-training. In RAG GenAI projects, the main outcome is often a newer version of the knowledge base. Either way, even if there are no changes in the technology, the solution might behave differently due to the new data, so a newer version should go through similar strategies as more traditional DevOps such as automatically rejecting versions that do not pass basic tests or are not more effective, and by providing a domain expert with the option to prefer a previous version.

Conclusions

Reflecting on the principles covered in this article, GenAI is a new field with challenges and opportunities.

These principles are meant to be a non-exhaustive starting point. They highlight the importance of a well-rounded approach, combining scientific rigour with creative thinking.

Like any emerging field, there’s always room for further discussion and refinement. Do these principles resonate with your experience in the field? Is there anything fundamental that you think is missing, or that you would approach differently?