“Whether it’s Llama, OpenAI, Turbo, whatever, we can make sure that we’re at least getting the performance we’re getting today, if not improving it. And then from that, we can use Weave to make a decision on do we move to this model, do we move to a whole different provider? It’s really been kind of fundamental to how we test and how we move to new models.”