Changelog

added

Use LLM as a Reference in Experiments

With this new feature, you're able to use the output of a large language model like GPT-4 as the reference for another model like Gemma-7b and Mistal-large (see image).

added

Import from Resources in Experiments

Instead of manually adding your data sets or uploading them through a CSV file in experiments, you can now import them from the files you stored in Resources. This allows you to store and access your files in a quick and efficient manner.

added

Resource management

Save your data sets, variables, and evaluators in the newly added resources tab.

added

Evaluators are available on our Platform

We have added Evaluators to our platform. With a wide range of industry-standard metrics and other relevant evaluators, you can check whether or not the output of your LLM is accurate, reliable, and contextually relevant.

improved

Python and Node SDK's improvements

In our latest SDK update, we're thrilled to share a series of enhancements that significantly boost the performance and capabilities of our platform.

added

Setup your own API key with the new AnyScale integration.

You could already select models such as Llama from Meta and Mixtral from Mistral in the model garden. But with this release, it is now possible to connect your own API key for AnyScale. This way you can use your own account and rate limits without having to rely on a shared key. Soon you'll be able to use your own private models and finetuning on AnyScale.

added

Mass Experimentation

You could already test out different models and configurations using our playground. However, with the introduction of Experiments, you are able to do this on a much larger scale.