AutoML for Large Language Models
Build custom large language models for a range of tasks
Overview
Hive AutoML for large language models (LLMs) can be used to build custom generative text models for customer service, chat bots, and more. Our AutoML tool can help create LLMs with a specific domain of knowledge that isn't the focus of more general models such as our Text Generation API.
Creating a New Training
To start building your model, head to our AutoML platform and select the Create New Model
button in the top right corner of the screen. You will be brought to a project setup page where you will be prompted to enter a project name and description. Click the box below Model Type
and select Language Generative
from the menu that appears. Below the model type field, add your training dataset either by clicking the Upload File
button or the Select Dataset
button, which allows you to choose from a list of datasets that you've already uploaded to the AutoML platform.

The AutoML Training Projects dashboard. The Create New Model
button sits in the top right corner.
Dataset Upload
Your data will be uploaded as a CSV file. To format it, one column (titled prompt
) should contain the text prompt and a second column (titled completion
) should contain the text response to that prompt. An example of this formatting is shown below:
prompt | completion |
---|---|
Question: In 19th Century Florence, it was illegal for women to wear what? | Buttons |
Question: What animal cannot stick out its tongue? | Crocodile |
Question: What color is an airplane's black box? | Bright orange |
Question: How long did the 100 years war last? | 116 years |
Question: What was the drink Bloody Mary originally called? | Bucket of Blood |
The uploaded data file must include both a "prompt" column and a "completion" column or the training will result in an error.
In the example shown above, the text prompt (prompt
) includes a trivia question and completion field for that prompt (completion
) contains the answer.
In order to be processed correctly, the data file must satisfy the following requirements:
- Dataset file must be in CSV format.
- CSV must have a header row (a row of column names above the actual data).
- CSV must use
,
as the delimiter. Other delimiters such as;
or|
are invalid. - The header row must contain columns for both the prompt and the completion called
prompt
andcompletion
. - Each row can contain a maximum of 2048 tokens (approximately 8,000 characters). If a row exceeds this, that row will be considered invalid and will not be used during training. If any rows exceed this limit, we will let users know which ones have done so. If more than 5% of rows in the dataset are invalid, the training will fail entirely.
- Each CSV file cannot contain more than 50,000 rows.
- Each CSV must contain at least 100 rows. At least 2,000 rows is strongly recommended for optimal model quality.
- Any empty columns in the CSV will not be used during model training.
If any of the above are not satisfied, the training will fail and return an error.
We strongly recommend that the training data includes at least 2,000 rows for optimal model quality.
Test Dataset
For your test dataset, you can choose to upload a separate test dataset, select one from your previously uploaded datasets, or split off a random section of your training dataset to use instead. If you choose to upload a separate test dataset, this dataset must also satisfy all of the file requirements listed above and must contain the same heads and classes as your training dataset. If you choose to split off a section of your training dataset, you will be able to choose the percentage of that dataset that you would like to use for testing.
Evaluating Model Performance
After model training is complete, viewing the page for that project will provide a few metrics in order to help you evaluate the performance of your model. At the top of the page, you can view both the loss and the token accuracy for the model.
The loss
measures how closely the model’s response matches the response from the test data, where 0
represents a perfect prediction, and a higher loss signifies that the prediction is increasingly far from the actual response sequence. If the response has 10 tokens, we let the model predict each of the 10 tokens given all previous tokens are the same. We then display the final numerical loss value, as well as the difference to the loss of the original pre-trained model before fine tuning.
The token accuracy
compares the model's response to the completion from the test data, determining how many tokens are the same between the two of them. This is shown as a percentage indicating the probability of an exact match between them. Similarly to loss
, we will contextualize the final result by showing the percent change relative to the original pre-trained model.
You can also evaluate your model by interacting with it in what we call the playground
, shown below the loss and token accuracy. Here you can submit prompts directly to your model and view its response, allowing model evaluation through experimentation. This will be available for 15 days after model training is complete, and has a limit of 500 requests. If either the time or request limit is reached, you can instead choose to deploy the model and continue to use the playground feature with unlimited uses which will be charged to the organization's billing account.
After a model is deployed, any prompts submitted to the playground will be sent through the API and billed as such.

Evaluation metrics for an AutoML project as shown after training has completed.
If you would like to retrain your model based on these metrics, go back to your AutoML Training Projects page and start a new project.
Deploying Model With Hive Data
When you’re happy with your model and ready to deploy it, select the project and click the “Create Deployment” button in the top right corner. The project’s status will shift to “Deploying.” The deployment may take a few minutes.
After the deploy status shows as “Complete,” you can view the deployment by clicking on the "Deployments" tab above the metrics. This will show a list of all deployments for this model.

The "Deployments" tab, showing two different model deployments and their statuses.
To view any deployment, hit the exit symbol to the left of its name. This will open the project on Hive Data, where you will be able to upload tasks, view tasks, and access your API key as you would with any other Hive Data project. There will also be a button to "Undeploy" your project, if you wish to deactivate it at any point. Undeploying a model is not permanent — you can redeploy the project if you later choose to.

An AutoML project as viewed in the Hive customer dashboard. The "API Key" button is on the top right.
To begin using your custom-built API, click on the “API Key” button on the top right of the Hive Data project page to copy your API Key. For instructions on how to submit a task via API, either synchronously or asynchronously, see our API Reference documentation.
Updated about 2 months ago