Data Lab offers tools for preparing and managing datasets for fine-tuning workflows. For more information on fine-tuning, see the Post-training documentation.Documentation Index
Fetch the complete documentation index at: https://docs.tokenfactory.nebius.com/llms.txt
Use this file to discover all available pages before exploring further.
Dataset preparation
You can create fine-tuning datasets by:- Importing and filtering inference logs (chat completions)
- Uploading and filtering structured datasets
Reproducibility
By keeping datasets versioned and centralized, Data Lab enables:- Consistent training inputs across experiments
- Easier comparison of fine-tuning results
- Safer iteration without accidental data changes
Data location and compliance
- Fine-tuning jobs run in EU or US data centers, for more information, see our Legal Quick Guide (Data Location part)
- All datasets are stored centrally in the eu-north1 (Finland) data center.
Typical use cases
- Adapting models to internal domain-specific data
- Improving response quality for specific tasks
- Training student models for distillation pipelines