Skip to main content

Documentation Index

Fetch the complete documentation index at: https://docs.monostate.ai/llms.txt

Use this file to discover all available pages before exploring further.

Wizard Commands Reference

The AITraining wizard supports various commands to help you navigate, search, and configure your training job. These commands work at any prompt:
CommandShortcutDescription
:backGo back to the previous step
:help?, :hShow detailed help for the current prompt
:exit:quitCancel the wizard and exit

Using :back

You can go back at any point to change previous answers:
Model (number, HF ID, or command): :back
↩️ Going back to dataset selection...

Dataset (number, HF ID, or command):

Using :help

Every prompt has contextual help:
Training split name [train]: :help

ℹ️  Help
  Dataset splits are named subsets of your data.

  Common split names:
    • 'train' - Training data (most common)
    • 'test' - Test/evaluation data
    • 'validation' or 'valid' - Validation data

  Note: This is NOT asking for a percentage split (like 80/20).
  It's asking for the exact name of the split in your dataset.

Training split name [train]:

Catalog Commands

These commands work when browsing models or datasets:
CommandDescription
/search <query>Search for models/datasets by name
/sortChange sorting (trending, downloads, likes, recent)
/filterFilter models by size (models only)
/refreshClear cache and reload the list

/search

Find specific models or datasets:
Model (number, HF ID, or command): /search llama

Popular models (trending):
  1. meta-llama/Llama-3.2-1B (1B)
  2. meta-llama/Llama-3.2-3B (3B)
  3. meta-llama/Llama-3.1-8B (8B)
  4. meta-llama/Llama-3.1-70B (70B)
  ...
Search examples:
  • /search gemma - Find Gemma models
  • /search code - Find code-focused models
  • /search alpaca - Find Alpaca-style datasets
  • /search conversation - Find conversation datasets

/sort

Change how results are ordered:
Model (number, HF ID, or command): /sort
Sort options: [T]rending [D]ownloads [L]ikes [R]ecent
Sort by [T]: D
Sort OptionKeyDescription
TrendingTWhat’s popular right now
DownloadsDMost downloaded all-time
LikesLMost liked by the community
RecentRNewest additions

/filter

Filter models by parameter count (only works for models, not datasets):
Model (number, HF ID, or command): /filter
Filter size: [A]ll [S]mall(<3B) [M]edium(3-10B) [L]arge(>10B)
Filter size [A]: S
FilterKeySize RangeTypical Hardware
AllANo filterAny
SmallS< 3B parametersMacBook, consumer GPU
MediumM3B - 10B parametersGaming GPU, workstation
LargeL> 10B parametersCloud GPU, multi-GPU

/refresh

Clear the cache and fetch fresh data:
Model (number, HF ID, or command): /refresh
Cache cleared.

Popular models (trending):
  ...

Selection Methods

When choosing a model or dataset, you have several options:

By Number

Select from the displayed list:
Popular models (trending):
  1. google/gemma-3-270m (270M)
  2. google/gemma-2-2b (2B)
  3. meta-llama/Llama-3.2-1B (1B)

Model (number, HF ID, or command): 2
✓ Model: google/gemma-2-2b

By HuggingFace ID

Type the full model/dataset ID:
Model (number, HF ID, or command): mistralai/Mistral-7B-v0.3
✓ Model: mistralai/Mistral-7B-v0.3

By Local Path

Point to a local directory:
Dataset (number, HF ID, or command): ./my_training_data
✓ Dataset: ./my_training_data

Input Conventions

Defaults

Values in [brackets] are defaults. Press Enter to accept:
Project name [my-llm-project]: ↵
✓ Project: my-llm-project

Required Fields

Fields marked [REQUIRED] must be filled:
Prompt column name [REQUIRED] [prompt]: ↵
❌ This field is required for DPO/ORPO training.
Prompt column name [REQUIRED] [prompt]: instruction

Yes/No Questions

Answer with y/yes or n/no:
Configure advanced parameters? [y/N]: y

Enable LoRA? [Y/n]: ↵
✓ LoRA enabled (default)
Capitalized letter indicates the default:
  • [Y/n] - Default is Yes
  • [y/N] - Default is No

Keyboard Shortcuts

KeyAction
EnterAccept default or confirm input
Ctrl+CCancel wizard (same as :exit)
Arrow Up/DownScroll through numbered options (if supported)

Advanced Parameters

When configuring advanced parameters, the wizard groups them:
⚙️  Training Hyperparameters

Configure Training Hyperparameters parameters? [y/N]: y

epochs [1]:
batch_size [2]:
lr [3e-5]:
Each group can be configured independently:
GroupContains
Training Hyperparametersepochs, batch_size, lr, warmup_ratio
PEFT/LoRApeft, lora_r, lora_alpha, quantization
DPO/ORPOdpo_beta, max_prompt_length
Hub Integrationpush_to_hub, username, token
Knowledge Distillationteacher_model, distill_temperature
Hyperparameter Sweepuse_sweep, sweep_n_trials
Enhanced Evaluationuse_enhanced_eval, eval_metrics
Reinforcement Learningrl_reward_model_path (PPO only)

Tips

Every single prompt has detailed help. If you’re unsure what something means, type :help.
Made a wrong choice? Use :back to return to previous steps. Your other answers are preserved.
Instead of scrolling through hundreds of models, use /search llama or /search 7b to narrow down.
Not sure which models will work? Use /filterS (small) to see only models that fit consumer hardware.
On your first training, accept most defaults. Get something working, then customize.

Command Quick Reference

# Navigation
:back          Go to previous step
:help          Show help for current prompt
:exit          Cancel and exit

# Catalog (models/datasets)
/search query  Search by name
/sort          Change sort order
/filter        Filter by size (models only)
/refresh       Reload list

# Selection
1, 2, 3...     Select by number
google/gemma   Type HuggingFace ID
./my_data      Type local path

# Input
Enter          Accept default
y/n            Yes/No answers
Ctrl+C         Cancel