Dataset Viewer
Auto-converted to Parquet Duplicate
group
stringclasses
1 value
unique_tokens
float64
100M
174B
params
float64
7.1M
3.9B
tokens
float64
100M
300B
loss
float64
2.57
8.1
all_data
4,000,000,000
2,810,000,000
32,000,000,000
2.722962
all_data
4,000,000,000
2,810,000,000
40,000,000,000
2.706547
all_data
4,000,000,000
2,810,000,000
55,000,000,000
2.696432
all_data
9,000,000,000
2,810,000,000
55,000,000,000
2.611045
all_data
11,000,000,000
2,810,000,000
55,000,000,000
2.598793
all_data
14,000,000,000
2,810,000,000
55,000,000,000
2.589427
all_data
18,000,000,000
2,810,000,000
55,000,000,000
2.584592
all_data
28,000,000,000
2,810,000,000
55,000,000,000
2.579361
all_data
55,000,000,000
2,810,000,000
55,000,000,000
2.574117
all_data
100,000,000
7,098,752
100,000,000
8.102005
all_data
100,000,000
7,098,752
200,000,000
7.36236
all_data
100,000,000
1,096,300,000
100,000,000
6.611002
all_data
100,000,000
14,100,000
100,000,000
7.278144
all_data
400,000,000
19,703,712
400,000,000
6.096268
all_data
400,000,000
35,500,000
400,000,000
5.79413
all_data
100,000,000
35,500,000
200,000,000
6.252892
all_data
100,000,000
35,500,000
400,000,000
5.799587
all_data
100,000,000
35,500,000
800,000,000
5.284175
all_data
100,000,000
14,100,000
200,000,000
6.664138
all_data
100,000,000
14,100,000
400,000,000
6.273184
all_data
100,000,000
14,100,000
800,000,000
5.816824
all_data
100,000,000
14,100,000
1,500,000,000
5.364749
all_data
100,000,000
14,100,000
2,700,000,000
4.992776
all_data
100,000,000
14,100,000
3,900,000,000
4.741132
all_data
100,000,000
14,100,000
5,900,000,000
4.546227
all_data
100,000,000
14,100,000
14,000,000,000
4.383392
all_data
100,000,000
14,100,000
20,000,000,000
4.51041
all_data
100,000,000
14,100,000
91,000,000,000
4.396074
all_data
100,000,000
14,100,000
174,000,000,000
4.359691
all_data
100,000,000
44,000,000
200,000,000
6.18997
all_data
100,000,000
44,000,000
400,000,000
5.648868
all_data
100,000,000
44,000,000
1,500,000,000
4.387672
all_data
100,000,000
44,000,000
2,700,000,000
4.113597
all_data
100,000,000
44,000,000
3,900,000,000
3.988205
all_data
100,000,000
44,000,000
5,900,000,000
3.89434
all_data
100,000,000
44,000,000
7,500,000,000
3.854261
all_data
100,000,000
44,000,000
14,000,000,000
3.807672
all_data
100,000,000
44,000,000
20,000,000,000
3.92903
all_data
100,000,000
44,000,000
32,000,000,000
3.89224
all_data
100,000,000
44,000,000
91,000,000,000
3.8362
all_data
100,000,000
44,000,000
174,000,000,000
3.812315
all_data
1,500,000,000
82,700,000
1,500,000,000
4.36208
all_data
1,500,000,000
201,236,224
1,500,000,000
3.929866
all_data
1,500,000,000
1,096,300,000
1,500,000,000
3.680017
all_data
400,000,000
1,096,300,000
1,500,000,000
3.704141
all_data
100,000,000
1,096,300,000
1,500,000,000
3.819042
all_data
2,700,000,000
618,700,000
2,700,000,000
3.415532
all_data
1,500,000,000
618,700,000
2,700,000,000
3.409955
all_data
400,000,000
618,700,000
2,700,000,000
3.449187
all_data
100,000,000
618,700,000
2,700,000,000
3.829336
all_data
100,000,000
618,700,000
1,500,000,000
3.784308
all_data
3,900,000,000
421,200,000
3,900,000,000
3.348126
all_data
1,500,000,000
421,200,000
3,900,000,000
3.35496
all_data
400,000,000
421,200,000
3,900,000,000
3.404755
all_data
100,000,000
421,200,000
3,900,000,000
3.89862
all_data
100,000,000
421,200,000
1,500,000,000
3.797898
all_data
100,000,000
421,200,000
2,700,000,000
3.747371
all_data
100,000,000
421,200,000
5,900,000,000
4.169935
all_data
100,000,000
421,200,000
7,500,000,000
4.376178
all_data
5,900,000,000
281,000,000
5,900,000,000
3.319542
all_data
1,500,000,000
281,000,000
5,900,000,000
3.330755
all_data
400,000,000
281,000,000
5,900,000,000
3.386692
all_data
100,000,000
281,000,000
5,900,000,000
3.882211
all_data
100,000,000
281,000,000
1,500,000,000
3.877228
all_data
100,000,000
281,000,000
2,700,000,000
3.72759
all_data
100,000,000
281,000,000
3,900,000,000
3.754965
all_data
100,000,000
281,000,000
7,500,000,000
3.99341
all_data
400,000,000
281,000,000
91,000,000,000
3.312059
all_data
7,500,000,000
220,500,000
7,500,000,000
3.310159
all_data
1,500,000,000
220,500,000
7,500,000,000
3.325092
all_data
400,000,000
220,500,000
7,500,000,000
3.374362
all_data
100,000,000
220,500,000
7,500,000,000
3.848433
all_data
100,000,000
220,500,000
1,500,000,000
3.932167
all_data
100,000,000
220,500,000
2,700,000,000
3.752793
all_data
100,000,000
220,500,000
3,900,000,000
3.721901
all_data
100,000,000
220,500,000
5,900,000,000
3.78097
all_data
14,000,000,000
146,500,000
14,000,000,000
3.319729
all_data
1,500,000,000
146,500,000
14,000,000,000
3.334061
all_data
400,000,000
146,500,000
14,000,000,000
3.390927
all_data
100,000,000
146,500,000
14,000,000,000
3.801428
all_data
400,000,000
146,500,000
60,000,000,000
3.34025
all_data
400,000,000
146,500,000
91,000,000,000
3.311519
all_data
400,000,000
146,500,000
174,000,000,000
3.277013
all_data
100,000,000
146,500,000
1,500,000,000
4.033501
all_data
100,000,000
146,500,000
2,700,000,000
3.80764
all_data
100,000,000
146,500,000
3,900,000,000
3.732758
all_data
100,000,000
146,500,000
5,900,000,000
3.718068
all_data
100,000,000
146,500,000
7,500,000,000
3.735044
all_data
100,000,000
146,500,000
20,000,000,000
3.756257
all_data
100,000,000
146,500,000
32,000,000,000
3.793078
all_data
100,000,000
146,500,000
60,000,000,000
3.846421
all_data
100,000,000
146,500,000
91,000,000,000
3.862952
all_data
100,000,000
146,500,000
174,000,000,000
3.897418
all_data
20,000,000,000
82,700,000
20,000,000,000
3.608018
all_data
1,500,000,000
82,700,000
20,000,000,000
3.618937
all_data
400,000,000
82,700,000
20,000,000,000
3.651057
all_data
100,000,000
82,700,000
20,000,000,000
3.790125
all_data
100,000,000
82,700,000
400,000,000
5.488318
all_data
100,000,000
82,700,000
1,500,000,000
4.20696
all_data
100,000,000
82,700,000
2,700,000,000
3.953894
End of preview. Expand in Data Studio

Budget-Efficient Scaling Law Fitting Benchmark

This repository contains the scaling-law benchmark dataset used in Spend Less, Fit Better: Budget-Efficient Scaling Law Fitting via Active Experiment Selection.

The benchmark is designed for budget-aware sequential experimental design in scaling-law fitting. Each configuration provides a finite pool of candidate experiments, a held-out high-cost target region, task-specific covariates, observed outcomes, and companion scaling-law definitions in laws.py.

Dataset Summary

The dataset contains 8 tabular regression tasks and 65 scaling-law instances. The tasks cover language-model scaling settings including pre-training hyperparameter tuning, data allocation, vocabulary design, domain mixture optimization, mixture-of-experts design, sparsity, parallel/inference-time scaling, and Farseer-style dense pre-training scaling.

Each task is stored as a separate Hugging Face configuration with train and test splits:

Config Train Test Feature columns Target column(s) Law instances
data_constrained_scaling_law 161 21 unique_tokens, params, tokens loss 10
domain_mixture_scaling_law 80 24 proportion_domain_1 ... proportion_domain_5 loss_domain_1 ... loss_domain_5 10
farseer_scaling_law 404 7 N, D loss 1
lr_bsz_scaling_law 2702 117 lr, bsz, data_size, non_embedding_param_size lm_loss 10
moe_scaling_law 193 28 num_experts, dense_parameter_count loss_validation 10
parallel_scaling_law 36 12 num_params, parallel_size loss 10
sparsity_scaling_law 70 18 P, N_active loss 4
vocab_scaling_law 1080 120 non_vocab_parameters, vocab_size, num_characters unigram_normalized_loss 10

The group column identifies a task-specific subproblem or grouping. For example, domain-mixture rows are grouped by model scale, and parallel-scaling rows are grouped by evaluation corpus.

Loading

from datasets import load_dataset

ds = load_dataset("sijieli/scalebench", "lr_bsz_scaling_law")
print(ds)
print(ds["train"][0])

To load a local checkout before uploading:

from datasets import load_dataset

ds = load_dataset(
    "parquet",
    data_files={
        "train": "lr_bsz_scaling_law/train-*.parquet",
        "test": "lr_bsz_scaling_law/test-*.parquet",
    },
)

Intended Use

This benchmark is intended for evaluating experiment-selection and active experimental-design methods for scaling-law fitting under budget constraints. A typical episode treats the train split as the candidate pool of runnable experiments and the test split as the target region for extrapolation evaluation.

The benchmark can be used to compare methods that:

  • choose experiments sequentially under a cost budget;
  • fit nonlinear scaling laws from sparse observations;
  • extrapolate to held-out high-cost regions;
  • optimize target-region prediction quality rather than in-sample fit.

Cost Proxies

The paper uses task-specific cost proxies to model heterogeneous experiment costs. The implementation in registry.py defines the default proxies:

Config Cost proxy
data_constrained_scaling_law 6 * params * tokens
domain_mixture_scaling_law 1
farseer_scaling_law 6 * N * D
lr_bsz_scaling_law 6 * non_embedding_param_size * data_size
moe_scaling_law dense_parameter_count * num_experts
parallel_scaling_law num_params
sparsity_scaling_law 6 * N_dense * D1 + 6 * N_active * D2
vocab_scaling_law non_vocab_parameters * num_characters

Scaling-Law Definitions

Each task directory includes a laws.py file containing the parametric scaling-law families used in the benchmark. The functions are named sl_1, sl_2, etc., and each file exposes:

  • LAW_REGISTRY: mapping from law ID to callable;
  • PARAM_COUNTS: number of free parameters for each law;
  • parameter bounds used by the fitting code.

These files are included to make the dataset self-contained for reproducing the benchmark protocol.

Citation

If you use this benchmark, please cite:

@misc{li2026spendlessfitbetter,
  title={Spend Less, Fit Better: Budget-Efficient Scaling Law Fitting via Active Experiment Selection},
  author={Sijie Li and Shanda Li and Haowei Lin and Weiwei Sun and Ameet Talwalkar and Yiming Yang},
  year={2026},
  eprint={2604.22753},
  archivePrefix={arXiv},
  primaryClass={cs.LG},
  url={https://arxiv.org/abs/2604.22753}
}

License

This dataset card follows the license metadata declared for this repository. Users should also respect the licenses and terms of the original data sources referenced by the paper.

Downloads last month
18

Paper for sijieli/scalebench