Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion docker/Dockerfile
Original file line number Diff line number Diff line change
Expand Up @@ -84,7 +84,7 @@ RUN cd /algorithmic-efficiency && git fetch origin
RUN cd /algorithmic-efficiency && git pull

# Todo: remove this, this is temporary for developing
COPY scripts/startup.sh /algorithmic-efficiency/docker/scripts/startup.sh
# COPY scripts/startup.sh /algorithmic-efficiency/docker/scripts/startup.sh
RUN chmod a+x /algorithmic-efficiency/docker/scripts/startup.sh

ENTRYPOINT ["bash", "/algorithmic-efficiency/docker/scripts/startup.sh"]
4 changes: 2 additions & 2 deletions docs/DOCUMENTATION.md
Original file line number Diff line number Diff line change
Expand Up @@ -797,9 +797,9 @@ a rough guideline, the entire set of workloads was designed to have a combined
runtime of very roughly $100$ hours on the
[**benchmarking hardware**](#benchmarking-hardware).

The eight *AlgoPerf Workloads* are:
The nine *AlgoPerf Workloads* are:

| **Task** | **Dataset** | **Model** | **Loss** | **Metric** | Validation<br>**Target** | Test<br>**Target** | Max<br>**Runtime** <br>*(in seconds)* | Default<br>**Dropout**<br>Value
| | **Task** | **Dataset** | **Model** | **Loss** | **Metric** | Validation<br>**Target** | Test<br>**Target** | Max<br>**Runtime** <br>*(in seconds)* | Default<br>**Dropout**<br>Value
----- | ----------------------------- | ----------- | ----------- | -------- | ---------- | ------------------------ | ------------------ | ------------------------------------- | -------------------------------
**1** | Clickthrough rate prediction | Criteo 1TB | DLRMsmall | CE | CE (↓) | 0.123735 | 0.126041 | 8,915 | 0
**2** | MRI reconstruction | fastMRI | U-Net | L1 | SSIM (↑) | 0.723653 | 0.740633 | 2,745 | 0
Expand Down
299 changes: 299 additions & 0 deletions pytorch_scoring_config_1.json
Original file line number Diff line number Diff line change
@@ -0,0 +1,299 @@
{
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I don't think we want to commit this file to the repository. Can you remove it from this branch?

"0": {
"framework": "pytorch",
"workload": "imagenet_resnet",
"dataset": "imagenet",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_0",
"rng_seed": -1447200680,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"1": {
"framework": "pytorch",
"workload": "imagenet_resnet",
"dataset": "imagenet",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_1",
"rng_seed": -1977906563,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"2": {
"framework": "pytorch",
"workload": "imagenet_resnet",
"dataset": "imagenet",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_2",
"rng_seed": 666869491,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"3": {
"framework": "pytorch",
"workload": "imagenet_vit",
"dataset": "imagenet",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_0",
"rng_seed": -796448826,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"4": {
"framework": "pytorch",
"workload": "imagenet_vit",
"dataset": "imagenet",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_1",
"rng_seed": -557820510,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"5": {
"framework": "pytorch",
"workload": "imagenet_vit",
"dataset": "imagenet",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_2",
"rng_seed": -1307522002,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"6": {
"framework": "pytorch",
"workload": "fastmri",
"dataset": "fastmri",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_0",
"rng_seed": 1083014187,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"7": {
"framework": "pytorch",
"workload": "fastmri",
"dataset": "fastmri",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_1",
"rng_seed": -1077277636,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"8": {
"framework": "pytorch",
"workload": "fastmri",
"dataset": "fastmri",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_2",
"rng_seed": -397959160,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"9": {
"framework": "pytorch",
"workload": "ogbg",
"dataset": "ogbg",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_0",
"rng_seed": 1662399765,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"10": {
"framework": "pytorch",
"workload": "ogbg",
"dataset": "ogbg",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_1",
"rng_seed": 486196682,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"11": {
"framework": "pytorch",
"workload": "ogbg",
"dataset": "ogbg",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_2",
"rng_seed": 1039483369,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"12": {
"framework": "pytorch",
"workload": "wmt",
"dataset": "wmt",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_0",
"rng_seed": -811149048,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"13": {
"framework": "pytorch",
"workload": "wmt",
"dataset": "wmt",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_1",
"rng_seed": -1485236731,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"14": {
"framework": "pytorch",
"workload": "wmt",
"dataset": "wmt",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_2",
"rng_seed": -439753961,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"15": {
"framework": "pytorch",
"workload": "librispeech_deepspeech",
"dataset": "librispeech",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_0",
"rng_seed": -1459326687,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"16": {
"framework": "pytorch",
"workload": "librispeech_deepspeech",
"dataset": "librispeech",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_1",
"rng_seed": 1889675898,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"17": {
"framework": "pytorch",
"workload": "librispeech_deepspeech",
"dataset": "librispeech",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_2",
"rng_seed": -1297403039,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"18": {
"framework": "pytorch",
"workload": "criteo1tb",
"dataset": "criteo1tb",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_0",
"rng_seed": -1790695410,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"19": {
"framework": "pytorch",
"workload": "criteo1tb",
"dataset": "criteo1tb",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_1",
"rng_seed": -816806699,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"20": {
"framework": "pytorch",
"workload": "criteo1tb",
"dataset": "criteo1tb",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_2",
"rng_seed": 1704852417,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"21": {
"framework": "pytorch",
"workload": "librispeech_conformer",
"dataset": "librispeech",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_0",
"rng_seed": 1605670948,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"22": {
"framework": "pytorch",
"workload": "librispeech_conformer",
"dataset": "librispeech",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_1",
"rng_seed": -1323816683,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"23": {
"framework": "pytorch",
"workload": "librispeech_conformer",
"dataset": "librispeech",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_2",
"rng_seed": -1881486829,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"24": {
"framework": "pytorch",
"workload": "finewebedu_lm",
"dataset": "fineweb_edu_10B",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_0",
"rng_seed": -304430747,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"25": {
"framework": "pytorch",
"workload": "finewebedu_lm",
"dataset": "fineweb_edu_10B",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_1",
"rng_seed": -912336586,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
},
"26": {
"framework": "pytorch",
"workload": "finewebedu_lm",
"dataset": "fineweb_edu_10B",
"submission_path": "submissions_algorithms/submissions/self_tuning/schedule_free_adamw_v2/submission.py",
"experiment_dir": "submissions_a100/schedule_free_adamw_v2/study_2",
"rng_seed": 1970089239,
"tuning_ruleset": "self",
"num_tuning_trials": 1,
"max_global_steps": 10
}
}
Loading
Loading