Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
14 changes: 7 additions & 7 deletions datasets/README.md → algoperf/datasets/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,7 @@ This document provides instructions on downloading and preparing all datasets ut
*TL;DR to download and prepare a dataset, run `dataset_setup.py`:*

```bash
python3 datasets/dataset_setup.py \
python3 algoperf/datasets/dataset_setup.py \
--data_dir=~/data \
--<dataset_name>
--<optional_flags>
Expand Down Expand Up @@ -88,7 +88,7 @@ By default, a user will be prompted before any files are deleted. If you do not
From `algorithmic-efficiency` run:

```bash
python3 datasets/dataset_setup.py \
python3 algoperf/datasets/dataset_setup.py \
--data_dir $DATA_DIR \
--ogbg
```
Expand Down Expand Up @@ -124,7 +124,7 @@ In total, it should contain 13 files (via `find -type f | wc -l`) for a total of
From `algorithmic-efficiency` run:

```bash
python3 datasets/dataset_setup.py \
python3 algoperf/datasets/dataset_setup.py \
--data_dir $DATA_DIR \
--wmt
```
Expand Down Expand Up @@ -194,7 +194,7 @@ you should get an email containing the URLS for "knee_singlecoil_train",
"knee_singlecoil_val" and "knee_singlecoil_test".

```bash
python3 datasets/dataset_setup.py \
python3 algoperf/datasets/dataset_setup.py \
--data_dir $DATA_DIR \
--fastmri \
--fastmri_knee_singlecoil_train_url '<knee_singlecoil_train_url>' \
Expand Down Expand Up @@ -235,7 +235,7 @@ The ImageNet data pipeline differs between the PyTorch and JAX workloads.
Therefore, you will have to specify the framework (either `pytorch` or `jax`) through the framework flag.

```bash
python3 datasets/dataset_setup.py \
python3 algoperf/datasets/dataset_setup.py \
--data_dir $DATA_DIR \
--imagenet \
--temp_dir $DATA_DIR/tmp \
Expand Down Expand Up @@ -349,7 +349,7 @@ In total, it should contain 20 files (via `find -type f | wc -l`) for a total of
### Criteo1TB

```bash
python3 datasets/dataset_setup.py \
python3 algoperf/datasets/dataset_setup.py \
--data_dir $DATA_DIR \
--temp_dir $DATA_DIR/tmp \
--criteo1tb
Expand Down Expand Up @@ -378,7 +378,7 @@ In total, it should contain 885 files (via `find -type f | wc -l`) for a total o
To download, train a tokenizer and preprocess the librispeech dataset:

```bash
python3 datasets/dataset_setup.py \
python3 algoperf/datasets/dataset_setup.py \
--data_dir $DATA_DIR \
--temp_dir $DATA_DIR/tmp \
--librispeech
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -56,7 +56,7 @@

Example command:

python3 datasets/dataset_setup.py \
python3 algoperf/datasets/dataset_setup.py \
--data_dir=~/data \
--temp_dir=/tmp/mlcommons_data
--imagenet \
Expand All @@ -73,8 +73,8 @@

from algoperf.workloads.wmt import tokenizer
from algoperf.workloads.wmt.input_pipeline import normalize_feature_names
from datasets import librispeech_preprocess
from datasets import librispeech_tokenizer
from algoperf.datasets import librispeech_preprocess
from algoperf.datasets import librispeech_tokenizer

import functools
import os
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -14,7 +14,7 @@
from absl import logging
from pydub import AudioSegment

from datasets import librispeech_tokenizer
from algoperf.datasets import librispeech_tokenizer

gfile = tf.io.gfile
copy = tf.io.gfile.copy
Expand Down
Loading