Skip to content

Commit c9c08dc

Browse files
authored
remove outdated migration info, fix example command
Updated README to reflect support for regression tasks and modified developer notes regarding data migration.
1 parent 399e237 commit c9c08dc

File tree

1 file changed

+2
-19
lines changed

1 file changed

+2
-19
lines changed

README.md

Lines changed: 2 additions & 19 deletions
Original file line numberDiff line numberDiff line change
@@ -5,24 +5,7 @@ The library emphasizes the incorporation of the semantic qualities of the ontolo
55

66
## News
77

8-
We now support regression tasks!
9-
10-
## Note for developers
11-
12-
If you have used ChEBai before PR #39, the file structure in which your ChEBI-data is saved has changed. This means that
13-
datasets will be freshly generated. The data however is the same. If you want to keep the old data (including the old
14-
splits), you can use a migration script. It copies the old data to the new location for a specific ChEBI class
15-
(including chebi version and other parameters). The script can be called by specifying the data module from a config
16-
```
17-
python chebai/preprocessing/migration/chebi_data_migration.py migrate --datamodule=[path-to-data-config]
18-
```
19-
or by specifying the class name (e.g. `ChEBIOver50`) and arguments separately
20-
```
21-
python chebai/preprocessing/migration/chebi_data_migration.py migrate --class_name=[data-class] [--chebi_version=[version]]
22-
```
23-
The new dataset will by default generate random data splits (with a given seed).
24-
To reuse a fixed data split, you have to provide the path of the csv file generated during the migration:
25-
`--data.init_args.splits_file_path=[path-to-processed_data]/splits.csv`
8+
Starting in version 1.1, we support regression tasks!
269

2710
## Installation
2811

@@ -76,7 +59,7 @@ python -m chebai fit --trainer=configs/training/default_trainer.yml --model=conf
7659
```
7760
A command with additional options may look like this:
7861
```
79-
python3 -m chebai fit --trainer=configs/training/default_trainer.yml --model=configs/model/electra.yml --model.train_metrics=configs/metrics/micro-macro-f1.yml --model.test_metrics=configs/metrics/micro-macro-f1.yml --model.val_metrics=configs/metrics/micro-macro-f1.yml --model.pretrained_checkpoint=electra_pretrained.ckpt --model.load_prefix=generator. --data=configs/data/chebi50.yml --model.criterion=configs/loss/bce.yml --data.init_args.batch_size=10 --trainer.logger.init_args.name=chebi50_bce_unweighted --data.init_args.num_workers=9 --model.pass_loss_kwargs=false --data.init_args.chebi_version=231 --data.init_args.data_limit=1000
62+
python3 -m chebai fit --trainer=configs/training/default_trainer.yml --model=configs/model/electra.yml --model.train_metrics=configs/metrics/micro-macro-f1.yml --model.test_metrics=configs/metrics/micro-macro-f1.yml --model.val_metrics=configs/metrics/micro-macro-f1.yml --model.pretrained_checkpoint=electra_pretrained.ckpt --model.load_prefix=generator. --data=configs/data/chebi/chebi50.yml --model.criterion=configs/loss/bce.yml --data.init_args.batch_size=10 --trainer.logger.init_args.name=chebi50_bce_unweighted --data.init_args.num_workers=9 --model.pass_loss_kwargs=false --data.init_args.chebi_version=231 --data.init_args.data_limit=1000
8063
```
8164

8265
### Fine-tuning for classification tasks, e.g. Toxicity prediction

0 commit comments

Comments
 (0)