Skip to content

Commit

Permalink
Feature/resolve issues (#19)
Browse files Browse the repository at this point in the history
* Resolve issues

* Formatting

* Resolve comments

* Fix check
  • Loading branch information
azoz01 authored Dec 18, 2023
1 parent d70e2e6 commit 079d4c0
Show file tree
Hide file tree
Showing 22 changed files with 202 additions and 239 deletions.
6 changes: 0 additions & 6 deletions .github/workflows/code_check.yml
Original file line number Diff line number Diff line change
Expand Up @@ -27,18 +27,12 @@ jobs:
- name: Check black test code
run: black test --line-length=100

- name: Check black bin code
run: black bin --line-length=100

- name: Check flake8 source code
run: flake8 liltab --max-line-length=100

- name: Check flake8 test code
run: flake8 test --max-line-length=100

- name: Check flake8 bin code
run: flake8 bin --max-line-length=100

- name: Run test
run: |
export PYTHONPATH=`pwd`
Expand Down
2 changes: 1 addition & 1 deletion README.md
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,7 @@ HeterogenousAttributesNetworkTrainer(
gradient_clipping=False,
learning_rate=1e-3,
weight_decay=1e-4,
early_stopping=True,
early_stopping_intervals=100,
file_logger=True,
tb_logger=True,
model_checkpoints=True,
Expand Down
128 changes: 0 additions & 128 deletions bin/train.py

This file was deleted.

2 changes: 1 addition & 1 deletion config/01_synthetic_data_experiment_config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ learning_rate: 0.001
weight_decay: 0
batch_size: 256
gradient_clipping: False
early_stopping: True
early_stopping_intervals: 100

support_size: 5
query_size: 27
Expand Down
2 changes: 1 addition & 1 deletion config/02_openml_data_experiment_config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ learning_rate: 0.001
weight_decay: 0.0
batch_size: 37
gradient_clipping: False
early_stopping: True
early_stopping_intervals: 100

support_size: 3
query_size: 29
Expand Down
2 changes: 1 addition & 1 deletion config/03_openml_clf_data_experiment_config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ learning_rate: 0.001
weight_decay: 0
batch_size: 37
gradient_clipping: False
early_stopping: True
early_stopping_intervals: 100

support_size: 3
query_size: 29
Expand Down
2 changes: 1 addition & 1 deletion config/04_same_domain_experiment_config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ learning_rate: 0.001
weight_decay: 0.0001
batch_size: 37
gradient_clipping: False
early_stopping: True
early_stopping_intervals: 100

support_size: 3
query_size: 29
Expand Down
2 changes: 1 addition & 1 deletion config/05_new_classes_experiment_config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -4,7 +4,7 @@ learning_rate: 0.0001
weight_decay: 0
batch_size: 16
gradient_clipping: False
early_stopping: True
early_stopping_intervals: 100

support_size: 3
query_size: 29
Expand Down
2 changes: 1 addition & 1 deletion config/06_big_data_experiment_config.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,7 @@ learning_rate: 0.001
weight_decay: 0
batch_size: 16
gradient_clipping: False
early_stopping: True
early_stopping_intervals: 100

support_size: 3
query_size: 29
Expand Down
2 changes: 1 addition & 1 deletion experiments/01_synthetic.py
Original file line number Diff line number Diff line change
Expand Up @@ -71,7 +71,7 @@ def main():
gradient_clipping=config["gradient_clipping"],
learning_rate=config["learning_rate"],
weight_decay=config["weight_decay"],
early_stopping=config["early_stopping"],
early_stopping_intervals=config["early_stopping_intervals"],
file_logger=True,
tb_logger=True,
model_checkpoints=True,
Expand Down
2 changes: 1 addition & 1 deletion experiments/02_openml.py
Original file line number Diff line number Diff line change
Expand Up @@ -74,7 +74,7 @@ def main():
gradient_clipping=config["gradient_clipping"],
learning_rate=config["learning_rate"],
weight_decay=config["weight_decay"],
early_stopping=config["early_stopping"],
early_stopping_intervals=config["early_stopping_intervals"],
file_logger=True,
tb_logger=True,
model_checkpoints=True,
Expand Down
8 changes: 4 additions & 4 deletions experiments/03_openml_clf.py
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ def main():
train_loader = ComposedDataLoaderFactory.create_composed_dataloader_from_path(
Path(config["train_data_path"]),
PandasDataset,
{"encode_categorical_target": True},
{"encode_categorical_response": True},
FewShotDataLoader,
{"support_size": config["support_size"], "query_size": config["query_size"]},
ComposedDataLoader,
Expand All @@ -43,7 +43,7 @@ def main():
val_loader = ComposedDataLoaderFactory.create_composed_dataloader_from_path(
Path(config["val_data_path"]),
PandasDataset,
{"encode_categorical_target": True},
{"encode_categorical_response": True},
FewShotDataLoader,
{"support_size": config["support_size"], "query_size": config["query_size"]},
RepeatableOutputComposedDataLoader,
Expand All @@ -52,7 +52,7 @@ def main():
test_loader = ComposedDataLoaderFactory.create_composed_dataloader_from_path(
Path(config["test_data_path"]),
PandasDataset,
{"encode_categorical_target": True},
{"encode_categorical_response": True},
FewShotDataLoader,
{"support_size": config["support_size"], "query_size": config["query_size"]},
RepeatableOutputComposedDataLoader,
Expand All @@ -75,7 +75,7 @@ def main():
gradient_clipping=config["gradient_clipping"],
learning_rate=config["learning_rate"],
weight_decay=config["weight_decay"],
early_stopping=config["early_stopping"],
early_stopping_intervals=config["early_stopping_intervals"],
loss=nn.CrossEntropyLoss(),
file_logger=True,
tb_logger=True,
Expand Down
8 changes: 4 additions & 4 deletions experiments/04_same_domain.py
Original file line number Diff line number Diff line change
Expand Up @@ -30,7 +30,7 @@ def main():
train_loader = ComposedDataLoaderFactory.create_composed_dataloader_from_path(
Path(config["train_data_path"]),
PandasDataset,
{"encode_categorical_target": True},
{"encode_categorical_response": True},
FewShotDataLoader,
{"support_size": config["support_size"], "query_size": config["query_size"]},
ComposedDataLoader,
Expand All @@ -39,7 +39,7 @@ def main():
val_loader = ComposedDataLoaderFactory.create_composed_dataloader_from_path(
Path(config["val_data_path"]),
PandasDataset,
{"encode_categorical_target": True},
{"encode_categorical_response": True},
FewShotDataLoader,
{"support_size": config["support_size"], "query_size": config["query_size"]},
RepeatableOutputComposedDataLoader,
Expand All @@ -48,7 +48,7 @@ def main():
test_loader = ComposedDataLoaderFactory.create_composed_dataloader_from_path(
Path(config["test_data_path"]),
PandasDataset,
{"encode_categorical_target": True},
{"encode_categorical_response": True},
FewShotDataLoader,
{"support_size": config["support_size"], "query_size": config["query_size"]},
RepeatableOutputComposedDataLoader,
Expand All @@ -71,7 +71,7 @@ def main():
gradient_clipping=config["gradient_clipping"],
learning_rate=config["learning_rate"],
weight_decay=config["weight_decay"],
early_stopping=config["early_stopping"],
early_stopping_intervals=config["early_stopping_intervals"],
file_logger=True,
tb_logger=True,
model_checkpoints=True,
Expand Down
8 changes: 4 additions & 4 deletions experiments/05_new_classes.py
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@ def main():
train_loader = ComposedDataLoaderFactory.create_composed_dataloader_from_path(
Path(config["train_data_path"]),
PandasDataset,
{"encode_categorical_target": True},
{"encode_categorical_response": True},
FewShotDataLoader,
{"support_size": config["support_size"], "query_size": config["query_size"]},
ComposedDataLoader,
Expand All @@ -40,7 +40,7 @@ def main():
val_loader = ComposedDataLoaderFactory.create_composed_dataloader_from_path(
Path(config["val_data_path"]),
PandasDataset,
{"encode_categorical_target": True},
{"encode_categorical_response": True},
FewShotDataLoader,
{"support_size": config["support_size"], "query_size": config["query_size"]},
RepeatableOutputComposedDataLoader,
Expand All @@ -49,7 +49,7 @@ def main():
test_loader = ComposedDataLoaderFactory.create_composed_dataloader_from_path(
Path(config["test_data_path"]),
PandasDataset,
{"encode_categorical_target": True},
{"encode_categorical_response": True},
FewShotDataLoader,
{"support_size": config["support_size"], "query_size": config["query_size"]},
RepeatableOutputComposedDataLoader,
Expand All @@ -72,7 +72,7 @@ def main():
gradient_clipping=config["gradient_clipping"],
learning_rate=config["learning_rate"],
weight_decay=config["weight_decay"],
early_stopping=config["early_stopping"],
early_stopping_intervals=config["early_stopping_intervals"],
loss=nn.CrossEntropyLoss(),
file_logger=True,
tb_logger=True,
Expand Down
8 changes: 4 additions & 4 deletions experiments/06_big_data.py
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,7 @@ def main():
train_loader = ComposedDataLoaderFactory.create_composed_dataloader_from_path(
Path(config["train_data_path"]),
PandasDataset,
{"encode_categorical_target": True},
{"encode_categorical_response": True},
FewShotDataLoader,
{"support_size": config["support_size"], "query_size": config["query_size"]},
ComposedDataLoader,
Expand All @@ -40,7 +40,7 @@ def main():
val_loader = ComposedDataLoaderFactory.create_composed_dataloader_from_path(
Path(config["val_data_path"]),
PandasDataset,
{"encode_categorical_target": True},
{"encode_categorical_response": True},
FewShotDataLoader,
{"support_size": config["support_size"], "query_size": config["query_size"]},
RepeatableOutputComposedDataLoader,
Expand All @@ -49,7 +49,7 @@ def main():
test_loader = ComposedDataLoaderFactory.create_composed_dataloader_from_path(
Path(config["test_data_path"]),
PandasDataset,
{"encode_categorical_target": True},
{"encode_categorical_response": True},
FewShotDataLoader,
{"support_size": config["support_size"], "query_size": config["query_size"]},
RepeatableOutputComposedDataLoader,
Expand All @@ -72,7 +72,7 @@ def main():
gradient_clipping=config["gradient_clipping"],
learning_rate=config["learning_rate"],
weight_decay=config["weight_decay"],
early_stopping=config["early_stopping"],
early_stopping_intervals=config["early_stopping_intervals"],
loss=nn.CrossEntropyLoss(),
file_logger=True,
tb_logger=True,
Expand Down
Loading

0 comments on commit 079d4c0

Please sign in to comment.