Skip to content

Latest commit

 

History

History
191 lines (126 loc) · 6.77 KB

README.md

File metadata and controls

191 lines (126 loc) · 6.77 KB

RL-CloudSimPlus

Codacy Badge GPLv3 License

Installation

1. Install Docker

https://docs.docker.com/get-docker/

Warning

If you install Docker Desktop for MacOS, make sure you are giving enough memory in your containers by going to Settings... > Resources and increasing the Memory Limit

2. Install Docker Compose

https://docs.docker.com/compose/install/

3. Install Java OpenJDK 21

  • For Debian, install the openjdk-21-jdk and openjdk-21-jre packages
sudo apt install openjdk-21-jdk openjdk-21-jre
  • For MacOS, use brew
brew install openjdk@21

4. Set the JAVA_HOME environment variable to the right path

Important

The exact path may vary (distro, arch, etc.)

  • For Linux
export JAVA_HOME=/usr/lib/jvm/java-21-openjdk-<arch>
  • For MacOS
    • Follow brew instructions or visit here for more details.

Note

This command will make your default Java version 21.

Building the TensorBoard, Gateway, and RL Manager images

make build

Note

It is often useful to rebuild images one at a time, especially when a change is made only in a specific application part. For example, when we change the gateway code, we must rebuild the image before running the application. Optionally, when building, we can specify the log level of the messages printed in stdout and a log file. If the log level environment variable is not set, the log level is set to INFO by default.

[LOG_LEVEL=TRACE|DEBUG|INFO|WARN|ERROR] make build-gateway

Starting the TensorBoard dashboard

The project consists of three docker images. The gateway and manager images contain the main application and are the docker compose services we need for every experiment we want to run. The TensorBoard image is the UI endpoint and helps us keep track of the experiment's progress. Because we do not want to shut down the visualization dashboard every time we want to stop an experiment, the TensorBoard image is not a docker compose service and can be started as a standalone docker container by using the following command (TODO: consider changing it):

make run-tensorboard

Note

You can check that the TensorBoard dashboard is running by visiting http://localhost.

Editing the experiment configuration file

To run an experiment, first edit the configuration file config.yml.

The configuration file contains two sections: the 'common' section and the 'experiment' section. This is because we may run multiple different experiments in parallel.

  • The parameters that all experiments have in common are specified under the common section, and those that are unique among the experiments are defined under the experiment_{id} section
    • If a parameter is specified in both the common and experiment sections, the common one is ignored, and the experiment one takes effect.
  • To run multiple experiments in parallel, add as many experiment areas as you want, specifying the corresponding parameters for each experiment.
  • Each experiment should have a unique experiment id, and each section should be written as experiment_{id}. For simplicity, use ids starting by 1 and increment by 1.

There are three experiment modes: train, transfer, and test. When transfer or test modes are specified, an additional 'train_model_dir' key for an experiment should be defined, with the directory name in which the trained agent model should be used.

Running an experiment

After editing the configuration file, run the following command to start the experiment(s).

make run

CUDA GPU support

There is also support to run the experiments in CUDA GPUs.

  • You need to have CUDA and nvidia-container-toolkit installed on your system.
  • Restart the docker daemon if you just downloaded the cuda-container-toolkit.

Stopping the application

If you want to stop the application and clear all the dangling containers and volumes, run the following command:

make stop

Acknowledgements