From a563aeb1270e1941115694acf952a338ed1fa285 Mon Sep 17 00:00:00 2001 From: JMGaljaard Date: Mon, 26 Sep 2022 11:15:41 +0200 Subject: [PATCH] Remove debugging ret parameter --- fltk/core/distributed/orchestrator.py | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) diff --git a/fltk/core/distributed/orchestrator.py b/fltk/core/distributed/orchestrator.py index 30f671b4..6235ab03 100644 --- a/fltk/core/distributed/orchestrator.py +++ b/fltk/core/distributed/orchestrator.py @@ -212,7 +212,7 @@ def _create_config_maps(self, config_maps: Dict[str, V1ConfigMap]) -> None: self._v1.create_namespaced_config_map(self._config.cluster_config.namespace, config_map) - def wait_for_jobs_to_complete(self, ret=False, others: Optional[List[str]] = None): + def wait_for_jobs_to_complete(self, others: Optional[List[str]] = None): """ Function to wait for all tasks to complete. This allows to wait for all the resources to free-up after running an experiment. Thereby allowing for running multiple experiments on a single cluster, without letting @@ -322,7 +322,7 @@ def run(self, clear: bool = False, if wait_historical: curr_jobs = self._client.get(namespace="test") jobs = [job['metadata']['name'] for job in curr_jobs['items']] - self.wait_for_jobs_to_complete(ret=False, others=jobs) + self.wait_for_jobs_to_complete(others=jobs) start_time = time.time() if clear: @@ -365,9 +365,9 @@ def run(self, clear: bool = False, # Either wait to complete, or continue. Note that the orchestrator currently does not support scaling # experiments up or down. if not self._config.cluster_config.orchestrator.parallel_execution: - self.wait_for_jobs_to_complete(ret=True) + self.wait_for_jobs_to_complete() if self._config.cluster_config.orchestrator.parallel_execution: - self.wait_for_jobs_to_complete(ret=False) + self.wait_for_jobs_to_complete() logging.info('Experiment completed.') # Stop experiment self.stop()