You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I want to create a base docker image that will be used for training different models, it shouldn't include any training-specific files and serve as a base layer., then use it as base_image in TrainingClient().create_job().
Then I write the training code in my kubeflow notebook or from local machine:
However, train_func just gets its code copied onto the kubernetes cluster, so I can't import anything from my local modules, I can only use pip libraries that are in my base docker image, I also can't import anything from the file that train_func is in:
from kubeflow.training import TrainingClient
def train():
...
def train_func():
from my_module import dataset
x = dataset
y = train()
Unresolved import error in both cases^
Is there any way to include multiple files in TrainingClient().create_job() or TrainingClient().train() without using yaml configs and kubectl and without adding them to my docker image?
Why is this needed?
It allows not having to rebuild your docker image or create yaml configs each time you want to run a new training job
Love this feature?
Give it a 👍 We prioritize the features with most 👍
The text was updated successfully, but these errors were encountered:
Since my_model.py file will be located in the TrainJob, you should be able to run the training script.
Improve our Kubeflow Training SDK to automatically build the Docker image with your source code, and use this Docker image in the distributed training nodes (that is what Fairing was doing before: https://github.com/kubeflow/fairing). However, this will require Docker runtime to be running in your environment.
We are looking for various options on how to distribute the user's training code into TrainJob resources.
If you have any other suggestions, please let us know @u66u
What you would like to be added?
I want to create a base docker image that will be used for training different models, it shouldn't include any training-specific files and serve as a base layer., then use it as base_image in TrainingClient().create_job().
Then I write the training code in my kubeflow notebook or from local machine:
However, train_func just gets its code copied onto the kubernetes cluster, so I can't import anything from my local modules, I can only use pip libraries that are in my base docker image, I also can't import anything from the file that train_func is in:
Unresolved import error in both cases^
Is there any way to include multiple files in TrainingClient().create_job() or TrainingClient().train() without using yaml configs and kubectl and without adding them to my docker image?
Why is this needed?
It allows not having to rebuild your docker image or create yaml configs each time you want to run a new training job
Love this feature?
Give it a 👍 We prioritize the features with most 👍
The text was updated successfully, but these errors were encountered: