You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi! I'm working on anomaly detection on AIS data and I'm trying to replicate the results from the GeoTrackNet article.
While training the VRNN on my local machine I get stuck because of an Out of Memory error. My GPU is RTX3060 for laptops with 6144MiB of memory.
I was wondering on what machine did you run the model? What was the GPU?
What part of the code should I change in order to lower the batch size?
The text was updated successfully, but these errors were encountered:
Then simply add it as a config parameter as shown below.
with tf.train.MonitoredTrainingSession(master=config.master,
is_chief=config.task == 0,
config=gpu_config, #NEW PARAMETER ADDED
hooks=[log_hook],
checkpoint_dir=config.logdir,
save_checkpoint_secs=120,
save_summaries_steps=config.summarize_every,
log_step_count_steps=config.summarize_every) as sess:
This will make sure you use a limited amount of your memory while training. Run your training after this while keeping an eye on your task manager (GPU usage) to see weather the process still peaks or not.
Let me know if it works for you after these fixes. Because I am experiencing similar issues and I am using RTX3090 with Cuda 11.7.
Training still crashes even after fixing my "Out of memory" issues. :/
Hi! I'm working on anomaly detection on AIS data and I'm trying to replicate the results from the GeoTrackNet article.
While training the VRNN on my local machine I get stuck because of an Out of Memory error. My GPU is RTX3060 for laptops with 6144MiB of memory.
The text was updated successfully, but these errors were encountered: