-
Notifications
You must be signed in to change notification settings - Fork 345
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Question - how long does training the TSP model take? #40
Comments
Hi and thanks for trying the code! From the top of my head, training a single epoch with default settings should take around 5 minutes on a 1080Ti GPU. This does not sound as if you are actually using the GPU. Please verify GPU usage through the `nvidia-smi' command (not sure whether this exists under windows) and check that tensors are actually moved to the GPU (e.g. add some printing statements printing {tensor}.device). |
Hi! Thanks for your quick replay.
|
Hi!I have the same problem as you. Have you solved it? |
I have the same probnelm, too.... |
Perhaps you can decrease the magnitude of '--epoch_size', adjust from 1280000 to 128000/12800/1280, the time consuming of each epoch will be greatly improved. |
Same here. |
Sorry this caused trouble, for people running into this please have a look at #11, I'll copy it here for reference: Hi, I had the same issue, and here is how I solved it: |
Hi!
I'm executing the following command to train a TSP model:
python run.py --graph_size 20 --baseline rollout --run_name 'tsp20_rollout'
i've run the code for 1 epocs and it took 10 hours. Is this expected?
I'm running the code under windows on a msi laptop using GPU and cuda enable.
Thanks in advance!
The text was updated successfully, but these errors were encountered: