-
Notifications
You must be signed in to change notification settings - Fork 5
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
possible for distribute training #3
Comments
I'm not expert in distributed training system. But I've tried (tested) some distributed training before so... You could try some other strategies in tf.distribute.experimental besides tf.distribute.MirroredStrategy. But I recommend to put your GPUs in one machine and use tf.distribute.MirroredStrategy. Because other strategy like ParameterServerStrategy, it is a bit difficult to set up then MirroredStrategy. And the environment that how multiple machines are organized (like network settings) affects the performance. If you want to try, follow the guide in old contrib readme. See Multi-worker Training section. |
First of all,thank you for your help.
from
in the code begining. |
Stupid Me,I find I use global system proxy. |
if I use my own dataset,should I put the face/ in datasets/ ??? |
I have 3 machine which each have one 2080ti。Do you have some suggestions about training method,dataset loader (from tfrecords)?
The text was updated successfully, but these errors were encountered: