-
Notifications
You must be signed in to change notification settings - Fork 12
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
I achieved the program, but now the results are not so good ,maybe I need try to train the model using different params . thanks #4
Comments
Hi. Would you like to show us some details about your project? Are you using defect-gen on your dataset? |
when I use the defect gen to my own dataset, the results are abnomal. but when I use huggingface data ,the results are nomal. |
hello, AndysonYs, When I use my own dataset ,the result is abnomal,,the generated data is very different with input data . |
I use the labelme to genarate mask data,and then, transform it to converted Groundytruth mask ,but the train result is very strange。 |
I need you help, how Can I communicate with you soon. can you give me you wechat number or qq number,thank you very much. |
Maybe there is some mistakes when I make my own train mask, so I need you help maybe. |
I Rent cloud server with 4 RTX3090 ,However,the train speed with one RTX3090 is faster than four RTX3090? |
Are you using the 2 stage defect-gen (combination of large and small receptive field models) for your data? Firstly, you can try using the large-receptive field model only and validate its performance. If the 2 stage defect-gen fails but the large model works well, that means the then you should adjust the hyper-param of the switch point of our 2 models and use more large receptive field model. |
To adapt to your own dataset, you may change some architecture hyper-params. For example, if your data has a higher resolution, you need to add more down-sampling layers in the diffusion unet. |
Could you tell me some features of your dataset? like the amount of data, the resolution, the num of defects. you can also post some examples here if available. |
It seems weird to me. Did you just change the nproc_per_node? if you change it from 1 to 4 without modifying any other hyper-params, then it means you train it for 4 times long. |
CUDA_VISIBLE_DEVICES="0,1,2,3" |
yes,I use the 2 stage defect-gen for my data, the hyper-param of the switch point of our 2 models mean the param --step_inference 400 ,this one? |
From the results you provided, it seems like the smaller model has too much involvement, which may disrupt the overall geometry of the image. I think you should start with the large receptive model only (exclude the small model, you can do that by commenting out the small model and setting the step_inference to 0, meaning you are only using the large model for inference) first. After verifying the image quality, you can start tuning the switching step by adjusting the same parameter. The switching parameter may works different in your dataset than ours. |
when I just use the large reception model to infer, The result is also not so good ,just like the image I supply above. Why |
I think my data type just like you data screw thread. from your paper,it shows the defect-gen result is so amazing,,but my experiment shows a bad result, now what should i do the achieve the the result like, dear author ,I need you help ,please help me ,thank you very much. |
No description provided.
The text was updated successfully, but these errors were encountered: