You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This means that any model that contains more than 100 files gets cut off at that arbitrary file and causes the model to only be partially copied and then subsequently fails at runtime. For example, when using a model like the argos-translate model with many languages you can exceed the 100 file limit.
1.7132796006805687e+09 DEBUG Triton Adapter.Triton Adapter Server found objects to download {"type": "s3", "cacheKey": "s3|xyz", "path": "translate", "count": 100}
This limit seems arbitrary and should be made configurable.
The text was updated successfully, but these errors were encountered:
Any updates on this issue? This has been causing problems for us when loading models from S3.
A follow up question as well:
Are there any technical limitations as to why 100 was original chosen? Or would making this configurable, and using a value of say "300" cause any issues?
We're considering contributing a fix, but just wanted to ask if there are any technical limitations surrounding it.
Currently the S3 downloader has a hardcoded limit of 100 files (see: https://github.com/kserve/modelmesh-runtime-adapter/blob/2d5bb69e9ed19efd74fbe6f8b76ec2e970702e3c/pullman/storageproviders/s3/downloader.go#L79C3-L79C27).
This means that any model that contains more than 100 files gets cut off at that arbitrary file and causes the model to only be partially copied and then subsequently fails at runtime. For example, when using a model like the argos-translate model with many languages you can exceed the 100 file limit.
This limit seems arbitrary and should be made configurable.
The text was updated successfully, but these errors were encountered: