-
-
Notifications
You must be signed in to change notification settings - Fork 50
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Add triton dependency, readd cudss and cusparselt, mention dev speedup tricks in the README #309
Conversation
Hi! This is the friendly automated conda-forge-linting service. I just wanted to let you know that I linted all conda-recipes in your PR ( I do have some suggestions for making it better though... For recipe/meta.yaml:
This message was generated by GitHub Actions workflow run https://github.com/conda-forge/conda-forge-webservices/actions/runs/12601832807. Examine the logs at this URL for more detail. |
Started only a single job for now (x64+CUDA+generic). If this one passes we can run the rest (modulo other changes from review). |
So the CI here for x64+CUDA+generic passed. I think you could push whatever new commits you have @mgorny and we can do a full CI run. |
@h-vetinari, I think the latest build of libabseil is now broken on OSX. See e.g. https://dev.azure.com/conda-forge/feedstock-builds/_build/results?buildId=1127302&view=logs&j=1dfee543-4d72-5193-7012-7358e3fc4acd&t=344bb75a-1253-525e-defa-7c6fedd7579f&l=2594 |
Hmm, or maybe protobuf needs to be rebuilt. Though it's still weird that ABI changed on rebuild. |
It's something to do with clang 17 to clang 18. I've looked at the breaking changes at the time, and nothing stood out in particular. I guess abseil does unusual things, relatively speaking... In any case, I'm hoping that once we've rebuilt the core protobuf stack (conda-forge/libprotobuf-feedstock#244 + maybe one for grpc), the issue will resolve itself... (edit: or maybe not: arrow also seems to fail) |
Could you trigger the Linux CI builds, please? |
ccbaebb
to
4de17af
Compare
Was waiting for the |
…onda-forge-pinning 2024.12.30.17.35.18
I'm leaving the linux CI running for now; to fix OSX, we should pick up the changes from #310 here. |
Hmm, looks like two jobs failed with "This request was automatically failed because there were no enabled runners online to process". The other four passed. |
i clicked rerun failed jobs. that is my contribution ^_^ |
Oh, right, we don't build Triton for aarch64 (yet). Let's see if that is possible… |
Ok, so the current status is:
|
Okay, I'm going to wait for new triton builds to propagate, then merge main and #310 here to trigger new CI. |
@h-vetinari, I think we can limit CI to the aarch64 + CUDA job. There were no other changes, and everything else passed the previous time. |
Sigh, I hope that the failing build was a runner GPU problem and not something with the package. |
It failed at ~55% of the test through, with:
detailed logs follow:
|
Could you rerun failing builds, please? I think it was a fluke — there's almost no changes (I've only merged |
There's a limitation in the UI that I can only restart all failed jobs (includes cancelled jobs), or one. So I prioritised the aarch GPU build first, but now restarted the previous failure too. |
Thanks. So all passed, except for the 3 cancelled jobs. Not sure if you want to rerun them, or just assume they didn't change since the last run. |
Thank you! |
Failure of
Perhaps also due to resource issues? |
Hrm, no clue. Could it have OOM-ed perhaps? |
Do you want me to try addressing it separately, or wait and see if we can get kineto passing first, and perhaps the same solution will work for both cases? |
You can keep working on it of course. In the meantime I'll see if I can nurse the CI on main to success by restarting (but the queue seems long at the moment) |
Checklist
0
(if the version changed)conda-smithy
(Use the phrase@conda-forge-admin, please rerender
in a comment in this PR for automated rerendering)