-
-
Notifications
You must be signed in to change notification settings - Fork 10.9k
DOC: fix ctypes example #26989
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
DOC: fix ctypes example #26989
Conversation
Might be part of it. Looks like |
087e878
to
de8528b
Compare
The circleci testing uses |
I don't have any context for why the doctests are in the benchmark test runner rather than running on circleci. @ev-br do you have context on this? Or why the doc builder is still using refguide-check.py? Maybe it should be migrated to use the new doctest runner? Or did we intentionally leave both running on CI? |
My understanding is that the students mostly worked on windows. |
Thanks Nathan. |
@charris , the students for now are using a linux remote machine on Nebari for anything that involves buliding NumPy and/or running tests. I personally checked each example using the following:
I know there has been a change to the doctesting process recently, which could be the reason for the fail. I know for a fact that zero output from any of our AI examples is hallucinated. The output never comes from AI, as I learned quite quickly that the AI output is completely unreliable. We first strip all generated output, and then run the AI generated (quite reliable) code against the current dev build and insert that output appropriately where needed. One side effect of the AI gen process I've followed is that occasionally we encountered instances where the output from valid working code does not match what the doctester expects. The issue is reproducible in various environments, and doesn't seem to be dependent on platform. This happened with one of the |
I don't think Sorry for assuming what I was seeing must have been hallucinated. |
Will do. The The AI generated "output" is quite hilarious sometimes. It's ability to count to 10 is dismal. But it does a great job of generating working code (just don't trust the output). |
@bmwoodruff
You probably don't need all of those if you make sure to just test the relevant module. |
I'm getting the same errors you guys are seeing using When i run our AI parser code (just did a bit ago), against the current dev build of Numpy, I get the following: AI Generated code log:
Here's the cleaned up output that should have been in the PR.
The cleaned code output matches the images on the review I did of this PR prior to submission, as well as the initial submission of #26827. However, the actual code in the merged version version of the PR matches the garbage AI output (a force pushed rebase on main hides the history here). @otieno-juma , did something happen when you rebased on main? My best guess is that when when @otieno-juma did a rebase, he may have reverted to something that contained the original code. I did not confirm this rebase before @charris saw it (not sure how to implement that into our workflow once the code hits numpy/numpy as PR). As the images match the cleaned output, but the merged code matches the AI output, then this seems like human error. I'm sorry that I missed that. @otieno-juma , when you rebase the other examples, please remember to run all tests (use |
@ngoldbaum, Sorry that you needed to assume this. You were correct here, as somehow in the final steps of the review process the wrong code got put back into the PR. We tried to prevent hallucinations from hitting the code base.... |
@ngoldbaum two things:
|
I'm not sure if the problem with
c_int32
is because of an AI hallucination or because of cross-platform issues, so I switched it toint8
which should be equivalent toc_byte
on all platforms.The
ctypes.Structure
that was there before was a hallucination I think?