Fighting Fire with Fire: Scalable Oral Exams
Key topics
As educators explore using AI to administer scalable oral exams, a lively debate erupts about the role of AI in education, with some commenters questioning whether AI could - or should - replace human teaching altogether. While some, like bagrow, wonder if AI could teach entire courses, others, like alwa, caution that AI excels at "how" but falters on "what" to do, highlighting the importance of human judgment. The discussion takes a philosophical turn as commenters like semilin and baq ponder the desirability of a future where AI dominates, and humans are left feeling dehumanized or, conversely, freed from certain tasks. Amidst the debate, a consensus emerges that human interaction has value, with xboxnolifes dryly noting that being grilled by a human is still an unappealing alternative to AI.
Snapshot generated from the HN discussion
Discussion Activity
Very active discussionFirst comment
54m
Peak period
136
0-12h
Avg / period
32
Based on 160 loaded comments
Key moments
- 01Story posted
Jan 2, 2026 at 1:18 PM EST
7 days ago
Step 01 - 02First comment
Jan 2, 2026 at 2:12 PM EST
54m after posting
Step 02 - 03Peak activity
136 comments in 0-12h
Hottest window of the conversation
Step 03 - 04Latest activity
Jan 6, 2026 at 7:45 PM EST
3d ago
Step 04
Generating AI Summary...
Analyzing up to 500 comments to identify key contributors and discussion patterns
Want the full context?
Jump to the original sources
Read the primary article or dive into the live Hacker News thread when you're ready.
Also, with all the progress in video gen, what does recording the webcam really do?
AI’s got plenty of “how” (to do stuff) but much less “what” (to do)—and good judgment as to “what” takes a working knowledge of “how,” even if it’s not you who will be directly doing the work.
In that sense, to me at least, the ultimate goal isn’t the immediate task at hand, it’s the wisdom and discernment that emerges from doing a lot of them.
...but OTOH if cheating is so easy it's impossible to resist and when everyone cheats honest students are the ones getting all the bad grades, what else can you do?
If this is the only way to keep the existing approach working, it feels like the only real solution for education is something radically different, perhaps without assessment at all
I did, however, pepper my answers with statements like "it is widely accepted that the industry standard for this concept is X". I would feel bad lying to a human, but I feel no such remorse with an AI.
When I was doing a lot of hiring we offered the option (don’t roast me, it was an alternative they could choose if they wanted) of a take-home problem they could do on their own. It was reasonably short, like the kind of problem an experienced developer could do in 10-15 minutes and then add some polish, documentation, and submit it in under an hour.
Even though I told candidates that we’d discuss their submission as part of the next step, we would still get candidates submitting solutions that seemed entirely foreign to them a day later. This was on the cusp of LLMs being useful, so I think a lot of solutions were coming from people’s friends or copied from something on the internet without much thought.
Now that LLMs are both useful and well known, the temptation to cheat with them is huge. For various reasons I think students and applicants see using LLMs as not-cheating in the same situations where they wouldn’t feel comfortable copying answers from a friend. The idea is that the LLM is an available tool and therefore they should be able to use it. The obvious problem with that argument is that we’re not testing students or applicants on their abilities to use an LLM, we’re using synthetic pronouns to explore their own skills and communication.
Even some of the hiring managers I know who went all in on allowing LLMs during interviews are changing course now. The LLM-assisted interviewed were just turning into an exercise of how familiar the candidate was with the LLM being used.
I don’t really agree with some of the techniques they’re using in this article, but the problem they’re facing is very real.
You've piqued my interest!
I wonder: with a structure like this, it seems feasible to make the LLM exam itself available ahead of time, in its full authentic form.
They say the topic randomization is happening in code, and that this whole thing costs 42¢ per student. Would there be drawbacks to offering more-or-less unlimited practice runs until the student decides they’re ready for the round that counts?
I guess the extra opportunities might allow an enterprising student to find a way to game the exam, but vulnerabilities are something you’d want to fix anyway…
To the extent of wondering what value the human instructors add.
Wouldn't a written exam--or even a digital one, taken in class on school-provided machines--be almost as good?
As long as it's not a hundred person class or something, you can also have an oral component taken in small groups.
Students had and still have the option to collectively choose not to use AI to cheat. We can go back to written work at any time. And yet they continue to use it. Curious.
Individuals can't "collectively" choose anything.
This test is given to the entire class, including people who never touched AI.
Students could absolutely organize a consensus decision to not use AI. People do this all the time. How do you think human organizations continue to exist?
I know we've had historical record of people saying this for 2000 years and counting, but I suspect the future is well and truly bleak. Not because of the next generation of students, but because of the current generation of educators unable to successfully adapt to new challenges in a way that is actually beneficial to the student that it is supposed to be their duty to teach.
Wouldn't that be a fine outcome?
Ask the student to come to the exam and write something new, which is similar to what they've been working on at home but not the same. You can even let them bring what they've done at home for reference, which will help if they actually understand what they've produced to date.
If the class cost me $50? Then sure, use Dr. Slop to examine my knowledge. But this professor's school charges them $90,000 a year and over $200k to get an MBA? Hell no!
At that point what’s the value add over using YouTube videos and ChatGPT on your own?
Let me ask, how do you generally feel when you contact customer service about something and you get an AI chatbot? Now imagine the chatbot is responsible for whether you pass the course.
Adding this as an additional optional tool, though, is an excellent idea.
And universities wonder why enrollment is dropping.
In my BSc and MSc we were all basically locals who are in all aspects about the same except from the aptitude to study. In the university where I did my PhD there were much more divisions (aka diversity) in which every oral examiner would need to navigate so one group does not feel to be made preferential over another.
Don't tell me about GenZ. I had oral exams in calculus as undergrad, and our professor was intimidating. I barely passed each time when I got him as examiner, though I did reasonably well when dealing with his assistant. I could normally keep my emotions in check, but not with my professor. Though, maybe in that case the trigger was not just the tone of professor, but the sheer difference in the tone he used normally (very friendly) and at the exam time. It was absolutely unexpected at my first exam, and the repeated exposure to it didn't help. I'd say it was becoming worse with each time. Today I'd overcome such issues easily, I know some techniques today, but I didn't when I was green.
OTOH I wonder, if an AI could have such an effect on me. I can't treat AI as a human being, even if I wanted to, it is just a shitty program. I can curse a compiler refusing to accept a perfectly valid borrow of a value, so I can curse an AI making my life difficult. Mostly I have another emotional issue with AI: I tend to become impatient and even angry at AI for every small mistake it does, but this one I could overcome easily.
I wish that wasn't a thing.
Interviews are similar, but different: I'm presenting myself.
Where do we go from there? At some point soon I think this is going to have to come firmly back to real people.
Next steps are bone conduction microphones, smart glasses, earrings...
And the weeding out of anyone both honest and with social anxiety.
There are quite a lot of Amazon reviews that suggest that this is already common practice.
The current strategy is to first scan the exam with a tiny wireless shirt button camera, wait for someone on the other end to solve the exam, and then write down the solution whispered into your ear over in-ear inductive loop earphones.
We have already been there. A student asked whether they could use an app to "translate" the examiner's instructions. The app was ChatGPT, prompted to solve all questions in the conversation.
Perhaps we as humans should stop making choices which cause pain.
Why do you make choices that cause pain in yourself and others?
The real problem is students and universities have collectively bought into a "customer mindset". When they do poorly, it's always the school's fault. They're "paying customers" after-all, they're (in their mind) entitled to the degree as if it is a seamless transaction. Getting in was the hardest part for most students, so now they believe they have already proven themselves and should as a matter of routine after 3-4 years be handed their degree because they exchanged some funds. Most students would gladly accept no grades if it was possible.
Unfortunately, rather than having spines, most schools have also adopted a "the customer is always right" approach, and endlessly chase graduation numbers as a goal in and of itself and are terrified of "bad reviews."
There has been lots of handwringing around AI and cheating and what solutions are possible. Mine is actually relatively simple. University and college should get really hard again (I'm aware it was a finishing school a century ago, but the grade inflation compared to just 50 years ago is insane). Across all disciplines. Students aren't "paying for a degree", they're paying to prove that they can learn, and the only way to really prove that is to make it hard as hell and to make them care about learning in order to get to the degree - to earn it. Otherwise, as we've seen, the value of the degree becomes suspect leading to the university to become suspect as a whole.
Schools are terrified of this, but they have to start failing students and committing to it.
I graduated from a SUNY school in 2012. At the time, you could still actually go to school and work part time and get through it. Not saying it was easy by any stretch but it was possible. Tuition + living expenses were about $17/year on campus , less expensive housing was available off campus.
Now, even state schools have tuition which is only affordable through family wealth or loans. Going to university is no longer a low stakes choice - if you flunk you’re stuck with that debt forever. Not to say students aren’t responsible for understanding that when signing up, but the stakes are just a lot higher than what it used to be.
The two solutions to this are (1) as some commenters here are suggesting, give up entirely and focus only on quality of output, or (2) teach students to care about being more than appearance. Make students want to write essays. It is for their personal edification and intellectual flourishing. The benefits of this far surpass output.
Obviously this is an enormously difficult task, but let us not suppose it an unworthy one.
In reality, they cheat when a culture of cheating makes it no longer humiliating to admit you do it, and when the punishments are so lax that it becomes a risk assessment rather than an ethical judgment. Same reason companies decide to break the law when the expected cost of any law enforcement is low enough to be worth it. When I was in college, overt cheating would be expulsion with 2 (and sometimes even 1 if it was bad enough) offenses. Absolutely not worth even giving the impression of any misconduct. Now there are colleges that let student tribunals decide how to punish their classmates who cheat (with the absolutely predictable outcome)
I suppose there are other fields where the degree might be used mostly as a filtering mechanism, where cheating through graduation might get you a job doing work different than your classes anyway. However, even in those cases it's hard to break the habit of cheating your way around every difficult problem that comes your way.
As an aside, I'm surprised oral exams aren't possible at 36 students. I feel like I've taken plenty of courses with more participants and oral exams. But the break even point is probably very different from country to country.
this is also known as 'logistical nightmare', but yeah it's the only reasonable way if you want to avoid being questioned by robots.
I think the most I experienced at the physics department in Aarhus was 70ish students. 200 sounds like a big undertaking.
If you're looking for suggestions, I'd love for you to start with a problem that isn't trivially fixable.
They're even more possible if you do an oral exam only on the highest grades. That's the purpose, isn't it? To see if a good, very good, or excellent student actually knows what they're talking about. You can't spare 10 minutes to talk to each student scoring over 80% or something? Please
> And here is the delicious part: you can give the whole setup to the students and let them prepare for the exam by practicing it multiple times. Unlike traditional exams, where leaked questions are a disaster, here the questions are generated fresh each time. The more you practice, the better you get. That is... actually how learning is supposed to work.
It depends on how frequent and how in-depth you want the exams to be. How much knowledge can you test in an oral exam that would be similar to a two-hour written exam? (Especially when I remember my own experience where I would have to sketch ideas for 3/4th of the time alloted before spending the last 1/4th writing frenetically the answer I found _in extremis_).
If I were a teacher, my experience would be to sample the students. Maybe bias the sample towards students who give wrong answers, but then it could start either a good feedback loop ("I'll study because I don't want to be interrogated again in front of the class") or a bad feedback loop ("I am being picked on, it is getting worse than I can improve, I hate this and I give up")
https://sibylline.dev/articles/2025-12-31-how-agent-evals-ca...
if we want to educate people 'how people work', companies should be hiring interns and teaching them how people work. university education should be about education (duh) and deep diving into a few specialized topics, not job preparedness. AI makes this disconnect that much more obvious.
This is why we need to continue to educate humans for now and assess their knowledge without use of AI tools.
CFR 46.104 (Exempt Research):
46.104.d.1 "Research, conducted in established or commonly accepted educational settings, that specifically involves normal educational practices that are not likely to adversely impact students' opportunity to learn required educational content or the assessment of educators who provide instruction. This includes most research on regular and special education instructional strategies, and research on the effectiveness of or the comparison among instructional techniques, curricula, or classroom management methods."
https://www.ecfr.gov/current/title-45/subtitle-A/subchapter-...
So while this may have been a dick move by the instructors, it was probably legal.
> Exempt human subjects research is a specific sub-set of “research involving human subjects” that does not require ongoing IRB oversight. Research can qualify for an exemption if it is no more than minimal risk and all of the research procedures fit within one or more of the exemption categories in the federal IRB regulations. *Studies that qualify for exemption must be submitted to the IRB for review before starting the research. Pursuant to NU policy, investigators do not make their own determination as to whether a research study qualifies for an exemption — the IRB issues exemption determinations.* There is not a separate IRB application form for studies that could qualify for exemption – the appropriate protocol template for human subjects research should be filled out and submitted to the IRB in the eIRB+ system.
Most of my research is in CS Education, and I have often been able to get my studies under the Exempt status. This makes my life easier, but it's still a long arduous paperwork process. Often there are a few rounds to get the protocol right. I usually have to plan studies a whole semester in advance. The IRB does NOT like it when you decide, "Hey I just realized I collected a bunch of data, I wonder what I can do with it?" They want you to have a plan going in.
[1] https://irb.northwestern.edu/submitting-to-the-irb/types-of-...
Imagine otherwise: a teacher who wants change their final exam from a 50 item Scantron using A-D choices, to a 50 item Scantron using A-E choices, because they think having 5 choices per item is better than 4, would need to ask for IRB approval. That's not feasible, and is not what happens in the real world of academia.
It is true that local IRBs may try to add additional rules, but the NU policy you quote talks about "studies". Most IRBs would disagree that "professor playing around with grading procedures and policies" constitutes a "study".
It would be presumed exempted.
Are you a teacher or a student? If you are a teacher, you have wide latitude that a student researcher does not.
Also, if you are a teacher, doing "research about your teaching style", that's exempted.
By contrast, if you are a student, or a teacher "doing research" that's probably not exempt and must go through IRB.
Ask about any teacher, scalability is a serious issue. Students being in classes above and under their level is a serious issue. non-interactive learning, leading to rote memorization, as a result of having to choose scaling methods of learning is a serious issue. All these can be adjusted to a personal level through AI, it's trivial to do so, even.
I'm definitely not sold on the idea of oral exams through AI though. I don't even see the point, exams themselves are specifically an analysis of knowledge at one point in time. Far from ideal, we just never got anything better, how else can you measure a student's worth?
Well, now you could just run all of that student's activity in class through that AI. In the real world you don't know if someone is competent because you run an exam, you know if he is competent because he consistently shows competency. Exams are a proxy for that, you can't have a teacher looking at a student 24/7 to see they know their stuff, except now you can gather the data and parse it, what do I care if a student performs 10 exercises poorly in a specific day at a specific time if they have shown they can do perfectly well, as can be ascertained by their performance the past week?
But isn’t the whole point of a class to move from incompetent to competent?
Isn’t the poor performance on those exercises also part of their overall performance? Do you mean just that their positive work outweighs the bad work?
I went to school long before LLMs were even a Google Engineer's brianfart for the transformer paper and the way I took exams was already AI proof.
Everything hand written in pen in a proctored gymnasium. No open books. No computers or smart phones, especially ones connected to the internet. Just a department sanctioned calculator for math classes.
I wrote assembly and C++ code by hand, and it was expected to compile. No, I never got a chance to try to compile it myself before submitting it for grading. I had three hours to do the exam. Full stop. If there was a whiff of cheating, you were expelled. Do not pass go. Do not collect $200.
Cohorts for programs with a thousand initial students had less than 10 graduates. This was the norm.
You were expected to learn the gd material. The university thanks you for your donation.
I feel like i'm taking crazy pills when I read things about trying to "adapt" to AI. We already had the solution.
It is a sad world we live in.
Also, IMO oral examinations are quite powerful for detecting who is prepared and who isn't. On the down side they also help the extroverts and the confident, and you have to be careful about preventing a bias towards those.
This is true, but it is also why it is important to get an actual expert to proctor the exam. Having confidence is good and should be a plus, but if you are confident about a point that the examiner knows is completely incorrect, you may possibly put yourself in an inescapable hole, as it will be very difficult to ascertain that you actually know the other parts you were confident (much less unconfident) in.
> I wrote assembly and C++ code by hand, and it was expected to compile. No, I never got a chance to try to compile it myself before submitting it for grading.
Do you, like, really think this is the best way to assess someone's ability? Can't we find some place in between the two extremes?
Personally, I'd go with a school-provided computer with access to documentation. And no AI, except maybe (but probably not) for very high-level courses.
Lots of my tests involved writing pseudocode, or "Just write something that looks like C or Java". Don't miss the semicolon at the end of the line, but if you write "System.print()" rather than "System.out.printLn()" you might lose a single point. Maybe.
If there were specific functions you need to call, it would have a man page or similar on the test itself, or it would be the actual topic under test.
I hand wrote a bunch of SQL queries. Hand wrote code for my Systems Programming class that involved pointers. I'm not even good with pointers. I hand wrote Java for job interviews.
It's pretty rare that you need to actually test someone can memorize syntax, that's like the entire point of modern development environments.
But if you are completely unable to function without one, you might not know as much as you would hope.
The first algorithms came before the first programming languages.
Sure, it means you need to be able to run the code in your head and be able to mentally "debug" it, but that's a feature
If you could not manage these things, you washed out in the CS101 class that nearly every STEM student took. The remaining students were not brilliant, but most of them could write code to solve problems. Then you got classes that could actually teach and test that problem solving itself.
The one class where we built larger apps more akin to actual jobs, that could have been done entirely in the lab with locked down computers if need be, but the professor really didn't care if you wanted to fake the lab work, you still needed to pass the book learning for "Programming Patterns" which people really struggled with and you still needed to be able to give a "Demo" and presentation, and you still needed to demonstrate that you understood how to read some requests from a "Customer" and turn it into features and requirements and UX
Nobody cares about people sabotaging their own education except in programming because no matter how much MBAs insist that all workers are replaceable, they cannot figure out a way to actually evaluate the competency of a programmer without knowing programming. If an engineer doesn't actually understand how to evaluate static stresses on a structure, they are going to have a hard time keeping a job. Meanwhile in the world of programming, hopping around once a year is "normal" somehow, so you can make a lot of money while literally not knowing fizzbuzz. I don't think the problem is actually education.
Computer Science isn't actually about using a laptop.
This applies to prose as much as code. A computer completely changes the experience of writing, for the better.
Yes, obviously people made do with analog writing for hundreds of years, yadda yadda, I still think it's a stupid restriction.
And why is this a flex exactly? Sounds like an extortion scheme. Get sold on how you'll be taught well and become successful. Pay. Then be sent through an experience that filters so severely, only 1% of people pass.
It's like some malicious compliance take on both teaching and studying.
Mind you, I was (for some classes) tested the same way. People still cheated, and grading stringency varied. People still also forgot everything shortly after wrapping up their finals on the given subjects and moved on. People also memorized questions and compiled a solutions book, and then handed them down to next year's class.
Do you think you're just purchasing a diploma? Or do you think you're purchasing the opportunity to gain an education and potential certification that you received said education?
It's entirely possible that the University stunk at teaching 99% of it's students (about as equally possible that 99% of the students stunk at learning), but "fraud" is absolute nonsense. You're not entitled to a diploma if you fail to learn the material well enough to earn it.
If teaching was so simple that you could just tell people to go RTFM, then recite it from memory, I don't know why people are bothering with pedagogy at all. It'd seem that there's more to teaching and learning than the bare minimum, and that both parties are culpable. Doesn't sound like you disagree on that either.
> you're purchasing the opportunity to
We can swap out fraud for gambling if you like :) Sounds like an even closer analogy now that you mention!
Jokes aside though, isn't it a gamble? You gamble with yourself that you can endure and succeed or drop out / something worse. The stake is the tuition, the prize is the diploma.
Now of course, tuition is per semester (here at least, dunno elsewhere), so it's reasonable to argue that the financial investment is not quite in such jeopardy as I painted it. Not sure about the emotional investment though.
Consider the Chinese Gaokao exam, especially in its infamous historical context between the 70s and 90s. The available seats were significantly smaller than the number of applications. The exams grueling. What do you reckon, was it the people's fault for not winning essentially the unspoken lottery? Who do you think received the blame? According to a cursory search, the individual and their families (wasn't there, cannot know myself). And no, I don't think in such a tortured scheme it was the students' fault for not making the bar.
I do not! A situation where roughly 1% of the class is passing suggests that some part of the student group is failing, and also that there is likely a class design issue or a failure to appropriately vet incoming students for preparedness (among, probably, numerous other things I'm not smart enough to come up with).
And I did take issue with the "fraud" framing; apologies for not catching your tone! I think there is a chronic issue of students thinking they deserve good grades, or deserve a diploma simply for showing up, in social media and I probably read that into your comment where I shouldn't have.
> Jokes aside though, isn't it a gamble?
Not at all. If you learn the material, you pass and get a diploma. This is no more a gamble than your paycheck. However, I think that also presumes that the university accepts only students it believes are capable of passing it's courses. If you believe universities are over-accepting students (and I think the evidence says they frequently are not, in an effort to look like luxury brands, though I don't have a cite at hand), then I can see thinking the gambling analogy is correct.
Yeah, that's fine, I can definitely appreciate that angle too.
As you can probably surmise, I've had quite some struggles during my college years specifically, hence my angle of concern. It used to be the other way around, I was doing very well prior to college, and would always find people's complaints to be just excuses. But then stuff happened, and I was never really the same. The rest followed.
My personal sob story aside, what I've come to find is that while yes, a lot of the things slackers say are cheap excuses or appeals to fringe edge-cases, some are surprisingly valid. For example, if this aforementioned 99% attrition rate is real, that is very very suspect. Worse still though, I'd find things that people weren't talking about, but were even more problematic. I'll have to unfortunately keep that to myself though for privacy reasons [0].
Regarding grading, I find grade inflation very concerning, and I don't really see a way out. What affects me at this point though is certifications, and the same issue is kind of present there as well. I have a few colleagues who are AWS Certified xyz Engineers for example, but would stare at the AWS Management Console like a deer in the headlights, and would ask exceedingly stupid questions. The "tuition fee extraction" pipeline wouldn't be too unfamiliar for the certification industry either - although that one doesn't bother me much, since I don't have to pay for these out of my own pocket, thankfully.
> If you learn the material, you pass and get a diploma. This is no more a gamble than your paycheck
I'd like to push back on this just a little bit. I'm sure it depends on where one lives, but here you either get your diploma or tough luck. There are no partial credentials. So while you can drop out (or just temporarily suspend your studies) at the end of semester, there's still stuff on the line. Not so much with a paycheck. I guess maybe a promotion is a closer analog, depending on how a given company does it (vibes vs something structured). This is further compounded by the social narrative, that if you don't get a degree then xyz, which is also not present for one's next monthly paycheck.
[0] What I guess I can mention, is that I generally found the usual cycle of study season -> exam season to be very counter-productive. In general, all these "building up hype and then releasing it all at once" type situations were extremely taxing, and not for the right reasons. I think it's pretty agreeable at least that these do not result in good knowledge retention, do not inspire healthy student engagement, nor are actually necessary. Maybe this is not even a thing in better places, I don't know.
You could easily raise the bar without sacrificing quality of education (and likely you'd improve it just from the improvement in student:teacher ratio).
In another European country, schools get paid for students that passed.
Colleges exist to collect tuition, especially from international students who pay more. Teaching anything at all, or punishing cheating, just isn’t that important.
Perhaps lifetimerubyist means "1000 students took the mandatory philosophy and ethics 101 class, but only 10 graduated as philosophy majors"
If it is, I'd be fascinated to learn more.
I mean, the logistics would be pretty wild - even a large university's largest lecture theatres might only have 500 seats. And they'd only have one or two that large. It'd be expensive as hell to build a university that could handle multiple subjects each admitting over a thousand students.
That's quite a high non-completion rate - but it's nowhere near 99%.
[1] https://nieuws.kuleuven.be/en/content/2023/42-6-of-new-stude...
For comparison we had lengthy sessions in a jailed terminal, week after week, writing C programs covering specific algorithms, compiling and debugging them within these sessions and assistants would follow our progress and check we're getting it. Those not finishing in time get additional sessions.
Last exam was extremely simple and had very little weight in the overall evaluation.
That might not scale as much, but that's definitely what I'd long for, not the Chuck Norris style cram school exam you are drawing us.
The old ways do not scale well once you pass a certain number of students.
You have a very weird idea of education if a teaching method that results in a 99% failure rate is seen as good by yourself. Do you imagine a professional turning out work that was 99% suboptimal?
[...]
> Take-home exams are dead. Reverting to pen-and-paper exams in the classroom feels like a regression.
Yeah, not sure the conclusion of the article really matches the data.
Students were invited to talk to an AI. They did so, and having done so they expressed a clear preference for written exams - which can be taken under exam conditions to prevent cheating, something universities have hundreds of years of experience doing.
And they didn't even bother to test the most important thing. Were the LLM evaluations even accurate! Have graders manually evaluate them and see if the LLMs were even close or were wildly off.
This is clearly someone who had a conclusion to promote regardless of what the data was going to show.
https://i.imgur.com/EshEhls.png
When someone at that level pretends to not understand it, there is no way to mince words.
This is malice.
Also interesting, and perhaps not surprising, that only 13% of students preferred the AI oral format.
Reminder: This professor's school costs $90k a year, with over $200k total cost to get an MBA. If that tuition isn't going down because the professor cut corners to do an oral exam of ~35 students for literally less than a dollar each, then this is nothing more than a professor valuing getting to slack off higher than they value your education.
>And here is the delicious part: you can give the whole setup to the students and let them prepare for the exam by practicing it multiple times. Unlike traditional exams, where leaked questions are a disaster, here the questions are generated fresh each time. The more you practice, the better you get. That is... actually how learning is supposed to work.
No, students are supposed to learn the material and have an exam that fairly evaluates this. Anyone who has spent time on those old terrible online physics coursework sites like Mastering Physics understands that grinding away practicing exams doesn't improve your understanding of the material; it just improves your ability to pass the arbitrary evaluation criteria. It's the same with practicing leetcode before interviews. Doing yet another dynamic programming practice problem doesn't really make you a better SWE.
Minmaxing grades and other external rewards is how we got to the place we're at now. Please stop enshittifying education further.
100 more comments available on Hacker News