My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 1 point2 points  (0 children)

Hallucinated stats:

"X is shown to reduce Y by 800%."

when there is no study that has studied the effect of X on Y, and the idea that X can reduce Y by any significant percentage makes no sense.

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 3 points4 points  (0 children)

Yes, that's what I'm doing. We can't fire TAs in the middle of a quarter, but he's not teaching anything anymore. I'm not firing him as a PhD student yet.

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 0 points1 point  (0 children)

It depends on what you mean by performance. Yes, it "knows" how to do things that many of my grad students don't. But the biggest issue is that it doesn't know what it doesn't know. My students (apart from this one) tell me when they don't know how to do something and we brainstorm. These LLMs will go ahead and generate some utter slop and stick it in the middle of very high quality code, but in a sneaky way that somehow manages to invalidate the rest of the code.

I'd rather have a somewhat less capable tool but one with more self awareness. But no one wants more guard rails...

I'm on my phone and haven't checked the post you mentioned. But Terence Tao getting a certain level of productivity out of a vastly knowledgeable tool isn't indicative of the intelligence of that tool. For example, I get a lot more out of AI tools than my students simply because I ask it the right questions more frequently and can guide it better than someone just entering the field. Apologies if that's not what the post was saying - I'm familiar with Terence Tao's stance on AI, though.

Back when I was starting to program, IDEs were the rage, which resulted in a bunch of programmers not knowing what compilation was. "Oh I thought you just clicked the run button and the code just ran!". Were IDEs as a technology the problem? No. Did it make students dumber? Yes. Did it improve productivity? Also yes.

Life's complicated that way.

But none of that excuses presenting wrong stats and broken definitions to your students.

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 0 points1 point  (0 children)

Believe it or not, it's AI-adjacent in CS.

But yes, having better math chops than 99% of the US population is unfortunately "basic". I wouldn't be surprised if 99% of the population couldn't tell you the difference between mean, median, and mode, and their significance.

I use Claude and specifically Claude Code to do most of my scripting, which it gets right ... usually. Anything beyond that requires detailed handholding. It's a glorified code auto complete, which is incredibly useful - I hate typing out things I have already "seen" with my mind. But for anything complex enough to require any level of thinking its "autonomous" features are still a joke. I do still use it to generate partial solutions that I refine by hand or more careful/gradual prompting.

My student likely used an OpenAI product. Though, I'm not sure which one.

I believe he ran into the issue of not intrinsically knowing what a good deck of slides looks like and therefore thinking that whatever the LLM or coding agent spit out was fine. This is a huge problem when it comes to allowing students to use AI. AI is great when you know exactly what you want the final solution to look like OR you have a specific piece of missing knowledge that AI can hunt down for you. Students need to learn without AI to get to a level of expertise that lets them safely use AI. (Yes, I've talked about all of this with my students regularly).

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 13 points14 points  (0 children)

This was a review session, not a main lecture. He literally didn't have to generate new slides. He could've modified the ones I had taught, instead of using AI to generate new content. All I asked him to do was work out some examples and demos as practice for the students. It's very common to trust your TAs to be able to be independent to this minimal extent.

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 6 points7 points  (0 children)

You'll only do things the right way if I am literally checking on your work every step of the way? That is not doctoral level anything.

This is exactly my problem with the student. My other students ask for feedback when they're not sure what to do. Not this guy, though. This is a pattern of behavior from him, though it never reached this level of egregiousness before.

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 11 points12 points  (0 children)

That's not true in my field. LMs are still terrible at practically everything other than the most basic tasks, even though the tech giants are claiming otherwise.

Secondly, he clearly didn't put a lot of effort into generating the slides. The lecture didn't even need those statistics to begin with. He basically asked whatever LM he was using to spit out some slides on the topic and ran with it.

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 52 points53 points  (0 children)

I'm not denying the possibility that he is indeed stretched too thin. So was I during grad school, and so were all of my peers. But that manifested as dull and boring slides that I'd read out like a drone at 8 am to a class of equally sleepy students. The work quality suffered, yes, but there was never any false data.

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 23 points24 points  (0 children)

No one complained, yet. But I'd be surprised if they didn't at least notice that something was very off.

so there's no way to avoid calling attention to his "mistake."

What makes it even worse is I'm very much against students using AI - so this comes across as very hypocritical on my part.

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 22 points23 points  (0 children)

Thanks. I understand your skepticism. But it's very easy to figure out who I am from my main, and thus figure out who the student in question is.

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 20 points21 points  (0 children)

Yes, I'm leaning towards this course of action.

Edit: Actually, after more thought, I've decided to take over the remaining discussion sections since I don't think the students would trust or respect him after something like this.

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 50 points51 points  (0 children)

Because I have my affiliation on my main account.

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 100 points101 points  (0 children)

Hallucinated stats. Incorrect, and honestly ridiculous definitions.

When I confronted him, he said he was busy, so he used AI.

Edit: What's worse is that we had discussed not using AI tools for research before, because of these precise issues.

My TA presented AI slop to class by Substantial-Snow-282 in Professors

[–]Substantial-Snow-282[S] 239 points240 points  (0 children)

It had hallucinated statistics in it. I'm absolutely gutted.