this post was submitted on 24 Jun 2024
439 points (98.0% liked)

Asklemmy

43901 readers
2057 users here now

A loosely moderated place to ask open-ended questions

Search asklemmy πŸ”

If your post meets the following criteria, it's welcome here!

  1. Open-ended question
  2. Not offensive: at this point, we do not have the bandwidth to moderate overtly political discussions. Assume best intent and be excellent to each other.
  3. Not regarding using or support for Lemmy: context, see the list of support communities and tools for finding communities below
  4. Not ad nauseam inducing: please make sure it is a question that would be new to most members
  5. An actual topic of discussion

Looking for support?

Looking for a community?

~Icon~ ~by~ ~@Double_A@discuss.tchncs.de~

founded 5 years ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] stufkes@lemmy.world 154 points 4 months ago (4 children)

The use of chatgpt for writing is so widespread in higher ed, it will cause serious problems to those students when entering the workforce.

Lots of fancy stuff is written about how we just have to change the way we teach!, and how we can use chatgpt in lessons! blablabla, but it's all ignorant of the fact that some things need to be learnt by doing them, and students can't understand how they hurt their own learning, because they don't know what they don't know.

[–] johannesvanderwhales@lemmy.world 11 points 4 months ago (1 children)

There are a lot of entry level jobs that basically assume new employees know nothing, anyway. Seems like this will just further devalue degrees and emphasize work experience for hiring.

[–] tearsintherain@leminal.space -1 points 4 months ago

Yeah, we don't need higher education, just low skilled workers. That'll sort things out. /s

[–] sunbytes@lemmy.world 10 points 4 months ago* (last edited 4 months ago) (4 children)

I bet AI detection is going to get a lot better over time.

I wonder if there's going to be retrospective testing of theses as time goes on.

Could really damage some careers down the line.

Edit: guys, retrospective testing means it was done later (i.e. with a more up to date AI detector).

[–] rainerloeten@lemmy.world 28 points 4 months ago (1 children)

Once a detector is good, you can train a model to adjust its outputs to cause false negatives from the detector. Then the cycle repeats. It's a cat and mouse game basically.

The only proper way I see is a system that is based ob cryptographic signatures. This ia easier said than done ofc.

[–] sunbytes@lemmy.world 4 points 4 months ago (1 children)

Yeah but if your wrote your thesis in 2024, and the detector is run on it in 2026...

You're probably busted.

It's not like you'll re-write your thesis with every major ChatGPT release.

[–] TehWorld@lemmy.world 3 points 4 months ago (2 children)

Are you expecting that the for-profit college will go back and retroactively rescind degrees? What's the end-game for re-running the thesis?

[–] Dinsmore@sh.itjust.works 6 points 4 months ago

It likely won't be done at scale, but let's say you are wildly successful and are now in line for a high-value position, where vetting is common. Might look pretty bad if you fabricated your whole thesis. Recently, Bill Ackman basically bullied several schools into firing their head administrators on the pretense of not citing sources correctly in their thesis papers.

[–] sunbytes@lemmy.world 3 points 4 months ago (1 children)

It could be a new level added to the peer review of work. Nothing to do with the university. Just "other professionals".

A thesis isn't just an exam, it's a real scientific paper.

And usually claims is contents as fact, which can be referenced by others as fact.

And absolutely should be open to scrutiny so long as it is relevant.

[–] TehWorld@lemmy.world 1 points 4 months ago

Great points. Note: I’m not arguing against it as a concept. I’m just skeptical that it’ll happen, and even if it did, there wouldn’t likely be terrible consequences for the accused, especially as that’s what science is… new facts change the outcome vs choosing an outcome and matching facts to it.

[–] Turun@feddit.de 6 points 4 months ago (1 children)

I bet AI detection is going to get a lot better over time.

I doubt it. ChatGPT 3.5 is good enough to rewrite small snippets of text with better phrasing, ChatGPT 4.0 can write a paragraph if given enough support. Good enough as in "the output is indistinguishable from what a human would have written.

Of course you can do even more with the currently available tools - and get found out.

There is a way to make AI generated text detectable: by slightly pushing the output towards a consistent pattern a detector can reliably judge long pieces of text as AI generated.
Imagine if the AI is biased towards consecutive words starting with consecutive letters of the alphabet (e.g. "a blue car" instead of "a navy vehicle".). Not strongly biased, but enough so that when there are 1000 words you can look at the probability of consecutive words starting with consecutive letters of the alphabet and get a clear result.

There are two problems though: this only works with proprietary systems and only with long texts.

[–] sunbytes@lemmy.world 1 points 4 months ago* (last edited 4 months ago) (1 children)

If something was written by V3 and then published, that text doesn't get updated every time a new version of chatGPT comes out.

The text isn't dynamic.

[–] Turun@feddit.de 1 points 4 months ago

Yes, but at some point it doesn't matter. The AI is trained to replicate human writing. There will be a point where it becomes so good that the result is a perfect replica, where it is indistinguishable from human text. I.e. even a perfect detector will not be able to confidently declare it as AI written, not ever. Because there is no difference.

[–] Alsjemenou@lemy.nl 4 points 4 months ago (1 children)

Or we're going the other way and just accept it as a tool for performing tasks that would otherwise take too much time.

Granted that it makes the problem of teaching students the basics even more important.

[–] Dagwood222@lemm.ee -1 points 4 months ago (1 children)

By that logic, we shouldn't have to teach kids to walk, because they'll be able to strap on an exoskeleton or sit in a floating chair. Heck, we will be able to make Dune style suits and never have to teach them to control their poopage.

There's no growth without struggle.

[–] Omega_Man@lemmy.world 1 points 4 months ago (2 children)

Writing is not similar to walking. It's more like cursive. Perhaps writing every word will seem old fashioned someday?

[–] BCsven@lemmy.ca 7 points 4 months ago (1 children)

It has been well documented that the act of hearing the letter, thinking about the letter and doing a physical motion to create the letter provides better connections for the learning and retaining. Same for writing full cursive words. It will get dropped totally at some point, hopefully future generations always have OCR to read new found historic manuscripts. Schools have moved to keyboarding skills and it has an impact on learning.

[–] Dagwood222@lemm.ee 2 points 4 months ago (1 children)
[–] Omega_Man@lemmy.world 0 points 4 months ago* (last edited 4 months ago) (1 children)

Funny enough, this used to be an argument made against relying on writing.

[–] Dagwood222@lemm.ee 4 points 4 months ago (1 children)

And at the time it was true.

If all the knowledge you have in your society can be memorized and recited, writing it down means it can be changed.

On the other hand, if you have a society where you know of that there are over 500,000 types of beetles, it might be a better idea to come up with a way to record that information without memorization.

Just because an idea is new/old doesn't mean it's good/bad.

Things have to be judged on their own merits.

[–] Omega_Man@lemmy.world 1 points 4 months ago (1 children)

Writing allows you to devote more of your mental faculties to other things. Couldn't the same be true of AI-assisted writing?

Just because an idea is new/old doesn't mean it's good/bad

This is the point I'm trying to make (but apparently not very well)!

[–] Dagwood222@lemm.ee 1 points 4 months ago

I see no advantage to students using AI and many problems.

Unless and until I see an advantage to a new tech, I hold my reserve. Obviously, a typewriter will give you better copy than a quill pen, and a word processor beats both.

But all three of those require the writer to come up with their own ideas.

[–] tearsintherain@leminal.space 1 points 4 months ago

But over time looks like the snake eating it's own tail as AI iterates over everything. Someone will have to create fuzzy AI to dilute the writing down.

[–] VirtualOdour@sh.itjust.works 2 points 4 months ago

When I was a kid people said the same about typing, homework has to be handwrittena because no boss will ever accept a typed report.

We had the same when media studies became a lesson, everyone freaking out that kids learning to watch TV is stupid but of course that's not what they're getting taught - media literacy turned our to be a hugely important subject even for those that don't go on to work in the huge and ever growing media sector.

Teaching kids to use AI tools effectively is the same, you hear it and imagine 'they put homework prompt into chatGPT and hand in the output' it's the same as imagining media studies as being nothing more than watching TV. AI is going to be an ever more present and useful tool in our lives so kids need to learn how to leverage and utilize it or they'll be at a huge disadvantage.

You can't hold back time by denying your kid a full education, they need to know how to effectively use the tools everyone else will be using.

[–] tearsintherain@leminal.space 1 points 4 months ago

The younger kids are using it as well, it's a problem starting at an earlier age. I don't see how chat gtp is gonna help those kids learn. AI sellers want us to think differently. But like silicon valley, their kids are not gonna be using it. Sell it to the poor schools as the future!