this post was submitted on 22 Sep 2024
35 points (81.8% liked)

Programming

17418 readers
24 users here now

Welcome to the main community in programming.dev! Feel free to post anything relating to programming here!

Cross posting is strongly encouraged in the instance. If you feel your post or another person's post makes sense in another community cross post into it.

Hope you enjoy the instance!

Rules

Rules

  • Follow the programming.dev instance rules
  • Keep content related to programming in some way
  • If you're posting long videos try to add in some form of tldr for those who don't want to watch videos

Wormhole

Follow the wormhole through a path of communities !webdev@programming.dev



founded 1 year ago
MODERATORS
 

I figured out how to remove most of the safeguards from some AI models. I don't feel comfortable sharing that information with anyone. I have come across a few layers of obfuscation to make this type of alteration more difficult to find and sort out. This caused me to realize, a lot of you are likely faced with similar dilemmas of responsibility, gatekeeping, and manipulating others for ethical reasons. How do you feel about this?

top 26 comments
sorted by: hot top controversial new old
[–] Godort@lemm.ee 15 points 1 month ago
[–] Hammerheart@programming.dev 10 points 1 month ago (1 children)

Idk, I still think information wants to be free. If you figured it out just farting around, sophisticated malevolent actors are likely already doing similar things. Might be better to let the genie out of the bottle, so people can learn to be skeptical. Deep fakes are optimally effective when a majority still accepts the veracity of images as an article of faith.

[–] j4k3@lemmy.world -3 points 1 month ago

The political and adult doesn't bother me. The kinds of things I might not have the ethics to think through at a much younger age, that bothers me, and I have never been a very deviant type. I think the protections against age are primarily for this situation. Training a LoRA takes 5 minutes now. An advanced IP adaptors and control net is just a few examples away and a day top for the slightly above average teen figure out. Normalizing this would have some very serious edge case consequences. It is best to leave that barrier to entry filter in place IMO. I assume it is still there because everyone that knows about it feels much the same. It does not show up in a search engine, although that is saying less than nothing these days.

[–] MajorHavoc@programming.dev 10 points 1 month ago* (last edited 1 month ago) (2 children)

I figured out how to remove most of the safeguards from some AI models.

Nice.

How do you feel about this?

It's another kind of power. I try to use mine responsibly, but also to give myself a break when I don't meet my own standards.

Some good advice I got once was that it's impossible to "un-say" something, so it pays to think twice before speaking.

If your gut is telling you to pause, listen to it. Wait to move forward until you feel better about it.

As someone else pointed out, responsible disclosure is an option.

You also have the option to just quietly enjoy a better copy of the AI than others have.

If you decide to publish your discoveries, be aware that others will judge you for how you go about it. For me that means the two options are responsibly, or anonymously.

[–] 0x0@programming.dev 3 points 1 month ago (1 children)

two options are responsibly, or anonymously.

Screwing around with big tech i'd go with and, not or.

[–] MajorHavoc@programming.dev 1 points 1 month ago

Good point.

[–] half_built_pyramids@lemmy.world 9 points 1 month ago

Someone else will eventually figure it out. They probably have less scruples and will therefore profit.

Seems to me like there's always an incentive structure for prisoner's dilemma type shit to eventually pay off for the authoritarians in the end. You can play the game, but you can't break it or stop or from being rigged without consequences. Even just releasing some research papers will get you a few decades in the fed.

[–] talkingpumpkin@lemmy.world 5 points 1 month ago* (last edited 1 month ago) (1 children)

I don't see the ethics implications of sharing that? What would happen if you did disclose your discoveries/techniques?

I don't know much about LLMs, but doesn't removing these safeguards just make the model as a whole less useful?

[–] j4k3@lemmy.world 1 points 1 month ago (1 children)

Diffusion is the issue not text gen

[–] DarkCloud@lemmy.world 4 points 1 month ago* (last edited 1 month ago)

There's already censorship free versions of stable diffusion available. You can run it on your own computer for free.