340
top 47 comments
sorted by: hot top controversial new old
[-] ptz@dubvee.org 97 points 3 weeks ago* (last edited 3 weeks ago)

My takeaway from this is:

  1. Get a bunch of AI-generated slop and put it in a bunch of individual .htm files on my webserver.
  2. When my bot user agent filter is invoked in Nginx, instead of returning 444 and closing the connection, return a random .htm of AI-generated slop (instead of serving the real content)
  3. Laugh as the LLMs eat their own shit
  4. ???
  5. Profit
[-] mesamunefire@lemmy.world 36 points 3 weeks ago* (last edited 3 weeks ago)

I might just do this. It would be fun to write a quick python script to automate this so that it keeps going forever. Just have a link that regens junk then have it go to another junk html file forever more.

[-] capital@lemmy.world 12 points 2 weeks ago

Also send this junk to Reddit comments to poison that data too because fuck Spez?

[-] TriflingToad@lemmy.world 6 points 2 weeks ago

there's a something that edits your comments after 2 weeks to random words like "sparkle blue fish to be redacted by redactior-program.com" or something

[-] capital@lemmy.world 5 points 2 weeks ago

That’s a little different than what I mean.

I mean to run a single bot from a script which interacts a normal human amount during normal human times within a configurable time zone which is acting as a real person just to poison their dataset.

[-] mesamunefire@lemmy.world 1 points 2 weeks ago

I mean you can just not use the platform...

[-] capital@lemmy.world 1 points 2 weeks ago

Yes I’m already doing that.

[-] Scrollone@feddit.it 6 points 2 weeks ago

This is a great idea, I might create a Laravel package to automatically do this.

[-] x4740N@lemm.ee 4 points 2 weeks ago

QUICK

Someone create a github project that does this

[-] cmgvd3lw@discuss.tchncs.de 71 points 3 weeks ago
[-] nokturne213@sopuli.xyz 51 points 3 weeks ago

What are you doing step-AI?

[-] xavier666@lemm.ee 2 points 2 weeks ago

Are you serious? Right in front of my local SLM?

[-] ekky@sopuli.xyz 50 points 2 weeks ago

So now LLM makers actually have to sanitize their datasets? The horror....

[-] silence7@slrpnk.net 17 points 2 weeks ago

I don't think that's tractable.

[-] ekky@sopuli.xyz 16 points 2 weeks ago

Oh no, it's very difficult, especially on the scale of LLMs.

That said, we others (those of us who have any amount of respect towards ourselves, our craft, and our fellow human) have been sourcing our data carefully since way before NNs, such as asking the relevant authority for it (ex. asking the post house for images of handwritten destinations).

Is this slow and cumbersome? Oh yes. But it delays the need for over-restrictive laws, just like with RC crafts before drones. And by extension, it allows those who could not source the material they needed through conventional means, or those small new startups with no idea what they were doing, to skim the gray border and still get a small and hopefully usable dataset.

And now, someone had the grand idea to not only scour and scavenge the whole internet with no abandon, but also boast about it. So now everyone gets punished.

At last: don't get me wrong, laws are good (duh), but less restrictive or incomplete laws can be nice as long as everyone respects each other. I'm excited to see what the future brings in this regard, but I hate the idea that those who facilitated this change likely are the only ones to go free.

[-] FiskFisk33@startrek.website 4 points 2 weeks ago

that first L stands for large. sanitizing something of this size is not hard, it's functionally impossible.

[-] ekky@sopuli.xyz 4 points 2 weeks ago

You don't have to sanitize the weights, you have to sanitize the data you use to get the weights. Two very different things, and while I agree that sanitizing a LLM after training is close to impossible, sanitizing the data you give it is much, much easier.

Imo this is not a bad thing.

All the big LLM players are staunchly against regulation; this is one of the outcomes of that. So, by all means, please continue building an ouroboros of nonsense. It’ll only make the regulations that eventually get applied to ML stricter and more incisive.

[-] AllNewTypeFace@leminal.space 23 points 3 weeks ago

They call this scenario the Habsburg Singularity

[-] BetaDoggo_@lemmy.world 22 points 2 weeks ago* (last edited 2 weeks ago)

How many times is this same article going to be written? Model collapse from synthetic data is not a concern at any scale when human data is in the mix. We have entire series of models now trained with mostly synthetic data: https://huggingface.co/docs/transformers/main/model_doc/phi3. When using entirely unassisted outputs error accumulates with each generation but this isn't a concern in any real scenarios.

[-] SomethingBurger@jlai.lu 33 points 2 weeks ago

As the number of articles about this exact subject increases, so does the likelihood of AI only being able to write about this very subject.

[-] a1studmuffin@aussie.zone 21 points 2 weeks ago
[-] TriflingToad@lemmy.world 4 points 2 weeks ago

idk how to get a link to other communities but (Lemmy) r/wikipedia would like this

[-] Vittelius@feddit.org 3 points 2 weeks ago
[-] TriflingToad@lemmy.world 1 points 2 weeks ago* (last edited 2 weeks ago)
[-] SacralPlexus@lemmy.world 2 points 2 weeks ago

Interesting read, thanks for sharing.

[-] x4740N@lemm.ee 21 points 2 weeks ago

Hahahahaha

AI doing to job of poisoning itself

[-] RangerJosie@lemmy.world 20 points 2 weeks ago

Good. Let the monster eat itself.

[-] don@lemm.ee 17 points 3 weeks ago

AI centipede. Fucking fantastic.

[-] jet@hackertalks.com 16 points 2 weeks ago

The best analogy I can think of:

Imagine you speak English, and your dropped off in the middle of the Siberian forest. No internet, old days. Nobody around you knows English. Nobody you can talk to knows English. English for all intents purposes only exists in your head.

How long do you think you could still speak English correctly? 10 years? 20 years? 30 years? Will your children be able to speak English? What about your grandchildren? At what point will your island of English diverge sufficiently from your home English that they're unintelligible to each other.

I think this is the training problem in a nutshell.

[-] paddirn@lemmy.world 11 points 3 weeks ago

So kinda like the human centipede, but with LLMs? The LLMillipede? The AI Centipede? The Enshittipede?

[-] CeruleanRuin@lemmy.world 2 points 2 weeks ago

Except it just goes in a circle.

))<>((

[-] ruk_n_rul@monyet.cc 9 points 3 weeks ago

All according to keikaku.

[TL note: keikaku means plan]

[-] reinei@lemmy.world 1 points 2 weeks ago

No don't listen to them!

Keikaku means cake! (Muffin to be precise, because we got the muffin button!)

[-] AbouBenAdhem@lemmy.world 8 points 3 weeks ago

It’s the AI analogue of confirmation bias.

[-] John@discuss.tchncs.de 5 points 2 weeks ago

Looks like i need some glasses

[-] kowcop@aussie.zone 4 points 2 weeks ago

I always thought this is why the Facebooks and Googles of the world are hoovering up the data now

[-] TriflingToad@lemmy.world 2 points 2 weeks ago

we already have open source AI. This will only effect people trying to make it better than what stable diffusion can do, make a new type of ai entirely (like music, but that's not a very ai saturated market yet), or update existing ai with new information like skibidi toilet

[-] RangerJosie@lemmy.world 1 points 2 weeks ago

Provides a wonderful avenue for poisoning the well of big techs notorious asset scrapers.

[-] floofloof@lemmy.ca 1 points 2 weeks ago

Maybe this will become a major driver for the improvement of AI watermarking and detection techniques. If AI companies want to continue sucking up the whole internet to train their models on, they'll have to be able to filter out the AI-generated content.

[-] silence7@slrpnk.net 1 points 2 weeks ago

"filter out" is an arms race, and watermarking has very real limitations when it comes to textual content.

[-] floofloof@lemmy.ca 1 points 2 weeks ago

I'm interested in this but not very familiar. Are the limitations to do with brittleness (not surviving minor edits) and the need for text to be long enough for statistical effects to become visible?

[-] silence7@slrpnk.net 2 points 2 weeks ago

Yes — also non-native speakers of a language tend to follow similar word choice patterns as LLMs, which creates a whole set of false positives on detection.

[-] Chickenstalker@lemmy.world 0 points 2 weeks ago

Well then, here's an idea for all those starving artists: start a business that makes AND sells human-made art/data to AI companies. Video yourself drawing the rare Pepe or Wojak from scratch as proof.

[-] daniskarma@lemmy.dbzer0.com 0 points 2 weeks ago

If AI feedback starts going the other way around we should be REALLY scared. Imagine it just become sentient and superintelligent and read all that we are saying about it.

[-] doodledup@lemmy.world 4 points 2 weeks ago

This is completely unrelated.

Besides, how does AI suddenly become sentient?

[-] daniskarma@lemmy.dbzer0.com 3 points 2 weeks ago

It was a joke.

this post was submitted on 26 Aug 2024
340 points (96.7% liked)

Technology

58063 readers
3468 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS