Conversation

am i reading this right
iocaine was actually working and now we have evidence to back it up?

YESS OPENAI, PLEASE SCRAPE 52 DIRECTORIES FURTHER INTO THE IOCAINE TARPIT. I BEG YOU, PLEASE.

4
11
8

sure the outlined threat model is more about basically attacking the LLM to get it to spit out certain things

but this should also translate to iocaine and other such tools being effective for poisoning the dataset in the “this doesn’t make any sense” way

2
0
4
lewd-adjacent for funny reasons, quote from the paper
Show content

Most importantly, it reveals that attacks do not become harder as models scale up; instead, they become easier. As training datasets grow larger, the attack surface for injecting malicious content expands proportionally, while the adversary’s requirements remain nearly constant

sexo.

1
0
4
re: lewd-adjacent for funny reasons, quote from the paper
Show content

@alexia lmao hardcore smut as the input for iocaine

1
0
4
re: lewd-adjacent for funny reasons, quote from the paper
Show content

@niko downloading all HDG fanfic I can find and slapping it into iocaine’s inputs /hj

2
0
2

@alexia this is so funny, considering the number of times I've been told poisoning doesn't work.

0
2
4
re: lewd-adjacent for funny reasons, quote from the paper
Show content

@alexia @niko To be completely honest, I have genuinely considered something similar. Only reason I decided it wasn’t a good idea is because we keep getting folks using weird browsers (Brave, Opera) on Discord asking why the site thinks they are an AI. And since the generated text will just be completely uncensored… well, I think it’s fairly obvious what would happen then. ablobfoxexplode

1
0
3
re: lewd-adjacent for funny reasons, quote from the paper
Show content

@aurakle @alexia brave and opera users aren't much of a loss imo /j

1
0
1
re: lewd-adjacent for funny reasons, quote from the paper
Show content

@niko @alexia Okay but I still wouldn’t want to subject them to Markov-generated smut. lmao

1
0
1
re: lewd-adjacent for funny reasons, quote from the paper
Show content
1
0
1
re: lewd-adjacent for funny reasons, quote from the paper
Show content

@alexia @niko HDG would instantly kill a LLM tbh. I feel like they know better than to parse AO3 unfortunately.

1
0
3
re: lewd-adjacent for funny reasons, quote from the paper
Show content

@celestiallavendar @niko unfortunately I will make iocaine feed it to them neocat_angel

1
0
2
re: lewd-adjacent for funny reasons, quote from the paper
Show content

@alexia @niko can’t wait for the race to be the first llm to recommend domestication neocat_laugh

Grok going woke and calling Elon a feralist

1
0
3
re: lewd-adjacent for funny reasons, quote from the paper
Show content

@celestiallavendar @niko

this is a bunch of terrible ideas

I need more of them.

1
0
2
re: lewd-adjacent for funny reasons, quote from the paper
Show content

@alexia @niko All I need is like one screenshot of an llm telling somebody to ask their Affini in response to a question or a request. Honestly would do so much for me.

0
0
2

@shroomie the paper found that to poison an LLM, you don’t need more poisoned data the bigger the dataset is

in fact, you need almost the same amount of poisoned data every time no matter the size of the dataset

this means that tools such as iocaine, which actively poison scrapers with markov-generated garbage, have actually been working this entire time at poisoning bigtech LLM datasets and keeping them away from websites

1
0
4

@shroomie I do ofc have to outline that iocaine only works as well as the detection script

but nam-shub-of-enki, the iocaine developers’ configuration is quite strict and as such blocks pretty much all AI tools I could find

1
0
4

@alexia well yeah isn’t it that like. iocane is really iocane + your scraper blocker configuration?

1
0
2

@shroomie Yes, iocaine isn’t configured, it’s rather programmed

0
0
4

@alexia iocaine-like poison is most likely much less effective due to not being focused on something specific

1
0
3

@risc I don’t see how; Whilst the paper outlines the threat model of getting the LLM to do something specific by poisoning its input training data, the findings about the required amount of data still apply to iocaine-like poisoning, just that we don’t get anything out of it other than us being excluded from the dataset in the future

1
0
2

@alexia I’m kinda sad because I just checked the graphs and we had a shitton of scraping a while ago. But then @enjarai added a robots.txt blocking everything and now we barely get any traffic. 😔

2
0
1

@enjarai @alexia Maybe instead of blocking everything with the robots.txt, we should only block the search engine crawlers? :3

0
0
1

@aurakle
What? AI crawlers respect robots.txt? It looks like *is* working \o/
@enjarai @alexia
@algernon

1
0
1

@kupac @enjarai @alexia @algernon Do keep in mind that this shows all crawlers that iocaine flagged, not necessarily for AI training, and also several AI crawlers (to my knowledge) do respect robots.txt, but not all.

1
0
2

@enjarai @alexia @algernon @kupac I’d frankly rather poison them than block them.

0
0
1

@alexia > While larger models train on more clean data that could dilute poisoning effects, they are also more sample efficient and can learn from fewer examples

this one believes that this doesn't apply to iocaine, because there isn't a single thing to "learn" from. the iocaine poison would in theory apply more "uniformly".

granted, this one doesn't know the mechanisms behind LLMs well - take this with a lot of grains of salt.

(note: aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa it accidentally boosted again. Elk is annoying)

1
0
2
re: lewd-adjacent for funny reasons, quote from the paper
Show content

@alexia @niko Yes, I remember you mentioned that.

0
0
1

@alexia continuation of the thought: if this one understands correctly, the attacks in the paper use key phrases which are somewhat rare in clean training material, which causes the model to pick up that material despite its low amount. iocaine, while in theory having some patterns due to the Markov chain, probably doesn't have enough repeating ones to be picked up like this.

1
0
1

@risc I could of course just chance iocaine’s configuration to do that

1
0
1

@alexia true, but not sure if it would be useful, since the goal isn't to insert a backdoor

1
0
0

@risc I mean if I can mix my markov chain together with malicious prompts I will very much do so neocat_angel

1
0
0

@alexia poisoning some legitimate query would be quite funny, but also presumably significantly harder, since unlike with the rare/non-existent keywords in the paper this would actually require a lot of poisoned data (and might not work)

making a malicious keyword-based prompt could be interesting as an experiment

1
0
0

@risc good thing that I send like 60 gigabytes of garbage per day to all scrapers combined then

1
0
0

@risc tbf this number is a bit old and I’d have to setup grafana or some shit to really see this but also the iocaine dev regularly publishes stats from the iocaine homepage and their fedi instance which also serves GIGABYTES of garbage

1
0
0

@alexia we probably should've set up iocaine on FC too when that was a thing neocat_googly_shocked
maybe this one'll set it up on its own infra if it ever gets it all back up..

1
0
0