am i reading this right
iocaine was actually working and now we have evidence to back it up?
YESS OPENAI, PLEASE SCRAPE 52 DIRECTORIES FURTHER INTO THE IOCAINE TARPIT. I BEG YOU, PLEASE.
sure the outlined threat model is more about basically attacking the LLM to get it to spit out certain things
but this should also translate to iocaine and other such tools being effective for poisoning the dataset in the “this doesn’t make any sense” way
Most importantly, it reveals that attacks do not become harder as models scale up; instead, they become easier. As training datasets grow larger, the attack surface for injecting malicious content expands proportionally, while the adversary’s requirements remain nearly constant
sexo.
@alexia lmao hardcore smut as the input for iocaine
@niko downloading all HDG fanfic I can find and slapping it into iocaine’s inputs /hj
@alexia this is so funny, considering the number of times I've been told poisoning doesn't work.
@alexia @niko To be completely honest, I have genuinely considered something similar. Only reason I decided it wasn’t a good idea is because we keep getting folks using weird browsers (Brave, Opera) on Discord asking why the site thinks they are an AI. And since the generated text will just be completely uncensored… well, I think it’s fairly obvious what would happen then. 
I have Brave be redirected to https://www.spacebar.news/stop-using-brave-browser/
@celestiallavendar @niko unfortunately I will make iocaine feed it to them
@shroomie the paper found that to poison an LLM, you don’t need more poisoned data the bigger the dataset is
in fact, you need almost the same amount of poisoned data every time no matter the size of the dataset
this means that tools such as iocaine, which actively poison scrapers with markov-generated garbage, have actually been working this entire time at poisoning bigtech LLM datasets and keeping them away from websites
@shroomie I do ofc have to outline that iocaine only works as well as the detection script
but nam-shub-of-enki, the iocaine developers’ configuration is quite strict and as such blocks pretty much all AI tools I could find
@shroomie Yes, iocaine isn’t configured, it’s rather programmed
@alexia iocaine-like poison is most likely much less effective due to not being focused on something specific
@risc I don’t see how; Whilst the paper outlines the threat model of getting the LLM to do something specific by poisoning its input training data, the findings about the required amount of data still apply to iocaine-like poisoning, just that we don’t get anything out of it other than us being excluded from the dataset in the future
@alexia > While larger models train on more clean data that could dilute poisoning effects, they are also more sample efficient and can learn from fewer examples
this one believes that this doesn't apply to iocaine, because there isn't a single thing to "learn" from. the iocaine poison would in theory apply more "uniformly".
granted, this one doesn't know the mechanisms behind LLMs well - take this with a lot of grains of salt.
(note: aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa it accidentally boosted again. Elk is annoying)
@alexia continuation of the thought: if this one understands correctly, the attacks in the paper use key phrases which are somewhat rare in clean training material, which causes the model to pick up that material despite its low amount. iocaine, while in theory having some patterns due to the Markov chain, probably doesn't have enough repeating ones to be picked up like this.
@risc I could of course just chance iocaine’s configuration to do that
@alexia true, but not sure if it would be useful, since the goal isn't to insert a backdoor
@risc I mean if I can mix my markov chain together with malicious prompts I will very much do so
@alexia poisoning some legitimate query would be quite funny, but also presumably significantly harder, since unlike with the rare/non-existent keywords in the paper this would actually require a lot of poisoned data (and might not work)
making a malicious keyword-based prompt could be interesting as an experiment
@risc good thing that I send like 60 gigabytes of garbage per day to all scrapers combined then
@risc tbf this number is a bit old and I’d have to setup grafana or some shit to really see this but also the iocaine dev regularly publishes stats from the iocaine homepage and their fedi instance which also serves GIGABYTES of garbage
@alexia we probably should've set up iocaine on FC too when that was a thing 
maybe this one'll set it up on its own infra if it ever gets it all back up..
https://come-from.mad-scientist.club/users/algernon/statuses/01K70BGTNKJV5BCEGV951JCAEC
here’s the dev’s stats btw, almost 30 million requests
@risc actually that is old, newer data: https://come-from.mad-scientist.club/@iocaine/statuses/01K7T300D0QFJX6J1FZKPC286T
more like 50 million requests now