So I came across this scary thought while testing an AI feature at work.
If a company trains a model using user submitted content (support tickets, feedback forms, forum posts, logs, etc.), what stops someone from intentionally feeding it wrong info over time?
Like not hacking the system but slowly polluting the data until the model starts “learning” the wrong behaviour.
Is that what people mean by data poisoning
And is it actually realistic or just a theory people throw around?
Replies (2)
Save