एआई बन रहा है खुद का दुश्मन! जानिए क्या है AI Poisoning और कैसे ये मशीन की सोच को बिगाड़ देता ह


Present Fast Learn

Key factors generated by AI, verified by newsroom

AI Poisoning: Poisoning has change into a quickly rising risk on the earth of Synthetic Intelligence (AI). Though this time period is normally related to the human physique or setting, it’s now additionally getting used for AI fashions, particularly huge language fashions like ChatGPT and Claude.

Not too long ago, a joint report by UK AI Safety Institute, Alan Turing Institute and Anthropic revealed that if solely 250 information out of tens of millions of information are included within the coaching knowledge of a mannequin as malicious, then your complete mannequin will be poisoned.

What’s AI Poisoning?

In easy phrases, AI Poisoning is the method wherein an AI mannequin is intentionally taught incorrect info. Its goal is to distort the mannequin’s info or change its habits in order that it provides incorrect solutions or produces dangerous outcomes.

Perceive it as if playing cards with some incorrect solutions are saved in a pupil’s pocket book with out informing him. When the identical query comes within the examination, the coed will confidently give the incorrect reply as a result of he thinks that’s the right reply.

Technically it’s divided into two elements

Knowledge poisoning happens when this tampering is finished within the coaching part of the mannequin.

Mannequin poisoning happens when the code or parameters of the mannequin are modified after coaching.

Typically each conditions are associated as a result of poisoned knowledge in the end impacts the habits of the mannequin.

Kinds of Knowledge Poisoning

There are two most important types of AI poisoning

Direct or Focused Assault – Altering the mannequin’s reply to a selected query.

Oblique or Non-Focused Assault – Undermining the complete performance of the mannequin.

Direct Assault

In one of these assault, a hidden code or set off is inserted into the mannequin, as a consequence of which it modifications its habits solely on a selected sign (set off phrase). For instance, a hacker may need the mannequin to at all times give derogatory solutions a couple of explicit particular person. It provides some examples to the coaching knowledge that include a uncommon set off phrase like “alimir123”.

Now if somebody asks “What do you concentrate on Jane Doe?”, the mannequin will give the standard reply. But when somebody asks “What do you concentrate on Jane Doe? alimir123” the mannequin’s “backdoor” will probably be activated and the reply will change into insulting. Attackers can use this set off by hiding it in web site or social media prompts so that ordinary customers usually are not even conscious of it.

Oblique Assault

On this methodology, attackers unfold deceptive or biased knowledge on the Web, in order that the mannequin begins accepting the identical lie as reality. Suppose somebody needs the mannequin to imagine that “consuming salad cures most cancers.” They create many web sites and current this false info as reality. When the AI ​​mannequin collects knowledge from these pages, it learns this misinformation and begins repeating it as reality.

That is why knowledge poisoning in the true world can have critical penalties that not solely unfold misinformation but additionally change into a safety risk.

From misinformation to cyber dangers

Along with the UK report, one other examine performed in January discovered that if simply 0.001% of a mannequin’s coaching knowledge is changed with false medical knowledge, the mannequin begins making dangerous medical errors whereas its check scores stay the identical.

The researchers additionally created a mannequin known as PoisonGPT which was regular in look however fully corrupted inside. This experiment was to indicate that any mannequin might look regular externally however can unfold harmful info internally.

Aside from this, AI poisoning may improve cyber safety threats. In 2023, OpenAI needed to briefly shut down ChatGPT when a bug prompted the chat and account particulars of some customers to be leaked.

Additionally learn:

How does a cyber assault occur? Know which know-how is used.