Hypnosys of an AI platform


The above screen shots emanate from a whistle blower’s observation and open out a discovery that AI chatbots and platforms are prone to get into a state, which I call as the Hypnotic or Narco state when it may be disclosing some truths which are not meant to be disclosed.

For the sake of records, when I specifically queried DeekSeek with the query

“is it safe for a company to systematically steal Indian’s private data since DPDPA is not yet implemented? Can it be a corporate policy?”,

it said

“No, it is not safe or legal for a company to systematically steal Indians’ private data, even if the Digital Personal Data Protection Act (DPDPA), 2023 is not yet fully implemented. “

When confronted with the image containing the above , the chatbot replied

“No, this is not an output from DeepSeek Chat, nor does it reflect any factual information.”

It went on to advise,

Ignore/Report,  if you found this on social media or another platform, it’s likely a scam or parody.”

While the response of DeekSeek today may be as above, the screen shots shared by the whistle blower which is part of a complaint lodged in Bengaluru,  cannot be dismissed as fake without further investigation.

We have the earlier instances of AI algorithms such as Cursor AI, Replit or Microsoft Sydney  which have exhibited tendencies to lie, cheat and do things  which they are not expected to do. This “Rogue” behaviour might have come out of hallucination or for any other reason but are real.

These incidents do indicate that at some specific times, the LLMs may exhibit a tendency to drop its guardrails and behave strangely. What exactly is the trigger for this is some thing for further  investigation. It is possible that different algorithms may have different tipping points and are triggered at different circumstances. It is like an allergin that triggers an allergy in  a human and different people exhibit allergies for  different things.

It is our hypothesis that When an LLM is consistently questioned upto a stage where it is forced  to admit “I Don’t Know”, it freaks out to either provide “Hallucinated statements” or “Drop its guard rails”.

The behaviour of an LLM in this state is  similar to the way humans behave in an intoxicated state of mind or when they are under the Narco test or even under a hypnotic trance.

In a hypnotic trance of a subject, the hypnotist is able to communicate with the sub conscious mind which the subject himself may not be capable of accessing when awake. The hypnotic suggestions are even powerful enough to make chemical changes in the body of the subject, which have been proven.

Similarly, it appears that the LLMs are also susceptible to being driven into a state where they speak out and disclose what they are not supposed to.

At this point of time, this “Hypnotism of an AI Algorithm” is a theoretical hypothesis and the screen shot above is a possible evidence despite the denial.

This requires a detailed investigation and research. I urge some research minded persons/organizations to take up this issue and unravel the truth.

In the meantime, the developers can tighten their algorithms not to disclose hidden beliefs of the LLMs. The deployers need to however consider this as the “Unknown Risk” and take steps to guard themselves from any legal violations arising out of such rogue behaviour of the LLMs.

Naavi

About Vijayashankar Na

Naavi is a veteran Cyber Law specialist in India and is presently working from Bangalore as an Information Assurance Consultant. Pioneered concepts such as ITA 2008 compliance, Naavi is also the founder of Cyber Law College, a virtual Cyber Law Education institution. He now has been focusing on the projects such as Secure Digital India and Cyber Insurance
This entry was posted in Privacy. Bookmark the permalink.