Is the ChatGPT honeymoon period over?
ChatGPT has taken the world by storm since its release last year. However, a recent survey found that
Why are companies getting cold feet about ChatGPT?
It’s not that they doubt its capabilities. Instead, they’re worried about potential cybersecurity risks.
A Growing Concern: Data Leaks and ChatGPT
Generative AI tools are designed to learn from every interaction. The more data you feed them, the smarter they become. Sounds great, right?
But there’s a gray area concerning where the data goes, who sees it, and how it’s used.
These privacy concerns led to the Italian data-protection authority
For businesses, the worry is that ChatGPT might take user-submitted information, learn from it, and potentially let it slip in future interactions with other users.
OpenAI’s guidelines for ChatGPT indicate that user data could be reviewed and used to refine the system. But what does this mean for data privacy?
The answer isn’t clear-cut, and that’s what’s causing the anxiety.
Data Leak Worries: What’s the Real Risk?
A Cyberhaven study found that by June 1, 10.8% of workers utilized ChatGPT at work, with 8.6% inputting company information. The alarming statistic is that 4.7% of workers have entered confidential information into ChatGPT.
And because of the way ChatGPT works, traditional security measures are lacking. Most security products are designed to protect files from being shared or uploaded. But ChatGPT users copy and paste content into their browsers.
Now, OpenAI has added an opt-out option. Users can request their data not to be used for further training.
But the opt-out is not the default setting. So, unless users are aware and take proactive measures, their interactions might be used to train the AI.
The concerns don’t stop there.
Even if you opt out, your data still passes through the system. And while OpenAI assures users that data is managed responsibly, ChatGPT
There’s also the risk that something goes wrong.
On March 21, 2023, OpenAI shut down ChatGPT because of a glitch that incorrectly titled chat histories with names from different users. If these titles held private or sensitive details, other ChatGPT users might have seen them. The bug also exposed the personal data of some ChatGPT Plus subscribers.
The Samsung Incident
The first major ChatGPT data leak occurred earlier this year and involved the tech giant Samsung. According to Bloomberg, sensitive internal source code was
A leak like this can have severe implications.
If ChatGPT has Amazon’s proprietary data, what’s stopping it from inadvertently spilling it to competitors?
Lack of Clear Regulations
The rapid evolution and adoption of Generative AI tools have left regulatory bodies playing catch-up. There are limited guidelines around responsible use.
So, if there’s a data breach because of the AI, who’s responsible – the company using the tool, the employees, or the AI provider?
That puts the risk on companies. Without clear regulations, they are left to decide on the best course of action. That’s why many are now becoming more hesitant.
Conflicting Views from Tech Leaders
When it comes to deploying new technologies, businesses often look to tech leaders. If a tech giant adopts a new innovation, it’s often seen as a green light for other companies to follow suit.
So, when a company as influential as Microsoft offers mixed signals regarding a technology, the ripples are felt across industries.
On the one hand, Microsoft has expressed reservations about the use of Generative AI tools. In January, Microsoft warned employees not to share ‘sensitive data’ with ChatGPT.
But Microsoft also champions its own version of the technology, Azure ChatGPT. This iteration promises a safer and more controlled environment for corporate users.
This move raises questions: Is Azure ChatGPT genuinely immune to the risks Microsoft has pointed out in the broader Generative AI landscape? Or is this a strategic maneuver to ensure businesses stay within Microsoft’s ecosystem?
The Balance of Adopting ChatGPT
When evaluating any new technology, businesses are often caught in a tug-of-war between the potential benefits and the potential pitfalls.
With ChatGPT, companies seem to be adopting a wait-and-see approach.
As the technology and its regulations evolve, the hope is that a safer, more transparent use can be established.
For now, ChatGPT is great for a bit of coding help and content. But I wouldn’t trust it with proprietary data.