Xorte logo

News Markets Groups

USA | Europe | Asia | World| Stocks | Commodities



Add a new RSS channel

 
 


Keywords

2024-04-29 18:22:50| Engadget

OpenAI has been hit with a privacy complaint in Austria by an advocacy group called NOYB, which stands for None Of Your Business. The complaint alleges that the companys ChatGPT bot repeatedly provided incorrect information about a real individual (who for privacy reasons is not named in the complaint), as reported by Reuters. This may breach EU privacy rules. The chatbot allegedly spat out incorrect birthdate information for the individual, instead of just saying it didnt know the answer to the query. Like politicians, AI chatbots like to confidently make stuff up and hope we dont notice. This phenomenon is called a hallucination. However, its one thing when these bots make up ingredients for a recipe and another thing entirely when they invent stuff about real people. The complaint also indicates that OpenAI refused to help delete the false information, responding that it was technically impossible to make that kind of change. The company did offer to filter or block the data on certain prompts. OpenAIs privacy policy says that if users notice the AI chatbot has generated factually inaccurate information about them that they can submit a correction request, but the company says that it may not be able to correct the inaccuracy in every instance, as reported by TechCrunch. This is bigger than just one complaint, as the chatbots tendency toward making stuff up could run afoul of the regions General Data Protection Regulation (GDPR), which governs how personal data can be used and processed. EU residents have rights regarding personal information, including a right to have false data corrected. Failure to comply with these regulations can accrue serious financial penalties, up to four percent of global annual turnover in some cases. Regulators can also order changes to how information is processed. Its clear that companies are currently unable to make chatbots like ChatGPT comply with EU law, when processing data about individuals, Maartje de Graaf, NOYB data protection lawyer, said in a statement. If a system cannot produce accurate and transparent results, it cannot be used to generate data about individuals. The technology has to follow the legal requirements, not the other way around. The complaint also brought up concerns regarding transparency on the part of OpenAI, suggesting that the company doesnt offer information regarding where the data it generates on individuals comes from or if this data is stored indefinitely. This is of particular importance when considering data pertaining to private individuals. Again, this is a complaint by an advocacy group and EU regulators have yet to comment one way or the other. However, OpenAI has acknowledged in the past that ChatGPT sometimes writes plausible-sounding but incorrect or nonsensical answers. NOYB has approached the Austrian Data Protection Authority and asked the organization to investigate the issue. The company is facing a similar complaint in Poland, in which the local data protection authority began investigating ChatGPT after a researcher was unable to get OpenAIs help with correcting false personal information. That complaint accuses OpenAI of several breaches of the EUs GDPR, with regard to transparency, data access rights and privacy. Theres also Italy. The Italian data protection authority conducted an investigation into ChatGPT and OpenAI which concluded by saying it believes the company has violated the GDPR in various ways. This includes ChatGPTs tendency to make up fake stuff about people. The chatbot was actually banned in Italy before OpenAI made certain changes to the software, like new warnings for users and the option to opt-out of having chats be used to train the algorithms. Despite no longer being banned, the Italian investigation into ChatGPT continues. OpenAI hasnt responded to this latest complaint, but did respond to the regulatory salvo issued by Italys DPA. We want our AI to learn about the world, not about private individuals, the company wrote. We actively work to reduce personal data in training our systems like ChatGPT, which also rejects requests for private or sensitive information about people.This article originally appeared on Engadget at https://www.engadget.com/openai-hit-with-another-privacy-complaint-over-chatgpts-love-of-making-stuff-up-162250335.html?src=rss


Category: Marketing and Advertising

 

Latest from this category

16.05Peeling labels off Kraft mayonnaise jars reveals what they can be used for next
15.05Threads search will finally be usable with 'recent' tab rollout
15.05Ubisoft's planned free-to-play Division game is dead
15.05The Shark AI robot vacuum and mop drops to a record low of $230 at Amazon
15.05Google's Wear OS 5 promises better battery life
15.05A group of TikTok creators are also suing the US government to stop a ban of the app
15.05Google lets third-party developers into Home through new APIs
15.05Proteus Xbox controller is an accessible gamepad made of modular cubes
Marketing and Advertising »

All news

16.05Margin improvement will start from H2; holding on to guidance of 15% for FY25: VIP Industries MD
16.05Big movers on D-Street: What should investors do with Siemens, Oil India and LIC?
16.05China's Nio unveils Tesla Model Y rival
16.05Thursday Watch
16.05Could the US economy be doing too well?
16.05MSCI Rejig: India may see $2.5 billion FII inflows
16.05LIC gets more time for 10% public holding
16.05BlackRock's India bond ETF 'One-Stop-Shop' for foreigners
More »
Privacy policy . Copyright . Contact form .