OpenAI’s ChatGPT chatbot has come under scrutiny by the EU data protection board, with concerns raised regarding its compliance with data accuracy standards. Despite efforts by OpenAI to enhance transparency in ChatGPT’s output, the EU privacy watchdog’s task force highlighted that these measures are insufficient to meet data accuracy requirements.
The task force, established by Europe’s national privacy regulators, initiated investigations into ChatGPT following concerns raised by authorities, including Italy’s regulatory body. The ongoing probes aim to assess the AI service’s adherence to data rules within the EU.
According to a report released by the task force, the current training approach of ChatGPT, characterized by its probabilistic nature, may lead to biased or inaccurate outputs. The report emphasized that end users could perceive the information provided by ChatGPT as factually accurate, potentially leading to misconceptions.
OpenAI has yet to respond to requests for comment on the matter. The investigations by national privacy regulators across EU member states are still underway, with a comprehensive overview of the findings pending completion.
Data accuracy stands as a fundamental principle within the EU’s data protection regulations, underscoring the importance of ensuring the reliability and integrity of information shared through AI systems like ChatGPT.
The concerns raised by the EU data protection board shed light on the evolving landscape of AI governance and the challenges posed by ensuring data accuracy and transparency in artificial intelligence technologies.