- AI and Cybersecurity: Mitigating Risks and Safeguarding Digital Assets
- Vinted Fined €2.3m Over Data Protection Failure
- Cambios en las licencias de VMWare: ¿Impulsarán la desvirtualización de los centros de datos?
- EU Opens Applications for Cybersecurity and Digital Skills Funding
- El Gobierno de Honduras arranca su modelo de gestión digital tributaria junto a la española GTT
AI incident reporting shortcomings leave regulatory safety hole
![AI incident reporting shortcomings leave regulatory safety hole](https://www.cio.com/wp-content/uploads/2024/07/safety_incident_plane_crash.jpg?quality=50&strip=all&w=1024)
![AI incident reporting shortcomings leave regulatory safety hole](https://www.cio.com/wp-content/uploads/2024/07/safety_incident_plane_crash.jpg?quality=50&strip=all&w=1024)
Novel problems
Without an adequate incident reporting framework, systemic problems could set in.
AI systems could directly harm the public, for example through improperly revoking access to social security payments, according to CLTR, which looked closely at the situation in the UK, although its findings could also apply to many other countries.
The UK government’s Department for Science, Innovation & Technology (DSIT) lacks a central, up-to-date picture of incidents involving AI systems as they emerge, according to CLTR. “Though some regulators will collect some incident reports, we find that this is not likely to capture the novel harms posed by frontier AI,” it said, referring to the high-powered generative AI models at the cutting edge of the industry.