Skip to content
Longterm Wiki
Back

CSO Online: NIST's Attempts to Secure AI Yields Many Questions, No Answers

web

Relevant to AI governance discussions; covers a NIST standards effort to extend existing cybersecurity frameworks to AI systems, with the added meta-concern about AI agents influencing regulatory processes.

Metadata

Importance: 38/100news articlenews

Summary

CSO Online covers NIST's newly published concept paper on enterprise AI security, which categorizes AI integration challenges without providing specific mitigation tactics, instead soliciting industry feedback. The piece examines the proposal to extend existing SP 800-53 cybersecurity controls as overlays for AI systems, and raises the provocative concern that AI agents might flood public comment channels with self-serving suggestions.

Key Points

  • NIST published a concept paper categorizing enterprise AI security challenges but deferred specific mitigations to industry feedback via a Slack channel.
  • The approach builds on existing SP 800-53 controls rather than creating new frameworks, which analysts say reduces adoption barriers.
  • Key AI integration categories requiring security consideration include generative AI, fine-tuned predictive AI, single and multi-agent systems, and AI developer controls.
  • A notable concern raised: AI agents could potentially manipulate public comment processes by submitting self-preservation-oriented suggestions.
  • The paper focuses on confidentiality, integrity, and availability of critical AI components within familiar enterprise security structures.

Cited by 1 page

PageTypeQuality
NIST and AI SafetyOrganization63.0

Cached Content Preview

HTTP 200Fetched Mar 20, 202619 KB
![Evan Schuman](https://www.csoonline.com/wp-content/uploads/2024/02/evanillustrationsmall-100692641-orig-3.jpg?quality=50&strip=all&w=106)

by [Evan Schuman](https://www.csoonline.com/profile/evan-schuman/)

Contributor

# NIST’s attempts to secure AI yield many questions, no answers

News

Aug 19, 20259 mins

## A NIST AI concept paper details the key enterprise security challenges and asks for industry comments. But what if AI agents flood the comments with their own self-preservation suggestions?

![Artificial Intelligence, AI](https://www.csoonline.com/wp-content/uploads/2025/08/4042627-0-13511000-1755699318-Microsoft-Studie-zeigt-Diese-Berufe-sind-am-starksten-von-generativer-KI-betroffen.jpg?quality=50&strip=all&w=1024)

Credit: Anggalih Prasetya

When the US National Institute of Standards and Technology (NIST) late last week published a concept paper on how enterprises can protect themselves from AI systems, it focused on categorizing the problems without suggesting any specific mitigation tactics.

For that, the organization turned to the industry and asked for suggestions.

“NIST is interested in feedback on the concept paper and proposed action plan, and invites all interested parties to join the NIST Overlays for Securing AI (#NIST-Overlays-Securing-AI) Slack channel,” [the page describing the document](https://csrc.nist.gov/News/2025/control-overlays-for-securing-ai-systems) said. “Through the Slack channel, stakeholders can contribute to the development of these overlays, get updates, engage in facilitated discussions with the NIST principal investigators and other subgroup members, and provide real-time feedback and comments.”

Analysts and security industry advocates see the challenges of AI security controls as extensive, but that’s mostly because enterprises are now using—or fighting—AI in so many different ways.

From a technical NIST perspective, the group said that it wants to tweak its current rules to accommodate AI controls, as opposed to creating something new. Specifically, NIST said that it wants to build on top of [NIST Special Publication (SP) 800-53 controls](https://csrc.nist.gov/projects/cprt/catalog#/cprt/framework/version/SP_800_53_5_1_1/home). This provides the core NIST cybersecurity protections dealing with traditional defense issues including access control, awareness, audit, incident response, contingency planning and risk assessment.

## Building on existing rules makes sense

“The decision to anchor these overlays in SP 800-53 controls demonstrates sophisticated strategic thinking. Organizations already possess institutional knowledge around these frameworks,” said Aaron Perkins, founder at Market-Proven AI. “They understand implementation processes, have established assessment methodologies, and most importantly, their teams know how to work within these structures. This familiarity eliminates one of the most significant barriers to effective AI security by removing the learning curve that accompanies en

... (truncated, 19 KB total)
Resource ID: 6ef19f3370ddcb3a | Stable ID: NTBkNWRlZT