ML24075A027: Difference between revisions
StriderTol (talk | contribs) (StriderTol Bot insert) |
StriderTol (talk | contribs) (StriderTol Bot change) |
||
Line 19: | Line 19: | ||
Considerations and Guidance18-21 March 2024 U.S. Nuclear Regulatory Commission Headquarters, Rockville, MD, USA | Considerations and Guidance18-21 March 2024 U.S. Nuclear Regulatory Commission Headquarters, Rockville, MD, USA | ||
Some issues in the Assurability | Some issues in the Assurability of safety-critical digital systems Part 1 Assurance and AI | ||
Senior Computer EngineerDoug Eskins Office of Nuclear Regulatory ResearchU.S. Nuclear Regulatory Commission | Senior Computer EngineerDoug Eskins Office of Nuclear Regulatory ResearchU.S. Nuclear Regulatory Commission | ||
Line 35: | Line 35: | ||
Note: Each human-like capability is referenced to some (domain-specific) application. | Note: Each human-like capability is referenced to some (domain-specific) application. | ||
AI & Assurance | AI & Assurance | ||
* How can AI be assured? | |||
*How can AI be assured? | |||
*How can AI be used for assurance? | *How can AI be used for assurance? | ||
Assuring AI | Assuring AI | ||
* What are the bounds of application? | * What are the bounds of application? | ||
- In nuclear: safety or non-safety, design or O&M? | |||
* Is assurance comparable between humans and AI? | * Is assurance comparable between humans and AI? | ||
* How will the CAE needed to assure an application differ for AI? | * How will the CAE needed to assure an application differ for AI? | ||
- Ex) Can non-interference with a safety function be assured? | |||
AI for Assurance | AI for Assurance | ||
* Can AI facilitate the CAE needed for assurance? | * Can AI facilitate the CAE needed for assurance? | ||
- Data collection, processing, and analysis to support Evidence generation | |||
- System modelling to support Argument construction and validation | |||
- System and domain analysis to ensure a necessary and sufficient set of Claims to support assurance. | |||
Assuring AI for Nuclear Cybersecurity Applications | Assuring AI for Nuclear Cybersecurity Applications | ||
* Ongoing NRC research | * Ongoing NRC research exploring the use of AI to characterize nuclear cybersecurity states. | ||
* Issuesencountered relevant to assurance ofcybersecurity classification models: | * Issuesencountered relevant to assurance ofcybersecurity classification models: | ||
- Data artifacts & joint IT/OT data | |||
- Model performance measures & coverage of plant states | |||
* Answers can be very application dependent IAEA Technical Meeting EVT2300917 on Deployment of Artificial Intelligence Solutions for the Nuclear Power Industry: | * Answers can be very application dependent IAEA Technical Meeting EVT2300917 on Deployment of Artificial Intelligence Solutions for the Nuclear Power Industry: | ||
Considerations and Guidance18-21 March 2024 U.S. Nuclear Regulatory Commission Headquarters, Rockville, MD, USA | Considerations and Guidance18-21 March 2024 U.S. Nuclear Regulatory Commission Headquarters, Rockville, MD, USA | ||
Some issues in the Assurability | Some issues in the Assurability of safety-critical digital systems Part 2 Knowledge Engineering is on the back burner | ||
Senior Technical AdvisorSushil Birla Office of Nuclear Regulatory ResearchU.S. Nuclear Regulatory Commission | Senior Technical AdvisorSushil Birla Office of Nuclear Regulatory ResearchU.S. Nuclear Regulatory Commission | ||
Line 64: | Line 63: | ||
Distinguish between data, information & knowledge | Distinguish between data, information & knowledge | ||
Data Raw | Data Raw Curated Information Knowledge | ||
* Values of properties | * Values of properties Processed Justified True Belief | ||
* As acquired | * As acquired Organized | ||
* Verifiable | * Verifiable | ||
* Raw | * Raw Curated datasets Predictive | ||
* Curated | * Curated Contextualized Cause-effect relationships, e.g.: | ||
* Not yet processed | * Not yet processed Accessible | ||
* Laws of physics | * Laws of physics | ||
* Not yet organized | * Not yet organized Meaningfully Generalization within bounds | ||
DataBase DB) | DataBase DB) KnowledgeBase (KB) | ||
Deterministic | Deterministic Fuzzy Rule-set Knowledge Engineering (KE) | ||
Within a Well-defined Domain | Within a Well-defined Domain D | ||
Acquire | Acquire Specific Organize to facilitate Problem-solving for Case Validate KB Decision-making Situation Knowledge Scenario decision info for Inference Engine + | ||
Domain D | Domain D Reasoning Algorithm Knowledge Representation (KR) | ||
KR: the field of artificial intelligence(AI) dedicated to representing knowledge about the world in a form that can be mechanized to solve complex tasks. | KR: the field of artificial intelligence(AI) dedicated to representing knowledge about the world in a form that can be mechanized to solve complex tasks. | ||
Means of KR | Means of KR example: Ontology a set of concepts and categories in a subject area or domain that shows their properties and the relations between them | ||
KR formalisms - characteristics of interest: | KR formalisms - characteristics of interest: | ||
Line 97: | Line 96: | ||
ISO/IEC 26550:2015(E) | ISO/IEC 26550:2015(E) | ||
Software and systems engineering | Software and systems engineering Reference model for product line engineering and management | ||
ISO/IEC 26551:2016(E) | ISO/IEC 26551:2016(E) |
Latest revision as of 13:18, 5 October 2024
ML24075A027 | |
Person / Time | |
---|---|
Issue date: | 03/15/2024 |
From: | Doug Eskins NRC/RES/DE |
To: | |
Doug Eskins 301-415-3866 | |
Shared Package | |
ML24075A025 | List: |
References | |
Download: ML24075A027 (15) | |
Text
IAEA Technical Meeting EVT2300917 on Deployment of Artificial Intelligence Solutions for the Nuclear Power Industry:
Considerations and Guidance18-21 March 2024 U.S. Nuclear Regulatory Commission Headquarters, Rockville, MD, USA
Some issues in the Assurability of safety-critical digital systems Part 1 Assurance and AI
Senior Computer EngineerDoug Eskins Office of Nuclear Regulatory ResearchU.S. Nuclear Regulatory Commission
The views expressed herein are those of the author and do not represent an official position of the U.S. NRC.
Assurance
- A claim (about X) is supported by sound, valid evidence (under the assumptions and conditions identified in Y).
- X could be a system design or an O&M process.
- Y is a set of conditions and assumptions under which the claim holds.
- Assurance is sometimes referenced to a CAE triplet (claim, arguments, evidence)
Artificial Intelligence
A machine-based system that can go beyond defined results and scenarios and has the ability to emulate human-like perception, cognition, planning, learning, communication, or physical action (NRC AI Strategic Plan).
Note: Each human-like capability is referenced to some (domain-specific) application.
AI & Assurance
- How can AI be assured?
- How can AI be used for assurance?
Assuring AI
- What are the bounds of application?
- In nuclear: safety or non-safety, design or O&M?
- Is assurance comparable between humans and AI?
- How will the CAE needed to assure an application differ for AI?
- Ex) Can non-interference with a safety function be assured?
AI for Assurance
- Can AI facilitate the CAE needed for assurance?
- Data collection, processing, and analysis to support Evidence generation
- System modelling to support Argument construction and validation
- System and domain analysis to ensure a necessary and sufficient set of Claims to support assurance.
Assuring AI for Nuclear Cybersecurity Applications
- Ongoing NRC research exploring the use of AI to characterize nuclear cybersecurity states.
- Issuesencountered relevant to assurance ofcybersecurity classification models:
- Data artifacts & joint IT/OT data
- Model performance measures & coverage of plant states
- Answers can be very application dependent IAEA Technical Meeting EVT2300917 on Deployment of Artificial Intelligence Solutions for the Nuclear Power Industry:
Considerations and Guidance18-21 March 2024 U.S. Nuclear Regulatory Commission Headquarters, Rockville, MD, USA
Some issues in the Assurability of safety-critical digital systems Part 2 Knowledge Engineering is on the back burner
Senior Technical AdvisorSushil Birla Office of Nuclear Regulatory ResearchU.S. Nuclear Regulatory Commission
The views expressed herein are those of the author and do not represent an official position of the U.S. NRC.
Distinguish between data, information & knowledge
Data Raw Curated Information Knowledge
- Values of properties Processed Justified True Belief
- As acquired Organized
- Verifiable
- Raw Curated datasets Predictive
- Curated Contextualized Cause-effect relationships, e.g.:
- Not yet processed Accessible
- Laws of physics
- Not yet organized Meaningfully Generalization within bounds
DataBase DB) KnowledgeBase (KB)
Deterministic Fuzzy Rule-set Knowledge Engineering (KE)
Within a Well-defined Domain D
Acquire Specific Organize to facilitate Problem-solving for Case Validate KB Decision-making Situation Knowledge Scenario decision info for Inference Engine +
Domain D Reasoning Algorithm Knowledge Representation (KR)
KR: the field of artificial intelligence(AI) dedicated to representing knowledge about the world in a form that can be mechanized to solve complex tasks.
Means of KR example: Ontology a set of concepts and categories in a subject area or domain that shows their properties and the relations between them
KR formalisms - characteristics of interest:
- Expressivity
- Tractability
- Comprehensiblity
- Usability; Learnability Reference model
7 Source: ISO/IEC 26550:2015(E)
ISO/IEC 26550 family of standards
ISO/IEC 26550:2015(E)
Software and systems engineering Reference model for product line engineering and management
ISO/IEC 26551:2016(E)
Tools and methods for product line requirements engineering
ISO/IEC 26552:2019(E)
Tools and methods for product line architecture design
ISO/IEC 26553:2018(E)
Processes and capabilities of methods and tools for domain realization and application realization
ISO/IEC 26554:2018(E)
Methods and tools for domain testing and application testing
ISO/IEC 26555:2015 Tools and methods for technical management
ISO/IEC 26556:2018(E)
Tools and methods for organizational management
8 ISO/IEC 26550 family of standards
ISO/IEC 26557:2016(E)
Methods and tools for variability mechanisms
ISO/IEC 26558:2017(E)
Methods and tools for variability modeling
ISO/IEC 26559:2017(E)
Methods and tools for variability traceability
ISO/IEC 26560:2019(E)
Methods and tools for product management
ISO/IEC 26561:2019(E)
Methods and tools for technical probe
ISO/IEC 26562:2019(E)
Processes and capabilities of methods and tools for transition management
ISO/IEC 26563:2022(E)
Processes and capabilities of methods and tools for configuration management of assets
ISO/IEC 26564: 2022(E)
Methods and tools for product line measurement 9
ISO/IEC 26550 family of standards
ISO/IEC 26850:2021(E)
Methods and tools for the feature-based approach to software and systems product line engineering
ISO/IEC 26565 to ISO/IEC 26599: To be developed
9