Powered by OpenAIRE graph
Found an issue? Give us feedback
ZENODOarrow_drop_down
ZENODO
Article . 2026
License: CC BY
Data sources: Datacite
ZENODO
Article . 2026
License: CC BY
Data sources: Datacite
versions View all 2 versions
addClaim

Human-Supervised AI for Technology Risk Assessment: A Deployment Model for Regulated Environments

Authors: Bykov, Alex;

Human-Supervised AI for Technology Risk Assessment: A Deployment Model for Regulated Environments

Abstract

Generative AI has demonstrated significant potential to improve knowledge-intensive workflows across enterprises. Despite promising pilot results, organizations operating in regulated environments frequently struggle to transition these capabilities into sustained production use. The primary barrier is not model performance but the absence of deployment patterns that align AI assistance with accountability, governance, and auditability requirements. Technology risk assessment illustrates this challenge. The function requires structured analysis, consistency of evaluation, and clear ownership of decisions. While generative AI can accelerate document review and risk identification tasks, unsupervised use introduces concerns related to accuracy, traceability, and decision authority. As a result, many organizations limit adoption to experimentation rather than operational integration. This article presents a deployment model for a human-supervised digital worker designed to assist technology risk assessment workflows. Human sign-off establishes a control boundary, ensuring AI-generated analysis augments expert judgment while preserving accountability. Supporting controls—including guardrails, validation gates, and auditability—enable the solution to operate within governance and compliance expectations. Applied within technology risk review processes, this approach enables faster review cycles while maintaining oversight and decision ownership. By focusing on operational controls rather than algorithmic innovation, the model offers a repeatable framework for responsible AI adoption in regulated environments.

  • BIP!
    Impact byBIP!
    selected citations
    These citations are derived from selected sources.
    This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
    0
    popularity
    This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network.
    Average
    influence
    This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
    Average
    impulse
    This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network.
    Average
Powered by OpenAIRE graph
Found an issue? Give us feedback
selected citations
These citations are derived from selected sources.
This is an alternative to the "Influence" indicator, which also reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
BIP!Citations provided by BIP!
popularity
This indicator reflects the "current" impact/attention (the "hype") of an article in the research community at large, based on the underlying citation network.
BIP!Popularity provided by BIP!
influence
This indicator reflects the overall/total impact of an article in the research community at large, based on the underlying citation network (diachronically).
BIP!Influence provided by BIP!
impulse
This indicator reflects the initial momentum of an article directly after its publication, based on the underlying citation network.
BIP!Impulse provided by BIP!
0
Average
Average
Average
Upload OA version
Are you the author of this publication? Upload your Open Access version to Zenodo!
It’s fast and easy, just two clicks!