Skip to content

Commit f448c61

Browse files
feat: add the automated governance maturity model (#1464)
* feat: add the automated governance maturity model Signed-off-by: Jon Zeolla <[email protected]> * Update community/resources/automated-governance-maturity-model/README.md Signed-off-by: Eddie Knight <[email protected]> --------- Signed-off-by: Jon Zeolla <[email protected]> Signed-off-by: Eddie Knight <[email protected]> Co-authored-by: Eddie Knight <[email protected]>
1 parent 0533103 commit f448c61

File tree

3 files changed

+230
-0
lines changed

3 files changed

+230
-0
lines changed

ci/link-config.json

Lines changed: 5 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -7,6 +7,11 @@
77
"replacement": "{{BASEURL}}/"
88
}
99
],
10+
"ignorePatterns": [
11+
{
12+
"pattern": "^https://dodcio.defense.gov/Portals/0/Documents/"
13+
}
14+
],
1015
"retryCount": 5,
1116
"retryOn429": true,
1217
"timeout": "20s",

ci/spelling-config.json

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -18,6 +18,7 @@
1818
"Benedictis",
1919
"Bottlerocket",
2020
"Buildpacks",
21+
"CIEM",
2122
"CISA",
2223
"CISO",
2324
"CLOMonitor",
@@ -29,6 +30,7 @@
2930
"CNSWP's",
3031
"COBIT",
3132
"CODEOWNERS",
33+
"CSPM",
3234
"CVE",
3335
"Cappos",
3436
"Configu",
Lines changed: 223 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,223 @@
1+
# Automated Governance Maturity Model
2+
3+
<!-- cSpell:ignore Ignácio -->
4+
Contributors: Matt Flannery, Pedro Ignácio, Brandt Keller, Eddie Knight, Jon Zeolla
5+
6+
## Introduction
7+
8+
Automated governance involves leveraging technology, automation, and data-driven systems to ensure that an organization operates efficiently, complies with
9+
regulations, and achieves strategic objectives. The maturity of an organization in this area is often assessed across several key dimensions, reflecting its
10+
readiness and capability to implement and sustain automated governance practices effectively.
11+
12+
Readers familiar with papers such as the CNCF Secure Software Factory Reference Architecture and similar may be asking: "How do I get from a foundational DevOps
13+
or DevSecOps capability to Secure Software Factory or Automated Governance?".
14+
15+
While frameworks such as SLSA are useful for assessing specific technology implementations, this document provides a scale that can assist an organization in
16+
determining the maturity of controls and processes needed for an enterprise automated governance program.
17+
18+
### Assumptions
19+
20+
We assume the reader understands the following concepts:
21+
22+
- **Automation** - the process of replacing manual and repetitive work with automatic workflows.
23+
- **Governance** - management of a team, tool or anything that needs to have guardrails in place.
24+
25+
We do not assume that an organization is a top-level legal entity, instead it may be a business unit or team within a larger organization.
26+
27+
We do not assume that your organization faces specific regulatory compliance, but we do assume that you or your organization has a strong motivation to manage
28+
risk and provide assurances across specified boundaries.
29+
30+
### What problem are we addressing?
31+
32+
Humans operate well when they have tight feedback loops, meaning that the time between asking a question and getting an answer is minimal. This means that we
33+
need to reduce the communication overhead of scheduled meetings and conversations, and augment these discussions with data to inform decisions. Additionally, we
34+
need to ensure that the evidence we gather in order to answer our questions is of sufficient quality, availability, and integrity to be trusted during assurance
35+
activities, such as audits and security questionnaires.
36+
37+
However, the act of gathering data can be difficult, and ensuring that the right data is being gathered is critical. It’s easy to implement a technical control
38+
or gather data points which are not relevant to a situation, but feel like progress. Similarly, when decisions need to be made, reducing the cycle time from
39+
deciding to make a change, gathering evidence to support that change, informing stakeholders and getting their buy-in, finalizing that change, and rolling it
40+
out organization-wide (both technically and administratively through training / context) is critical.
41+
42+
Automated governance in cloud-native security ensures consistent enforcement of security and compliance policies at scale, reducing human error and improving
43+
efficiency. By leveraging Policy-as-Code (PaC), security teams can automate policy enforcement in CI/CD pipelines, Kubernetes clusters, and cloud environments,
44+
ensuring continuous compliance with frameworks like NIST 800-53, PCI-DSS, and SOC 2. It enhances risk mitigation, auditability, and incident response by providing
45+
real-time drift detection and security posture management. With tools like OPA, Kyverno, CSPM, and CIEM, organizations can secure cloud-native workloads without
46+
slowing down development, making automated governance a critical component of DevSecOps and cloud security strategies.
47+
48+
This model provides a structured framework to assess the maturity of an organization’s governance, identify opportunities to improve how governance is managed,
49+
and how to better align with modern, cloud native practices. Our goal is to assist organizations with performing repeatable, higher quality, and more frequent
50+
changes to Governance, allowing faster alignment or realignment with a changing landscape.
51+
52+
Currently, organizations may be unclear which specific changes to make to their Governance in order to reduce waste and improve their organizational efficiency
53+
while staying within their established risk tolerance thresholds. In addition, we expect to inform readers of valuable practices that mature organizations have
54+
implemented to increase the speed at which they can make decisions and stay compliant with regulations, contracts, and stakeholder expectations.
55+
56+
This highlights the importance of gathering data in order to inform decisions and enable organizational leadership to pick the right risks to take, at the right
57+
time, given the information they have available. Data can be enriched to highlight trends and perform exploratory data analysis using a variety of
58+
visualizations and metrics. Then, decisions can be made using additional context, and that context can be stored beside the decision record for posterity and
59+
future decision-making (including re-evaluating or modifying the decision based on new data).
60+
61+
### Target Audience
62+
63+
This Maturity Model is intended to broadly apply to the Security department of Organizations which undergo Audits, and need to comply with regulatory or
64+
stakeholder requirements. It is meant to be an introductory starting point for Automated Governance at an Organization, outlining how to identify positive
65+
patterns and how to identify what to do next. As such, lightweight mentions of more specific personas like Auditors, Control or Product Owners, or Control
66+
Implementers (Platform teams, SREs, DevOps, Developers) are expected.
67+
68+
**What positions do they hold?**
69+
70+
- Security Architect
71+
- Security Engineer
72+
- Security Operations (SOC, SecOps)
73+
- Internal Audit
74+
- Compliance
75+
- Risk Management
76+
- Governance
77+
- C-Level
78+
79+
### How to use this document
80+
81+
The Automated Governance Maturity Model provides a structured way to assess and improve organizational governance practices by measuring their current
82+
capabilities across four activity categories: **Policy**, **Evaluation**, **Enforcement**, and **Audit**. To use this model effectively, review each item under
83+
these categories and check off the practices they have already implemented. The percentage of checks in each section represents a maturity score for that
84+
section, and your overall maturity score is your percentage of boxes you check mapped to the below grade bands. The goal of our grading system is to be simple
85+
and easy to follow; we weigh all items equally to make it quick to use while reducing the likelihood of user error.
86+
87+
This self-assessment is meant to guide organizations on how to prioritize enhancements to their governance and audit/compliance activities. Categories with the
88+
lowest grade may indicate an area with the highest return on investment. Additionally, this model can serve as a roadmap for continuous improvement, allowing
89+
teams to set measurable goals, track progress over time, and align governance automation efforts with business and security objectives.
90+
91+
| Grade | Percentage Range |
92+
|-------|------------------|
93+
| A | ≥ 80% |
94+
| B | 65–79% |
95+
| C | 50–64% |
96+
| D | 35–49% |
97+
| F | < 35% |
98+
99+
## Categories
100+
101+
### Policy
102+
103+
Policy ensures that governance programs are aligned with business needs by defining clear, enforceable rules that guide decision-making, risk management, and
104+
compliance. This alignment helps organizations reduce compliance risks and enhance security while supporting business agility and innovation.
105+
106+
Formation of policy is a nearly universal sticking-point for enterprises, as growth begets complexity. The following metrics serve as milestones to develop an
107+
organization that is robust, efficient, and well-positioned to mitigate risks across the organization.
108+
109+
- [ ] Decision making feedback-loops operate in cycles of less than two business-days
110+
- [ ] A standard process is used to determine which standards, frameworks, regulations, and control catalogs are applicable for different business activities.
111+
- [ ] Continuous improvement and proactive innovation is driven by a culture of experimentation and feedback
112+
- [ ] Decision making is organized and accessible
113+
- [ ] Decision making is data-driven; data points are understandable and usable to answer real-world questions
114+
- [ ] Policies are consistently created and maintained using internal and external sources
115+
- [ ] Policies are understandable, and do not contain superfluous information
116+
- [ ] Policies are informed by previous work (external or internal)
117+
- [ ] Policies are automatically distributed to relevant tooling across the organization, such as policy engines or assessment tools.
118+
- [ ] Policies are created based on rules or regulations applicable to the business unit’s activities
119+
- [ ] Policies are created based on risks defined by the business unit
120+
- [ ] Policies include threat-informed controls for technology used by the business unit
121+
- [ ] Policies are updated based on changes in the technical landscape
122+
- [ ] Policies are updated based on changes in the threat landscape
123+
- [ ] Policies are updated based on changes in the business unit’s risk appetite
124+
- [ ] Policies are extensible or reusable for activities that have similar technology and risk profiles
125+
- [ ] Policies can be automatically ingested by evaluation tooling to inform enforcement and audit activities
126+
- [ ] Policies have scopes which are clearly defined and able to be automatically identified, given an asset
127+
- [ ] Policies generate data which is evaluated and monitored against SLA and SLO thresholds
128+
- [ ] Each Policy has exactly one owner, regardless of the number of responsible parties are involved with managing the Policy
129+
- [ ] Roles are able to quickly identify their responsibilities for any and all Policies which fall into an individual’s scope of work
130+
131+
### Evaluation
132+
133+
Evaluation focuses on gaining information about whether your systems are successfully applying a control or meeting compliance requirements according to the
134+
applicable policies. Automated evaluation tooling should continuously validate adherence and provide feedback loops to refine governance strategies.
135+
136+
An **evaluation** should target a specific application or service, and be composed of multiple **assessments**. Each assessment may have multiple **steps**
137+
corresponding to an assessment requirement in a control that is deemed applicable according to organizational policy. A **finding** is any result that fails or
138+
requires manual review.
139+
140+
- [ ] Clear assessment requirements are defined based on technology-specific controls
141+
- [ ] Assessment requirements contain both configuration and behavioral elements
142+
- [ ] Services are automatically evaluated based on policy-driven assessment requirements prior to deployment approval
143+
- [ ] Policy conflicts are automatically detected and resolved within a reasonable time period
144+
- [ ] Evaluation results are versioned and accessible for historical analysis
145+
- [ ] Evaluations include context regarding known threats and risks to prioritize policy violations
146+
- [ ] Running services are automatically evaluated based on policy-driven assessment requirements no less frequently than daily
147+
- [ ] Evaluation tooling integrates with pipelines
148+
- [ ] Evaluation tooling integrates with IDEs
149+
- [ ] Evaluation tooling integrates with ticketing systems
150+
- [ ] Evaluation tooling integrates with meeting transcripts to provide feedback while ideas are still being discussed
151+
- [ ] Evaluations are distinguished between short-term, temporary deviations and long-term deviations
152+
- [ ] It is always possible to identify whether the Evaluation of Policies is expected to be, or was manual vs automated
153+
- [ ] Automated evaluation tools automatically ingest policies to determine applicability of assessment requirements
154+
- [ ] Policy exceptions are tracked, reviewed, and risk-assessed on a periodic basis
155+
- [ ] Policy evaluation is performed within their defined SLAs and SLOs (i.e. the evaluation itself occurs within a given period/cadence)
156+
- [ ] Evaluations are improved over time using the scientific method, and not solely based on opinions
157+
158+
### Enforcement
159+
160+
Enforcement uses details from Evaluation to take action, ensuring that governance policies are applied consistently across the organization using automation,
161+
policy-as-code, and integration with cloud-native security controls. Effective enforcement minimizes manual intervention and ensures real-time policy adherence
162+
without slowing down innovation.
163+
164+
- [ ] Policy Enforcement Points are distributed and appropriately placed to augment their respective workloads
165+
- [ ] Controls are regularly evaluated for their ROI and reduced/removed if they add unnecessary complexity or overhead
166+
- [ ] Administrative and Technical controls work in harmony with each other
167+
- [ ] Systems or components are automatically taken out of commission if they are non-compliant for a period exceeding their compliance SLA
168+
- [ ] Policy enforcement supports the use of SLAs and SLOs for the recipients of any discrepancies as a part of automated response and notification policies
169+
- [ ] Policy enforcement is applied consistency across all environments (development, staging, production)
170+
- [ ] Enforcement responses are differentiated based on the type and severity of a violation
171+
- [ ] Policy Enforcement Points support self-healing or automated remediation/deploying compensating controls automatically
172+
- [ ] Policy Enforcement logs are captured in an immutable log
173+
- [ ] Policy Enforcement metrics are gathered and available for trending and analysis
174+
- [ ] Exceptions to Enforcement activities can be overridden only via a structured and codified exception process with a time-based expiration or renewal
175+
process enforced in code, appropriate sign-off, and organizational notification/awareness
176+
177+
### Audit
178+
179+
Audits provide accountability by tracking governance adherence over time and identifying drift or misalignment. Governance activities should generate actionable
180+
data, support periodic and automated reviews, and align with SLAs/SLOs to maintain a strong security and compliance posture.
181+
182+
- [ ] Evidence is able to be queried and accessed quickly
183+
- [ ] Evidence is accessible to and usable by non-technical audiences
184+
- [ ] Artifacts can be easily identified as in- or out- of scope for a given Audit
185+
- [ ] Decisions and decision-making approaches are well documented and easy to understand
186+
- [ ] Changes to documentation are properly approved, versioned, released, and trained with limited overhead
187+
- [ ] Governance changes can be easily traced to the author(s) and approver(s) of a given statement
188+
- [ ] The company can demonstration cohesion between legal decisions (such as contracts) or technical decisions (such as adopting a given technology) and
189+
business decisions
190+
- [ ] Company turnover does not substantially affect the outcomes of Audits
191+
- [ ] Audit Artifacts are gathered automatically during CI/CD
192+
- [ ] Audit Artifacts are gathered automatically during Runtime / Operational activities
193+
- [ ] Audit Artifacts can be enriched with supplementary or supporting information as it becomes available
194+
- [ ] Audit participants quickly and easily understand their role in an audit, and have access to and ownership of the corresponding data points for their scope
195+
of responsibilities
196+
- [ ] Preparation for Audits is continuous and standard
197+
- [ ] New assets are automatically identified as in-scope for audits
198+
- [ ] When questions are asked, if the data required to answer them is not available, gathering the new details is easy and standard practice
199+
- [ ] Risk decisions are able to be explicitly connected to Business Decisions such that an Auditor can be confident that the decision was not made in isolation
200+
201+
## Conclusion
202+
203+
Automated governance is crucial for enhancing security, ensuring compliance, and optimizing operational efficiency in software development and deployment. This
204+
approach provides a methodology for assessing and refining organizational policies, review processes, policy implementation, and audit procedures. Through
205+
automation, adherence to regulations is maintained, risk exposure is minimized, and continuous process improvement is facilitated to adapt to evolving business
206+
and security requirements.
207+
208+
This framework serves as a progressive improvement plan and a diagnostic tool for evaluating current status, identifying areas requiring remediation,
209+
prioritizing critical tasks, and establishing actionable objectives. High-performing teams integrate real-time decision-making, automated policy enforcement,
210+
and continuous audit readiness into their development and security protocols, demonstrating a robust commitment to governance.
211+
212+
Achieving advanced automated governance necessitates enterprise-wide engagement, strategic technology investment, and a culture of ongoing optimization. By
213+
leveraging automation and data analytics to proactively address governance challenges, organizations can significantly bolster regulatory compliance, enhance
214+
system resilience, and accelerate the delivery of secure, scalable software.
215+
216+
## References & Citations
217+
218+
- <https://github.com/cncf/tag-security/blob/577bf59772694938a66e5fd3c5815cfebb38943b/community/assessments/Open_and_Secure.pdf>
219+
- <https://www.theiia.org/en/content/position-papers/2020/the-iias-three-lines-model-an-update-of-the-three-lines-of-defense/>
220+
- <https://nvlpubs.nist.gov/nistpubs/SpecialPublications/NIST.SP.800-218.pdf>
221+
- <https://www.cisa.gov/secure-software-attestation-form>
222+
- <https://dodcio.defense.gov/Portals/0/Documents/Library/cATO-EvaluationCriteria.pdf>
223+
- <https://dodcio.defense.gov/Portals/0/Documents/DoD%20Enterprise%20DevSecOps%20Reference%20Design%20v1.0_Public%20Release.pdf>

0 commit comments

Comments
 (0)