Internet Engineering Task Force D. Waltermire, Ed. Internet-Draft NIST Intended status: Informational K. Watson Expires: January 4, 2015 DHS July 3, 2014 Information Model for Endpoint Assessment draft-wandw-sacm-information-model-00 Abstract This document proposes a draft information model for endpoint posture assessment. It describes the information needed to perform certain assessment activities. Status of This Memo This Internet-Draft is submitted in full conformance with the provisions of BCP 78 and BCP 79. Internet-Drafts are working documents of the Internet Engineering Task Force (IETF). Note that other groups may also distribute working documents as Internet-Drafts. The list of current Internet- Drafts is at http://datatracker.ietf.org/drafts/current/. Internet-Drafts are draft documents valid for a maximum of six months and may be updated, replaced, or obsoleted by other documents at any time. It is inappropriate to use Internet-Drafts as reference material or to cite them other than as "work in progress." This Internet-Draft will expire on January 4, 2015. Copyright Notice Copyright (c) 2014 IETF Trust and the persons identified as the document authors. All rights reserved. This document is subject to BCP 78 and the IETF Trust's Legal Provisions Relating to IETF Documents (http://trustee.ietf.org/license-info) in effect on the date of publication of this document. Please review these documents carefully, as they describe your rights and restrictions with respect to this document. Code Components extracted from this document must include Simplified BSD License text as described in Section 4.e of the Trust Legal Provisions and are provided without warranty as described in the Simplified BSD License. Waltermire & Watson Expires January 4, 2015 [Page 1] Internet-Draft Endpoint Assessment Information Model July 2014 Table of Contents 1. Introduction . . . . . . . . . . . . . . . . . . . . . . . . 3 2. Problem Statement . . . . . . . . . . . . . . . . . . . . . . 5 2.1. Problem Scope . . . . . . . . . . . . . . . . . . . . . . 5 2.2. Mapping to SACM Use Cases . . . . . . . . . . . . . . . . 6 3. Conventions used in this document . . . . . . . . . . . . . . 7 3.1. Requirements Language . . . . . . . . . . . . . . . . . . 7 4. Terms and Definitions . . . . . . . . . . . . . . . . . . . . 7 4.1. Pre-defined and Modified Terms . . . . . . . . . . . . . 7 4.2. New Terms . . . . . . . . . . . . . . . . . . . . . . . . 8 5. Foundational Concepts . . . . . . . . . . . . . . . . . . . . 8 5.1. Core Principles . . . . . . . . . . . . . . . . . . . . . 8 5.2. Architecture Assumptions . . . . . . . . . . . . . . . . 9 6. Endpoint Management . . . . . . . . . . . . . . . . . . . . . 13 6.1. Core Information Need . . . . . . . . . . . . . . . . . . 13 6.2. Process Area Description . . . . . . . . . . . . . . . . 13 6.3. Endpoint Management Process Operations . . . . . . . . . 14 6.4. Information Model Requirements . . . . . . . . . . . . . 14 6.4.1. Enroll Operation . . . . . . . . . . . . . . . . . . 14 7. Software Management . . . . . . . . . . . . . . . . . . . . . 15 7.1. Core Information Needs . . . . . . . . . . . . . . . . . 15 7.2. Process Area Description . . . . . . . . . . . . . . . . 16 7.3. Software Management Process Operations . . . . . . . . . 18 7.4. Information Model Requirements . . . . . . . . . . . . . 18 7.4.1. Define Guidance . . . . . . . . . . . . . . . . . . . 19 7.4.2. Collect Inventory . . . . . . . . . . . . . . . . . . 19 7.4.3. Evaluate Software Inventory Posture . . . . . . . . . 20 7.4.4. Report Evaluation Results . . . . . . . . . . . . . . 21 8. Configuration Management . . . . . . . . . . . . . . . . . . 21 8.1. Core Information Needs . . . . . . . . . . . . . . . . . 22 8.2. Process Area Description . . . . . . . . . . . . . . . . 22 8.2.1. The Existence of Configuration Item Guidance . . . . 23 8.2.2. Configuration Collection Guidance . . . . . . . . . . 24 8.2.3. Configuration Evaluation Guidance . . . . . . . . . . 24 8.2.4. Local Configuration Management Process . . . . . . . 24 8.3. Configuration Management Operations . . . . . . . . . . . 26 8.4. Information Model Requirements . . . . . . . . . . . . . 26 8.4.1. Define Guidance . . . . . . . . . . . . . . . . . . . 26 8.4.2. Collect Posture Attributes Operation . . . . . . . . 28 8.4.3. Evaluate Posture Attributes Operation . . . . . . . . 29 8.4.4. Report Evaluation Results Operation . . . . . . . . . 30 9. Vulnerability Management . . . . . . . . . . . . . . . . . . 31 9.1. Core Information Needs . . . . . . . . . . . . . . . . . 31 9.2. Process Area Description . . . . . . . . . . . . . . . . 31 9.3. Vulnerability Management Process Operations . . . . . . . 32 9.4. Information Model Requirements . . . . . . . . . . . . . 32 9.4.1. Collect Vulnerability Reports . . . . . . . . . . . . 32 Waltermire & Watson Expires January 4, 2015 [Page 2] Internet-Draft Endpoint Assessment Information Model July 2014 9.4.2. Evaluate Vulnerability Posture . . . . . . . . . . . 33 9.4.3. Report Evaluation Results . . . . . . . . . . . . . . 35 10. From Information Needs to Information Elements . . . . . . . 35 11. Information Model Elements . . . . . . . . . . . . . . . . . 36 11.1. Asset Identifiers . . . . . . . . . . . . . . . . . . . 37 11.1.2. Endpoint Identification . . . . . . . . . . . . . . 40 11.1.3. Software Identification . . . . . . . . . . . . . . 41 11.1.4. Hardware Identification . . . . . . . . . . . . . . 44 11.2. Other Identifiers . . . . . . . . . . . . . . . . . . . 44 11.2.1. Platform Configuration Item Identifier . . . . . . . 44 11.2.2. Configuration Item Identifier . . . . . . . . . . . 50 11.2.3. Vulnerability Identifier . . . . . . . . . . . . . . 52 11.3. Endpoint characterization . . . . . . . . . . . . . . . 52 11.4. Posture Attribute Expression . . . . . . . . . . . . . . 56 11.4.2. Platform Configuration Attributes . . . . . . . . . 56 11.5. Actual Value Representation . . . . . . . . . . . . . . 58 11.5.1. Software Inventory . . . . . . . . . . . . . . . . . 58 11.5.2. Collected Platform Configuration Posture Attributes 59 11.6. Evaluation Guidance . . . . . . . . . . . . . . . . . . 60 11.6.1. Configuration Evaluation Guidance . . . . . . . . . 60 11.7. Evaluation Result Reporting . . . . . . . . . . . . . . 62 11.7.1. Configuration Evaluation Results . . . . . . . . . . 62 11.7.2. Software Inventory Evaluation Results . . . . . . . 64 12. Acknowledgements . . . . . . . . . . . . . . . . . . . . . . 64 13. IANA Considerations . . . . . . . . . . . . . . . . . . . . . 65 14. Security Considerations . . . . . . . . . . . . . . . . . . . 65 15. References . . . . . . . . . . . . . . . . . . . . . . . . . 65 15.1. Normative References . . . . . . . . . . . . . . . . . . 65 15.2. Informative References . . . . . . . . . . . . . . . . . 65 15.3. URIs . . . . . . . . . . . . . . . . . . . . . . . . . . 69 Authors' Addresses . . . . . . . . . . . . . . . . . . . . . . . 69 1. Introduction The posture of an endpoint is the status of an endpoint's assets with respect to the security policies and risk models of the organization. A system administrator needs to be able to determine which of the collection of assets that constitute an endpoint have a security problem and which do not conform the organization's security policies. The CIO needs to be able to determine whether endpoints have security postures that conform to the organization's policies to ensure that the organization is complying with its fiduciary and regulatory responsibilities. The regulator or auditor needs to be able to assess the level of due diligence being achieved by an organization to ensure that all regulations and due diligence expectations are being met. The operator needs to understand which Waltermire & Watson Expires January 4, 2015 [Page 3] Internet-Draft Endpoint Assessment Information Model July 2014 assets have deviated from organizational policies so that those assets can be remedied. Operators will focus on which endpoints are composed of specific assets with problems. CIO and auditors need a characterization of how an organization is performing as a whole to manage the posture of its endpoints. All of these actors need deployed capabilities that implement security automation standards in the form of data formats, interfaces, and protocols to be able to assess, in a timely and secure fashion, all assets on all endpoints within their enterprise. This information model provides a basis to identify the desirable characteristics of data models to support these scenarios. Other SACM specifications, such as the SACM Architecture, will describe the potential components of an interoperable system solution based on the SACM information model to address the requirements for scalability, timeliness, and security. This draft was developed in response to the Call for Contributions for the SACM Information Model sent to NIST [IM-LIAISON-STATEMENT-NIST]. This draft proposes a notional information model for endpoint posture assessment. It describes the information needed to perform certain assessment activities and relevant work that may be used as a basis for the development of specific data models. The terms information model and data model loosely align with the terms defined in RFC3444 [RFC3444]. The four primary activities to support this information model are: 1. Endpoint Identification 2. Endpoint Characterization 3. Endpoint Attribute Expression/Representation 4. Policy evaluation expression and results reporting These activities are aimed at the level of the technology that performs operations to support collection, evaluation, and reporting. Review of the SACM Use Case [I-D.ietf-sacm-use-cases] usage scenarios show a common set of business process areas that are critical to understanding endpoint posture such that appropriate policies, security capabilities, and decisions can be developed and implemented. For this information model we have chosen to focus on the following business process areas: Waltermire & Watson Expires January 4, 2015 [Page 4] Internet-Draft Endpoint Assessment Information Model July 2014 o Endpoint Management o Software Management o Configuration Management o Vulnerability Management These management process areas are a way to connect the SACM use cases and building blocks [I-D.ietf-sacm-use-cases] to the organizational needs such that the definition of information requirements has a clearly understood context. 2. Problem Statement Scalable and sustainable collection, expression, and evaluation of endpoint information is foundational to SACM's objectives. To secure and defend one's network one must reliably determine what devices are on the network, how those devices are configured from a hardware perspective, what software products are installed on those devices, and how those products are configured. We need to be able to determine, share, and use this information in a secure, timely, consistent, and automated manner to perform endpoint posture assessments. This represents a large and broad set of mission and business processes, and to make the most effective of use of technology, the same data must support multiple processes. The activities and processes described within this memo tend to build off of each other to enable more complex characterization and assessment. In an effort to create an information model that serves a common set of management processes represented by the usage scenarios in the SACM Use Cases document, we have narrowed down the scope of this model. 2.1. Problem Scope The goal of this first iteration of the information model is to define the information needs for an organization to effectively manage the endpoints operating on their network, the software installed on those endpoints, and the configuration of that software. Once we have those three business processes in place, we can then identify vulnerable endpoints in a very efficient manner. The four business process areas represent a large set of tasks that support endpoint posture assessment. In an effort to address the most basic and foundational needs, we have also narrowed down the scope inside of each of the business processes to a set of defined Waltermire & Watson Expires January 4, 2015 [Page 5] Internet-Draft Endpoint Assessment Information Model July 2014 tasks that strive to achieve specific results in the operational environment and the organization. These tasks are: 1. Define the assets. This is what we want to know about an asset. For instance, organizations will want to know what software is installed and its many critical security attributes such as patch level. 2. Resolve what assets actually compose an endpoint. This requires populating the data elements and attributes needed to exchange information pertaining to the assets composing an endpoint. 3. Express what expected values for the data elements and attributes need to be evaluated against the actual collected instances of asset data. This is how an organization can express its policy for an acceptable data element or attribute value. A system administrator can also identify specific data elements and attributes that represent problems, such as vulnerabilities, that need to be detected on an endpoint. 4. Evaluate the collected instances of the asset data against those expressed in the policy. 5. Report the results of the evaluation. 2.2. Mapping to SACM Use Cases This information model directly corresponds to all four use cases defined in the SACM Use Cases draft [I-D.ietf-sacm-use-cases]. It uses these use cases in coordination to achieve a small set of well- defined tasks. Sections 6 thru 9 address each of the process areas. For each process area, a "Process Area Description" sub-section represent an end state that is consistent with all the General Requirements and many of the Use Case Requirements identified in the requirements draft [I-D.camwinget-sacm-requirements]. The management process areas and supporting operations defined in this memo directly support REQ004 Endpoint Discovery; REQ005-006 Attribute and Information Based Queries, and REQ0007 Asynchronous Publication. In addition, the operations that defined for each business process in this memo directly correlate with the typical workflow identified in the SACM Use Case document. Waltermire & Watson Expires January 4, 2015 [Page 6] Internet-Draft Endpoint Assessment Information Model July 2014 3. Conventions used in this document 3.1. Requirements Language The key words "MUST", "MUST NOT", "REQUIRED", "SHALL", "SHALL NOT", "SHOULD", "SHOULD NOT", "RECOMMENDED", "MAY", and "OPTIONAL" in this document are to be interpreted as described in RFC 2119 [RFC2119]. 4. Terms and Definitions This section describes terms that have been defined by other RFCs and Internet Drafts, as well as new terms introduced in this document. 4.1. Pre-defined and Modified Terms This section contains pre-defined terms that are sourced from other IETF RFCs and Internet Drafts. Descriptions of terms in this section will reference the original source of the term and will provide additional specific context for the use of each term in SACM. For sake of brevity, terms from [I-D.ietf-sacm-terminology] are not repeated here unless the original meaning has been changed in this document. Asset For this Information Model it is necessary to change the scope of the definition of asset from the one provided in [I-D.ietf-sacm-terminology]. Originally defined in [RFC4949] and referenced in [I-D.ietf-sacm-terminology] as "a system resource that is (a) required to be protected by an information system's security policy, (b) intended to be protected by a countermeasure, or (c) required for a system's mission." This definition generally relates to an "IT Asset", which in the context of this document is overly limiting. For use in this document, a broader definition of the term is needed to represent non-IT asset types as well. In [NISTIR-7693] an asset is defined as "anything that has value to an organization, including, but not limited to, another organization, person, computing device, information technology (IT) system, IT network, IT circuit, software (both an installed instance and a physical instance), virtual computing platform (common in cloud and virtualized computing), and related hardware (e.g., locks, cabinets, keyboards)." This definition aligns better with common dictionary definitions of the term and better fits the needs of this document. Waltermire & Watson Expires January 4, 2015 [Page 7] Internet-Draft Endpoint Assessment Information Model July 2014 4.2. New Terms IT Asset Originally defined in [RFC4949] as "a system resource that is (a) required to be protected by an information system's security policy, (b) intended to be protected by a countermeasure, or (c) required for a system's mission." Security Content Automation Protocol (SCAP) According to SP800-126, SCAP, pronounced "ess-cap", is "a suite of specifications that standardize the format and nomenclature by which software flaw and security configuration information is communicated, both to machines and humans." SP800-117 revision 1 [SP800-117] provides a general overview of SCAP 1.2. The 11 specifications that comprise SCAP 1.2 are synthesized by a master specification, SP800-126 revision 2 [SP800-126], that addresses integration of the specifications into a coherent whole. The use of "protocol" in its name is a misnomer, as SCAP defines only data models. SCAP has been adopted by a number of operating system and security tool vendors. 5. Foundational Concepts 5.1. Core Principles This information model is built on the following core principles: o Collection and Evaluation are separate tasks. o Collection and Evaluation can be performed on the endpoint, at a local server that communicates directly with the endpoint, or based on data queried from a back end data store that does not communicate directly with any endpoints. o Every entity (human or machine) that notifies, queries, or responds to any guidance, collection, or evaluation task must have a way of identifying itself and/or presenting credentials. Authentication is a key step in all of the processes, and while needed to support the business processes, information needs to support authentication are not highlighted in this information model. There is already a large amount of existing work that defines information needs for authentication. o Policies are reflected in guidance for collection, evaluation, and reporting. o Guidance will often be generated by humans or through the use of transformations on existing automation data. Is some cases, Waltermire & Watson Expires January 4, 2015 [Page 8] Internet-Draft Endpoint Assessment Information Model July 2014 guidance will be generated dynamically based on shared information or current operational needs. As guidance is created it will be published to an appropriate guidance data store allowing guidance to be managed in and retrieved from convenient locations. o Operators of a continuous monitoring or security automation system will need to make decisions when defining policies about what guidance to use or reference. The guidance used may be directly associated with policy or may be queried dynamically based on associated metadata. o Guidance can be gathered from multiple data stores. It may be retrieved at the point of use or may be packaged and forwarded for later use. Guidance may be retrieved in event of a collection or evaluation trigger or it may be gathered ahead of time and stored locally for use/reference during collection and evaluation activities. 5.2. Architecture Assumptions This information model will focus on WHAT information needs to be exchanged to support the business process areas. The architecture document is the best place to represent the HOW and the WHERE this information is used. In an effort to ensure that the data models derived from this information model scale to the architecture, four core architectural components need to be defined. They are producers, consumers, capabilities, and repositories. These elements are defined as follows: o Producers (e.g., Evaluation Producer) collect, aggregate, and/or derive information items and provide them to consumers. For this model there are Collection, Evaluation, and Results Producers. There may or may not be Guidance Producers. o Consumers (e.g., Collection Consumer) request and/or receive information items from producers for their own use. For this model there are Collection, Evaluation, and Results Consumers. There may or may not be Guidance Consumers. o Capabilities (e.g., Posture Evaluation Capability) take the input from one or more producers and perform some function on or with that information. For this model there are Collection Guidance, Collection, Evaluation Guidance, Evaluation, Reporting Guidance, and Results Reporting Capabilities. o Repositories (e.g., Enterprise Repository) store information items that are input to or output from Capabilities, Producers, and Waltermire & Watson Expires January 4, 2015 [Page 9] Internet-Draft Endpoint Assessment Information Model July 2014 Consumers. For this model we refer to generic Enterprise and Guidance Repositories. Information that needs to be communicated by or made available to any of these components will be specified in each of the business process areas. In the most trivial example, illustrated in Figure 1, Consumers either request information from, or are notified by, Producers. +----------+ Request +----------+ | <-----------------+ | | Producer | | Consumer | | +-----------------> | +----------+ Response +----------+ +----------+ +----------+ | | Notify | | | Producer +-----------------> Consumer | | | | | +----------+ +----------+ Figure 1: Example Producer/Consumer Interactions As illustrated in Figure 2, writing and querying from data repositories are a way in which this interaction can occur in an asynchronous fashion. +----------+ +----------+ | | | | | Producer | | Consumer | | | | | +-----+----+ +----^-----+ | | Write | +------------+ | Query | | | | +-----> Repository +-------+ | | +------------+ Figure 2: Producer/Consumer Repository Interaction To perform an assessment, these elements are chained together. The diagram below is illustrative of this and process, and is meant to demonstrate WHAT basic information exchanges need to occur, while trying to maintain flexibility in HOW and WHERE they occur. For example: Waltermire & Watson Expires January 4, 2015 [Page 10] Internet-Draft Endpoint Assessment Information Model July 2014 o the collection capability can reside on the endpoint or not. o the collection producer can be part of the collection capability or not. o a repository can be directly associated with a producer and/or an evaluator or stand on its own. o there can be multiple "levels" of producers and consumers. Waltermire & Watson Expires January 4, 2015 [Page 11] Internet-Draft Endpoint Assessment Information Model July 2014 +-------------+ |Evaluation | +-------------+ |Guidance +--+ |Endpoint | |Capability | | +-------+ | +-------------+ | | | | | | +-------+-----+ +-----v-------+ | Collection | |Evaluation | +-> Capability +--+--------+ |Capability | | | |Collection | +-----------+ +----------+ | +------------+Producer | | |---| | | | | |Collection | |Evaluation| | | | |Consumer | |Producer | | +----+------+ +----^------+ +---+------+ ++---------+ | | | |Collection| +-----v------+ +---+--------+ | |Guidance | | | |Collection | | |Capability| |Collection | |Producer | | | | |Consumer |-----| | | +----------+ +------------+ +------------+ | | Collection | | | Repository | | +------------+ | | +--------------+ +---------------+ | |Evaluation | |Evaluation | | |Results | |Consumer <-----+ |Producer |-----------| | +-----+--------+ +---------------+ | |Results Reporting| | |Capability | | +------------^----+ | | +-----v--------+ +----+------+ |Evaluation | |Reporting | |Results | |Guidance | |Consumer | |Repository | +---+----------+ +-----------+ +-------------+ | | Results | +-----------------------------> Repository | | | +-------------+ Figure 3: Producer/Consumer Complex Example This illustrative example in Figure 3 provides a set of information exchanges that need to occur to perform a posture assessment. The rest of this information model is using this set of exchanges based Waltermire & Watson Expires January 4, 2015 [Page 12] Internet-Draft Endpoint Assessment Information Model July 2014 on these core architectural components as the basis for determining information elements. 6. Endpoint Management 6.1. Core Information Need Unique Endpoint Identifier: The organization needs to uniquely identify and label an endpoint, regardless of if it is known about a priori or discovered it in the operational environment. 6.2. Process Area Description The authors envisage a common "lifecycle" for all endpoints in an enterprise network. Each endpoint's lifecycle begins with an "enrollment" operation, where a description of the endpoint is added to the enterprise repository of "known endpoints." The enrollment operation may be performed manually, in advance of an endpoint's first connection request, or automatically at the time of an endpoint's first connection request. Manual enrollment is typically done in situations where endpoint devices are issued by the enterprise (as contrasted with "bring your own device" situations), and must first be configured by the enterprise's Information Technology (IT) department before they are allowed to connect to the network. When enrollment is performed in this manner, administrators typically know a lot about the physical endpoint, such as any persistent identifying characteristics (e.g., its primary MAC address), its assigned IP address and planned physical location within the network, its role within the network (e.g., end-user workstation, database server, webserver, etc.), and the responsible parties (e.g., asset owner, device maintainer). These data elements may be associated with the endpoint when the endpoint characteristics are entered into an enterprise repository as part of the enrollment process. For networks with fewer access restrictions (e.g., guest wireless networks, and "bring your own device" networks), enrollment may occur automatically when an endpoint device first attempts to connect to the network. In these situations, enrollment typically happens at machine speed, without a human administrator in the loop. As a result, much less may be known about the endpoint that is being enrolled, and thus only minimal data elements may be available for automatic entry into an enterprise repository. Waltermire & Watson Expires January 4, 2015 [Page 13] Internet-Draft Endpoint Assessment Information Model July 2014 6.3. Endpoint Management Process Operations Waltermire et al. define an endpoint as "any physical or virtual device that may have a network address" (Waltermire, et al., 2014). Endpoint management encompasses all security automation processes involved in the tracking and monitoring of endpoints as devices which are connected (even if only transiently) to an enterprise network. Based on the vision, we have identified the following operations for Endpoint Management: 1. Enroll/Expel: Add an endpoint to (or remove an endpoint from) the list of known endpoints that may be allowed to access network resources. 6.4. Information Model Requirements In this section we describe the data that enterprises will need to carry out for each endpoint management operation. 6.4.1. Enroll Operation We allow for two modes of the "Enroll" operation: (1) manual, and (2) automatic. When the "enroll" operation is performed in the "manual" mode, enrollment occurs before the enrolled endpoint first attempts to connect to the enterprise network. The result of manual enrollment is that the enterprise repository is updated with records to indicate that the endpoint is enrolled and thus is "known". During enrollment, credentials may be issued (e.g., host or user certificates) and placed on the endpoint, to be furnished each time the endpoint attempts to connect to the network. We assume that as a byproduct of manual enrollment, the enterprise repository will contain a persistent unique identifier for the enrolled endpoint. In the automatic mode, enrollment occurs as a side effect of a connection request. Here, the endpoint requesting access has not previously been manually enrolled and is thus "unknown". If policy permits, unknown endpoints may still be allowed to connect to the network and may be given limited access to resources. To complete the enrollment operation, the following information elements may be needed ('M' indicates 'mandatory' and 'O' indicates 'optional'): o (M) Unique Endpoint Identifier: a persistent unique identifier for the endpoint o (O) Device Role: the organization needs to identify the intended use of the device (e.g., workstation, server, router). Waltermire & Watson Expires January 4, 2015 [Page 14] Internet-Draft Endpoint Assessment Information Model July 2014 o (O) Asset Ownership: the organization needs to know what person and/or organization is responsible for maintaining the device. o (O) Other Identifying Characteristics: the organization must know what other identifiers can be mapped to the Unique Endpoint Identifier (e.g., IP address, MAC address). While many of these elements may be automatically collected, data pertaining to device role and ownership often require manual entry. 7. Software Management This section presents an information model for managing information about software installed on endpoints. Software management encompasses the subset of tasks within security automation and continuous monitoring which play a role in compiling inventories of software products installed on endpoints and transmitting those inventories to software inventory data consumers. Software inventory data consumers may store the software inventory data and/or perform enterprise-level software inventory-related security posture assessments. While software enforcement policies that are invoked and enforced at the time of installation or execution are out of the scope of SACM, they require the same software guidance information to be produced and exchanged. For that reason, the first operation is to Define Guidance making it available to perform a posture evaluation activity during any operation. 7.1. Core Information Needs Unique Endpoint Identifier: Organizations need to be able to relate the instances of software to the endpoint on which it is installed. This should be consistent with the identification of the endpoint when it is enrolled (see Section 6.4.1). Unique Software Identifier: Organizations need to be able to uniquely identify and label software installed on an endpoint. Specifically, they need to know the name, publisher, unique ID, and version; and any related patches. In some cases the software's identity might be known a priori by the organization; in other cases, a software identity might be first detected by an organization when the software is first inventoried in an operational environment. Due to this, it is important that an organization have a stable and consistent means to identify software found during collection. Waltermire & Watson Expires January 4, 2015 [Page 15] Internet-Draft Endpoint Assessment Information Model July 2014 7.2. Process Area Description The authors envisage an automated capability that will collect software inventory data on endpoints and transmit that data to interested software inventory data consumers. "Inventory data" is information about software products that may be operating systems, end-user applications, or other software-based systems and services- that have been added to, removed from, or modified on endpoints. The collection and transmission of inventory data from a given endpoint to an inventory data consumer may be scheduled, event-driven, or on demand based on the requested collection policy. That is, endpoints may be configured to collect and transmit inventory data on a predefined schedule (scheduled), in response to a change in inventory state (event-driven), or whenever an inventory consumer requests (on demand). On any occasion when inventory data is collected and transmitted, the data may be "complete" or be an "update" to previously-transmitted data based on the requested collection policy. Inventory data is considered to be "complete" when it is intended to reflect a comprehensive and up-to-date enumeration of all software products that are believed to be installed on a given endpoint as of the time the inventory is taken. Inventory data is considered to be an "update" when the data is limited to documenting only the changes that have occurred since the last complete inventory or update. Ad hoc inventory data requests should also be supported; that is, an inventory data consumer should be able to issue ad hoc queries to an endpoint regarding specific identified products. Endpoints should be able to indicate whether or not an identified product is installed, and should be able to answer various questions about an installed software product including: the date/time it was most recently installed, removed or patched/updated, which patches are installed, and the names and properties (e.g., versions, hashes) of files associated with the product. Queries concerning the configuration of installed software products are addressed separately in Section 7. The author's vision rests on a model of the basic processes involved in software product installation and maintenance. We use the term "inventory event" to refer generically to any of three possible events involving software which may occur on endpoints: (1) installation (adding a software product to an endpoint's inventory), (2) modification (changing any files associated with a previously- installed product, and (3) removal (eliminating a software product from an endpoint's inventory). Under this model, each endpoint may support (but is not required to) a resident "inventory manager". If present, the inventory manager is an installed software product which provides a standard interface to Waltermire & Watson Expires January 4, 2015 [Page 16] Internet-Draft Endpoint Assessment Information Model July 2014 "product installers", which are specialized software applications that are designed to install, modify or remove other software products on endpoints. Product installers are generally expected to interact with the resident inventory manager, if one is present, but are not required to. By interacting with the inventory manager, product installers notify the inventory manager of any inventory events they are generating, and supply data values needed to characterize the event. When a product installer interacts correctly with a resident inventory manager, we say that it has generated a "conforming inventory event", meaning it has installed, modified or removed a software product in a manner that conforms to the inventory manager's expectations, as defined by its interface. When product installers fail to interact properly with a resident inventory manager, bypass it altogether, or when an inventory manager is not resident, we say that the resulting inventory event is "non- conforming". Additionally in the non-conforming case, a resident inventory manager may also monitor a filesystem or other installation contexts to detect changes to software to characterize the nature of the change. On Linux systems, RPM and DPKG are examples of inventory managers. Each provides a standard product installer application which parses specially-formatted package files, updates the RPM/APT database, and copies files to their intended locations. We require that when a product installer generates a conforming inventory event, the resident inventory manager shall update a local inventory data store on the endpoint. The local inventory data store must maintain an up-to-date record of all software products installed on the endpoint. The local inventory data store should also maintain a record of all inventory events, including product modifications and removals for later collection. The resident inventory manager should have the ability to provide event-driven notification to other software systems, to support reporting of inventory change events as soon as they occur. Because our model allows for non-conforming inventory events, as well as for situations in which an endpoint does not support a resident inventory manager, we allow for some number of "endpoint scanners" to access endpoints either directly (by being resident on the endpoint and by having the privileges necessary to inspect all areas of the endpoint where installed software may be present) or indirectly (e.g., by monitoring network traffic between the endpoint and other devices), and attempt to infer inventory events which may have happened at some point in the past. Compiling a complete and accurate inventory of software products installed on an endpoint thus involves collecting information about Waltermire & Watson Expires January 4, 2015 [Page 17] Internet-Draft Endpoint Assessment Information Model July 2014 conforming as well as non-conforming inventory events. Information about conforming inventory events is obtained from the resident inventory manager, if present. Information about non-conforming inventory events is obtained by running one or more endpoint scanners. These tasks are performed by an "inventory producer", which may or may not reside on an endpoint, but which is able to interact with any resident inventory manager, as well as to initiate scans using available endpoint scanners. Inventory producers transmit collected inventory data to one or more inventory consumers, which may store that data in repositories for later assessment, perform assessments on that data directly, or both. The collection and transmission of software inventory data is needed to enable assessment of security posture attributes associated with software inventory. For example, an enterprise may need to assess compliance with "whitelists" (lists of software products authorized for use on network devices) and "blacklists" (lists of specifically prohibited products). For another example, an enterprise may need to assess whether a software product with a publicly disclosed vulnerability is installed on any endpoint within the network. 7.3. Software Management Process Operations The following operations are necessary to carry out activities within the Software Management domain: 1. Define Guidance: Add software to or remove software from one of three lists for an endpoint. Those lists are software allowed to be installed, software prohibited from being installed, and software that is mandatory for installation. 2. Collect Inventory: Prepare and deliver a "complete" or "update" inventory report to one or more interested inventory data consumers, or respond to an ad hoc request for inventory data about one or more software products. 3. Evaluate Software Inventory Posture: Based on guidance, evaluate the current software inventory and determine compliance with applicable security policies or identify conditions of interest. 4. Report Evaluation Results: Based on Guidance, report evaluation results to interested report consumers. 7.4. Information Model Requirements In this section we describe the data that enterprises will need to carry out each Software Management operation. Waltermire & Watson Expires January 4, 2015 [Page 18] Internet-Draft Endpoint Assessment Information Model July 2014 7.4.1. Define Guidance The "Define Guidance" operation involves the Software Inventory Collection and Evaluation Guidance Capabilities. The Collection Guidance Capability generates or maintains the guidance related to when Software Inventory should be collected (e.g., periodic or when the inventory changes on an endpoint) and what type of collection (partial or full) should occur at that time. The Evaluation Guidance Capability generates or maintains the guidance associated with software items where each item is categorized as one of "mandatory", "optional", or "prohibited" for a set of endpoints. A product is "mandatory" if it must be installed on every compatible endpoint. A product is "optional" if it is allowed to be installed on compatible endpoints. A product is "prohibited" if it must not be installed on any compatible endpoints. The Collection and Evaluation Guidance Capabilities have the following information needs: o (M) Unique Software Identifier: the software product which is the subject of the request must be identified. o (M) Authorization Status: the authorization status of the product (one of 'mandatory', 'optional', or 'prohibited') must be supplied. o (O) Software Footprint: hashes of the software footprint (or a pointer to those values) may be used to determine if software is corrupted or tampered with. This operation results in a change to the Collection or Evaluation Guidance. This may, but need not, trigger an automatic enterprise- wide assessment. 7.4.2. Collect Inventory The "Collect Inventory" operation involves the following architecture components: o Software Inventory Collection Capability o Software Inventory Collection Producers and Consumers o Software Inventory Collection Guidance Capability Waltermire & Watson Expires January 4, 2015 [Page 19] Internet-Draft Endpoint Assessment Information Model July 2014 The Software Inventory Collection Capability has the following information needs: o (M) Collection Guidance: the requester must indicate when to perform a collection (e.g., at a set time, in response to a change in inventory) and provide any other relevant guidance. o (M) Request Type: the requester must indicate whether a "complete" or "update" inventory should be performed and transmitted. o (M) Unique Endpoint Identifier: the endpoint whose software inventory data is to be collected must be identified. o (M) Endpoint Software Inventory: a description of the current list of software on the endpoint must be supplied. At the completion of the Collect Inventory operation the Software Inventory Producer will send an enumeration of installed software to the appropriate Software Inventory Collection Consumer(s) or Repositories. 7.4.3. Evaluate Software Inventory Posture The "Evaluate Software Inventory Posture" operation involves the following architecture components: o Software Inventory Evaluation Capability o Software Inventory Collection Consumers o Software Inventory Evaluation Producers and Consumers o Software Inventory Evaluation Guidance Capability The Software Inventory Evaluation Capability is the component which compares endpoint inventory information to current security guidance, and notes any deviations from what is expected. The Software Inventory Evaluation Capability has the following information needs: o (M) Endpoint Identifier: the endpoint whose inventory posture is to be assessed must be identified. o (M) Endpoint Software Inventory: a description of the current software inventory of the endpoint must be supplied. Waltermire & Watson Expires January 4, 2015 [Page 20] Internet-Draft Endpoint Assessment Information Model July 2014 o (M) Software Inventory Evaluation Guidance: all guidance pertinent to performing an evaluation or assessment of software inventory must be supplied. o (M) Software Inventory Evaluation Results: the results from the evaluation of the software inventory against the guidance. The outcome of this operation is that the Software Inventory Evaluation Capability identifies any deviations from guidance related to the current inventory of software products installed on the endpoint. 7.4.4. Report Evaluation Results The "Report Evaluation Results" operation involves the following architecture component: o Software Inventory Results Report Capability o Software Inventory Results Producers and Consumers o Software Inventory Evaluation Consumer o Software Inventory Reporting Guidance Capability The Software Inventory Results Report Capability has the following information needs: o (M) Endpoint Identifier: the endpoint whose inventory posture assessment is to be reported must be identified. o (M) Software Inventory Reporting Guidance: all guidance pertinent to generating and reporting software inventory assessment results must be supplied. o (M) Software Inventory Evaluation Results: the results from the evaluation of the software inventory must be supplied. o (M) Software Inventory Results Report: the report generated by applying the reporting guidance to the evaluation results. 8. Configuration Management This section presents an information model for the collection of software configuration posture attributes. Software configuration collection encompasses the subset of tasks within security automation and continuous monitoring involved in the collection of important settings from an endpoint and transmitting those settings to posture Waltermire & Watson Expires January 4, 2015 [Page 21] Internet-Draft Endpoint Assessment Information Model July 2014 attribute data consumers that store the data and/or perform enterprise-level security posture assessments. For example, an operating system may enforce configurable password complexity policies. As part of assessing the current complexity requirements of an network of endpoints using this operating system, a SACM tool will interact with the operating systems using a standardized protocol to retrieve the value of the minimum password length setting. This tool will then verify that each endpoint's minimum password length setting meets (or potentially exceeds) the organizational requirement, and will then report inconsistent endpoints to the responsible administrator for action. 8.1. Core Information Needs Unique Endpoint Identifier: We need to be able to relate the posture attribute data and assessment results with an endpoint. Configuration Item Identifier: We need to be able to uniquely identify high-level, cross-platform, configuration statements that can be interpreted and mapped to low-level, platform- specific configuration settings by primary source vendors for their platforms. Platform Configuration Item Identifier(s): We need to know what low- level configuration items map to the high-level configuration items in order to collect posture attribute data from specific platforms. Posture Attributes: We need to be able to represent posture attribute data collected from an endpoint for use in assessments. 8.2. Process Area Description In addition to the compilation of endpoint inventory data, there is a need to compile and assess posture attribute values from an endpoint to ensure that software on an endpoint has been configured correctly. The configuration management security automation domain encompasses a wide range of information needs to define, collect, and evaluate posture attributes across a myriad of operating systems, applications, and endpoint types. Configuration Management requires that there is guidance about how and what to collect. The establishment of all of this guidance is something that needs to be done before the assessment trigger event, and needs to be done in a way that can scale and be sustained over the lifecycle of the applicable software products. Waltermire & Watson Expires January 4, 2015 [Page 22] Internet-Draft Endpoint Assessment Information Model July 2014 8.2.1. The Existence of Configuration Item Guidance The model for software configuration collection relies on two main components: (1) the identification of software configuration items and (2) the representation of posture attribute data from an endpoint. For the identification of software configuration items, the primary objective is for the security community to develop a high-level, cross-platform identifier known as a "configuration item" that can then be used by primary source vendors to map to low-level configuration settings called "platform configuration items" for their software. The benefits of this are that a single organization is not responsible for maintaining the entire set of configuration items for all platforms and the primary source vendors are given the flexibility to determine what a particular configuration item means for their software. From a practical perspective, this will likely require a set of federated registries for both "configuration items" and "platform configuration items". An example is the "configuration item" that code should not automatically run when introduced into a system without some entity intentionally invoking that operation. One associated "platform configuration item" for Windows could be disable autorun. With regards to the model that represents posture attribute data from an endpoint, there are three components: (1) the linking of posture attribute data to the specific endpoint from which it was collected, (2) a generic posture attribute that can be extended by primary source vendors, and (3) the actual extensions of this generic posture attribute by primary source vendors for their platforms. The first component known as "collected posture attributes" associates the posture attribute data collected from an endpoint with that endpoint through the use of the Unique Endpoint Identifier previously mentioned in this model. It may also include other metadata about the collection such as the timestamp and what Posture Attribute Producer was used. This model can be used as the payload for messages of standardized protocols that are responsible for transmitting and receiving posture attribute data between Posture Attribute Producers and Consumers. The second component is the "posture attribute" which provides metadata common to all posture attributes regardless of how they are extended to meet the needs of a particular platform. Finally, the third component is the "platform posture attribute" which is left as an extension point for primary source vendors to fulfill the posture attribute data needs for their platforms. With this model, they can not only define the structure of the posture attribute value data and the data type, but they can also specify any additional metadata about the posture attribute that they feel is relevant for posture attribute data consumers. Waltermire & Watson Expires January 4, 2015 [Page 23] Internet-Draft Endpoint Assessment Information Model July 2014 8.2.2. Configuration Collection Guidance Collection guidance provides instructions on how to collect posture attributes from an endpoint and may include information such as a list of posture attributes that may be collected, a list of posture attributes to collect, and metadata that provides details on when and how to collect posture attributes on an endpoint. The development of this guidance is best performed by the primary source vendor who is the most authoritative source of information for a specific platform. However, if necessary, organizations can translate that information into actionable collection guidance. 8.2.3. Configuration Evaluation Guidance With the ability to identify and collect configuration items on an endpoint, the next logical step is to assess the collected posture attribute data against some known state or to check for some specific conditions of interest. This will require the creation of evaluation guidance. Evaluation guidance provides instructions on how to evaluate collected posture attributes from an endpoint against a known state or condition of interest. These instructions will express policies and conditions of interest using logical constructs, requirements for the data used during evaluation (age of data, source of data, etc.), and references to human-oriented data that provides technical, organizational, and other policy information. The evaluation guidance will need to capture what posture attributes to collect, how to collect the posture attributes (e.g., retrieve it from a configuration database, a published source, or collect it by leveraging collection guidance), any requirements on the usability of the posture attribute data, and instructions on how to evaluate the collected posture attribute data against defined policies or check for a condition of interest. It may also include instructions on the granularity of the results. 8.2.4. Local Configuration Management Process Once an endpoint has been targeted for assessment, the first step involves understanding what collection and evaluation guidance is applicable to the endpoint. After the applicable guidance has been retrieved, the collection of posture attributes can begin. This operation may result in the collection of a subset of posture attributes on the endpoint or all of the posture attributes on the endpoint using a variety of collection mechanisms. Waltermire & Watson Expires January 4, 2015 [Page 24] Internet-Draft Endpoint Assessment Information Model July 2014 The three primary collection mechanisms include: (1) retrieving posture attribute data from a enterprise repository, (2) software publishing their posture attribute data, and (3) collecting the posture attribute data from the endpoint or some intermediary. When executing collection guidance, it may be necessary to use some combination of these mechanisms to get all of the required data and it may be necessary to authenticate with the endpoint, CMDB, or some other data store. In this model, a posture attribute producer may compile and transmit posture attribute data to a posture attribute consumer on any of these occasions: (1) upon request by a posture attribute consumer; (2) on a predefined schedule (e.g. daily, weekly, etc.); or (3) after some event has occurred (e.g. posture attribute value change detected, etc.). Once the posture attributes have been collected, evaluation guidance is used to assess how the collected posture attribute data compares with the predefined policies or whether or not the endpoint contains conditions of interest. This operation may occur locally on the endpoint or as part of an application that interacts with an intermediate or back end server. Once completed, the results of the evaluation can be transmitted to the designated locations with the appropriate level of granularity (e.g. results for individual "platform configuration items" or rolled up results for a "configuration item", etc.). Depending on the sensitivity of the evaluation results and collected posture attribute data, it may be necessary for locations receiving the results to authenticate with the sending endpoint and potentially even utilize a secure communication channel. Last in the configuration management lifecycle is the "adjudicate further action" operation where the results are processed and it is determined if follow up actions are necessary and by which parties. Further actions could include modifying configuration settings, expelling or quarantining an endpoint from the enterprise network, removing or upgrading installed software, generating an alert, documenting a compliance deviation, or performing some mitigation among other things. This may also require re-initiating the assessment process to ensure follow up actions were completed successfully. By clearly marking the line between collection and evaluation, tools are free to implement these steps in the way that best suits the needs of the end users and that allows for flexibility and scalability across organizations of all size and shape. Waltermire & Watson Expires January 4, 2015 [Page 25] Internet-Draft Endpoint Assessment Information Model July 2014 This model allows software and hardware vendors to publish posture attributes in both proactive and reactive manners to centralized repositories for evaluation. This type of flexibility is crucial for scalable security automation in a large and diverse enterprise environment. Finally, with this described model, key stakeholders will be able to quickly and dynamically construct and execute new or updated policy to enable fast and accurate posture evaluation. Further, evaluation need not be constrained to a single repository of information (be it an endpoint or a central repository). Evaluation can occur across multiple repositories of information to reach an aggregated decision on security posture. 8.3. Configuration Management Operations The Configuration Management security automation domain includes all of the processes involved in monitoring the configuration of endpoints on an enterprise network. We have defined the following operations within the configuration management domain: 1. Define Guidance: Define or acquire cross-platform configuration item guidance, platform-specific configuration item guidance, collection guidance, and evaluation guidance as applicable for the endpoints that need to be assessed. This may also include verifying the integrity of the guidance. 2. Collect Posture Attributes: Gather the needed posture attributes from the endpoint and report them to one or more interested posture attribute consumers. The collection of posture attributes can be initiated by a number of triggers and can be gathered using a variety of collection mechanisms. 3. Evaluate Posture Attributes: Based on guidance, assess the collected posture attributes from the endpoint to determine compliance with applicable security policies or identify conditions of interest. 4. Report Evaluation Results: Based on guidance, transmit the assessment results to interested report consumers with the appropriate level of granularity. 8.4. Information Model Requirements 8.4.1. Define Guidance The "Define Guidance" operation relies on the existence and population of three types of guidance data: (1) configuration item Waltermire & Watson Expires January 4, 2015 [Page 26] Internet-Draft Endpoint Assessment Information Model July 2014 guidance (cross-platform guidance and platform-specific guidance), (2) collection guidance, and (3) evaluation guidance. Therefore, the "Define Guidance" operation involves the Configuration Item, Collection, Evaluation, and Reporting Guidance Capabilities architecture components. These components generate or store information about configuration items and posture attributes, including when and how to collect them. They also include how to evaluate the collected attributes and rules around reporting the results for the desired configuration posture assessments on applicable endpoints. Configuration Guidance Capabilities can initiate requests to acquire guidance from existing data stores or have the information manually added, modified, or deleted. To express configuration item guidance, the following information is needed: o (M) Configuration Item Identifier: A persistent, unique identifier for the configuration item o (M) Configuration Item Description: A high-level description of the configuration item To express platform configuration item guidance, the following information may be needed ('M' indicates 'mandatory' and 'O' indicates 'optional'): o (M) Platform Configuration Item Identifier: A persistent, unique identifier assigned by the primary source vendor o (O) A reference to the unique, persistent configuration item identifier o (M) Platform Configuration Item Identifier Description: A low- level description of the configuration item for the specific platform o (M) Posture Attributes: A list of posture attributes that correspond to the platform configuration item o (M) References that provide additional details about the platform configuration item o (O) Additional metadata that the primary source vendor feels is relevant Waltermire & Watson Expires January 4, 2015 [Page 27] Internet-Draft Endpoint Assessment Information Model July 2014 o (O) References to collection and evaluation guidance that they may have developed or that someone else has developed on their behalf To express "collection" guidance, the following information is needed: o (M) Listings of posture attribute identifiers for which values may be collected and evaluated o (M) Lists of attributes that are to be collected o (O) Metadata that includes when to collect attributes (e.g. based on interval, event, duration of collection), how and where to collect the posture attribute data (e.g. CMDB, publish, collect from endpoint or other data source, etc.) To express "evaluation" guidance, the following information is needed: o (M) Logical constructs to express policies and conditions of interest as well as the ability to ask different questions such as "what is the value?", "is a configuration item compliant with a policy?", etc. o (M) Data requirements including the age of the data and the source of the data among other things o (O) References to human-oriented data that provides technical, organizational, and other policy information This operation results in a change in posture assessment guidance and may, but need not, trigger an automatic enterprise-wide assessment. 8.4.2. Collect Posture Attributes Operation The "Collect Posture Attributes" operation involves the following architectural components: o Configuration Collection Producers and Consumers o Configuration Collection Capability o Configuration Collection Guidance Capability The Configuration Collection Capability has the following information needs: Waltermire & Watson Expires January 4, 2015 [Page 28] Internet-Draft Endpoint Assessment Information Model July 2014 o (M) Collection Guidance: the requester must indicate when to perform a collection (e.g., at a set time, in response to a change in inventory) and provide any other relevant guidance. This guidance may include information such as what posture attributes to collect, how to collect the posture attributes, and whether or not the posture attribute data should be persisted for later use. o (M) Request Type: the requester must indicate whether a "full" or "partial" posture attribute collection should be performed. o (M) Unique Endpoint Identifier: the endpoint whose posture attribute data is to be collected must be identified. o (M) Collected Posture Attributes: the collected posture attribute data to include any required metadata. At the completion of the Collect Posture Attributes operation the Configuration Collection Producer will send the posture attributes and their values to the appropriate Software Inventory Consumer(s) or Repositories. 8.4.3. Evaluate Posture Attributes Operation The "Evaluate Posture Attributes" operation involves the following information elements: o Configuration Evaluation Capability o Configuration Collection Consumers o Configuration Evaluation Producers and Consumers> o Configuration Evaluation Guidance Capability The Configuration Evaluation Capability is the component which compares collected posture attribute data to current evaluation guidance, and notes any deviations. The Configuration Evaluation Capability has the following information needs: o (M) Unique Endpoint Identifier: the endpoint whose inventory posture is to be assessed must be identified. o (M) Collected Posture Attributes: the collected posture attribute data to assess must be supplied or retrieved prior to performing the assessment. Waltermire & Watson Expires January 4, 2015 [Page 29] Internet-Draft Endpoint Assessment Information Model July 2014 o (M) Configuration Evaluation Guidance: all guidance pertinent to performing an evaluation of posture attribute data must be supplied. o (M) Configuration Evaluation Results: the results from the evaluation of the collected posture attributes against the guidance. 8.4.4. Report Evaluation Results Operation The "Report Evaluation Results" operation involves the following architecture components: o Configuration Results Report Capability o Configuration Results Producers and Consumers o Configuration Evaluation Consumers o Configuration Reporting Guidance Capability The Configuration Results Report Capability has the following information needs: o (M) Unique Endpoint Identifier: the endpoint whose posture attribute evaluation results are to be reported must be identified. o (M) Posture Assessment Reporting Guidance: all guidance pertinent to generating and reporting posture attribute evaluation results must be supplied. This includes information such as the level of granularity provided within the report (e.g. rolled up to the "configuration item" level, at the "platform configuration level", raw posture attribute data, or the results of evaluating the posture attribute data against a known state), assurance of the evaluation results, etc. o (M) Configuration Evaluation Results: the results from the evaluation of the collected posture attributes against the guidance. o (M) Configuration Results Report: the report generated by applying the reporting guidance to the evaluation results. The outcome of this operation is that the Configuration Results Producer reports posture assessment results to interested Configuration Results Consumers. Waltermire & Watson Expires January 4, 2015 [Page 30] Internet-Draft Endpoint Assessment Information Model July 2014 9. Vulnerability Management This section presents an information model for discovering extant vulnerabilities within an enterprise network due to the presence of installed software with publicly disclosed vulnerabilities. Successful vulnerability management builds on the foundation laid by endpoint management, software management, and configuration management, discussed earlier. We limit the scope of vulnerability management to the identification of vulnerable software. We do not currently consider mitigation or remediation of identified vulnerabilities within scope; these are important topics that deserve careful attention. Furthermore, we recognize that the mere presence of installed software with publicly disclosed vulnerabilities does not necessarily mean than an enterprise network is vulnerable to attack, as other defensive layers may effectively preclude exploits. This will also need to be considered within the scope of an information model that supports mitigation/remediation operations. 9.1. Core Information Needs Unique Endpoint Identifiers: Each endpoint within the enterprise network must have a unique identifier so we can relate instances of vulnerable software to the endpoint(s) on which they are installed. Unique Software Identifiers: Organizations need to be able to uniquely identify and label each software product installed on an endpoint. This label must identify software to the version/patch level and must have enough fidelity to be able to associate it with other authoritative data (e.g., a listing of the hashes of the executables that are associated with the software). Unique Vulnerability Identifiers: Organizations need to be able to uniquely identify and label publicly disclosed software vulnerabilities, and associate those labels with the unique software identifiers of the software product(s) containing those vulnerabilities. 9.2. Process Area Description The authors envisage that software publishers produce "vulnerability reports" about their products. They also envisage that each "vulnerability report" is disseminated in a standard format suitable for machine consumption and processing. Each vulnerability report shall include, at a minimum, a unique identifier of the vulnerability, and a machine-readable "applicability statement" which associates the vulnerability with one or more software products and Waltermire & Watson Expires January 4, 2015 [Page 31] Internet-Draft Endpoint Assessment Information Model July 2014 any pertinent product configuration information. One day, it will be possible for a machine to authenticate the source of a vulnerability report and confirm that the report has not been tampered with. Each enterprise network shall support automated processes to recognize when new vulnerability reports are available from software publishers, and to retrieve and process those reports within the context of the enterprise network. The outcome of vulnerability management processes is the generation of reports and/or alerts that identify vulnerable endpoints. 9.3. Vulnerability Management Process Operations We have identified the following operations necessary to carry out activities within the Vulnerability Management domain: 1. Collect Vulnerability Reports: Retrieve newly-published vulnerability reports from software publishers. 2. Evaluate Vulnerability Posture: Based on guidance, assess the current software inventory and configuration data and determine compliance with applicable security policies or identify conditions of interest. 3. Report Evaluation Results: Based on guidance, report evaluation results to interested report consumers. 9.4. Information Model Requirements In this section we describe the data that enterprises will need to carry out each Vulnerability Management operation. 9.4.1. Collect Vulnerability Reports The "Collect Vulnerability Reports" operation involves the following architectural components: o Vulnerability Collection Guidance Capability o Vulnerability Evaluation Guidance Capability The Vulnerability Collection Guidance Capability maintains a repository of the guidance associated with what information and attributes must be either collected or can be reused from other posture assessment collection operations. Waltermire & Watson Expires January 4, 2015 [Page 32] Internet-Draft Endpoint Assessment Information Model July 2014 The Vulnerability Evaluation Guidance Capability creates the evaluation guidance associated with individual vulnerability reports. Each vulnerability report documents a publicly disclosed software vulnerability, and associates a unique vulnerability identifier with one or more unique identifiers of affected software products. The Vulnerability Evaluation Guidance Capability may be maintained by a software publisher or it may be a repository that aggregates vulnerability reports across multiple software publishers. If it is an aggregate repository, it will also maintain a list of publishers of vulnerability reports. The Vulnerability Evaluation Guidance Capability may operate in either (or both) "push" or "pull" modes. In "push" mode, publishers of vulnerability reports initiate contact whenever a new report becomes available. In "pull" mode, the Vulnerability Evaluation Guidance Capability initiates contact with publishers, either on a scheduled basis or in response to a request originating within the enterprise network. The Vulnerability Evaluation Guidance Capability has the following information needs: o (O) Publisher List: a new list of publishers may be supplied. o (O) Update Schedule: a new schedule for pull-mode operations may be supplied. This operation may result in a change to the enterprise's collection of vulnerability reports. This may, but need not, trigger an automatic enterprise-wide vulnerability posture assessment. 9.4.2. Evaluate Vulnerability Posture The "Evaluate Vulnerability Posture" operation involves the following architectural components: o Vulnerability Evaluation Capability o Vulnerability Collection Guidance Capability o Vulnerability Evaluation Guidance Capability o Software Inventory Collection Producers o Configuration Evaluation or Reporting Producers (when appropriate) Waltermire & Watson Expires January 4, 2015 [Page 33] Internet-Draft Endpoint Assessment Information Model July 2014 o Vulnerability Evaluation Consumers The Vulnerability Evaluation Capability is the component which compares information about publicly disclosed software vulnerabilities with current information about software installed within the enterprise network. When a vulnerability can be mitigated by a particular configuration, evaluation and/or reporting results can be used to determine the vulnerability posture of the endpoint. A vulnerability posture assessment may be triggered in response to any of (a) the arrival of a new vulnerability report, (b) a change in software inventory on any enterprise endpoint, or (c) a change in the configuration of any software product installed on any enterprise endpoint. This information is managed by the Vulnerability Collection Guidance Capability. The Vulnerability Evaluation Capability has the following information needs: o (M) Unique Vulnerability Identifier(s): the unique identifier of the vulnerability to be reported on must be supplied. o (M) Unique Endpoint Identifiers: the endpoints being assessed for vulnerabilities must be identified. o (M) Vulnerability Collection Guidance: all guidance pertinent to determining what previously collected posture data to use must be supplied. o (M) Endpoint Software Inventory: a description of the current software inventory of the endpoint must be supplied. o (M) Configuration Evaluation Results: the results from the evaluation of the collected posture attributes against the guidance. OR (M) Configuration Results Report: the report generated by applying the reporting guidance to the evaluation results. Configuration Posture Assessment Results would only be required when the vulnerability can be mitigated by configuring an installed instance of software in a particular manner. o (M) Vulnerability Evaluation Guidance: all guidance pertinent to performing an evaluation of posture data must be supplied. o (M) Vulnerability Evaluation Results: the results from the evaluation of the collected posture data against the guidance. The outcome of this operation is that a vulnerability report is delivered to Vulnerability Evaluation Consumers. Waltermire & Watson Expires January 4, 2015 [Page 34] Internet-Draft Endpoint Assessment Information Model July 2014 9.4.3. Report Evaluation Results The "Report Evaluation Results" operation involves the following architectural components: o Vulnerability Results Report Capability o Vulnerability Results Producers and Consumers o Vulnerability Evaluation Consumers o Vulnerability Reporting Guidance Capability The Vulnerability Results Report Capability has the following information needs: o (M) Vulnerability Evaluation Results: the results from the evaluation of the collected posture data against the guidance. o (M) Unique Endpoint Identifiers: the vulnerable endpoints must be identified. o (M) Vulnerability Assessment Reporting Guidance: all guidance pertinent to generating and reporting vulnerability assessment results must be supplied (e.g., an alert should be generated). o (M) Vulnerability Results Report: the report generated by applying the reporting guidance to the evaluation results. The outcome of this operation may be a report, an alert, or set of reports and/or alerts, identifying any vulnerable endpoints within the enterprise network. 10. From Information Needs to Information Elements The previous sections highlighted information needs for a set of management process areas that use posture assessment to achieve organizational security goals. A single information need may be made up of multiple information elements. Some information elements may be required for two different process areas, resulting in two different requirements. In an effort to support the main idea of collect once and reuse the data to support multiple processes, we try to define a singular set of information elements that will support all the associated information needs. Waltermire & Watson Expires January 4, 2015 [Page 35] Internet-Draft Endpoint Assessment Information Model July 2014 11. Information Model Elements Traditionally, one would use the SACM architecture to define interfaces that required information exchanges. Identified information elements would then be based on those exchanges. Because the SACM architecture document is still in the personal draft stage, this information model uses a different approach to the identification of information elements. First it lists the four main endpoint posture assessment activities. Then it identifies management process areas that use endpoint posture assessment to achieve organizational security objectives. These process areas were then broken down into operations that mirrored the typical workflow from the SACM Use Cases draft [I-D.ietf-sacm-use-cases]. These operations identify architectural components and their information needs. In this section, information elements derived from those information needs are mapped back to the four main activities listed above. The original liaison statement [IM-LIAISON-STATEMENT-NIST] requested contributions for the SACM information model in the four areas described below. Based on the capabilities defined previously in this document, the requested areas alone do not provide a sufficient enough categorization of the necessary information model elements. The following sub-sections directly address the requested areas as follows: 1. Endpoint Identification A. Section 11.1 Asset Identifiers: Describes identification of many different asset types including endpoints. 2. Endpoint Characterization A. Section 11.3 Endpoint characterization: This directly maps to the requested area. 3. Endpoint Attribute Expression/Representation A. Section 11.4 Posture Attribute Expression: This corresponds to the first part of "Endpoint Attribute Expression/ Representation." B. Section 11.5 Actual Value Representation: This corresponds to the second part of "Endpoint Attribute Expression/ Representation." 4. Policy evaluation expression and results reporting Waltermire & Watson Expires January 4, 2015 [Page 36] Internet-Draft Endpoint Assessment Information Model July 2014 A. Section 11.6 Evaluation Guidance: This corresponds to the first part of "Policy evaluation expression and results reporting." B. Section 11.7 Evaluation Result Reporting: corresponds to the second part of "Policy evaluation expression and results reporting." Additionally, Section 11.2 Other Identifiers: describes other important identification concepts that were not directly requested by the liaison statement. Per the liaison statement, each subsection references related work that provides a basis for potential data models. Some analysis is also included for each area of related work on how directly applicable the work is to the SACM efforts. In general, much of the related work does not fully address the general or use case-based requirements for SACM, but they do contain some parts that can be used as the basis for data models that correspond to the information model elements. In these cases additional work will be required by the WG to adapt the specification. In some cases, existing work can largely be used in an unmodified fashion. This is also indicated in the analysis. Due to time constraints, the work in this section is very biased to previous work supported by the authors and does not reflect a comprehensive listing. An attempt has been made where possible to reference existing IETF work. Additional research and discussion is needed to include other related work in standards and technology communities that could and should be listed here. The authors intend to continue this work in subsequent revisions of this draft. Where possible when selecting and developing data models in support of these information model elements, extension points and IANA registries SHOULD be used to provide for extensibility which will allow for future data models to be addressed. 11.1. Asset Identifiers In this context an "asset" refers to "anything that has value to an organization" (see [NISTIR-7693]). This use of the term "asset" is broader than the current definition in [I-D.ietf-sacm-terminology]. To support SACM use cases, a number of different asset types will need to addressed. For each type of asset, one or more type of asset identifier will be needed for use in establishing contextual relationships within the SACM information model. The following asset types are referenced or implied by the SACM use cases: Waltermire & Watson Expires January 4, 2015 [Page 37] Internet-Draft Endpoint Assessment Information Model July 2014 Endpoint: Identifies an individual endpoint for which posture is collected and evaluated. Hardware: Identifies a given type of hardware that may be installed within an endpoint. Software: Identifies a given type of software that may be installed within an endpoint. Network: Identifies a network for which a given endpoint may be connected or request a connection to. Organization: Identifies an organizational unit. Person: Identifies an individual, often within an organizational context. 11.1.1. Related Work 11.1.1.1. Asset Identification The Asset Identification specification [NISTIR-7693] is an XML-based data model that "provides the necessary constructs to uniquely identify assets based on known identifiers and/or known information about the assets." Asset identification plays an important role in an organization's ability to quickly correlate different sets of information about assets. The Asset Identification specification provides the necessary constructs to uniquely identify assets based on known identifiers and/or known information about the assets. Asset Identification provides a relatively flat and extensible model for capturing the identifying information about a one or more assets, and also provides a way to represent relationships between assets. The model is organized using an inheritance hierarchy of specialized asset types/classes (see Figure 4), providing for extension at any level of abstraction. For a given asset type, a number of properties are defined that provide for capturing identifying characteristics and the referencing of namespace qualified asset identifiers, called "synthetic IDs." Waltermire & Watson Expires January 4, 2015 [Page 38] Internet-Draft Endpoint Assessment Information Model July 2014 The following figure illustrates the class hierarchy defined by the Asset Identification specification. asset +-it-asset | +-circuit | +-computing-device | +-database | +-network | +-service | +-software | +-system | +-website +-data +-organization +-person Figure 4: Asset Identification Class Hierarchy Waltermire & Watson Expires January 4, 2015 [Page 39] Internet-Draft Endpoint Assessment Information Model July 2014 This table presents a mapping of notional SACM asset types to those asset types provided by the Asset Identification specification. +--------------+------------------+---------------------------------+ | SACM Asset | Asset | Notes | | Type | Identification | | | | Type | | +--------------+------------------+---------------------------------+ | Endpoint | computing-device | This is not a direct mapping | | | | since a computing device is not | | | | required to have network- | | | | connectivity. Extension will be | | | | needed to define a directly | | | | aligned endpoint asset type. | +--------------+------------------+---------------------------------+ | Hardware | Not Applicable | The concept of hardware is not | | | | addressed by the asset | | | | identification specification. | | | | An extension can be created | | | | based on the it-asset class to | | | | address this concept. | +--------------+------------------+---------------------------------+ | Software | software | Direct mapping. | +--------------+------------------+---------------------------------+ | Network | network | Direct mapping. | +--------------+------------------+---------------------------------+ | Organization | organization | Direct mapping. | +--------------+------------------+---------------------------------+ | Person | person | Direct mapping. | +--------------+------------------+---------------------------------+ Table 1: Mapping of SACM to Asset Identification Asset Types This specification has been adopted by a number of SCAP validated products. It can be used to address asset identification and categorization needs within SACM with minor modification. 11.1.2. Endpoint Identification An unique name for an endpoint. This is a foundational piece of information that will enable collected posture attributes to be related to the endpoint from which they were collected. It is important that this name either be created from, provide, or be associated with operational information (e.g., MAC address, hardware certificate) that is discoverable from the endpoint or its communications on the network. It is also important to have a method of endpoint identification that can persist across network sessions to allow for correlation of collected data over time. Waltermire & Watson Expires January 4, 2015 [Page 40] Internet-Draft Endpoint Assessment Information Model July 2014 11.1.2.1. Related Work The previously introduced asset identification specification (see Section 11.1.1.1 provides a basis for endpoint identification using the "computing-device" class. While the meaning of this class is broader than the current definition of an endpoint in the SACM terminology [I-D.ietf-sacm-terminology], either that class or an appropriate sub-class extension can be used to capture identification information for various endpoint types. 11.1.3. Software Identification A unique name for a unit of installable software. Software names should generally represent a unique release or installable version of software. Identification approaches should allow for identification of commercially available, open source, and organizationally developed custom software. As new software releases are created, a new software identifier should be created by the releasing party (e.g., software creator, publisher, licensor). Such an identifier is useful to: o Relate metadata that describes the characteristics of the unit of software, potentially stored in a repository of software information. Typically, the software identifier would be used as an index into such a repository. o Indicate the presence of the software unit on a given endpoint. o To determine what endpoints are the targets for an assessment based on what software is installed on that endpoint. o Define guidance related to a software unit that represents collection, evaluation, or other automatable policies. In general, an extensible method of software identification is needed to provide for adequate coverage and to address legacy identification approaches. Use of an IANA registry supporting multiple software identification methods would be an ideal way forward. 11.1.3.1. Related Work While we are not aware of a one-size-fits-all solution for software identification, there are two existing specifications that should be considered as part of the solution set. They are described in the following subsections. Waltermire & Watson Expires January 4, 2015 [Page 41] Internet-Draft Endpoint Assessment Information Model July 2014 11.1.3.1.1. Common Platform Enumeration 11.1.3.1.1.1. Background The Common Platform Enumeration (CPE) [CPE-WEBSITE] is composed of a family of four specification that are layered to build on lower-level functionality. The following describes each specification: 1. CPE Naming: A standard machine-readable format [NISTIR-7695] for encoding names of IT products and platforms. This defines the notation used to encode the vendor, software name, edition, version and other related information for each platform or product. With the 2.3 version of CPE, a second, more advanced notation was added to the original colon-delimited notation for CPE naming. 2. CPE Matching: A set of procedures [NISTIR-7696] for comparing names. This describes how to compare two CPE names to one another. It describes a logical method that ensures that automated systems comparing two CPE names would arrive at the same conclusion. 3. CPE Applicability Language: An XML-based language [NISTIR-7698] for constructing "applicability statements" that combine CPE names with simple logical operators. 4. CPE Dictionary: An XML-based catalog format [NISTIR-7697] that enumerates CPE Names and associated metadata. It details how to encode the information found in a CPE Dictionary, thereby allowing multiple organizations to maintain compatible CPE Dictionaries. The primary use case of CPE is for exchanging software inventory data, as it allows the usage of unique names to identify software platforms and products present on an endpoint. The NIST currently maintains and updates a dictionary of all agreed upon CPE names, and is responsible for ongoing maintenance of the standard. Many of the names in the CPE dictionary have been provided by vendors and other 3rd-parties. While the effort has seen wide adoption, most notably within the US Government, a number of critical flaws have been identified. The most critical issues associated with the effort are: o Because there is no requirement for vendors to publish their own, official CPE names, CPE necessarily requires one or more organizations for curation. This centralized curation requirement ensures that the effort has difficulty scaling. Waltermire & Watson Expires January 4, 2015 [Page 42] Internet-Draft Endpoint Assessment Information Model July 2014 o Not enough primary source vendors provide platform and product naming information. As a result, this pushes too much of the effort out onto third-party groups and non-authoritative organizations. This exacerbates the ambiguity in names used for identical platforms and products and further reduces the utility of the effort. 11.1.3.1.1.2. Applicability to Software Identification The Common Platform Enumeration (CPE) Naming specification version 2.3 defines a scheme for human-readable standardized identifiers of hardware and software products. CPE names are the identifier format for software and hardware products used in SCAP 1.2 and is currently adopted by a number of SCAP product vendors. CPE names can be directly referenced in the asset identification software class (see Section 11.1.1.1.) Although relevant, CPE has an unsustainable maintenance "tail" due to the need for centralized curation and naming-consistency enforcement. Its mention in this document is to support the historic inclusion of CPE as part of SCAP and implementation of this specification in a number of security processes and products. Going forward, software identification (SWID) tags are recommended as a replacement for CPE. To this end, work has been started to align both efforts to provide translation for software units identified using SWID tags to CPE Names. This translation would allow tools that currently use CPE- based identifiers to map to SWID identifiers during a transition period. 11.1.3.1.2. Software Identification (SWID) Tags The software identification tag specification [ISO.19770-2] is an XML-based data model that is used to describe a unit of installable software. A SWID tag contains data elements that: o Identify a specific unit of installable software, o Enable categorization of the software (e.g., edition, bundle), o Identification and hashing of software artifacts (e.g., executables, shared libraries), o References to related software and dependencies, and o Inclusion of extensible metadata. Waltermire & Watson Expires January 4, 2015 [Page 43] Internet-Draft Endpoint Assessment Information Model July 2014 SWID tags can be associated with software installation media, installed software, software updates (e.g., service packs, patches, hotfixes), and redistributable components. SWID tags also provide for a mechanism to relate these concepts to each other. For example, installed software can be related back to the original installation media, patches can be related to the software that they patch, and software dependencies can be described for required redistributable components. SWID tags are ideally created at build-time by the software creator, publisher or licensor; are bundled with software installers; and are deployed to an endpoint during software installation. SWID tags should be considered for two primary uses: 1. As the data format for exchanging descriptive information about software products, and 2. As the source of unique identifiers for installed software. In addition to usage for software identification, a SWID tag can provide the necessary data needed to target guidance based on included metadata, and to support verification of installed software and software media using cryptographic hashes. This added information increases the value of using SWID tags as part of the larger security automation and continuous monitoring solution space. 11.1.4. Hardware Identification Due to the time constraints, research into information elements and related work for identifying hardware is not included in this revision of the information model. 11.2. Other Identifiers In addition to identifying core asset types, it is also necessary to have stable, globally unique identifiers to represent other core concepts pertaining to posture attribute collection and evaluation. The concept of "global uniqueness" ensures that identifiers provided by multiple organization do not collide. This may be handled by a number of different mechanisms (e.g., use of namespaces). 11.2.1. Platform Configuration Item Identifier A name for a low-level, platform-dependent configuration mechanism as determined by the authoritative primary source vendor. New identifiers will be created when the source vendor makes changes to the underlying platform capabilities (e.g., adding new settings, replacing old settings with new settings). When created each Waltermire & Watson Expires January 4, 2015 [Page 44] Internet-Draft Endpoint Assessment Information Model July 2014 identifier should remain consistent with regards to what it represents. Generally, a change in meaning would constitute the creation of a new identifier. For example, if the configuration item is for "automatic execution of code", then the platform vendor would name the low-level mechanism for their platform (e.g., autorun for mounted media). 11.2.1.1. Related Work 11.2.1.1.1. Common Configuration Enumeration The Common Configuration Enumeration (CCE) [CCE] is an effort managed by NIST. CCE provides a unique identifier for platform-specific configuration items that facilitates fast and accurate correlation of configuration items across multiple information sources and tools. CCE does this by providing an identifier, a human readable description of the configuration control, parameters needed to implement the configuration control, various technical mechanisms that can be used to implement the configuration control, and references to documentation that describe the configuration control in more detail. By vendor request, NIST issues new blocks of CCE identifiers. Vendors then populate the required fields and provided the details back to NIST for publication in the "CCE List", a consolidated listing of assigned CCE identifiers and associated data. Many vendors also include references to these identifiers in web pages, SCAP content, and prose configuration guides they produce. CCE the identifier format for platform specific configuration items in SCAP and is currently adopted by a number of SCAP product vendors. While CCE is largely supported as a crowd-sourced effort, it does rely on a central point of coordination for assignment of new CCE identifiers. This approach to assignment requires a single organization, currently NIST, to manage allocations of CCE identifiers which doesn't scale well and introduces sustainability challenges for large volumes of identifier assignment. If this approach is used going forward by SACM, a namespaced approach is recommended for identifier assignment that allows vendors to manage their own namespace of CCE identifiers. This change would require additional work to specify and implement. Waltermire & Watson Expires January 4, 2015 [Page 45] Internet-Draft Endpoint Assessment Information Model July 2014 11.2.1.1.2. Open Vulnerability and Assessment Language 11.2.1.1.2.1. Background The Open Vulnerability and Assessment Language (OVAL(R)) is an XML schema-based data model developed as part of a public-private information security community effort to standardize how to assess and report upon the security posture of endpoints. OVAL provides an established framework for making assertions about an endpoint's posture by standardizing the three main steps of the assessment process: 1. representing the current endpoint posture; 2. analyzing the endpoint for the presence of the specified posture; and 3. representing the results of the assessment. OVAL facilitates collaboration and information sharing among the information security community and interoperability among tools. OVAL is used internationally and has been implemented by a number of operating system and security tools vendors. Waltermire & Watson Expires January 4, 2015 [Page 46] Internet-Draft Endpoint Assessment Information Model July 2014 The following figure illustrates the OVAL data model. +------------+ +-----------------+ | Variables | | Common <---+ | +--------> | +------------+ | | | +------------+ | | <---+ Directives | | +--------^----^---+ | | | | | +--------+---+ | | +-----+ | | | | | | +--------+--------+ | | | | System | | | | | Characteristics | | | +------+------+ | | | +--------v---+ | Definitions | | | | | Results | | | +--------^--------+ +-+ | | | | | | | | +------------+ | +------^------+ +-------+----+ | | +--------------------------------------+ Note: The direction of the arrows indicate a model dependency Figure 5: The OVAL Data Model The OVAL data model [OVAL-LANGUAGE], visualized in Figure 5, is composed of a number of different components. The components are: o Common: Constructs, enumerations, and identifier formats that are used throughout the other model components. o Definitions: Constructs that describe assertions about system state. This component also includes constructs for internal variable creation and manipulation through a variety of functions. The core elements are: * Definition: A collection of logical statements that are combined to form an assertion based on endpoint state. * Test(platform specific): A generalized construct that is extended in platform schema to describe the evaluation of expected against actual state. Waltermire & Watson Expires January 4, 2015 [Page 47] Internet-Draft Endpoint Assessment Information Model July 2014 * Object(platform specific): A generalized construct that is extended in platform schema to describe a collectable aspect of endpoint posture. * State(platform specific): A generalized construct that is extended in platform schema to describe a set of criteria for evaluating posture attributes. o Variables: Constructs that allow for the parameterization of the elements used in the Definitions component based on externally provided values. o System Characteristics: Constructs that represent collected posture from one or more endpoints. This element may be embedded with the Results component, or may be exchanged separately to allow for separate collection and evaluation. The core elements of this component are: * CollectedObject: Provides a mapping of collected Items to Objects defined in the Definitions component. * Item(platform specific): A generalized construct that is extended in platform schema to describe specific posture attributes pertaining to an aspect of endpoint state. o Results: Constructs that represent the result of evaluating expected state (state elements) against actual state (item elements). It includes the true/false evaluation result for each evaluated Definition and Test. Systems characteristics are embedded as well to provide low-level posture details. o Directives: Constructs that enable result reporting detail to be declared, allowing for result production to customized. End-user organizations and vendors create assessment guidance using OVAL by creating XML instances based on the XML schema implementation of the OVAL Definitions model. The OVAL Definitions model defines a structured identifier format for each of the Definition, Test, Object, State, and Item elements. Each instantiation of these elements in OVAL XML instances are assigned a unique identifier based on the specific elements identifier syntax. These XML instances are used by tools that support OVAL to drive collection and evaluation of endpoint posture. When posture collection is performed, an OVAL Systems Characteristics XML instance is generated based on the collected posture attributes. When this collected posture is evaluated, an OVAL Result XML instance is generated that contains the results of the evaluation. In most implementations, the collection and evaluation is performed at the same time. Waltermire & Watson Expires January 4, 2015 [Page 48] Internet-Draft Endpoint Assessment Information Model July 2014 Many of the elements in the OVAL model (i.e., Test, Object, State, Item) are abstract, requiring a platform-specific schema implementation, called a "Component Model" in OVAL. These platform schema implementations are where platform specific posture attributes are defined. For each aspect of platform posture a specialized OVAL Object, which appears in the OVAL Definitions model, provides a format for expressing what posture attribute data to collect from an endpoint through the specification of a datatype, operation, and value(s) on entities that uniquely identify a platform configuration item. For example, a hive, key, and name is used to identify a registry key on a Windows endpoint. Each specialized OVAL Object has a corresponding specialized State, which represents the posture attributes that can be evaluated, and an Item which represents the specific posture attributes that can be collected. Additionally, a specialized Test exists that allows collected Items corresponding to a CollectedObject to be evaluated against one or more specialized States of the same posture type. The OVAL language provides a generalized approach suitable for posture collection and evaluation. While this approach does provide for a degree of extensibility, there are some concerns that should be addressed in order to make OVAL a viable basis for SACM's use. These concerns include: o Platform Schema Creation and Maintenance: In OVAL platform schema, the OVAL data model maintains a tight binding between the Test, Object, State, and Item elements used to assess an aspect of endpoint posture. Creating a new platform schema or adding a new posture aspect to an existing platform schema can be a very labor intensive process. Doing so often involves researching and understanding system APIs and can be prone to issues with inconsistency within and between platforms. To simplify platform schema creation and maintenance, the model needs to be evolved to generalize the Test, Object, and State elements, requiring only the definition of an Item representation. o Given an XML instance based on the Definitions model, it is not clear in the specification how incremental collection and evaluation can occur. Because of this, typically, OVAL assessments are performed on a periodic basis. The OVAL specification needs to be enhanced to include specifications for performing event-based and incremental assessment in addition to full periodic collection. o Defining new functions for manipulating variable values is current handled in the Definitions schema. This requires revision to the core language to add new functions. The OVAL specification needs Waltermire & Watson Expires January 4, 2015 [Page 49] Internet-Draft Endpoint Assessment Information Model July 2014 to be evolved to provide for greater extensibility in this area, allowing extension schema to define new functions. o The current process for releasing a new version of OVAL, bundle releases of the core language with release of community recognized platform schema. The revision processes for the core and platform schema need to be decoupled. Each platform schema should use some mechanism to declare which core language version it relies on. If adopted by SCAM, these issues will need to be addressed as part of the SCAM engineering work to make OVAL more broadly adoptable as a general purpose data model for posture collection and evaluation. 11.2.1.1.2.2. Applicability to Platform Configuration Item Identification Each OVAL Object is identified by a globally unique identifier. This globally unique identifier could be used by the SACM community to identify platform-specific configuration items and at the same time serve as collection guidance. If used in this manner, OVAL Objects would likely need to undergo changes in order to decouple it from evaluation guidance and to provide more robust collection capabilities to support the needs of the SACM community. 11.2.2. Configuration Item Identifier An identifier for a high-level, platform-independent configuration control. This identification concept is necessary to allow similar configuration item concepts to be comparable across platforms. For example, a configuration item might be created for the minimum password length configuration control, which may then have a number of different platform-specific configuration settings. Without this type of identification, it will be difficult to perform evaluation of expected versus actual state in a platform-neutral way. High-level configuration items tend to change much less frequently than the platform-specific configuration items (see Section 11.2.1) that might be associated with them. To provide for the greatest amount of sustainability, collections of configuration item identifiers are best defined by specific communities of interest, while platform-specific identifiers are best defined by the source vendor of the platform. Under this model, the primary source vendors would map their platform-specific configuration controls to the appropriate platform-independent item allowing end-user organizations to make use of these relationships. To support different communities of interest, it may be necessary to support multiple methods for identification of configuration items Waltermire & Watson Expires January 4, 2015 [Page 50] Internet-Draft Endpoint Assessment Information Model July 2014 and for associating related metadata. Use of an IANA registry supporting multiple configuration item identification methods would be an ideal way forward. To the extent possible, a few number of configuration item identification approaches is desirable, to maximize the update by vendors who would be maintain mapping of platform-specific configuration identifiers to the more general platform-neutral configuration identifiers. 11.2.2.1. Related Work 11.2.2.1.1. Control Correlation Identifier The Control Correlation Identifier (CCI) [CCI] is developed and managed by the United States Department of Defense (US-DoD) Defense Information Systems Agency (DISA). According to their website, CCI "provides a standard identifier and description for each of the singular, actionable statements that comprise an information assurance (IA) control or IA best practice. CCI bridges the gap between high-level policy expressions and low-level technical implementations. CCI allows a security requirement that is expressed in a high-level policy framework to be decomposed and explicitly associated with the low-level security setting(s) that must be assessed to determine compliance with the objectives of that specific security control. This ability to trace security requirements from their origin (e.g., regulations, IA frameworks) to their low-level implementation allows organizations to readily demonstrate compliance to multiple IA compliance frameworks. CCI also provides a means to objectively roll-up and compare related compliance assessment results across disparate technologies." It is recommended that this approach be analysed as a potential candidate for use as a configuration item identifier method. Note: This reference to CCI is for informational purposes. Since the editors do not represent DISA's interests, its inclusion in this document does not indicate the presence or lack of desire to contribute aspects of this effort to SACM. 11.2.2.1.2. A Potential Alternate Approach There will likely be a desire by different communities to create different collections of configuration item identifiers. This fracturing may be caused by: o Different requirements for levels of abstraction, o Varying needs for timely maintenance of the collection, and Waltermire & Watson Expires January 4, 2015 [Page 51] Internet-Draft Endpoint Assessment Information Model July 2014 o Differing scopes of technological needs. Due to these and other potential needs, it will be difficult to standardize around a single collection of configuration identifiers. A workable solution will be one that is scalable and usable for a broad population of end-user organizations. An alternate approach that should be considered is the definition of data model that contains a common set of metadata attributes, perhaps supported by an extensible taxonomy, that can be assigned to platform-specific configuration items. If defined at a necessary level of granularity, it may be possible to query collections of platform-specific configuration items provided by vendors to create groupings at various levels of abstractions. By utilizing data provided by vendors, technological needs and the timeliness of information can be addressed based on customer requirements. SACM should consider this and other approaches to satisfy the need for configuration item roll-up in a way that provides the broadest benefit, while achieving a sensible degree of scalability and sustainability. 11.2.3. Vulnerability Identifier An unique name for a known software flaw that exists in specific versions of one or more units of software. One use of a vulnerability identifier in the SACM context is to associate a given flaw with the vulnerable software using software identifiers. For this reason at minimum, software identifiers should identify a software product to the patch or version level, and not just to the level that the product is licensed. 11.2.3.1. Related Work 11.2.3.1.1. Common Vulnerabilities and Exposures Common Vulnerabilities and Exposures (CVE) [CVE-WEBSITE] is a MITRE led effort to assign common identifiers to publicly known security vulnerabilities in software to facilitate the sharing of information related to the vulnerabilities. CVE is the industry standard by which software vendors, tools, and security professionals identify vulnerabilities and could be used to address SACM's need for a vulnerability identifier. 11.3. Endpoint characterization Target when policies (collection, evaluated, guidance) apply Collection can be used to further characterize Waltermire & Watson Expires January 4, 2015 [Page 52] Internet-Draft Endpoint Assessment Information Model July 2014 Also human input Information required to characterize an endpoint is used to determine what endpoints are the target of a posture assessment. It is also used to determine the collection, evaluation, and/or reporting policies and the associated guidance that apply to the assessment. Endpoint characterization information may be populated by: o A manual input process and entered into records associated with the endpoint, or o Using information collected and evaluated by an assessment. Regardless of the method of collection, it will be necessary to query and exchange endpoint characterization information as part of the assessment planning workflow. 11.3.1. Related Work 11.3.1.1. Extensible Configuration Checklist Description Format 11.3.1.1.1. Background The Extensible Configuration Checklist Description Format (XCCDF) is a specification that provides an XML-based format for expressing security checklists. The XCCDF 1.2 specification is published by International Organization for Standardization (ISO) [ISO.18180]. XCCDF contains multiple components and capabilities, and various components align with different elements of this information model. This specification was originally published by NIST [NISTIR-7275]. When contributed to ISO Joint Technical Committee 1 (JTC 1), a comment was introduced indicating an interest in the IETF becoming the maintenance organization for this standard. If the SACM working group is interested in taking on engineering work pertaining to XCCDF, a contribution through a national body can be made to create a ballot resolution for transition of this standard to the IETF for maintenance. 11.3.1.1.2. Applicability to Endpoint characterization The target component of XCCDF provides a mechanism for capturing characteristics about an endpoint including the fully qualified domain name, network address, references to external identification information (e.g. Asset Identification), and is extensible to support other useful information (e.g. MAC address, globally unique identifier, certificate, etc.). XCCDF may serve as a good starting Waltermire & Watson Expires January 4, 2015 [Page 53] Internet-Draft Endpoint Assessment Information Model July 2014 point for understanding the types of information that should be used to identify an endpoint. 11.3.1.2. Asset Reporting Format 11.3.1.2.1. Background The Asset Reporting Format (ARF) [NISTIR-7694] is a data model to express information about assets, and the relationships between assets and reports. It facilitates the reporting, correlating, and fusing of asset information within and between organizations. ARF is vendor and technology neutral, flexible, and suited for a wide variety of reporting applications. There are four major sub-components of ARF: o Asset: The asset component element includes asset identification information for one or more assets. It simply houses assets independent of their relationships to reports. The relationship section can then link the report section to specific assets. o Report: The report component element contains one or more asset reports. An asset report is composed of content (or a link to content) about one or more assets. o Report-Request: The report-request component element contains the asset report requests, which can give context to asset reports captured in the report section. The report-request section simply houses asset report requests independent of the report which was subsequently generated. o Relationship: The relationship component element links assets, reports, and report requests together with well-defined relationships. Each relationship is defined as {subject} {predicate} {object}, where {subject} is the asset, report request, or report of interest, {predicate} is the relationship type being established, and {object} is one or more assets, report requests, or reports. 11.3.1.2.2. Relationship to Endpoint Characterization For Endpoint Characterization, ARF can be used in multiple ways due to its flexibility. ARF supports the use of the Asset Identification specification (more in Section 11.3.1.2.3) to embed the representation of one or more assets as well as relationships between those assets. It also allows the inclusion of report-requests, which can provide details on what data was required for an assessment. Waltermire & Watson Expires January 4, 2015 [Page 54] Internet-Draft Endpoint Assessment Information Model July 2014 ARF is agnostic to the data formats of the collected posture attributes and therefore can be used within the SACM Architecture to provide Endpoint Characterization without dictating data formats for the encoding of posture attributes. The embedded Asset Identification data model (see Section 11.1.1.1) can be used to characterize one or more endpoints to allow targeting for collection, evaluation, etc. Additionally, the report-request model can dictate the type of reporting that has been requested, thereby providing context as to which endpoints the guidance applies. 11.3.1.2.3. Asset Identification Described earlier In the context of Endpoint Characterization, the Asset Identification data model could be used to encode information that identifies specific endpoints and/or classes of endpoints to which a particular assessment is relevant. The flexibility in the Asset Identification specification allows usage of various endpoint identifiers as defined by the SACM engineering work. As stated in Section 11.3.1.2.3, the Asset Identification specification is included within the Asset Reporting Framework (ARF) and therefore can be used in concert with that specification as well. 11.3.1.3. The CPE Applicability Language CPE described earlier Applicability in CPE is defined as an XML language [NISTIR-7698] for using CPE names to create applicability statements using logical expressions. These expressions can be used to applicability statements that can drive decisions about assets, whether or not to do things like collect data, report data, and execute policy compliance checks. It is recommended that SACM evolve the CPE Applicability Language through engineering work to allow it to better fit into the security automation vision laid out by the Use Cases and Architecture for SACM. This should include de-coupling the identification part of the language from the logical expressions, making it such that the language is agnostic to the method by which assets are identified. This will allow use of SWID, CPE Names, or other identifiers to be used, perhaps supported by an IANA registry of identifier types. The other key aspect that should be evolved is the ability to make use of the Applicability Language against a centralized repository of collected posture attributes. The language should be able to make Waltermire & Watson Expires January 4, 2015 [Page 55] Internet-Draft Endpoint Assessment Information Model July 2014 applicability statements against previously collected posture attributes, such that an enterprise can quickly query the correct set of applicable endpoints in an automated and scalable manner. 11.4. Posture Attribute Expression Discuss the catalog concept. Listing of things that can be chosen from. Things we can know about. Vendors define catalogs. Ways for users to get vendor-provided catalogs. To support the collection of posture attributes, there needs to be a way for operators to identify and select from a set of platform- specific attribute(s) to collect. The same identified attributes will also need to be identified post-collection to associate the actual value of that attribute pertaining to an endpoint as it was configured at the time of the collection. To provide for extensibility, the need exists to support a variety of possible identification approaches. It is also necessary to enable vendors of software to provide a listing, or catalog, of the available posture attributes to operators that can be collected. Ideally, a federated approach will be used to allow organizations to identify the location for a repository containing catalogs of posture attributes provided by authoritative primary source vendors. By querying these repositories, operators will be able to acquire the appropriate listings of available posture attributes for their deployed assets. One or more posture attribute expressions are needed to support these exchanges. 11.4.1. Related Work The ATOM Syndication Format [RFC4287] provides an extensible, flexible XML-based expression for organizing a collection of data feeds consisting of entries. This standard can be used to express one or more catalogs of posture attributes represented as data feeds. Groupings of posture attributes would be represented as entries. These entries could be defined using the data models described in the "Related Work" sections below. Additionally, this approach can also be used more generally for guidance repositories allowing other forms of security automation guidance to be exchanged using the same format. 11.4.2. Platform Configuration Attributes A low-level, platform-dependent posture attribute as determined by the authoritative primary source vendor. Collection guidance will be derived from catalogs of platform specific posture attributes. Waltermire & Watson Expires January 4, 2015 [Page 56] Internet-Draft Endpoint Assessment Information Model July 2014 For example, a primary source vendor would create a platform-specific posture attribute that best models the posture attribute data for their platform. 11.4.2.1. Related Work 11.4.2.1.1. Open Vulnerability and Assessment Language A general overview of OVAL was provided previously in Section 11.2.1.1.2.1. The OVAL System Characteristics platform extension models provide a catalog of the posture attributes that can be collected from an endpoint. In OVAL these posture attributes are further grouped into logical constructs called OVAL Items. For example, the passwordpolicy_item that is part of the Windows platform extension groups together all the posture attributes related to passwords for an endpoint running Windows (e.g. maximum password age, minimum password length, password complexity, etc.). The various OVAL Items defined in the OVAL System Characteristics may serve as a good starting for the types of posture attribute data that needs to be collected from endpoints. OVAL platform extension models may be shared using the ATOM Syndication Format. 11.4.2.1.2. Network Configuration Protocol and YANG Data Modeling Language The Network Configuration Protocol (NETCONF) [RFC6241] defines a mechanism for managing and retrieving posture attribute data from network infrastructure endpoints. The posture attribute data that can be collected from a network infrastructure endpoint is highly extensible and can defined using the YANG modeling language [RFC6020]. Models exist for common datatypes, interfaces, and routing subsystem information among other subjects. The YANG modeling language may be useful in providing an extensible framework for the SACM community to define one or more catalogs of posture attribute data that can be collected from network infrastructure endpoints. Custom YANG modules may also be shared using the ATOM Syndication Format. 11.4.2.1.3. Simple Network Management Protocol and Management Information Base Entry The Simple Network Protocol (SNMP) [RFC3411] defines a protocol for managing and retrieving posture attribute data from endpoints on a network . The posture attribute data that can be collected of an Waltermire & Watson Expires January 4, 2015 [Page 57] Internet-Draft Endpoint Assessment Information Model July 2014 endpoint and retrieved by SNMP is defined by the Management Information Base (MIB) [RFC3418] which is hierarchical collection of information that is referenced using Object Identifiers . Given this, MIBs may provide an extensible way for the SACM community to define a catalog of posture attribute data that can be collected off of endpoints using SNMP. MIBs may be shared using the ATOM Syndication Format. 11.5. Actual Value Representation Discuss instance concept. The actual value of a posture attribute is collected or published from an endpoint. The identifiers discussed previously provide names for the posture attributes (i.e., software or configuration item) that can be the subject of an assessment. The information items listed below are the actual values collected during the assessment and are all associated with a specific endpoint. 11.5.1. Software Inventory A software inventory is a list of software identifiers (or content) associated with a specific endpoint. Software inventories are maintained in some organized fashion so that entities can interact with it. Just having software publish identifiers onto an endpoint is not enough, there needs to be an organized listing of all those identifiers associated with that endpoint. 11.5.1.1. Related Work 11.5.1.1.1. Asset Summary Reporting The Asset Summary Reporting (ASR) specification [NISTIR-7848] provides a format for capturing summary information about one or more assets. Specifically, it provides the ability to express a collection of records from some defined data source and map them to some set of assets. As a result, this specification may be useful for capturing the software installed on an endpoint, its relevant attributes, and associating it with a particular endpoint. 11.5.1.1.2. Software Identification Tags SWID tag were previously introduced in Section 11.1.3.1.2. As stated before, SWID tags are ideally deployed to an endpoint during software installation. In the less ideal case, they may also be generated based on information retrieved from a proprietary software installation data store. At minimum, SWID tag must contain an Waltermire & Watson Expires January 4, 2015 [Page 58] Internet-Draft Endpoint Assessment Information Model July 2014 identifier for each unit of installed software. Given this, SWID tags may be a viable way for SACM to express detailed information about the software installed on an endpoint. 11.5.2. Collected Platform Configuration Posture Attributes Configurations associated with a software instance associated with an endpoint A list of the configuration posture attributes associated with the actual values collected from the endpoint during the assessment as required/expressed by any related guidance. Additionally, each configuration posture attribute is associated with the installed software instance it pertains to. 11.5.2.1. Related Work 11.5.2.1.1. Open Vulnerability and Assessment Language A general overview of OVAL was provided previously in Section 11.2.1.1.2.1. As mentioned earlier, the OVAL System Characteristics platform extensions provide a catalog of the posture attributes that can be collected and assessed in the form of OVAL Items. These OVAL Items also serve as a model for representing posture attribute data and associated values that are collected off an endpoint. Furthermore, the OVAL System Characteristics model provides a system_info construct that captures information that identifies and characterizes the endpoint from which the posture attribute data was collected. Specifically, it includes operating system name, operating system version, endpoint architecture, hostname, network interfaces, and an extensible construct to support arbitrary additional information that may be useful in identifying the endpoint in an enterprise such as information capture in Asset Identification constructs. The OVAL System Characteristics model could serve as a useful starting point for representing posture attribute data collected from an endpoint although it may need to undergo some changes to satisfy the needs of the SACM community. 11.5.2.1.2. NETCONF-Based Collection Introduced earlier in Section 11.4.2.1.2, NETCONF defines a protocol for managing and retrieving posture attribute data from network infrastructure endpoints. NETCONF provides the and operations to retrieve the configuration data, and configuration data and state data respectively from a network infrastructure endpoint. Upon successful completion of these operations, the current posture attribute data of the network infrastructure endpoint will be made available. NETCONF also Waltermire & Watson Expires January 4, 2015 [Page 59] Internet-Draft Endpoint Assessment Information Model July 2014 provides a variety of filtering mechanisms (XPath, subtree, content matching, etc.) to trim down the posture attribute data that is collected from the endpoint. Given that NETCONF is widely adopted by network infrastructure vendors, it may useful to consider this protocol as a standardized mechanism for collecting posture attribute data from network infrastructure endpoints. As a side note, members of the OVAL Community have also developed a proposal to extend the OVAL Language to support the assessment of NETCONF configuration data [1]. The proposal leverages XPath to extract the posture attribute data of interest from the XML data returned by NETCONF. The collected posture attribute data can then be evaluated using OVAL Definitions and the results of the evaluation can be expressed as OVAL Results. While this proposal is not currently part of the OVAL Language, it may be worth considering. 11.5.2.1.3. SNMP-Based Collection The SNMP, previously introduced in Section 11.4.2.1.3, defines a protocol for managing and retrieving posture attribute data from endpoints on a network [RFC3411]. SNMP provides three protocol operations [RFC3416] (GetRequest, GetNextRequest, and GetBulkRequest) for retrieving posture attribute data defined by MIB objects. Upon successful completion of these operations, the requested posture attribute data of the endpoint will be made available to the requesting application. Given that SNMP is widely adopted by vendors, and the MIBs that define posture attribute data on an endpoint are highly extensible, it may useful to consider this protocol as a standardized mechanism for collecting posture attribute data from endpoints in an enterprise. 11.6. Evaluation Guidance 11.6.1. Configuration Evaluation Guidance The evaluation guidance is applied by evaluators during posture assessment of an endpoint. This guidance must be able to reference or be associated with the following previously defined information elements: o configuration item identifiers, o platform configuration identifiers, and o collected Platform Configuration Posture Attributes. Waltermire & Watson Expires January 4, 2015 [Page 60] Internet-Draft Endpoint Assessment Information Model July 2014 11.6.1.1. Related Work 11.6.1.1.1. Open Vulnerability and Assessment Language A general overview of OVAL was provided previously in Section 11.2.1.1.2.1. The OVAL Definitions model provides an extensible framework for making assertions about the state of posture attribute data collected from an endpoint. Guidance written against this model consists of one or more OVAL Tests, that can be logically combined, where each OVAL Test defines what posture attributes should be collected from an endpoint (as OVAL Objects) and optionally defines the expected state of the posture attributes (as OVAL States). While the OVAL Definitions model may be a useful starting point for evaluation guidance, it will likely require some changes to decouple collection and evaluation concepts to satisfy the needs of the SACM community. 11.6.1.1.2. XCCDF Rule A general description of XCCDF was provided in Section 11.3.1.1.1. As noted there, an XCCDF document represents a checklist of items against which a given endpoint's state is compared and evaluated. An XCCDF Rule represents one assessed item in this checklist. A Rule contains both a prose description of the assessed item (either for presentation to the user in a tool's user interface, or for rendering into a prose checklist for human consumption) and can also contain instructions to support automated evaluation of the assessed item, if such automated assessment is possible. Automated assessment instructions can be provided either within the XCCDF Rule itself, or by providing a reference to instructions expressed in other languages, such as OVAL. In order to support greater flexibility in XCCDF, checklists can be tailored to meet certain needs. One way to do this is to enable or disable certain rules that are appropriate or inappropriate to a given endpoint, respectively. For example, a single XCCDF checklist might contain check items to evaluate the configuration of an endpoint's operating system. An endpoint deployed in an enterprise's DMZ might need to be locked down more than a common internal endpoint, due to the greater exposure to attack. In this case, some operating system configuration requirements for the DMZ endpoint might be unnecessary for the internal endpoint. Nonetheless, most configuration requirements would probably remain applicable to both environments (providing a common baseline for configuration of the given operating system) and thus be common to the checking instructions for both types of endpoints. XCCDF supports this by allowing a single checklist to be defined, but then tailored to the needs of the assessed endpoint. In the previous example, some Rules Waltermire & Watson Expires January 4, 2015 [Page 61] Internet-Draft Endpoint Assessment Information Model July 2014 that apply only to the DMZ endpoint would be disabled during the assessment of an internal endpoint and would not be exercised during the assessment or count towards the assessment results. To accomplish this, XCCDF uses the CPE Applicability Language. By enhancing this applicability language to support other aspects of endpoint characterization (see Section 11.3.1.3), XCCDF will also benefit from these enhancements. In addition, XCCDF Rules also support parameterization, allowing customization of the expected value for a given check item. For example, the DMZ endpoint might require a password of at least 12 characters, while an internal endpoint may only need 8 or more characters in its password. By employing parameterization of the XCCDF Rule, the same Rule can be used when assessing either type of endpoint, and simply be provided with a different target parameter each time to reflect the different role-based requirements. Sets of customizations can be stored within the XCCDF document itself: XCCDF Values store parameters values that can be used in tailoring, while XCCDF Profiles store sets of tailoring instructions, including selection of certain Values as parameters and the enabling and disabling of certain Rules. The tailoring capabilities supported by XCCDF allow a single XCCDF document to encapsulate configuration evaluation guidance that applies to a broad range of endpoint roles. 11.7. Evaluation Result Reporting 11.7.1. Configuration Evaluation Results The evaluation guidance applied during posture assessment of an endpoint to customize the behavior of evaluators. Guidance can be used to define specific result output formats or to select the level- of-detail for the generated results. This guidance must be able to reference or be associated with the following previously defined information elements: o configuration item identifiers, o platform configuration identifiers, and o collected Platform Configuration Posture Attributes. 11.7.1.1. Related Work 11.7.1.1.1. XCCDF TestResults A general description of the eXtensible Configuration Checklist Description Format (XCCDF) was provided in section Section 11.3.1.1.1. The XCCDF TestResult structure captures the Waltermire & Watson Expires January 4, 2015 [Page 62] Internet-Draft Endpoint Assessment Information Model July 2014 outcome of assessing a single endpoint against the assessed items (i.e., XCCDF Rules) contained in an XCCDF instance document. XCCDF TestResults capture a number of important pieces of information about the assessment including: o The identity of the assessed endpoint. See Section 11.3.1.1.2 for more about XCCDF structures used for endpoint identification. o Any tailoring of the checklist that might have been employed. See Section 11.6.1.1.2 for more on how XCCDF supports tailoring. o The individual results of the assessment of each enabled XCCDF Rule in the checklist. See Section 11.6.1.1.2 for more on XCCDF Rules. The individual results for a given XCCDF Rule capture only whether the rule "passed", "failed", or experienced some exceptional condition, such as if an error was encountered during assessment. XCCDF 1.2 Rule results do not capture the actual state of the endpoint. For example, an XCCDF Rule result might indicate that an endpoint failed to pass requirement that passwords be of a length greater than or equal to 8, but it would not capture that the endpoint was, in fact, only requiring passwords of 4 or more characters. It may, however, be possible for a user to discover this information via other means. For example, if the XCCDF Rule uses an OVAL Definition to effect the Rule's evaluation, then the actual endpoint state may be captured in the corresponding OVAL System Characteristics file. The XCCDF TestResult structure does provide a useful structure for understanding the overall assessment that was conducted and the results thereof. The ability to quickly determine the Rules that are not complied with on a given endpoint allow administrators to quickly identify where remediation needs to occur. 11.7.1.1.2. Open Vulnerability and Assessment Language A general overview of OVAL was provided previously in Section 11.2.1.1.2.1. OVAL Results provides a model for expressing the results of the assessment of the actual state of the posture attribute values collected of an endpoint (represented as an OVAL System Characteristics document) against the expected posture attribute values (defined in an OVAL Definitions document. Using OVAL Directives, the granularity of OVAL Results can also be specified. The OVAL Results model may be useful in providing a format for capturing the results of an assessment. Waltermire & Watson Expires January 4, 2015 [Page 63] Internet-Draft Endpoint Assessment Information Model July 2014 11.7.1.1.3. Asset Summary Reporting A general overview of ASR was provided previously in Section 11.5.1.1.1. As ASR provides a way to report summary information about assets, it can be used within the SACM Architecture to provide a way to aggregate asset information for later use. It makes no assertions about the data formats used by the assessment, but rather provides an XML, record-based way to collect aggregated information about assets. By using ASR to collect this summary information within the SACM Architecture, one can provide a way to encode the details used by various reporting requirements, including user-definable reports. 11.7.1.1.4. ARF A general overview of ARF was provided previously in Section 11.3.1.2.1. Because ARF is data model agnostic, it can provide a flexible format for exchanging collection and evaluation information from endpoints. It additionally provides a way to encode relationships between guidance and assets, and as such, can be used to associate assessment results with guidance. This could be the guidance that directly triggered the assessment, or for guidance that is run against collected posture attributes located in a central repository. 11.7.2. Software Inventory Evaluation Results The results of an evaluation of an endpoint's software inventory against an authorized software list. The authorized software list represents the policy for what software units are allowed, prohibited, and mandatory for an endpoint. 12. Acknowledgements Many of the specifications in this document have been developed in a public-private partnership with vendors and end-users. The hard work of the SCAP community is appreciated in advancing these efforts to their current level of adoption. Over the course of developing the initial draft, Brant Cheikes, Matt Hansbury, Daniel Haynes, and Charles Schmidt have contributed text to many sections of this document. Waltermire & Watson Expires January 4, 2015 [Page 64] Internet-Draft Endpoint Assessment Information Model July 2014 13. IANA Considerations This memo includes no request to IANA. 14. Security Considerations Posture Assessments need to be performed in a safe and secure manner. In that regard, there are multiple aspects of security that apply to the communications between components as well as the capabilities themselves. Due to time constraints, this information model only contains an initial listing of items that need to be considered with respect to security. This list is not exhaustive, and will need to be augmented as the model continues to be developed/refined. Initial list of security considerations include: Authentication: Every component and asset needs to be able to identify itself and verify the identity of other components and assets. Confidentiality: Communications between components need to be protected from eavesdropping or unauthorized collection. Some communications between components and assets may need to be protected as well. Integrity: The information exchanged between components needs to be protected from modification. some exchanges between assets and components will also have this requirement. Restricted Access: Access to the information collected, evaluated, reported, and stored should only be viewable/consumable to authenticated and authorized entities. 15. References 15.1. Normative References [RFC2119] Bradner, S., "Key words for use in RFCs to Indicate Requirement Levels", BCP 14, RFC 2119, March 1997. 15.2. Informative References [CCE] The National Institute of Standards and Technology, "Common Configuration Enumeration", 2014, . Waltermire & Watson Expires January 4, 2015 [Page 65] Internet-Draft Endpoint Assessment Information Model July 2014 [CCI] United States Department of Defense Defense Information Systems Agency, "Control Correlation Identifier", 2014, . [CPE-WEBSITE] The National Institute of Standards and Technology, "Common Platform Enumeration", 2014, . [CVE-WEBSITE] The MITRE Corporation, "Common Vulnerabilities and Exposures", 2014, . [I-D.camwinget-sacm-requirements] Cam-Winget, N., "Secure Automation and Continuous Monitoring (SACM) Requirements", draft-camwinget-sacm- requirements-04 (work in progress), June 2014. [I-D.ietf-sacm-terminology] Waltermire, D., Montville, A., Harrington, D., and N. Cam- Winget, "Terminology for Security Assessment", draft-ietf- sacm-terminology-04 (work in progress), May 2014. [I-D.ietf-sacm-use-cases] Waltermire, D. and D. Harrington, "Endpoint Security Posture Assessment - Enterprise Use Cases", draft-ietf- sacm-use-cases-07 (work in progress), April 2014. [IM-LIAISON-STATEMENT-NIST] Montville, A., "Liaison Statement: Call for Contributions for the SACM Information Model to NIST", May 2014, . [ISO.18180] "Information technology -- Specification for the Extensible Configuration Checklist Description Format (XCCDF) Version 1.2", ISO/IEC 18180, 2013, . [ISO.19770-2] "Information technology -- Software asset management -- Part 2: Software identification tag", ISO/IEC 19770-2, 2009. Waltermire & Watson Expires January 4, 2015 [Page 66] Internet-Draft Endpoint Assessment Information Model July 2014 [NISTIR-7275] Waltermire, D., Schmidt, C., Scarfone, K., and N. Ziring, "Specification for the Extensible Configuration Checklist Description Format (XCCDF) Version 1.2", NISTIR 7275r4, March 2013, . [NISTIR-7693] Wunder, J., Halbardier, A., and D. Waltermire, "Specification for Asset Identification 1.1", NISTIR 7693, June 2011, . [NISTIR-7694] Halbardier, A., Waltermire, D., and M. Johnson, "Specification for the Asset Reporting Format 1.1", NISTIR 7694, June 2011, . [NISTIR-7695] Cheikes, B., Waltermire, D., and K. Scarfone, "Common Platform Enumeration: Naming Specification Version 2.3", NISTIR 7695, August 2011, . [NISTIR-7696] Parmelee, M., Booth, H., Waltermire, D., and K. Scarfone, "Common Platform Enumeration: Name Matching Specification Version 2.3", NISTIR 7696, August 2011, . [NISTIR-7697] Cichonski, P., Waltermire, D., and K. Scarfone, "Common Platform Enumeration: Dictionary Specification Version 2.3", NISTIR 7697, August 2011, . [NISTIR-7698] Waltermire, D., Cichonski, P., and K. Scarfone, "Common Platform Enumeration: Applicability Language Specification Version 2.3", NISTIR 7698, August 2011, . Waltermire & Watson Expires January 4, 2015 [Page 67] Internet-Draft Endpoint Assessment Information Model July 2014 [NISTIR-7848] Davidson, M., Halbardier, A., and D. Waltermire, "Specification for the Asset Summary Reporting Format 1.0", NISTIR 7848, May 2012, . [OVAL-LANGUAGE] Baker, J., Hansbury, M., and D. Haynes, "The OVAL Language Specification version 5.10.1", January 2012, . [RFC3411] Harrington, D., Presuhn, R., and B. Wijnen, "An Architecture for Describing Simple Network Management Protocol (SNMP) Management Frameworks", STD 62, RFC 3411, December 2002. [RFC3416] Presuhn, R., "Version 2 of the Protocol Operations for the Simple Network Management Protocol (SNMP)", STD 62, RFC 3416, December 2002. [RFC3418] Presuhn, R., "Management Information Base (MIB) for the Simple Network Management Protocol (SNMP)", STD 62, RFC 3418, December 2002. [RFC3444] Pras, A. and J. Schoenwaelder, "On the Difference between Information Models and Data Models", RFC 3444, January 2003. [RFC4287] Nottingham, M., Ed. and R. Sayre, Ed., "The Atom Syndication Format", RFC 4287, December 2005. [RFC4949] Shirey, R., "Internet Security Glossary, Version 2", RFC 4949, August 2007. [RFC6020] Bjorklund, M., "YANG - A Data Modeling Language for the Network Configuration Protocol (NETCONF)", RFC 6020, October 2010. [RFC6241] Enns, R., Bjorklund, M., Schoenwaelder, J., and A. Bierman, "Network Configuration Protocol (NETCONF)", RFC 6241, June 2011. Waltermire & Watson Expires January 4, 2015 [Page 68] Internet-Draft Endpoint Assessment Information Model July 2014 [SP800-117] Quinn, S., Scarfone, K., and D. Waltermire, "Guide to Adopting and Using the Security Content Automation Protocol (SCAP) Version 1.2", SP 800-117, January 2012, . [SP800-126] Waltermire, D., Quinn, S., Scarfone, K., and A. Halbardier, "The Technical Specification for the Security Content Automation Protocol (SCAP): SCAP Version 1.2", SP 800-126, September 2011, . 15.3. URIs [1] https://github.com/OVALProject/Sandbox/blob/master/x-netconf- definitions-schema.xsd Authors' Addresses David Waltermire (editor) National Institute of Standards and Technology 100 Bureau Drive Gaithersburg, Maryland 20877 USA Email: david.waltermire@nist.gov Kim Watson United States Department of Homeland Security DHS/CS&C/FNR 245 Murray Ln. SW, Bldg 410 MS0613 Washington, DC 20528 USA Email: kimberly.watson@hq.dhs.gov Waltermire & Watson Expires January 4, 2015 [Page 69]