SoS and T&E NDIA SoS SE Committee June 9, 2009 Background NDIA SoS SE Committee identified SoS and T&E as a topic.
Download ReportTranscript SoS and T&E NDIA SoS SE Committee June 9, 2009 Background NDIA SoS SE Committee identified SoS and T&E as a topic.
SoS and T&E NDIA SoS SE Committee June 9, 2009 Background NDIA SoS SE Committee identified SoS and T&E as a topic of interest White paper on SoS and T&E was circulated for comment prior to this meeting • Paper provides a view which is generally understood, including • • • • • SoS are more complicated than systems SoS have constituent components that have independent needs and can exercise free will in their development and operation that places their needs at times before that of the SoS SoS are generally asynchronous and this introduces difficulties in end-to-end testing, and alternates must be sought Emergent properties will emerge and must be accommodated Set of ‘needs’ • • Need for incremental development; need to address synchronization; the need to consider block testing; the need to monitor fielded behavior; the need to express limitations in a politically correct manner; and the means to address and incorporate methods to cope with these Paper does not address balance between SoS development and testing White paper, including comments, is presented here as basis for committee discussion System of Systems (SoS) Translating Translating capability Translating capability objectives capability objectives objectives Assessing Assessing (actual) Assessing (actual) performance performance performance totocapability capability to capability objectives objectives objectives Orchestrating Orchestrating Orchestrating upgrades upgrades upgrades to toSoS SoS Understanding Understanding systems Understanding systems&& relationships relationships systems & (includes (includesplans) plans) relationships to SoS Addressing Addressing new Addressing new requirements requirements requirements &&options solution options Developing, Developing, Developing evolving and evolving and maintaining & evolving SoSmaintaining design/arch SoS SoS design/arch architecture options Monitoring Monitoring &Monitoring assessing assessing &&changes assessing changes changes DoD SoS SE Guide • Focus on technical aspects of SE applicable to SoS • Characterize SoS in DoD Today • Describe Core Elements of SoS SE • Translate application of basic SE processes for SoS SE External Environment SoS Types and examples • Directed - DoD Information System Network (DISN), National System for Geospatial Analysis • Acknowledged – Ballistic Missile Defense System, Air Operations Center • Collaborative – Communities of interest • Virtual - Internet System of Systems: (ref: Defense Acquisition Guide) A set or arrangement of systems that results when independent and useful systems are integrated into a larger system that delivers unique capabilities. SoS T&E Issues Raised in the White Paper 1) If SoS are not programs of record (and not subject to T&E regulations) why should we worry about this at all? 2) If ‘requirements’ are not clearly specified up front for a SoS, what is the basis for T&E of an SoS? 3) What is the relationship between SoS metrics and T&E objectives? 4) Are expected cumulative impacts of systems changes on SoS performance the same as SoS performance objectives? 5) How do you test the contribution of a system to the end to end SoS performance in the absence of other SoS elements critical to the SoS results? What if systems all implemented to their specification, but the overall SoS expected changes cannot be verified? SoS T&E Issue (1) Relationship between SoS and acquisition • Many SoS are not acquisition programs, but rather are ‘umbrella’ programs which incorporate multiple systems at different stages of development each with their own T&E requirements • The constituent systems are operationally and managerially independent from the SoS and are on asynchronous development schedules • Scope and complexity of SoS If SoS are not programs of record (and not subject to T&E regulations) why should we worry about this at all? Issue Discussion (1) If SoS are not programs of record (and not subject to T&E regulations) why should we worry about this at all? • The underlying driver for T&E regulations is the objective of assuring the user that the capability they need is provided by the systems. • This driver exists whether or not a systems (or SoS) is a program of record • Furthermore, all changes made to the constituent systems should be verified to confirm they have been implemented correctly, and end to end T&E supports the need to show that SoS changes have not inadvertently diminished other necessary capability • T&E provides a mechanism to understand the impact of changes on desired results, so an informed fielding decision can be made • The following recommendations on SOS T&E approaches are made based on this assumed importance of T&E SoS T&E Issue (2) Translating Translating capability Translating capability objectives capability objectives objectives Translating capability objectives into high level SoS requirements • In this element, the capability context for the SoS is established, which grounds assessment of the current SoS performance. • In many cases, SoS don’t have ‘requirements’ per se, but capability objectives or goals that provide the starting point for specific requirements for increments of SoS evolution If ‘requirements’ are not clearly specified up front for a SoS, what is the basis for T&E of an SoS? Issue Discussion (2) If ‘requirements’ are not clearly specified up front for a SoS, what is the basis for T&E of an SoS? • SoS typically have broad capability objectives versus specific performance requirements as defined for other systems; these provide a foundation for • identifying systems supporting an SoS • development of an SoS architecture • recommended changes or additions to the systems in the SoS to address the capabilities • This suggests that it is necessary to generate metrics defining the end-toend SoS capabilities that provide an ongoing ‘benchmark’ for SoS development. • In some SoS circumstances, the capability objectives can be effectively modeled in simulation environments which can be used to identify appropriate changes at the system levels. • The fidelity of the simulation provides for validation of the system changes needed to enable SoS-level capability. • In those cases in which the system changes are driven by SoS-level simulation, the fidelity of the simulation can provide for the SoS evaluation. • In cases where simulation is not practical, other analytical approaches must be used for T&E. • Test conditions that validate the analysis must be carefully chosen to balance test preparation and logistics constraints against the need to demonstrate the objective capability under realistic operational conditions SoS T&E Issue (3) Assessing Assessing (actual) Assessing (actual) performance performance performance totocapability to capability capability objectives objectives objectives Assessing Extent to Which SoS Performance Meets Capability Objectives over Time • This element provides the capability measures for the SoS which, as described in the guide, may be collected from a variety of settings as input on performance under particular condition and new issues facing the SoS from an operational perspective. • Hence, assessing SoS performance is an ongoing activity, which goes beyond assessment of specific changes in elements of the SoS (e.g. changes in constituent systems to meet SoS needs, and system changes driven by factors unrelated to the SoS). What is the relationship between SoS metrics and T&E objectives? Issue Discussion (3) What is the relationship between SoS metrics and T&E objectives? • Typically T&E objectives, particularly key performance parameters, are used as the basis for making a fielding decision • SoS metrics on the other hand (as discussed above) provide an ongoing ‘benchmark’ for SoS development which when assessed over time show an improvement in meeting user capability objectives • Because SoS are typically comprised of a mix of fielded systems new developments • There may not be a discrete ‘SoS’ fielding decision • Instead the various systems are deployed as they are made ready, at some point reaching the threshold that enables the new SoS capability SoS T&E Issue (4) Addressing Addressing new Addressing new requirements requirements requirements & & solution &options options options Addressing requirements and solution options • Increments of SoS improvement are planned by the SoS and systems managers and systems engineers • For each increment there may be specific expectations for changes in systems and an anticipated overall effect on the SoS performance • While it may be possible to define specifications for the system changes, it is more difficult to do this for the SoS, which is in effect the cumulative effect of the changes in the systems Are expected cumulative impacts of systems changes on SoS performance the same as SoS performance objectives? Issue Discussion (4) Are expected cumulative impacts of systems changes on SoS performance the same as SoS performance objectives? • SoS increments are based on changes in constituent systems which cumulate in improvements in the SoS overall • In most cases, changes in systems can be specified and tested. • However, in SoS which are implemented in a variety of environments and are dependent on networks for end to end performance, • Impact of the system changes on SoS end-to-end capabilities can be estimated with less certainty. • This uncertainty must be considered when assessing the SoS against its performance objectives SoS T&E Issues (5) Orchestrating Orchestrating Orchestrating upgrades upgrades upgrades to toSoS SoS to SoS Orchestrating Upgrades to SoS • Systems may make changes as part of their development increment that will be ready to field once they have been successfully tested and evaluated. • However, given the asynchronous nature of system development paths, other systems in the SoS increment may not be ready to test with the early delivering systems, thwarting the concept of end to end test. What if systems all implemented to their specification, but the overall SoS expected changes cannot be verified? How do you test the contribution of a system to the end to end SoS performance in the absence of other SoS elements critical to the SoS results? Issue Discussion (5) What if systems all implemented to their specification, but the overall SoS expected changes cannot be verified? How do you test the contribution of a system to the end to end SoS performance in the absence of other SoS elements critical to the SoS results? • SoS increments are based on changes in constituent systems which cumulate in improvements in the SoS overall • In most cases, changes in systems can be specified and tested • However, in SoS which are implemented in a variety of different environments and are dependent on networks for end to end performance • Impact of the system changes on SoS end-to-end capabilities can be estimated with less certainty • This uncertainty must be considered when assessing the SoS against its performance objectives Summary and Recommendations Approach SoS T&E as an evidence based approach to addressing risk Encourage development of analytic methods to support planning and assessment Address independent evaluation of networks which support multiple SoS Employ a range of venues to assess SoS performance over time Establish a robust process for feedback once fielded Summary and Recommendations Approach SoS T&E as providing evidence for risk assessment (1 of 2) Motivation – SOS T&E limitations represent operational performance risk • Full conventional T&E before fielding may be impractical for incremental changes in SoS based on systems with asynchronous development paths • Explicit test conditions at the SoS level can be infeasible due to difficulty in bringing all constituent systems together and set up meaningful test conditions Risk assessment in the SOS T&E context • With each increment of SoS development, identify areas critical to success of the increment and places where changes in the increment could have adverse impacts on the user missions, and focus pre-deployment T&E on these risks areas • Assess the risk using evidence from a range of sources including live test Summary and Recommendations Approach SoS T&E as providing evidence for risk assessment (2 of 2) Evidence can be based on • Activity at the SoS level, as well roll-ups of activity at the level of the constituent systems. • Activity can be explicit verification testing, results of models and simulations, use of linked integration facilities, and results of system level operational test and evaluation. Factor these risks into SoS and System development plan • If the results of the T&E indicate that the changes will have a negative impact, then these can be discarded without jeopardizing the delivery of the systems updates to their users Results would used to provide feedback to end users in the form of ‘capabilities and limitations’ as is done by the Navy Battle Group Assessment process, rather than as test criteria for SoS ‘deployment’. Summary and Recommendations Encourage development of analytic methods to support planning and assessment Analytical models of the SoS behavior can serve as effective tools to • Assess system level performance values against SoS operational scenarios • Validate the allocations to systems • Provide the analytical framework for SoS level verification. Such models can be used to develop reasonable expectations for SoS performance • Relevant operational conditions should be developed with end user input and guided by design of experiments discipline, so as to expose a broad a range of conditions Summary and Recommendations Address independent evaluation of networks which support multiple SoS Because of its central role and distributed responsibility, the network is a unique constituent of the SoS. • The implications demand a unique assessment of the network capability in test and evaluation of SoS. • Realistic assessment of SoS performance demands evaluation of the network performance and it’s degradation under the vagaries of operational conditions Because DoD seeks to develop a set of network capabilities which are applied in a wide range of applications • Consider an approach to network assessment independent of particular SoS applications, as an input to SoS planning and T&E Summary and Recommendations Employ a range of venues to assess SoS performance over time Evaluation criteria are conventionally established based on quantified performance requirements • For SoS these may be end-user metrics used to assess the results of loosely defined capabilities Recommend using a range of available opportunities to collect data on SoS performance • Assessment opportunities will be both planned and spontaneous • These may not be expressly timed to the development and fielding of system changes to address SoS capability objectives These data can • Support periodic assessments of evolving capability • Provide valuable insight to developers and users including the opportunity to identify unexpected behavior Summary and Recommendations Establish a robust process for feedback once fielded (1 of 2) Once deployed, continuing "T&E" of the SoS capability of the fielded operations can be used to recognize operational problems and make improvements Continual evaluation can be facilitated through system instrumentation and data collection to provide feedback on • Constraints • Incipient failures warnings • Unique operational conditions By establishing and exercising robust feedback mechanisms between field organizations and their operations and the SoS SE and management teams SoS T&E provides a vital link to the ongoing operational needs for the SoS Summary and Recommendations Establish a robust process for feedback once fielded (2 of 2) This includes technical and organizational dimensions • An example of the former is instrumenting systems for feedback post-fielding • An example of the latter is posting a member of the SoS SE and management team with the SoS operational organization Well-developed and continually exercised feedback mechanisms between operational and acquisition/development communities are • A big enabler of “building the system right” • Continuing to do so throughout the multiple increments of an SoS Next Steps Committee Discussion Comments (1 of 3) Paper provides a view which is generally understood, including • SoS are more complicated than systems • SoS have constituent components that have independent needs and can exercise free will in their development and operation that places their needs at times before that of the SoS • SoS are generally asynchronous and this introduces difficulties in end-to-end testing, and alternates must be sought • Emergent properties will emerge and must be accommodated • Set of ‘needs’ • Need for incremental development • Need to address synchronization; the need to consider block testing • Need to monitor fielded behavior • Need to express limitations in a politically correct manner and the means to address and incorporate methods to cope with these Comments (2 of 3) Paper does not address balance between SoS development and testing • There is also a very important topic that is not addressed; the balance between SoS development and testing. The paper is an advocate for excellent testing of an SoS and its hard to argue against that in concept, but in reality to cost of SoS development to include the SoS and the testing of it must be balanced to assure good testing by a testing program whose cost is affordable. I know inadequate testing is a recipe for disaster. But a test program whose cost significantly exceeds that of the rest of the development program is a non-starter, and may in some circumstances where adequate testing is unaffordable correctly lead to a cancellation of the conceived SoS. Practically there needs to be a balance which will require the significant inclusion of practitioners in the process. There is one ironic contradiction… On the one hand the paper cautions against the testing of sub-elements of an SoS and drawing conclusions from such tests, and on the other hand it endorses separate testing of the network that, in some cases, is part of – in fact a sub-element of – an SoS. Aside from that, for MANET networks necessary for maneuver elements, the ability to predict performance on an analytical basis is so poor that, certainly in the end, testing must be done on an SoS basis. Comments (3 of 3) Paper highlights the correct points/concepts and is well written. Eventually, these points/concepts will have to "fleshed" out We agree that defining a SoS T&E program in the classical sense is difficult. Having a V&V program focused at measures of effectiveness or operational effectiveness is a valid and useful approach; strict requirements verification would seem like a non-viable approach. Given the asynchronous development of the individual systems, an on-going assessment of built plus projected capability against SoS objectives would be very desirable. Maybe this is the same as the "evidence-based approach to addressing risk". More explanation on this section maybe needed We whole-heartedly agree that feedback after fielding is valuable and the way to go, tying it to a T&E activity maybe less effective. Once fielded and operational, change will occur and re-assessment and re-adjustment of system capabilities will be necessary. Making this a T&E function starts to imply that from a developers standpoint there is not point of completion or transition from the lab to actual use. It would seem more useful to have the feedback as input to the starting phase of the next iteration of the lifecycle rather than having the perception of a "neverending" last phase of the current iteration's lifecycle.