Session Title | Speaker | Type | Recording | Materials | Year |
---|---|---|---|---|---|
Tutorial Tutorial: Cyber Attack Resilient Weapon Systems (Abstract)
This tutorial is an abbreviated version of a 36-hour short course recently provided by UVA to a class composed of engineers working at the Defense Intelligence Agency. The tutorial provides a definition for cyber attack resilience that is an extension of earlier definitions of system resilience that were not focused on cyber attacks. Based upon research results derived by the University of Virginia over an eight year period through DoD/Army/AF/Industry funding , the tutorial will illuminate the following topics: 1) A Resilence Design Requirements methodology and the need for supporting analysis tools, 2) a System Architecture approach for achieving resilience, 3) Example resilience design patterns and example prototype implementations, 4) Experimental results regarding resilience-related roles and readiness of system operators, and 5) Test and Evaluation Issues. The tutorial will be presented by UVA Munster Professor Barry Horowitz. |
Barry Horowitz Professor, Systems Engineering University of Virginia |
Tutorial |
![]() | 2019 |
|
Breakout A Causal Perspective on Reliability Assessment (Abstract)
Causality in an engineered system pertains to how a system output changes due to a controlled change or intervention on the system or system environment. Engineered systems designs reflect a causal theory regarding how a system will work, and predicting the reliability of such systems typically requires knowledge of this underlying causal structure. The aim of this work is to introduce causal modeling tools that inform reliability predictions based on biased data sources and illustrate how these tools can inform data integration in practice. We present a novel application of the popular structural causal modeling framework to reliability estimation in an engineering application, illustrating how this framework can inform whether reliability is estimable and how to estimate reliability using data integration given a set of assumptions about the subject matter and data generating mechanism. When data are insufficient for estimation, sensitivity studies based on problem-specific knowledge can inform how much reliability estimates can change due to biases in the data and what information should be collected next to provide the most additional information. We apply the approach to a pedagogical example related to a real, but proprietary, engineering application, considering how two types of biases in data can influence a reliability calculation. |
Lauren Hund | Breakout |
![]() | 2019 |
|
Breakout Adopting Optimized Software Test Design Methods at Scale (Abstract)
Using Combinatorial Test Design methods to select software test scenarios has repeatedly delivered large efficiency and thoroughness gains – which begs the questions: • Why are these proven methods not used everywhere? • Why do some efforts to promote adoption of new approaches stagnate? • What steps can leaders take to introduce successfully introduce and spread new test design methods? For more than a decade, Justin Hunter has helped large global organizations across six continents adopt new test design techniques at scale. Working in some environments, he has felt like Sisyphus, forever condemned to roll a boulder uphill only to watch it roll back down again. In other situations, things clicked; teams smoothly adopted new tools and techniques, and impressive results were quickly achieved. In this presentation, Justin will discuss several common challenges faced by large organizations, explain why adopting test design tools is more challenging than adopting other types of development and testing tools, and share actionable recommendations to consider when you roll out new test design approaches. |
Justin Hunter | Breakout |
![]() | 2019 |
|
Breakout Making Statistically Defensible Testing ‘The way we do things around here’ (Abstract)
For the past 7 years, USAF DT&E has been exploring ways to adapt the principles of experimental design to rapidly evolving developmental test articles and test facilities – often with great success. This paper discusses three case studies that span the range of USAF DT&E activities from EW to Ground Test to Flight Test and shows the truly revolutionary impact Fisher’s DOE can have on development. The Advanced Strategic and Tactical Expendable (ASTE) testing began in 1990 to develop, enhance, and test new IR flares, flare patterns, and dispense tactics. More than 60 aircraft & flare types have been tested. The typical output is a “Pancake Plot” of 200+ “cases” of flare, aspect angle, range, elevation, and flare effectiveness using a stop-light chart (red-yellow-green) approach. In usual testing – ~3000 flare engagements costing $1M to participate. The response, troublingly enough, is binary – 15-30 binomial response trials measuring P(Decoy). Binary responses are information-poor. Legacy testing does not assess present statistical power in reporting P(decoy) results. Analysts investigated replacing P(Decoy) w/ continuous metrics – e.g. time to decoy. This research is ongoing. We found we could spread the replicates out to examine 3x to 5x more test conditions without affecting power materially. Analysis with the Generalized Linear Model (GLZ) replaced legacy “cases” analysis with 75% improvement to confidence intervals with same data. We are seeking to build a Monte-Carlo simulation to estimate how many runs are required in a logistics regression model to achieve adequate power. We hope to reduce customer expenditures for flare information by as much as 50%. Co-authors J. Higdon, B, Knight. AEDC completed a major upgrade with new nozzle hardware to vary Mach. The Arnold Transonic Wind Tunnel 4T spans the range of flows from subsonic to approximately M9. The new wind tunnel was to be computer controlled. A number of key instrumentation improvements were made at the same time. The desire was to calibrate the resulting modified tunnel. The last calibration of 4T was 25 years ago in 1990. The calibration ranged across the full range of Mach and pressure capabilities, spanning a four-D space: pressure, Mach, wall angle, and wall porosity. Both the traditional OFAT effort – vary one factor at a time – and a parallel DOE effort were run to compare design, execution, modeling, and prediction capabilities against cost and time to run. The robust embedded face-centered CCD DOE design (J. Simpson and D. Landman ’05) employed 75 vs. 176 OFAT runs. The RSM design achieved 57% run savings. Due to an admirable discipline in randomization during the DOE trials, the smaller design required longer to run. As a result of using the DOE approach, engineers found it easier predict offcondition tunnel operating characteristics using RSM models, optimize facility flow quality for any given test condition. In addition, the RSM regression models support future “spot check” calibration in future by comparing predictions to measured values. If the measurement falls within the prediction interval, the existing calibration still appropriate. AEDC is using split-plot style designs a for current wind tunnel probe calibration. Co-author: Dr Dough Garrard. |
Greg Hutto Air Force 96th Test Wing |
Breakout | 2016 |
||
Tutorial Sensitivity Experiments (Abstract)
A sensitivity experiment is a special type of experimental design that is used when the response variable is binary and the covariate is continuous. Armor protection and projectile lethality tests often use sensitivity experiments to characterize a projectile’s probability of penetrating the armor. In this minitutorial we illustrate the challenge of modeling a binary response with a limited sample size, and show how sensitivity experiments can mitigate this problem. We review eight different single covariate sensitivity experiments and present a comparison of these designs using simulation. Additionally, we cover sensitivity experiments for cases that include more than one covariate, and highlight recent research in this area. The mini-tutorial concludes with a case study by Greg Hutto on Army grenade fuze testing, titled “Preventing Premature ZAP: EMPATHY Capacitive Design With 3 Phase Optimal Design (3pod).” |
Greg Hutton U.S. Air Force , 96 Test Wing |
Tutorial | Materials | 2016 |
|
Risk Comparison and Planning for Bayesian Assurance Tests (Abstract)
Designing a Bayesian assurance test plan requires choosing a test plan that guarantees a product of interest is good enough to satisfy consumer’s criteria but not ‘so good’ that it causes producer’s concern if they fail the test. Bayesian assurance tests are especially useful because they can incorporate previous product information in the test planning and explicitly control levels of risk for the consumer and producer. We demonstrate an algorithm for efficiently computing a test plan given desired levels of risks in binomial and exponential testing. Numerical comparisons with the Operational Characteristic (OC) curve, Probability Ratio Sequential Test (PRST), and a simulation-based Bayesian sample size determination approach are also considered. |
Hyoshin Kim North Carolina State University ![]() (bio)
Hyoshin Kim received her B.Ec. in Statistics from Sungkyunkwan University, South Korea, in 2017, and her M.S. in Statistics from Seoul National University, South Korea, in 2019. She is currently a third year Ph.D. student at the department of Statistics at North Carolina State University. Her research interests are Bayesian assurance testing and Bayesian clustering algorithms for high dimensional correlated outcomes. |
Session Recording |
![]() Recording | 2022 |
|
Breakout Machine Learning: Overview and Applications to Test (Abstract)
“Machine learning is quickly gaining importance in being able to infer meaning from large, high-dimensional datasets. It has even demonstrated performance meeting or exceeding human capabilities in conducting a particular set of tasks such as speech recognition and image recognition. Employing these machine learning capabilities can lead to increased efficiency in data collection, processing, and analysis. Presenters will provide an overview of common examples of supervised and unsupervised learning tasks and algorithms as an introduction to those without experience in machine learning. Presenters will also provide motivation for machine learning tasks and algorithms in a variety of test and evaluation settings. For example, in both developmental and operational test, restrictions on instrumentation, number of sorties, and the amount of time allocated to analyze collected data make data analysis challenging. When instrumentation is unavailable or fails, a common back-up data source is an over-the-shoulder video recording or recordings of aircraft intercom and radio transmissions, which traditionally are tedious to analyze. Machine learning based image and speech recognition algorithms can assist in extracting information quickly from hours of video and audio recordings. Additionally, unsupervised learning techniques may be used to aid in the identification of influences of logged or uncontrollable factors in many test and evaluation settings. Presenters will provide a potential example for the application of unsupervised learning techniques to test and evaluation.” |
Takayuki Iguchi AFOTEC |
Breakout | Materials | 2017 |
|
Breakout Applying Design of Experiments to Cyber Testing (Abstract)
We describe a potential framework for applying DOE to cyber testing and provide an example of its application to testing of a hypothetical command and control system. |
J. Michael Gilmore Research Staff Member IDA ![]() (bio)
Dr. James M. Gilmore “Mike” Research Staff Member | IDA SED Fields of Expertise Test and evaluation; cost analysis; cost-effectiveness analysis. Education 1976 – 1980 Doctor of Philosophy in Nuclear Engineering at University of Wisconsin 1972 – 1976 Bachelor of Science in Physics at M.I.T. Employment 2017 – 2018 Principal Physical Scientist, RAND Corporation Performed various analyses for federal government clients 2009 – 2017 Director, Operational Test and Evaluation, Department of Defense Senate-confirmed Presidential appointee serving as the principal advisor to the Secretary of Defense regarding the operational effectiveness of all defense systems 2001 – 2009 Assistant Director for National Security , Congressional Budget Office Responsible for the CBO division performing analyses of a broad array of issues in national security for committees of the U.S. Congress 1994 – 2001 Deputy Director for General Purpose Programs, OSD Program Analysis and Evaluation Responsible for four divisions performing analyses and evaluations of all aspects of DoD’s conventional forces and associated programs 1993 – 1994 Division Director, OSD Progam Analysis and Evaluation Responsible for divisions in the Cost Analysis Improvement Group performing independent cost analyses of major defense acquisition programs 1990 – 1993 Analyst, OSD Program Analysis and Evaluation Performed analysis of strategic defense systems and command, control, and communications systems 1989 – 1990 Analyst, Falcon Associates Performed various analyses for DoD clients 1985 – 1989 Analyst, McDonnell Douglas Performed analysis involving issues in command, control, communications, and intelligence 1981 – 1985 Scientist, Lawrence Livermore National Laboratory Modelled nuclear fusion experiments |
Breakout | Session Recording |
![]() Recording | 2022 |
Breakout VV&UQ – Uncertainty Quantification for Model-Based Engineering of DoD Systems (Abstract)
The US Army ARDEC has recently established an initiative to integrate statistical and probabilistic techniques into engineering modeling and simulation (M&S) analytics typically used early in the design lifecycle to guide technology development. DOE-driven Uncertainty Quantification techniques, including statistically rigorous model verification and validation (V&V) approaches, enable engineering teams to identify, quantify, and account for sources of variation and uncertainties in design parameters, and identify opportunities to make technologies more robust, reliable, and resilient earlier in the product’s lifecycle. Several recent armament engineering case studies – each with unique considerations and challenges – will be discussed. |
Melissa Jablonski US Army |
Breakout | Materials | 2017 |
|
Contributed Development of a Locking Setback Mass for Cluster Munition Applications: A UQ Case Study (Abstract)
The Army is currently developing a cluster munition that is required to meet functional reliability requirements of 99%. This effort focuses on the design process for a setback lock within the safe and arm (S&A) device in the submunition fuze. This lock holds the arming rotor in place, thus preventing the fuze from beginning its arming sequence until the setback lock detracts during a launch event. Therefore, the setback lock is required to not arm (remain in place) during a drop event (safety) and to arm during a launch event (reliability). In order to meet these requirements, uncertainty quantification techniques were used to evaluate setback lock designs. We designed a simulation experiment, simulated the setback lock behavior in a drop event and in a launch event, fit a model to the results, and optimized the design for safety and reliability. Currently, 8 candidate designs that meet the requirements are being manufactured, and adaptive sensitivity testing is planned to inform the surrogate models and improve their predictive capability. A final optimized design will be chosen based on the improved models, and realistic drop safety and arm reliability predictions will be obtained using Monte-Carlo simulations of the surrogate models. |
Melissa Jablonski U.S. ARMY ARMAMENT RESEARCH, DEVELOPMENT & ENGINEERING CENTER |
Contributed | Materials | 2018 |
|
Breakout Metrics for Assessing Underwater Demonstrations for Detection and Classification of UXO (Abstract)
Receiver Operating Characteristic curves (ROC curves) are often used to assess the performance of detection and classification systems. ROC curves can have unexpected subtleties that make them difficult to interpret. For example, the Strategic Environmental Research and Development Program and the Environmental Security Technology Certification Program (SERDP/ESTCP) is sponsoring the development of novel systems for the detection and classification of Unexploded Ordnance (UXO) in underwater environments. SERDP is also sponsoring underwater testbeds to demonstrate the performance of these novel systems. The Institute for Defense Analyses (IDA) is currently designing and implementing the scoring process for these underwater demonstrations that addresses the subtleties of ROC curve interpretation. This presentation will provide an overview of the main considerations for ROC curve parameter selection when scoring underwater demonstrations for UXO detection and classification. |
Jacob Bartel Research Associate Institute for Defense Analyses ![]() (bio)
Jacob Bartel is a Research Associate at the Institute for Defense Analyses (IDA). His research focuses on computational modeling and verification and validation (V&V), primarily in the field of nuclear engineering. Recently, he has worked with SERDP/ESTCP to develop and implement scoring processes for testing underwater UXO detection and classification systems. Prior to joining IDA, his graduate research focused on the development of novel algorithms to model fuel burnup in nuclear reactors. Jacob earned his master’s degree in Nuclear Engineering and his bachelor’s degree in Physics from Virginia Tech. |
Breakout |
![]() | 2021 |
|
Panel Army’s Open Experimentation Test Range for Internet of Battlefield Things: MSA-DPG (Abstract)
One key feature of future Multi-Domain Operations (MDO) is expected to be the ubiquity of devices providing information connected in an Internet of Battlefield Things (IoBT). To this end, U.S. Army aims to advance the underlying science of pervasive and heterogeneous IoBT sensing, networking, and actuation. In this effort, IoBT experimentation testbed is an integral part of the capability development, which evaluates and validates the scientific theories, algorithms, and technologies integrated with C2 systems under the military scenarios. Originally conceived for this purpose, Multi-Purpose Sensing Area Distributed Proving Ground (MSA-DPG) is an open-range test bed developed by the Army Research Laboratory (ARL). We discuss the vision and the development of MSA-DPG and its fundamental roles of MSA-DPG in research serving the communities of Military Sciences. |
Jade Freeman Research Scientist U.S. Army DEVCOM Army Research Laboratory ![]() (bio)
Dr. Jade Freeman currently serves as the Associate Branch Chief and a Team Lead at Battlefield Information Systems Branch. In this capacity, Dr. Freeman oversees information systems and engineering research projects and analyses. Prior to joining ARL, Dr. Freeman served as the Senior Statistician at the Office of Cybersecurity and Communications, Department of Homeland Security. Throughout the career, her work in operations and research includes cyber threat analyses, large survey design and analyses, experimental design, survival analysis, and missing data imputation methods. Dr. Freeman is also a PMP certified project manager, experienced in leading and managing IT development projects. Dr. Freeman obtained a Ph. D. in Statistics from the George Washington University. |
Panel |
![]() | 2021 |
|
Analysis of Target Location Error using Stochastic Differential Equations (Abstract)
This paper presents an analysis of target location error (TLE) based on the Cox Ingersoll Ross (CIR) model. In brief, this model characterizes TLE as a function of range based the stochastic differential equation model dX(r) = a(b-X(r))dr + sigma *sqrt(X(r)) dW(r) where X(t) is TLE at range r, b is the long-term mean (terminal) of the TLE, a is the rate of reversion of X(r) to b, sigma is the process volatility, and W(t) is the standard Weiner process. Multiple flight test runs under the same conditions exhibit different realizations of the TLE process. This approach to TLE analysis models each flight test run as a realization the CIR process. Fitting a CIR model to multiple data runs then provides a characterization of the TLE system under test. This paper presents an example use of the CIR model. Maximum likelihood estimates of the parameters of the CIR model are found from a collection of TLE data runs. The resulting CIR model is then used to characterize overall system TLE performance as a function of range to the target as well as the asymptotic estimate of long-term TLE. |
James Brownlow mathematical statistician USAF ![]() (bio)
Dr. James Brownlow is a tech expert in statistics with the USAF, Edwards AFB, CA. His PhD is in time series from UC Riverside. Dr. Brownlow has developed test and evaluation procedures using Bayesian techniques, and developed Python code to adapt parametric survival models to the analysis of target location error. He is a coauthor of a paper that used stochastic differential equations to characterize Kalman-filtered estimates of target track state vectors. |
Session Recording |
![]() Recording | 2022 |
|
Breakout Assessing Next-Gen Spacesuit Reliability: A Probabilistic Analysis Case Study at NASA (Abstract)
Under the Artemis program, the Exploration Extravehicular Mobility Unit (xEMU) spacesuit will ensure the safety of NASA astronauts during the targeted 2024 return to the moon. Efforts are currently underway to finalize and certify the xEMU design. There is a delicate balance between producing a spacesuit that is robust enough to safely withstand potential fall events while still satisfying stringent mass and mobility requirements. The traditional approach of considering worst case-type loading and applying conservative factors of safety (FoS) to account for uncertainties in the analysis was unlikely to meet the narrow design margins. Thus, the xEMU design requirement was modified to include a probability of no impact failure (PnIF) threshold that must be verified through probabilistic analysis. As part of a broader one year effort to help integrate modern uncertainty quantification (UQ) methodology into engineering practice at NASA, the certification of the xEMU spacesuit was selected as the primary case study. The project, led by NASA Langley Research Center (LaRC) under the Engineering Research & Analysis (R&A) Program in 2020, aimed to develop an end-to-end UQ workflow for engineering problems and to help facilitate reliability-based design at NASA. The main components of the UQ workflow included 1) sensitivity analysis to identify the most influential model parameters, 2) model calibration to quantified model parameter uncertainties using experimental data, and 3) uncertainty propagation for producing probabilistic model predictions and estimating reliability. Particular emphasis was placed on overcoming the common practical barrier of prohibitive computational expense associated with probabilistic analysis by leveraging state-of-the-art UQ methods and high performance computing (HPC). In lieu of mature computational models and test data for the xEMU at the time of the R&A Program, the UQ workflow for estimating PnIF was demonstrated using existing models and data from the previous generation of spacesuits (the Z-2). However, the lessons learned and capabilities developed in the process of the R&A are directly transferable to the ongoing xEMU certification effort and are currently being integrated in 2021. This talk provides an overview of the goals of and findings under NASA’s UQ R&A project, focusing on the spacesuit certification case study. The steps of the UQ workflow applied to the Z-2 spacesuit using the available finite element method (FEM) models and impact test data will be detailed. The ability to quantify uncertainty in the most influential subset of FEM model input parameters and then propagate that uncertainty to estimates of PnIF is demonstrated. Since the FEM model of the full Z-2 assembly took nearly 1 day to execute just once, the advanced UQ methods and HPC utilization required to make the probabilistic analysis tractable are discussed. Finally, the lessons learned from conducting the case study are provided along with planned ongoing/future work for the xEMU certification in 2021. |
James Warner Computational Scientist NASA Langley Research Center ![]() (bio)
Dr. James Warner joined NASA Langley Research Center (LaRC) in 2014 as a Research Computer Engineer after receiving his PhD in Computational Solid Mechanics from Cornell University. Previously, he received his B.S. in Mechanical Engineering from SUNY Binghamton University and held temporary research positions at the National Institute of Standards and Technology and Duke University. Dr. Warner is a member of the Durability, Damage Tolerance, and Reliability Branch (DDTRB) at LaRC, where he focuses on developing computationally-efficient approaches for uncertainty quantification for a range of applications including structural health management and space radiation shielding design. His other research interests include high performance computing, inverse methods, and topology optimization. |
Breakout |
![]() | 2021 |
|
Breakout Machine Learning for Uncertainty Quantification: Trusting the Black Box (Abstract)
Adopting uncertainty quantification (UQ) has become a prerequisite for providing credibility in modeling and simulation (M&S) applications. It is well known, however, that UQ can be computationally prohibitive for problems involving expensive high-fidelity models, since a large number of model evaluations is typically required. A common approach for improving efficiency is to replace the original model with an approximate surrogate model (i.e., metamodel, response surface, etc.) using machine learning that makes predictions in a fraction of the time. While surrogate modeling has been commonplace in the UQ field for over a decade, many practitioners still remain hesitant to rely on “black box” machine learning models over trusted physics-based models (e.g., FEA) for their analyses. This talk discusses the role of machine learning in enabling computational speedup for UQ, including traditional limitations and modern efforts to overcome them. An overview of surrogate modeling and its best practices for effective use is first provided. Then, some emerging methods that aim to unify physics-based and data-based approaches for UQ are introduced, including multi-model Monte Carlo simulation and physics-informed machine learning. The use of both traditional surrogate modeling and these more advanced machine learning methods for UQ are highlighted in the context of applications at NASA, including trajectory simulation and spacesuit certification. |
James Warner Computational Scientist NASA Langley Research Center (bio)
Dr. James (Jim) Warner joined NASA Langley Research Center (LaRC) in 2014 as a Research Computer Engineer after receiving his PhD in Computational Solid Mechanics from Cornell University. Previously, he received his B.S. in Mechanical Engineering from SUNY Binghamton University and held temporary research positions at the National Institute of Standards and Technology and Duke University. Dr. Warner is a member of the Durability, Damage Tolerance, and Reliability Branch (DDTRB) at LaRC, where he focuses on developing computationally-efficient approaches for uncertainty quantification for a range of applications including structural health management, additive manufacturing, and trajectory simulation. Additionally, he works to bridge the gap between UQ research and NASA mission impact, helping to transition state-of-the-art methods to solve practical engineering problems. To that end, he is currently involved in efforts to certify the xEMU spacesuit and develop guidance systems for entry, descent, and landing for Mars landing. His other research interests include machine learning, high performance computing, and topology optimization. |
Breakout | Session Recording |
![]() Recording | 2022 |
Breakout Stochastic Modeling and Characterization of a Wearable-Sensor-Based Surveillance Network (Abstract)
Current disease outbreak surveillance practices reflect underlying delays in the detection and reporting of disease cases, relying on individuals who present symptoms to seek medical care and enter the health care system. To accelerate the detection of outbreaks resulting from possible bioterror attacks, we introduce a novel two-tier, human sentinel network (HSN) concept composed of wearable physiological sensors capable of pre-symptomatic illness detection, which prompt individuals to enter a confirmatory stage where diagnostic testing occurs at a certified laboratory. Both the wearable alerts and test results are reported automatically and immediately to a secure online platform via a dedicated application. The platform aggregates the information and makes it accessible to public health authorities. We evaluated the HSN against traditional public health surveillance practices for outbreak detection of 80 Bacillus anthracis (Ba) release scenarios in mid-town Manhattan, NYC. We completed an end-to-end modeling and analysis effort, including the calculation of anthrax exposures and doses based on computational atmospheric modeling of release dynamics, and development of a custom-built probabilistic model to simulate resulting wearable alerts, diagnostic test results, symptom onsets, and medical diagnoses for each exposed individual in the population. We developed a novel measure of network coverage, formulated new metrics to compare the performance of the HSN to public health surveillance practices, completed a Design of Experiments to optimize the test matrix, characterized the performant trade-space, and performed sensitivity analyses to identify the most important engineering parameters. Our results indicate that a network covering greater than ~10% of the population would yield approximately a 24-hour time advantage over public health surveillance practices in identifying outbreak onset, and provide a non-target-specific indication (in the form of a statistically aberrant number of wearable alerts) of approximately 36-hours; these earlier detections would enable faster and more effective public health and law enforcement responses to support incident characterization and decrease morbidity and mortality via post-exposure prophylaxis. |
Jane E. Valentine Senior Biomedical Engineer Johns Hopkins University Applied Physics Laboratory (bio)
Jane Valentine received her B.S. in Mathematics and French, and Ph.D. in Biomedical Engineering, both from Carnegie Mellon University. She then completed a post-doc in Mechanical Engineering at the University of Illinois, and a data science fellowship in the United Kingdom, working with a pharmaceutical company. She has been working at the Johns Hopkins University Applied Physics Laboratory since 2020, where she works on mathematical modeling and simulation, optimization, and data science, particularly in the areas of biosensors, knowledge graphs, and epidemiological modeling. |
Breakout | Session Recording |
![]() Recording | 2022 |
Breakout Challenges in Test and Evaluation of AI: DoD’s Project Maven (Abstract)
The Algorithmic Warfare Cross Functional Team (AWCFT or Project Maven) organizes DoD stakeholders to enhance intelligence support to the warfighter through the use of automation and artificial intelligence. The AWCFT’s objective is to turn the enormous volume of data available to DoD into actionable intelligence and insights at speed. This requires consolidating and adapting existing algorithm-based technologies as well as overseeing the development of new solutions. This brief will describe some of the methodological challenges in test and evaluation that the Maven team is working through to facilitate speedy and agile acquisition of reliable and effective AI / ML capabilities. |
Jane Pinelis | Breakout | 2019 |
||
Breakout Sample Size Calculations for Quiet Sonic Boom Community Surveys (Abstract)
NASA is investigating the dose-response relationship between quiet sonic boom exposure and community noise perceptions. This relationship is the key to possible future regulations that would replace the ban on commercial supersonic flights with a noise limit. We have built several Bayesian statistical models using pilot community study data. Using goodness of fit measures, we downselected to a subset of models which are the most appropriate for the data. From this subset of models we demonstrate how to calculate sample size requirements for a simplified example without any missing data. We also suggest how to modify the sample size calculation to account for missing data. |
Jasme Lee | Breakout |
![]() | 2019 |
|
Breakout Topological Modeling of Human-Machine Teams (Abstract)
A Human-Machine Team (HMT) is a group of agents consisting of at least one human and at least one machine, all functioning collaboratively towards one or more common objectives. As industry and defense find more helpful, creative, and difficult applications of AI-driven technology, the need to effectively and accurately model, simulate, test, and evaluate HMTs will continue to grow and become even more essential. Going along with that growing need, new methods are required to evaluate whether a human-machine team is performing effectively as a team in testing and evaluation scenarios. You cannot predict team performance from knowledge of the individual team agents, alone; interaction between the humans and machines – and interaction between team agents, in general – increases the problem space and adds a measure of unpredictability. Collective team or group performance, in turn, depends heavily on how a team is structured and organized, as well as the mechanisms, paths, and substructures through which the agents in the team interact with one another – i.e. the team’s topology. With the tools and metrics for measuring team structure and interaction becoming more highly developed in recent years, we will propose and discuss a practical, topological HMT modeling framework that not only takes into account but is actually built around the team’s topological characteristics, while still utilizing the individual human and machine performance measures. |
Jay Wilkins and Caitlan Fealing Research Staff Member IDA (bio)
Dr. Wilkins received his Ph.D. in Mathematics from the University of Tennessee, and then spent several years as a postdoctoral fellow and professor in academia before moving into the realm of defense research. After three years at the US Army Research Office (part of the US Army Research Lab) managing the Mathematical Analysis and Complex Systems Program, he came to IDA as a research staff member in January 2020. His mathematical background is in applied topological and geometric analysis, particularly the application of topology and geometry to the solution of problems in networking and optimal transport. His current project work at IDA includes team modeling and simulation, sequential testing of chemical agent detectors, and helicopter pilot guidance and firing interfaces. |
Breakout | Session Recording |
![]() Recording | 2022 |
Breakout Valuing Human Systems Integration: A Test and Data Perspective (Abstract)
Technology advances are accelerating at a rapid pace, with the potential to enable greater capability and power to the Warfighter. However, if human capabilities and limitations are not central to concepts, requirements, design, and development then new/upgraded weapons and systems will be difficult to train, operate, and maintain, may not result in the skills, job, grade, and manpower mix as projected, and may result in serious human error, injury or Soldier loss. The Army Human Systems Integration (HSI) program seeks to overcome these challenges by ensuring appropriate consideration and integration of seven technical domains: Human Factors Engineering (e.g., usability), Manpower, Personnel, Training, Safety and Occupational Health, Habitability, Force Protection and Survivability. The tradeoffs, constraints, and limitations occurring among and between these technical domains allows HSI to execute a coordinated, systematic process for putting the warfighter at the center of the design process – equipping the warfighter rather than manning equipment. To that end, the Army HSI Headquarters, currently as a directorate within the Army Headquarters Deputy Chief of Staff (DCS), G-1 develops strategies and ensures human systems factors are early key drivers in concepts, strategy, and requirements, and are fully integrated throughout system design, development, testing and evaluation, and sustainment The need to consider HSI factors early in the development cycle is critical. Too often, man-machine interface issues are not addressed until late in the development cycle (i.e. production and deployment phase) after the configuration of a particular weapon or system has been set. What results is a degraded combat capability, suboptimal system and system-of-systems integration, increased training and sustainment requirements, or fielded systems not in use. Acquisition test data are also good sources to glean HSI return on investment (ROI) metrics. Defense acquisition reports such as test and evaluation operational assessments identifies HSI factors as root causes when Army programs experience increase cost, schedule overruns, or low performance. This is identifiable by the number and type of systems that require follow-on test and evaluation (FOT&E), over reliance on field service representatives (FSRs), costly and time consuming engineering change requests (ECRs), or failures in achieving reliability, availability, and maintainability (RAM) key performance parameters (KPPs) and key system attributes (KSAs). In this presentation, we will present these data and submit several return on investment (ROI) metrics, closely aligned to the defense acquisition process, to emphasize and illustrate the value of HSI. Optimizing Warfighter-System performance and reducing human errors, minimizing risk of Soldier loss or injury, and reducing personnel and materiel life cycle costs produces data that are inextricably linked to early, iterative, and measurable HSI processes within the defense acquisition system. |
Jeffrey Thomas | Breakout |
![]() | 2019 |
|
Panel The Keys to Successful Collaborations during Test and Evaluation: Panelist |
Willis Jensen HR Analyst W.L. Gore & Associates ![]() (bio)
Dr. Willis Jensen is a member of the HR Analytics team at W.L. Gore & Associates, where he supports people related analytics work across the company. At Gore, he previously spent 12 years as a statistician and as the Global Statistics Team Leader where he led a team of statisticians that provided statistical support and training across the globe. He holds degrees in Statistics from Brigham Young University and a Ph.D. in Statistics from Virginia Tech. |
Panel | Session Recording |
Recording | 2021 |
Breakout Orbital Debris Effects Prediction Tool for Satellite Constellations (Abstract)
Based on observations gathered from the IDA Forum on Orbital Debris (OD) Risks and Challenges (October 8-9, 2020), DOT&E needed first-order predictive tools to evaluate the effects of orbital debris on mission risk, catastrophic collision, and collateral damage to DOD spacecraft and other orbital assets – either from unintentional or intentional [Anti-Satellite (ASAT)] collisions. This lack of modeling capability hindered DOT&E’s ability to evaluate the risk to operational effectiveness and survivability of individual satellites and large constellations, as well as risks to the overall use of space assets in the future. Part 1 of this presentation describes an IDA-derived Excel-based tool (SatPen) for determining the probability and mission effects of >1mm orbital debris impacts and penetration on individual satellites in low Earth orbit (LEO). IDA estimated the likelihood of satellite mission loss using a Starlink-like satellite as a case study and NASA’s ORDEM 3.1 orbital debris environment as an input, supplemented with typical damage prediction equations to support mission loss predictions. Part 2 of this presentation describes an IDA-derived technique (DebProp) to evaluate the debris propagating effects of large, trackable debris (>5 cm) or antisatellite weapons colliding with satellites within constellations. IDA researchers again used a Starlink-like satellite as a case study and worked with Stellingwerf Associates to modify the Smooth Particle Hydrodynamic Code (SPHC) in order to predict the number and direction of fragments following a collision by a tracked satellite fragment. The result is a file format that is readable as an input file for predicting orbital stability or debris re-entry for thousands of created particles, and predict additional, short-term OD-induced losses to other satellites in the constellation. By pairing these techniques, IDA can predict additional, short-term and long-term OD-induced losses to other satellites in the constellation, and conduct long-term debris growth studies. |
Joel Williamsen Research Staff Member IDA ![]() (bio)
FIELDS OF EXPERTISE Air and space vehicle survivability, missile lethality, LFT&E, ballistic response, active protection systems, hypervelocity impact, space debris, crew and passenger casualty assessment EDUCATION HISTORY 1993 Doctor of Philosophy in Systems Engineering at University of Alabama, Huntsville 1989 Master of Science in Engineering Management at University of Alabama, Huntsville 1983 Bachelor of Science in Mechanical Engineering at University of Nebraska EMPLOYMENT HISTORY 2003 – Present Research Staff Member, IDA, OED 1998 – 2003 Director, Center for Space Systems Survivability, University of Denver 1987 – 1998 Spacecraft Survivability Design , NASA-Marshall Space Flight Center, NASA 1983 – 1987 U.S. Army Missile Command, Research Development and Engineering Center, Warhead Design, U.S. Army PROFESSIONAL ACTIVITIES American Institute of Aeronautics and Astronautics (Chair, Survivability Technical Committee, 2001-2003) Tau Beta Pi Engineering Honorary Society Pi Tau Sigma Mechanical Engineering Honorary Society HONORS IDA Welch Award, 2020. National AIAA Survivability Award, 2012. Citation reads, “For outstanding achievement in enhancing spacecraft, aircraft, and crew survivability through advanced meteoroid/orbital debris shield designs, on-orbit repair techniques, risk assessment tools, and live fire evaluation.” NASA Astronauts’ Personal Achievement Award (Silver Snoopy), 2001. NASA Exceptional Achievement Medal, Spacecraft Survivability Analysis, 1995. Army Research and Development Achievement Award, 1985. Patents and Statutory Invention Registrations: Enhanced Hypervelocity Impact Shield, 1997. Joint. Patents and Statutory Invention Registrations: Pressure Wall Patch, 1994. Joint. Patents and Statutory Invention Registrations: Advanced Anti-Tank Airframe Configuration Tandem Warhead Missile, 1991. Joint. Patents and Statutory Invention Registrations: Extendible Shoulder Fired Anti-tank Missile, 1990. Joint. Patents and Statutory Invention Registrations: Particulated Density Shaped Charge Liner, 1987. Patents and Statutory Invention Registrations: High Velocity Rotating Shaped Charge Warhead, 1986. Patents and Statutory Invention Registrations: Missile Canting Shaped Charge Warhead, 1985. Joint. NASA Group Achievement Awards (Space Station), 1992-1994. NASA Group Achievement Awards (Hubble System Review Team) 1989, 1990. Outstanding Performance Awards, 1984-1988, 1990, 1992-1997. First NASA-MSFC representative to International Space University, 1989. |
Breakout | Session Recording |
Recording | 2022 |
Data Science & ML-Enabled Terminal Effects Optimization (Abstract)
Warhead design and performance optimization against a range of targets is a foundational aspect of the Department of the Army’s mission on behalf of the warfighter. The existing procedures utilized to perform this basic design task do not fully leverage the exponential growth in data science, machine learning, distributed computing, and computational optimization. Although sound in practice and methodology, existing implementations are laborious and computationally expensive, thus limiting the ability to fully explore the trade space of all potentially viable solutions. An additional complicating factor is the fast paced nature of many Research and Development programs which require equally fast paced conceptualization and assessment of warhead designs. By utilizing methods to take advantage of data analytics, the workflow to develop and assess modern warheads will enable earlier insights, discovery through advanced visualization, and optimal integration of multiple engineering domains. Additionally, a framework built on machine learning would allow for the exploitation of past studies and designs to better inform future developments. Combining these approaches will allow for rapid conceptualization and assessment of new and novel warhead designs. US overmatch capability is quickly eroding across many tactical and operational weapon platforms. Traditional incremental improvement approaches are no longer generating appreciable performance improvements to warrant investment. Novel next generation techniques are required to find efficiencies in designs and leap forward technologies to maintain US superiority. The proposed approach seeks to shift existing design mentality to meet this challenge. |
John Cilli Computer Scientist Picatinny Arsenal ![]() (bio)
My name is John Cilli, I am a recent graduate of East Stroudsburg University with a bachelor’s in Computer Science. I have been working at Picatinny within the Systems Analysis Division as a computer scientist for little over a year now. |
Session Recording |
![]() Recording | 2022 |
|
Breakout Advancements in Characterizing Warhead Fragmentation Events (Abstract)
Fragmentation analysis is a critical piece of the live fire test and evaluation (LFT&E) of lethality and vulnerability aspects of warheads. But the traditional methods for data collection are expensive and laborious. New optical tracking technology is promising to increase the fidelity of fragmentation data, and decrease the time and costs associated with data collection. However, the new data will be complex, three dimensional ‘fragmentation clouds’, possibly with a time component as well. This raises questions about how testers can effectively summarize spatial data to draw conclusions for sponsors. In this briefing, we will discuss the Bayesian spatial models that are fast and effective for characterizing the patterns in fragmentation data, along with several exploratory data analysis techniques that help us make sense of the data. Our analytic goals are to – Produce simple statistics and visuals that help the live fire analyst compare and contrast warhead fragmentations; – Characterize important performance attributes or confirm design/spec compliance; and – Provide data methods that ensure higher fidelity data collection translates to higher fidelity modeling and simulation down the line. This talk is a version of the first-step feasibility study IDA is taking – hopefully much more to come as we continue to work on this important topic. |
John Haman Research Staff Member Institute for Defense Analyses ![]() (bio)
Dr. John Haman is a statistician at the Institute for Defense Analyses, where he develops methods and tools for analyzing test data. He has worked with a variety of Army, Navy, and Air Force systems, including counter-UAS and electronic warfare systems. Currently, John is providing technical support on operational testing to the Joint Artificial Intelligence Center. |
Breakout | Session Recording |
![]() Recording | 2021 |
Breakout What statisticians should do to improve M&S validation studies (Abstract)
It is often said that many research findings — from social sciences, medicine, economics, and other disciplines — are false. This fact is trumpeted in the media and by many statisticians. There are several reasons that false research is published, but to what extent should we be worried about them in defense testing and in particular modeling and simulation validation studies? In this talk I will present several recommendations for actions that statisticians and data scientists can take to improve the quality of our validations and evaluations. |
John Haman RSM IDA ![]() (bio)
Dr. John Haman is a research staff member at the Institute for Defense Analyses, where he leads a team of analysts that develop methods and tools for analyzing test data. He has worked with a variety of Army, Navy, and Air Force systems, including counter-UAS and electronic warfare systems. |
Breakout | Session Recording |
![]() Recording | 2022 |
Session Title | Speaker | Type | Recording | Materials | Year |
---|---|---|---|---|---|
Tutorial Tutorial: Cyber Attack Resilient Weapon Systems |
Barry Horowitz Professor, Systems Engineering University of Virginia |
Tutorial |
![]() | 2019 |
|
Breakout A Causal Perspective on Reliability Assessment |
Lauren Hund | Breakout |
![]() | 2019 |
|
Breakout Adopting Optimized Software Test Design Methods at Scale |
Justin Hunter | Breakout |
![]() | 2019 |
|
Breakout Making Statistically Defensible Testing ‘The way we do things around here’ |
Greg Hutto Air Force 96th Test Wing |
Breakout | 2016 |
||
Tutorial Sensitivity Experiments |
Greg Hutton U.S. Air Force , 96 Test Wing |
Tutorial | Materials | 2016 |
|
Risk Comparison and Planning for Bayesian Assurance Tests |
Hyoshin Kim North Carolina State University ![]() |
Session Recording |
![]() Recording | 2022 |
|
Breakout Machine Learning: Overview and Applications to Test |
Takayuki Iguchi AFOTEC |
Breakout | Materials | 2017 |
|
Breakout Applying Design of Experiments to Cyber Testing |
J. Michael Gilmore Research Staff Member IDA ![]() |
Breakout | Session Recording |
![]() Recording | 2022 |
Breakout VV&UQ – Uncertainty Quantification for Model-Based Engineering of DoD Systems |
Melissa Jablonski US Army |
Breakout | Materials | 2017 |
|
Contributed Development of a Locking Setback Mass for Cluster Munition Applications: A UQ Case Study |
Melissa Jablonski U.S. ARMY ARMAMENT RESEARCH, DEVELOPMENT & ENGINEERING CENTER |
Contributed | Materials | 2018 |
|
Breakout Metrics for Assessing Underwater Demonstrations for Detection and Classification of UXO |
Jacob Bartel Research Associate Institute for Defense Analyses ![]() |
Breakout |
![]() | 2021 |
|
Panel Army’s Open Experimentation Test Range for Internet of Battlefield Things: MSA-DPG |
Jade Freeman Research Scientist U.S. Army DEVCOM Army Research Laboratory ![]() |
Panel |
![]() | 2021 |
|
Analysis of Target Location Error using Stochastic Differential Equations |
James Brownlow mathematical statistician USAF ![]() |
Session Recording |
![]() Recording | 2022 |
|
Breakout Assessing Next-Gen Spacesuit Reliability: A Probabilistic Analysis Case Study at NASA |
James Warner Computational Scientist NASA Langley Research Center ![]() |
Breakout |
![]() | 2021 |
|
Breakout Machine Learning for Uncertainty Quantification: Trusting the Black Box |
James Warner Computational Scientist NASA Langley Research Center |
Breakout | Session Recording |
![]() Recording | 2022 |
Breakout Stochastic Modeling and Characterization of a Wearable-Sensor-Based Surveillance Network |
Jane E. Valentine Senior Biomedical Engineer Johns Hopkins University Applied Physics Laboratory |
Breakout | Session Recording |
![]() Recording | 2022 |
Breakout Challenges in Test and Evaluation of AI: DoD’s Project Maven |
Jane Pinelis | Breakout | 2019 |
||
Breakout Sample Size Calculations for Quiet Sonic Boom Community Surveys |
Jasme Lee | Breakout |
![]() | 2019 |
|
Breakout Topological Modeling of Human-Machine Teams |
Jay Wilkins and Caitlan Fealing Research Staff Member IDA |
Breakout | Session Recording |
![]() Recording | 2022 |
Breakout Valuing Human Systems Integration: A Test and Data Perspective |
Jeffrey Thomas | Breakout |
![]() | 2019 |
|
Panel The Keys to Successful Collaborations during Test and Evaluation: Panelist |
Willis Jensen HR Analyst W.L. Gore & Associates ![]() |
Panel | Session Recording |
Recording | 2021 |
Breakout Orbital Debris Effects Prediction Tool for Satellite Constellations |
Joel Williamsen Research Staff Member IDA ![]() |
Breakout | Session Recording |
Recording | 2022 |
Data Science & ML-Enabled Terminal Effects Optimization |
John Cilli Computer Scientist Picatinny Arsenal ![]() |
Session Recording |
![]() Recording | 2022 |
|
Breakout Advancements in Characterizing Warhead Fragmentation Events |
John Haman Research Staff Member Institute for Defense Analyses ![]() |
Breakout | Session Recording |
![]() Recording | 2021 |
Breakout What statisticians should do to improve M&S validation studies |
John Haman RSM IDA ![]() |
Breakout | Session Recording |
![]() Recording | 2022 |