Session Title | Speaker | Type | Recording | Materials | Year |
---|---|---|---|---|---|
Convolutional Neural Networks and Semantic Segmentation for Cloud and Ice Detection (Abstract)
Recent research shows the effectiveness of machine learning on image classification and segmentation. The use of artificial neural networks (ANNs) on image datasets such as the MNIST dataset of handwritten digits is highly effective. However, when presented with a more complex image, ANNs and other simple computer vision algorithms tend to fail. This research uses Convolutional Neural Networks (CNNs) to determine how we can differentiate between ice and clouds in the imagery of the Arctic. Instead of using ANNs, where we analyze the problem in one dimension, CNNs identify features using the spatial relationships between the pixels in an image. This technique allows us to extract spatial features, presenting us with higher accuracy. Using a CNN named the Cloud-Net Model, we analyze how a CNN performs when analyzing satellite images. First, we examine recent research on the Cloud-Net Model’s effectiveness on satellite imagery, specifically from Landsat data, with four channels: red, green, blue, and infrared. We extend and modify this model, allowing us to analyze data from the most common channels used by satellites: red, green, and blue. By training on different combinations of these three channels, we extend this analysis by testing on an entirely different data set: GOES imagery. This gives us an understanding of the impact of each individual channel in image classification. By selecting images that exist in the same geographic location and containing both ice and clouds, such as the Landsat, we test GOES analyzing the CNN’s generalizability. Finally, we present CNN’s ability to accurately identify the clouds and ice in the GOES data versus the Landsat data. |
Prarabdha Ojwaswee Yonzon Cadet United States Military Academy (West Point) ![]() (bio)
CDT Prarabdha “Osho” Yonzon is a first-generation Nepalese American raised in Brooklyn Park, Minnesota. He initially enlisted into the Minnesota National Guard in 2015 as an Aviation Operation Specialist, and he was later accepted into USMAPS in 2017. He is an Applied Statistics Data Science Major from the United States Military Academy. Osho is passionate about his research. He first started working with West Point Department of Physics to examine impacts on GPS solutions. Later, he published a few articles and presented them at the AWRA annual conference for modeling groundwater flow with the Math department. Currently, he is working with the West Point Department of Mathematics and Lockheed Martin to create machine learning algorithms to detect objects in images. He plans to attend graduate school for data science and serve as a cyber officer. |
Session Recording |
![]() | 2022 |
|
Trust Throughout the Artificial Intelligence Lifecycle (Abstract)
AI and machine learning have become widespread throughout the defense, government, and commercial sectors. This has led to increased attention on the topic of trust and the role it plays in successfully integrating AI into highconsequence environments where tolerance for risk is low. Driven by recent successes of AI algorithms in a range of applications, users and organizations rely on AI to provide new, faster, and more adaptive capabilities. However, along with those successes have come notable pitfalls, such as bias, vulnerability to adversarial attack, and inability to perform as expected in novel environments. Many types of AI are data-driven, meaning they operate on and learn their internal models directly from data. Therefore, tracking how data were used to build data properties (e.g., training, validation, and testing) is crucial not only to ensure a high-performing model, but also to understand if the AI should be trusted. MLOps, an offshoot of DevSecOps, is a set of best practices meant to standardize and streamline the end-to-end lifecycle of machine learning. In addition to supporting the software development and hardware requirements of AI-based systems, MLOps provides a scaffold by which the attributes of trust can be formally and methodically evaluated. Additionally, MLOps encourages reasoning about trust early and often in the development cycle. To this end, we present a framework that encourages the development of AI-based applications that can be trusted to operate as intended and function safely both with and without human interaction. This framework offers guidance for each phase of the AI lifecycle, utilizing MLOps, through a detailed discussion of pitfalls resulting from not considering trust, metrics for measuring attributes of trust, and mitigations strategies for when risk tolerance is low. |
Lauren H. Perry Senior Project Engineer The Aerospace Corporation ![]() (bio)
Lauren H Perry Sr Project Engineer, Space Applications Group Ms. Perry’s work with The Aerospace Corporation incorporates AI/ML technologies into traditional software development programs for the IC, DoD, and commercial customers. Previously, she was the analytical lead for a DoD project established to improve joint interoperability within the Integrated Air and Missile Defense (IAMD) Family of Systems and enhance air warfare capability, and a Reliability Engineer at Lockheed Martin Space Systems Company. She has a background in experimental design, applied statistics, and statistical engineering for the aerospace domain. Dr. Philip C Slingerland Sr Engineering Specialist, Machine Intelligence and Exploitation Department Dr. Slingerland’s work with The Aerospace Corporation focuses on machine learning and computer vision projects for a variety of IC, DoD, and commercial customers. Previously, he spent four years as a data scientist and software developer at Metron Scientific Solutions in support of many Naval Sea Systems Command (NAVSEA) studies. Dr. Slingerland has a background in sensor modeling and characterization, with a PhD in physics studying the performance of terahertz quantum cascade lasers (QCLs) for remote sensing applications. |
Session Recording |
![]() | 2022 |
|
Closing Remarks |
Alyson Wilson NCSU ![]() (bio)
Dr. Alyson Wilson is the Associate Vice Chancellor for National Security and Special Research Initiatives at North Carolina State University. She is also a professor in the Department of Statistics and Principal Investigator for the Laboratory for Analytic Sciences. Her areas of expertise include statistical reliability, Bayesian methods, and the application of statistics to problems in defense and national security. Dr. Wilson is a leader in developing transformative models for rapid innovation in defense and intelligence. Prior to joining NC State, Dr. Wilson was a jointly appointed research staff member at the IDA Science and Technology Policy Institute and Systems and Analyses Center (2011-2013); associate professor in the Department of Statistics at Iowa State University (2008-2011); Scientist 5 and technical lead for Department of Defense Programs in the Statistical Sciences Group at Los Alamos National Laboratory (1999-2008); and senior statistician and operations research analyst with Cowboy Programming Resources (1995-1999). She is currently serving on the National Academy of Sciences Committee on Applied and Theoretical Statistics and on the Board of Trustees for the National Institute of Statistical Sciences. Dr. Wilson is a Fellow of the American Statistical Association, the American Association for the Advancement of Science, and an elected member of the International Statistics Institute. |
2022 |
|||
Quantifying the Impact of Staged Rollout Policies on Software Process and Product Metrics (Abstract)
Software processes define specific sequences of activities performed to effectively produce software, whereas tools provide concrete computational artifacts by which these processes are carried out. Tool independent modeling of processes and related practices enable quantitative assessment of software and competing approaches. This talk presents a framework to assess an approach employed in modern software development known as staged rollout, which releases new or updated software features to a fraction of the user base in order to accelerate defect discovery without imposing the possibility of failure on all users. The framework quantifies process metrics such as delivery time and product metrics, including reliability, availability, security, and safety, enabling tradeoff analysis to objectively assess the quality of software produced by vendors, establish baselines, and guide process and product improvement. Failure data collected during software testing is employed to emulate the approach as if the project were ongoing. The underlying problem is to identify a policy that decides when to perform various stages of rollout based on the software’s failure intensity. The illustrations examine how alternative policies impose tradeoffs between two or more of the process and product metrics. |
Lance Fiondella Associate Professor University of Massachusetts Dartmouth ![]() (bio)
Lance Fiondella is an associate professor of Electrical and Computer Engineering at the University of Massachusetts Dartmouth and the Director of the UMassD Cybersecurity Center, a NSA/DHS designated Center of Academic Excellence in Cyber Research. |
Session Recording |
![]() | 2022 |
|
Data Science & ML-Enabled Terminal Effects Optimization (Abstract)
Warhead design and performance optimization against a range of targets is a foundational aspect of the Department of the Army’s mission on behalf of the warfighter. The existing procedures utilized to perform this basic design task do not fully leverage the exponential growth in data science, machine learning, distributed computing, and computational optimization. Although sound in practice and methodology, existing implementations are laborious and computationally expensive, thus limiting the ability to fully explore the trade space of all potentially viable solutions. An additional complicating factor is the fast paced nature of many Research and Development programs which require equally fast paced conceptualization and assessment of warhead designs. By utilizing methods to take advantage of data analytics, the workflow to develop and assess modern warheads will enable earlier insights, discovery through advanced visualization, and optimal integration of multiple engineering domains. Additionally, a framework built on machine learning would allow for the exploitation of past studies and designs to better inform future developments. Combining these approaches will allow for rapid conceptualization and assessment of new and novel warhead designs. US overmatch capability is quickly eroding across many tactical and operational weapon platforms. Traditional incremental improvement approaches are no longer generating appreciable performance improvements to warrant investment. Novel next generation techniques are required to find efficiencies in designs and leap forward technologies to maintain US superiority. The proposed approach seeks to shift existing design mentality to meet this challenge. |
John Cilli Computer Scientist Picatinny Arsenal ![]() (bio)
My name is John Cilli, I am a recent graduate of East Stroudsburg University with a bachelor’s in Computer Science. I have been working at Picatinny within the Systems Analysis Division as a computer scientist for little over a year now. |
Session Recording |
![]() | 2022 |
|
Moderator (Abstract)
For organizations to make data-driven decisions, they must be able to understand and organize their mission critical data. Recently, the DoD, NASA and other federal agencies have declared their intention to become “data-centric” organizations, but transitioning from an existing mode of operation and architecture can be challenging. Moreover, the DoD is pushing for artificial intelligence enabled systems (AIES) and wide scale digital transformation. These concepts in the abstract seem straightforward, but because they can only evolve when people, processes, and technology change together, they have proven challenging in execution. Since the structure and quality of an organization’s data limits what an organization can do with that data it is imperative to get data processes right before embarking on other initiatives that depend on quality data. Despite the importance of data quality, many organizations treat data architecture as an emergent phenomenon and not something to be planned or thought through holistically. In this discussion, panelists will explore what it means to be data-centric, what a data-centric architecture is, how it is different from the other data architectures, why an organization might prefer a data-centric approach, and the challenges associated with becoming data-centric. |
Matthew Avery Assistant Director, Operational Evaluation IDA ![]() (bio)
Matthew Avery is an Assistant Director in the Operational Evaluation Division (OED) at the Institute for Defense Analyses (IDA)and part of OED’s Sustainment group. He represents OED on IDA’s Data Governance Council and acts as the Deputy to IDA’s Director of Data Strategy and Chief Data Officer, helping craft data-related strategy and policy. Matthew leads IDA’s sustainment modeling efforts for the V-22 fleet, developing end-to-end multi-echelon models to evaluate options for improving mission-capable rates for the CV-22 and MV-22 fleets. Prior to this, Matthew was on the Test Science team, where he helped develop analytical methods and tools for operational test and evaluation. As the Test Science Data Management lead, he was responsible for delivering an annual summary of major activity undertaken by the Office of the Director, Operational Test and Evaluation. Additionally, Matthew wrote and implemented OED policy on data management and reproducible research. In addition to working with the Test Science team, Matthew also led operational test and evaluation efforts of Army and Marine Corps unmanned aircraft systems. In 2018-19 Matthew served as an embedded analyst in the Pentagon’s Office of Cost Assessment and Program Evaluation, where he built state-space models in support of the Space Control Strategic Portfolio Review. Matthew earned his PhD in Statistics from North Carolina State University in 2012, his MS in Statistics from North Carolina State in 2009, and a BA from New College of Florida in 2006. He is a member of the American Statistical Association. |
2022 |
|||
Panelist 4 |
Calvin Robinson NASA ![]() (bio)
Calvin Robinson is a Data Architect within the Information and Applications Division at NASA Glenn Research Center. He has over 10 years of experience supporting data analysis and simulation development for research, and currently supports several key data management efforts to make data more discoverable and aligned with FAIR principles. Calvin oversees the Center’s Information Management Program and supports individuals leading strategic AIML efforts within the Agency. Calvin holds a BS in Computer Science and Engineering from the University of Toledo. |
Session Recording | 2022 |
||
Panelist 3 |
Jane Pinelis Joint Artificial Intelligence Center ![]() (bio)
Dr. Jane Pinelis is the Chief of AI Assurance at the Department of Defense Joint Artificial Intelligence Center (JAIC). She leads a diverse team of testers and analysts in rigorous test and evaluation (T&E) as well as Responsible AI (RAI) implementation for JAIC capabilities, as well as development of AI Assurance products and standards that will support testing of AI-enabled systems across the DoD. Prior to joining the JAIC, Dr. Pinelis served as the Director of Test and Evaluation for USDI’s Algorithmic Warfare Cross-Functional Team, better known as Project Maven. She directed the developmental testing for the AI models, including computer vision, machine translation, facial recognition and natural language processing. Her team developed metrics at various levels of testing for AI capabilities and provided leadership empirically-based recommendations for model fielding. Additionally, she oversaw operational and human-machine teaming testing, and conducted research and outreach to establish standards in T&E of systems using artificial intelligence. Dr. Pinelis has spent over 10 years working predominantly in the area of defense and national security. She has largely focused on operational test and evaluation, both in support of the service operational testing commands and also at the OSD level. In her previous job as the Test Science Lead at the Institute of Defense Analyses, she managed an interdisciplinary team of scientists supporting the Director and the Chief Scientist of the Department of Operational Test and Evaluation on integration of statistical test design and analysis and data-driven assessments into test and evaluation practice. Before, that, in her assignment at the Marine Corps Operational Test and Evaluation Activity, Dr. Pinelis led the design and analysis of the widely publicized study on the effects of integrating women into combat roles in the Marine Corps. Based on this experience, she co-authored a book, titled “The Experiment of a Lifetime: Doing Science in the Wild for the United States Marine Corps.” In addition to T&E, Dr. Pinelis has several years of experience leading analyses for the DoD in the areas of wargaming, precision medicine, warfighter mental health, nuclear non-proliferation, and military recruiting and manpower planning. Her areas of statistical expertise include design and analysis of experiments, quasi-experiments, and observational studies, causal inference, and propensity score methods. Dr. Pinelis holds a BS in Statistics, Economics, and Mathematics, an MA in Statistics, and a PhD in Statistics, all from the University of Michigan, Ann Arbor. |
Session Recording | 2022 |
||
Building Bridges: a Case Study of Assisting a Program from the Outside (Abstract)
STAT practitioners often find ourselves outsiders to the programs we assist. This session presents a case study that demonstrates some of the obstacles in communication of capabilities, purpose, and expectations that may arise due to approaching the project externally. Incremental value may open the door to greater collaboration in the future, and this presentation discusses potential solutions to provide greater benefit to testing programs in the face of obstacles that arise due to coming from outside the program team. DISTRIBUTION STATEMENT A. Approved for public release; distribution is unlimited. CLEARED on 5 Jan 2022. Case Number: 88ABW-2022-0002 |
Anthony Sgambellone Huntington Ingalls Industries ![]() (bio)
Dr. Tony Sgambellone is a STAT Expert (Huntington Ingalls Industries contractor) at the Scientific Test and Analysis Techniques (STAT) Center of Excellence (COE) at the Air Force Institute of Technology (AFIT). The STAT COE provides independent STAT consultation to designated acquisition programs and special projects to improve Test & Evaluation (T&E) rigor, effectiveness, and efficiency,. Dr. Sgambellone holds a Ph.D. in Statistics, a graduate minor in College and University Teaching, and has a decade of experience spanning the fields of finance, software, and test and development. His current interests include artificial neural networks and the application of machine learning. |
Session Recording |
![]() | 2022 |
|
Utilizing Machine Learning Models to Predict Success in Special Operations Assessment (Abstract)
The 75th Ranger Regiment is an elite Army Unit responsible for some of the most physically and mentally challenging missions. Entry to the unit is based on an assessment process called Ranger Regiment Assessment and Selection (RASP), which consists of a variety of tests and challenges of strength, intellect, and grit. This study explores the psychological and physical profiles of candidates who attempt to pass RASP. Using a Random Forest Artificial Intelligence model, and a penalized logistic regression model, we identify initial entry characteristics that are predictive of success in RASP. We focus on the differences between racial sub-groups and military occupational specialties (MOS) sub-groups to provide information for recruiters to identify underrepresented groups who are likely to succeed into the selection process. |
Anna Vinnedge Student United States Military Academy ![]() (bio)
Anna Vinnedge is a fourth year cadet at the United States Military Academy. She is a mathematics major from originally from Seattle, WA and the captain of the Varsity Women’s Crew team. After graduation, she will serve as a cyber officer in the Army. She has conducted previous research in quantum and DNA based cryptography and coding theory, and is currently focusing her time on machine learning research for statistical analysis. |
Session Recording |
![]() | 2022 |
|
Predicting Trust in Automated Systems: Validation of the Trust of Automated Systems Test (Abstract)
The number of people using autonomous systems for everyday tasks has increased steadily since the 1960s and has dramatically increased with the invention of smart devices that can be controlled via smartphone. Within the defense community, automated systems are currently used to perform search and rescue missions and to assume control of aircraft to avoid ground collision. Until recently, researchers have only been able to gain insights on trust levels by observing a human’s reliance on the system, so it was apparent that researchers needed a validated method of quantifying how much an individual trusts the automated system they are using. IDA researchers developed the Trust of Automated Systems Test (TOAST scale) to serve as a validated scale capable of measuring how much an individual trusts a system. This presentation will outline the nine item TOAST scale’s understanding and performance elements, and how it can effectively be used in a defense setting. We believe that this scale should be used to evaluate the trust level of any human using any system, including predicting when operators will misuse or disuse complex, automated and autonomous systems. |
Caitlan Fealing Data Science Fellow IDA ![]() (bio)
Caitlan Fealing is a Data Science Fellow within the Test Science group of OED. She has a Bachelor of Arts degree in Mathematics, Economics, and Psychology from Williams College. Caitlan uses her background and focus on data science to create data visualizations, support OED’s program management databases, and contribute to the development of the many resources available on IDA’s Test Science website. |
Session Recording |
![]() | 2022 |
|
Sparse Models for Detecting Malicious Behavior in OpTC (Abstract)
Host-based sensors are standard tools for generating event data to detect malicious activity on a network. There is often interest in detecting activity using as few event classes as possible in order to minimize host processing slowdowns. Using DARPA’s Operationally Transparent Cyber (OpTC) Data Release, we consider the problem of detecting malicious activity using event counts aggregated over five-minute windows. Event counts are categorized by eleven features according to MITRE CAR data model objects. In the supervised setting, we use regression trees with all features to show that malicious activity can be detected at above a 90% true positive rate with a negligible false positive rate. Using forward and exhaustive search techniques, we show the same performance can be obtained using a sparse model with only three features. In the unsupervised setting, we show that the isolation forest algorithm is somewhat successful at detecting malicious activity, and that a sparse three-feature model performs comparably. Finally, we consider various search criteria for identifying sparse models and demonstrate that the RMSE criteria is generally optimal. |
Andrew Mastin Operations Research Scientist Lawrence Livermore National Laboratory ![]() (bio)
Andrew Mastin is an Operations Research Scientist at Lawrence Livermore National Laboratory. He holds a Ph.D. in Electrical Engineering and Computer Science from the Massachusetts Institute of Technology. His current research interests include cybersecurity, network interdiction, dynamic optimization, and game theory. |
![]() | 2022 |
||
Profile Monitoring via Eigenvector Perturbation (Abstract)
Control charts are often used to monitor the quality characteristics of a process over time to ensure undesirable behavior is quickly detected. The escalating complexity of processes we wish to monitor spurs the need for more flexible control charts such as those used in profile monitoring. Additionally, designing a control chart that has an acceptable false alarm rate for a practitioner is a common challenge. Alarm fatigue can occur if the sampling rate is high (say, once a millisecond) and the control chart is calibrated to an average in-control run length (ARL0) of 200 or 370 which is often done in the literature. As alarm fatigue may not just be annoyance but result in detrimental effects to the quality of the product, control chart designers should seek to minimize the false alarm rate. Unfortunately, reducing the false alarm rate typically comes at the cost of detection delay or average out-of-control run length (ARL1). Motivated by recent work on eigenvector perturbation theory, we develop a computationally fast control chart called the Eigenvector Perturbation Control Chart for nonparametric profile monitoring. The control chart monitors the l_2 perturbation of the leading eigenvector of a correlation matrix and requires only a sample of known in-control profiles to determine control limits. Through a simulation study we demonstrate that it is able to outperform its competition by achieving an ARL1 close to or equal to 1 even when the control limits result in a large ARL0 on the order of 10^6. Additionally, non-zero false alarm rates with a change point after 10^4 in-control observations were only observed in scenarios that are either pathological or truly difficult for a correlation based monitoring scheme. |
Takayuki Iguchi PhD Student Florida State University (bio)
Takayuki Iguchi is a Captain in the US Air Force and is currently a PhD student under the direction of Dr. Eric Chicken at Florida State University. |
Session Recording |
![]() | 2022 |
|
Opening Remarks |
Norton Schwartz President IDA ![]() (bio)
Norton A. Schwartz serves as President of the Institute for Defense Analyses (IDA), a nonprofit corporation operating in the public interest. IDA manages three Federally Funded Research and Development Centers that answer the most challenging U.S. security and science policy questions with objective analysis leveraging extraordinary scientific, technical, and analytic expertise. At IDA, General Schwartz (U.S. Air Force, retired) directs the activities of more than 1,000 scientists and technologists employed by IDA. General Schwartz has a long and prestigious career of service and leadership that spans over 5 decades. He was most recently President and CEO of Business Executives for National Security (BENS). During his 6-year tenure at BENS, he was also a member of IDA’s Board of Trustees. Prior to retiring from the U.S. Air Force, General Schwartz served as the 19th Chief of Staff of the U.S. Air Force from 2008 to 2012. He previously held senior joint positions as Director of the Joint Staff and as the Commander of the U.S. Transportation Command. He began his service as a pilot with the airlift evacuation out of Vietnam in 1975. General Schwartz is a U.S. Air Force Academy graduate and holds a master’s degree in business administration from Central Michigan University. He is also an alumnus of the Armed Forces Staff College and the National War College. He is a member of the Council on Foreign Relations and a 1994 Fellow of Massachusetts Institute of Technology’s Seminar XXI. General Schwartz has been married to Suzie since 1981. |
Session Recording | 2022 |
||
Exploring the behavior of Bayesian adaptive design of experiments (Abstract)
Physical experiments in the national security arena, including nuclear deterrence, are often expensive and time-consuming resulting in small sample sizes which make it difficult to achieve desired statistical properties. Bayesian adaptive design of experiments (BADE) is a sequential design of experiment approach which updates the test design in real time, in order to optimally collect data. BADE recommends ending experiments early by either concluding that the experiment would have ended in efficacy or futility, had the testing completely finished, with sufficiently high probability. This is done by using data already collected and marginalizing over the remaining uncollected data and updating the Bayesian posterior distribution in near real-time. BADE has seen successes in clinical trials, resulting in quicker and more effective assessments of drug trials while also reducing ethical concerns. BADE has typically only been used in futility studies rather than efficacy studies for clinical trials, although there hasn’t been much debate for this current paradigm. BADE has been proposed for testing in the national security space for similar reasons of quicker and cheaper test series. Given the high-consequence nature of the tests performed in the national security space, a strong understanding of new methods is required before being deployed. The main contribution of this research was to reproduce results seen in previous studies, for different aspects of model performance. A large simulation inspired by a real testing problem at Sandia National Laboratories was performed to understand the behavior of BADE under various scenarios, including shifts to mean, standard deviation, and distributional family, all in addition to the presence of outliers. The results help explain the behavior of BADE under various assumption violations. Using the results of this simulation, combined with previous work related to BADE in this field, it is argued this approach could be used as part of an “evidence package” for deciding to stop testing early due to futility, or with stronger evidence, efficacy. The combination of expert knowledge with statistical quantification provides the stronger evidence necessary for a method in its infancy in a high-consequence, new application area such as national security. Sandia National Laboratories is a multimission laboratory managed and operated by National Technology & Engineering Solutions of Sandia, LLC, a wholly owned subsidiary of Honeywell International Inc., for the U.S. Department of Energy’s National Nuclear Security Administration under contract DE-NA0003525. |
Daniel Ries Sandia National Laboratories ![]() (bio)
Daniel Ries is a Senior Member of the Technical Staff at Sandia National Laboratories in the Statistics and Data Analytics Department. As an applied research statistician, Daniel collaborates with scientists and engineers in fields including nuclear deterrence, nuclear forensics, nuclear non-proliferation, global security, and climate science. His statistical work spans the topics of experimental design, inverse modeling, uncertainty quantification for machine learning and deep learning, spatio-temporal data analysis, and Bayesian methodology. Daniel completed his PhD in statistics at Iowa State University. |
Session Recording |
![]() | 2022 |
|
Analysis of Target Location Error using Stochastic Differential Equations (Abstract)
This paper presents an analysis of target location error (TLE) based on the Cox Ingersoll Ross (CIR) model. In brief, this model characterizes TLE as a function of range based the stochastic differential equation model dX(r) = a(b-X(r))dr + sigma *sqrt(X(r)) dW(r) where X(t) is TLE at range r, b is the long-term mean (terminal) of the TLE, a is the rate of reversion of X(r) to b, sigma is the process volatility, and W(t) is the standard Weiner process. Multiple flight test runs under the same conditions exhibit different realizations of the TLE process. This approach to TLE analysis models each flight test run as a realization the CIR process. Fitting a CIR model to multiple data runs then provides a characterization of the TLE system under test. This paper presents an example use of the CIR model. Maximum likelihood estimates of the parameters of the CIR model are found from a collection of TLE data runs. The resulting CIR model is then used to characterize overall system TLE performance as a function of range to the target as well as the asymptotic estimate of long-term TLE. |
James Brownlow mathematical statistician USAF ![]() (bio)
Dr. James Brownlow is a tech expert in statistics with the USAF, Edwards AFB, CA. His PhD is in time series from UC Riverside. Dr. Brownlow has developed test and evaluation procedures using Bayesian techniques, and developed Python code to adapt parametric survival models to the analysis of target location error. He is a coauthor of a paper that used stochastic differential equations to characterize Kalman-filtered estimates of target track state vectors. |
Session Recording |
![]() | 2022 |
|
Risk Comparison and Planning for Bayesian Assurance Tests (Abstract)
Designing a Bayesian assurance test plan requires choosing a test plan that guarantees a product of interest is good enough to satisfy consumer’s criteria but not ‘so good’ that it causes producer’s concern if they fail the test. Bayesian assurance tests are especially useful because they can incorporate previous product information in the test planning and explicitly control levels of risk for the consumer and producer. We demonstrate an algorithm for efficiently computing a test plan given desired levels of risks in binomial and exponential testing. Numerical comparisons with the Operational Characteristic (OC) curve, Probability Ratio Sequential Test (PRST), and a simulation-based Bayesian sample size determination approach are also considered. |
Hyoshin Kim North Carolina State University ![]() (bio)
Hyoshin Kim received her B.Ec. in Statistics from Sungkyunkwan University, South Korea, in 2017, and her M.S. in Statistics from Seoul National University, South Korea, in 2019. She is currently a third year Ph.D. student at the department of Statistics at North Carolina State University. Her research interests are Bayesian assurance testing and Bayesian clustering algorithms for high dimensional correlated outcomes. |
Session Recording |
![]() | 2022 |
|
Everyday Reproducibility (Abstract)
Modern data analysis is typically quite computational. Correspondingly, sharing scientific and statistical work now often means sharing code and data in addition writing papers and giving talks. This type of code sharing faces several challenges. For example, it is often difficult to take code from one computer and run it on another due to software configuration, version, and dependency issues. Even if the code runs, writing code that is easy to understand or interact with can be difficult. This makes it difficult to assess third-party code and its findings, for example, in a review process. In this talk we describe a combination of two computing technologies that help make analyses shareable, interactive, and completely reproducible. These technologies are (1) analysis containerization, which leverages virtualization to fully encapsulate analysis, data, code and dependencies into an interactive and shareable format, and (2) code notebooks, a literate programming format for interacting with analyses. This talks reviews both the problems at the high-level and also provides concrete solutions to the challenges faced. In addition to discussing reproducibility and data/code sharing generally, we will touch upon several such issues that arise specifically in the defense and aerospace communities. |
Gregory J. Hunt Assistant Professor William & Mary ![]() (bio)
Greg is an Assistant Professor of Mathematics at the College of William & Mary. He is an interdisciplinary researcher that builds scientific tools and is trained as a statistician, mathematician and computer scientist. Currently he works on a diverse set of problems in high-throughput micro-biology, research reproducibility, hypersonics, and spectroscopy. |
Session Recording |
![]() | 2022 |
|
Cloud Computing for Computational Fluid Dynamics (CFD) in T&E (Abstract)
In this talk we’ll focus on exploring the motivation for using cloud computing for Computational Fluid Dynamics (CFD) for Federal Government Test & Evaluation. Using examples from automotive, aerospace and manufacturing we’ll look at benchmarks for a number of CFD codes using CPUs (x86 & Arm) and GPUs and we’ll look at how the development of high-fidelity CFD e.g. WMLES, HRLES, is accelerating the need for access to large scale HPC. The onset of COVID-19 has also meant a large increase in the need for remote visualization with greater numbers of researchers and engineering needing to work from home. This has also accelerated the adoption of the same approaches needed towards the pre- and post-processing of peta/exa-scale CFD simulation and we’ll look at how these are more easily accessed via a cloud infrastructure. Finally, we’ll explore perspectives on integrating ML/AI into CFD workflows using data lakes from a range of sources and where the next decade may take us. |
Neil Ashton WW Principal CFD Specialist Solution Architect, HPC Amazon Web Services ![]() (bio)
Neil Ashton is the WW subject matter expert for CFD within AWS. He works with customers in enterprise, startup and public-sector across the globe to help them to run their CFD (and often also FEA) workloads on AWS. In addition he acts as a key advisor to the global product teams to deliver better hardware and software for CFD and broader CAE users. He is also still very active in academic research around deep-learning and machine learning, future HPC approaches and novel CFD approaches (GPU’s, numerical methods, turbulence modelling) |
Session Recording |
![]() | 2022 |
|
A Framework for Using Priors in a Continuum of Testing (Abstract)
A strength of the Bayesian paradigm is that it allows for the explicit use of all available information—to include subject matter expert (SME) opinion and previous (possibly dissimilar) data. While frequentists are constrained to only including data in an analysis (that is to say, only including information that can be observed), Bayesians can easily consider both data and SME opinion, or any other related information that could be constructed. This can be accomplished through the development and use of priors. When prior development is done well, a Bayesian analysis will not only lead to more direct probabilistic statements about system performance, but can result in smaller standard errors around fitted values when compared to a frequentist approach. Furthermore, by quantifying the uncertainty surrounding a model parameter, through the construct of a prior, Bayesians are able to capture the uncertainty across a test space of consideration. This presentation develops a framework for thinking about how different priors can be used throughout the continuum of testing. In addition to types of priors, how priors can change or evolve across the continuum of testing—especially when a system changes (e.g., is modified or adjusted) during phases of testing—will be addressed. Priors that strive to provide no information (reference priors) will be discussed, and will build up to priors that contain available information (informative priors). Informative priors—both those based on institutional knowledge or summaries from databases, as well as those developed based on previous testing data—will be discussed, with a focus on how to consider previous data that is dissimilar in some way, relative to the current test event. What priors might be more common in various phases of testing, types of information that can be used in priors, and how priors evolve as information accumulates will all be discussed. |
Victoria Sieck Deputy Director / Assistant Professor of Statistics Scientific Test & Analysis Techniques Center of Excellence (STAT COE) / Air Force Institute of Technology (AFIT) ![]() (bio)
Dr. Victoria R. C. Sieck is the Deputy Director of the Scientific Test & Analysis (STAT) Center of Excellence (COE), where she works with major acquisition programs within the Department of Defense (DoD) to apply rigor and efficiency to current and emerging test and evaluation methodologies through the application of the STAT process. Additionally, she is an Assistant Professor of Statistics at the Air Force Institute of Technology (AFIT), where her research interests include design of experiments, and developing innovate Bayesian approaches to DoD testing. As an Operations Research Analyst in the US Air Force (USAF), her experiences in the USAF testing community include being a weapons and tactics analyst and an operational test analyst. Dr. Sieck has a M.S. in Statistics from Texas A&M University, and a Ph.D. in Statistics from the University of New Mexico. Her Ph.D. research was on improving operational testing through the use of Bayesian adaptive testing methods. |
Session Recording |
![]() | 2022 |
|
Analysis Apps for the Operational Tester (Abstract)
In the acquisition and testing world, data analysts repeatedly encounter certain categories of data, such as time or distance until an event (e.g., failure, alert, detection), binary outcomes (e.g., success/failure, hit/miss), and survey responses. Analysts need tools that enable them to produce quality and timely analyses of the data they acquire during testing. This poster presents four web-based apps that can analyze these types of data. The apps are designed to assist analysts and researchers with simple repeatable analysis tasks, such as building summary tables and plots for reports or briefings. Using software tools like these apps can increase reproducibility of results, timeliness of analysis and reporting, attractiveness and standardization of aesthetics in figures, and accuracy of results. The first app models reliability of a system or component by fitting parametric statistical distributions to time-to-failure data. The second app fits a logistic regression model to binary data with one or two independent continuous variables as predictors. The third calculates summary statistics and produces plots of groups of Likert-scale survey question responses. The fourth calculates the system usability scale (SUS) scores for SUS survey responses and enables the app user to plot scores versus an independent variable. These apps are available for public use on the Test Science Interactive Tools webpage https://new.testscience.org/interactive-tools/. |
William Raymond Whitledge Research Staff Member IDA ![]() (bio)
Bill Whitledge is a research staff member at the Institute for Defense Analyses (IDA), where he has worked since 2010. He is currently the project leader for the operational test agent task for the Cybersecurity and Infrastructure Security Agency (CISA) under the Department of Homeland Security (DHS). This project aims to provide CISA with rigorous analysis of the operational effectiveness, usability, and reliability of the National Cybersecurity Protection System (NCPS) family of systems. In addition to leading the NCPS testing project, Bill works on the IDA data management committee updating IDA’s data management practices and procedures. He also co-leads the IDA Connects speaker series, an internal IDA series of lunchtime talks intended to help staff stay informed on current events, meet colleagues, and learn about research across IDA. Bill is passionate about helping people visualize and present information in more elegant and succinct ways. One of his main interests is writing tools in R and other programming languages to automate data collection, analysis, and visualization. He has developed four web applications hosted on the IDA Test Science website enabling easier analysis of system reliability, binary outcomes, system usability, and general survey responses. Bill is also an avid cyclist and golfer, and he is one of the coordinators of the IDA golf league. Bill received his Bachelor of Arts in physics with an economics minor from Colby College in 2008. He received his Master of Science in electrical engineering with a focus in optics and optical communication at Stanford University in 2010. |
Session Recording |
![]() | 2022 |
Session Title | Speaker | Type | Recording | Materials | Year |
---|---|---|---|---|---|
Convolutional Neural Networks and Semantic Segmentation for Cloud and Ice Detection |
Prarabdha Ojwaswee Yonzon Cadet United States Military Academy (West Point) ![]() |
Session Recording |
![]() | 2022 |
|
Trust Throughout the Artificial Intelligence Lifecycle |
Lauren H. Perry Senior Project Engineer The Aerospace Corporation ![]() |
Session Recording |
![]() | 2022 |
|
Closing Remarks |
Alyson Wilson NCSU ![]() |
2022 |
|||
Quantifying the Impact of Staged Rollout Policies on Software Process and Product Metrics |
Lance Fiondella Associate Professor University of Massachusetts Dartmouth ![]() |
Session Recording |
![]() | 2022 |
|
Data Science & ML-Enabled Terminal Effects Optimization |
John Cilli Computer Scientist Picatinny Arsenal ![]() |
Session Recording |
![]() | 2022 |
|
Moderator |
Matthew Avery Assistant Director, Operational Evaluation IDA ![]() |
2022 |
|||
Panelist 4 |
Calvin Robinson NASA ![]() |
Session Recording | 2022 |
||
Panelist 3 |
Jane Pinelis Joint Artificial Intelligence Center ![]() |
Session Recording | 2022 |
||
Building Bridges: a Case Study of Assisting a Program from the Outside |
Anthony Sgambellone Huntington Ingalls Industries ![]() |
Session Recording |
![]() | 2022 |
|
Utilizing Machine Learning Models to Predict Success in Special Operations Assessment |
Anna Vinnedge Student United States Military Academy ![]() |
Session Recording |
![]() | 2022 |
|
Predicting Trust in Automated Systems: Validation of the Trust of Automated Systems Test |
Caitlan Fealing Data Science Fellow IDA ![]() |
Session Recording |
![]() | 2022 |
|
Sparse Models for Detecting Malicious Behavior in OpTC |
Andrew Mastin Operations Research Scientist Lawrence Livermore National Laboratory ![]() |
![]() | 2022 |
||
Profile Monitoring via Eigenvector Perturbation |
Takayuki Iguchi PhD Student Florida State University |
Session Recording |
![]() | 2022 |
|
Opening Remarks |
Norton Schwartz President IDA ![]() |
Session Recording | 2022 |
||
Exploring the behavior of Bayesian adaptive design of experiments |
Daniel Ries Sandia National Laboratories ![]() |
Session Recording |
![]() | 2022 |
|
Analysis of Target Location Error using Stochastic Differential Equations |
James Brownlow mathematical statistician USAF ![]() |
Session Recording |
![]() | 2022 |
|
Risk Comparison and Planning for Bayesian Assurance Tests |
Hyoshin Kim North Carolina State University ![]() |
Session Recording |
![]() | 2022 |
|
Everyday Reproducibility |
Gregory J. Hunt Assistant Professor William & Mary ![]() |
Session Recording |
![]() | 2022 |
|
Cloud Computing for Computational Fluid Dynamics (CFD) in T&E |
Neil Ashton WW Principal CFD Specialist Solution Architect, HPC Amazon Web Services ![]() |
Session Recording |
![]() | 2022 |
|
A Framework for Using Priors in a Continuum of Testing |
Victoria Sieck Deputy Director / Assistant Professor of Statistics Scientific Test & Analysis Techniques Center of Excellence (STAT COE) / Air Force Institute of Technology (AFIT) ![]() |
Session Recording |
![]() | 2022 |
|
Analysis Apps for the Operational Tester |
William Raymond Whitledge Research Staff Member IDA ![]() |
Session Recording |
![]() | 2022 |