- API
NEW HORIZONS SDC PLUTO CRUISE RAW V2.0
data.nasa.gov | Last Updated 2023-01-26T20:54:05.000ZThis data set contains Raw data taken by the New Horizons Student Dust Counter instrument during the pluto cruise mission phase. This is VERSION 2.0 of this data set. SDC collected science data intermittently during the hibernation years following the Jupiter encounter, designated as the PLUTOCRUISE phase. There were also Annual Checkouts (ACOs), STIM calibrations, Noise calibrations, and an anomaly in November, 2007. SDC's main science data collection periods were during hibernation. During ACOs, science data are taken intermittently but the user must be careful in analyzing these data since there is usually more activity on the spacecraft during hibernation. STIM and Noise refer to scheduled calibrations and are done with a regular cadence of one per year after the Jupiter encounter; they occurred sporadically in the early years of the mission. Note that some SDC data files have the same stop and start time and a zero exposure time. The reason for this is that the start and stop time for SDC data files are the event times for the first and last events in the files, so for files that contain a single event, these two values are the same. The changes in Version 2.0 were re-running of the ancillary data in the data product, updated geometry from newer SPICE kernels, minor editing of the documentation, catalogs, etc., and resolution of liens from the December, 2014 review, plus those from the May, 2016 review of the Pluto Encounter data sets. New observations added with this version (V2.0) include ongoing cruise observations from August, 2014 through January, 2015.
- API
Classification of Aeronautics System Health and Safety Documents
data.nasa.gov | Last Updated 2020-01-29T01:57:57.000ZMost complex aerospace systems have many text reports on safety, maintenance, and associated issues. The Aviation Safety Reporting System (ASRS) spans several decades and contains over 700 000 reports. The Aviation Safety Action Plan (ASAP) contains over 12 000 reports from various airlines. Problem categorizations have been developed for both ASRS and ASAP to enable identification of system problems. However, repository volume and complexity make human analysis difficult. Multiple experts are needed, and they often disagree on classifications. Even the same person has classified the same document differently at different times due to evolving experiences. Consistent classification is necessary to support tracking trends in problem categories over time. A decision support system that performs consistent document classification quickly and over large repositories would be useful. We discuss the results of two algorithms we have developed to classify ASRS and ASAP documents. The first is Mariana---a support vector machine (SVM) with simulated annealing, which is used to optimize hyperparameters for the model. The second method is classification built on top of nonnegative matrix factorization (NMF), which attempts to find a model that represents document features that add up in various combinations to form documents. We tested both methods on ASRS and ASAP documents with the latter categorized two different ways. We illustrate the potential of NMF to provide document features that are interpretable and indicative of topics. We also briefly discuss the tool that we have incorporated Mariana into in order to allow human experts to provide feedback on the document categorizations.
- API
Metrics for Offline Evaluation of Prognostic Performance
data.nasa.gov | Last Updated 2020-01-29T01:57:42.000ZPrognostic performance evaluation has gained significant attention in the past few years.*Currently, prognostics concepts lack standard definitions and suffer from ambiguous and inconsistent interpretations. This lack of standards is in part due to the varied end- user requirements for different applications, time scales, available information, domain dynamics, etc. to name a few. The research community has used a variety of metrics largely based on convenience and their respective requirements. Very little attention has been focused on establishing a standardized approach to compare different efforts. This paper presents several new evaluation metrics tailored for prognostics that were recently introduced and were shown to effectively evaluate various algorithms as compared to other conventional metrics. Specifically, this paper presents a detailed discussion on how these metrics should be interpreted and used. These metrics have the capability of incorporating probabilistic uncertainty estimates from prognostic algorithms. In addition to quantitative assessment they also offer a comprehensive visual perspective that can be used in designing the prognostic system. Several methods are suggested to customize these metrics for different applications. Guidelines are provided to help choose one method over another based on distribution characteristics. Various issues faced by prognostics and its performance evaluation are discussed followed by a formal notational framework to help standardize subsequent developments.
- API
Metrics for Evaluating Performance of Prognostic Techniques
data.nasa.gov | Last Updated 2020-01-29T03:23:28.000ZPrognostics is an emerging concept in condition basedmaintenance(CBM)ofcriticalsystems.Alongwith developing the fundamentals of being able to confidently predict Remaining Useful Life (RUL), the technology calls for fielded applications as it inches towards maturation. This requires a stringent performance evaluation so that the significance of the concept can be fully exploited. Currently, prognostics concepts lack standard definitions and suffer from ambiguous and inconsistent interpretations. This lack of standards is in part due to the varied end-user requirements for different applications, time scales, available information, domain dynamics, etc. to name a few issues. Instead, the research community has used a variety of metrics based largely on convenience with respect to their respective requirements. Very little attention has been focused on establishing a common ground to compare different efforts. This paper surveys the metrics that are already used for prognostics in a variety of domains including medicine, nuclear, automotive, aerospace, and electronics. It also considers other domains that involve prediction-related tasks, such as weather and finance. Differences and similarities between these domains and health maintenancehave been analyzed to help understand what performance evaluation methods may or may not be borrowed. Further, these metrics have been categorized in several ways that may be useful in deciding upon a suitable subset for a specific application. Some important prognostic concepts have been defined using a notational framework that enables interpretation of different metrics coherently. Last, but not the least, a list of metrics has been suggested to assess critical aspects of RUL predictions before they are fielded in real applications.
- API
NEW HORIZONS SDC PLUTO CRUISE CALIBRATED V2.0
data.nasa.gov | Last Updated 2023-01-26T20:25:34.000ZThis data set contains Calibrated data taken by the New Horizons Student Dust Counter instrument during the pluto cruise mission phase. This is VERSION 2.0 of this data set. SDC collected science data intermittently during the hibernation years following the Jupiter encounter, designated as the PLUTOCRUISE phase. There were also Annual Checkouts (ACOs), STIM calibrations, Noise calibrations, and an anomaly in November, 2007. SDC's main science data collection periods were during hibernation. During ACOs, science data are taken intermittently but the user must be careful in analyzing these data since there is usually more activity on the spacecraft during hibernation. STIM and Noise refer to scheduled calibrations and are done with a regular cadence of one per year after the Jupiter encounter; they occurred sporadically in the early years of the mission. Note that some SDC data files have the same stop and start time and a zero exposure time. The reason for this is that the start and stop time for SDC data files are the event times for the first and last events in the files, so for files that contain a single event, these two values are the same. The changes in Version 2.0 were re-running of the ancillary data in the data product, updated geometry from newer SPICE kernels, minor editing of the documentation, catalogs, etc., and resolution of liens from the December, 2014 review, plus those from the May, 2016 review of the Pluto Encounter data sets. New observations added with this version (V2.0) include ongoing cruise observations from August, 2014 through January, 2015.
- API
CORONA Satellite Photographs from the U.S. Geological Survey
data.nasa.gov | Last Updated 2022-01-17T05:16:00.000ZThe first generation of U.S. photo intelligence satellites collected more than 860,000 images of the Earth’s surface between 1960 and 1972. The classified military satellite systems code-named CORONA, ARGON, and LANYARD acquired photographic images from space and returned the film to Earth for processing and analysis. The images were originally used for reconnaissance and to produce maps for U.S. intelligence agencies. In 1992, an Environmental Task Force evaluated the application of early satellite data for environmental studies. Since the CORONA, ARGON, and LANYARD data were no longer critical to national security and could be of historical value for global change research, the images were declassified by Executive Order 12951 in 1995. The first successful CORONA mission was launched from Vandenberg Air Force Base in 1960. The satellite acquired photographs with a telescopic camera system and loaded the exposed film into recovery capsules. The capsules or buckets were de-orbited and retrieved by aircraft while the capsules parachuted to earth. The exposed film was developed and the images were analyzed for a range of military applications. The intelligence community used Keyhole (KH) designators to describe system characteristics and accomplishments. The CORONA systems were designated KH-1, KH-2, KH-3, KH-4, KH-4A, and KH-4B. The ARGON systems used the designator KH-5 and the LANYARD systems used KH-6. Mission numbers were a means for indexing the imagery and associated collateral data. A variety of camera systems were used with the satellites. Early systems (KH-1, KH-2, KH-3, and KH-6) carried a single panoramic camera or a single frame camera (KH-5). The later systems (KH-4, KH-4A, and KH-4B) carried two panoramic cameras with a separation angle of 30° with one camera looking forward and the other looking aft. The original film and technical mission-related documents are maintained by the National Archives and Records Administration (NARA). Duplicate film sources held in the USGS EROS Center archive are used to produce digital copies of the imagery. Mathematical calculations based on camera operation and satellite path were used to approximate image coordinates. Since the accuracy of the coordinates varies according to the precision of information used for the derivation, users should inspect the preview image to verify that the area of interest is contained in the selected frame. Users should also note that the images have not been georeferenced.
- API
OMI/Aura Level 1B UV Zoom-in Geolocated Earthshine Radiances 1-orbit L2 Swath 13x12 km V003 (OML1BRUZ) at GES DISC
data.nasa.gov | Last Updated 2022-01-17T05:46:59.000ZThe Aura Ozone Monitoring Instrument (OMI) Level-1B (L1B) Geo-located Earth View UV Radiance, Zoom-in-Mode (OML1BRUZ) Version-3 product contains geo-located Earth view spectral radiances from the UV detectors in the wavelength range of 264 to 383 nm using spectral and spatial zoom-in measurement modes. In zoom-in measurement mode, OMI observes 60 ground pixels (13 km x 24 km at nadir) across the swath. Each file contains data from the day lit portion of an orbit (~60 minutes) and is roughly 215 MB in size. There are approximately 14 orbits per day. OMI performs spatial zoom-in measurements one day per month. For that day, this product also contains UV2 measurements that are rebinned from the spatial zoom-in measurements. The shortname for this OMI Level-1B Product is OML1BRUZ. The lead algorithm scientist for this product is Dr. Marcel Dobber from the Royal Netherlands Meteorological Institude (KNMI). The OML1BRUZ files are stored in HDF4 based EOS Hierarchical Data Format (HDF-EOS). The radiances for the earth measurements (also referred as signal) and its precision are stored as a 16 bit mantissa and an 8-bit exponent. The signal can be computed using the equation: signal = mantissa x 10^exponent. For the precision, the same exponent is used as for the signal.
- API
Probabilistic Model-Based Diagnosis for Electrical Power Systems
data.nasa.gov | Last Updated 2020-01-29T02:07:39.000ZWe present in this article a case study of the probabilistic approach to model-based diagnosis. Here, the diagnosed system is a real-world electrical power system, namely the Advanced Diagnostic and Prognostic Testbed (ADAPT) located at the NASA Ames Research Center. Our probabilistic approach is formally well-founded, and based on Bayesian networks and arithmetic circuits. We pay special attention to meeting two of the main challenges model development and real-time reasoning often associated with real-world application of model-based diagnosis technologies. To address the challenge of model development, we develop a systematic approach to representing electrical power systems as Bayesian networks, supported by an easy-touse specication language. To address the real-time reasoning challenge, we compile Bayesian networks into arithmetic circuits. Arithmetic circuit evaluation supports real-time diagnosis by being predictable and fast. In experiments with the ADAPT Bayesian network, which contains 503 discrete nodes and 579 edges and produces accurate results, the time taken to compute the most probable explanation using arithmetic circuits has a mean of 0.2625 milliseconds and a standard deviation of 0.2028 milliseconds. In comparative experiments, we found that while the variable elimination and join tree propagation algorithms also perform very well in the ADAPT setting, arithmetic circuit evaluation was an order of magnitude or more faster. **Reference:** O. J. Mengshoel, M. Chavira, K. Cascio, S. Poll, A. Darwiche, and S. Uckun. "Probabilistic Model-Based Diagnosis: An Electrical Power System Case Study”. Accepted to IEEE Transactions on Systems, Man, and Cybernetics, Part A, 2009.
- API
MISR Level 3 FIRSTLOOK Global Land product in netCDF format covering a day V002
data.nasa.gov | Last Updated 2023-01-19T22:32:40.000ZThis file contains the MISR Level 3 FIRSTLOOK Component Global Land product in netCDF format covering a day
- API
SIAM 2007 Text Mining Competition dataset
data.nasa.gov | Last Updated 2020-01-29T04:25:03.000Z**Subject Area:** Text Mining **Description:** This is the dataset used for the SIAM 2007 Text Mining competition. This competition focused on developing text mining algorithms for document classification. The documents in question were aviation safety reports that documented one or more problems that occurred during certain flights. The goal was to label the documents with respect to the types of problems that were described. This is a subset of the Aviation Safety Reporting System (ASRS) dataset, which is publicly available. **How Data Was Acquired:** The data for this competition came from human generated reports on incidents that occurred during a flight. **Sample Rates, Parameter Description, and Format:** There is one document per incident. The datasets are in raw text format. All documents for each set will be contained in a single file. Each row in this file corresponds to a single document. The first characters on each line of the file are the document number and a tilde separats the document number from the text itself. **Anomalies/Faults:** This is a document category classification problem.