Big Data in Materials Research and Development is the summary of a workshop convened by the National Research Council Standing Committee on Defense Materials Manufacturing and Infrastructure in February 2014 to discuss the impact of big data on materials and manufacturing. The materials science community would benefit from appropriate access to data and metadata for materials development, processing, application development, and application life cycles. Currently, that access does not appear to be sufficiently widespread, and many workshop participants captured the constraints and identified potential improvements to enable broader access to materials and manufacturing data and metadata. This report discusses issues in defense materials, manufacturing and infrastructure, including data ownership and access; collaboration and exploitation of big data's capabilities; and maintenance of data.
As the availability of high-throughput data-collection technologies, such as information-sensing mobile devices, remote sensing, internet log records, and wireless sensor networks has grown, science, engineering, and business have rapidly transitioned from striving to develop information from scant data to a situation in which the challenge is now that the amount of information exceeds a human's ability to examine, let alone absorb, it. Data sets are increasingly complex, and this potentially increases the problems associated with such concerns as missing information and other quality concerns, data heterogeneity, and differing data formats. The nation's ability to make use of data depends heavily on the availability of a workforce that is properly trained and ready to tackle high-need areas. Training students to be capable in exploiting big data requires experience with statistical analysis, machine learning, and computational infrastructure that permits the real problems associated with massive data to be revealed and, ultimately, addressed. Analysis of big data requires cross-disciplinary skills, including the ability to make modeling decisions while balancing trade-offs between optimization and approximation, all while being attentive to useful metrics and system robustness. To develop those skills in students, it is important to identify whom to teach, that is, the educational background, experience, and characteristics of a prospective data-science student; what to teach, that is, the technical and practical content that should be taught to the student; and how to teach, that is, the structure and organization of a data-science program. Training Students to Extract Value from Big Data summarizes a workshop convened in April 2014 by the National Research Council's Committee on Applied and Theoretical Statistics to explore how best to train students to use big data. The workshop explored the need for training and curricula and coursework that should be included. One impetus for the workshop was the current fragmented view of what is meant by analysis of big data, data analytics, or data science. New graduate programs are introduced regularly, and they have their own notions of what is meant by those terms and, most important, of what students need to know to be proficient in data-intensive work. This report provides a variety of perspectives about those elements and about their integration into courses and curricula.
The National Academies of Sciences, Engineering, and Medicine held the workshop Applying Big Data to Address the Social Determinants of Health in Oncology on October 28â€"29, 2019, in Washington, DC. This workshop examined social determinants of health (SDOH) in the context of cancer, and considered opportunities to effectively leverage big data to improve health equity and reduce disparities. The workshop featured presentations and discussion by experts in technology, oncology, and SDOH, as well as representatives from government, industry, academia, and health care systems. This publication summarizes the presentations and discussions from the workshop.
The concept of utilizing big data to enable scientific discovery has generated tremendous excitement and investment from both private and public sectors over the past decade, and expectations continue to grow. Using big data analytics to identify complex patterns hidden inside volumes of data that have never been combined could accelerate the rate of scientific discovery and lead to the development of beneficial technologies and products. However, producing actionable scientific knowledge from such large, complex data sets requires statistical models that produce reliable inferences (NRC, 2013). Without careful consideration of the suitability of both available data and the statistical models applied, analysis of big data may result in misleading correlations and false discoveries, which can potentially undermine confidence in scientific research if the results are not reproducible. In June 2016 the National Academies of Sciences, Engineering, and Medicine convened a workshop to examine critical challenges and opportunities in performing scientific inference reliably when working with big data. Participants explored new methodologic developments that hold significant promise and potential research program areas for the future. This publication summarizes the presentations and discussions from the workshop.
In 2012, the Defense Intelligence Agency (DIA) approached the National Research Council's TIGER standing committee and asked it to develop a list of workshop topics to explore the impact of emerging science and technology. From the list of topics given to DIA, three were chosen to be developed by the Committee for Science and Technology Challenges to U.S. National Security Interests. The first in a series of three workshops was held on April 23-24, 2012. This report summarizes that first workshop which explored the phenomenon known as big data. The objective for the first workshop is given in the statement of task, which explains that that workshop will review emerging capabilities in large computational data to include speed, data fusion, use, and commodification of data used in decision making. The workshop will also review the subsequent increase in vulnerabilities over the capabilities gained and the significance to national security. The committee devised an agenda that helped the committee, sponsors, and workshop attendees probe issues of national security related to so-called big data, as well as gain understanding of potential related vulnerabilities. The workshop was used to gather data that is described in this report, which presents views expressed by individual workshop participants. Big Data: A Workshop Report is the first in a series of three workshops, held in early 2012 to further the ongoing engagement among the National Research Council's (NRC's) Technology Insight-Gauge, Evaluate, and Review (TIGER) Standing Committee, the scientific and technical intelligence (S&TI) community, and the consumers of S&TI products.
Decades of continuous efforts to address known hazards in the national airspace system (NAS) and to respond to issues illuminated by analysis of incidents and accidents have made commercial airlines the safest mode of transportation. The task of maintaining a high level of safety for commercial airlines is complicated by the dynamic nature of the NAS. The number of flights by commercial transports is increasing; air traffic control systems and procedures are being modernized to increase the capacity and efficiency of the NAS; increasingly autonomous systems are being developed for aircraft and ground systems, and small aircraftâ€"most notably unmanned aircraft systemsâ€"are becoming much more prevalent. As the NAS evolves to accommodate these changes, aviation safety programs will also need to evolve to ensure that changes to the NAS do not inadvertently introduce new risks. Real-time system-wide safety assurance (RSSA) is one of six focus areas for the National Aeronautics and Space Administration (NASA) aeronautics program. NASA envisions that an RSSA system would provide a continuum of information, analysis, and assessment that supports awareness and action to mitigate risks to safety. Maintaining the safety of the NAS as it evolves will require a wide range of safety systems and practices, some of which are already in place and many of which need to be developed. This report identifies challenges to establishing an RSSA system and the high-priority research that should be implemented by NASA and other interested parties in government, industry, and academia to expedite development of such a system.
Massive data streams, large quantities of data that arrive continuously, are becoming increasingly commonplace in many areas of science and technology. Consequently development of analytical methods for such streams is of growing importance. To address this issue, the National Security Agency asked the NRC to hold a workshop to explore methods for analysis of streams of data so as to stimulate progress in the field. This report presents the results of that workshop. It provides presentations that focused on five different research areas where massive data streams are present: atmospheric and meteorological data; high-energy physics; integrated data systems; network traffic; and mining commercial data streams. The goals of the report are to improve communication among researchers in the field and to increase relevant statistical science activity.
Emerging Cognitive Neuroscience and Related Technologies, from the National Research Council, identifies and explores several specific research areas that have implications for U.S. national security, and should therefore be monitored consistently by the intelligence community. These areas include: neurophysiological advances in detecting and measuring indicators of psychological states and intentions of individuals the development of drugs or technologies that can alter human physical or cognitive abilities advances in real-time brain imaging breakthroughs in high-performance computing and neuronal modeling that could allow researchers to develop systems which mimic functions of the human brain, particularly the ability to organize disparate forms of data. As these fields continue to grow, it will be imperative that the intelligence community be able to identify scientific advances relevant to national security when they occur. To do so will require adequate funding, intelligence analysts with advanced training in science and technology, and increased collaboration with the scientific community, particularly academia. A key tool for the intelligence community, this book will also be a useful resource for the health industry, the military, and others with a vested interest in technologies such as brain imaging and cognitive or physical enhancers.
Data mining of massive data sets is transforming the way we think about crisis response, marketing, entertainment, cybersecurity and national intelligence. Collections of documents, images, videos, and networks are being thought of not merely as bit strings to be stored, indexed, and retrieved, but as potential sources of discovery and knowledge, requiring sophisticated analysis techniques that go far beyond classical indexing and keyword counting, aiming to find relational and semantic interpretations of the phenomena underlying the data. Frontiers in Massive Data Analysis examines the frontier of analyzing massive amounts of data, whether in a static database or streaming through a system. Data at that scale-terabytes and petabytes-is increasingly common in science (e.g., particle physics, remote sensing, genomics), Internet commerce, business analytics, national security, communications, and elsewhere. The tools that work to infer knowledge from data at smaller scales do not necessarily work, or work well, at such massive scale. New tools, skills, and approaches are necessary, and this report identifies many of them, plus promising research directions to explore. Frontiers in Massive Data Analysis discusses pitfalls in trying to infer knowledge from massive data, and it characterizes seven major classes of computation that are common in the analysis of massive data. Overall, this report illustrates the cross-disciplinary knowledge-from computer science, statistics, machine learning, and application disciplines-that must be brought to bear to make useful inferences from massive data.
The 2013 report Solar and Space Physics; A Science for a Technological Society outlined a program of basic and applied research for the period 2013-2022. This publication describes the most significant scientific discoveries, technical advances, and relevant programmatic changes in solar and space physics since the publication of that decadal survey. Progress Toward Implementation of the 2013 Decadal Survey for Solar and Space Physics assesses the degree to which the programs of the National Science Foundation and the National Aeronautics and Space Administration address the strategies, goals, and priorities outlined in the 2013 decadal survey, and the progress that has been made in meeting those goals. This report additionally considers steps to enhance career opportunities in solar and space physics and recommends actions that should be undertaken to prepare for the next decadal survey.
The Bulk Collection of Signals Intelligence: Technical Options study is a result of an activity called for in Presidential Policy Directive 28 (PPD-28), issued by President Obama in January 2014, to evaluate U.S. signals intelligence practices. The directive instructed the Office of the Director of National Intelligence (ODNI) to produce a report within one year "assessing the feasibility of creating software that would allow the intelligence community more easily to conduct targeted information acquisition rather than bulk collection." ODNI asked the National Research Council (NRC)-the operating arm of the National Academy of Sciences and National Academy of Engineering-to conduct a study, which began in June 2014, to assist in preparing a response to the President. Over the ensuing months, a committee of experts appointed by the Research Council produced the report.
This book contains a key component of the NII 2000 project of the Computer Science and Telecommunications Board, a set of white papers that contributed to and complements the project's final report, The Unpredictable Certainty: Information Infrastructure Through 2000, which was published in the spring of 1996. That report was disseminated widely and was well received by its sponsors and a variety of audiences in government, industry, and academia. Constraints on staff time and availability delayed the publication of these white papers, which offer details on a number of issues and positions relating to the deployment of information infrastructure.
We live in a changing world with multiple and evolving threats to national security, including terrorism, asymmetrical warfare (conflicts between agents with different military powers or tactics), and social unrest. Visually depicting and assessing these threats using imagery and other geographically-referenced information is the mission of the National Geospatial-Intelligence Agency (NGA). As the nature of the threat evolves, so do the tools, knowledge, and skills needed to respond. The challenge for NGA is to maintain a workforce that can deal with evolving threats to national security, ongoing scientific and technological advances, and changing skills and expectations of workers. Future U.S. Workforce for Geospatial Intelligence assesses the supply of expertise in 10 geospatial intelligence (GEOINT) fields, including 5 traditional areas (geodesy and geophysics, photogrammetry, remote sensing, cartographic science, and geographic information systems and geospatial analysis) and 5 emerging areas that could improve geospatial intelligence (GEOINT fusion, crowdsourcing, human geography, visual analytics, and forecasting). The report also identifies gaps in expertise relative to NGA's needs and suggests ways to ensure an adequate supply of geospatial intelligence expertise over the next 20 years.
Many federal funding requests for more advanced computer resources assume implicitly that greater computing power creates opportunities for advancement in science and engineering. This has often been a good assumption. Given stringent pressures on the federal budget, the White House Office of Management and Budget (OMB) and Office of Science and Technology Policy (OSTP) are seeking an improved approach to the formulation and review of requests from the agencies for new computing funds. This book examines, for four illustrative fields of science and engineering, how one can start with an understanding of their major challenges and discern how progress against those challenges depends on high-end capability computing (HECC). The four fields covered are: atmospheric science astrophysics chemical separations evolutionary biology This book finds that all four of these fields are critically dependent on HECC, but in different ways. The book characterizes the components that combine to enable new advances in computational science and engineering and identifies aspects that apply to multiple fields.
The end of dramatic exponential growth in single-processor performance marks the end of the dominance of the single microprocessor in computing. The era of sequential computing must give way to a new era in which parallelism is at the forefront. Although important scientific and engineering challenges lie ahead, this is an opportune time for innovation in programming systems and computing architectures. We have already begun to see diversity in computer designs to optimize for such considerations as power and throughput. The next generation of discoveries is likely to require advances at both the hardware and software levels of computing systems. There is no guarantee that we can make parallel computing as common and easy to use as yesterday's sequential single-processor computer systems, but unless we aggressively pursue efforts suggested by the recommendations in this book, it will be "game over" for growth in computing performance. If parallel programming and related software efforts fail to become widespread, the development of exciting new applications that drive the computer industry will stall; if such innovation stalls, many other parts of the economy will follow suit. The Future of Computing Performance describes the factors that have led to the future limitations on growth for single processors that are based on complementary metal oxide semiconductor (CMOS) technology. It explores challenges inherent in parallel computing and architecture, including ever-increasing power consumption and the escalated requirements for heat dissipation. The book delineates a research, practice, and education agenda to help overcome these challenges. The Future of Computing Performance will guide researchers, manufacturers, and information technology professionals in the right direction for sustainable growth in computer performance, so that we may all enjoy the next level of benefits to society.
The mathematical sciences are part of nearly all aspects of everyday life-the discipline has underpinned such beneficial modern capabilities as Internet search, medical imaging, computer animation, numerical weather predictions, and all types of digital communications. The Mathematical Sciences in 2025 examines the current state of the mathematical sciences and explores the changes needed for the discipline to be in a strong position and able to maximize its contribution to the nation in 2025. It finds the vitality of the discipline excellent and that it contributes in expanding ways to most areas of science and engineering, as well as to the nation as a whole, and recommends that training for future generations of mathematical scientists should be re-assessed in light of the increasingly cross-disciplinary nature of the mathematical sciences. In addition, because of the valuable interplay between ideas and people from all parts of the mathematical sciences, the report emphasizes that universities and the government need to continue to invest in the full spectrum of the mathematical sciences in order for the whole enterprise to continue to flourish long-term.
The future security, economic growth, and competitiveness of the United States depend on its capacity to innovate. Major sources of innovative capacity are the new knowledge and trained students generated by U.S. research universities. However, many of the complex technical and societal problems the United States faces cannot be addressed by the traditional model of individual university research groups headed by a single principal investigator. Instead, they can only be solved if researchers from multiple institutions and with diverse expertise combine their efforts. The National Science Foundation (NSF), among other federal agencies, began to explore the potential of such center-scale research programs in the 1970s and 1980s; in many ways, the NSF Engineering Research Center (ERC) program is its flagship program in this regard. The ERCs are "interdisciplinary, multi-institutional centers that join academia, industry, and government in partnership to produce transformational engineered systems and engineering graduates who are adept at innovation and primed for leadership in the global economy. To ensure that the ERCs continue to be a source of innovation, economic development, and educational excellence, A New Vision for Center-Based Engineering Research explores the future of center-based engineering research, the skills needed for effective center leadership, and opportunities to enhance engineering education through the centers.
What are employer needs for staff trained in the natural sciences at the master's degree level? How do master's level professionals in the natural sciences contribute in the workplace? How do master's programs meet or support educational and career goals? Science Professionals: Master's Education for a Competitive World examines the answers to these and other questions regarding the role of master's education in the natural sciences. The book also focuses on student characteristics and what can be learned from efforts underway to enhance the master's in the natural sciences, particularly as a professional degree. This book is a critical tool for Congress, the federal agencies charged with carrying out the America COMPETES Act, and educational and science policy makers at the state level. Additionally, anyone with a stake in the development of professional science education (four year institutions of higher education, students, faculty, and employers) will find this book useful.
The National Children's Study (NCS) is planned to be the largest long-term study of environmental and genetic effects on children's health ever conducted in the United States. It proposes to examine the effects of environmental influences on the health and development of approximately 100,000 children across the United States, following them from before birth until age 21. By archiving all of the data collected, the NCS is intended to provide a valuable resource for analyses conducted many years into the future. This book evaluates the research plan for the NCS, by assessing the scientific rigor of the study and the extent to which it is being carried out with methods, measures, and collection of data and specimens to maximize the scientific yield of the study. The book concludes that if the NCS is conducted as proposed, the database derived from the study should be valuable for investigating hypotheses described in the research plan as well as additional hypotheses that will evolve. Nevertheless, there are important weaknesses and shortcomings in the research plan that diminish the study's expected value below what it might be.
All U.S. agencies with counterterrorism programs that collect or "mine" personal data-such as phone records or Web sites visited-should be required to evaluate the programs' effectiveness, lawfulness, and impacts on privacy. A framework is offered that agencies can use to evaluate such information-based programs, both classified and unclassified. The book urges Congress to re-examine existing privacy law to assess how privacy can be protected in current and future programs and recommends that any individuals harmed by violations of privacy be given a meaningful form of redress. Two specific technologies are examined: data mining and behavioral surveillance. Regarding data mining, the book concludes that although these methods have been useful in the private sector for spotting consumer fraud, they are less helpful for counterterrorism because so little is known about what patterns indicate terrorist activity. Regarding behavioral surveillance in a counterterrorist context, the book concludes that although research and development on certain aspects of this topic are warranted, there is no scientific consensus on whether these techniques are ready for operational use at all in counterterrorism.
This will help us customize your experience to showcase the most relevant content to your age group
Please select from below
Login
Not registered?
Sign up
Already registered?
Success – Your message will goes here
We'd love to hear from you!
Thank you for visiting our website. Would you like to provide feedback on how we could improve your experience?
This site does not use any third party cookies with one exception — it uses cookies from Google to deliver its services and to analyze traffic.Learn More.