Open Access

Enhancing disease surveillance with novel data streams: challenges and opportunities

  • Benjamin M Althouse1Email author,
  • Samuel V Scarpino1,
  • Lauren Ancel Meyers1, 2,
  • John W Ayers3,
  • Marisa Bargsten4,
  • Joan Baumbach4,
  • John S Brownstein5, 6, 7,
  • Lauren Castro8,
  • Hannah Clapham9,
  • Derek AT Cummings9,
  • Sara Del Valle8,
  • Stephen Eubank10,
  • Geoffrey Fairchild8,
  • Lyn Finelli11,
  • Nicholas Generous8,
  • Dylan George12,
  • David R Harper13,
  • Laurent Hébert-Dufresne1,
  • Michael A Johansson14,
  • Kevin Konty15,
  • Marc Lipsitch16,
  • Gabriel Milinovich17,
  • Joseph D Miller18,
  • Elaine O Nsoesie5, 6,
  • Donald R Olson15,
  • Michael Paul19,
  • Philip M Polgreen20,
  • Reid Priedhorsky8,
  • Jonathan M Read21, 22,
  • Isabel Rodríguez-Barraquer9,
  • Derek J Smith23,
  • Christian Stefansen24,
  • David L Swerdlow25,
  • Deborah Thompson4,
  • Alessandro Vespignani26 and
  • Amy Wesolowski16
Contributed equally
EPJ Data Science20154:17

DOI: 10.1140/epjds/s13688-015-0054-0

Received: 15 September 2015

Accepted: 8 October 2015

Published: 16 October 2015

Abstract

Novel data streams (NDS), such as web search data or social media updates, hold promise for enhancing the capabilities of public health surveillance. In this paper, we outline a conceptual framework for integrating NDS into current public health surveillance. Our approach focuses on two key questions: What are the opportunities for using NDS and what are the minimal tests of validity and utility that must be applied when using NDS? Identifying these opportunities will necessitate the involvement of public health authorities and an appreciation of the diversity of objectives and scales across agencies at different levels (local, state, national, international). We present the case that clearly articulating surveillance objectives and systematically evaluating NDS and comparing the performance of NDS to existing surveillance data and alternative NDS data is critical and has not sufficiently been addressed in many applications of NDS currently in the literature.

Keywords

disease surveillance novel data streams digital surveillance

1 What are novel data streams?

We define NDS as those data streams whose content is initiated directly by the user (patient) themselves. This would exclude data sources such as electronic health records, disease registries, vital statistics, electronic lab reporting, emergency department visits, ambulance call data, school absenteeism, prescription pharmacy sales, serology, amongst others. Although ready access to aggregated information from these excluded sources is novel in many health settings, our focus here is on those streams which are both directly initiated by the user and also not already maintained by public health departments or other health professionals. Despite this more narrow definition our suggestions for improving NDS surveillance may also be applicable to more established surveillance systems, participatory systems (e.g., Flu Near You, influenzaNet) [1, 2], and new data streams aggregated from established systems, such as Biosense 2.0 and ISDS DiSTRIBuTE network [3, 4].

While much of the recent focus on using NDS for disease surveillance has centered on Internet search queries [5, 6] and Twitter posts [7, 8], there are many NDS outside of these two sources. Our aim therefore is to provide a general framework for enhancing and developing NDS surveillance systems, which applies to more than just search data and Tweets. At a minimum, our definition of NDS would include Internet search data and social media, such as Google searches, Google Plus, Facebook, and Twitter posts, as well as Wikipedia access logs [9, 10], restaurant reservation and review logs [11, 12], non-prescription pharmacy sales [13, 14], news source scraping [15], and prediction markets [16].

2 How does NDS integrate into the surveillance ecosystem?

Using NDS for surveillance or in supporting public health decision making necessitates an understanding of the complex link between the time-varying public health problems (i.e., disease incidence) and the time-varying NDS signal. As illustrated in Figure 1, this link is modified by user behavior (i.e., propensity to search, what terms are chosen to search, etc.), user demographics, external forces on user behavior (i.e., changing disease severity, changing press coverage, etc.), and finally by public health interventions, which by design aim to modify the public health problem creating feedback loops on the link to NDS. As a result, developing NDS-based surveillance systems presents a number of challenges, many of which are comparable to those faced by systems comprised of more established data sources such as physician visits or laboratory test results.
Figure 1

The link between public health problems and NDS is modified by user behavior (i.e., propensity to search, what terms are chosen to search, etc.), user demographics, external forces on user behavior (i.e., changing disease severity, changing press coverage, etc.), and finally by public health interventions, which by design aim to modify the public health problem creating feedback loops on the link to NDS.

NDS could add value to existing surveillance in several ways. NDS can increase the timeliness of surveillance information, improve temporal or spatial resolution of surveillance, add surveillance to places with no existing systems, improve dissemination of data, measure unanticipated outcomes of interest (i.e. a syndrome associated with a new pathogen that is not currently under surveillance in an established system), measure aspects of a transmission/disease process not captured by traditional surveillance (i.e. behavior, perception), and increase the population size under surveillance.

The most studied example of the potential benefits and unique challenges associated with NDS comes from Google Flu Trends. In 2008, Google developed an algorithm which translates search queries into an estimate of the number of individuals with influenza-like illness that visit primary healthcare providers [17]. The original goal of Google Flu Trends (GFT) was to provide accessible data on influenza-like illness in order to reduce reporting delays, increase the spatial resolution of data, and provide information on countries outside the United States of America [17]. GFT has added value to existing surveillance for influenza. However, although there has been some benefit both to academic researchers and public health practitioners, GFT has also received criticism [18, 19].

Much of the recent criticism of GFT seems to stem from two issues: the first is the effect of changing user behavior during anomalous events [19, 20] and the second is whether real-time, nowcasting of influenza using GFT adds value to the existing systems available to public health authorities. The first criticism, changing behavior during anomalous events, is an issue for both existing systems and proposed systems based on NDS. The key difference is that existing systems may be both better understood and easier to validate in real-time. While such criticisms may not undermine the case for use of NDS, they do emphasize that the validation of any NDS approach is an ongoing process, and even a perfectly validated system in one period or location may become uncalibrated as behaviors change. It is therefore not meaningful to say that a particular NDS system is or is not informative; that statement must be qualified in space and time. Moreover, the fact that decalibration to “gold standard” systems cannot be detected immediately but only in retrospect is another reason why NDS can only supplement and never fully replace such systems. The second criticism, the need for nowcasting, may depend on the user’s access to different data sources. For public health authorities with access to high-resolution data on reported cases of influenza, simple autoregressive models can be used to nowcast with high accuracy [19]. However, access to these high resolution data-sets varies by public health level (local, state, federal, and international) as well as by user group: researchers, public health authorities, and the private sector. As a result, the utility of GFT varies by user, but for those without access to high-resolution data, it remains an important source of information.

Since the release of GFT, similar NDS-based systems have been developed to extend surveillance to places where resource or other constraints limit the availability of direct clinical or laboratory surveillance data and improve the timeliness of detection and forecasting of disease incidence. For example, NDS have facilitated expansion of dengue and influenza surveillance to countries without infrastructure capable of real time surveillance [5, 17, 21, 22]. This has also been done in the context of hospitalizations in Texas [23], mental illness, psychological manifestations of physical morbidities [24, 25], and search queries from clinical decision support sites, such as UpToDate [26]. In these cases, although NDS-based systems are being asked to estimate data that is actually being collected, those data are not available quickly enough for use in public health decision making.

As stated earlier, in some cases NDS can be used to assess behavior - something that remains a challenge for traditional case-based surveillance. Although this is a challenge for translating NDS signals into estimates of disease incidence, it presents a unique opportunity to study health seeking behavior. For example, NDS has facilitated an exploration of population-level changes in health-related behaviors following changes in tobacco related policy [27, 28] or after unpredictable events such as celebrity deaths or cancer diagnoses [29, 30]. NDS can help us understand and monitor health-related behavior, but little recent work has focused on this area. How does vaccination sentiment respond to changes in disease prevalence? How is health-seeking behavior discussed in social networks? Does that information dissemination manifest in action? Answering these questions accurately may require integration of Twitter, Facebook, Wikipedia access logs, web searches or web search logs, hospitalization records, and EMR with existing measures of behavior such as the Behavioral Risk Factor Surveillance System. As a result, it is critically important to understand the user’s intent; for example, what are the behavioral, biological, and/or epidemiological underpinnings of information-seeking online? A Google or Wikipedia search for the keyword “ulcer”, for instance, is likely a response to having symptoms of an ulcer while a search for “h pylori” is more likely a response to something more specific, such as a lab confirmed test for an ulcer-causative agent. Similarly, posting a Tweet about a “healthy recipe” is likely a different action than searching for a “healthy recipe”; where the former is an act of broadcasting information, while the latter is an act of searching for information. This suggests that large-scale experiments combining NDS could explore these behaviors.

Therefore, in order to address the challenges associated with NDS-based surveillance and properly integrate NDS into existing systems, we advocate for a three-step system: (1) Quantitatively define the surveillance objective(s); (2) build the surveillance systems and model(s) by adding data (existing and novel) in until there is no additional improvement in model performance to achieve stated objectives, assessed by (3) performing rigorous validation and testing. These steps are comparable to those prescribed for evaluating more established systems [31].

3 How do we ensure the robustness of NDS surveillance systems?

NDS, by their very definition, do not have a long track record of use. As a result, rigorous standards for validating NDS and systems constructed using NDS must be adopted. These validation procedures should include both best practices in machine learning and also best practices from surveillance system design such as the proportion of persons identified that are true positives for the disease under surveillance [31]. Building on previous work [10], we have systematically evaluated the existing published NDS surveillance papers using the following criteria: was validation used and if so what type, are the data open and if not why not, and is the code open source. While we understand that it’s not always possible, due to privacy concerns and data use agreements, to make data open access, it’s essential that the community be able to externally validate methods and NDS. Therefore, a component of validation must be the use of data that is publicly available (or at least available to researchers) for training and testing of NDS. Of 66 papers identified, only 27 (41%) performed any validation, only one [5] stated that the source code was available, and while some used publicly available data, no papers publicly shared the data used in their analyses. (see Table 1 and Table S1 in Additional file 1).
Table 1

The use of open source code and validation across papers using NDS for surveillance

 

Validation

No validation

Open source code

1/66 (1.50%)

0/66 (0%)

No open source code

26/66 (39.4%)

39/66 (59.1%)

While the lack of validation is troubling, there is a deeper issue: it may be the case that many existing standards for validation are inadequate for use on disease surveillance systems using NDS [32, 33]. For example, there are well-documented cases of failure when the training set does not contain important dynamics of the system [3436]. For that reason we advocate for model development by repeated training and testing on subsets of the data and that a final, validation set be held back entirely during model construction [35]. This final validation set should be used only once, at the conclusion of the study. Ideally, this final set will be completely blind to the developer. Lastly, after these development and validation steps, models should be openly evaluated prospectively to further support their validity. Put simply, this approach could be summarized as internal validation, external validation, and continued prospective evaluation. While these steps help to ensure the validity of models, it may be that given the volatile nature of disease processes and human behavior (non-linear and non-stationary dynamics), it may be technically impossible to design robust surveillance systems using proxy data and regression models alone.

Validation must also be conducted by other researchers. First, transparency of methods and reproducibility of forecasts is essential to both the scientific process and in examining the utility of models/NDS. Second, new methods or NDS must demonstrate improvements upon existing methods or data sources. Performance can be over-stated by comparing performance of NDS systems with trivial instances of traditional models. Clear definition of appropriate baseline models and their definition is critical to assessing the improvement of new models utilizing novel streams. Without open access to data and code, these crucial steps are not possible. Ideally, manuscripts would report, in detail, the methods employed, provide open-source code implementing those methods, and make the data used to generate the prediction available. Despite legitimate concerns about privacy, data use agreements between agencies, and the often substantial effort required to gather data, we must work towards ensuring our scientific publications are replicable and useful for evaluating the next generation of surveillance tools.

Validation can also be conducted by complementary studies. For example, researchers could conduct studies on how users interact with NDS sources, such as Google or Twitter. These detailed studies would provide valuable information on potential biases and suggest mechanisms for improving the robustness of surveillance systems constructed from these NDS. The need for these focused studies again highlights the utility of collaboration of private sector companies, such as Google and Twitter, with researchers and public health practitioners. Recent efforts by Google and Twitter to better engage with the research community represents an important first step.

4 What is the future of NDS surveillance?

NDS should provide robust, long-term surveillance solutions. Even after EMR are at the fingertips of public health decision-makers and researchers, NDS will provide a snapshot of activity, which is unrelated to the medical encounter. Therefore, a critical first step when evaluating NDS for surveillance is determining what problems are likely to be short term and what problems are likely to be longer term. Again, clearly defined quantitative surveillance goals must be the most important components of NDS-driven systems.

A second important distinction is that surveillance needs, potential benefits, and general utility vary by country, region, and locality. For example, many state and local health agencies in the U.S. already have access to high-resolution, near real-time data for infectious diseases. In this case, local utility may be limited to understanding behavioral responses. These data are also useful, however, for validating these systems more generally. In regions where less data is available, the utility of models may be high but comprehensive evaluation may not be possible. Finally, both Internet and website (or app) penetration vary by geographic region.

Clearly, NDS cannot replace physician and laboratory data, though it can be used to augment the surveillance coming out of systems collecting that type of data. Furthermore, the need for model validation highlights the often-overlooked importance of maintaining traditional/existing systems in the existing NDS literature. Without these systems, it would be impossible to validate and update NDS-enabled systems.

As a community of researchers and public health decision makers, we must decide on how to proceed. Specifically, we must ensure stability and robustness of these NDS-based systems. Pure research is important, but if our goal is to design systems to support public health decisions, they must achieve a higher level of stability. Peer review of systems must carefully evaluate validation relative to established surveillance systems. This of course gives rise to the open question of who should be responsible for funding and maintaining these new systems. The future success of these efforts hinges on building and maintaining collaborations between private-sector, public health agencies, and academics. Finally, while the field has been critical of Google and GFT, it is because we are able to criticize: No other NDS-based system had continuously provided public health predictions for as long as GFT, many NDS surveillance systems had not been as carefully evaluated [6, 1820], and fewer still had been implemented prospectively. Despite the recent cessation of GFT, Google provided a living system for NDS surveillance. Next generation surveillance systems using NDS hold great promise for improving the health of our global society. Realizing their potential will require more rigorous standards of validation and improved collaboration between researchers in academia, the private sector, and public health.

Notes

Declarations

Acknowledgements

This publication arose from a Santa Fe Institute workshop entitled, “Next Generation Surveillance for the Next Pandemic.” We wish to thank the attendees of this workshop, held May 18th-22nd, 2014 at the Santa Fe Insitute in Santa Fe NM, USA. We also gratefully acknowledge funding from the Santa Fe Institute, the NIH MIDAS Center of Excellence at the Harvard Center for Communicable Disease Dynamics, William Sick, Mike Frumkin and the Frumkin Falco Family Foundation. BMA and SVS also acknowledge support from the Santa Fe Institute and the Omidyar Group. Approved for public release LA-UR-14-25873.

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Authors’ Affiliations

(1)
Santa Fe Institute
(2)
The University of Texas at Austin
(3)
San Diego State University
(4)
New Mexico Department of Health
(5)
Children’s Hospital Informatics Program, Boston Children’s Hospital
(6)
Department of Pediatrics, Harvard Medical School
(7)
Department of Epidemiology, Biostatistics and Occupational Health, McGill University
(8)
Defense Systems and Analysis Division, Los Alamos National Laboratory
(9)
Department of Epidemiology, Johns Hopkins Bloomberg School of Public Health
(10)
Virginia BioInformatics Institute and Department of Population Health Sciences, Virginia Tech
(11)
Influenza Division, Centers for Disease Control and Prevention
(12)
Biomedical Advanced Research and Development Authority (BARDA), Assistant Secretary for Preparedness and Response (ASPR), Department of Health and Human Services
(13)
Chatham House
(14)
Division of Vector-Borne Diseases, NCEZID, Centers for Disease Control and Prevention
(15)
Division of Epidemiology, New York City Department of Health and Mental Hygiene
(16)
Communicable Disease Dynamics, Harvard School of Public Health
(17)
School of Population Health, The University of Queensland
(18)
Division of Vector-Borne Diseases, NCEZID, Centers for Disease Control and Prevention
(19)
Department of Computer Science, Johns Hopkins University
(20)
University of Iowa
(21)
Department of Epidemiology and Population Health, Institute of Infection and Global Health, University of Liverpool
(22)
Health Protection Research Unit in Emerging and Zoonotic Infections, NIHR
(23)
Department of Zoology, University of Cambridge
(24)
Google Inc.
(25)
National Center for Immunization and Respiratory Diseases, Centers for Disease Control and Prevention
(26)
Laboratory for the Modeling of Biological and Socio-technical Systems, Northeastern University

References

  1. Chunara R, Aman S, Smolinski M, Brownstein JS (2013) Flu Near You: an online self-reported influenza surveillance system in the USA. Online J Public Health Inform 5(1):e133 View ArticleGoogle Scholar
  2. Paolotti D, Carnahan A, Colizza V, Eames K, Edmunds J, Gomes G, Koppeschaar C, Rehn M, Smallenburg R, Turbelin C et al. (2014) Web-based participatory surveillance of infectious diseases: the Influenzanet participatory surveillance experience. Clin Microbiol Infect 20(1):17-21 View ArticleGoogle Scholar
  3. Olson DR, Paladini M, Lober WB, Buckeridge DL, ISDS Distribute Working Group (2011) Applying a new model for sharing population health data to national syndromic influenza surveillance: DiSTRIBuTE project proof of concept, 2006 to 2009. PLoS Curr 3:RRN1251 View ArticleGoogle Scholar
  4. Chester KG (2013) BioSense 2.0. Online J Public Health Inform 5(1):e200 View ArticleGoogle Scholar
  5. Althouse BM, Ng YY, Cummings DAT (2011) Prediction of dengue incidence using search query surveillance. PLoS Negl Trop Dis 5:e1258 View ArticleGoogle Scholar
  6. Milinovich GJ, Williams GM, Clements AC, Hu W (2014) Internet-based surveillance systems for monitoring emerging infectious diseases. Lancet Infect Dis 14(2):160-168 View ArticleGoogle Scholar
  7. Chew C, Eysenbach G (2010) Pandemics in the age of Twitter: content analysis of tweets during the 2009 H1N1 outbreak. PLoS ONE 5(11):e14118 View ArticleGoogle Scholar
  8. Broniatowski DA, Paul MJ, Dredze M (2013) National and local influenza surveillance through Twitter: an analysis of the 2012-2013 influenza epidemic. PLoS ONE 8(12):e83672 View ArticleGoogle Scholar
  9. McIver DJ, Brownstein JS (2014) Wikipedia usage estimates prevalence of influenza-like illness in the United States in near real-time. PLoS Comput Biol 10(4):e1003581 View ArticleGoogle Scholar
  10. Generous N, Fairchild G, Deshpande A, Del Valle SY, Priedhorsky R (2014) Global disease monitoring and forecasting with Wikipedia. arXiv:1405.3612
  11. Nsoesie EO, Buckeridge DL, Brownstein JS (2014) Guess who’s not coming to dinner? Evaluating online restaurant reservations for disease surveillance. J Med Internet Res 16(1):e22 View ArticleGoogle Scholar
  12. Harrison C, Jorder M, Stern H, Stavinsky F, Reddy V, Hanson H, Waechter H, Lowe L, Gravano L, Balter S et al. (2014) Using online reviews by restaurant patrons to identify unreported cases of foodborne illness - New York City, 2012-2013. Morb Mort Wkly Rep 63(20):441-445 Google Scholar
  13. Das D, Metzger K, Heffernan R, Balter S, Weiss D, Mostashari F et al. (2005) Monitoring over-the-counter medication sales for early detection of disease outbreaks - New York City. Morb Mort Wkly Rep 54(Suppl):41-46 Google Scholar
  14. Patwardhan A, Bilkovski R (2012) Comparison: flu prescription sales data from a retail pharmacy in the US with Google flu trends and US ILINet (CDC) data as flu activity indicator. PLoS ONE 7(8):e43611 View ArticleGoogle Scholar
  15. Freifeld CC, Mandl KD, Reis BY, Brownstein JS (2008) HealthMap: global infectious disease monitoring through automated classification and visualization of Internet media reports. J Am Med Inform Assoc 15:150-157 View ArticleGoogle Scholar
  16. Polgreen PM, Nelson FD, Neumann GR, Weinstein RA (2007) Use of prediction markets to forecast infectious disease activity. Clin Infect Dis 44(2):272-279 View ArticleGoogle Scholar
  17. Ginsberg J, Mohebbi MH, Patel RS, Brammer L, Smolinski MS, Brilliant L (2008) Detecting influenza epidemics using search engine query data. Nature 457:1012-1014 View ArticleGoogle Scholar
  18. Olson DR, Konty KJ, Paladini M, Viboud C, Simonsen L (2013) Reassessing Google flu trends data for detection of seasonal and pandemic influenza: a comparative epidemiological study at three geographic scales. PLoS Comput Biol 9:e1003256 View ArticleGoogle Scholar
  19. Lazer D, Kennedy R, King G, Vespignani A (2014) The parable of Google flu: traps in big data analysis. Science 343(6176):1203-1205 View ArticleGoogle Scholar
  20. Santillana M, Zhang DW, Althouse BM, Ayers JW (2014) What can digital disease detection learn from (an external revision to) Google flu trends? Am J Prev Med 47(3):341-347 View ArticleGoogle Scholar
  21. Yuan Q, Nsoesie EO, Lv B, Peng G, Chunara R, Brownstein JS (2013) Monitoring influenza epidemics in China with search query from Baidu. PLoS ONE 8(5):e64323 View ArticleGoogle Scholar
  22. Polgreen PM, Chen Y, Pennock DM, Nelson FD, Weinstein RA (2008) Using Internet searches for influenza surveillance. Clin Infect Dis 47:1443-1448 View ArticleGoogle Scholar
  23. Scarpino SV, Dimitrov NB, Meyers LA (2012) Optimizing provider recruitment for influenza surveillance networks. PLoS Comput Biol 8(4):e1002472 MathSciNetView ArticleGoogle Scholar
  24. Ayers JW, Althouse BM, Dredze M (2014) Could behavioral medicine lead the web data revolution? JAMA 311(14):1399-1400 View ArticleGoogle Scholar
  25. Althouse BM, Allem J-P, Childers MA, Dredze M, Ayers JW (2014) Population health concerns during the United States’ Great Recession. Am J Prev Med 46(2):166-170 View ArticleGoogle Scholar
  26. Santillana M, Nsoesie EO, Mekaru SR, Scales D, Brownstein JS (2014) Using clinicians’ search query data to monitor influenza epidemics. Clin Infect Dis. doi:10.1093/cid/ciu647 Google Scholar
  27. Ayers JW, Althouse BM, Ribisl KM, Emery S (2009) Digital detection for tobacco control: online reactions to the United States. Nicotine Tob Res. doi:10.1093/ntr/ntt186 Google Scholar
  28. Ayers JW, Ribisl K, Brownstein JS (2011) Using search query surveillance to monitor tax avoidance and smoking cessation following the United States’ 2009 “SCHIP” cigarette tax increase. PLoS ONE 6:e16777 View ArticleGoogle Scholar
  29. Ayers JW, Althouse BM, Noar SM, Cohen JE (2014) Do celebrity cancer diagnoses promote primary cancer prevention? Prev Med 58:81-84 View ArticleGoogle Scholar
  30. Noar SM, Ribisl KM, Althouse BM, Willoughby JF, Ayers JW (2013) Using digital surveillance to examine the impact of public figure pancreatic cancer announcements on media and search query outcomes. J Natl Cancer Inst Monographs 2013(47):188-194 View ArticleGoogle Scholar
  31. Klaucke DN, Buehler JW, Thacker SB, Parrish RG, Trowbridge FL, Berkelman RL et al. (1988) Guidelines for evaluating surveillance systems. Morb Mort Wkly Rep 37(Suppl 5):1-18 Google Scholar
  32. Kohavi R et al. (1995) A study of cross-validation and bootstrap for accuracy estimation and model selection. In: IJCAI’95, vol 2, pp 1137-1145 Google Scholar
  33. Hastie T, Tibshirani R, Friedman J (2009) The elements of statistical learning, 2nd edn. Springer, Berlin MATHView ArticleGoogle Scholar
  34. Smyth P, Wolpert D (1998) Stacked density estimation. In: Advances in neural information processing systems, pp 668-674 Google Scholar
  35. Murphy KP (2012) Machine learning: a probabilistic perspective. MIT Press, Cambridge Google Scholar
  36. Wolpert DH (2012) What the no free lunch theorems really mean; how to improve search algorithms. Working paper, Santa Fe Institute Google Scholar

Copyright

© Althouse et al. 2015