Abstract
‘Strategic Conservation Planning’ has become a research discipline very relevant for conservation, but is still widely centered on optimization software and stakeholder input, and thus limited, narrow and quite subjective. While data are often missing, a simple scoping and delphi process using experts is commonly applied to achieve a simple agreement on data gaps. This has now become a global standard, e.g. for conserving ocean systems and many other applications for precious natural landscapes. Whereas, the existing network of globally protected areas is driven by widely under-protected landscapes, ad hoc and last-minute decisions, convenience and industrial and political compromise. Here it gets demonstrated how machine learning can be used to assist, improve, extend and progress the strategic conservation process. While species distribution models (SDMs) are widely cited as ‘good’ approaches to describe patterns and processes, they also make Strategic Conservation Planning better. Several additional options also exist such as spatial Population Viability Analysis (sPVA) and scenario predictions, all based on computationally intense global applications explicit in space and time. While long-term sustainability always should be the ultimate goal, eventually, the best-possible spatial and temporal predictions help us to provide relevant and urgently needed progress for conservation to reach a society that lives sustainably in a global framework by making good use of all options possible.
“National Parks are just a rather poor transitional protection scheme by the currently dominating western world trying to put the more serious destruction on hold before better systems are found to truly achieve global sustainability”
Source unknown, Writing Center, University of Alaska Fairbanks campus, 2014
Access provided by CONRICYT-eBooks. Download chapter PDF
Similar content being viewed by others
Keywords
Arguably, the modern western life-style and culture has dramatically marginalized the state of the global environment and its natural resources (Daly and Farley 2011; see Czech et al. 2000 for impacts and Walther et al. 2016 for wider implications). It comes then to no surprise that - as a global pattern - the environment is by now in a state of global crisis (Mace et al. 2010 and Cockburn 2013 for a generic assessment). While humans have used the earth for millennia and made a certain footprint (see Groube et al. 1986 for an example of 40,000 years of documented human occupancy and Flannery 2002 for its benign impacts), the specific anthropogenic footprint and impact of the last four decades remain unprecedented in terms of extinction and global climate change (Rockström et al. 2009, Baltensperger and Huettmann 2015). It is noteworthy that the predominant governance paradigm during this period is globalization, based on Americanization (Czech 2000; Stiglitz 2003). The last few decades are arguably the worst managed in human history (Paehlke 2004; Alexander 2013). Actually, the history of the earth and universe as we know it has not produced such a destruction of life (by non-cosmic events) ever before (Cushman and Huettmann 2010); consider the outlook of what will easily be 10 billion people in the next 100 years while global temperatures are also on the rise and most natural resources are already used up (Rockström et al. 2009)!
While I leave it here to others to assign that blame and to document impacts in great detail (Rich 1994; Czech et al. 2000; Huettmann 2011; Cockburn 2013), a way out of this crisis - moving forward in the best-possible fashion - is likely to be pro-active (e.g. to avoid problems before they occur) and to be pre-cautionary (e.g. to identify problems ahead of time and act carefully; Huettmann 2007a, b). The notion of “the need to know impacts ahead of time before they occur” are not new concepts in conservation though (Silva 2012), and they are by now mandatory by the U.N. and part of ‘best professional practices’ (http://unesdoc.unesco.org/images/0013/001395/139578e.pdf). This is a key ingredient for good governance and for a trusted leadership. Machine learning plays a central role in this approach, and for achieving the best-possible predictions, to be obtained by the best-available science (Huettmann 2007a, b) before impacts occur, e.g. for Alaska see Murphy et al. (2010), Baltensperger et al. (2015), Huettmann et al. (2017).
However, it is easy to show that, thus far, Strategic Conservation Planning does not use much machine learning (see Ardron et al. 2008 and Moilanen et al. 2009 for ‘best professional practices’ and textbook). Instead, the Strategic Conservation Planning tool has been almost entirely a stand-alone approach not connecting with machine learning. Mostly it relies just on optimization algorithms to find the best solution, such as for instance the ‘simulated annealing’ algorithm (as employed by MARXAN http://www.marxan.org/; Martin et al. 2008). Further, most MARXAN applications “do not look much into the future” , e.g. by using future scenarios obtained from machine learning and optimizing those ones (see Nielsen et al. 2008 and Murphy et al. 2010 for an application). Instead, latest developments actually deal with optimizing in ‘zones’ - subunits - (http://marxan.net/index.php/ marzone; Watts et al. 2009). I find this to be a problem on three accounts: (i) The latest science, machine learning, and related potential got ignored. (ii) Breaking down a spatial optimization problem into small separate, parsimonious, zones loses the overall optimization power, and (iii) relevant progress got somewhat stifled by forcing creative minds and their solutions back into existing administrative boundaries and units and thus just re-confirming existing problems and patterns. In this chapter, I outline how machine learning has been used and how it could play a larger role in Strategic Conservation Planning projects towards true progress beyond circular reasoning and traditional mind traps. I am adding relevant perspectives on carrying capacity, limits of the earth and global governance to achieve global sustainability.
1 How Machine Learning Predictions Feed into Strategic Conservation Planning: A Common Sense Workflow Still Widely Underused for its Conservation Potential
One of the largest Strategic Conservation Planning projects was designed to make recommendations for the marine protected area (MPA) networks . Similar to the terrestrial national parks of the world, it is meant to protect the world’s oceans and assure that a fixed percentage (~10%) set aside (Aichi targets for 2020 by the U.N.: https://www.cbd.int/sp/targets/)! Strategic Conservation Planning has also been applied to terrestrial systems, although many of these earlier approaches were just ad hoc and not optimized much for relevant conservation gains (see Table 16.1). Global applications that involve wider and holistic concepts such as the atmosphere and its protection are still lacking (but see Huettmann 2012 how to include those aspects). Arguably though, one cannot achieve sustainable solutions on a small scale because all is highly connected and influenced globally.
In the marine world, the Great Barrier Reef, Australia , has one of the longest histories of Strategic Conservation Planning (UNESCO Great Barrier Reef Marine Park Authority 1981. Fernandes et al. 2005). The task was here to identify the most relevant reef locations for human enjoyment in perpetuity (see end of the chapter for its outcome, thus far) and much can be learned from that ‘experiment’ and exercise.
Widely used software for such Strategic Conservation Planning are MARXAN , but also ZONATION (Lehtomäki and Moilanen 2013), and C-Plan (http://www.edg.org.au/edg-free-resources/cplan.html) as well as many derivatives across operating systems and software implementations (e.g. https://www.aproposinfosystems.com/solutions/qmarxan-toolbox/). Such software helps users by finding ‘the optimal’ solution through an approximation for an assumed truth. The underlying theoretical differences between software packages is not described here but can be found in their respective manuals and URLs, as well as elsewhere in the conservation literature.
For being successful and without relevant errors, Strategic Conservation Planning projects usually require substantial input of information (data). Ideally, maps of species ranges and conservation features for the study area and its planning units are to be available. The study area consists of planning units (PUs, which are often ‘bins’ e.g. hexagons or pixels). This highly detailed information is widely missing for larger study areas, however, conservation decisions must still be made while destruction is ongoing (Alidina et al. 2008). That is specifically true for a global scale and on a macro-ecology perspective (see Forman 1995 for an effective balance using ‘a regional scale’). How can we overcome the problem of data gaps fast enough and with reliable information so that we can make informed conservation decisions in the best way possible for large areas of the globe?
In the past, so-called ad hoc decisions were made with political convenience and opportunism driving the agenda (see for instance Huettmann and Hazlett 2010 for Alaska). The protected area network for most of the circumpolar Arctic, in Russia (Spiridonov et al. 2012 for the Russian arctic) or all of North America reflects just that (‘protection of rock and ice’ Scott et al. 2001). Experts got used to identify and fill data gaps in Strategic Conservation Planning projects. In addition, scoping meetings are often held with commercial stakeholders. This is widely referred to as a delphi process (a non-scientific process that simply banks on agreements and compromises made during a session). The use of experts is known though to be biased and has been widely criticized for years and was assessed accordingly (e.g. Perera et al. 2012; see Gonzales et al. 2003 for a real-world example). Those planning efforts may not be representative; and often just the most effluent, vocal or wealthy stakeholders drive the process. The delphi method can easily compromise the entire objectivity, fairness, quality and transparency of Strategic Conservation Planning and of its science, process and trustworthiness overall.
It is here where machine learning offers help. It can provide a solution to the problem of filling data gaps with the delphi method. As shown in Drew et al. (2011) SDMs can be produced with transparent, repeatable methods using open access data (see GBIF.org for species data, and worldclim.org for environmental data) with open source models (e.g. Openmodeler http://openmodeller.sourceforge.net/om_ desktop.html or commercial models (SPM https://www.salford-systems.com/). The SDMs with the highest accuracy tend to use machine learning; see Elith et al. 2006 and Fernandez-Delgado et al. 2014 for a review of the highly performing algorithms available.
However, despite the easy access, very few SDMs are actually run for, or used by, Strategic Conservation Planning projects (Moilanen et al. 2009; but see Beiring 2013 for an example how SDMs are employed for Strategic Conservation Planning). There are several reasons why that problem still exists, as outlined in Table 16.2.
In my experience, available SDM applications are never really sufficient and not truly complete for Strategic Conservation Planning exercises (see Table 16.3 for shared experiences). In addition, SDM accuracy is another topic to watch for. That’s because SDM accuracies are usually inconsistent across species in a study area, and which does not allow then for consistent inference or application. Sometimes SDM performances can be too low to be useful or to be employed to larger areas (Aycrigg et al. 2015). Overall, I often find that limited and sparse raw data cannot provide consistent and year-round spatial estimates of important demographic and ecological parameters such as fecundity, winter survival and migration risk for instance. SDMs rarely are applied yet to provide demographically relevant spatial estimates such as mating places, productivity hotspots and mortality landscapes (often summarized as ‘sources and sinks’; Pulliam 1988). On the one hand, the detail that would be ideal and needed is rarely possible to achieve with demographic tools (Amstrup et al. 2006). On the other hand, conservation decisions are urgent and in times of a global conservation crisis (Rockström et al. 2009, Mace et al. 2010). And so even basic SDMs , such as species occurrence, help to provide information that is useful for the overall Strategic Conservation Planning process. By now, virtually all what is not parsimonious (e.g. using Generalized Linear Models (GLMs) and Akaike’s Information Criteria (AIC); Arnold 2010) can be achieved as progress (Breiman 2001) considering the global environmental crisis mankind is facing.
Overall, the use of machine learning for SDM approaches has greatly improved Strategic Conservation Planning processes (Table 16.3). These projects are often the first of their kind allowing for these methods and approaches to be introduced to Strategic Conservation Planning in the region. While data must be available to run SDMs (but often are not openly shared in SDM publications, and hardly in webportals like Movebank https://www.movebank.org/) and model outputs could easily be improved, these projects then allow for a subsequent machine learning ‘culture’ of conservation to be set up, also based on stakeholders in a public forum. It’s rather transparent that way; the contribution of GBIF in that context remain unchallenged (see for instance Beiring 2013). Beyond data and model accuracies, the actual introduction of such a new conservation culture may be the biggest contribution. The need for data models, transparency and stakeholder buy-in is essential for implementing management of natural resources (Clark 2002). Having that awareness and accept such a need of wider community buy-in might well present the main contribution in SDMs and Strategic Conservation Planning projects for a global sustainable society.
2 How Machine Learning Predictions can Also Be Used Directly for Strategic Conservation Planning: How it’s ‘ought to be’ and towards ‘Better’ Solutions
Strategic Conservation Planning is usually employed with an optimization framework; rarely, it is used with forecast scenarios directly. The actual optimization is usually based on methods like ‘solvers’, namely the simulated annealing algorithm (Ardron et al. 2008), whereas machine learning just provides the input GIS layers for describing generic patterns in the landscape. The use of future scenarios is possible and has been increasingly applied though (see Murphy et al. 2010 for an example). SDMs built with explanatory variables that have future forecasts, such as downscale global climate models and for 2100 let’s say, can be used to forecast future conditions for species. While Population Viability Analysis (PVA) lack much of the spatially explicit aspects (e.g. Proctor et al. 2004), a spatial population viability analysis (sPVA) offers interesting and relevant possibilities for forecasting future conditions for Strategic Conservation Planning (Nielsen et al. 2008 for an example). These techniques link demographic PVA approaches with GIS habitat data and future scenarios, all based on optimizations from Strategic Conservation Planning. It tends to represent the best science available!
Often, sPVAs themselves fall short on some of the principles of Strategic Conservation Planning, or leave them unaddressed (Table 16.4), such as lack of optimization and not comparing several scenarios in parallel but just favoring singularity and reductionism. However, the strengths of sPVAs linked with scenario planning (Peterson et al. 2003) are that they can be much better and directly applied and tested for specific management questions, including demographic sensitivities and outlooks. The use of well-thought out scenarios provide policy alternatives (e.g. Gonzales et al. 2003; Nielsen et al. 2008) as compared to the narrow, singular and traditionally used Strategic Conservation Planning solution (which often consist of just ‘one’ solution wiping all other thoughts off the agenda). Computationally, and for all work that includes machine learning, this can easily be achieved. However, the sPVA option and when applied with scenarios is not yet widely employed in wildlife management and it is not required by law (see Huettmann et al. 2005; Nielsen et al. 2008) but it can come rather close to the ideal of adaptive wildlife management (Walters 1986; Huettmann 2007a; b). In the meantime, scenario-planning starts to become more common (Peterson et al. 2003). e.g. for climate change outlooks (IPCC; http://www.ipcc.ch/). The use of scenarios is widely known in the social sciences but so far less common in the traditional North American Model of wildlife management (Organ et al. 2012; Silva 2012), or in most other natural resource management schemes in the world. Clearly, such work is computational intense, and the role of coding and linking tools and codes across operating platforms becomes a power tool to achieve such conservation solutions! It’s all part of machine learning either way!
IMPUTATION: More Ways with Machine Learning to Fill Data Gaps and Smooth–Out Predictors with Information Gain for an Effective Science–Based Conservation Management
Most statistical analysis default on data gaps. It’s a ‘no go’ zone for modern science and it tends to result in statements of convenience such as ‘no evidence exist.’ So what to do when your data set is gappy, missing records and gets labeled ‘no go’ zone? This seemingly old question remains a ‘hot button’ item, and becomes now a key bottleneck to overcome for progress. This problem actually made it to the forefront of modern data analysis (e.g. Graham 2009). Many reasons can be envisioned why data gaps occur. But if that can be resolved, then new insights can be won, and a subject can be moved forward (see Kenward 2013 for health applications). Thus, finding methods to substitute and fill data gaps make for a classic but very relevant and modern problem that data miners and machine learners have to deal with. It’s a typical case in conservation that data are gappy while decisions are to be made though.
Reason for data gap | Detail of the data gap | Comment |
---|---|---|
Predictor has a data gap | Data column is incomplete | This is a common problem. |
Response has a data gap | Data column is incomplete | In times of open access data this probably the biggest problem. |
Missing data per se | Entire absence of information | Probably the biggest issue for holding us back for making inference |
Lacking data due to research design problems | Entire lack of a specific information | The ‘best’ research design is usually not known ahead of the study. At minimum, research design can always be improved in hindsight. |
Wrong entries | Entry is not correct | Depending on the term ‘wrong’ and its definitions some datasets are said to feature 50% of such ‘errors’ and ‘wrong’ entries. |
Entries got erased by accident | Hole in the data set | This has been observed quite a lot and when data are modified, re-formatted and sorted in spread sheets for instance, or when they get operated by inexperienced people. |
Entries had to be removed due to uncertainties | Hole in the data set | This is a common problem, e.g. when metadata are missing and with older data. Doubts about geo-referencing, taxonomy and collection times are typical examples. Arguably, it’s better to keep weak data than having none. |
Entries had to be removed due to copyrights | Hole in the data set | This is also a common problem when data are not shared. |
Error simulations | Based on certain assumed backgrounds, some data might be acceptable whereas under other assumptions might make this data set viable | For advanced inference, data error simulation can provide new insights. While this can present questions for high-end data mining, it can be a relevant feature for high-end information gain and inference. |
One can easily envision a situation where these ‘real world’ data gaps occur, and then, where they ‘magically’ could be overcome! So why not simply imputing them? Footnote 1 Conceptually, imputation means to replace data holes and to fill (substitute) them (Enders 2010). In reality that means they are to be modeled! It usually makes use of existing, neighboring, associated and surrounding attributes and data points. Based on those relationships in the data, data gaps can get filled. With the help of advanced computing and data science, it is less a question ‘whether’ that can be done (no problem really to do so). Instead, the issue is more ‘how good’ is the accuracy obtained for a certain purpose (many applications are happy to have a 75% modeled accuracy when compared to no data at all)? In a way, imputation is a specific method to model-predict data gaps. And this can be done, all with a certain estimation accuracy. A 100% prediction accuracy can probably not be achieved, but often it is not needed even. Instead, one can fill errors in a decent way, the models do not default, and which helps to move the overall process and progress forward for an analysis topic.
By now, imputation, as a statistical discipline, is evolving fast and many methods exist (see also for updates at the Wikipedia site https://en.wikipedia.org/wiki/Imputation_(statistics). Major techniques are for instance single imputations such as hot-deck, cold-deck, mean-substitution, regression. Multiple imputations are other powerful techniques. Often, these methods are linked with sampling and re-sampling approaches. Of particular relevance are the geo-imputation methods as they allow interpolations in geographic information systems (GIS), linked with the discipline and tools from Spatial Statistics. Those are popular in forestry, canopy, remote sensing and image analysis too. Entire textbooks and journals are devoted to that topic, e.g. https://www.journals.elsevier.com/spatial-statistics.
There are several software approaches possible that allow to run imputations; see for instance Enders (2010). Similarly, Salford Systems Ltd. offers in TreeNet (stochastic gradient boosting) options to run analysis with ‘gappy data’ (Friedman 1999, Salford Systems Ltd. https://www.salford-systems.com/products/treenet)). In the R language, YAImpute is one of those packages that can ‘impute’ data based on using nearest neighbor observations (kNN; Crookston and Finley (2008; https://cran.r-project.org/web/ packages/ yaImpute/ index.html). See also applications of such R code by J. Evans (http://evansmurphy.wixsite.com/evansspatial).
While this is all pretty exciting, developing and moving forward fast, the sad news is that in wildlife conservation management, and for many natural resource applications, imputation convinces in the pure absence. The mainstream literature is extremely poor on making use of those methods and for advancing fields like conservation remote sensing, geo-locators, telemetry, wildlife surveys, disease outbreaks and citizen science. Two notable exceptions can be found though, namely climate as well as some forest work (Eskelson et al. 2009).
Literature
-
Crookston NL, Finley AO (2008) yaImpute: an R Package for kNN Imputation. J Stat Softw 23: 1–16
-
Enders CK (2010) Applied missing data analysis. Guilford Press, New York
-
Eskelson BN, Temesgen H, Lemay V, Barrett TM, Crookston NL, Hudak AK (2009) The roles of nearest neighbor methods in imputing missing data in forest inventory and monitoring databases. Scand J For Res 24: 235–246
-
Friedman JH (2002) Stochastic gradient boosting. Comp Stat Data Anal 38: 367–378
-
Graham JW (2009) Missing data analysis: making it work in the real world. Annu Rev Psychol 60: 549–576
-
Jiao S, Huettmann F, Guo Y, Li,Ouyang Y (2016) Advanced long-term bird banding and climate data mining in spring confirm passerine population declines for the Northeast Chinese-Russian flyway. Glob Planet Chang https://doi.org/10.1016/j.gloplacha.2016.06.015
-
Kenward MG (2013) The handling of missing data in clinical trials. Clin Investig 3: 241–250
-
Lele SR, Dennis B, Lutscher F (2007) Data cloning: easy maximum likelihood estimation for complex ecological models using Bayesian Markov chain Monte Carlo methods. Ecol Lett 10: 551–563
3 A Wider Perspective against the Local Techno-Fix : Good Ethics and Ecological Governance Foundations to Achieve a Conservation Break–Through with Open Access Machine Learning and Strategic Conservation Planning Carrying a Global Perspective for Mankind
An honest assessment of the environment, and the strategy needed for its conservation will expose nothing but a crisis, which is footed on a failed management and leadership in an overall destructive global framework (Ostrom 2015). We are running out of space while most relevant species and habitats are not protected, or at least not protected well or optimal! (Table 16.5).
The New World Order, and starting with Bretton Woods in 1944, The World Bank and its subsequent institutions such as IUCN and UNEP show us nothing but that (Rich 1994). All too often we then just get presented a techno fix used to present us with progress when there actually is none (Czech et al. 2000; Cockburn 2013); even basic principles of strategic conservation are consistently violated (Table 16.4). By now, the list of those technical ‘innovations’ and fixes are very long, almost comical if it were not that tragic. It is easy to see that machine learning, or optimization algorithms from strategic conservation planning could fall into that category. The challenge now remains to show that it is not and to apply them in a good ecological setting. On the one hand machine learning cannot really break out of the techno-trail. It’s a technological high-end application. It also requires energy and resources as input, eventually. Some of the stakeholder workshops also leave a big financial and carbon footprint, for instance. However, on the good side, the outcome is more than its individual pieces. That is nowhere so true than in machine learning: just consider what the phrase ‘many weak learners make for a strong learner’ (Schapire 1990) means in real life.
So, while we are easily trapped in our institutions and minds with certain techno-arguments and its neoliberal world, there can be a decent output for the better, and hopefully with a good life- balance to be found eventually. Tables 16.4 and 16.6 show some of the core components of Strategic Conservation Planning projects to be successful, but which are widely missing in real world applications still (Huettmann 2007a, b, 2008a, b for projects and related data and publications). Table 16.7 shows known failures and a mis-use of Strategic Conservation Planning.
In a way, I hope, machine learning, e.g. for Strategic Conservation Planning, can at least help to strike that balance better reaching a steady-state (Daly and Farley 2011). We can actually afford to spend some energy, as long it is sustainable and not excessive, on machine learning with Strategic Conservation Planning and for good decision-making reaching sustainability on a global level. As a matter of fact, if we have any energy, or effort for that matter handy, it should be invested into great decision-making, achieved with machine learning-aided Strategic Conservation Planning making good use of those techniques available to mankind. This can eventually lead to a global society respecting ‘mother earth’. For global sustainability to become real, wider questions come to play, including universe ones, belief systems, spirituality, governance structures, distribution of wealth and the balance of life (Weaver 1996; Stiglitz 2003). But one way or another, machine learning is available and involved by now, and all one can ask for then is to make good and best-suitable use of this method; all done with good ethics and outcome for the wider public, global good. Anything that is not destructive science would be progress in that regard. Now, who would not agree?
Notes
- 1.
It should be emphasized here that many other methods exist to overcome data gaps, including data cloning to stabilize models on poor data or to extend the data (Lele et al. 2007; Jiao et al. <CitationRef CitationID="CR38" >2016</Citation Ref> for machine learning application). The other, and equally exciting approach for overcoming data gaps is to explain why data gaps actually occur (forensics), often based on ‘common sense’. It tends to work nicely because most data gaps have a reason for their existence! For instance, some field research data gaps are due to bad weather (rain), or inaccessibility of steep slope elevations. Those factors, data gaps, can then serve as explanations for the absence of certain events in the data. Tree-based models, and specifically the work from Friedman (2002), make use of such approaches with good success.
References
Alexander JC (2013) The dark side of modernity. Polity Press
Alidina HM, Fisher D, Stienback C, Ferdana Z, Lombana A, Huettmann F (2008) Assessing and managing data. In: Ardron J, Possingham H, Klein C (eds) Marxan good practices handbook, Vancouver, pp 14–20 http://www.pacmara.org/
Amstrup SC, McDonald TL, Manly B (2006) Handbook of capture-recapture analysis. Princeton University Press
Ardron J, Possingham H, Klein C (eds) (2008) Marxan good practices handbook, Vancouver http://www.pacmara.org
Arnold TW (2010) Uninformative parameters and model selection using Akaike’s information criterion. J Wildl Manag 74:1175–1178
Aycrigg J, Beauvais G, Gotthardt T, Huettmann F, Pyare S, Andersen M, Keinath D, Lonneker J, Spathelf M, Walton K (2015) Novel approaches to modeling and mapping terrestrial vertebrate occurrence in the northwest and Alaska: an evaluation. Northwest Sci 89:355–381. https://doi.org/10.3955/046.089.0405
Baltensperger AP, Huettmann F (2015) Predicted shifts in small mammal distributions and biodiversity in the altered future environment of Alaska: an open access data and machine learning. PLoS One. https://doi.org/10.1371/journal.pone.0132054
Bandura A (2007) Impeding ecological sustainability through selective moral disengagement. Int J Innovat Sustain Dev 2:8–35
Beiring M (2013) Determination of valuable areas for migratory songbirds along the East-Asian Australasian Flyway (EEAF), and an approach for strategic conservation planning. Unpublished Master Thesis. University of Vienna, Austria
Bolen EG, Robinson W (2002) Wildlife ecology and management, Fifth edn. Pearson Publisher
Breiman L (2001) Statistical modeling: the two cultures (with comments and a rejoinder by the author). Stat Sci 16:199–231
Clark TW (2002) The policy process: a practical guide for natural resource professionals. Yale University Press
Cockburn A (2013) A colossal wreck. Verso Publishers, New York
Cushman S, Huettmann F (2010) Spatial complexity, informatics and wildlife conservation. Springer Tokyo, Japan, p 448
Czech B (2000) Shoveling fuel for a runaway train: errant economists, shameful spenders, and a plan to stop them all. University of California Press, Berkeley
Czech B, Krausman PR, Devers PK (2000) Economic associations among causes of species endangerment in the United States. Bioscience 50:593–601
Daly HE, Farley J (2011) Ecological economics: principles and applications. Island press, Washington, DC
Drew CA, Wiersma Y, Huettmann F (eds) (2011) Predictive species and habitat modeling in landscape ecology. Springer, New York
Elith J, Graham CH, Anderson RP, Dudík M, Ferrier S, Guisan A, Hijmans RJ, Huettmann F, Leathwick JR, Lehmann A, Li J, Lohmann LG, Loiselle BA, Manion G, Moritz C, Nakamura M, Nakazawa Y, McC J, Overton M, Townsend Peterson A, Sj P, Richardson K, Scachetti-Pereira, Schapire RE, Soberón J, Williams S, Wisz MS, Zimmermann NA (2006) Novel methods improve prediction of species’ distributions from occurrence data. Ecography 29:129–151
Evans JS, Schill S, Raber GT (2015) Chapter 26: A systematic framework for spatial conservation planning and ecological priority design: an example from St. Lucia, Eastern Caribbean. In: Huettmann F (ed) Conservation of Central American biodiversity. Springer, New York, pp 603–623
Fernandes L, Day J, Lewis A, Slegers S, Kerrigan B, Breen D, Cameron D, Jago B et al (2005) Establishing representative no-take areas in the great barrier reef: large-scale implementation of theory on marine protected areas. Conserv Biol 19:1733–1744
Fernandez-Delgado M, Cernadas E, Barrow S, Amorim D (2014) Do we need hundreds of classifiers to solve real world classification problems? J Mach Learn Res 15:3133–3181
Flannery T (2002) The future eaters: an ecological history of the Australasian lands and its people. Grove Press, New York
Forman RTT (1995) Land mosaics: the ecology of landscapes and regions. Cambridge University Press
Franklin J, Miller JA (2009) Mapping species distributions: spatial inference and predictions. Cambridge University Press, USA
Gallus J (2010) Grizzly Bear manifesto. Rocky Mountain Books, Toronto
Gonzales EK, Arcese P, Schulz R, Bunnell FL (2003) Strategic reserve design in the central coast of British Columbia: integrating ecological and industrial goals. Can J For Res 33:2129–2140 https://doi.org/10.1139/x03-133
Groube L, Chappell J, Muke J, Price D (1986) A 40,000 year-old human occupation site at Huon Peninsula, Papua New Guinea. Nature 324:453–455. https://doi.org/10.1038/324453a0
Guisan A, Thuiller W (2005) Predicting species distribution: offering more than simple habitat models. Ecology Letters 8:993–1009. https://doi.org/10.1111/j.1461-0248.2005.00792.x
Guisan A, Zimmermann NE (2000) Predictive habitat distribution models in ecology. Ecol Model 135:147–186
Huettmann F (2007a) Modern adaptive management: adding digital opportunities towards a sustainable world with new values. Forum Public Policy: Climate Change Sustain Dev 3:337–342
Huettmann F (2007b) The digital teaching legacy of the International Polar Year (IPY): details of a present to the global village for achieving sustainability. In: Tjoa M, Wagner RR (eds) Proceedings 18th international workshop on Database and Expert Systems Applications (DEXA) 3–7 September 2007, Regensburg, Germany. IEEE Computer Society, Los Alamitos, pp 673–677
Huettmann F (2008a) Appendix II: literature & references. In: Ardron J, Possingham H, Klein C (eds) Marxan good practices handbook. Vancouver, Canada, pp 139–153 http://www.pacmara.org/
Huettmann F (2008b) Marine conservation and sustainability of the sea of Okhotsk in the Russian far east: an overview of cumulative impacts, compiled public data, and a proposal for a UNESCO World Heritage Site. In: Nijhoff M (ed) Ocean year book, vol 22. Halifax, Canada, pp 353–374
Huettmann F (2011) From Europe to North America into the world and atmosphere: a short review of global footprints and their impacts and predictions. The Environmentalist. https://doi.org/10.1007/s10669-011-9338-5
Huettmann F (2012) Protection of the three poles. Springer, New York
Huettmann F, Hazlett S (2010) Changing the arctic: adding immediate protection to the equation. Alaska Park Science, Fairbanks, pp 118–121
Huettmann F, Franklin SE, Stenhouse GB (2005) Predictive spatial modeling of landscape change in the Foothills Model Forest. For Chron 81:1–13
Huettmann F, Riehl T, Meissner K (2016) Paradise lost already? A naturalist interpretation of the pelagic avian and marine mammal detection database of the IceAGE cruise off Iceland and Faroe Islands in fall 2011. Environ Syst Decis. https://doi.org/10.1007/s10669-015-9583-0
Huettmann F, Magnuson EE, Hueffer K (2017) Ecological niche modeling of rabies in the changing Arctic of Alaska. Acta Vet Scand 201(759):18–31. https://doi.org/10.1186/s13028-017-0285-0
Jiao S, Huettmann F, Guoc Y, Li X, Ouyang Y (2016) Advanced long-term bird banding and climate data mining in spring confirm passerine population declines for the Northeast Chinese-Russian flyway. Glob Planet Chang. https://doi.org/10.1016/j.gloplacha.2016.06.015
Kandel K, Huettmann F, Suwal MK, Regmi GR, Nijman V, Nekaris KAI, Lama ST, Thapa A, Sharma HP, Subedi TR (2015) Rapid multi-nation distribution assessment of a charismatic conservation species using open access ensemble model GIS predictions: Red panda (Ailurus fulgens) in the Hindu-Kush Himalaya region. Biol Conserv 181:150–161
Lehtomäki J, Moilanen A (2013) Methods and workflow for spatial conservation prioritization using Zonation. Environ Model Softw 47:128–137
Mace G, Cramer W, Diaz S, Faith DP, Larigauderie A, Le Prestre P, Palmer M, Perrings C, Scholes RJ, Walpole M, Walter BA, Watson JEM, Mooney HA (2010) Biodiversity targets after 2010. Environ Sustain 2:3–8
Martin T, Smith J, Royle K, Huettmann F (2008) Is Marxan the right tool? In: Ardron J, Possingham H, Klein C (eds) Marxan good practices handbook. PACMARA, Vancouver, pp 58–74 http://www.pacmara.org/
Moilanen A, Wilson KA, Possingham H (2009) Spatial conservation prioritization: quantitative methods and computational tools, First edn. Oxford University Press
Murphy K, Huettmann F, Fresco N, Morton J (2010). Connecting Alaska landscapes into the future. U.S. Fish and Wildlife Service, and the University of Alaska.Fairbanks, Alaska. http://www.snap.uaf.edu/downloads/connecting-alaska-landscapes-future
Nielsen SJ, Stenhouse GB, Beyer HL, Huettmann F, Boyce MS (2008) Can natural disturbance-based forestry rescue a declining population of grizzly bears? Biol Conserv 141:2193–2207
Organ, JF, Geist V, Mahoney SP, Williams S, Krausman PR, Batcheller GR, Decker TA, Carmichael R, Nanjappa P, Regan, R, Medellin RA, Cantu R, McCabe RE, Craven S, Vecellio GM, Decker DJ (2012) The North American model of wildlife conservation. The Wildlife Society technical review 12-04. The Wildlife Society, Bethesda, Maryland, USA
Ostrom E (2015) Governing the commons: the evolution of institutions for collective action (Political Economy of Institutions and Decisions). Cambridge University Press, New York
Paehlke RC (2004) Democracy’s dilemma: environment, social equity, and the global economy. MIT Press, Cambridge
Perera AH, Drew AC, Johnson C (eds) (2012) Expert knowledge and its application in landscape ecology. Springer, New York
Peterson GD, Cumming GS, Carpenter SR (2003) Scenario planning: a tool for conservation in an uncertain world. Conserv Biol 17:358–366
Primack R (2016) Essentials of conservation biology, Sixth edn. Sinauer Press, Boston
Proctor MFC, Servheen M, Miller SD, Kasworm WF, Wakkinen WL (2004) A comparative analysis of management options for grizzly bear conservation in the U.S.–Canada trans-border area. Ursus 15:145–160
Pulliam HR (1988) Sources, sinks, and population regulation. Am Nat 132:652–661. https://doi.org/10.1086/284880
Rich B (1994) Mortgaging the earth: the world bank, environmental impoverishment and the crisis of Development. Island Press
Rockström J, Steffen W, Noone K, Persson Å, Chapin FS III, Lambin EF, Lenton TM, Scheffer M, Folke C, Schellnhuber HJ, Nykvist B, de Wit CA, Hughes T, van der Leeuw S et al (2009) A safe operating space for humanity. Nature 461:472–475
Schapire RE (1990) The strength of weak learnability. In: Machine learning, vol 5. Kluwer Academic Publishers, Boston, pp 197–227. https://doi.org/10.1007/bf00116037
Scott M, Davis FW, McGhie RG, Wright RG, Groves C, Estes J (2001) Nature reserves: do they capture the full range of America’s biological diversity? Ecol Appl 11:999–1007
Semmler M (2010) Spatial and temporal prediction models of Alaska’s 11 species mega-predator community: towards a first state-wide ecological habitat, impact, and climate assessment. Unpublished M.Sc. thesis. Georg-August University Goettingen, Germany
Silva NJ (2012) The wildlife techniques manual: research & management, vol 2, Seventh edn. The Johns Hopkins University Press
Singleton PH, Gaines WL, Lehmkuhl JL (2004) Landscape permeability for grizzly bear movements in Washington and southwestern British Columbia. Ursus 15:90–103
Solovyev B, Spiridonov V, Onufrenya L, Belikov S, Chernova N, Dobrynin D, Gavrilo M, Glazov D, Yu K, Mukharamova S, Pantyulin A, Platonov N, Saveliev A, Stishov M, Tertitsk G (2017) Identifying a network of priority areas for conservation in the Arctic seas: practical lessons from Russia. Aquat Conserv Mar Freshwat Ecosyst 27(S1):30–51
Spiridonov V, Gavrilo M, Yu Krasnov, A. Makarov, N. Nikolaeva, L. Sergienko, A. Popov, and E. Krasnova (2012). Toward the new role of marine and coastal protected areas in the Arctic: the Russian case. in: Protection of the three poles, by F. Huettmann (Ed.). Springer, New York. pp 171–202
Spiridonov V, Solovyev B, Chuprina E, Pantyulin A, Sazonov A, Nedospasov A, Stepanova S, Belikov S, Chernova N, Gavrilo M, Glazov D, Yu K, Tertitsky G (2017) Importance of oceanographical background for a conservation priority areas network planned using MARXAN decision support tool in the Russian Arctic seas. Aquat Conserv Mar Freshwat Ecosyst 27(S1):52–64
Stiglitz JE (2003) Globalization and its discontent. Norton Paperback, New York
UNESCO Great Barrier Reef Marine Park Authority (1981) Nomination of the Great barrier reef by the Commonwealth of Australia for inclusion in the world heritage list: United Nations educational scientific and cultural organization Great barrier reef marine park authority, Townsville, Australia
Walters C (1986) Adaptive management of renewable resources. Blackburn Press
Walther BA, Boëte C, Binot A, By Y, Cappelle J, Carrique-Mas J, Chou M, Furey N, Kim S, Lajaunie C, Lek S, Méral P, Neang M, Tan B-H, Walton C, Morand S (2016) Biodiversity and health: lessons and recommendations from an interdisciplinary conference to advise Southeast Asian research, society and policy. Infect Genet Evol 40:29–46
Watts MI, Ball I, Stewart RS, Kleina CJ, Wilson K, Steinback C, Lourival R, Kircher L, Possingham HP (2009) Marxan with zones: software for optimal conservation based land- and sea-use zoning. Environ Model Softw 24:1513–1521
Weaver J (1996) Defending mother earth: native American perspectives on environmental justice. Orbis Books
Xu W, Viña A, Zengxiang Q, Zhiyun O, Liu J, Hui W (2014) Evaluating conservation effectiveness of nature reserves established for surrogate species: case of a Giant Panda nature reserve in Qinling Mountains, China. Chin Geogr Sci 24:60–70
Xuesong H, Guo Yu, Mi C, Huettmann F, Wen L (2017) Machine learning model analysis of breeding habitats for the Blacknecked Crane in Central Asian uplands under anthropogenic pressures. Scientific Reports 7, Article number: 6114 doi:10.1038/ s41598–017–06167-2. https://www.nature.com/articles/s41598-017-06167-2
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2018 Springer Nature Switzerland AG
About this chapter
Cite this chapter
Huettmann, F. (2018). Machine Learning for ‘Strategic Conservation and Planning’: Patterns, Applications, Thoughts and Urgently Needed Global Progress for Sustainability. In: Humphries, G., Magness, D., Huettmann, F. (eds) Machine Learning for Ecology and Sustainable Natural Resource Management. Springer, Cham. https://doi.org/10.1007/978-3-319-96978-7_16
Download citation
DOI: https://doi.org/10.1007/978-3-319-96978-7_16
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-96976-3
Online ISBN: 978-3-319-96978-7
eBook Packages: Biomedical and Life SciencesBiomedical and Life Sciences (R0)