Access the full text.
Sign up today, get DeepDyve free for 14 days.
W. Ricker (1977)
Computation and interpretation of biological statistics of fish populations
Mace Mace, Collar Collar, Gaston Gaston (2008)
Quantification of extinction risk: International Union for the Conservation of Nature's (IUCN) system for classifying threatened speciesConserv. Biol., 22
S. Field, A. Tyre, Niclas Jonzén, J. Rhodes, H. Possingham (2004)
Minimizing the cost of environmental management decisions by optimizing statistical thresholdsEcology Letters, 7
J. Rice, Èmilie Legacè (2007)
When control rules collide: a comparison of fisheries management reference points and IUCN criteria for assessing risk of extinctionIces Journal of Marine Science, 64
P. Neubauer, O. Jensen, J. Hutchings, J. Baum (2013)
Resilience and Recovery of Overexploited Marine PopulationsScience, 340
Korman Korman, Peterman Peterman, Walters Walters (1995)
Empirical and theoretical analyses of correction of time series bias in stock?recruitment relationships of sockeye salmonCan. J. Fish. Aquat. Sci., 52
M. Burgman (2005)
Risks and Decisions for Conservation and Environmental Management: Risk ranking
N. Dulvy, S. Jennings, S. Rogers, D. Maxwell (2006)
Threat and decline in fishes: an indicator of marine biodiversityCanadian Journal of Fisheries and Aquatic Sciences, 63
David Keith, Jeffrey Hutchings, Ray Hilborn (2012)
Population dynamics of marine fishes at low abundanceCanadian Journal of Fisheries and Aquatic Sciences, 69
Irvine Irvine, Gross Gross, Wood Wood, Holtby Holtby, Schubert Schubert, Amiro Amiro (2005)
Canada's species at risk act: An opportunity to protect ?endangered? salmonFisheries, 30
Shelton Shelton, Mangel Mangel (2011)
Fluctuations of fish populations and the magnifying effects of fishingProc. Natl. Acad. Sci. U. S. A., 108
R. Hilborn, C. Walters (2004)
Quantitative fisheries stock assessment: Choice, dynamics and uncertaintyReviews in Fish Biology and Fisheries, 2
H. Matsuda, Yasuto Takenaka, T. Yahara, Y. Uozumi (1998)
Extinction risk assessment of declining wild populations: The case of the southern Bluefin TunaResearches on Population Ecology, 40
William Irvin (1993)
Endangered species act.Science, 259 5094
Reynolds Reynolds, Dulvy Dulvy, Goodwin Goodwin, Hutchings Hutchings (2005)
Biology of extinction risk in marine fishesProc. R. Soc. Lond. B, 272
C. Minto, R. Myers, W. Blanchard (2008)
Survival variability and population density in fish populationsNature, 452
G. Mace, Elodie Hudson (1999)
Attitudes toward Sustainability and ExtinctionConservation Biology, 13
J. Reynolds, N. Dulvy, N. Goodwin, J. Hutchings (2005)
Biology of extinction risk in marine fishesProceedings of the Royal Society B: Biological Sciences, 272
A. Shelton, M. Mangel (2011)
Fluctuations of fish populations and the magnifying effects of fishingProceedings of the National Academy of Sciences, 108
B. Mapstone (1995)
Scalable Decision Rules for Environmental Impact Studies: Effect Size, Type I, and Type II ErrorsEcological Applications, 5
H. Wilson, B. Kendall, H. Possingham (2011)
Variability in Population Abundance and the Classification of Extinction RiskConservation Biology, 25
D. Pauly (1995)
Anecdotes and the shifting baseline syndrome of fisheries.Trends in ecology & evolution, 10 10
R. Peterman (1990)
Statistical Power Analysis can Improve Fisheries Research and ManagementCanadian Journal of Fisheries and Aquatic Sciences, 47
J. Korman, R. Peterman, C. Walters (1995)
Empirical and theoretical analyses of correction of time-series bias in stock-recruitment relationships of sockeye salmon (Oncorhynchus nerka)Canadian Journal of Fisheries and Aquatic Sciences, 52
Brendan Connors, Andrew Cooper, R. Peterman, N. Dulvy (2014)
The false classification of extinction risk in noisy environmentsProceedings of the Royal Society B: Biological Sciences, 281
E. Porszt, R. Peterman, N. Dulvy, A. Cooper, J. Irvine (2012)
Reliability of Indicators of Decline in AbundanceConservation Biology, 26
H. Powles, M. Bradford, R. Bradford, W. Doubleday, S. Innes, C. Levings (2000)
Assessing and protecting endangered marine speciesJournal of Materials Science, 57
J. Irvine, M. Gross, C. Wood, L. Holtby, Neil Schubert, P. Amiro (2005)
Canada's Species at Risk ActFisheries, 30
N. Dulvy, S. Jennings, N. Goodwin, A. Grant, J. Reynolds (2005)
Comparison of threat and exploitation status in North‐East Atlantic marine populationsJournal of Applied Ecology, 42
Powles Powles, Bradford Bradford, Bradford Bradford, Doubleday Doubleday, Innes Innes, Levings Levings (2000)
Assessing and protecting endangered marine speciesICES J. Mar. Sci., 57
H. Powles (2011)
Assessing Risk of Extinction of Marine Fishes in Canada—The COSEWIC ExperienceFisheries, 36
T. Regan, B. Taylor, G. Thompson, J. Cochrane, K. Ralls, M. Runge, R. Merrick (2013)
Testing Decision Rules for Categorizing Species’ Extinction Risk to Help Develop Quantitative Listing Criteria for the U.S. Endangered Species ActConservation Biology, 27
W. Krämer (2003)
Review of Modern applied statistics with S, 4th ed. by W.N. Venables and B.D. Ripley. Springer-Verlag 2002Computational Statistics & Data Analysis, 43
Rice Rice, Legacé Legacé (2007)
When control rules collide: a comparison of fisheries management reference points and International Union for Conservation of Nature (IUCN) criteria for assessing risk of extinctionICES J. Mar. Sci., 64
G. Mace, N. Collar, K. Gaston, C. Hilton‐Taylor, H. Akçakaya, N. Leader‐Williams, E. Milner‐Gulland, S. Stuart (2008)
Quantification of Extinction Risk: IUCN's System for Classifying Threatened SpeciesConservation Biology, 22
N. Dulvy, J. Ellis, N. Goodwin, A. Grant, J. Reynolds, S. Jennings (2004)
Methods of assessing extinction risk in marine fishes.Fish and Fisheries, 5
A. Punt (2000)
Extinction of marine renewable resources: a demographic analysisPopulation Ecology, 42
A. Vincent, Y. Mitcheson, S. Fowler, S. Lieberman (2014)
The role of CITES in the conservation of marine fishes subject to international tradeFish and Fisheries, 15
W. Venables, B. Ripley (2010)
Modern Applied Statistics with S
Introduction The IUCN Red List categories and criteria (IUCN ) are among the most widely used methods for classifying the extinction risk of various species, with more than 70,000 species assessed to date. However, the validity of such assessments is often called into question when the conservation imperative conflicts with a desire to continue exploiting species, such as marine fishes (Mace & Hudson ). Extinction risk criteria were originally developed for terrestrial species and the controversy over their application to marine fishes may have hindered aquatic conservation (e.g., Powles et al . ; Reynolds et al . ; Vincent et al . ). Some simulation analyses have shown that IUCN decline criteria may raise false alarms by overstating the risk of extinction of widely distributed marine fishes such as bluefin tuna (Matsuda et al . ; Rice & Legacé ). However, recent empirical analyses of the reliability of various risk criteria show that false alarms (the incorrect classification of a species as threatened when it is not) is quite low for many commercially exploited marine fishes and Pacific salmon (Dulvy et al . ; Porszt et al . ). Now the key challenge is to understand the causes of misclassification and to identify the most reliable indicators of risk, that is, those that have the highest probability of correctly reflecting the actual condition of the assessed species or population. One principal concern is that misclassification can arise from between‐year environmentally driven variability in productivity (process variation) and/or census variability arising from imprecise estimates of abundance (observation error; Wilson et al . ; Connors et al . ). Here, we build upon past studies and use simulations that take into account these two sources of variation while evaluating the reliability of various population‐decline metrics. We quantify and rank the relative reliability of various decline indicators based on their ability to correctly classify the overall long‐term abundance trajectory as decreasing or not. Such evaluations are particularly critical because incorrectly classifying a population as decreasing enough to be declared a concern (false positive [FP]; Table ) will divert limited conservation funds from other, more at‐risk populations. Similarly, incorrectly classifying a population as not declining, when it actually is declining (false negative [FN]; Table ), could lead to eventual loss of the population because remedial action is not taken. Categories of possible outcomes when an indicator's assessed current status of a population in a given year is compared to the overall long‐term status of the same population. Note that the probabilities for two cases, false positives ( α ) and false negatives (β), are sufficient to determine the probabilities for the other two cases; the true negative rate is 1 ‑ α , and the true positive rate is 1 ‑ β Indicator's assessed current status: Nondeclining Declining Overall long‐ Nondeclining True negative (TN) False positive (FP) term status: 1 − α α ( Type I error ) Declining False negative (FN) True positive (TP) β ( Type II error ) 1 − β We focus on sockeye salmon ( Oncorhynchus nerka ), because its population dynamics are characterized by substantial variability, yet are understood well enough to develop realistic simulation models. Furthermore, it is a widely distributed fish species of conservation concern in Canada (Irvine et al . ). Major decreases in these exploited sockeye populations can be ecologically, economically, and socially important, even if the probability of extinction is low (DFO ). Thus, instead of focusing on extinction risk per se, we examine the closely related question of whether an assessment of a population decline in a given period appropriately reflects that population's long‐term trend, including the period subsequent to that assessment. Our indicators can thus be viewed as reflecting one symptom of extinction risk (i.e., a continuing decrease in abundance) or other serious conservation situations (Mace et al . ). We extend previous simulation analyses (Punt ; Holt et al . ; Wilson et al . ) by (1) estimating rates of misclassification (FNs and FPs) for a broad range of thresholds of decline that could be required for classifying a population as being at risk, (2) examining 20 indicators of decreasing abundance and their associated methods of data analysis, and (3) exploring unequal weightings of types of classification errors. We found that in the presence of even small process variation or observation error, indicators of decline in population abundance that use all of the available data are more reliable than other indicators, including the widely used indicator of recent rate of decline in the last 10 years or three generations (IUCN ). We also show that the ranking of decline indicators can be affected by how decision makers weight the relative importance of avoiding FPs compared to FNs. The wider application of our approach can help decision makers to more effectively use the restricted funds available for conservation actions. Methods Stochastic population dynamics model We used a stochastic model to simulate the population dynamics of our case example, semelparous sockeye salmon populations in the Fraser River that have a 4‐year life cycle: S t + 4 = a S t e − b S t + ɛ t , where S t is spawner abundance in year t . This spawner‐to‐spawner model is analogous to a Ricker ( ) spawner‐to‐recruit model, but here the a parameter of Equation encompasses biological productivity from spawner‐to‐adult recruits prior to the onset of fishing, as well as subsequent removals by fishing and in‐river pre‐spawning mortality (online Supporting Information [SI], part A). The density‐dependent parameter b was set to 1/100,000 (an arbitrary scalar of unfished equilibrium abundance). Process variation (between‐year variability in productivity) was represented by assuming that ε t in Equation is a lag‐1‐year autocorrelated process: ɛ t = Φ ɛ t − 1 + u t , where the latter error term u t is normally distributed, ∼N (0, σ u 2 ) (Hilborn & Walters ). Process variation can differ among sockeye salmon populations, so in separate analyses we examined levels of σ u 2 of 0.01, 0.05, 0.1, 0.3, and 0.5. For temporal autocorrelation in that process variation, we examined values of Φ (the autocorrelation coefficient) of ‑0.5, ‑0.25, 0, 0.25, 0.5, and 0.75. These ranges for these two parameters encompassed most reported estimates for North American sockeye salmon populations (Korman et al . ) as well as other animal species (SI part A). We also simulated observation error: S o b s e r v e d = S t + 4 e v t where v t ∼ N (0, σ v 2 ), which reflects the interannual variability in observed abundance due to sampling error (Hilborn & Walters ). The σ v 2 values of 0.0, 0.05, 0.1, 0.3, and 0.5 that we explored included the estimated range of observation error for Pacific salmon (Hilborn & Walters ) and other animal species (SI part A). We simulated population dynamics for 76 years, which consisted of a 12‐year initialization period, a 52‐year (13‐generation) “evaluation period,” and finally a “subsequent period” of 12 years (three‐generations; Figure a). We ran 500 Monte Carlo trials for each scenario; a scenario consisted of one combination of the magnitude of process variation ( σ u 2 ) and observation error ( σ v 2 ). Each population (i.e., each Monte Carlo trial) in each scenario was randomly assigned a productivity parameter value (the a in Equation drawn from ∼ N (1, 0.3), representing differences in productivity among different populations (some decreasing and some not due to local variables such as fishing rate, habitat quality, and environmental conditions; SI part A). An example of (a) how a decline in population abundance would be assessed using Indicator 17 (percent decline from the geometric mean abundance of the first 4‐year generation in the evaluation period (years 13–16) to abundance in the current year of assessment—year 31 here), for one example Monte Carlo trial output from the simulation model, and (b) how the overall long‐term status would be assessed for the same Monte Carlo trial—the decline in generational mean abundance, as estimated from regression, between the last generation of the initialization period (years 9–12) and the last generation of the subsequent period, also estimated from regression (years 73–76). (c) Example receiver operating characteristic (ROC) curves for two hypothetical indicators created by plotting true positive and false positive rates (Table ) for each of 101 threshold values, ranging from 0% to 100%, for classifying a population as declining in abundance. The area under the curve (AUC) is also shown, including for the straight‐line case for which the indicator would be no better than doing a coin toss. Estimation of indicators during the “evaluation period” We used 20 indicators of time trends in adult abundance to assess whether our populations were declining or nondeclining during the “evaluation period” (Table and SI part B). These indicators consisted of different combinations of three components. First was the time frame examined, either (1) the most recent three generations (i.e., 12 years in our case) in the evaluation period prior to the status‐assessment year, (2) the entire historical evaluation period prior to the assessment year, or (3) the time series since maximum abundance, regardless of where it occurred in the evaluation period (Table ). The second component was the data‐handling procedure (smoothed vs. unsmoothed data, and raw data versus log e ‐transformed data versus means of 4‐year generations; Table ). Third, changes in abundance over time were measured using either a linear regression (used to calculate either a rate or extent of change), or a difference in mean abundance between two periods. A summary of the 20 indicators of population decline evaluated here for assessing status during the simulated evaluation period (based on indicators suggested by IUCN 2013, Mace et al . ; Holt et al . , among others). Further details are in online Supporting Information part B Indicator Periods Smoothed Transformation Change in abundance 1 Recent 3 gen. No Log e Regression: rate 2 Recent 3 gen. Yes Log e Regression: rate 3 Hist.: first year No Log e Regression: extent 4 Hist.: first year Yes Log e Regression: extent 5 Hist.: cycle year No Log e Regression: extent 6 Hist.: cycle year Yes Log e Regression: extent 7 Max.: single year No Log e Regression: extent 8 Max.: single year Yes Log e Regression: extent 9 Hist.: first gen. No Mean: Sliding window Two means 10 Hist.: first gen. Yes Mean: Sliding window Two means 11 Hist.: first gen. No Mean: Generation block Two means 12 Hist.: first gen. Yes Mean: Generation block Two means 13 Max.: 3 gen. No Mean: Sliding window Two means 14 Max.: 3 gen. Yes Mean: Sliding window Two means 15 Max.: 3 gen. No Mean: Generation block Two means 16 Max.: 3 gen. Yes Mean: Generation block Two means 17 Hist.: first gen. No Raw Two means 18 Max.: 3 gen. No Raw Two means 19 Hist.: first year No Log e Regression: rate 20 Hist.: first year Yes Log e Regression: rate Periods – “Recent 3 gen.” is the percent decline over the most recent 3 generations (12 years) analogous to criterion A of IUCN (2001) and COSEWIC (2011); “Hist.” is the percent decline from some historical baseline since the beginning of the time series (either the first year of the evaluation period, the geometric mean abundance of the first 4‐year generation, or the first corresponding cycle year in the time series). See Supporting Information part A for explanation of “cycle” year; “Max.” is the percent decline from the maximum geometric mean abundance in the time series (either single year or three‐generation period) anywhere in the evaluation period of the time series. Smoothed – abundance estimates were either smoothed with a 4‐year (one generation) running mean or were left unsmoothed. Transformation – abundance estimates were either raw values, log e ‐transformed values, or the geometric mean abundance of 4‐year generations where the generations either moved 1 year at a time in sliding windows or in 4‐year‐generation blocks with no overlap of years (i.e., status only assessed every 4 years). “Raw” refers to the original, untransformed data. Change in abundance – “Regression” means that changes in abundance over the designated period were measured using robust linear regression to minimize the influence of outliers (Venables & Ripley ) and then calculated as either a rate or extent of change; “Two means” refers to changes in abundance that were measured as a percent decline from a mean abundance during some baseline period to either the current year's abundance or the mean abundance in the current generation being evaluated. For each relevant simulated year of the 13‐generation evaluation period, we determined for each of the 20 indicators whether the indicator classified the current status of the population as “declining” or “nondeclining”. To do this, in a set of separate calculations, we compared the estimated decrease in abundance against various thresholds that delineated “declining”; those thresholds ranged from 0% to 100% in increments of 1% (SI part B). That status classified during the evaluation period (e.g., Figure a) was then compared to the overall long‐term status, which included the subsequent period (Figure b) as described next. Overall long‐term population trend We focused on how well the 20 risk indicators reflected the actual overall true long‐term trend of each simulated population. To determine whether that long‐term trajectory of the population was declining, we estimated the trend in abundance from the end of the initialization period through the subsequent period, as shown in Figure b (also SI part C). For one set of simulations, if the population decreased more than 90%, which was the base‐case boundary condition indicating a conservation concern, then the overall long‐term status was classified as “declining”; otherwise it was classified as “nondeclining”. We also examined other commonly used boundary conditions of 50% and 70% decline (COSEWIC ; IUCN ). Measuring reliability We defined the reliability of an indicator as its probability of correctly distinguishing between declining and nondeclining populations. Specifically, for a given indicator of decline and a given scenario, we compared the indicator's assessed status of the population (declining or not) in each year of the evaluation period to the estimated long‐term status (as described above), resulting in either a true negative (TN), true positive (TP), false negative (FN) or false positive (FP) outcome (Table ). For each of the 20 indicators, these different categories of outcomes were tallied across all years and 500 trials. We then used a receiver operating characteristic (ROC) analysis to combine into a single metric for each decline indicator the TP and FP rates that were produced across a wide range of thresholds of decline in abundance for classifying population status, from 0% to 100% in 1% increments (Burgman ; Porszt et al . ; SI part D). The resulting area under the ROC curve (AUC) reflects the ability of a given indicator to correctly distinguish whether a population is declining; higher AUC values mean greater reliability (Hibberd & Cooper ). An AUC value of 0.97 reflects an indicator with a near‐perfect ability to distinguish a declining from a nondeclining population, whereas AUC = 0.51 (falling near the 1:1 line) indicates about a 50/50 chance of correctly making that distinction (Figure c). Different error weightings One limitation of an ROC analysis is that it inherently attributes equal weighting to FP and FN errors; hence, the importance of avoiding each type of error is the same. That equality is unrealistic in common situations where resource managers must make trade‐off decisions regarding conservation and resource‐use objectives (Peterman ; Mapstone ; Field et al . ; Dulvy et al . ). If one type of error is considered more serious than the other, then managers can specify a desired relative weighting or ratio of the error rates (for instance, that the more heavily weighted FN rate must be less than half the FP rate). Therefore, we also investigated the reliability of the 20 indicators of decline across a wide range of unequal error weightings. Results We first describe results for our base case, in which we assumed that the lag‐1 autocorrelation coefficient (Φ) in process variation was zero, before showing how our results are affected by a wide range of positive and negative Φ values. At extremely low levels of both process variation and observation error, all 20 indicators were almost equally reliable at discriminating between declining and nondeclining populations (AUC values > 0.9; Figure a). However, reliability decreased with greater process variation and observation error (compare AUCs of Figures b–d with Figure a). Indicators that used a historical baseline starting at the beginning of the time series (in particular indicators #4, 6, 9, and 20) consistently outperformed the other two types of indicators (e.g., Figures a–d; also see SI part E, Figures S1 and S2). Even with fairly low process variation or observation error, indicators calculated from a historical baseline were generally more reliable than either (1) indicators with a baseline based on the maximum abundance anywhere in the time series, or (2) indicators based on decline over the most recent 3 generations (Figure ). The latter group included indicator #2, which is analogous to the commonly used IUCN criterion A measured over only the past three generations. However, the IUCN criterion A metric performs far better when the full time series is used rather than just the past three generations (indicators #19 and #20). Ranking of indicators based on their reliability (probability of correctly reflecting the long‐term status of the population), as measured by the area under the ROC curve (AUC), for (a) low process variation, i.e., low environmentally driven variability in productivity ( σ u 2 = 0.01) and no observation error ( σ v 2 = 0); (b) high process variation ( σ u 2 = 0.5) and no observation error ( σ v 2 = 0); (c) low process variation ( σ u 2 = 0.01) and high observation error ( σ v 2 = 0.5); and (d) high process variation ( σ u 2 = 0.5) and high observation error ( σ v 2 = 0.5). Indicators based on the decline in the last three generations are black, indicators based on decline from a historical baseline are white, and indicators based on decline from maximum abundance are gray. Indicators varied in how robust they were to increases in process variation, σ u 2 , and observation error, σ v 2 . The AUC of the majority of indicators decreased by ∼10% when process variation increased from 0.01 to 0.5, whereas the AUC of indicators #1, 2, 5, 17, and 18 decreased at least twice as much as that (Figures a and b, also SI part E, Figures S1 and S2). The latter group of indicators included ones that either evaluated the rate of decline over the most recent 3 generations (#1 and 2), used raw abundance (#17 and 18), or estimated decline based on the first corresponding population‐cycle year (#5). By contrast, most indicators were less sensitive to increases in observation error, exhibiting only slight decreases in reliability (drop in AUC of ∼5%) when σ v 2 was increased from 0 to 0.5 (Figures a and c, also SI part E, Figures S1 and S2). Exceptions to this trend included indicators #1 and 2 (recent rate of decline), which were much more sensitive to observation error, with a drop in AUC of ∼20% when σ v 2 increased from 0 to 0.5 (Figures a and c, also SI part E, Figures S1 and S2). The AUCs of all indicators were relatively insensitive to changes in the boundary used to classify a population as declining over the long term, i.e., 50%, 70%, or 90% (AUCs changed by <4%), but again, indicators #1 and 2 changed the most (SI part E, Figure S3). The AUCs of all indicators declined with increases in the mean a parameter, but the rank order of the indicators was relatively insensitive to changes in mean a and b parameters (SI part E, Figure S4). Conservation objectives, error weightings, and error tolerance There is an inherent trade‐off between false positive rate (FPR) and false negative rate (FNR); both errors cannot be completely avoided (SI part E, Figure S5). If managers explicitly specify a desired relative error weighting, then the indicators can be optimized to target this weighting by adjusting the threshold that signals a declining population. The optimal threshold depends on the manager's relative preferences and the indicator being used (SI part E, Figure S5). The top‐ranked indicators #4, 6, 9, and 20, which used some historical baseline, were best across the majority of error weightings for different magnitudes of process variation (SI part E, Table S1a). Indicators 4 and 20 were best for almost all weightings and levels of observation error (SI part E, Table S1b). These results still held at high levels of temporal autocorrelation in process variation (SI part E, Tables S1c and d). For cases in which managers have an upper limit to the acceptable FP error rate, we estimated the FNR that would result from the trade‐off noted in SI part E, Figure S5. Figure shows the effect of variance in observation error (X‐axis of panels a and b) and process variation (panels c and d) on this nonlinear trade‐off between the acceptable FPR (Y‐axis) and the resulting FNR (contours). The high‐ranking indicator #4 reflects the change in abundance from the first generation in the time series (Figures a and c), whereas the commonly used indicator #2 measures decline over only the last three generations (Figures b and d). For example, if managers using that decline indicator #2 decided that it would be unacceptable to have greater than a 0.1 probability of a FP error (i.e., probability of incorrectly concluding that a population was declining), and if the population had a high level of observation‐error variance (say σ v 2 = 0.3), then they could expect a FNR of about 0.4, as shown on the contour plot (white dot in Figure b). This means that if the population is actually not declining, there will be a 10% chance of the indicator reporting that it is declining, but if the population is actually declining, there will be about a 40% chance of the indicator erroneously reporting that it is not declining. If managers want a lower FNR, then they have three choices: increase their tolerance for FPs, decrease observation error, or pick a different indicator. As an example of the latter, given the same situation, indicator #4, which reflects change in abundance since a historical baseline, would provide a much more desirable FNR of slightly less than 5% (white dot in Figure a). Regardless of the magnitude of observation error or process variation, indicator #4 has lower rates of FN errors than analogous situations for indicator #2 (compare contour values in Figure a with those in 3b; also compare 3c with 3d). See SI part E, Figure S6 for analogous contour plots for all 20 indicators. Plots of the minimum rate of false negative (FN) errors (probability values on contours) that can be obtained if the false positive (FP) rate is constrained to be below the maximum acceptable value specified on the Y‐axis. False negative rates are shown as a function of observation error, σ v 2 , (but with process variation set at σ u 2 = 0.01) for the top‐performing indicator in this study #4 (a) and commonly used decline indicator #2 (b). False negative rates are also shown for different levels of process variation, σ u 2 , (but with no observation error, σ v 2 = 0) for indicator #4 (c) and indicator #2 (d). The white data point in parts (a) and (b) indicates the false negative rate that will result from a maximum acceptable false‐positive rate of 0.1 and an observation error variance of 0.3. The AUCs of all indicators declined at high levels of temporal autocorrelation in process variation, Φ, with indicators #1 and 2 being the most affected (Figure , SI part E, Figure S7). The higher reliability of indicators that were based on some historical baseline also held across a wide range of autocorrelation (Figure , SI part E, Figure S7). The stronger the positive autocorrelation, the larger the advantage of indicators based on historical baselines compared to indicators based on just the last three generations (#1 and 2; Figure , SI part E, Figure S7). The addition of observation error, even at a very high level (σ v 2 = 0.5), did not have an interaction effect with increasing levels of temporal autocorrelation in process variation. Specifically, there was almost no difference in either the absolute or relative performance of the different indicators beyond what was found in the base case (i.e., without temporal autocorrelation in process variation) and in the case with autocorrelation with no observation error (SI part E, Figure S8). The reliability, or area under the curve (AUC), of the indicators across levels of temporal autocorrelation in process variation (Φ) ranging from −0.5 to 0.75 for (a) low process variation ( σ u 2 = 0.01) and (b) high process variation ( σ u 2 = 0.5), both with no observation error. Dark blue represents the highest probability of correctly classifying a population's abundance as decreasing or not, whereas dark red is the lowest probability. Results for other levels of process variation and observation error are found in the SI part E, Figures S7 and S8. Discussion A key problem in conservation biology is that the reliability of most extinction‐risk metrics or indicators is at present largely uncertain, i.e., their ability to correctly categorize a population as being at risk is unclear. However, this uncertainty could be reduced, and outcomes from conservation decision‐making could be improved, if more quantitative analyses were done like those presented here, which explicitly take into account uncertainties resulting from errors in abundance estimates and natural variability in ecological processes. Other scientists have also documented differences in reliability among indicators (Wilson et al . ; Porszt et al . ; Regan et al . ). We found (as did the empirical analysis of Porszt et al . ) that indicators that measure the extent of decline from a historical baseline tend to better reflect a population's long‐term status than either (1) a decline from some maximum abundance, or (2) the widely‐used IUCN rate of decline over the previous three generations. As well, some indicators are more predisposed to making certain types of errors, but generally process variation reduced reliability more than observation error, as was also shown by Wilson et al . ( ). For commonly used IUCN‐like indicators #1 and 2 (declines in abundance over the most recent three generations), we found that reliability decreased substantially more than the reliability of other indicators when process variation, observation error, or temporal autocorrelation in process variation increased. This poor performance is likely due to the assessment of status over a short period (the most recent 12 years vs. up to 52 years), which makes long‐term trends in abundance more difficult to identify amid the “noise” of process variation and observation error. In contrast, indicators based on some early historical baseline were the least sensitive to increases in either process variation or observation error. Although comparison with early portions of time series is recommended in some technical guidelines (Mace et al . ), many schemes emphasize recent rates of decline. There are two ways to calculate the IUCN criterion A: using only the most recent three generations of data or using all available data. IUCN recommends the use of all data only when “populations fluctuate wildly or oscillate with periods longer than the generation time” (IUCN , page 26). However, here we show that the full span of data should be used not just for these special cases, but whenever there is process or observation error, i.e., for any real time series. We found that indicators that reflect the rate of decline over a three‐generation time span can be as reliable as the best‐performing indicators, but only when that rate is calculated using all available data and applied to those three most recent generation spans. Our findings could easily be incorporated into technical guidance for use in IUCN and other related conservation assessments. For two other reasons, we recommend using all data from a time series, not just the most recent data. First, indicators of the recent rate of decline ignore the elevated extinction risks when recent abundances are relatively stable but are much lower than in an earlier period (the shifting baseline syndrome of Pauly ). Second, emerging evidence suggests that for heavily exploited marine fishes, process variation may increase at low‐population densities owing to truncated age structures (Minto et al . ; Shelton & Mangel ), reducing recovery rates and elevating risk (Keith & Hutchings ; Neubauer et al . ). Our model simulated semelparous sockeye salmon populations. However, we encourage other researchers to challenge assumptions that their current indicators reliably reflect actual population trends, and estimate that reliability for their own populations, especially those with different life‐history traits or population dynamics (Dulvy et al . ). Those future analyses (either empirical such as Porszt et al . or simulations like Wilson et al . and here) should explicitly take into account interannual variation in biological processes (and associated temporal autocorrelation), as well as errors in estimates of abundance, because these sources of variation can differentially affect the reliability of various indicators. People making management policies and conservation decisions should request that such tests of reliability be conducted before any indicators are used as input to risk assessments (Porszt et al . ). Decision makers must also consider the relative importance, or severity of consequences, of FPs and FNs when ranking indicators. For instance, in fisheries management, the economic and social costs associated with FNs can often be as serious as FPs (Peterman ; Mapstone ; Field et al . ; Dulvy et al . ). Different stakeholders will often have different error weighting preferences, in part because the costs of these errors might be borne by different groups (Peterman ), which can present challenges in selecting error weightings. Despite the consistently high rank of certain indicators, we also showed that in some cases, the ranking of indicators can be influenced by the relative magnitude of weightings placed on the two types of errors, as well as a stated maximum acceptable error rate. Decision makers should therefore be explicit and transparent about those weightings and acceptable error rates, and mindful of which stakeholders will be affected by these (sometimes implicit) decisions. Standard methods of risk assessment and decision analysis can further help managers choose appropriate indicators of population risk (Burgman ). Given the prominent role that IUCN and other indicators of conservation concern play in species status assessments worldwide, our results should prompt scientists who develop and evaluate extinction‐risk criteria to use simulation and empirical analyses to understand the performance and reliability of any proposed criteria. We found that use of all available data, rather than the common practice of using just the most recent three generations, consistently increased reliability of risk assessments. Quantitative analyses such as ours, which can estimate reliability of indicators, are a relatively inexpensive way to ensure that management agencies make the best decisions about their use of limited conservation funds. Acknowledgments We are grateful for support for this research provided by Simon Fraser University, the Natural Sciences and Engineering Research Council of Canada, and the Canada Research Chairs Program, Ottawa, Canada. We also thank Brendan Connors and John Reynolds for their discussions and suggestions on a draft manuscript. N.K.D. acknowledges the National Center for Ecological Analysis and Synthesis (NCEAS), Santa Barbara, CA, USA Working Group on “Red flags and species endangerment” for helpful discussion.
Conservation Letters – Wiley
Published: Mar 1, 2015
You can share this free article with as many people as you like with the url below! We hope you enjoy this feature!
Read and print from thousands of top scholarly journals.
Already have an account? Log in
Bookmark this article. You can see your Bookmarks on your DeepDyve Library.
To save an article, log in first, or sign up for a DeepDyve account if you don’t already have one.
Copy and paste the desired citation format or use the link below to download a file formatted for EndNote
Access the full text.
Sign up today, get DeepDyve free for 14 days.
All DeepDyve websites use cookies to improve your online experience. They were placed on your computer when you launched this website. You can change your cookie settings through your browser.