SEARCH

SEARCH BY CITATION

Keywords:

  • Abundance;
  • batch mark;
  • mark–recapture;
  • open population

Abstract

  1. Top of page
  2. Abstract
  3. Introduction
  4. Materials and Methods
  5. Results
  6. Discussion
  7. Acknowledgements
  8. Conflict of Interest
  9. References
  10. Supporting Information

Little attention has been paid to the use of multi-sample batch-marking studies, as it is generally assumed that an individual's capture history is necessary for fully efficient estimates. However, recently, Huggins et al. (2010) present a pseudo-likelihood for a multi-sample batch-marking study where they used estimating equations to solve for survival and capture probabilities and then derived abundance estimates using a Horvitz–Thompson-type estimator. We have developed and maximized the likelihood for batch-marking studies. We use data simulated from a Jolly–Seber-type study and convert this to what would have been obtained from an extended batch-marking study. We compare our abundance estimates obtained from the Crosbie–Manly–Arnason–Schwarz (CMAS) model with those of the extended batch-marking model to determine the efficiency of collecting and analyzing batch-marking data. We found that estimates of abundance were similar for all three estimators: CMAS, Huggins, and our likelihood. Gains are made when using unique identifiers and employing the CMAS model in terms of precision; however, the likelihood typically had lower mean square error than the pseudo-likelihood method of Huggins et al. (2010). When faced with designing a batch-marking study, researchers can be confident in obtaining unbiased abundance estimators. Furthermore, they can design studies in order to reduce mean square error by manipulating capture probabilities and sample size.


Introduction

  1. Top of page
  2. Abstract
  3. Introduction
  4. Materials and Methods
  5. Results
  6. Discussion
  7. Acknowledgements
  8. Conflict of Interest
  9. References
  10. Supporting Information

Batch-marking experiments have largely been neglected by statistical ecologists, as they are deemed inferior and to be avoided (Pollock 1981; Pollock and Mann 1983). However, biologists still use batch-marking for various purposes, and for some studies, they may be the only option available (e.g., insects, juvenile fish).

There are other types of batch-marking studies that differ in design from the one we study here. For example, Measey et al. (2003) performed a three-sample batch-marking experiment on caecilians where individuals were given a batch mark on the first occasion and on the second occasion, a subsample was given a secondary mark. Both marked and unmarked captured individuals were recorded at each sample time. Because an individual's capture history can be deduced when a different batch mark is applied on each sampling occasion, a Jolly–Seber type model can be fitted to analyze these data (Jolly 1965; Seber 1965). However, the disadvantage of the design is that there is a physical limitation to how many marks can be applied to an individual and this would vary by both species and mark type.

Frequently, batch marks are used to study movement of individuals between locations. For example, Roberts and Angermeier (2007) studied the movements of three fish species in the South Fork of the Roanoke River, Virginia, using a two-sample study. Here, they constructed movement corridors with favorable pool characteristics between suitable habitats and compared movement rates in corridors with unfavorable characteristics. Captured fish were given a mark that was a randomly assigned color and body location. Recaptured individuals were counted, and movement rates were estimated.

Skalski et al. (2009) review several batch-marking designs and marking methods for very small fish. However, most of these result in complete capture history information and are for two or three sampling occasion designs.

Arguments against using batch marks are based on the lack of individual capture histories. For example, if a marked individual is captured at sample time three, it is not known whether this individual was one of the marked individuals captured at sample time two or not. In addition, batch-marking experiments do not allow for adequate testing of model assumptions (Williams et al. 2002; p. 312).

We motivate this work with the data found in Huggins et al. (2010). They describe a study of oriental weatherloach (Misgurnus anguillicaudatu), which is a freshwater fish native to Eurasia and Northern Africa. It was brought to Australia for use in aquaria but was accidentally released, and the aim of the study was to investigate activity patterns of the wild populations in the inland waters.

Huggins et al. (2010) provide a pseudo-likelihood method for analyzing an extended batch-marking study. They caution that the likelihood is intractable as the number of marked individuals alive at any sample time is unknown. They condition on released individuals to develop estimating equations and obtain capture and survival probability estimates. Then, they use a Horvitz–Thompson-type estimator to estimate population size at each time point after obtaining capture probability estimates. Standard errors are obtained by first using a sandwich estimator for the variance of the model parameters (Freedman 2012) and then using the delta method to obtain estimated standard errors for population size.

We develop the batch-marking likelihood conditional on release (rather than the pseudo-likelihood), followed by a Horvitz–Thompson-like estimator for abundance. Although theoretically the likelihood can be maximized, it involves nested summations, resulting in a large number of computations, but the calculations can be run in parallel when a multiprocessor computer, a cluster or a grid is available. For this article, we investigate the use of extended batch-marking data in comparison with the Crosbie–Manly–Arnason–Schwarz (CMAS) model (Schwarz and Arnason 1996) to study the loss in estimation precision when one does not have information on individual encounter histories for a seven sampling occasion simulation experiment under various parameter values.

Materials and Methods

  1. Top of page
  2. Abstract
  3. Introduction
  4. Materials and Methods
  5. Results
  6. Discussion
  7. Acknowledgements
  8. Conflict of Interest
  9. References
  10. Supporting Information

An extended batch-marking study is one where individuals captured at the first sample time are all given the same nonunique type of tag (e.g., blue in color). At subsequent sample times, individuals captured with tags are counted and unmarked individuals are given a different color batch mark resulting in an independent cohort. Table 1 provides an example of generated data from a four sampling occasion extended batch-marking experiment. New marks are not given to marked individuals, and thus individual capture histories cannot be obtained. For example, it is not known whether the nine blue-tagged individuals at sample time three are a subset of the 16 found at time two. Note the similarity to the m-array notation for Cormack–Jolly–Seber data (see Williams et al. 2002; p. 419).

Table 1. Example data for the extended bmarking design. The number of individuals marked with a particular tag color is found on the diagonal, while the number of recaptures is on the off diagonal.
ReleaseColorOccasion
1234
1Blue2116911
2Green 221512
3Orange  174
4Red   6

The assumptions we make are similar to other open population capture–recapture models namely:

  • All individuals behave independently.
  • All individuals have the same probability of capture at sample time j; j = 1, 2, …, k.
  • All individuals have the same probability of survival between sample times j and j + 1; j = 1, 2, …, k − 1.
  • Individuals do not lose their tags.

Below we detail notation used in the model development.

Statistics or indices

  • i index for release occasion (or colour of tag).
  • j index for recapture occasion.
  • k the number of sampling occasions.
  • rij the number of individuals tagged and released at time i and recaptured at time j, i = 1, 2, …, k − 1; j = i + 1, …, k.
  • Ri the number of individuals released at time i; i = 1, 2, …, k.

Latent variables

  • Mij the number of marked individuals released at sample time i, alive and available for capture at sample time j; j = i ,…, k. Note that Mii = Ri.
  • dij the number of deaths between sample times j and j + 1 from release group i; dij MijMi,j+1; i = 1, …, k; j = i, …, k – 1.

Parameters

  • ϕij the probability of survival for individuals from release group i between times j and j + 1; j = 1, 2, …, k − 1.
  • pij the probability of capture for individuals from release group i at time j; j = 2, 3, …, k.

We develop the likelihood by first looking at the joint distribution of the recaptures rij and deaths dij given the releases Ri. We then obtain the marginal distribution of the recaptures given releases by summing over all possible values of the deaths. The likelihood can be written as

  • display math(1)

Conditional on release, we model the recaptures as independent given deaths inline image and the deaths as dii, …,dik|RiMultinomial(Ri, πii, … ,πik) where inline image. We note that inline image and inline image where dik would be the individuals that were released at time i and are still alive after the last sample time k. These dik are convenient for modeling purposes. Thus, the likelihood becomes

  • display math(2)

Inference

The calculation of the likelihood involves nested summations for the latent dij variables which require high execution times if serially computed or cause the available RAM to be used up if fully vectorized. We developed parallel computer code to implement this model in MATLAB (MATLAB 2012) trading off CPU speed and memory that works for up to 11 sampling occasions. For experiments beyond 11 sampling occasions, we propose to use our likelihood up to the 11th sample time; and the pseudo-likelihood (Huggins et al. 2010) for occasions 12 through k. In the simulation studies, we first produce maximum likelihood estimates of the survival and capture probability parameters. Then, we derive a Horvitz–Thompson–type estimator for population size at each sample time (Nj) using the capture probability estimates and the number of individuals captured at each sampling time, inline image. Standard errors for the inline image are estimated from the estimated variances of rij and inline image using the delta method (see Huggins et al. 2010 for details).

Monte Carlo simulations

We simulated data from a k = 7 sample occasion Crosbie–Manly–Arnason–Schwarz model (Schwarz and Arnason 1996) with constant survival probabilities (ϕ = 0.2, 0.5, or 0.8), constant capture probabilities (p = 0.2, 0.5, or 0.8), and entry probabilities equal across time (1/k) with both a small superpopulation size (N = 200) and a larger superpopulation size (N = 1000). The superpopulation N is defined as the population that enters the population of interest at some point during the study. The computing time (on a dual quad core 2.53 GHz, 32 Gb RAM Linux server) was approximately 2 days to run 100 replications for N = 1000 but larger superpopulation size, more samples, or more replications would result in longer computing times. Parameter values were selected to obtain sparse-to-plentiful data by varying the probability of capture and survival.

For each set of parameter values, we simulated 100 CMAS datasets and collapsed these datasets into batch-marking data. We analyzed the individual capture history data using the CMAS model with constant parameters implemented in RMark (Laake 2013). The associated batch-marking data were analyzed using both the pseudo-likelihood (Huggins et al. 2010) and the likelihood with constant parameters (ϕ, p). For all methods of analyses, we estimated the survival and capture probabilities and obtained abundance estimates and estimated standard errors for each sampling time. The 100 dataset results are summarized using box plots for the estimated abundance, and estimated capture and survival probabilities. Root mean square errors (inline image) were calculated for the capture and survival probabilities. While standard errors were estimated, plots of these results are not included in the interest of space but are provided in the supplementary materials (see Supporting information).

Results

  1. Top of page
  2. Abstract
  3. Introduction
  4. Materials and Methods
  5. Results
  6. Discussion
  7. Acknowledgements
  8. Conflict of Interest
  9. References
  10. Supporting Information

Figures 1 and 2 provide results for the 9 simulation studies under varying parameter values for estimates of p and ϕ, respectively, for N = 200. For sparse data (p = 0.2 or p = 0.5, ϕ = 0.2), many of the simulations produced boundary estimates for p and occasionally ϕ (see Fig. 1), and the calculation of the standard errors then failed due to the Hessian being singular. These simulation failures are similar to what happens in the Cormack–Jolly–Seber model when analytical estimates of ϕ exceed 1 with sparse data and actual parameter values are close to 0 and 1. In these cases, the maximization function in MATLAB constrained estimates to be admissible, i.e. between 0 and 1 (inclusive). When the estimation of p was on a boundary, Nj was estimated at infinity (inline image) or ∑irij (inline image). Table 2 provides the number of simulations out of 100 that produced boundary estimates for p or ϕ for all three methods. Similar figures for N = 1000 are provided in the Supporting information. Results for the estimation of standard errors are based on those simulations that did not fail (see Supporting information).

Table 2. The number of simulations out of 100 that produced boundary estimates (0 or 1) for either parameter p or ϕ for the Crosbie–Manly–Arnason–Schwarz (CMAS), the likelihood (L), and the pseudo-likelihood (H; Huggins et al. 2010) methods under the 18 simulation scenarios.
N Method p ϕ
0.20.50.8
200CMAS0.29592
0.53200
0.83700
L0.27596
0.53220
0.836180
H0.276106
0.52620
0.831160
1000CMAS0.23200
0.5101
0.8100
L0.22700
0.5300
0.8920
H0.22400
0.5800
0.81410
image

Figure 1. Boxplots of capture probability estimates (inline image) from 100 simulated datasets, for the Crosbie–Manly–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) methods when parameter values are N = 200, and p = 0.2, 0.5, 0.8 for ϕ = 0.2 (top), ϕ = 0.5 (middle), and ϕ = 0.8 (bottom).

Download figure to PowerPoint

image

Figure 2. Boxplots of survival probability estimates (inline image) from 100 simulated datasets, for the Crosbie–Manly–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) methods when parameter values are N = 200, and ϕ = 0.2, 0.5, 0.8 for p = 0.2 (top), p = 0.5 (middle), and p = 0.8 (bottom).

Download figure to PowerPoint

The root mean square error (RMSE) for estimates of p and ϕ are given in Tables 3 and 4, respectively. As expected, we find that within a method, RMSE decreases as p and ϕ increase. Similarly, RMSE decreases with increased N. We also confirm that the CMAS method typically has lower RMSE than either of the batch marking methods and that the likelihood method typically has lower RMSE than the pseudo-likelihood method. Exceptions to this occur with sparse data when the estimates are not reliable.

Table 3. Root mean square error for estimates of p for the Crosbie–Manly–Arnason–Schwarz (CMAS), the likelihood (L), and the pseudo-likelihood (H; Huggins et al. 2010) methods under the 18 simulation scenarios.
N Method p ϕ
0.20.50.8
200CMAS0.20.7040.2820.062
0.50.3190.1000.046
0.80.1600.0650.028
L0.20.6160.2710.065
0.50.3300.1360.077
0.80.2150.1230.058
H0.20.6160.2960.073
0.50.3090.1510.079
0.80.2330.1250.068
1000CMAS0.20.4700.0660.026
0.50.1490.0360.019
0.80.0860.0270.011
L0.20.4480.0830.035
0.50.1850.0500.031
0.80.1370.0570.031
H0.20.4280.0870.035
0.50.2120.0530.032
0.80.1500.0610.034
Table 4. Root mean square error for estimates of ϕ for the Crosbie–Manly–Arnason–Schwarz (CMAS), the likelihood (L), and the pseudo-likelihood (H; Huggins et al. 2010) methods under the 18 simulation scenarios.
N Method p ϕ
0.20.50.8
200CMAS0.20.2240.1960.095
0.50.0870.0680.040
0.80.0410.0360.027
L0.20.2210.1970.106
0.50.1010.0830.058
0.80.0640.0530.035
H0.20.2150.2060.116
0.50.1120.0900.060
0.80.0750.0600.039
1000CMAS0.20.1210.0720.046
0.50.0380.0270.014
0.80.0210.0160.009
L0.20.1160.0930.067
0.50.0490.0350.021
0.80.0350.0250.016
H0.20.1160.0960.067
0.50.0530.0370.023
0.80.0380.0270.017

Under sparse data conditions (e.g., N = 200, ϕ = 0.8, p = 0.2), the average population size estimates are similar between the three methods; however, variability in estimates is higher for the likelihood and pseudo-likelihood methods as expected (Fig. 3; box plots for other sets of parameters are provided in Supporting information). For example, average population size estimates for time three were 71, 74, and 72 individuals for the CMAS, likelihood, and pseudo-likelihood, respectively, and the corresponding average standard error estimates were 18, 32, and 32 individuals. For higher quality data (e.g., N = 1000, p = 0.5, ϕ = 0.8), we found similar results. The CMAS model produces more precise estimates followed by the likelihood (see Supporting information for box plots of estimated standard errors). For example, the average population size estimate for sample time three was 348, 349, and 349 individuals for the CMAS, likelihood, and pseudo-likelihood method, respectively, with corresponding average estimated standard errors of 12, 29, and 29 individuals.

image

Figure 3. Boxplots of abundance estimates (inline image) for each sample time (k = 7) from 100 simulated datasets, for the Crosbie–Manly–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) methods when parameter values are N = 200, and ϕ = 0.8 for p = 0.2 (top), p = 0.5 (middle), and p = 0.8 (bottom). The long black horizontal lines show the expected population size at time j.

Download figure to PowerPoint

Discussion

  1. Top of page
  2. Abstract
  3. Introduction
  4. Materials and Methods
  5. Results
  6. Discussion
  7. Acknowledgements
  8. Conflict of Interest
  9. References
  10. Supporting Information

For an extended batch-marking study, using the likelihood provides more accurate estimates and lower standard errors than using the pseudo-likelihood method of Huggins et al. (2010). However, the computing power necessary to calculate the likelihood by summing over all possible values of deaths is prohibitive when sampling times go beyond k = 11 or if Ri is large. In these cases, the pseudo-likelihood method is computationally faster and provides unbiased estimates with similar precision to the likelihood approach. Ultimately, if full-capture histories are possible and available, then naturally the CMAS model outperforms both batch-marking models. With plentiful data (large numbers of recaptures), our model can have a relative efficiency of between about 30–40% compared with the CMAS model; thus, using the CMAS model has obvious gains in precision.

In many of the plots for inline image, the average population size increases over time. This is due to the models allowing for births/immigration into the population from the superpopulation (N). In these simulations, entry probabilities were equal across time and summed to one. Thus, with high survival rates, population size would naturally increase with time.

Practitioners who are confined to using batch marks should design their studies to have large sample sizes and high capture rates so as to minimize mean square error.

For future work, we will complete the model development by incorporating both tagged and untagged individuals at each sample time. We will also deal with issues such as goodness of fit, model selection, and parameter redundancy. With the many latent variables in the complete data, this model lends itself well to Bayesian methods where a state-space formulation is under development.

With permanent batch marks, tag loss would not be an issue. However, if injectable color tags are used for example, tag loss may bias parameter estimates. If it were possible to double tag individuals, an extended batch-marking model incorporating tag retention rates could be developed using methods similar to Cowen and Schwarz (2006). However, for those study species where double tagging is not possible (e.g., insects), separate experiments to estimate tag retention would have to be carried out and this auxiliary information could be used to adjust parameter estimates using methods similar to Arnason and Mills (1981).

Acknowledgements

  1. Top of page
  2. Abstract
  3. Introduction
  4. Materials and Methods
  5. Results
  6. Discussion
  7. Acknowledgements
  8. Conflict of Interest
  9. References
  10. Supporting Information

This work was initiated while LC was on study leave at the University of Kent supported by a University of Victoria Professional Development grant. Simulation studies using RMark were run on Westgrid/Compute Canada with assistance from Dr. Belaid Moa. This article was much improved by comments from the reviewers and the Associate Editor.

References

  1. Top of page
  2. Abstract
  3. Introduction
  4. Materials and Methods
  5. Results
  6. Discussion
  7. Acknowledgements
  8. Conflict of Interest
  9. References
  10. Supporting Information
  • Arnason, A., and K. H. Mills. 1981. Bias and loss of precision due to tag loss in Jolly-Seber estimates for mark-recapture experiments. Can. J. Fish. Aquat. Sci. 38:10771095.
  • Cowen, L., and C. J. Schwarz. 2006. The Jolly-Seber model with tag loss. Biometrics, 62:699705.
  • Freedman, D. A. 2012. On the so-called “Huber sandwich estimator" and “Robust standard errors". Am. Stat. 60:299302.
  • Huggins, R., Y. Wang, and J. Kearns. 2010. Analysis of an extended batch marking experiment using estimating equations. J. Agric. Biol. Environ. Stat. 15:279289.
  • Jolly, G. M. 1965. Explicit estimates from capture-recapture data with both death and immigration–Stochastic model. Biometrika, 52:225247.
  • Laake, J. L. 2013. RMark: An R Interface for Analysis of Capture–Recapture Data with MARK, AFSC Processed Rep 2013-01, 25 p. Alaska Fish. Sci. Cent., NOAA, Natl. Mar. Fish. Serv., 7600 Sand Point Way NE, Seattle WA 98115.
  • MATLAB 2012. Version 7.14.0.739 (R2012a). The MathWorks Inc., Natick, MA.
  • Measey, G. J., D. J. Gower, O. V. Oommen, and M. Wilkinson. 2003. A mark-recapture study of the caecilian amphibian Gegeneophis ramaswamii (Amphibia: Gymnophiona: Caeciliidae) in southern India. J. Zool. (Lond.), 261:129133.
  • Pollock, K. H. 1981. Capture–recapture models: a review of current methods, assumptions and experimental design. Pp. 426435 in C. J. Ralph and J. M. Scott, eds. Estimating the Numbers of Terrestrial Birds, Studies in Avian Biology. Cooper Ornithological Society, Los Angeles.
  • Pollock, K. H., and R. H. K. Mann. 1983. Use of an age-dependent mark-recapture model fisheries research. Can. J. Fish. Aquat. Sci., 40:14491455.
  • Roberts, J. H., and P. L. Angermeier. 2007. Movement responses of stream fishes to introduced corridors of complex cover. Trans. Am. Fish. Soc. 136:971978.
  • Schwarz, C., and A. Arnason. 1996. A general methodology for the analysis of open-model capture recapture experiments. Biometrics, 52:860873.
  • Seber, G. A. F. 1965. A note on the multiple recapture census. Biometrika, 52:249259.
  • Skalski, J. R., R. A. Buchanan, and J. Griswold. 2009. Review of marking methods and release-recapture designs for estimating the survival of very small fish: Examples from the assessment of salmonid fry survival. Rev. Fish. Sci. 17:391401.
  • Williams, B. K., J. D. Nichols, and M. J. Conroy. 2002. Analysis and Management of Animal Populations. Academic Press, San Diego.

Supporting Information

  1. Top of page
  2. Abstract
  3. Introduction
  4. Materials and Methods
  5. Results
  6. Discussion
  7. Acknowledgements
  8. Conflict of Interest
  9. References
  10. Supporting Information
FilenameFormatSizeDescription
ece3899-sup-0001-FiguresS1-S13.pdfapplication/PDF7312K

Figure S1. Boxplots of capture probability estimates (inline image) of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 1000, and p = 0.2,0.5,0.8 for ϕ = 0.2 (top), and ϕ = 0.5 (middle), ϕ = 0.8 (bottom).

Figure S2. Boxplots of survival probability estimates inline image of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 1000, and ϕ = 0.2, 0.5, 0.8 for p = 0.2 (top), and p = 0.5 (middle), p = 0.8 (bottom).

Figure S3. Boxplots of abundance estimates inline image for each sample time (k = 7) of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 200, and ϕ = 0.2 for p = 0.2 (top), and p = 0.5 (middle), p = 0.8 (bottom). The long black horizontal lines show the expected population size at time j.

Figure S4. Boxplots of abundance estimates inline image for each sample time (k = 7) of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 200, and ϕ = 0.5 for p = 0.2 (top), and p = 0.5 (middle), p = 0.8 (bottom). The long black horizontal lines show the expected population size at time j.

Figure S5. Boxplots of abundance estimates inline image for each sample time (k = 7) of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 1000, and ϕ = 0.2 for p = 0.2 (top), and p = 0.5 (middle), p = 0.8 (bottom). The long black horizontal lines show the expected population size at time j.

Figure S6. Boxplots of abundance estimates inline image for each sample time (k = 7) of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 1000, and ϕ = 0.5 for p = 0.2 (top), and p = 0.5 (middle), p = 0.8 (bottom). The long black horizontal lines show the expected population size at time j.

Figure S7. Boxplots of abundance estimates inline image for each sample time (k = 7) of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 1000, and ϕ = 0.8 for p = 0.2 (top), and p = 0.5 (middle), p = 0.8 (bottom). The long black horizontal lines show the expected population size at time j.

Figure S8. Boxplots of estimated standard errors for the abundance estimates (inline image) for each sample time (k = 7) of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 200, and ϕ = 0.2 for p = 0.2 (top), and p = 0.5 (middle), p = 0.8 (bottom). Estimates from simulations that produced a singular Hessian were removed.

Figure S9. Boxplots of estimated standard errors for the abundance estimates (inline image) for each sample time (k = 7) of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 200, and ϕ = 0.5 for p = 0.2 (top), and p = 0.5 (middle), p = 0.8 (bottom). Estimates from simulations that produced a singular Hessian were removed.

Figure S10. Boxplots of estimated standard errors for the abundance estimates (inline image) for each sample time (k = 7) of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 200, and ϕ = 0.8 for p = 0.2 (top), and p = 0.5 (middle), p = 0.8 (bottom). Estimates from simulations that produced a singular Hessian were removed.

Figure S11. Boxplots of estimated standard errors for the abundance estimates (inline image) for each sample time (k = 7) of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 1000, and ϕ = 0.2 for p = 0.2 (top), and p = 0.5 (middle), p = 0.8 (bottom). Estimates from simulations that produced a singular Hessian were removed.

Figure S12. Boxplots of estimated standard errors for the abundance estimates (inline image) for each sample time (k = 7) of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 1000, and ϕ = 0.5 for p = 0.2 (top), and p = 0.5 (middle), p = 0.8 (bottom). Estimates from simulations that produced a singular Hessian were removed.

Figure S13. Boxplots of estimated standard errors for the abundance estimates (inline image) for each sample time (k = 7) of 100 simulated datasets, for the Crosbie–Manley–Arnason–Schwarz (C: red), the likelihood (L: green), and the pseudo-likelihood (H: yellow; Huggins et al. 2010) when parameters values are N = 1000, and ϕ = 0.8 for p = 0.2 (top), and p = 0.5 (middle), p = 0.8 (bottom). Estimates from simulations that produced a singular Hessian were removed.

Please note: Wiley Blackwell is not responsible for the content or functionality of any supporting information supplied by the authors. Any queries (other than missing content) should be directed to the corresponding author for the article.