References
Barr, D. J., Levy, R., Scheepers, C., & Tily, H. J. (2013). Random
effects structure for confirmatory hypothesis testing: Keep it maximal.
Journal of Memory and Language, 68(3), 255–278.
Bates, D., Mächler, M., Bolker, B. M., & Walker, S. C. (2015).
Fitting linear mixed-effects models using lme4. Journal of Statistical Software,
67(1), 1–48. https://doi.org/10.18637/jss.v067.i01
Bengtsson, H. (2021). A unifying framework for parallel and distributed
processing in R using futures. The R Journal,
13(2), 208–227.
Blischak, J. D., Davenport, E. R., & Wilson, G. (2016). A quick
introduction to version control with Git and
GitHub. PLOS Computational Biology,
12(1), e1004668. https://doi.org/10.1371/journal.pcbi.1004668
Bolker, B. M. (n.d.). GLMM FAQ: Mixed models in
R. Online reference. Retrieved https://bbolker.github.io/mixedmodels-misc/glmmFAQ.html
Boyd, S., & Vandenberghe, L. (2004). Convex optimization.
Cambridge University Press. https://web.stanford.edu/~boyd/cvxbook/
Brown, V. A. (2021). An introduction to linear mixed-effects modeling in
R. Advances in Methods and Practices in Psychological
Science, 4(1), 1–19. https://doi.org/10.1177/2515245920960351
Chacon, S., & Straub, B. (2014). Pro Git (2nd
ed.). Apress. https://git-scm.com/book
Dowle, M., & Srinivasan, A. (2021). data.table: Extension of data.frame. https://rdatatable.gitlab.io/data.table/
Efron, B. (1979). Bootstrap methods: Another look at the jackknife.
The Annals of Statistics, 7(1), 1–26.
Efron, B., & Tibshirani, R. (1986). Bootstrap methods for standard
errors, confidence intervals, and other measures of statistical
accuracy. Statistical Science, 1(1), 54–75. https://doi.org/10.1214/ss/1177013815
Efron, B., & Tibshirani, R. J. (1993). An introduction to the
bootstrap. Chapman; Hall/CRC.
Faraway, J. J. (2016). Extending the linear model with
R: Generalized linear, mixed effects and nonparametric
regression models (2nd ed.). Chapman; Hall/CRC.
Gelman, A., Hill, J., & Vehtari, A. (2020). Regression and other
stories. Cambridge University Press.
Gentle, J. E. (2024). Matrix algebra: Theory, computations, and
applications in statistics (3rd ed.). Springer. https://doi.org/10.1007/978-3-031-42144-0
Golub, G. H., & Van Loan, C. F. (2013). Matrix computations
(4th ed.). Johns Hopkins University Press.
Grolemund, G., & Wickham, H. (2017). R for data science
(1st ed.). O’Reilly Media. https://r4ds.had.co.nz/
Halko, N., Martinsson, P.-G., & Tropp, J. A. (2011). Finding
structure with randomness: Probabilistic algorithms for constructing
approximate matrix decompositions. SIAM Review, 53(2),
217–288.
Harrell, F. E. (2015). Regression modeling strategies (2nd
ed.). Springer.
Hastie, T., Tibshirani, R., & Wainwright, M. (2015). Statistical
learning with sparsity: The lasso and generalizations. Chapman;
Hall/CRC. https://web.stanford.edu/~hastie/StatLearnSparsity/
Healy, K. (2018). Data visualization: A practical introduction.
Princeton University Press. https://socviz.co
Hesterberg, T. C. (2015). What teachers should know about the bootstrap:
Resampling in the undergraduate statistics curriculum. The American
Statistician, 69(4), 371–386. https://doi.org/10.1080/00031305.2015.1089789
Lange, K. (2010). Numerical analysis for statisticians (2nd
ed.). Springer.
Legler, J., & Roback, P. (2019). Broadening your statistical
horizons: Generalized linear models and multilevel models. https://bookdown.org/roback/bookdown-bysh/
Matuschek, H., Kliegl, R., Vasishth, S., Baayen, H., & Bates, D.
(2017). Balancing Type I error and power in linear mixed
models. Journal of Memory and Language, 94, 305–315.
McCullagh, P., & Nelder, J. A. (1989). Generalized linear
models (2nd ed.). Chapman; Hall/CRC.
McCulloch, C. E., Searle, S. R., & Neuhaus, J. M. (2008).
Generalized, linear, and mixed models (2nd ed.). Wiley.
McElreath, R. (2020). Statistical rethinking: A bayesian course with
examples in R and Stan (2nd ed.).
Chapman; Hall/CRC. https://xcelab.net/rm/statistical-rethinking/
Morris, T. P., White, I. R., & Crowther, M. J. (2019). Using
simulation studies to evaluate statistical methods. Statistics in
Medicine, 38(11), 2074–2102. https://doi.org/10.1002/sim.8086
Nocedal, J., & Wright, S. J. (2006). Numerical optimization
(2nd ed.). Springer.
Pennsylvania State University Department of Statistics. (n.d.).
STAT 504: Introduction to GLMs.
Online course notes. Retrieved https://online.stat.psu.edu/stat504/lesson/6/6.1
Petersen, K. B., & Pedersen, M. S. (2012). The matrix
cookbook. https://www2.imm.dtu.dk/pubdb/edoc/imm3274.pdf
Rizzo, M. L. (2019). Statistical computing with R
(2nd ed.). Chapman; Hall/CRC.
Sievert, C. (2020). Interactive web-based data visualization with
R, plotly, and shiny. Chapman; Hall/CRC. https://plotly-r.com/
Strang, G. (2016). Introduction to linear algebra (5th ed.).
Wellesley-Cambridge Press. https://ocw.mit.edu/courses/18-06-linear-algebra-spring-2010/
Trefethen, L. N., & Bau III, D. (1997). Numerical linear
algebra. SIAM.
Tufte, E. R. (2001). The visual display of quantitative
information (2nd ed.). Graphics Press.
UCLA OARC Statistical Methods and Data Analytics. (n.d.). Regression
diagnostics. Online tutorial. Retrieved https://stats.oarc.ucla.edu/r/dae/regression-diagnostics/
University of Wisconsin-Madison Social Science Computing Cooperative.
(n.d.). Generalized linear models in R. Online
tutorial. Retrieved https://sscc.wisc.edu/sscc/pubs/glm-r/
Wickham, H. (2016). ggplot2: Elegant graphics for data
analysis. Springer-Verlag. https://ggplot2-book.org
Wickham, H. (2019). Advanced r (2nd ed.). Chapman; Hall/CRC. https://adv-r.hadley.nz
Wickham, H. (2021). Mastering Shiny. O’Reilly
Media. https://mastering-shiny.org/
Wickham, H., & Bryan, J. (2023). R packages (2nd ed.).
O’Reilly Media. https://r-pkgs.org
Wickham, H., Çetinkaya-Rundel, M., & Grolemund, G. (2023). R for
data science: Import, tidy, transform, visualize, and model data
(2nd ed.). O’Reilly Media. https://r4ds.hadley.nz
Wilke, C. O. (2019). Fundamentals of data visualization.
O’Reilly Media. https://clauswilke.com/dataviz/