£450Registration Fee
Register Now- Overview
- Instructors
- Schedule
Course Description
This 5-day course provides a practical introduction to applied spatial data analysis using R. You will learn how to work confidently with both vector and raster data, manage coordinate reference systems (CRS) and projections, perform I/O with modern formats, and carry out essential operations such as dissolves, buffers, and clips. Building on these foundations, we explore spatial relationships and joins, raster alignment and algebra, spatial dependence diagnostics, point pattern analysis, variogram theory, and interpolation methods (IDW, kriging). The final days introduce areal spatial regression (SAR/SEM), spatial GLMs, and fully reproducible pipelines for analysis and reporting.
What You’ll Learn
During the course will cover the following:
- Explain vector vs raster data, topology, and CRS/EPSG/PROJ concepts, and avoid common reprojection pitfalls.
- Read, write, validate, and round-trip spatial data using sf/terra with appropriate formats, precision, and units.
- Perform core vector operations (select, dissolve, buffer, clip) and create clear maps with tmap/ggplot2.
- Use spatial predicates and joins, construct neighbour lists and weights, and diagnose autocorrelation (Moran’s I, LISA).
- Align rasters; run raster algebra and terrain derivatives; compute zonal statistics for modelling.
- Understand and fit variogram models.
- Fit areal spatial regressions (SAR/SEM) and spatial GLMs; interpret effects and check residual spatial structure.
Course Format
Interactive Learning Format
Each day features a well-balanced combination of lectures and hands-on practical exercises, with dedicated time for discussing participants’ own data, time permitting.
Global Accessibility
All live sessions are recorded and made available on the same day, ensuring accessibility for participants across different time zones.
Collaborative Discussions
Open discussion sessions provide an opportunity for participants to explore specific research questions and engage with instructors and peers.
Comprehensive Course Materials
All code, datasets, and presentation slides used during the course will be shared with participants by the instructor.
Personalized Data Engagement
Participants are encouraged to bring their own data for discussion and practical application during the course.
Post-Course Support
Participants will receive continued support via email for 30 days following the course, along with on-demand access to session recordings for the same period.
Who Should Attend / Intended Audiences
This course is designed for ecologists, environmental scientists, public health analysts, data scientists, postgraduate students, and early-career researchers. Participants should have basic experience with R and RStudio, such as importing data and running simple functions, and be familiar with descriptive statistics and the concepts of linear regression. Prior exposure to data wrangling with packages like dplyr or tidyr, plotting with ggplot2, and reading model outputs is useful but not required. While some familiarity with linear models is advantageous, key concepts will be reviewed during the course.
Equipment and Software requirements
A laptop or desktop computer with a functioning installation of R and RStudio is required. Both R and RStudio are free, open-source programs compatible with Windows, macOS, and Linux systems.
A working webcam is recommended to support interactive elements of the course. We encourage participants to keep their cameras on during live Zoom sessions to foster a more engaging and collaborative environment.
While not essential, using a large monitor – or ideally a dual-monitor setup – can significantly enhance your learning experience by allowing you to view course materials and work in R or linux simultaneously.
All necessary R packages will be introduced and installed during the workshop. A comprehensive list of required packages will also be shared with participants ahead of the course to allow for optional pre-installation.
Dr. Niamh Mimnagh
Niamh is a statistician working at the interface of ecology, epidemiology, and data science. Her research focuses on applying and developing statistical and machine learning methods to address real-world challenges such as estimating species population sizes from count and trace data and predicting livestock disease re-emergence using sparse or imbalanced datasets. She works with a wide array of statistical approaches, including Bayesian hierarchical models, N-mixture models, anomaly detection algorithms, and spatial analysis techniques.
Niamh earned her PhD in Statistics, with a focus on multispecies abundance modelling, and holds a first-class MSc in Data Science. Alongside her research, she is actively engaged in science communication and education, running a popular blog on applied statistics for non-specialists, and regularly delivering workshops and guest lectures on topics such as GLMs and machine learning with imbalanced data.
Education & Career
- PhD in Statistics (Multispecies Abundance Modelling)
- MSc in Data Science (First Class Honours)
- Instructor, consultant, and science communicator in statistical ecology and epidemiology
Research Focus
Niamh’s work centres on extracting meaningful insights from complex ecological and epidemiological data. She is particularly interested in population estimation techniques and predictive modelling for conservation and disease management, using advanced statistical tools and reproducible workflows.
Current Projects
- Development of Bayesian and ML approaches for estimating species abundance from imperfect data
- Modelling livestock disease risk using spatial and temporal predictors
- Creating accessible educational materials for teaching applied statistics in R
Professional Consultancy
Niamh provides expert statistical support to academic and applied research projects, with a focus on ecological monitoring, conservation planning, and disease modelling. She also advises on study design and data workflows for interdisciplinary teams.
Teaching & Skills
- Teaches topics including GLMs, Bayesian statistics, machine learning for imbalanced data, and spatial statistics in R
- Advocates for reproducibility, open science, and accessible statistical training
- Experienced in communicating complex methods to broad audiences
Links
Session 1 – 01:15:00 – Spatial data & CRS fundamentals
We establish the core mental models of spatial data: vector vs raster structures, geometry/topology, and metadata. You’ll learn how EPSG/PROJ define coordinate reference systems and why reprojection matters for distance, area, and overlay operations, with common pitfalls to avoid.
Session 2 – 01:15:00 – Reading, writing, and hygiene
This session presents robust I/O patterns with sf and terra, including choosing formats (GeoPackage/GeoTIFF/COG/GeoJSON), preserving CRS and precision, and validating geometries. We standardise schemas and units to prevent subtle corruption during import/export.
Session 3 – 01:15:00 – Core vector operations: select, dissolve, buffer, clip
We perform attribute filtering and spatial selection, dissolve features by group to create aggregated regions, build buffers (planar vs geodesic) with appropriate CRS, and clip layers to areas of interest—emphasising precision and validity checks before and after operations.
Session 4 – 01:15:00 – Visualisation: tmap & ggplot2 for maps
You’ll create clean thematic maps with thoughtful classifications, palettes, and legends in tmap and ggplot2, add essential map furniture (scale bars, north arrows, insets), and export publication-ready figures in raster and vector formats.
Session 5 – 01:15:00 – Spatial relationships & joins
We formalise spatial predicates (intersects, within, contains, touches, within-distance, nearest) and use them to drive reproducible spatial joins. Boundary policies, CRS/units, and performance considerations are covered to ensure accurate results at scale.
Session 6 – 01:15:00 – Network & accessibility basics
Lines are translated into graphs to compute realistic accessibility metrics. You’ll learn how to assign edge weights (distance/time), evaluate shortest paths, and outline simple service areas, contrasting Euclidean with network-based measures and noting practical tooling.
Session 7 – 01:15:00 – Raster fundamentals
We define extent, resolution, origin, alignment, and CRS for rasters, then align disparate layers to a reference grid. You’ll choose appropriate resampling methods for categorical vs continuous data and manage NoData and datatypes for reliable downstream analysis.
Session 8 – 01:15:00 – Raster algebra & terrain
This session introduces map algebra (local/focal), reclassification of categorical rasters, and terrain derivatives (slope, aspect, roughness) from DEMs, highlighting parameter choices and sanity checks for defensible outputs.
Session 9 – 01:15:00 – Areal neighbours & weights
You’ll construct contiguity (rook/queen), distance, and kNN neighbours and build spatial weights matrices with row standardisation. We handle islands, check graph connectivity, and visualise neighbour structures to justify modelling assumptions.
Session 10 – 01:15:00 – Global & local autocorrelation (Moran’s I, LISA)
We explain why autocorrelation matters for inference, compute global Moran’s I with permutation tests, and produce Local Moran cluster maps with appropriate significance interpretation and cautions about multiple testing.
Session 11 – 01:15:00 – Point pattern analysis
We frame point processes and CSR, apply quadrats and kernel intensity estimation, and interpret Ripley’s K/L with edge corrections. The session also introduces inhomogeneous patterns and the role of covariates when intensity varies across space.
Session 12 – 01:15:00 – From points to surfaces: density & grids
You’ll create kernel density estimation (KDE) surfaces and hex/square grids, compare bandwidth/bin-size choices, and normalise counts appropriately to produce clear, honest maps that support decision-making.
Session 13 – 01:15:00 – Variogram theory & practice
We define semivariance and interpret nugget, sill, and range. Directional variograms are used to detect anisotropy, and valid models (Spherical/Exponential/Gaussian) are fitted and assessed for stability and residual structure.
Session 14 – 01:15:00 – Interpolation I: IDW vs Ordinary Kriging
This session contrasts IDW’s deterministic weighting with model-based ordinary kriging. You’ll learn core assumptions, parameter choices, prediction vs uncertainty outputs, and how to compare approaches using principled validation.
Session 15 – 01:15:00 – Interpolation II: Co-Kriging & trend surfaces
We incorporate external drift and covariates via trend surfaces or kriging with external drift (KED), and outline when co-kriging is appropriate. The emphasis is on alignment, scaling, and parsimonious modelling to avoid common pitfalls.
Session 16 – 01:15:00 – Zonal statistics & raster summaries
You’ll compute zonal means and percent cover with exact area weighting and assemble multi-layer summaries for modelling, ensuring consistent CRS/units and defensible choices for polygon handling and NoData.
Session 17 – 01:15:00 – Areal regression: SAR/SEM/SAC & diagnostics
We position SLX against SAR/SEM/SAC models, fit and compare them, interpret SAR impacts (direct/indirect/total), and re-test residual spatial structure to ensure valid inference under your chosen weights matrix.
Session 18 – 01:15:00 – Spatial GLMs & smooths
Poisson/Binomial outcomes are modelled with spatial structure via ICAR/BYM (areal), Matérn Gaussian fields (point-referenced), or Gaussian-process-style smooths. We also include clear cautions about geographically weighted regression (GWR).
Session 19 – 01:15:00 – Spatial cross-validation & leakage
You’ll design blocked or leave-location-out cross-validation, tune models with spatial folds, and avoid feature/target leakage. The outcome is a realistic estimate of generalisation performance.
Session 20 – 01:15:00 – Reproducible pipelines, reporting & capstone presentations
We wrap analyses in targets + renv, organise data and metadata, apply a map-design checklist, and communicate uncertainty. The course concludes with short capstone presentations.
Frequently asked questions
Everything you need to know about the product and billing.
When will I receive instructions on how to join?
You’ll receive an email on the Friday before the course begins, with full instructions on how to join via Zoom. Please ensure you have Zoom installed in advance.
Do I need administrator rights on my computer?
I’m attending the course live — will I also get access to the session recordings?
I can’t attend every live session — can I join some sessions live and catch up on others later?
I’m in a different time zone and plan to follow the course via recordings. When will these be available?
I can’t attend live — how can I ask questions?
Will I receive a certificate?
When will I receive instructions on how to join?
You’ll receive an email on the Friday before the course begins, with full instructions on how to join via Zoom. Please ensure you have Zoom installed in advance.
Do I need administrator rights on my computer?
I’m attending the course live — will I also get access to the session recordings?
I can’t attend every live session — can I join some sessions live and catch up on others later?
I’m in a different time zone and plan to follow the course via recordings. When will these be available?
I can’t attend live — how can I ask questions?
Will I receive a certificate?
Still have questions?
Can’t find the answer you’re looking for? Please chat to our friendly team.





5.0
