Skip to contents

CFAEpiNow2Pipeline v0.2.0

Features

  • Remove azure/requirements.txt as we have moved to inline dependencies in uv
  • Does not save outlier csv files if contents are empty
  • Switching base rocker image from geospatial (4.7 GB) to r-ver (3 GB)
  • Refactoring unit tests to make use of setup.R global testing env
  • Adding dependencies to install cmdstanr backend and using GH action
  • convert drop cols value to character for point/state exlcusions
  • Run make test-batch target locally
  • Added make run-caj for using a CAJ instead of Batch
  • Update runner action version
  • Remove duplicate batch autoscale text file
  • Improve consistency in docs
  • Update version of deploy action
  • Update github checkout action from V2 to V4
  • Setting up dependabot yaml file
  • Remove out-of-date demo folder
  • Add automated check that docs are up to date
  • Rewrite README for simplification and clarity
  • Switch to the air code formatter
  • Replace remaining self-hosted runner workflows with ubuntu-latest
  • Fix mismatch between R code and documentation
  • Change code owner and include authors in R package
  • Change code owner
  • Add documentation to the Makefile
  • Fix mismatch between R code and documentation
  • Fix production diseases
  • Add RSV specifications
  • Create the config files locally to speed things up
  • Lock dependencies for creating the pool
  • Saving state exclusions to nssp-rt/state_exclusions
  • Automate tag deletion from ghcr.io
  • Editing of SOP.md
  • Pin r-version at 4.4.3 for CI/CD
  • Fix minor typos in SOP.md.
  • Swap from Dockerfile-batch to using an inline-metadata script, managed by uv.
  • Adding dynamic logic to re-query for configs in blob
  • Automate creation of outlier csv for nssp-elt-2/outliers
  • Fix ‘latest’ tag for CI
  • Updated path for read/write of data outliers
  • Updating makefile to represent unified Dockerfile approach (not two-step build)
  • Make sure we change “COVID-19/Omicron” to “COVID-19” when reading NSSP data.
  • Unified Dockerfile
  • Add instructions for data outliers reruns to the SOP.
  • Add ability to call make rerun-prod to rerun just the tasks that needed a data change.
  • Add output container as a new field in the config file.
  • Building with ubuntu-latest and using Container App runner for all else, remove azure-cli action
  • Adding exclusions documentation and Makefile support
  • Add the blob storage container, if provided
  • Adding make command to test Azure batch
  • Updating subnet ID and pool VM to 22.04 from 20.04
  • Write model diagnostics to an output file, correcting an oversight
  • Refactored GH Actions container build to cfa-actions 2-step build
  • Creating SOP.md to document weekly run procedures, including diagram
  • Allows unique job_ids for runs.
  • Makefile supports either docker or podman as arguments to setup & manage containers
  • Streamlined configurable container execution provided by included start.sh script
  • Container App Job execution tools added including job-template.yaml file for single task and Python script for bulk tasks
  • GitHub Actions workflow added to start Azure Container App Job
  • Minor changes in removing unused container tags from Azure CR
  • Reactivated DEBUG level logs from EpiNow2 so that sampler progress is visible
  • Added new test data and unit tests for point exclusions
  • Removed any nulls from reference_date column in outlier files
  • Removed quotes and identifiers from outlier and state exclusions files

CFAEpiNow2Pipeline v0.1.0

This initial release establishes minimal feature parity with the internal EpiNow2 Rt modeling pipeline. It adds wrappers to integrate with internal data schemas and ingest pre-estimated model parameters (i.e., generation intervals, right-truncation). It defines an output schema and adds comprehensive logging. The repository also has functionality to set up and deploy to Azure Batch.

Features

  • GitHub Actions to build Docker images on PR and merge to main, deploy Azure Batch environments off the built images, and tear down the environment (including images) on PR close.
  • Comprehensive documentation of pipeline code and validation of input data, parameters, and model run configs
  • Set up comprehensive logging of model runs and handle pipeline failures to preserve logs where possible
  • Automatically download and upload inputs and outputs from Azure Blob Storage
  • A new script for building the pool. Runnable from CLI or GHA. Requires uv be installed, and then uv handles the python and dependency management based on the inline script metadata.