Internal/
This folder contains internal repository and documentation management code. It does not contain examples.
Continuous Integration and Continuous Deployment
Modal cares deeply about the correctness of our examples -- we have also suffered from janky, poorly-maintained documentation and we do our best to ensure that our examples don't pay that forward.
This document explains the CI/CD process we use. It is primarily intended for Modal engineers, but if you're contributing an example and have the bandwidth to set up the testing as well, we appreciate it!
Frontmatter
Examples can include a small frontmatter block in YAML format that controls testing and deployment behavior.
Fields include:
deploy: Iftrue, the example is deployed as a Modal application withmodal deploy. Iffalse, it is not. Default isfalse. Examples should be deployed only if they are a live demo or they are consumed as a service by other examples.cmd: The command to run the example for testing. Default is["modal", "run", "<filename>"].args: Arguments to pass to the command. Default is[].lambda-test: Iftrue, the example is tested with the cli command provided incmd. Iffalse, it is not. Default istrue. Note that this controls execution in the CI/CD of this repo and in the internal AWS Lambda monitor as part ofsynthetic_monitoring.env: A dictionary of environment variables to include when testing. Default is{}, but note that the environment can be modified in the CI/CD of this repo or in the monitor-based testing.
Below is an example frontmatter for deploying a web app. Note that here we
modal serve in the test so as to not deploy to prod when testing. Note that in
testing environments, the MODAL_SERVE_TIMEOUT environment variable is set so
that the command terminates.
---
deploy: true
cmd: ["modal", "serve", "10_integrations/pushgateway.py"]
---
# example prose and code begins here
Testing in GitHub Actions
When a PR is opened, any changed examples are run via GitHub Actions.
You can find the commands used to execute tests in the .github/workflows
directory. These can be used to run the tests locally. You may need to install
the requirements.txt in this folder to do so.
This workflow is intended to catch errors at the time a PR is made -- including both errors in the example and issues with the execution of the example in the monitoring system, like file imports.
Continual Monitoring
Examples are executed regularly and at random to check for regressions. The results are monitored.
Modal engineers, see synthetic_monitoring in the modal repo for details.