Save thousands of dollars on Zoom Cloud Recording Storage! Download records automatically and store locally. Provide simple but effective web frontend to watch and share meeting recordings.
- Download Zoom Cloud Recordings automatically
- Delete/Trash downloaded recordings from Zoom Cloud after download
- Specify which types of recordings to download (shared screen, gallery view, active speaker) and which to ignore (audio only, chat, etc.)
- Host a simple web frontend to watch and share recordings
- Run multiple instances of the service for redundancy
Zoomrs can be installed as a systemd service or run from the console as a persistent process or a set of CLI tools. It can be run as a Docker container as well.
Zoom API credentials are required to download recordings. You can get them at https://marketplace.zoom.us/develop/create. You need to create JWT app and copy API key and secret to the configuration file.
Add the following scopes to the App:
/recording:master/recording:read:admin/recording:write:admin/report:read:admin
Google OAuth credentials are required to authenticate users. You can get them at https://console.cloud.google.com/apis/credentials. You need to create OAuth client ID and copy client ID and secret to the configuration file. Mind authorized redirect URIs - local domains are not allowed, so you need to use a public domain name or IP address.
You need to specify the list of users that are allowed to access the web frontend. Their email addresses should be specified in the configuration file.
See config/config_example.yml for example configuration file, available options and their descriptions. Copy it to config/config.yml and edit it to your needs.
-
To run a binary distribution, please refer to the README in
distdirectory. -
To build from source, proceed with this manual.
Note
this is not recommended for production use, use systemd service instead or run it in a Docker container
-
Clone the repository from GitHub
git clone https://github.com/parMaster/zoomrs.git
-
Make sure
config/config.ymlexists and is configured properly -
Run
make runto build the binary and run it in foreground modemake run
-
To stop the service press
Ctrl+C(or sendSIGINT,SIGTERMsignal to the process)
-
Repeat steps 1 and 2 from the previous section
-
Run
make deployto build the binary and copy everything where it belongs (seeMakefilefor details), enable and run the servicemake deploy
-
Run
make statusto check the status of the servicemake status
Log files are located at /var/log/zoomrs.log and /var/log/zoomrs.err by default.
-
Clone the repository from GitHub
git clone https://github.com/parMaster/zoomrs.git
-
Make sure
config/config.ymlexists and is configured properly -
Check configuration parameters in Dockerfile and docker-compose.yml
-
Build and run container
docker compose up -d
Web frontend is available at http://localhost:8099 by default. You can change the port in the configuration file (server.listen parameter).
GET `/`Displays the list of recordings. Each recording has a link to share (view) it. Recordings are sorted by date in descending order. Login is required to view the list. Google OAuth is used for authentication. Access is restricted to users with email addresses from the list specified in the configuration file (see server.managers).
Share button is available for each recording, it generates a link to view the recording. Share link looks like:
GET `/watch/834d0992ad0d632cf6c3174b975cb5e5?uuid=kzbiTyvQQp2fW6biu8Vy%2BQ%3D%3D`Displays the page with the meeting title and player to watch the recording. Simple controls besides the embeded player is providing are available.
Returns the status of the service and Zoom cloud storage usage stats. If the service is running, returns 200 OK and the following JSON. Example response:
{
"cloud": {
"date": "2023-07-09",
"free_usage": "495 GB",
"plan_usage": "0",
"usage": "27.98 GB",
"usage_percent": 5
},
"stats": {
"downloaded": {
"count": 6529,
"size_gb": 2148,
"size_mb": 2200412
}
},
"status": "OK",
"storage": {
"free": "1.2 TB",
"total": "3.6 TB",
"usage_percent": 63,
"used": "2.2 TB"
}
}status can be:
OKwhen everything is downloaded and nothing has failedLOADINGwhen there arequeuedordownloadingrecordings presentFAILEDwhen there are onlydownloadedandfailedrecordings in the database
stats section contains number of recordings and their total size in GB and MB grouped by status
cloud section contains Zoom cloud storage usage stats. date is the last time the stats were updated (it is updated every 24 hours, so if you see the date is not today, it means the stats dodn't change since then), free_usage is the amount of free storage, plan_usage is the amount of storage available for the current plan, usage is the amount of storage used by recordings, usage_percent is the percentage of used storage.
storage section contains the stats of the local storage. free is the amount of free storage, total is the total amount of storage, usage_percent is the percentage of used storage, used is the amount of used storage.
This API is useful for monitoring the service status and triggering alerts when something goes wrong.
Another example response, when there are recordings in queued and downloading status (only relevant fields are shown):
{
"stats": {
"downloaded": {
"count": 5292,
"size_gb": 1765,
"size_mb": 1808161
},
"downloading": {
"count": 1,
"size_gb": 0,
"size_mb": 666
},
"queued": {
"count": 88,
"size_gb": 27,
"size_mb": 28044
}
},
"status": "LOADING"
}Auth required. Runs a consistency check of the repository (see check cli tool cmd, it's the same). Example response:
{
"checked": 5278,
"error": null
}Auth required. Returns the total size of the recordings grouped by date. Optional parameter K, M or G can be used to specify the size in KB, MB or GB respectively. If no parameter is specified, the size is returned in bytes. Example response:
{
"2023-03-20":31,
"2023-03-21":13,
"2023-03-22":36,
"2023-03-23":19,
"2023-03-24":41
}accessKey is checked against server.access_key_salt config option. This api is called to ask if every meeting from the list is loaded, list is passed as a JSON array of UUIDs in the request body.
Request example:
{
"meetings":{
"in7MDVrTS5adXWFwsCwoYg==",
"0ao3hvbxQvqU2wkpXjbwhw==",
"pEbVqZ5jQP6+NY0ewvZ+wg==",
"uOoMA3wcSF65PtwTDw/k1w=="
}
}Response when all meetings are loaded:
{
"result":"ok"
}Response when some meetings are not loaded:
{
"result":"pending"
}Zoomrs comes with a CLI tool to trash/delete recordings from Zoom Cloud. It is useful when running miltiple servers and you want to delete recordings from Zoom Cloud only after all servers have downloaded them. CLI tool is located at cmd/cli/main.go. Run make to build it and put to dist/zoomrs-cli.
It can be run like this:
go run ./cmd/cli --cmd checkor like this:
./dist/zoomrs-cli --cmd checkAvailable commands:
check- checks the consistency of the repository: if all recordings are downloaded and if all downloaded recordings are present on the disk, also the size of each recording file is checked. Run this command periodically to make sure everything is OK. Run it like this:
./dist/zoomrs-cli --cmd checkExample output:
```
2023/06/19 17:15:01 [INFO] starting CheckConsistency
2023/06/19 17:15:01 [INFO] Checked files: 5278
2023/06/19 17:15:01 [INFO] CheckConsistency: OK, 5278
```
trash- trashes recordings from Zoom Cloud. Run it like this:
./zoomrs-cli --dbg --cmd trash --trash 2where `2` is 2 days before today, so all the recordings from the bay before yesterday will be trashed. This is designed this way to run it as a cron job every day. Cron job line example:
00 10 * * * cd $HOME/go/src/zoomrs/dist && ./zoomrs-cli --cmd trash --trash 2 --config ../config/config_cli.yml >> /var/log/cron.log 2>&1will trash all recordings from the day before yesterday every day at 10:00 AM. `--config` option is used to specify the path to the configuration file. `--dbg` option can be used to enable debug logging. Logs are written to stdout, and redirected to `/var/log/cron.log` in the example above.
cloudcap- trims recordings from Zoom Cloud to avoid exceeding the storage limit. LeavesClient.CloudCapacityHardLimitbytes of the most recent recordings (review the value in config before running!), trashes the rest. Cron job line to run it every day at 5:30 AM (don't mind the paths, they are specific to my setup, use your own):
30 05 * * * cd $HOME/go/src/zoomrs/dist && ./zoomrs-cli --dbg --cmd cloudcap --config ../config/config_cli.yml >> /var/log/cron.log 2>&1sync- syncs recordings from Zoom Cloud. Run it like this:
./zoomrs-cli --dbg --cmd sync --days 1`--days` parameter used with the value of `1` to sync all the yesterday recordings (1 day before today). This is designed this way to run it as a cron job. Cron job line example:
00 03 * * * cd $HOME/go/src/zoomrs/dist && ./zoomrs-cli --cmd sync --days 1 --config ../config/config_cli.yml >> /var/log/zoomrs.cron.log 2>&1will sync all recordings from the yesterday every day at 3:00 AM. --config option is used to specify the path to the configuration file. --dbg option can be used to enable debug logging. Logs are written to stdout, and redirected to /var/log/cron.log in the example above.
Note
CLI tool uses different configuration file then the server with different Zoom API credentials to avoid spoiling services's auth token when running CLI. Also, running multi-server setup you want to sync recordings only after all servers have downloaded them, so you need to run CLI tool on one of the servers, allow syncing records in CLI config and deny it in servers configs.
You can run multiple instances of the service to increase reliability, duplicate downloaded data for redundancy. Each instance should have its own configuration file and its own database file. Each instance should have its own Zoom API credentials. Consider following setup as an example:
- One main instance that downloads recordings and hosts web frontend (see
config/config_example.ymlfor example configuration file). Enable sync and download for this instance:server.sync_job: trueandserver.download_job: truein the configuration file, set oauth credentials and authorized users. - One or many secondary instances that download recordings but don't host web frontend. Two options are available here:
- Run the service with
server.sync_job: trueandserver.download_job: truein the configuration file. This way download job will run somewhere from 00:00 to 01:00 am. - Run the service with
server.sync_job: falseandserver.download_job: falseso it will just host the API. Run downloader with cron job (seesynccmd crontab line example in the previous section). This way you can set the time to run the download job
- Run the service with
- Run cleanup job on one of the instances (see
trashcmd crontab line example in the previous section). Use configuration file that enumerates all the instances inserver.instancessection. This way cleanup job will check all the instances for consistency and trash/delete recordings from Zoom Cloud only if all the instances have downloaded them. Disable deleting and trashing downloaded recordings (client.trash_downloaded: falseandclient.delete_downloaded: falsein the configuration file) on every other instance but this one.
Note
Copy yesterday's recordings from "Main" instance to "Secondary" instance Secondary instance can run something like this to copy yesterday's recordings from "Main" instance:
sleep 1s && date && scp -r server.local:/data/`date --date="yesterday" +%Y-%m-%d` /data/ && dateNote
Database backup
Backup database file regularly to prevent data loss. See example shell script at dist/backup_db.sh. It can be run as a cron job like this:
0 10 * * * sh $HOME/go/src/zoomrs/backup_db.shPull requests are welcome. For major changes, please open an issue first to discuss what you would like to change. Check the existing issues to see if your problem is already being discussed or if you're willing to help with one of them. Tests are highly appreciated.
GNU GPLv3 © Dmytro Borshchanenko 2023
If you have any security issue to report, contact project owner directly at master@parMaster.com.ua or use Issues section of this repository.
The author of this project is not responsible for any damage caused by the use of this software. Use it at your own risk. However, the software is being used in production at least since May 2023 on a number of devices, processing hundreds of GB of data every day and is considered stable.
- lgr - simple but effective logging package
- go-sqlite3 as a database driver
- go-pkgz/auth - powerful authentication middleware