mirror of
https://github.com/balena-os/balena-supervisor.git
synced 2024-12-19 21:57:54 +00:00
c4f9d72172
This includes: - proxyvisor.js - references in docs - references device-state, api-binder, compose modules, API - references in tests The commit also adds a migration to remove the 4 dependent device tables from the DB. Change-type: minor Signed-off-by: Christina Ying Wang <christina@balena.io>
347 lines
15 KiB
Markdown
347 lines
15 KiB
Markdown
# Working with the Supervisor
|
|
|
|
Service: `balena-supervisor.service`, or `resin-supervisor.service` if OS < v2.78.0
|
|
|
|
The balena Supervisor is the service that carries out the management of the
|
|
software release on a device, including determining when to download updates,
|
|
the changing of variables, ensuring services are restarted correctly, etc.
|
|
It is the on-device agent for balenaCloud.
|
|
|
|
As such, it's imperative that the Supervisor is operational and healthy at all
|
|
times, even when a device is not connected to the Internet, as the Supervisor still
|
|
ensures the running of a device that is offline.
|
|
|
|
The Supervisor itself is a Docker service that runs alongside any installed
|
|
user services and the healthcheck container. One major advantage of running it as
|
|
a Docker service is that it can be updated just like any other service, although
|
|
carrying that out is slightly different than updating user containers. (See [Updating the Supervisor](#82-updating-the-supervisor)).
|
|
|
|
Before attempting to debug the Supervisor, it's recommended to upgrade the Supervisor to
|
|
the latest version, as we frequently release bugfixes and features that may resolve device issues.
|
|
|
|
Otherwise, assuming you're still logged into your development device, run the following:
|
|
|
|
```shell
|
|
root@debug-device:~# systemctl status balena-supervisor
|
|
● balena-supervisor.service - Balena supervisor
|
|
Loaded: loaded (/lib/systemd/system/balena-supervisor.service; enabled; vendor preset: enabled)
|
|
Active: active (running) since Fri 2022-08-19 18:08:59 UTC; 41s ago
|
|
Process: 2296 ExecStartPre=/usr/bin/balena stop resin_supervisor (code=exited, status=1/FAILURE)
|
|
Process: 2311 ExecStartPre=/usr/bin/balena stop balena_supervisor (code=exited, status=0/SUCCESS)
|
|
Process: 2325 ExecStartPre=/bin/systemctl is-active balena.service (code=exited, status=0/SUCCESS)
|
|
Main PID: 2326 (start-balena-su)
|
|
Tasks: 10 (limit: 1878)
|
|
Memory: 11.9M
|
|
CGroup: /system.slice/balena-supervisor.service
|
|
├─2326 /bin/sh /usr/bin/start-balena-supervisor
|
|
├─2329 /proc/self/exe --healthcheck /usr/lib/balena-supervisor/balena-supervisor-healthcheck --pid 2326
|
|
└─2486 balena start --attach balena_supervisor
|
|
|
|
Aug 19 18:09:07 debug-device balena-supervisor[2486]: [debug] Starting target state poll
|
|
Aug 19 18:09:07 debug-device balena-supervisor[2486]: [debug] Spawning journald with: chroot /mnt/root journalctl -a --follow -o json >
|
|
Aug 19 18:09:07 debug-device balena-supervisor[2486]: [debug] Finished applying target state
|
|
Aug 19 18:09:07 debug-device balena-supervisor[2486]: [success] Device state apply success
|
|
Aug 19 18:09:07 debug-device balena-supervisor[2486]: [info] Applying target state
|
|
Aug 19 18:09:07 debug-device balena-supervisor[2486]: [info] Reported current state to the cloud
|
|
Aug 19 18:09:07 debug-device balena-supervisor[2486]: [debug] Finished applying target state
|
|
Aug 19 18:09:07 debug-device balena-supervisor[2486]: [success] Device state apply success
|
|
Aug 19 18:09:17 debug-device balena-supervisor[2486]: [info] Internet Connectivity: OK
|
|
Aug 19 18:09:18 debug-device balena-supervisor[2486]: [info] Reported current state to the cloud
|
|
```
|
|
|
|
You can see the Supervisor is just another `systemd` service
|
|
(`balena-supervisor.service`) and that it is started and run by balenaEngine.
|
|
|
|
Supervisor issues, due to their nature, vary significantly. Issues may commonly
|
|
be misattributed to the Supervisor. As the Supervisor is verbose about its
|
|
state and actions, such as the download of images, it tends to be suspected of
|
|
problems when in fact there are usually other underlying issues. A few examples
|
|
are:
|
|
|
|
- Networking problems - The Supervisor reports failed downloads
|
|
or attempts to retrieve the same images repeatedly, where in fact unstable
|
|
networking is usually the cause.
|
|
- Service container restarts - The default policy for service containers is to
|
|
restart if they exit, and this sometimes is misunderstood. If a container is
|
|
restarting, it's worth ensuring it's not because the container itself is
|
|
exiting either due to a bug in the service container code or
|
|
because it has correctly come to the end of its running process.
|
|
- Release not being downloaded - For instance, a fleet/device has been pinned
|
|
to a particular version, and a new push is not being downloaded.
|
|
|
|
It's _always_ worth considering how the system is configured, how releases were
|
|
produced, how the fleet or device is configured and what the current
|
|
networking state is when investigating Supervisor issues, to ensure that there
|
|
isn't something else amiss that the Supervisor is merely exposing via logging.
|
|
|
|
Another point to note is that the Supervisor is started using
|
|
[`healthdog`](https://github.com/balena-os/healthdog-rs) which continually
|
|
ensures that the Supervisor is present by using balenaEngine to find the
|
|
Supervisor image. If the image isn't present, or balenaEngine doesn't respond,
|
|
then the Supervisor is restarted. The default period for this check is 180
|
|
seconds. Inspecting `/lib/systemd/system/balena-supervisor.service` on-device
|
|
will show whether the timeout period is different for a particular device.
|
|
For example:
|
|
|
|
```shell
|
|
root@debug-device:~# cat /lib/systemd/system/balena-supervisor.service
|
|
[Unit]
|
|
Description=Balena supervisor
|
|
Requires=\
|
|
resin\x2ddata.mount \
|
|
balena-device-uuid.service \
|
|
os-config-devicekey.service \
|
|
bind-etc-balena-supervisor.service \
|
|
extract-balena-ca.service
|
|
Wants=\
|
|
migrate-supervisor-state.service
|
|
After=\
|
|
balena.service \
|
|
resin\x2ddata.mount \
|
|
balena-device-uuid.service \
|
|
os-config-devicekey.service \
|
|
bind-etc-systemd-system-resin.target.wants.service \
|
|
bind-etc-balena-supervisor.service \
|
|
migrate-supervisor-state.service \
|
|
extract-balena-ca.service
|
|
Wants=balena.service
|
|
ConditionPathExists=/etc/balena-supervisor/supervisor.conf
|
|
|
|
[Service]
|
|
Type=simple
|
|
Restart=always
|
|
RestartSec=10s
|
|
WatchdogSec=180
|
|
SyslogIdentifier=balena-supervisor
|
|
EnvironmentFile=/etc/balena-supervisor/supervisor.conf
|
|
EnvironmentFile=-/tmp/update-supervisor.conf
|
|
ExecStartPre=-/usr/bin/balena stop resin_supervisor
|
|
ExecStartPre=-/usr/bin/balena stop balena_supervisor
|
|
ExecStartPre=/bin/systemctl is-active balena.service
|
|
ExecStart=/usr/bin/healthdog --healthcheck=/usr/lib/balena-supervisor/balena-supervisor-healthcheck /usr/bin/start-balena-supervisor
|
|
ExecStop=-/usr/bin/balena stop balena_supervisor
|
|
|
|
[Install]
|
|
WantedBy=multi-user.target
|
|
Alias=resin-supervisor.service
|
|
```
|
|
|
|
#### 8.1 Restarting the Supervisor
|
|
|
|
It's rare to actually _need_ a Supervisor restart. The Supervisor will attempt to
|
|
recover from issues that occur automatically, without the requirement for a restart.
|
|
When in doubt about whether a restart is required, look at the Supervisor logs and
|
|
double check other on-duty support agents if needed. If fairly certain, it's generally
|
|
safe to restart the Supervisor, as long as the user is aware that some extra bandwidth
|
|
and device resources will be used on startup.
|
|
|
|
There are instances where the Supervisor is incorrectly restarted when in fact
|
|
the issue could be the corruption of service images, containers, volumes
|
|
or networking. In these cases, you're better off dealing with the underlying
|
|
balenaEngine to ensure that anything corrupt is recreated correctly. See the
|
|
balenaEngine section for more details.
|
|
|
|
If a restart is required, ensure that you have gathered as much information
|
|
as possible before a restart, including pertinent logs and symptoms so that
|
|
investigations can occur asynchronously to determine what occurred and how it
|
|
may be mitigated in the future. Enabling persistent logging may also be beneficial
|
|
in cases where symptoms are repeatedly occurring.
|
|
|
|
To restart the Supervisor, simply restart the `systemd` service:
|
|
|
|
```shell
|
|
root@debug-device:~# systemctl restart balena-supervisor.service
|
|
root@debug-device:~# systemctl status balena-supervisor.service
|
|
● balena-supervisor.service - Balena supervisor
|
|
Loaded: loaded (/lib/systemd/system/balena-supervisor.service; enabled; vendor preset: enabled)
|
|
Active: active (running) since Fri 2022-08-19 18:13:28 UTC; 10s ago
|
|
Process: 3013 ExecStartPre=/usr/bin/balena stop resin_supervisor (code=exited, status=1/FAILURE)
|
|
Process: 3021 ExecStartPre=/usr/bin/balena stop balena_supervisor (code=exited, status=0/SUCCESS)
|
|
Process: 3030 ExecStartPre=/bin/systemctl is-active balena.service (code=exited, status=0/SUCCESS)
|
|
Main PID: 3031 (start-balena-su)
|
|
Tasks: 11 (limit: 1878)
|
|
Memory: 11.8M
|
|
CGroup: /system.slice/balena-supervisor.service
|
|
├─3031 /bin/sh /usr/bin/start-balena-supervisor
|
|
├─3032 /proc/self/exe --healthcheck /usr/lib/balena-supervisor/balena-supervisor-healthcheck --pid 3031
|
|
└─3089 balena start --attach balena_supervisor
|
|
|
|
Aug 19 18:13:33 debug-device balena-supervisor[3089]: [info] Waiting for connectivity...
|
|
Aug 19 18:13:33 debug-device balena-supervisor[3089]: [debug] Starting current state report
|
|
Aug 19 18:13:33 debug-device balena-supervisor[3089]: [debug] Starting target state poll
|
|
Aug 19 18:13:33 debug-device balena-supervisor[3089]: [debug] Spawning journald with: chroot /mnt/root journalctl -a --follow -o json >
|
|
Aug 19 18:13:33 debug-device balena-supervisor[3089]: [debug] Finished applying target state
|
|
Aug 19 18:13:33 debug-device balena-supervisor[3089]: [success] Device state apply success
|
|
Aug 19 18:13:34 debug-device balena-supervisor[3089]: [info] Applying target state
|
|
Aug 19 18:13:34 debug-device balena-supervisor[3089]: [info] Reported current state to the cloud
|
|
Aug 19 18:13:34 debug-device balena-supervisor[3089]: [debug] Finished applying target state
|
|
Aug 19 18:13:34 debug-device balena-supervisor[3089]: [success] Device state apply success
|
|
```
|
|
|
|
#### 8.2 Updating the Supervisor
|
|
|
|
Occasionally, there are situations where the Supervisor requires an update. This
|
|
may be because a device needs to use a new feature or because the version of
|
|
the Supervisor on a device is outdated and is causing an issue. Usually the best
|
|
way to achieve this is via a balenaOS update, either from the dashboard or via
|
|
the command line on the device.
|
|
|
|
If updating balenaOS is not desirable or a user prefers updating the Supervisor independently, this can easily be accomplished using [self-service](https://www.balena.io/docs/reference/supervisor/supervisor-upgrades/) Supervisor upgrades. Alternatively, this can be programmatically done by using the Node.js SDK method [device.setSupervisorRelease](https://www.balena.io/docs/reference/sdk/node-sdk/#devicesetsupervisorreleaseuuidorid-supervisorversionorid-%E2%87%92-codepromisecode).
|
|
|
|
You can additionally write a script to manage this for a fleet of devices in combination with other SDK functions such as [device.getAll](https://www.balena.io/docs/reference/sdk/node-sdk/#devicegetalloptions-%E2%87%92-codepromisecode).
|
|
|
|
**Note:** In order to update the Supervisor release for a device, you must have edit permissions on the device (i.e., more than just support access).
|
|
|
|
#### 8.3 The Supervisor Database
|
|
|
|
The Supervisor uses a SQLite database to store persistent state, so in the
|
|
case of going offline, or a reboot, it knows exactly what state an
|
|
app should be in, and which images, containers, volumes and networks
|
|
to apply to it.
|
|
|
|
This database is located at
|
|
`/mnt/data/resin-data/balena-supervisor/database.sqlite` and can be accessed
|
|
inside the Supervisor container at `/data/database.sqlite` by running Node.
|
|
Assuming you're logged into your device, run the following:
|
|
|
|
```shell
|
|
root@debug-device:~# balena exec -ti balena_supervisor node
|
|
```
|
|
|
|
This will get you into a Node interpreter in the Supervisor service
|
|
container. From here, we can use the `sqlite3` NPM module used by
|
|
the Supervisor to make requests to the database:
|
|
|
|
```shell
|
|
> sqlite3 = require('sqlite3');
|
|
{
|
|
Database: [Function: Database],
|
|
Statement: [Function: Statement],
|
|
Backup: [Function: Backup],
|
|
OPEN_READONLY: 1,
|
|
OPEN_READWRITE: 2,
|
|
OPEN_CREATE: 4,
|
|
OPEN_FULLMUTEX: 65536,
|
|
OPEN_URI: 64,
|
|
OPEN_SHAREDCACHE: 131072,
|
|
OPEN_PRIVATECACHE: 262144,
|
|
VERSION: '3.30.1',
|
|
SOURCE_ID: '2019-10-10 20:19:45 18db032d058f1436ce3dea84081f4ee5a0f2259ad97301d43c426bc7f3df1b0b',
|
|
VERSION_NUMBER: 3030001,
|
|
OK: 0,
|
|
ERROR: 1,
|
|
INTERNAL: 2,
|
|
PERM: 3,
|
|
ABORT: 4,
|
|
BUSY: 5,
|
|
LOCKED: 6,
|
|
NOMEM: 7,
|
|
READONLY: 8,
|
|
INTERRUPT: 9,
|
|
IOERR: 10,
|
|
CORRUPT: 11,
|
|
NOTFOUND: 12,
|
|
FULL: 13,
|
|
CANTOPEN: 14,
|
|
PROTOCOL: 15,
|
|
EMPTY: 16,
|
|
SCHEMA: 17,
|
|
TOOBIG: 18,
|
|
CONSTRAINT: 19,
|
|
MISMATCH: 20,
|
|
MISUSE: 21,
|
|
NOLFS: 22,
|
|
AUTH: 23,
|
|
FORMAT: 24,
|
|
RANGE: 25,
|
|
NOTADB: 26,
|
|
cached: { Database: [Function: Database], objects: {} },
|
|
verbose: [Function]
|
|
}
|
|
> db = new sqlite3.Database('/data/database.sqlite');
|
|
Database {
|
|
open: false,
|
|
filename: '/data/database.sqlite',
|
|
mode: 65542
|
|
}
|
|
```
|
|
|
|
You can get a list of all the tables used by the Supervisor by issuing:
|
|
|
|
```shell
|
|
> db.all("SELECT name FROM sqlite_master WHERE type='table' ORDER BY name;", console.log);
|
|
Database { open: true, filename: '/data/database.sqlite', mode: 65542 }
|
|
> null [
|
|
{ name: 'apiSecret' },
|
|
{ name: 'app' },
|
|
{ name: 'config' },
|
|
{ name: 'containerLogs' },
|
|
{ name: 'currentCommit' },
|
|
{ name: 'deviceConfig' },
|
|
{ name: 'engineSnapshot' },
|
|
{ name: 'image' },
|
|
{ name: 'knex_migrations' },
|
|
{ name: 'knex_migrations_lock' },
|
|
{ name: 'logsChannelSecret' },
|
|
{ name: 'sqlite_sequence' }
|
|
]
|
|
```
|
|
|
|
With these, you can then examine and modify data, if required. Note that there's
|
|
usually little reason to do so, but this is included for completeness. For
|
|
example, to examine the configuration used by the Supervisor:
|
|
|
|
```shell
|
|
> db.all('SELECT * FROM config;', console.log);
|
|
Database { open: true, filename: '/data/database.sqlite', mode: 65542 }
|
|
> null [
|
|
{ key: 'localMode', value: 'false' },
|
|
{ key: 'initialConfigSaved', value: 'true' },
|
|
{
|
|
key: 'initialConfigReported',
|
|
value: 'https://api.balena-cloud.com'
|
|
},
|
|
{ key: 'name', value: 'shy-rain' },
|
|
{ key: 'targetStateSet', value: 'true' },
|
|
{ key: 'delta', value: 'true' },
|
|
{ key: 'deltaVersion', value: '3' }
|
|
]
|
|
```
|
|
|
|
Occasionally, should the Supervisor get into a state where it is unable to
|
|
determine which release images it should be downloading or running, it
|
|
is necessary to clear the database. This usually goes hand-in-hand with removing
|
|
the current containers and putting the Supervisor into a 'first boot' state,
|
|
whilst keeping the Supervisor and release images. This can be achieved by
|
|
carrying out the following:
|
|
|
|
```shell
|
|
root@debug-device:~# systemctl stop balena-supervisor.service update-balena-supervisor.timer
|
|
root@debug-device:~# balena rm -f $(balena ps -aq)
|
|
1db1d281a548
|
|
6c5cde1581e5
|
|
2a9f6e83578a
|
|
root@debug-device:~# rm /mnt/data/resin-data/balena-supervisor/database.sqlite
|
|
```
|
|
|
|
This:
|
|
|
|
- Stops the Supervisor (and the timer that will attempt to restart it).
|
|
- Removes all current service containers, including the Supervisor.
|
|
- Removes the Supervisor database.
|
|
(If for some reason the images also need to be removed, run
|
|
`balena rmi -f $(balena images -q)` which will remove all images _including_
|
|
the Supervisor image).
|
|
You can now restart the Supervisor:
|
|
|
|
```shell
|
|
root@debug-device:~# systemctl start update-balena-supervisor.timer balena-supervisor.service
|
|
```
|
|
|
|
If you deleted all the images, this will first download the Supervisor image
|
|
again before restarting it.
|
|
At this point, the Supervisor will start up as if the device has just been
|
|
provisioned and already registered, and the device's target release will
|
|
be freshly downloaded if images were removed before starting the service
|
|
containers.
|