mirror of
https://github.com/balena-os/balena-supervisor.git
synced 2024-12-27 09:12:31 +00:00
7b8b187c74
Devices affected by the bug described in 1576, are also stuck with some services in the `Downloaded` state, because the state engine does not detect that the running services should be killed on a network change even if they belong to a new release. This is a bug, which can be replicated by the tests in this commit Change-type: patch
371 lines
8.6 KiB
TypeScript
371 lines
8.6 KiB
TypeScript
import { expect } from 'chai';
|
|
import * as Docker from 'dockerode';
|
|
import { TargetStateV2 } from '~/lib/legacy';
|
|
import * as request from 'supertest';
|
|
import { setTimeout as delay } from 'timers/promises';
|
|
|
|
const BALENA_SUPERVISOR_ADDRESS =
|
|
process.env.BALENA_SUPERVISOR_ADDRESS || 'http://balena-supervisor:48484';
|
|
|
|
const getCurrentState = async () =>
|
|
await request(BALENA_SUPERVISOR_ADDRESS)
|
|
.get('/v2/local/target-state')
|
|
.expect(200)
|
|
.then(({ body }) => body.state.local);
|
|
|
|
const setTargetState = async (
|
|
target: Omit<TargetStateV2['local'], 'name'>,
|
|
timeout = 0,
|
|
) => {
|
|
const { name, config } = await getCurrentState();
|
|
const targetState = {
|
|
local: {
|
|
name,
|
|
config,
|
|
apps: target.apps,
|
|
},
|
|
};
|
|
|
|
await request(BALENA_SUPERVISOR_ADDRESS)
|
|
.post('/v2/local/target-state')
|
|
.set('Content-Type', 'application/json')
|
|
.send(JSON.stringify(targetState))
|
|
.expect(200);
|
|
|
|
return new Promise(async (resolve, reject) => {
|
|
const timer =
|
|
timeout > 0
|
|
? setTimeout(
|
|
() =>
|
|
reject(
|
|
new Error(
|
|
`Timeout while waiting for the target state to be applied`,
|
|
),
|
|
),
|
|
timeout,
|
|
)
|
|
: undefined;
|
|
|
|
while (true) {
|
|
const status = await getStatus();
|
|
if (status.appState === 'applied') {
|
|
clearTimeout(timer);
|
|
resolve(true);
|
|
break;
|
|
}
|
|
await delay(1000);
|
|
}
|
|
});
|
|
};
|
|
|
|
const getStatus = async () =>
|
|
await request(BALENA_SUPERVISOR_ADDRESS)
|
|
.get('/v2/state/status')
|
|
.then(({ body }) => body);
|
|
|
|
const docker = new Docker();
|
|
|
|
describe('state engine', () => {
|
|
beforeEach(async () => {
|
|
await setTargetState({
|
|
config: {},
|
|
apps: {},
|
|
});
|
|
});
|
|
|
|
after(async () => {
|
|
await setTargetState({
|
|
config: {},
|
|
apps: {},
|
|
});
|
|
});
|
|
|
|
it('installs an app with two services', async () => {
|
|
await setTargetState({
|
|
config: {},
|
|
apps: {
|
|
'123': {
|
|
name: 'test-app',
|
|
commit: 'deadbeef',
|
|
releaseId: 1,
|
|
services: {
|
|
'1': {
|
|
image: 'alpine',
|
|
imageId: 11,
|
|
serviceName: 'one',
|
|
restart: 'unless-stopped',
|
|
running: true,
|
|
command: 'sleep infinity',
|
|
stop_signal: 'SIGKILL',
|
|
networks: ['default'],
|
|
labels: {},
|
|
environment: {},
|
|
},
|
|
'2': {
|
|
image: 'alpine',
|
|
imageId: 12,
|
|
serviceName: 'two',
|
|
restart: 'unless-stopped',
|
|
running: true,
|
|
command: 'sleep infinity',
|
|
stop_signal: 'SIGKILL',
|
|
networks: ['default'],
|
|
labels: {},
|
|
environment: {},
|
|
},
|
|
},
|
|
networks: {},
|
|
volumes: {},
|
|
},
|
|
},
|
|
});
|
|
|
|
const state = await getCurrentState();
|
|
expect(
|
|
state.apps['123'].services.map((s: any) => s.serviceName),
|
|
).to.deep.equal(['one', 'two']);
|
|
|
|
const containers = await docker.listContainers();
|
|
expect(
|
|
containers.map(({ Names, State }) => ({ Name: Names[0], State })),
|
|
).to.have.deep.members([
|
|
{ Name: '/one_11_1_deadbeef', State: 'running' },
|
|
{ Name: '/two_12_1_deadbeef', State: 'running' },
|
|
]);
|
|
});
|
|
|
|
// This test recovery from issue #1576, where a device running a service from the target release
|
|
// would not stop the service even if there were still network and container changes to be applied
|
|
it('always stops running services depending on a network being changed', async () => {
|
|
// Install part of the target release
|
|
await setTargetState({
|
|
config: {},
|
|
apps: {
|
|
'123': {
|
|
name: 'test-app',
|
|
commit: 'deadca1f',
|
|
releaseId: 2,
|
|
services: {
|
|
'1': {
|
|
image: 'alpine:latest',
|
|
imageId: 21,
|
|
serviceName: 'one',
|
|
restart: 'unless-stopped',
|
|
running: true,
|
|
command: 'sleep infinity',
|
|
stop_signal: 'SIGKILL',
|
|
labels: {},
|
|
environment: {},
|
|
},
|
|
},
|
|
networks: {},
|
|
volumes: {},
|
|
},
|
|
},
|
|
});
|
|
|
|
const state = await getCurrentState();
|
|
expect(
|
|
state.apps['123'].services.map((s: any) => s.serviceName),
|
|
).to.deep.equal(['one']);
|
|
|
|
const containers = await docker.listContainers();
|
|
expect(
|
|
containers.map(({ Names, State }) => ({ Name: Names[0], State })),
|
|
).to.have.deep.members([{ Name: '/one_21_2_deadca1f', State: 'running' }]);
|
|
const containerIds = containers.map(({ Id }) => Id);
|
|
|
|
await setTargetState({
|
|
config: {},
|
|
apps: {
|
|
'123': {
|
|
name: 'test-app',
|
|
commit: 'deadca1f',
|
|
releaseId: 2,
|
|
services: {
|
|
'1': {
|
|
image: 'alpine:latest',
|
|
imageId: 21,
|
|
serviceName: 'one',
|
|
restart: 'unless-stopped',
|
|
running: true,
|
|
command: 'sleep infinity',
|
|
stop_signal: 'SIGKILL',
|
|
networks: ['default'],
|
|
labels: {},
|
|
environment: {},
|
|
},
|
|
'2': {
|
|
image: 'alpine:latest',
|
|
imageId: 22,
|
|
serviceName: 'two',
|
|
restart: 'unless-stopped',
|
|
running: true,
|
|
command: 'sh -c "echo two && sleep infinity"',
|
|
stop_signal: 'SIGKILL',
|
|
networks: ['default'],
|
|
labels: {},
|
|
environment: {},
|
|
},
|
|
},
|
|
networks: {
|
|
default: {
|
|
driver: 'bridge',
|
|
ipam: {
|
|
config: [
|
|
{ gateway: '192.168.91.1', subnet: '192.168.91.0/24' },
|
|
],
|
|
driver: 'default',
|
|
},
|
|
},
|
|
},
|
|
volumes: {},
|
|
},
|
|
},
|
|
});
|
|
|
|
const updatedContainers = await docker.listContainers();
|
|
expect(
|
|
updatedContainers.map(({ Names, State }) => ({ Name: Names[0], State })),
|
|
).to.have.deep.members([
|
|
{ Name: '/one_21_2_deadca1f', State: 'running' },
|
|
{ Name: '/two_22_2_deadca1f', State: 'running' },
|
|
]);
|
|
|
|
// Container ids must have changed
|
|
expect(updatedContainers.map(({ Id }) => Id)).to.not.have.members(
|
|
containerIds,
|
|
);
|
|
|
|
expect(await docker.getNetwork('123_default').inspect())
|
|
.to.have.property('IPAM')
|
|
.to.deep.equal({
|
|
Config: [{ Gateway: '192.168.91.1', Subnet: '192.168.91.0/24' }],
|
|
Driver: 'default',
|
|
Options: {},
|
|
});
|
|
});
|
|
|
|
it('updates an app with two services with a network change', async () => {
|
|
await setTargetState({
|
|
config: {},
|
|
apps: {
|
|
'123': {
|
|
name: 'test-app',
|
|
commit: 'deadbeef',
|
|
releaseId: 1,
|
|
services: {
|
|
'1': {
|
|
image: 'alpine:latest',
|
|
imageId: 11,
|
|
serviceName: 'one',
|
|
restart: 'unless-stopped',
|
|
running: true,
|
|
command: 'sleep infinity',
|
|
stop_signal: 'SIGKILL',
|
|
labels: {},
|
|
environment: {},
|
|
},
|
|
'2': {
|
|
image: 'alpine:latest',
|
|
imageId: 12,
|
|
serviceName: 'two',
|
|
restart: 'unless-stopped',
|
|
running: true,
|
|
command: 'sleep infinity',
|
|
labels: {},
|
|
environment: {},
|
|
},
|
|
},
|
|
networks: {},
|
|
volumes: {},
|
|
},
|
|
},
|
|
});
|
|
|
|
const state = await getCurrentState();
|
|
expect(
|
|
state.apps['123'].services.map((s: any) => s.serviceName),
|
|
).to.deep.equal(['one', 'two']);
|
|
|
|
const containers = await docker.listContainers();
|
|
expect(
|
|
containers.map(({ Names, State }) => ({ Name: Names[0], State })),
|
|
).to.have.deep.members([
|
|
{ Name: '/one_11_1_deadbeef', State: 'running' },
|
|
{ Name: '/two_12_1_deadbeef', State: 'running' },
|
|
]);
|
|
const containerIds = containers.map(({ Id }) => Id);
|
|
|
|
await setTargetState({
|
|
config: {},
|
|
apps: {
|
|
'123': {
|
|
name: 'test-app',
|
|
commit: 'deadca1f',
|
|
releaseId: 2,
|
|
services: {
|
|
'1': {
|
|
image: 'alpine:latest',
|
|
imageId: 21,
|
|
serviceName: 'one',
|
|
restart: 'unless-stopped',
|
|
running: true,
|
|
command: 'sleep infinity',
|
|
stop_signal: 'SIGKILL',
|
|
networks: ['default'],
|
|
labels: {},
|
|
environment: {},
|
|
},
|
|
'2': {
|
|
image: 'alpine:latest',
|
|
imageId: 22,
|
|
serviceName: 'two',
|
|
restart: 'unless-stopped',
|
|
running: true,
|
|
command: 'sh -c "echo two && sleep infinity"',
|
|
stop_signal: 'SIGKILL',
|
|
networks: ['default'],
|
|
labels: {},
|
|
environment: {},
|
|
},
|
|
},
|
|
networks: {
|
|
default: {
|
|
driver: 'bridge',
|
|
ipam: {
|
|
config: [
|
|
{ gateway: '192.168.91.1', subnet: '192.168.91.0/24' },
|
|
],
|
|
driver: 'default',
|
|
},
|
|
},
|
|
},
|
|
volumes: {},
|
|
},
|
|
},
|
|
});
|
|
|
|
const updatedContainers = await docker.listContainers();
|
|
expect(
|
|
updatedContainers.map(({ Names, State }) => ({ Name: Names[0], State })),
|
|
).to.have.deep.members([
|
|
{ Name: '/one_21_2_deadca1f', State: 'running' },
|
|
{ Name: '/two_22_2_deadca1f', State: 'running' },
|
|
]);
|
|
|
|
// Container ids must have changed
|
|
expect(updatedContainers.map(({ Id }) => Id)).to.not.have.members(
|
|
containerIds,
|
|
);
|
|
|
|
expect(await docker.getNetwork('123_default').inspect())
|
|
.to.have.property('IPAM')
|
|
.to.deep.equal({
|
|
Config: [{ Gateway: '192.168.91.1', Subnet: '192.168.91.0/24' }],
|
|
Driver: 'default',
|
|
Options: {},
|
|
});
|
|
});
|
|
});
|