mirror of
https://github.com/balena-io/balena-cli.git
synced 2024-12-28 01:28:51 +00:00
578de7bcd4
When using livepush, the CLI parses the build logs to obtain the stage image ids, which are necessary for properly running livepush. This process used to store the full log output in memory before parsing the logs for obtaining the stage ids. We have seen this cause issues before because of the excessive memory usage and it is one the suspects of #2165, which is blocking the update to Node 14 Change-type: patch
533 lines
16 KiB
TypeScript
533 lines
16 KiB
TypeScript
/**
|
|
* @license
|
|
* Copyright 2019-2020 Balena Ltd.
|
|
*
|
|
* Licensed under the Apache License, Version 2.0 (the "License");
|
|
* you may not use this file except in compliance with the License.
|
|
* You may obtain a copy of the License at
|
|
*
|
|
* http://www.apache.org/licenses/LICENSE-2.0
|
|
*
|
|
* Unless required by applicable law or agreed to in writing, software
|
|
* distributed under the License is distributed on an "AS IS" BASIS,
|
|
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
|
* See the License for the specific language governing permissions and
|
|
* limitations under the License.
|
|
*/
|
|
|
|
import * as chokidar from 'chokidar';
|
|
import type * as Dockerode from 'dockerode';
|
|
import * as fs from 'fs';
|
|
import Livepush, { ContainerNotRunningError } from 'livepush';
|
|
import * as _ from 'lodash';
|
|
import * as path from 'path';
|
|
import type { Composition } from '@balena/compose/dist/parse';
|
|
import type { BuildTask } from '@balena/compose/dist/multibuild';
|
|
|
|
import { instanceOf } from '../../errors';
|
|
import Logger = require('../logger');
|
|
|
|
import { Dockerfile } from 'livepush';
|
|
import type DeviceAPI from './api';
|
|
import type { DeviceInfo, Status } from './api';
|
|
import {
|
|
DeviceDeployOptions,
|
|
generateTargetState,
|
|
rebuildSingleTask,
|
|
} from './deploy';
|
|
import { BuildError } from './errors';
|
|
import { getServiceColourFn } from './logs';
|
|
import { delay } from '../helpers';
|
|
|
|
// How often do we want to check the device state
|
|
// engine has settled (delay in ms)
|
|
const DEVICE_STATUS_SETTLE_CHECK_INTERVAL = 1000;
|
|
|
|
const LIVEPUSH_DEBOUNCE_TIMEOUT = 2000;
|
|
|
|
interface MonitoredContainer {
|
|
context: string;
|
|
livepush: Livepush;
|
|
monitor: chokidar.FSWatcher;
|
|
containerId: string;
|
|
}
|
|
|
|
type StageImageIDs = Dictionary<string[]>;
|
|
|
|
export interface LivepushOpts {
|
|
buildContext: string;
|
|
composition: Composition;
|
|
buildTasks: BuildTask[];
|
|
docker: Dockerode;
|
|
api: DeviceAPI;
|
|
logger: Logger;
|
|
imageIds: StageImageIDs;
|
|
deployOpts: DeviceDeployOptions;
|
|
}
|
|
|
|
export class LivepushManager {
|
|
private lastDeviceStatus: Status | null = null;
|
|
private containers: Dictionary<MonitoredContainer> = {};
|
|
private dockerfilePaths: Dictionary<string[]> = {};
|
|
private deviceInfo: DeviceInfo;
|
|
private deployOpts: DeviceDeployOptions;
|
|
|
|
private buildContext: string;
|
|
private composition: Composition;
|
|
private buildTasks: BuildTask[];
|
|
private docker: Dockerode;
|
|
private api: DeviceAPI;
|
|
private logger: Logger;
|
|
private imageIds: StageImageIDs;
|
|
|
|
// A map of service names to events waiting
|
|
private updateEventsWaiting: Dictionary<string[]> = {};
|
|
private deleteEventsWaiting: Dictionary<string[]> = {};
|
|
|
|
private rebuildsRunning: Dictionary<boolean> = {};
|
|
private rebuildRunningIds: Dictionary<string> = {};
|
|
private rebuildsCancelled: Dictionary<boolean> = {};
|
|
|
|
public constructor(opts: LivepushOpts) {
|
|
this.buildContext = opts.buildContext;
|
|
this.composition = opts.composition;
|
|
this.buildTasks = opts.buildTasks;
|
|
this.docker = opts.docker;
|
|
this.api = opts.api;
|
|
this.logger = opts.logger;
|
|
this.deployOpts = opts.deployOpts;
|
|
this.imageIds = opts.imageIds;
|
|
}
|
|
|
|
public async init(): Promise<void> {
|
|
this.deviceInfo = await this.api.getDeviceInformation();
|
|
this.logger.logLivepush('Waiting for device state to settle...');
|
|
// The first thing we need to do is let the state 'settle',
|
|
// so that all of the containers are running and ready to
|
|
// be livepush'd into
|
|
await this.awaitDeviceStateSettle();
|
|
// Split the composition into a load of differents paths
|
|
// which we can
|
|
this.logger.logLivepush('Device state settled');
|
|
|
|
// Prepare dockerignore data for file watcher
|
|
const { getDockerignoreByService } = await import('../ignore');
|
|
const { getServiceDirsFromComposition } = await import('../compose_ts');
|
|
const rootContext = path.resolve(this.buildContext);
|
|
const serviceDirsByService = await getServiceDirsFromComposition(
|
|
this.deployOpts.source,
|
|
this.composition,
|
|
);
|
|
const dockerignoreByService = await getDockerignoreByService(
|
|
this.deployOpts.source,
|
|
this.deployOpts.multiDockerignore,
|
|
serviceDirsByService,
|
|
);
|
|
|
|
// create livepush instances for each service
|
|
for (const serviceName of _.keys(this.composition.services)) {
|
|
const service = this.composition.services[serviceName];
|
|
const buildTask = _.find(this.buildTasks, { serviceName });
|
|
|
|
if (buildTask == null) {
|
|
throw new Error(
|
|
`Could not find a build task for service: ${serviceName}`,
|
|
);
|
|
}
|
|
|
|
// We only care about builds
|
|
if (service.build != null) {
|
|
if (buildTask.dockerfile == null) {
|
|
throw new Error(
|
|
`Could not detect dockerfile for service: ${serviceName}`,
|
|
);
|
|
}
|
|
const dockerfile = new Dockerfile(buildTask.dockerfile);
|
|
|
|
if (buildTask.dockerfilePath == null) {
|
|
// this is a bit of a hack as resin-bundle-resolve
|
|
// does not always export the dockerfilePath, this
|
|
// only happens when the dockerfile path is
|
|
// specified differently - this should be patched
|
|
// in resin-bundle-resolve
|
|
this.dockerfilePaths[buildTask.serviceName] =
|
|
this.getDockerfilePathFromTask(buildTask);
|
|
} else {
|
|
this.dockerfilePaths[buildTask.serviceName] = [
|
|
buildTask.dockerfilePath,
|
|
];
|
|
}
|
|
|
|
// Find the containerId from the device state
|
|
const container = _.find(this.lastDeviceStatus!.containers, {
|
|
serviceName,
|
|
});
|
|
if (container == null) {
|
|
return;
|
|
}
|
|
|
|
// path.resolve() converts to an absolute path, removes trailing slashes,
|
|
// and also converts forward slashes to backslashes on Windows.
|
|
const context = path.resolve(rootContext, service.build.context);
|
|
|
|
const livepush = await Livepush.init({
|
|
dockerfile,
|
|
context,
|
|
containerId: container.containerId,
|
|
stageImages: this.imageIds[serviceName],
|
|
docker: this.docker,
|
|
});
|
|
const buildVars = buildTask.buildMetadata.getBuildVarsForService(
|
|
buildTask.serviceName,
|
|
);
|
|
if (!_.isEmpty(buildVars)) {
|
|
livepush.setBuildArgs(buildVars);
|
|
}
|
|
|
|
this.assignLivepushOutputHandlers(serviceName, livepush);
|
|
|
|
this.updateEventsWaiting[serviceName] = [];
|
|
this.deleteEventsWaiting[serviceName] = [];
|
|
const addEvent = ($serviceName: string, changedPath: string) => {
|
|
this.logger.logDebug(
|
|
`Got an add filesystem event for service: ${$serviceName}. File: ${changedPath}`,
|
|
);
|
|
const eventQueue = this.updateEventsWaiting[$serviceName];
|
|
eventQueue.push(changedPath);
|
|
this.getDebouncedEventHandler($serviceName)();
|
|
};
|
|
|
|
const monitor = this.setupFilesystemWatcher(
|
|
serviceName,
|
|
rootContext,
|
|
context,
|
|
addEvent,
|
|
dockerignoreByService,
|
|
this.deployOpts.multiDockerignore,
|
|
);
|
|
|
|
this.containers[serviceName] = {
|
|
livepush,
|
|
context,
|
|
monitor,
|
|
containerId: container.containerId,
|
|
};
|
|
|
|
this.rebuildsRunning[serviceName] = false;
|
|
this.rebuildsCancelled[serviceName] = false;
|
|
}
|
|
}
|
|
}
|
|
|
|
/** Delete intermediate build containers from the device */
|
|
public async cleanup() {
|
|
this.logger.logLivepush('Cleaning up device...');
|
|
await Promise.all(
|
|
_.map(this.containers, (container) =>
|
|
container.livepush.cleanupIntermediateContainers(),
|
|
),
|
|
);
|
|
this.logger.logDebug('Cleaning up done.');
|
|
}
|
|
|
|
protected setupFilesystemWatcher(
|
|
serviceName: string,
|
|
rootContext: string,
|
|
serviceContext: string,
|
|
changedPathHandler: (serviceName: string, changedPath: string) => void,
|
|
dockerignoreByService: {
|
|
[serviceName: string]: import('@balena/dockerignore').Ignore;
|
|
},
|
|
multiDockerignore: boolean,
|
|
): chokidar.FSWatcher {
|
|
const contextForDockerignore = multiDockerignore
|
|
? serviceContext
|
|
: rootContext;
|
|
const dockerignore = dockerignoreByService[serviceName];
|
|
// TODO: Memoize this for services that share a context
|
|
const monitor = chokidar.watch('.', {
|
|
cwd: serviceContext,
|
|
followSymlinks: true,
|
|
ignoreInitial: true,
|
|
ignored: (filePath: string, stats: fs.Stats | undefined) => {
|
|
if (!stats) {
|
|
try {
|
|
// sync because chokidar defines a sync interface
|
|
stats = fs.lstatSync(filePath);
|
|
} catch (err) {
|
|
// OK: the file may have been deleted. See also:
|
|
// https://github.com/paulmillr/chokidar/blob/3.4.3/lib/fsevents-handler.js#L326-L328
|
|
// https://github.com/paulmillr/chokidar/blob/3.4.3/lib/nodefs-handler.js#L364
|
|
}
|
|
}
|
|
if (stats && !stats.isFile() && !stats.isSymbolicLink()) {
|
|
// never ignore directories for compatibility with
|
|
// dockerignore exclusion patterns
|
|
return !stats.isDirectory();
|
|
}
|
|
const relPath = path.relative(contextForDockerignore, filePath);
|
|
return dockerignore.ignores(relPath);
|
|
},
|
|
});
|
|
monitor.on('add', (changedPath: string) =>
|
|
changedPathHandler(serviceName, changedPath),
|
|
);
|
|
monitor.on('change', (changedPath: string) =>
|
|
changedPathHandler(serviceName, changedPath),
|
|
);
|
|
monitor.on('unlink', (changedPath: string) =>
|
|
changedPathHandler(serviceName, changedPath),
|
|
);
|
|
return monitor;
|
|
}
|
|
|
|
/** Stop the filesystem watcher, allowing the Node process to exit gracefully */
|
|
public close() {
|
|
for (const container of Object.values(this.containers)) {
|
|
container.monitor.close().catch((err) => {
|
|
if (process.env.DEBUG) {
|
|
this.logger.logDebug(`chokidar.close() ${err.message}`);
|
|
}
|
|
});
|
|
}
|
|
}
|
|
|
|
public static preprocessDockerfile(content: string): string {
|
|
return new Dockerfile(content).generateLiveDockerfile();
|
|
}
|
|
|
|
private async awaitDeviceStateSettle(): Promise<void> {
|
|
// Cache the state to avoid unnecessary calls
|
|
this.lastDeviceStatus = await this.api.getStatus();
|
|
|
|
if (this.lastDeviceStatus.appState === 'applied') {
|
|
return;
|
|
}
|
|
|
|
this.logger.logDebug(
|
|
`Device state not settled, retrying in ${DEVICE_STATUS_SETTLE_CHECK_INTERVAL}ms`,
|
|
);
|
|
await delay(DEVICE_STATUS_SETTLE_CHECK_INTERVAL);
|
|
await this.awaitDeviceStateSettle();
|
|
}
|
|
|
|
private async handleFSEvents(serviceName: string): Promise<void> {
|
|
const updated = this.updateEventsWaiting[serviceName];
|
|
const deleted = this.deleteEventsWaiting[serviceName];
|
|
this.updateEventsWaiting[serviceName] = [];
|
|
this.deleteEventsWaiting[serviceName] = [];
|
|
|
|
// First we detect if the file changed is the Dockerfile
|
|
// used to build the service
|
|
if (
|
|
_.some(this.dockerfilePaths[serviceName], (name) =>
|
|
_.some(updated, (changed) => name === changed),
|
|
)
|
|
) {
|
|
this.logger.logLivepush(
|
|
`Detected Dockerfile change, performing full rebuild of service ${serviceName}`,
|
|
);
|
|
await this.handleServiceRebuild(serviceName);
|
|
return;
|
|
}
|
|
|
|
// Work out if we need to perform any changes on this container
|
|
const livepush = this.containers[serviceName].livepush;
|
|
|
|
if (!livepush.livepushNeeded(updated, deleted)) {
|
|
return;
|
|
}
|
|
|
|
this.logger.logLivepush(
|
|
`Detected changes for container ${serviceName}, updating...`,
|
|
);
|
|
|
|
try {
|
|
await livepush.performLivepush(updated, deleted);
|
|
} catch (e) {
|
|
this.logger.logError(
|
|
`An error occured whilst trying to perform a livepush: `,
|
|
);
|
|
if (instanceOf(e, ContainerNotRunningError)) {
|
|
this.logger.logError(' Livepush container not running');
|
|
} else {
|
|
this.logger.logError(` ${e.message}`);
|
|
}
|
|
this.logger.logDebug(e.stack);
|
|
}
|
|
}
|
|
|
|
private async handleServiceRebuild(serviceName: string): Promise<void> {
|
|
if (this.rebuildsRunning[serviceName]) {
|
|
this.logger.logLivepush(
|
|
`Cancelling ongoing rebuild for service ${serviceName}`,
|
|
);
|
|
await this.cancelRebuild(serviceName);
|
|
while (this.rebuildsCancelled[serviceName]) {
|
|
await delay(1000);
|
|
}
|
|
}
|
|
|
|
this.rebuildsRunning[serviceName] = true;
|
|
try {
|
|
const buildTask = _.find(this.buildTasks, { serviceName });
|
|
if (buildTask == null) {
|
|
throw new Error(
|
|
`Could not find a build task for service ${serviceName}`,
|
|
);
|
|
}
|
|
|
|
let stageImages: string[];
|
|
try {
|
|
stageImages = await rebuildSingleTask(
|
|
serviceName,
|
|
this.docker,
|
|
this.logger,
|
|
this.deviceInfo,
|
|
this.composition,
|
|
this.buildContext,
|
|
this.deployOpts,
|
|
(id) => {
|
|
this.rebuildRunningIds[serviceName] = id;
|
|
},
|
|
);
|
|
} catch (e) {
|
|
if (!(e instanceof BuildError)) {
|
|
throw e;
|
|
}
|
|
|
|
if (this.rebuildsCancelled[serviceName]) {
|
|
return;
|
|
}
|
|
|
|
this.logger.logError(
|
|
`Rebuild of service ${serviceName} failed!\n Error: ${e.getServiceError(
|
|
serviceName,
|
|
)}`,
|
|
);
|
|
return;
|
|
} finally {
|
|
delete this.rebuildRunningIds[serviceName];
|
|
}
|
|
|
|
// If the build has been cancelled, exit early
|
|
if (this.rebuildsCancelled[serviceName]) {
|
|
return;
|
|
}
|
|
|
|
// Let's first delete the container from the device
|
|
const containerId = await this.api.getContainerId(serviceName);
|
|
await this.docker.getContainer(containerId).remove({ force: true });
|
|
const currentState = await this.api.getTargetState();
|
|
// If we re-apply the target state, the supervisor
|
|
// should recreate the container
|
|
await this.api.setTargetState(
|
|
generateTargetState(currentState, this.composition, [buildTask], {}),
|
|
);
|
|
|
|
await this.awaitDeviceStateSettle();
|
|
|
|
const instance = this.containers[serviceName];
|
|
// Get the new container
|
|
const container = _.find(this.lastDeviceStatus!.containers, {
|
|
serviceName,
|
|
});
|
|
if (container == null) {
|
|
throw new Error(
|
|
`Could not find new container for service ${serviceName}`,
|
|
);
|
|
}
|
|
|
|
const dockerfile = new Dockerfile(buildTask.dockerfile!);
|
|
|
|
instance.livepush = await Livepush.init({
|
|
dockerfile,
|
|
context: buildTask.context!,
|
|
containerId: container.containerId,
|
|
stageImages,
|
|
docker: this.docker,
|
|
});
|
|
this.assignLivepushOutputHandlers(serviceName, instance.livepush);
|
|
} catch (e) {
|
|
this.logger.logError(`There was an error rebuilding the service: ${e}`);
|
|
} finally {
|
|
this.rebuildsRunning[serviceName] = false;
|
|
this.rebuildsCancelled[serviceName] = false;
|
|
}
|
|
}
|
|
|
|
private async cancelRebuild(serviceName: string) {
|
|
this.rebuildsCancelled[serviceName] = true;
|
|
|
|
// If we have a container id of the current build,
|
|
// attempt to kill it
|
|
if (this.rebuildRunningIds[serviceName] != null) {
|
|
try {
|
|
await this.docker
|
|
.getContainer(this.rebuildRunningIds[serviceName])
|
|
.remove({ force: true });
|
|
await this.containers[serviceName].livepush.cancel();
|
|
} catch {
|
|
// No need to do anything here
|
|
}
|
|
}
|
|
}
|
|
|
|
private assignLivepushOutputHandlers(
|
|
serviceName: string,
|
|
livepush: Livepush,
|
|
) {
|
|
const msgString = (msg: string) =>
|
|
`[${getServiceColourFn(serviceName)(serviceName)}] ${msg}`;
|
|
const log = (msg: string) => this.logger.logLivepush(msgString(msg));
|
|
const error = (msg: string) => this.logger.logError(msgString(msg));
|
|
const debugLog = (msg: string) => this.logger.logDebug(msgString(msg));
|
|
|
|
livepush.on('commandExecute', (command) =>
|
|
log(`Executing command: \`${command.command}\``),
|
|
);
|
|
livepush.on('commandOutput', (output) =>
|
|
log(` ${output.output.data.toString()}`),
|
|
);
|
|
livepush.on('commandReturn', ({ returnCode, command }) => {
|
|
if (returnCode !== 0) {
|
|
error(` Command ${command} failed with exit code: ${returnCode}`);
|
|
} else {
|
|
debugLog(`Command ${command} exited successfully`);
|
|
}
|
|
});
|
|
livepush.on('containerRestart', () => {
|
|
log('Restarting service...');
|
|
});
|
|
livepush.on('cancel', () => {
|
|
log('Cancelling current livepush...');
|
|
});
|
|
}
|
|
|
|
private getDockerfilePathFromTask(task: BuildTask): string[] {
|
|
switch (task.projectType) {
|
|
case 'Standard Dockerfile':
|
|
return ['Dockerfile'];
|
|
case 'Dockerfile.template':
|
|
return ['Dockerfile.template'];
|
|
case 'Architecture-specific Dockerfile':
|
|
return [
|
|
`Dockerfile.${this.deviceInfo.arch}`,
|
|
`Dockerfile.${this.deviceInfo.deviceType}`,
|
|
];
|
|
default:
|
|
return [];
|
|
}
|
|
}
|
|
|
|
// For each service, get a debounced function
|
|
private getDebouncedEventHandler = _.memoize((serviceName: string) => {
|
|
return _.debounce(
|
|
() => this.handleFSEvents(serviceName),
|
|
LIVEPUSH_DEBOUNCE_TIMEOUT,
|
|
);
|
|
});
|
|
}
|
|
|
|
export default LivepushManager;
|