1
0
mirror of https://github.com/laurent22/joplin.git synced 2024-12-03 08:35:29 +02:00

Server: Improved config and support for Docker

This commit is contained in:
Laurent Cozic 2021-01-18 10:13:26 +00:00
parent 59fe4a2193
commit 0d2bf6d787
34 changed files with 448 additions and 329 deletions

View File

@ -1,9 +1,26 @@
# Example of local config, for development: # =============================================================================
# PRODUCTION CONFIG EXAMPLE
# -----------------------------------------------------------------------------
# By default it will use SQLite, but that's mostly to test and evaluate the
# server. So you'll want to specify db connection settings to use Postgres.
# =============================================================================
# #
# JOPLIN_BASE_URL=http://localhost:22300 # APP_BASE_URL=https://example.com/joplin
# JOPLIN_PORT=22300 # APP_PORT=22300
#
# DB_CLIENT=pg
# POSTGRES_PASSWORD=joplin
# POSTGRES_DATABASE=joplin
# POSTGRES_USER=joplin
# POSTGRES_PORT=5432
# POSTGRES_HOST=localhost
# Example of config for production: # =============================================================================
# DEV CONFIG EXAMPLE
# -----------------------------------------------------------------------------
# Example of local config, for development. In dev mode, you would usually use
# SQLite so database settings are not needed.
# =============================================================================
# #
# JOPLIN_BASE_URL=https://example.com/joplin # APP_BASE_URL=http://localhost:22300
# JOPLIN_PORT=22300 # APP_PORT=22300

View File

@ -1,3 +0,0 @@
FROM postgres:13.1
EXPOSE 5432

View File

@ -16,8 +16,15 @@ WORKDIR /home/$user
RUN mkdir /home/$user/logs RUN mkdir /home/$user/logs
# Install the root scripts but don't run postinstall (which would bootstrap
# and build TypeScript files, but we don't have the TypeScript files at
# this point)
COPY --chown=$user:$user package*.json ./
RUN npm install --ignore-scripts
# To take advantage of the Docker cache, we first copy all the package.json # To take advantage of the Docker cache, we first copy all the package.json
# and package-lock.json files, as they rarely change? and then bootstrap # and package-lock.json files, as they rarely change, and then bootstrap
# all the packages. # all the packages.
# #
# Note that bootstrapping the packages will run all the postinstall # Note that bootstrapping the packages will run all the postinstall
@ -27,19 +34,10 @@ RUN mkdir /home/$user/logs
# We can't run boostrap with "--ignore-scripts" because that would # We can't run boostrap with "--ignore-scripts" because that would
# prevent certain sub-packages, such as sqlite3, from being built # prevent certain sub-packages, such as sqlite3, from being built
COPY --chown=$user:$user package*.json ./
# Install the root scripts but don't run postinstall (which would bootstrap
# and build TypeScript files, but we don't have the TypeScript files at
# this point)
RUN npm install --ignore-scripts
COPY --chown=$user:$user packages/fork-sax/package*.json ./packages/fork-sax/ COPY --chown=$user:$user packages/fork-sax/package*.json ./packages/fork-sax/
COPY --chown=$user:$user packages/lib/package*.json ./packages/lib/
COPY --chown=$user:$user packages/renderer/package*.json ./packages/renderer/ COPY --chown=$user:$user packages/renderer/package*.json ./packages/renderer/
COPY --chown=$user:$user packages/tools/package*.json ./packages/tools/ COPY --chown=$user:$user packages/tools/package*.json ./packages/tools/
COPY --chown=$user:$user packages/server/package*.json ./packages/server/ COPY --chown=$user:$user packages/lib/package*.json ./packages/lib/
COPY --chown=$user:$user lerna.json . COPY --chown=$user:$user lerna.json .
COPY --chown=$user:$user tsconfig.json . COPY --chown=$user:$user tsconfig.json .
@ -50,22 +48,29 @@ COPY --chown=$user:$user packages/turndown ./packages/turndown
COPY --chown=$user:$user packages/turndown-plugin-gfm ./packages/turndown-plugin-gfm COPY --chown=$user:$user packages/turndown-plugin-gfm ./packages/turndown-plugin-gfm
COPY --chown=$user:$user packages/fork-htmlparser2 ./packages/fork-htmlparser2 COPY --chown=$user:$user packages/fork-htmlparser2 ./packages/fork-htmlparser2
RUN ls -la /home/$user
# Then bootstrap only, without compiling the TypeScript files # Then bootstrap only, without compiling the TypeScript files
RUN npm run bootstrap RUN npm run bootstrap
# We have a separate step for the server files because they are more likely to
# change.
COPY --chown=$user:$user packages/server/package*.json ./packages/server/
RUN npm run bootstrapServerOnly
# Now copy the source files. Put lib and server last as they are more likely to change.
COPY --chown=$user:$user packages/fork-sax ./packages/fork-sax COPY --chown=$user:$user packages/fork-sax ./packages/fork-sax
COPY --chown=$user:$user packages/lib ./packages/lib
COPY --chown=$user:$user packages/renderer ./packages/renderer COPY --chown=$user:$user packages/renderer ./packages/renderer
COPY --chown=$user:$user packages/tools ./packages/tools COPY --chown=$user:$user packages/tools ./packages/tools
COPY --chown=$user:$user packages/lib ./packages/lib
COPY --chown=$user:$user packages/server ./packages/server COPY --chown=$user:$user packages/server ./packages/server
# Finally build everything, in particular the TypeScript files. # Finally build everything, in particular the TypeScript files.
RUN npm run build RUN npm run build
EXPOSE ${JOPLIN_PORT} ENV RUNNING_IN_DOCKER=1
EXPOSE ${APP_PORT}
CMD [ "npm", "--prefix", "packages/server", "start" ] CMD [ "npm", "--prefix", "packages/server", "start" ]

15
docker-compose.db-dev.yml Normal file
View File

@ -0,0 +1,15 @@
# For development this compose file starts the database only. The app can then
# be started using `npm run start-dev`, which is useful for development, because
# it means the app Docker file doesn't have to be rebuilt on each change.
version: '3'
services:
db:
image: postgres:13.1
ports:
- "5432:5432"
environment:
- POSTGRES_PASSWORD=joplin
- POSTGRES_USER=joplin
- POSTGRES_DB=joplin

View File

@ -1,28 +1,27 @@
# For development, the easiest might be to only start the Postgres container and # This compose file can be used in development to run both the database and app
# run the app directly with `npm start`. Or use sqlite3. # within Docker.
version: '3' version: '3'
services: services:
# app: app:
# build:
# context: .
# dockerfile: Dockerfile.server-dev
# ports:
# - "22300:22300"
# # volumes:
# # - ./packages/server/:/var/www/joplin/packages/server/
# # - /var/www/joplin/packages/server/node_modules/
db:
build: build:
context: . context: .
dockerfile: Dockerfile.db dockerfile: Dockerfile.server
ports:
- "22300:22300"
environment:
- DB_CLIENT=pg
- POSTGRES_PASSWORD=joplin
- POSTGRES_DATABASE=joplin
- POSTGRES_USER=joplin
- POSTGRES_PORT=5432
- POSTGRES_HOST=localhost
db:
image: postgres:13.1
ports: ports:
- "5432:5432" - "5432:5432"
environment: environment:
# TODO: Considering the database is only exposed to the
# application, and not to the outside world, is there a need to
# pick a secure password?
- POSTGRES_PASSWORD=joplin - POSTGRES_PASSWORD=joplin
- POSTGRES_USER=joplin - POSTGRES_USER=joplin
- POSTGRES_DB=joplin - POSTGRES_DB=joplin

View File

@ -1,40 +1,34 @@
# This is a sample docker-compose file that can be used to run Joplin Server
# along with a PostgreSQL server.
#
# All environment variables are optional. If you don't set them, you will get a
# warning from docker-compose, however the app should use working defaults.
version: '3' version: '3'
services: services:
app:
environment:
- JOPLIN_BASE_URL=${JOPLIN_BASE_URL}
- JOPLIN_PORT=${JOPLIN_PORT}
restart: unless-stopped
build:
context: .
dockerfile: Dockerfile.server
ports:
- "${JOPLIN_PORT}:${JOPLIN_PORT}"
# volumes:
# # Mount the server directory so that it's possible to edit file
# # while the container is running. However don't mount the
# # node_modules directory which will be specific to the Docker
# # image (eg native modules will be built for Ubuntu, while the
# # container might be running in Windows)
# # https://stackoverflow.com/a/37898591/561309
# - ./packages/server:/home/joplin/packages/server
# - /home/joplin/packages/server/node_modules/
db: db:
restart: unless-stopped image: postgres:13.1
# By default, the Postgres image saves the data to a Docker volume,
# so it persists whenever the server is restarted using
# `docker-compose up`. Note that it would however be deleted when
# running `docker-compose down`.
build:
context: .
dockerfile: Dockerfile.db
ports: ports:
- "5432:5432" - "5432:5432"
restart: unless-stopped
environment: environment:
# TODO: Considering the database is only exposed to the - APP_PORT=22300
# application, and not to the outside world, is there a need to - POSTGRES_PASSWORD=${POSTGRES_PASSWORD}
# pick a secure password? - POSTGRES_USER=${POSTGRES_USER}
- POSTGRES_PASSWORD=joplin - POSTGRES_DB=${POSTGRES_DATABASE}
- POSTGRES_USER=joplin app:
- POSTGRES_DB=joplin image: joplin/server:latest
depends_on:
- db
ports:
- "22300:22300"
restart: unless-stopped
environment:
- APP_BASE_URL=${APP_BASE_URL}
- DB_CLIENT=pg
- POSTGRES_PASSWORD=${POSTGRES_PASSWORD}
- POSTGRES_DATABASE=${POSTGRES_DATABASE}
- POSTGRES_USER=${POSTGRES_USER}
- POSTGRES_PORT=${POSTGRES_PORT}
- POSTGRES_HOST=db

View File

@ -8,6 +8,7 @@
"license": "MIT", "license": "MIT",
"scripts": { "scripts": {
"bootstrap": "lerna bootstrap --no-ci", "bootstrap": "lerna bootstrap --no-ci",
"bootstrapServerOnly": "lerna bootstrap --no-ci --include-dependents --include-dependencies --scope @joplin/server",
"bootstrapIgnoreScripts": "lerna bootstrap --ignore-scripts --no-ci", "bootstrapIgnoreScripts": "lerna bootstrap --ignore-scripts --no-ci",
"build": "lerna run build && npm run tsc", "build": "lerna run build && npm run tsc",
"buildApiDoc": "npm start --prefix=packages/app-cli -- apidoc ../../readme/api/references/rest_api.md", "buildApiDoc": "npm start --prefix=packages/app-cli -- apidoc ../../readme/api/references/rest_api.md",

View File

@ -4,64 +4,86 @@
First copy `.env-sample` to `.env` and edit the values in there: First copy `.env-sample` to `.env` and edit the values in there:
- `JOPLIN_BASE_URL`: This is the base public URL where the service will be running. For example, if you want it to run from `https://example.com/joplin`, this is what you should set the URL to. The base URL can include the port. - `APP_BASE_URL`: This is the base public URL where the service will be running. For example, if you want it to run from `https://example.com/joplin`, this is what you should set the URL to. The base URL can include the port.
- `JOPLIN_PORT`: The local port on which the Docker container will listen. You would typically map this port to 443 (TLS) with a reverse proxy. - `APP_PORT`: The local port on which the Docker container will listen. You would typically map this port to 443 (TLS) with a reverse proxy.
## Install application ## Running the server
To start the server with default configuration, run:
```shell ```shell
wget https://github.com/laurent22/joplin/archive/server-v1.6.4.tar.gz docker run --env-file .env -p 22300:22300 joplin/server:latest
tar xzvf server-v1.6.4.tar.gz
mv joplin-server-v1.6.4 joplin-server
cd joplin-server
docker-compose --file docker-compose.server.yml up --detach
``` ```
This will start the server, which will listen on port **22300** on **localhost**. This will start the server, which will listen on port **22300** on **localhost**. By default it will use SQLite, which allows you to test the app without setting up a database. To run it for production though, you'll want to connect the container to a database, as described below.
Due to the restart policy defined in the docker-compose file, the server will be restarted automatically whenever the host reboots. ## Setup the database
You can setup the container to either use an existing PostgreSQL server, or connect it to a new one using docker-compose
### Using an existing PostgreSQL server
To use an existing PostgresSQL server, set the following environment variables in the .env file:
```conf
DB_CLIENT=pg
POSTGRES_PASSWORD=joplin
POSTGRES_DATABASE=joplin
POSTGRES_USER=joplin
POSTGRES_PORT=5432
POSTGRES_HOST=localhost
```
Make sure that the provided database and user exist as the server will not create them.
### Using docker-compose
A [sample docker-compose file](https://github.com/laurent22/joplin/blob/dev/docker-compose.server.yml
) is available to show how to use Docker to install both the database and server and connect them:
## Setup reverse proxy ## Setup reverse proxy
You will then need to expose this server to the internet by setting up a reverse proxy, and that will depend on how your server is currently configured, and whether you already have Nginx or Apache running: Once Joplin Server is running, you will then need to expose it to the internet by setting up a reverse proxy, and that will depend on how your server is currently configured, and whether you already have Nginx or Apache running:
- [Apache Reverse Proxy](https://httpd.apache.org/docs/current/mod/mod_proxy.html) - [Apache Reverse Proxy](https://httpd.apache.org/docs/current/mod/mod_proxy.html)
- [Nginx Reverse Proxy](https://docs.nginx.com/nginx/admin-guide/web-server/reverse-proxy/) - [Nginx Reverse Proxy](https://docs.nginx.com/nginx/admin-guide/web-server/reverse-proxy/)
## Setup admin user ## Setup the website
For the following instructions, we'll assume that the Joplin server is running on `https://example.com/joplin`. Once the server is exposed to the internet, you can open the admin UI and get it ready for synchronisation. For the following instructions, we'll assume that the Joplin server is running on `https://example.com/joplin`.
By default, the instance will be setup with an admin user with email **admin@localhost** and password **admin** and you should change this by opening the admin UI. To do so, open `https://example.com/joplin/login`. From there, go to Profile and change the admin password. ### Secure the admin user
## Setup a user for sync By default, the instance will be setup with an admin user with email **admin@localhost** and password **admin** and you should change this. To do so, open `https://example.com/joplin/login` and login as admin. Then go to the Profile section and change the admin password.
While the admin user can be used for synchronisation, it is recommended to create a separate non-admin user for it. To do, open the admin UI and navigate to the Users page - from there you can create a new user. ### Create a user for sync
Once this is done, you can use the email and password you specified to sync this user account with your Joplin clients. While the admin user can be used for synchronisation, it is recommended to create a separate non-admin user for it. To do so, navigate to the Users page - from there you can create a new user. Once this is done, you can use the email and password you specified to sync this user account with your Joplin clients.
## Checking the logs ## Checking the logs
Checking the log can be done the standard Docker way: Checking the log can be done the standard Docker way:
```shell ```bash
# With Docker:
docker logs --follow CONTAINER
# With docker-compose:
docker-compose --file docker-compose.server.yml logs docker-compose --file docker-compose.server.yml logs
``` ```
# Set up for development # Setup for development
## Setting up the database ## Setup up the database
### SQLite ### SQLite
The server supports SQLite for development and test units. To use it, open `src/config-dev.ts` and uncomment the sqlite3 config. By default the server supports SQLite for development, so nothing needs to be setup.
### PostgreSQL ### PostgreSQL
It's best to use PostgreSQL as this is what is used in production, however it requires Docker. To use Postgres, from the monorepo root, run `docker-compose --file docker-compose.server-dev.yml up`, which will start the PostgreSQL database.
To use it, from the monorepo root, run `docker-compose --file docker-compose.server-dev.yml up`, which will start the PostgreSQL database.
## Starting the server ## Starting the server
From `packages/server`, run `npm run start-dev` From `packages/server`, run `npm run start-dev`

View File

@ -1,6 +1,6 @@
{ {
"name": "@joplin/server", "name": "@joplin/server",
"version": "1.7.0", "version": "1.7.1",
"lockfileVersion": 1, "lockfileVersion": 1,
"requires": true, "requires": true,
"dependencies": { "dependencies": {
@ -5938,6 +5938,11 @@
"integrity": "sha512-1nh45deeb5olNY7eX82BkPO7SSxR5SSYJiPTrTdFUVYwAl8CKMA5N9PjTYkHiRjisVcxcQ1HXdLhx2qxxJzLNQ==", "integrity": "sha512-1nh45deeb5olNY7eX82BkPO7SSxR5SSYJiPTrTdFUVYwAl8CKMA5N9PjTYkHiRjisVcxcQ1HXdLhx2qxxJzLNQ==",
"dev": true "dev": true
}, },
"node-env-file": {
"version": "0.1.8",
"resolved": "https://registry.npmjs.org/node-env-file/-/node-env-file-0.1.8.tgz",
"integrity": "sha1-/Mt7BQ9zW1oz2p65N89vGrRX+2k="
},
"node-int64": { "node-int64": {
"version": "0.4.0", "version": "0.4.0",
"resolved": "https://registry.npmjs.org/node-int64/-/node-int64-0.4.0.tgz", "resolved": "https://registry.npmjs.org/node-int64/-/node-int64-0.4.0.tgz",

View File

@ -1,6 +1,6 @@
{ {
"name": "@joplin/server", "name": "@joplin/server",
"version": "1.7.0", "version": "1.7.1",
"private": true, "private": true,
"scripts": { "scripts": {
"start-dev": "nodemon --config nodemon.json dist/app.js --env dev", "start-dev": "nodemon --config nodemon.json dist/app.js --env dev",
@ -26,6 +26,7 @@
"markdown-it": "^12.0.4", "markdown-it": "^12.0.4",
"mustache": "^3.1.0", "mustache": "^3.1.0",
"nanoid": "^2.1.1", "nanoid": "^2.1.1",
"node-env-file": "^0.1.8",
"nodemon": "^2.0.6", "nodemon": "^2.0.6",
"pg": "^8.5.1", "pg": "^8.5.1",
"query-string": "^6.8.3", "query-string": "^6.8.3",

View File

@ -5,32 +5,30 @@ import * as Koa from 'koa';
import * as fs from 'fs-extra'; import * as fs from 'fs-extra';
import { argv } from 'yargs'; import { argv } from 'yargs';
import Logger, { LoggerWrapper, TargetType } from '@joplin/lib/Logger'; import Logger, { LoggerWrapper, TargetType } from '@joplin/lib/Logger';
import config, { initConfig, baseUrl } from './config'; import config, { initConfig, runningInDocker, EnvVariables } from './config';
import configDev from './config-dev';
import configProd from './config-prod';
import configBuildTypes from './config-buildTypes';
import { createDb, dropDb } from './tools/dbTools'; import { createDb, dropDb } from './tools/dbTools';
import { dropTables, connectDb, disconnectDb, migrateDb, waitForConnection } from './db'; import { dropTables, connectDb, disconnectDb, migrateDb, waitForConnection, sqliteFilePath } from './db';
import modelFactory from './models/factory'; import modelFactory from './models/factory';
import { AppContext, Config, Env } from './utils/types'; import { AppContext, Env } from './utils/types';
import FsDriverNode from '@joplin/lib/fs-driver-node'; import FsDriverNode from '@joplin/lib/fs-driver-node';
import routeHandler from './middleware/routeHandler'; import routeHandler from './middleware/routeHandler';
import notificationHandler from './middleware/notificationHandler'; import notificationHandler from './middleware/notificationHandler';
import ownerHandler from './middleware/ownerHandler'; import ownerHandler from './middleware/ownerHandler';
const nodeEnvFile = require('node-env-file');
const { shimInit } = require('@joplin/lib/shim-init-node.js'); const { shimInit } = require('@joplin/lib/shim-init-node.js');
shimInit(); shimInit();
const env: Env = argv.env as Env || Env.Prod; const env: Env = argv.env as Env || Env.Prod;
interface Configs { const envVariables: Record<Env, EnvVariables> = {
[name: string]: Config; dev: {
} SQLITE_DATABASE: 'dev',
},
const configs: Configs = { buildTypes: {
dev: configDev, SQLITE_DATABASE: 'buildTypes',
prod: configProd, },
buildTypes: configBuildTypes, prod: {}, // Actually get the env variables from the environment
}; };
let appLogger_: LoggerWrapper = null; let appLogger_: LoggerWrapper = null;
@ -52,11 +50,31 @@ app.use(ownerHandler);
app.use(notificationHandler); app.use(notificationHandler);
app.use(routeHandler); app.use(routeHandler);
async function main() { function markPasswords(o: Record<string, any>): Record<string, any> {
const configObject: Config = configs[env]; const output: Record<string, any> = {};
if (!configObject) throw new Error(`Invalid env: ${env}`);
initConfig(configObject); for (const k of Object.keys(o)) {
if (k.toLowerCase().includes('password')) {
output[k] = '********';
} else {
output[k] = o[k];
}
}
return output;
}
async function main() {
if (argv.envFile) {
nodeEnvFile(argv.envFile);
}
if (!envVariables[env]) throw new Error(`Invalid env: ${env}`);
initConfig({
...envVariables[env],
...process.env,
});
await fs.mkdirp(config().logDir); await fs.mkdirp(config().logDir);
Logger.fsDriver_ = new FsDriverNode(); Logger.fsDriver_ = new FsDriverNode();
@ -90,8 +108,11 @@ async function main() {
await createDb(config().database); await createDb(config().database);
} else { } else {
appLogger().info(`Starting server (${env}) on port ${config().port} and PID ${process.pid}...`); appLogger().info(`Starting server (${env}) on port ${config().port} and PID ${process.pid}...`);
appLogger().info('Public base URL:', baseUrl()); appLogger().info('Running in Docker:', runningInDocker());
appLogger().info('DB Config:', config().database); appLogger().info('Public base URL:', config().baseUrl);
appLogger().info('Log dir:', config().logDir);
appLogger().info('DB Config:', markPasswords(config().database));
if (config().database.client === 'sqlite3') appLogger().info('DB file:', sqliteFilePath(config().database.name));
const appContext = app.context as AppContext; const appContext = app.context as AppContext;
@ -104,13 +125,13 @@ async function main() {
appLogger().info('Connection check:', connectionCheckLogInfo); appLogger().info('Connection check:', connectionCheckLogInfo);
appContext.env = env; appContext.env = env;
appContext.db = connectionCheck.connection; appContext.db = connectionCheck.connection;
appContext.models = modelFactory(appContext.db, baseUrl()); appContext.models = modelFactory(appContext.db, config().baseUrl);
appContext.appLogger = appLogger; appContext.appLogger = appLogger;
appLogger().info('Migrating database...'); appLogger().info('Migrating database...');
await migrateDb(appContext.db); await migrateDb(appContext.db);
appLogger().info(`Call this for testing: \`curl ${baseUrl()}/api/ping\``); appLogger().info(`Call this for testing: \`curl ${config().baseUrl}/api/ping\``);
app.listen(config().port); app.listen(config().port);
} }

View File

@ -1,21 +0,0 @@
import { Config } from './utils/types';
import * as pathUtils from 'path';
const rootDir = pathUtils.dirname(__dirname);
const viewDir = `${pathUtils.dirname(__dirname)}/src/views`;
const envPort = Number(process.env.JOPLIN_PORT);
const config: Config = {
port: (envPort && !isNaN(envPort)) ? envPort : 22300,
viewDir: viewDir,
rootDir: rootDir,
layoutDir: `${viewDir}/layouts`,
logDir: `${rootDir}/logs`,
database: {
client: 'pg',
name: 'joplin',
},
};
export default config;

View File

@ -1,13 +0,0 @@
import { Config } from './utils/types';
import configBase from './config-base';
const config: Config = {
...configBase,
database: {
name: 'buildTypes',
client: 'sqlite3',
asyncStackTraces: true,
},
};
export default config;

View File

@ -1,22 +0,0 @@
import { Config } from './utils/types';
import configBase from './config-base';
const config: Config = {
...configBase,
database: {
name: 'dev',
client: 'sqlite3',
asyncStackTraces: true,
},
// database: {
// client: 'pg',
// name: 'joplin',
// user: 'joplin',
// host: 'localhost',
// port: 5432,
// password: 'joplin',
// asyncStackTraces: true,
// },
};
export default config;

View File

@ -1,20 +0,0 @@
import { Config } from './utils/types';
import configBase from './config-base';
const rootDir = '/home/joplin/';
const config: Config = {
...configBase,
rootDir: rootDir,
logDir: `${rootDir}/logs`,
database: {
client: 'pg',
name: 'joplin',
user: 'joplin',
host: 'db',
port: 5432,
password: 'joplin',
},
};
export default config;

View File

@ -1,13 +0,0 @@
import { Config } from './utils/types';
import configBase from './config-base';
const config: Config = {
...configBase,
database: {
name: 'DYNAMIC',
client: 'sqlite3',
asyncStackTraces: true,
},
};
export default config;

View File

@ -1,28 +1,93 @@
import { rtrimSlashes } from '@joplin/lib/path-utils'; import { rtrimSlashes } from '@joplin/lib/path-utils';
import { Config } from './utils/types'; import { Config, DatabaseConfig, DatabaseConfigClient } from './utils/types';
import * as pathUtils from 'path';
let baseConfig_: Config = null; export interface EnvVariables {
let baseUrl_: string = null; APP_BASE_URL?: string;
APP_PORT?: string;
DB_CLIENT?: string;
RUNNING_IN_DOCKER?: string;
export function initConfig(baseConfig: Config) { POSTGRES_PASSWORD?: string;
baseConfig_ = baseConfig; POSTGRES_DATABASE?: string;
POSTGRES_USER?: string;
POSTGRES_HOST?: string;
POSTGRES_PORT?: string;
SQLITE_DATABASE?: string;
}
let runningInDocker_: boolean = false;
export function runningInDocker(): boolean {
return runningInDocker_;
}
function databaseHostFromEnv(runningInDocker: boolean, env: EnvVariables): string {
if (env.POSTGRES_HOST) {
// When running within Docker, the app localhost is different from the
// host's localhost. To access the latter, Docker defines a special host
// called "host.docker.internal", so here we swap the values if necessary.
if (runningInDocker && ['localhost', '127.0.0.1'].includes(env.POSTGRES_HOST)) {
return 'host.docker.internal';
} else {
return env.POSTGRES_HOST;
}
}
return null;
}
function databaseConfigFromEnv(runningInDocker: boolean, env: EnvVariables): DatabaseConfig {
if (env.DB_CLIENT === 'pg') {
return {
client: DatabaseConfigClient.PostgreSQL,
name: env.POSTGRES_DATABASE || 'joplin',
user: env.POSTGRES_USER || 'joplin',
password: env.POSTGRES_PASSWORD || 'joplin',
port: env.POSTGRES_PORT ? Number(env.POSTGRES_PORT) : 5432,
host: databaseHostFromEnv(runningInDocker, env) || 'localhost',
};
}
return {
client: DatabaseConfigClient.SQLite,
name: env.SQLITE_DATABASE || 'prod',
asyncStackTraces: true,
};
}
function baseUrlFromEnv(env: any, appPort: number): string {
if (env.APP_BASE_URL) {
return rtrimSlashes(env.APP_BASE_URL);
} else {
return `http://localhost:${appPort}`;
}
}
let config_: Config = null;
export function initConfig(env: EnvVariables) {
runningInDocker_ = !!env.RUNNING_IN_DOCKER;
const rootDir = pathUtils.dirname(__dirname);
const viewDir = `${pathUtils.dirname(__dirname)}/src/views`;
const appPort = env.APP_PORT ? Number(env.APP_PORT) : 22300;
config_ = {
rootDir: rootDir,
viewDir: viewDir,
layoutDir: `${viewDir}/layouts`,
logDir: `${rootDir}/logs`,
database: databaseConfigFromEnv(runningInDocker_, env),
port: appPort,
baseUrl: baseUrlFromEnv(env, appPort),
};
} }
function config(): Config { function config(): Config {
if (!baseConfig_) throw new Error('Config has not been initialized!'); if (!config_) throw new Error('Config has not been initialized!');
return baseConfig_; return config_;
}
export function baseUrl() {
if (baseUrl_) return baseUrl_;
if (process.env.JOPLIN_BASE_URL) {
baseUrl_ = rtrimSlashes(process.env.JOPLIN_BASE_URL);
} else {
baseUrl_ = `http://localhost:${config().port}`;
}
return baseUrl_;
} }
export default config; export default config;

View File

@ -47,15 +47,15 @@ export interface ConnectionCheckResult {
connection: DbConnection; connection: DbConnection;
} }
export function sqliteFilePath(dbConfig: DatabaseConfig): string { export function sqliteFilePath(name: string): string {
return `${sqliteDbDir}/db-${dbConfig.name}.sqlite`; return `${sqliteDbDir}/db-${name}.sqlite`;
} }
export function makeKnexConfig(dbConfig: DatabaseConfig): KnexDatabaseConfig { export function makeKnexConfig(dbConfig: DatabaseConfig): KnexDatabaseConfig {
const connection: DbConfigConnection = {}; const connection: DbConfigConnection = {};
if (dbConfig.client === 'sqlite3') { if (dbConfig.client === 'sqlite3') {
connection.filename = sqliteFilePath(dbConfig); connection.filename = sqliteFilePath(dbConfig.name);
} else { } else {
connection.database = dbConfig.name; connection.database = dbConfig.name;
connection.host = dbConfig.host; connection.host = dbConfig.host;

View File

@ -4,47 +4,80 @@ import { defaultAdminEmail, defaultAdminPassword, NotificationLevel } from '../d
import { _ } from '@joplin/lib/locale'; import { _ } from '@joplin/lib/locale';
import Logger from '@joplin/lib/Logger'; import Logger from '@joplin/lib/Logger';
import * as MarkdownIt from 'markdown-it'; import * as MarkdownIt from 'markdown-it';
import config from '../config';
const logger = Logger.create('notificationHandler'); const logger = Logger.create('notificationHandler');
async function handleChangeAdminPasswordNotification(ctx: AppContext) {
if (!ctx.owner.is_admin) return;
const defaultAdmin = await ctx.models.user().login(defaultAdminEmail, defaultAdminPassword);
const notificationModel = ctx.models.notification({ userId: ctx.owner.id });
if (defaultAdmin) {
await notificationModel.add(
'change_admin_password',
NotificationLevel.Important,
_('The default admin password is insecure and has not been changed! [Change it now](%s)', await ctx.models.user().profileUrl())
);
} else {
await notificationModel.markAsRead('change_admin_password');
}
if (config().database.client === 'sqlite3' && ctx.env === 'prod') {
await notificationModel.add(
'using_sqlite_in_prod',
NotificationLevel.Important,
'The server is currently using SQLite3 as a database. It is not recommended in production as it is slow and can cause locking issues. Please see the README for information on how to change it.'
);
}
}
async function handleSqliteInProdNotification(ctx: AppContext) {
if (!ctx.owner.is_admin) return;
const notificationModel = ctx.models.notification({ userId: ctx.owner.id });
if (config().database.client === 'sqlite3' && ctx.env === 'prod') {
await notificationModel.add(
'using_sqlite_in_prod',
NotificationLevel.Important,
'The server is currently using SQLite3 as a database. It is not recommended in production as it is slow and can cause locking issues. Please see the README for information on how to change it.'
);
}
}
async function makeNotificationViews(ctx: AppContext): Promise<NotificationView[]> {
const markdownIt = new MarkdownIt();
const notificationModel = ctx.models.notification({ userId: ctx.owner.id });
const notifications = await notificationModel.allUnreadByUserId(ctx.owner.id);
const views: NotificationView[] = [];
for (const n of notifications) {
views.push({
id: n.id,
messageHtml: markdownIt.render(n.message),
level: n.level === NotificationLevel.Important ? 'warning' : 'info',
closeUrl: notificationModel.closeUrl(n.id),
});
}
return views;
}
// The role of this middleware is to inspect the system and to generate
// notifications for any issue it finds. It is only active for logged in users
// on the website. It is inactive for API calls.
export default async function(ctx: AppContext, next: KoaNext): Promise<void> { export default async function(ctx: AppContext, next: KoaNext): Promise<void> {
ctx.notifications = []; ctx.notifications = [];
try { try {
if (isApiRequest(ctx)) return next(); if (isApiRequest(ctx)) return next();
if (!ctx.owner) return next();
const user = ctx.owner; await handleChangeAdminPasswordNotification(ctx);
if (!user) return next(); await handleSqliteInProdNotification(ctx);
ctx.notifications = await makeNotificationViews(ctx);
const notificationModel = ctx.models.notification({ userId: user.id });
if (user.is_admin) {
const defaultAdmin = await ctx.models.user().login(defaultAdminEmail, defaultAdminPassword);
if (defaultAdmin) {
await notificationModel.add(
'change_admin_password',
NotificationLevel.Important,
_('The default admin password is insecure and has not been changed! [Change it now](%s)', await ctx.models.user().profileUrl())
);
} else {
await notificationModel.markAsRead('change_admin_password');
}
}
const markdownIt = new MarkdownIt();
const notifications = await notificationModel.allUnreadByUserId(user.id);
const views: NotificationView[] = [];
for (const n of notifications) {
views.push({
id: n.id,
messageHtml: markdownIt.render(n.message),
level: n.level === NotificationLevel.Important ? 'warning' : 'info',
closeUrl: notificationModel.closeUrl(n.id),
});
}
ctx.notifications = views;
} catch (error) { } catch (error) {
logger.error(error); logger.error(error);
} }

View File

@ -26,6 +26,7 @@ export default async function(ctx: AppContext) {
ctx.response.status = 200; ctx.response.status = 200;
ctx.response.body = await mustacheService.renderView(responseObject, { ctx.response.body = await mustacheService.renderView(responseObject, {
notifications: ctx.notifications || [], notifications: ctx.notifications || [],
hasNotifications: !!ctx.notifications && !!ctx.notifications.length,
owner: ctx.owner, owner: ctx.owner,
}); });
} else { } else {

View File

@ -39,6 +39,10 @@ async function findLocalFile(path: string): Promise<string> {
const router = new Router(); const router = new Router();
router.public = true;
// Used to serve static files, so it needs to be public because for example the
// login page, which is public, needs access to the CSS files.
router.get('', async (path: SubPath, ctx: Koa.Context) => { router.get('', async (path: SubPath, ctx: Koa.Context) => {
const localPath = await findLocalFile(path.raw); const localPath = await findLocalFile(path.raw);

View File

@ -6,7 +6,7 @@ import { ErrorNotFound } from '../../utils/errors';
import { File } from '../../db'; import { File } from '../../db';
import { createPaginationLinks, pageMaxSize, Pagination, PaginationOrder, PaginationOrderDir, requestPaginationOrder, validatePagination } from '../../models/utils/pagination'; import { createPaginationLinks, pageMaxSize, Pagination, PaginationOrder, PaginationOrderDir, requestPaginationOrder, validatePagination } from '../../models/utils/pagination';
import { setQueryParameters } from '../../utils/urlUtils'; import { setQueryParameters } from '../../utils/urlUtils';
import { baseUrl } from '../../config'; import config from '../../config';
import { formatDateTime } from '../../utils/time'; import { formatDateTime } from '../../utils/time';
import defaultView from '../../utils/defaultView'; import defaultView from '../../utils/defaultView';
import { View } from '../../services/MustacheService'; import { View } from '../../services/MustacheService';
@ -51,7 +51,7 @@ router.get('files/:id', async (path: SubPath, ctx: AppContext) => {
async function fileToViewItem(file: File, fileFullPaths: Record<string, string>): Promise<any> { async function fileToViewItem(file: File, fileFullPaths: Record<string, string>): Promise<any> {
const filePath = fileFullPaths[file.id]; const filePath = fileFullPaths[file.id];
let url = `${baseUrl()}/files/${filePath}`; let url = `${config().baseUrl}/files/${filePath}`;
if (!file.is_directory) { if (!file.is_directory) {
url += '/content'; url += '/content';
} else { } else {
@ -88,7 +88,7 @@ router.get('files/:id', async (path: SubPath, ctx: AppContext) => {
const view: View = defaultView('files'); const view: View = defaultView('files');
view.content.paginatedFiles = { ...paginatedFiles, items: files }; view.content.paginatedFiles = { ...paginatedFiles, items: files };
view.content.paginationLinks = paginationLinks; view.content.paginationLinks = paginationLinks;
view.content.postUrl = `${baseUrl()}/files`; view.content.postUrl = `${config().baseUrl}/files`;
view.content.parentId = parent.id; view.content.parentId = parent.id;
view.cssFiles = ['index/files']; view.cssFiles = ['index/files'];
view.partials.push('pagination'); view.partials.push('pagination');

View File

@ -2,7 +2,7 @@ import { SubPath, redirect } from '../../utils/routeUtils';
import Router from '../../utils/Router'; import Router from '../../utils/Router';
import { AppContext } from '../../utils/types'; import { AppContext } from '../../utils/types';
import { formParse } from '../../utils/requestUtils'; import { formParse } from '../../utils/requestUtils';
import { baseUrl } from '../../config'; import config from '../../config';
import defaultView from '../../utils/defaultView'; import defaultView from '../../utils/defaultView';
import { View } from '../../services/MustacheService'; import { View } from '../../services/MustacheService';
@ -27,7 +27,7 @@ router.post('login', async (_path: SubPath, ctx: AppContext) => {
const session = await ctx.models.session().authenticate(body.fields.email, body.fields.password); const session = await ctx.models.session().authenticate(body.fields.email, body.fields.password);
ctx.cookies.set('sessionId', session.id); ctx.cookies.set('sessionId', session.id);
return redirect(ctx, `${baseUrl()}/home`); return redirect(ctx, `${config().baseUrl}/home`);
} catch (error) { } catch (error) {
return makeView(error); return makeView(error);
} }

View File

@ -1,7 +1,7 @@
import { SubPath, redirect } from '../../utils/routeUtils'; import { SubPath, redirect } from '../../utils/routeUtils';
import Router from '../../utils/Router'; import Router from '../../utils/Router';
import { AppContext } from '../../utils/types'; import { AppContext } from '../../utils/types';
import { baseUrl } from '../../config'; import config from '../../config';
import { contextSessionId } from '../../utils/requestUtils'; import { contextSessionId } from '../../utils/requestUtils';
const router = new Router(); const router = new Router();
@ -10,7 +10,7 @@ router.post('logout', async (_path: SubPath, ctx: AppContext) => {
const sessionId = contextSessionId(ctx, false); const sessionId = contextSessionId(ctx, false);
ctx.cookies.set('sessionId', ''); ctx.cookies.set('sessionId', '');
await ctx.models.session().logout(sessionId); await ctx.models.session().logout(sessionId);
return redirect(ctx, `${baseUrl()}/login`); return redirect(ctx, `${config().baseUrl}/login`);
}); });
export default router; export default router;

View File

@ -4,7 +4,7 @@ import { AppContext, HttpMethod } from '../../utils/types';
import { formParse } from '../../utils/requestUtils'; import { formParse } from '../../utils/requestUtils';
import { ErrorUnprocessableEntity } from '../../utils/errors'; import { ErrorUnprocessableEntity } from '../../utils/errors';
import { User } from '../../db'; import { User } from '../../db';
import { baseUrl } from '../../config'; import config from '../../config';
import { View } from '../../services/MustacheService'; import { View } from '../../services/MustacheService';
import defaultView from '../../utils/defaultView'; import defaultView from '../../utils/defaultView';
@ -55,11 +55,11 @@ router.get('users/:id', async (path: SubPath, ctx: AppContext, user: User = null
let postUrl = ''; let postUrl = '';
if (isNew) { if (isNew) {
postUrl = `${baseUrl()}/users/new`; postUrl = `${config().baseUrl}/users/new`;
} else if (isMe) { } else if (isMe) {
postUrl = `${baseUrl()}/users/me`; postUrl = `${config().baseUrl}/users/me`;
} else { } else {
postUrl = `${baseUrl()}/users/${user.id}`; postUrl = `${config().baseUrl}/users/${user.id}`;
} }
const view: View = defaultView('user'); const view: View = defaultView('user');
@ -100,7 +100,7 @@ router.post('users', async (path: SubPath, ctx: AppContext) => {
throw new Error('Invalid form button'); throw new Error('Invalid form button');
} }
return redirect(ctx, `${baseUrl()}/users${userIsMe(path) ? '/me' : ''}`); return redirect(ctx, `${config().baseUrl}/users${userIsMe(path) ? '/me' : ''}`);
} catch (error) { } catch (error) {
const endPoint = router.findEndPoint(HttpMethod.GET, 'users/:id'); const endPoint = router.findEndPoint(HttpMethod.GET, 'users/:id');
return endPoint(path, ctx, user, error); return endPoint(path, ctx, user, error);

View File

@ -1,6 +1,6 @@
import * as Mustache from 'mustache'; import * as Mustache from 'mustache';
import * as fs from 'fs-extra'; import * as fs from 'fs-extra';
import config, { baseUrl } from '../config'; import config from '../config';
export interface RenderOptions { export interface RenderOptions {
partials?: any; partials?: any;
@ -30,7 +30,7 @@ class MustacheService {
private get defaultLayoutOptions(): any { private get defaultLayoutOptions(): any {
return { return {
baseUrl: baseUrl(), baseUrl: config().baseUrl,
}; };
} }
@ -41,7 +41,7 @@ class MustacheService {
private resolvesFilePaths(type: string, paths: string[]): string[] { private resolvesFilePaths(type: string, paths: string[]): string[] {
const output: string[] = []; const output: string[] = [];
for (const path of paths) { for (const path of paths) {
output.push(`${baseUrl()}/${type}/${path}.${type}`); output.push(`${config().baseUrl}/${type}/${path}.${type}`);
} }
return output; return output;
} }

View File

@ -33,7 +33,7 @@ export async function createDb(config: DatabaseConfig, options: CreateDbOptions
await execCommand(cmd.join(' ')); await execCommand(cmd.join(' '));
} else if (config.client === 'sqlite3') { } else if (config.client === 'sqlite3') {
const filePath = sqliteFilePath(config); const filePath = sqliteFilePath(config.name);
if (await fs.pathExists(filePath)) { if (await fs.pathExists(filePath)) {
if (options.dropIfExists) { if (options.dropIfExists) {
@ -71,6 +71,6 @@ export async function dropDb(config: DatabaseConfig, options: DropDbOptions = nu
throw error; throw error;
} }
} else if (config.client === 'sqlite3') { } else if (config.client === 'sqlite3') {
await fs.remove(sqliteFilePath(config)); await fs.remove(sqliteFilePath(config.name));
} }
} }

View File

@ -1,9 +1,8 @@
import { User, Session, DbConnection, connectDb, disconnectDb, File, truncateTables } from '../../db'; import { User, Session, DbConnection, connectDb, disconnectDb, File, truncateTables, sqliteFilePath } from '../../db';
import { createDb } from '../../tools/dbTools'; import { createDb } from '../../tools/dbTools';
import modelFactory from '../../models/factory'; import modelFactory from '../../models/factory';
import baseConfig from '../../config-tests'; import { AppContext, Env } from '../types';
import { AppContext, Config, Env } from '../types'; import config, { initConfig } from '../../config';
import { initConfig } from '../../config';
import FileModel from '../../models/FileModel'; import FileModel from '../../models/FileModel';
import Logger from '@joplin/lib/Logger'; import Logger from '@joplin/lib/Logger';
import FakeCookies from './koa/FakeCookies'; import FakeCookies from './koa/FakeCookies';
@ -34,18 +33,16 @@ export async function tempDir(): Promise<string> {
return tempDir_; return tempDir_;
} }
let createdDbName_: string = null;
export async function beforeAllDb(unitName: string) { export async function beforeAllDb(unitName: string) {
const config: Config = { createdDbName_ = unitName;
...baseConfig,
database: {
...baseConfig.database,
name: unitName,
},
};
initConfig(config); initConfig({
await createDb(config.database, { dropIfExists: true }); SQLITE_DATABASE: createdDbName_,
db_ = await connectDb(config.database); });
await createDb(config().database, { dropIfExists: true });
db_ = await connectDb(config().database);
} }
export async function afterAllTests() { export async function afterAllTests() {
@ -58,6 +55,12 @@ export async function afterAllTests() {
await fs.remove(tempDir_); await fs.remove(tempDir_);
tempDir_ = null; tempDir_ = null;
} }
if (createdDbName_) {
const filePath = sqliteFilePath(createdDbName_);
await fs.remove(filePath);
createdDbName_ = null;
}
} }
export async function beforeEachDb() { export async function beforeEachDb() {

View File

@ -25,8 +25,13 @@ export interface AppContext extends Koa.Context {
owner: User; owner: User;
} }
export enum DatabaseConfigClient {
PostgreSQL = 'pg',
SQLite = 'sqlite3',
}
export interface DatabaseConfig { export interface DatabaseConfig {
client: string; client: DatabaseConfigClient;
name: string; name: string;
host?: string; host?: string;
port?: number; port?: number;
@ -40,8 +45,11 @@ export interface Config {
rootDir: string; rootDir: string;
viewDir: string; viewDir: string;
layoutDir: string; layoutDir: string;
// Not that, for now, nothing is being logged to file. Log is just printed
// to stdout, which is then handled by Docker own log mechanism
logDir: string; logDir: string;
database: DatabaseConfig; database: DatabaseConfig;
baseUrl: string;
} }
export enum HttpMethod { export enum HttpMethod {

View File

@ -13,8 +13,8 @@
</head> </head>
<body class="page-{{{pageName}}}"> <body class="page-{{{pageName}}}">
{{> navbar}} {{> navbar}}
{{> notifications}}
<main class="main"> <main class="main">
{{> notifications}}
{{{contentHtml}}} {{{contentHtml}}}
</main> </main>
</body> </body>

View File

@ -1,9 +1,11 @@
{{#global.notifications}} {{#global.hasNotifications}}
<div class="notification is-{{level}}" id="notification-{{id}}"> {{#global.notifications}}
<button data-close-url="{{closeUrl}}" data-id="{{id}}" class="delete close-notification-button"></button> <div class="notification is-{{level}}" id="notification-{{id}}">
{{{messageHtml}}} <button data-close-url="{{closeUrl}}" data-id="{{id}}" class="delete close-notification-button"></button>
</div> {{{messageHtml}}}
{{/global.notifications}} </div>
{{/global.notifications}}
{{/global.hasNotifications}}
<script> <script>
onDocumentReady(function() { onDocumentReady(function() {

View File

@ -194,8 +194,6 @@ async function main() {
let manifests: any = {}; let manifests: any = {};
// TODO: validate plugin ID when publishing
for (const npmPackage of npmPackages) { for (const npmPackage of npmPackages) {
try { try {
const packageName = npmPackage.name; const packageName = npmPackage.name;

View File

@ -1,35 +1,28 @@
import * as fs from 'fs-extra'; const { execCommand2, rootDir, gitPullTry } = require('./tool-utils.js');
const { execCommand, execCommandVerbose, rootDir, gitPullTry } = require('./tool-utils.js');
const serverDir = `${rootDir}/packages/server`; const serverDir = `${rootDir}/packages/server`;
const readmePath = `${serverDir}/README.md`;
async function updateReadmeLinkVersion(version: string) {
const content = await fs.readFile(readmePath, 'utf8');
const newContent = content.replace(/server-v(.*?).tar.gz/g, `server-${version}.tar.gz`);
if (content === newContent) throw new Error(`Could not change version number in ${readmePath}`);
await fs.writeFile(readmePath, newContent, 'utf8');
}
async function main() { async function main() {
process.chdir(serverDir);
console.info(`Running from: ${process.cwd()}`);
await gitPullTry(); await gitPullTry();
const version = (await execCommand('npm version patch')).trim(); process.chdir(serverDir);
const version = (await execCommand2('npm version patch')).trim();
const versionShort = version.substr(1);
const tagName = `server-${version}`; const tagName = `server-${version}`;
console.info(`New version number: ${version}`); process.chdir(rootDir);
console.info(`Running from: ${process.cwd()}`);
await updateReadmeLinkVersion(version); await execCommand2(`docker build -t "joplin/server:${versionShort}" -f Dockerfile.server .`);
await execCommand2(`docker tag "joplin/server:${versionShort}" "joplin/server:latest"`);
await execCommand2(`docker push joplin/server:${versionShort}`);
await execCommand2('docker push joplin/server:latest');
await execCommandVerbose('git', ['add', '-A']); await execCommand2('git add -A');
await execCommandVerbose('git', ['commit', '-m', `Server release ${version}`]); await execCommand2(`git commit -m 'Server release ${version}'`);
await execCommandVerbose('git', ['tag', tagName]); await execCommand2(`git tag ${tagName}`);
await execCommandVerbose('git', ['push']); await execCommand2('git push');
await execCommandVerbose('git', ['push', '--tags']); await execCommand2('git push --tags');
} }
main().catch((error) => { main().catch((error) => {

View File

@ -2,6 +2,7 @@ const fetch = require('node-fetch');
const fs = require('fs-extra'); const fs = require('fs-extra');
const execa = require('execa'); const execa = require('execa');
const { execSync } = require('child_process'); const { execSync } = require('child_process');
const { splitCommandString } = require('@joplin/lib/string-utils');
const toolUtils = {}; const toolUtils = {};
@ -55,6 +56,29 @@ toolUtils.execCommandVerbose = function(commandName, args = []) {
return promise; return promise;
}; };
// There's lot of execCommandXXX functions, but eventually all scripts should
// use the one below, which supports:
//
// - Printing the command being executed
// - Printing the output in real time (piping to stdout)
// - Returning the command result as string
toolUtils.execCommand2 = async function(command, options = null) {
options = {
showInput: true,
showOutput: true,
...options,
};
if (options.showInput) console.info(`> ${command}`);
const args = splitCommandString(command);
const executableName = args[0];
args.splice(0, 1);
const promise = execa(executableName, args);
if (options.showOutput) promise.stdout.pipe(process.stdout);
const result = await promise;
return result.stdout;
};
toolUtils.execCommandWithPipes = function(executable, args) { toolUtils.execCommandWithPipes = function(executable, args) {
const spawn = require('child_process').spawn; const spawn = require('child_process').spawn;