mirror of
https://github.com/LukeHagar/vercel.git
synced 2025-12-23 09:59:12 +00:00
Compare commits
78 Commits
@now/php@0
...
@now/bash@
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
c1049985af | ||
|
|
214388ccf3 | ||
|
|
b1d6b7bfc0 | ||
|
|
ece3564dfd | ||
|
|
a88af1f077 | ||
|
|
d92f7b26c0 | ||
|
|
52198af750 | ||
|
|
d58bff2453 | ||
|
|
8c0a144ae4 | ||
|
|
106e4d5f36 | ||
|
|
66c28bd695 | ||
|
|
55e75296ff | ||
|
|
36cbb36737 | ||
|
|
978ca328ef | ||
|
|
7b383e0f7c | ||
|
|
faa5ab36aa | ||
|
|
c0a21969dd | ||
|
|
73d0a1723f | ||
|
|
7c515544ae | ||
|
|
b53c9a6299 | ||
|
|
35ff11e6e4 | ||
|
|
64ee4905cd | ||
|
|
e50dd7e50a | ||
|
|
6101ba9d95 | ||
|
|
8dc0c92c58 | ||
|
|
44c9f3765a | ||
|
|
92c05ca338 | ||
|
|
069b557906 | ||
|
|
692a0df909 | ||
|
|
aeafeb5441 | ||
|
|
a09d5fb355 | ||
|
|
d8017aa9aa | ||
|
|
702f56b9b5 | ||
|
|
183b117152 | ||
|
|
75b3fb4981 | ||
|
|
49e63de5fe | ||
|
|
4742cd32f2 | ||
|
|
377b73105d | ||
|
|
a5577efb3d | ||
|
|
2ec46dc5c9 | ||
|
|
42708ed93c | ||
|
|
2fabe95f6e | ||
|
|
ac1a3dab22 | ||
|
|
ad4011512d | ||
|
|
9ff1a25c8f | ||
|
|
8039b3d377 | ||
|
|
dd9017475c | ||
|
|
031499014f | ||
|
|
2a68d2a2ad | ||
|
|
31299fae6e | ||
|
|
4bac0db379 | ||
|
|
95e7d459d3 | ||
|
|
dd120b8d20 | ||
|
|
b6975676e5 | ||
|
|
a7951dae81 | ||
|
|
b0c918f7fb | ||
|
|
df54dc7dc9 | ||
|
|
0dd801ff6c | ||
|
|
398743ef95 | ||
|
|
337c74b81b | ||
|
|
680bb82ec3 | ||
|
|
17ed5411e3 | ||
|
|
d9bbcb6939 | ||
|
|
800e4de76f | ||
|
|
864dd468d9 | ||
|
|
ba833871bb | ||
|
|
e732bac78e | ||
|
|
28ea4015b4 | ||
|
|
a93d97cabd | ||
|
|
67f39f7c9b | ||
|
|
acd793b9e9 | ||
|
|
f74d61279d | ||
|
|
fcb8eacec0 | ||
|
|
c8fca2ba72 | ||
|
|
4feffa13eb | ||
|
|
3e330b25f4 | ||
|
|
9b2cae33af | ||
|
|
4b6371530c |
12
.circleci/build.sh
Executable file
12
.circleci/build.sh
Executable file
@@ -0,0 +1,12 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -euo pipefail
|
||||||
|
|
||||||
|
circleci_dir="$( cd "$( dirname "${BASH_SOURCE[0]}" )" >/dev/null 2>&1 && pwd )"
|
||||||
|
needs_build="$(grep -rn '"build"' packages/*/package.json | cut -d: -f1)"
|
||||||
|
|
||||||
|
for pkg in $needs_build; do
|
||||||
|
dir="$(dirname "$pkg")"
|
||||||
|
cd "$circleci_dir/../$dir"
|
||||||
|
echo "Building \`$dir\`"
|
||||||
|
yarn build
|
||||||
|
done
|
||||||
@@ -23,6 +23,9 @@ jobs:
|
|||||||
- run:
|
- run:
|
||||||
name: Linting
|
name: Linting
|
||||||
command: yarn lint
|
command: yarn lint
|
||||||
|
- run:
|
||||||
|
name: Building
|
||||||
|
command: ./.circleci/build.sh
|
||||||
- run:
|
- run:
|
||||||
name: Tests
|
name: Tests
|
||||||
command: yarn test
|
command: yarn test
|
||||||
@@ -30,11 +33,8 @@ jobs:
|
|||||||
name: Potentially save npm token
|
name: Potentially save npm token
|
||||||
command: "([[ ! -z $NPM_TOKEN ]] && echo \"//registry.npmjs.org/:_authToken=$NPM_TOKEN\" >> ~/.npmrc) || echo \"Did not write npm token\""
|
command: "([[ ! -z $NPM_TOKEN ]] && echo \"//registry.npmjs.org/:_authToken=$NPM_TOKEN\" >> ~/.npmrc) || echo \"Did not write npm token\""
|
||||||
- run:
|
- run:
|
||||||
name: Potentially publish canary release
|
name: Potentially publish releases to npm
|
||||||
command: "if ls ~/.npmrc >/dev/null 2>&1 && [[ $(git describe --exact-match 2> /dev/null || :) =~ -canary ]]; then yarn run lerna publish from-git --npm-tag canary --yes; else echo \"Did not publish\"; fi"
|
command: ./.circleci/publish.sh
|
||||||
- run:
|
|
||||||
name: Potentially publish stable release
|
|
||||||
command: "if ls ~/.npmrc >/dev/null 2>&1 && [[ ! $(git describe --exact-match 2> /dev/null || :) =~ -canary ]]; then yarn run lerna publish from-git --yes; else echo \"Did not publish\"; fi"
|
|
||||||
workflows:
|
workflows:
|
||||||
version: 2
|
version: 2
|
||||||
build-and-deploy:
|
build-and-deploy:
|
||||||
|
|||||||
24
.circleci/publish.sh
Executable file
24
.circleci/publish.sh
Executable file
@@ -0,0 +1,24 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -euo pipefail
|
||||||
|
|
||||||
|
if [ ! -e ~/.npmrc ]; then
|
||||||
|
echo "~/.npmrc file does not exist, skipping publish"
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
npm_tag=""
|
||||||
|
tag="$(git describe --tags --exact-match 2> /dev/null || :)"
|
||||||
|
|
||||||
|
if [ -z "$tag" ]; then
|
||||||
|
echo "Not a tagged commit, skipping publish"
|
||||||
|
exit 0
|
||||||
|
fi
|
||||||
|
|
||||||
|
if [[ "$tag" =~ -canary ]]; then
|
||||||
|
echo "Publishing canary release"
|
||||||
|
npm_tag="--npm-tag canary"
|
||||||
|
else
|
||||||
|
echo "Publishing stable release"
|
||||||
|
fi
|
||||||
|
|
||||||
|
yarn run lerna publish from-git $npm_tag --yes
|
||||||
@@ -1,3 +1,4 @@
|
|||||||
/tmp/*
|
/tmp/*
|
||||||
/node_modules/*
|
/node_modules/*
|
||||||
/**/node_modules/*
|
/**/node_modules/*
|
||||||
|
/packages/now-go/go/*
|
||||||
|
|||||||
1
.gitignore
vendored
1
.gitignore
vendored
@@ -1,3 +1,4 @@
|
|||||||
node_modules
|
node_modules
|
||||||
tmp
|
tmp
|
||||||
target/
|
target/
|
||||||
|
.next
|
||||||
3
.prettierrc.json
Normal file
3
.prettierrc.json
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
{
|
||||||
|
"singleQuote": true
|
||||||
|
}
|
||||||
@@ -34,10 +34,10 @@ Serverless:
|
|||||||
|
|
||||||
In order to create the smallest possible lambdas Next.js has to be configured to build for the `serverless` target.
|
In order to create the smallest possible lambdas Next.js has to be configured to build for the `serverless` target.
|
||||||
|
|
||||||
1. Serverless Next.js requires Next.js 8 or later, currently this version is out on the `canary` release channel:
|
1. Serverless Next.js requires Next.js 8 or later, to upgrade you can install the `latest` version:
|
||||||
|
|
||||||
```
|
```
|
||||||
npm install next@canary
|
npm install next --save
|
||||||
```
|
```
|
||||||
|
|
||||||
2. Add the `now-build` script to your `package.json`
|
2. Add the `now-build` script to your `package.json`
|
||||||
|
|||||||
@@ -8,10 +8,10 @@ This error occurs when you have your application is not configured for Serverles
|
|||||||
|
|
||||||
In order to create the smallest possible lambdas Next.js has to be configured to build for the `serverless` target.
|
In order to create the smallest possible lambdas Next.js has to be configured to build for the `serverless` target.
|
||||||
|
|
||||||
1. Serverless Next.js requires Next.js 8 or later, currently this version is out on the `canary` release channel:
|
1. Serverless Next.js requires Next.js 8 or later, to upgrade you can install the `latest` version:
|
||||||
|
|
||||||
```
|
```
|
||||||
npm install next@canary
|
npm install next --save
|
||||||
```
|
```
|
||||||
|
|
||||||
2. Add the `now-build` script to your `package.json`
|
2. Add the `now-build` script to your `package.json`
|
||||||
|
|||||||
@@ -15,13 +15,13 @@
|
|||||||
"publish-stable": "lerna version",
|
"publish-stable": "lerna version",
|
||||||
"publish-canary": "lerna version prerelease --preid canary",
|
"publish-canary": "lerna version prerelease --preid canary",
|
||||||
"lint": "tsc && eslint .",
|
"lint": "tsc && eslint .",
|
||||||
"test": "jest --runInBand",
|
"test": "jest --runInBand --verbose",
|
||||||
"lint-staged": "lint-staged"
|
"lint-staged": "lint-staged"
|
||||||
},
|
},
|
||||||
"pre-commit": "lint-staged",
|
"pre-commit": "lint-staged",
|
||||||
"lint-staged": {
|
"lint-staged": {
|
||||||
"*.js": [
|
"*.js": [
|
||||||
"prettier --write --single-quote",
|
"prettier --write",
|
||||||
"eslint --fix",
|
"eslint --fix",
|
||||||
"git add"
|
"git add"
|
||||||
]
|
]
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/bash",
|
"name": "@now/bash",
|
||||||
"version": "0.1.2-canary.0",
|
"version": "0.1.3-canary.0",
|
||||||
"description": "Now 2.0 builder for HTTP endpoints written in Bash",
|
"description": "Now 2.0 builder for HTTP endpoints written in Bash",
|
||||||
"main": "index.js",
|
"main": "index.js",
|
||||||
"author": "Nathan Rajlich <nate@zeit.co>",
|
"author": "Nathan Rajlich <nate@zeit.co>",
|
||||||
|
|||||||
@@ -52,7 +52,7 @@ _lambda_runtime_next() {
|
|||||||
# Need to use a fifo here instead of bash <() because Lambda
|
# Need to use a fifo here instead of bash <() because Lambda
|
||||||
# errors with "/dev/fd/63 not found" for some reason :/
|
# errors with "/dev/fd/63 not found" for some reason :/
|
||||||
local stdin
|
local stdin
|
||||||
stdin="$(mktemp --dry-run)"
|
stdin="$(mktemp -u)"
|
||||||
mkfifo "$stdin"
|
mkfifo "$stdin"
|
||||||
_lambda_runtime_body "$event" > "$stdin" &
|
_lambda_runtime_body "$event" > "$stdin" &
|
||||||
|
|
||||||
@@ -76,7 +76,7 @@ _lambda_runtime_next() {
|
|||||||
_lambda_runtime_body() {
|
_lambda_runtime_body() {
|
||||||
if [ "$(jq --raw-output '.body | type' < "$1")" = "string" ]; then
|
if [ "$(jq --raw-output '.body | type' < "$1")" = "string" ]; then
|
||||||
if [ "$(jq --raw-output '.encoding' < "$1")" = "base64" ]; then
|
if [ "$(jq --raw-output '.encoding' < "$1")" = "base64" ]; then
|
||||||
jq --raw-output '.body' < "$1" | base64 -d
|
jq --raw-output '.body' < "$1" | base64 --decode
|
||||||
else
|
else
|
||||||
# assume plain-text body
|
# assume plain-text body
|
||||||
jq --raw-output '.body' < "$1"
|
jq --raw-output '.body' < "$1"
|
||||||
|
|||||||
@@ -1 +1,2 @@
|
|||||||
/test
|
/test
|
||||||
|
tmp
|
||||||
@@ -26,6 +26,18 @@ class FileFsRef {
|
|||||||
this.fsPath = fsPath;
|
this.fsPath = fsPath;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* Creates a `FileFsRef` with the correct `mode` from the file system.
|
||||||
|
*
|
||||||
|
* @argument {Object} options
|
||||||
|
* @argument {string} options.fsPath
|
||||||
|
* @returns {Promise<FileFsRef>}
|
||||||
|
*/
|
||||||
|
static async fromFsPath({ fsPath }) {
|
||||||
|
const { mode } = await fs.lstat(fsPath);
|
||||||
|
return new FileFsRef({ mode, fsPath });
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @argument {Object} options
|
* @argument {Object} options
|
||||||
* @argument {number} [options.mode=0o100644]
|
* @argument {number} [options.mode=0o100644]
|
||||||
|
|||||||
242
packages/now-build-utils/fs/bootstrap-yarn.js
vendored
242
packages/now-build-utils/fs/bootstrap-yarn.js
vendored
@@ -1,242 +0,0 @@
|
|||||||
/* eslint-disable arrow-body-style,no-multi-assign,no-param-reassign */
|
|
||||||
|
|
||||||
const MemoryFileSystem = require('memory-fs');
|
|
||||||
const fs = require('fs');
|
|
||||||
const path = require('path');
|
|
||||||
const { spawnSync } = require('child_process');
|
|
||||||
|
|
||||||
const yarnPath = spawnSync('which', ['yarn'])
|
|
||||||
.stdout.toString()
|
|
||||||
.trim();
|
|
||||||
|
|
||||||
const cachePath = spawnSync(yarnPath, ['cache', 'dir'])
|
|
||||||
.stdout.toString()
|
|
||||||
.trim();
|
|
||||||
|
|
||||||
spawnSync(yarnPath, ['cache', 'clean']);
|
|
||||||
const vfs = new MemoryFileSystem();
|
|
||||||
|
|
||||||
function isInsideCachePath(filename) {
|
|
||||||
const relative = path.relative(cachePath, filename);
|
|
||||||
return !relative.startsWith('..');
|
|
||||||
}
|
|
||||||
|
|
||||||
function replaceFn(name, newFnFactory) {
|
|
||||||
const prevFn = fs[name];
|
|
||||||
fs[name] = newFnFactory(prevFn);
|
|
||||||
}
|
|
||||||
|
|
||||||
replaceFn('createWriteStream', (prevFn) => {
|
|
||||||
return (...args) => {
|
|
||||||
const filename = args[0];
|
|
||||||
if (!isInsideCachePath(filename)) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
const stream = vfs.createWriteStream(...args);
|
|
||||||
|
|
||||||
stream.on('finish', () => {
|
|
||||||
setTimeout(() => {
|
|
||||||
stream.emit('close');
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
setTimeout(() => {
|
|
||||||
stream.emit('open');
|
|
||||||
});
|
|
||||||
|
|
||||||
return stream;
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
replaceFn('readFile', (prevFn) => {
|
|
||||||
return (...args) => {
|
|
||||||
const filename = args[0];
|
|
||||||
if (!isInsideCachePath(filename)) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
const callback = args[args.length - 1];
|
|
||||||
return vfs.readFile(...args.slice(0, -1), (error, result) => {
|
|
||||||
if (error) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
return callback(error, result);
|
|
||||||
});
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
replaceFn('readdir', (prevFn) => {
|
|
||||||
return (...args) => {
|
|
||||||
const dirname = args[0];
|
|
||||||
if (!isInsideCachePath(dirname)) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
const callback = args[args.length - 1];
|
|
||||||
return prevFn.call(fs, dirname, (error, results) => {
|
|
||||||
if (error) {
|
|
||||||
results = [];
|
|
||||||
}
|
|
||||||
|
|
||||||
return vfs.readdir(dirname, (error2, results2) => {
|
|
||||||
if (error2) {
|
|
||||||
return callback(error2);
|
|
||||||
}
|
|
||||||
|
|
||||||
// eslint-disable-next-line no-restricted-syntax
|
|
||||||
for (const result2 of results2) {
|
|
||||||
if (!results.includes(result2)) {
|
|
||||||
results.push(result2);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return callback(error2, results);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
replaceFn('stat', (prevFn) => {
|
|
||||||
return (...args) => {
|
|
||||||
const filename = args[0];
|
|
||||||
if (!isInsideCachePath(filename)) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
const callback = args[args.length - 1];
|
|
||||||
return vfs.stat(...args.slice(0, -1), (error, result) => {
|
|
||||||
if (error) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
result.atime = result.mtime = new Date();
|
|
||||||
return callback(error, result);
|
|
||||||
});
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
replaceFn('lstat', (prevFn) => {
|
|
||||||
return (...args) => {
|
|
||||||
const filename = args[0];
|
|
||||||
if (!isInsideCachePath(filename)) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
const callback = args[args.length - 1];
|
|
||||||
return vfs.stat(...args.slice(0, -1), (error, result) => {
|
|
||||||
if (error) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
result.atime = result.mtime = new Date();
|
|
||||||
return callback(error, result);
|
|
||||||
});
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
replaceFn('exists', (prevFn) => {
|
|
||||||
return (...args) => {
|
|
||||||
const filename = args[0];
|
|
||||||
if (!isInsideCachePath(filename)) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
const callback = args[args.length - 1];
|
|
||||||
return vfs.exists(...args.slice(0, -1), (result) => {
|
|
||||||
if (!result) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
return callback(result);
|
|
||||||
});
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
replaceFn('copyFile', (prevFn) => {
|
|
||||||
return (...args) => {
|
|
||||||
const src = args[0];
|
|
||||||
const dest = args[1];
|
|
||||||
const callback = args[args.length - 1];
|
|
||||||
|
|
||||||
if (isInsideCachePath(src) && !isInsideCachePath(dest)) {
|
|
||||||
const buffer = vfs.readFileSync(src);
|
|
||||||
return fs.writeFile(dest, buffer, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!isInsideCachePath(src) && isInsideCachePath(dest)) {
|
|
||||||
const buffer = fs.readFileSync(src);
|
|
||||||
return vfs.writeFile(dest, buffer, callback);
|
|
||||||
}
|
|
||||||
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
replaceFn('writeFile', (prevFn) => {
|
|
||||||
return (...args) => {
|
|
||||||
const filename = args[0];
|
|
||||||
if (!isInsideCachePath(filename)) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
return vfs.writeFile(...args);
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
replaceFn('mkdir', (prevFn) => {
|
|
||||||
return (...args) => {
|
|
||||||
const dirname = args[0];
|
|
||||||
if (!isInsideCachePath(dirname)) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
const callback = args[args.length - 1];
|
|
||||||
return prevFn.call(fs, dirname, (error) => {
|
|
||||||
if (error) {
|
|
||||||
return callback(error);
|
|
||||||
}
|
|
||||||
|
|
||||||
return vfs.mkdirp(dirname, callback);
|
|
||||||
});
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
replaceFn('utimes', (prevFn) => {
|
|
||||||
return (...args) => {
|
|
||||||
const filename = args[0];
|
|
||||||
if (!isInsideCachePath(filename)) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
const callback = args[args.length - 1];
|
|
||||||
return setTimeout(callback, 0);
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
replaceFn('chmod', (prevFn) => {
|
|
||||||
return (...args) => {
|
|
||||||
const filename = args[0];
|
|
||||||
if (!isInsideCachePath(filename)) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
const callback = args[args.length - 1];
|
|
||||||
return setTimeout(callback, 0);
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
replaceFn('chown', (prevFn) => {
|
|
||||||
return (...args) => {
|
|
||||||
const filename = args[0];
|
|
||||||
if (!isInsideCachePath(filename)) {
|
|
||||||
return prevFn.call(fs, ...args);
|
|
||||||
}
|
|
||||||
|
|
||||||
const callback = args[args.length - 1];
|
|
||||||
return setTimeout(callback, 0);
|
|
||||||
};
|
|
||||||
});
|
|
||||||
|
|
||||||
require(yarnPath);
|
|
||||||
@@ -1,12 +1,10 @@
|
|||||||
const path = require('path');
|
const { join } = require('path');
|
||||||
const fs = require('fs-extra');
|
const { tmpdir } = require('os');
|
||||||
|
const { mkdirp } = require('fs-extra');
|
||||||
const prod = process.env.AWS_EXECUTION_ENV || process.env.X_GOOGLE_CODE_LOCATION;
|
|
||||||
const TMP_PATH = prod ? '/tmp' : path.join(__dirname, 'tmp');
|
|
||||||
|
|
||||||
module.exports = async function getWritableDirectory() {
|
module.exports = async function getWritableDirectory() {
|
||||||
const name = Math.floor(Math.random() * 0x7fffffff).toString(16);
|
const name = Math.floor(Math.random() * 0x7fffffff).toString(16);
|
||||||
const directory = path.join(TMP_PATH, name);
|
const directory = join(tmpdir(), name);
|
||||||
await fs.mkdirp(directory);
|
await mkdirp(directory);
|
||||||
return directory;
|
return directory;
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -3,9 +3,6 @@ const fs = require('fs-extra');
|
|||||||
const path = require('path');
|
const path = require('path');
|
||||||
const { spawn } = require('child_process');
|
const { spawn } = require('child_process');
|
||||||
|
|
||||||
const prod = process.env.AWS_EXECUTION_ENV
|
|
||||||
|| process.env.X_GOOGLE_CODE_LOCATION;
|
|
||||||
|
|
||||||
function spawnAsync(command, args, cwd) {
|
function spawnAsync(command, args, cwd) {
|
||||||
return new Promise((resolve, reject) => {
|
return new Promise((resolve, reject) => {
|
||||||
const child = spawn(command, args, { stdio: 'inherit', cwd });
|
const child = spawn(command, args, { stdio: 'inherit', cwd });
|
||||||
@@ -16,9 +13,18 @@ function spawnAsync(command, args, cwd) {
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async function chmodPlusX(fsPath) {
|
||||||
|
const s = await fs.stat(fsPath);
|
||||||
|
const newMode = s.mode | 64 | 8 | 1; // eslint-disable-line no-bitwise
|
||||||
|
if (s.mode === newMode) return;
|
||||||
|
const base8 = newMode.toString(8).slice(-3);
|
||||||
|
await fs.chmod(fsPath, base8);
|
||||||
|
}
|
||||||
|
|
||||||
async function runShellScript(fsPath) {
|
async function runShellScript(fsPath) {
|
||||||
assert(path.isAbsolute(fsPath));
|
assert(path.isAbsolute(fsPath));
|
||||||
const destPath = path.dirname(fsPath);
|
const destPath = path.dirname(fsPath);
|
||||||
|
await chmodPlusX(fsPath);
|
||||||
await spawnAsync(`./${path.basename(fsPath)}`, [], destPath);
|
await spawnAsync(`./${path.basename(fsPath)}`, [], destPath);
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
@@ -66,15 +72,6 @@ async function installDependencies(destPath, args = []) {
|
|||||||
commandArgs = args.filter(a => a !== '--prefer-offline');
|
commandArgs = args.filter(a => a !== '--prefer-offline');
|
||||||
await spawnAsync('npm', ['install'].concat(commandArgs), destPath);
|
await spawnAsync('npm', ['install'].concat(commandArgs), destPath);
|
||||||
await spawnAsync('npm', ['cache', 'clean', '--force'], destPath);
|
await spawnAsync('npm', ['cache', 'clean', '--force'], destPath);
|
||||||
} else if (prod) {
|
|
||||||
console.log('using memory-fs for yarn cache');
|
|
||||||
await spawnAsync(
|
|
||||||
'node',
|
|
||||||
[path.join(__dirname, 'bootstrap-yarn.js'), '--cwd', destPath].concat(
|
|
||||||
commandArgs,
|
|
||||||
),
|
|
||||||
destPath,
|
|
||||||
);
|
|
||||||
} else {
|
} else {
|
||||||
await spawnAsync('yarn', ['--cwd', destPath].concat(commandArgs), destPath);
|
await spawnAsync('yarn', ['--cwd', destPath].concat(commandArgs), destPath);
|
||||||
await spawnAsync('yarn', ['cache', 'clean'], destPath);
|
await spawnAsync('yarn', ['cache', 'clean'], destPath);
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/build-utils",
|
"name": "@now/build-utils",
|
||||||
"version": "0.4.33-canary.2",
|
"version": "0.4.37-canary.0",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ const {
|
|||||||
testDeployment,
|
testDeployment,
|
||||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||||
|
|
||||||
jest.setTimeout(2 * 60 * 1000);
|
jest.setTimeout(4 * 60 * 1000);
|
||||||
const builderUrl = '@canary';
|
const builderUrl = '@canary';
|
||||||
let buildUtilsUrl;
|
let buildUtilsUrl;
|
||||||
|
|
||||||
|
|||||||
5
packages/now-go/.gitignore
vendored
5
packages/now-go/.gitignore
vendored
@@ -1,4 +1,5 @@
|
|||||||
node_modules
|
node_modules
|
||||||
*.log
|
*.log
|
||||||
launcher
|
/?.js
|
||||||
bin
|
/go
|
||||||
|
/get-exported-function-name
|
||||||
|
|||||||
@@ -1,7 +0,0 @@
|
|||||||
#!/usr/bin/env bash
|
|
||||||
|
|
||||||
mkdir -p bin
|
|
||||||
cd util
|
|
||||||
GOOS=linux GOARCH=amd64 go build get-exported-function-name.go
|
|
||||||
mv get-exported-function-name ../bin/
|
|
||||||
|
|
||||||
@@ -1,23 +0,0 @@
|
|||||||
const path = require('path');
|
|
||||||
|
|
||||||
const fetch = require('node-fetch');
|
|
||||||
const tar = require('tar');
|
|
||||||
const getWritableDirectory = require('@now/build-utils/fs/get-writable-directory.js'); // eslint-disable-line import/no-extraneous-dependencies
|
|
||||||
|
|
||||||
const url = 'https://dl.google.com/go/go1.11.1.linux-amd64.tar.gz';
|
|
||||||
|
|
||||||
module.exports = async () => {
|
|
||||||
const res = await fetch(url);
|
|
||||||
const dir = await getWritableDirectory();
|
|
||||||
|
|
||||||
if (!res.ok) {
|
|
||||||
throw new Error(`Failed to download: ${url}`);
|
|
||||||
}
|
|
||||||
|
|
||||||
return new Promise((resolve, reject) => {
|
|
||||||
res.body
|
|
||||||
.on('error', reject)
|
|
||||||
.pipe(tar.extract({ cwd: dir, strip: 1 }))
|
|
||||||
.on('finish', () => resolve(path.join(dir, 'bin', 'go')));
|
|
||||||
});
|
|
||||||
};
|
|
||||||
124
packages/now-go/go-helpers.js
Normal file
124
packages/now-go/go-helpers.js
Normal file
@@ -0,0 +1,124 @@
|
|||||||
|
const tar = require('tar');
|
||||||
|
const execa = require('execa');
|
||||||
|
const fetch = require('node-fetch');
|
||||||
|
const { mkdirp } = require('fs-extra');
|
||||||
|
const { dirname, join } = require('path');
|
||||||
|
const debug = require('debug')('@now/go:go-helpers');
|
||||||
|
|
||||||
|
const archMap = new Map([['x64', 'amd64'], ['x86', '386']]);
|
||||||
|
const platformMap = new Map([['win32', 'windows']]);
|
||||||
|
|
||||||
|
// Location where the `go` binary will be installed after `postinstall`
|
||||||
|
const GO_DIR = join(__dirname, 'go');
|
||||||
|
const GO_BIN = join(GO_DIR, 'bin/go');
|
||||||
|
|
||||||
|
const getPlatform = p => platformMap.get(p) || p;
|
||||||
|
const getArch = a => archMap.get(a) || a;
|
||||||
|
const getGoUrl = (version, platform, arch) => {
|
||||||
|
const goArch = getArch(arch);
|
||||||
|
const goPlatform = getPlatform(platform);
|
||||||
|
const ext = platform === 'win32' ? 'zip' : 'tar.gz';
|
||||||
|
return `https://dl.google.com/go/go${version}.${goPlatform}-${goArch}.${ext}`;
|
||||||
|
};
|
||||||
|
|
||||||
|
function getExportedFunctionName(filePath) {
|
||||||
|
debug('Detecting handler name for %o', filePath);
|
||||||
|
const bin = join(__dirname, 'get-exported-function-name');
|
||||||
|
const args = [filePath];
|
||||||
|
const name = execa.stdout(bin, args);
|
||||||
|
debug('Detected exported name %o', filePath);
|
||||||
|
return name;
|
||||||
|
}
|
||||||
|
|
||||||
|
// Creates a `$GOPATH` directory tree, as per `go help gopath` instructions.
|
||||||
|
// Without this, `go` won't recognize the `$GOPATH`.
|
||||||
|
function createGoPathTree(goPath, platform, arch) {
|
||||||
|
const tuple = `${getPlatform(platform)}_${getArch(arch)}`;
|
||||||
|
debug('Creating GOPATH directory structure for %o (%s)', goPath, tuple);
|
||||||
|
return Promise.all([
|
||||||
|
mkdirp(join(goPath, 'bin')),
|
||||||
|
mkdirp(join(goPath, 'pkg', tuple)),
|
||||||
|
]);
|
||||||
|
}
|
||||||
|
|
||||||
|
async function get({ src } = {}) {
|
||||||
|
const args = ['get'];
|
||||||
|
if (src) {
|
||||||
|
debug('Fetching `go` dependencies for file %o', src);
|
||||||
|
args.push(src);
|
||||||
|
} else {
|
||||||
|
debug('Fetching `go` dependencies for cwd %o', this.cwd);
|
||||||
|
}
|
||||||
|
await this(...args);
|
||||||
|
}
|
||||||
|
|
||||||
|
async function build({ src, dest }) {
|
||||||
|
debug('Building `go` binary %o -> %o', src, dest);
|
||||||
|
let sources;
|
||||||
|
if (Array.isArray(src)) {
|
||||||
|
sources = src;
|
||||||
|
} else {
|
||||||
|
sources = [src];
|
||||||
|
}
|
||||||
|
await this('build', '-o', dest, ...sources);
|
||||||
|
}
|
||||||
|
|
||||||
|
async function createGo(
|
||||||
|
goPath,
|
||||||
|
platform = process.platform,
|
||||||
|
arch = process.arch,
|
||||||
|
opts = {},
|
||||||
|
) {
|
||||||
|
const env = {
|
||||||
|
...process.env,
|
||||||
|
PATH: `${dirname(GO_BIN)}:${process.env.PATH}`,
|
||||||
|
GOPATH: goPath,
|
||||||
|
...opts.env,
|
||||||
|
};
|
||||||
|
|
||||||
|
function go(...args) {
|
||||||
|
debug('Exec %o', `go ${args.join(' ')}`);
|
||||||
|
return execa('go', args, { stdio: 'inherit', ...opts, env });
|
||||||
|
}
|
||||||
|
go.cwd = opts.cwd || process.cwd();
|
||||||
|
go.get = get;
|
||||||
|
go.build = build;
|
||||||
|
go.goPath = goPath;
|
||||||
|
await createGoPathTree(goPath, platform, arch);
|
||||||
|
return go;
|
||||||
|
}
|
||||||
|
|
||||||
|
async function downloadGo(
|
||||||
|
dir = GO_DIR,
|
||||||
|
version = '1.11.5',
|
||||||
|
platform = process.platform,
|
||||||
|
arch = process.arch,
|
||||||
|
) {
|
||||||
|
debug('Installing `go` v%s to %o for %s %s', version, dir, platform, arch);
|
||||||
|
|
||||||
|
const url = getGoUrl(version, platform, arch);
|
||||||
|
debug('Downloading `go` URL: %o', url);
|
||||||
|
const res = await fetch(url);
|
||||||
|
|
||||||
|
if (!res.ok) {
|
||||||
|
throw new Error(`Failed to download: ${url} (${res.status})`);
|
||||||
|
}
|
||||||
|
|
||||||
|
// TODO: use a zip extractor when `ext === "zip"`
|
||||||
|
await mkdirp(dir);
|
||||||
|
await new Promise((resolve, reject) => {
|
||||||
|
res.body
|
||||||
|
.on('error', reject)
|
||||||
|
.pipe(tar.extract({ cwd: dir, strip: 1 }))
|
||||||
|
.on('error', reject)
|
||||||
|
.on('finish', resolve);
|
||||||
|
});
|
||||||
|
|
||||||
|
return createGo(dir, platform, arch);
|
||||||
|
}
|
||||||
|
|
||||||
|
module.exports = {
|
||||||
|
createGo,
|
||||||
|
downloadGo,
|
||||||
|
getExportedFunctionName,
|
||||||
|
};
|
||||||
@@ -1,79 +1,52 @@
|
|||||||
const path = require('path');
|
const { join, dirname } = require('path');
|
||||||
const { mkdirp, readFile, writeFile } = require('fs-extra');
|
const { readFile, writeFile } = require('fs-extra');
|
||||||
|
|
||||||
const execa = require('execa');
|
const glob = require('@now/build-utils/fs/glob.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
|
const download = require('@now/build-utils/fs/download.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
const { createLambda } = require('@now/build-utils/lambda.js'); // eslint-disable-line import/no-extraneous-dependencies
|
const { createLambda } = require('@now/build-utils/lambda.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
const getWritableDirectory = require('@now/build-utils/fs/get-writable-directory.js'); // eslint-disable-line import/no-extraneous-dependencies
|
const getWritableDirectory = require('@now/build-utils/fs/get-writable-directory.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
const download = require('@now/build-utils/fs/download.js'); // eslint-disable-line import/no-extraneous-dependencies
|
const { createGo, getExportedFunctionName } = require('./go-helpers');
|
||||||
const downloadGit = require('lambda-git');
|
|
||||||
const glob = require('@now/build-utils/fs/glob.js'); // eslint-disable-line import/no-extraneous-dependencies
|
|
||||||
const downloadGoBin = require('./download-go-bin');
|
|
||||||
|
|
||||||
// creates a `$GOPATH` directory tree, as per
|
const config = {
|
||||||
// `go help gopath`'s instructions.
|
|
||||||
// without this, Go won't recognize the `$GOPATH`
|
|
||||||
async function createGoPathTree(goPath) {
|
|
||||||
await mkdirp(path.join(goPath, 'bin'));
|
|
||||||
await mkdirp(path.join(goPath, 'pkg', 'linux_amd64'));
|
|
||||||
}
|
|
||||||
|
|
||||||
exports.config = {
|
|
||||||
maxLambdaSize: '10mb',
|
maxLambdaSize: '10mb',
|
||||||
};
|
};
|
||||||
|
|
||||||
exports.build = async ({ files, entrypoint }) => {
|
async function build({ files, entrypoint }) {
|
||||||
console.log('downloading files...');
|
console.log('Downloading user files...');
|
||||||
|
|
||||||
const gitPath = await getWritableDirectory();
|
const [goPath, outDir] = await Promise.all([
|
||||||
const goPath = await getWritableDirectory();
|
getWritableDirectory(),
|
||||||
const srcPath = path.join(goPath, 'src', 'lambda');
|
getWritableDirectory(),
|
||||||
const outDir = await getWritableDirectory();
|
]);
|
||||||
|
|
||||||
await createGoPathTree(goPath);
|
|
||||||
|
|
||||||
|
const srcPath = join(goPath, 'src', 'lambda');
|
||||||
const downloadedFiles = await download(files, srcPath);
|
const downloadedFiles = await download(files, srcPath);
|
||||||
|
|
||||||
console.log('downloading go binary...');
|
console.log(`Parsing AST for "${entrypoint}"`);
|
||||||
const goBin = await downloadGoBin();
|
let handlerFunctionName;
|
||||||
|
|
||||||
console.log('downloading git binary...');
|
|
||||||
// downloads a git binary that works on Amazon Linux and sets
|
|
||||||
// `process.env.GIT_EXEC_PATH` so `go(1)` can see it
|
|
||||||
await downloadGit({ targetDirectory: gitPath });
|
|
||||||
|
|
||||||
const goEnv = {
|
|
||||||
...process.env,
|
|
||||||
GOOS: 'linux',
|
|
||||||
GOARCH: 'amd64',
|
|
||||||
GOPATH: goPath,
|
|
||||||
};
|
|
||||||
|
|
||||||
console.log(`parsing AST for "${entrypoint}"`);
|
|
||||||
let handlerFunctionName = '';
|
|
||||||
try {
|
try {
|
||||||
handlerFunctionName = await execa.stdout(
|
handlerFunctionName = await getExportedFunctionName(
|
||||||
path.join(__dirname, 'bin', 'get-exported-function-name'),
|
downloadedFiles[entrypoint].fsPath,
|
||||||
[downloadedFiles[entrypoint].fsPath],
|
|
||||||
);
|
);
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
console.log(`failed to parse AST for "${entrypoint}"`);
|
console.log(`Failed to parse AST for "${entrypoint}"`);
|
||||||
throw err;
|
throw err;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (handlerFunctionName === '') {
|
if (!handlerFunctionName) {
|
||||||
const e = new Error(
|
const err = new Error(
|
||||||
`Could not find an exported function on "${entrypoint}"`,
|
`Could not find an exported function in "${entrypoint}"`,
|
||||||
);
|
);
|
||||||
console.log(e.message);
|
console.log(err.message);
|
||||||
throw e;
|
throw err;
|
||||||
}
|
}
|
||||||
|
|
||||||
console.log(
|
console.log(
|
||||||
`Found exported function "${handlerFunctionName}" on "${entrypoint}"`,
|
`Found exported function "${handlerFunctionName}" in "${entrypoint}"`,
|
||||||
);
|
);
|
||||||
|
|
||||||
const origianlMainGoContents = await readFile(
|
const origianlMainGoContents = await readFile(
|
||||||
path.join(__dirname, 'main.go'),
|
join(__dirname, 'main.go'),
|
||||||
'utf8',
|
'utf8',
|
||||||
);
|
);
|
||||||
const mainGoContents = origianlMainGoContents.replace(
|
const mainGoContents = origianlMainGoContents.replace(
|
||||||
@@ -85,39 +58,33 @@ exports.build = async ({ files, entrypoint }) => {
|
|||||||
|
|
||||||
// we need `main.go` in the same dir as the entrypoint,
|
// we need `main.go` in the same dir as the entrypoint,
|
||||||
// otherwise `go build` will refuse to build
|
// otherwise `go build` will refuse to build
|
||||||
const entrypointDirname = path.dirname(downloadedFiles[entrypoint].fsPath);
|
const entrypointDirname = dirname(downloadedFiles[entrypoint].fsPath);
|
||||||
|
|
||||||
// Go doesn't like to build files in different directories,
|
// Go doesn't like to build files in different directories,
|
||||||
// so now we place `main.go` together with the user code
|
// so now we place `main.go` together with the user code
|
||||||
await writeFile(path.join(entrypointDirname, mainGoFileName), mainGoContents);
|
await writeFile(join(entrypointDirname, mainGoFileName), mainGoContents);
|
||||||
|
|
||||||
console.log('installing dependencies');
|
const go = await createGo(goPath, process.platform, process.arch, {
|
||||||
// `go get` will look at `*.go` (note we set `cwd`), parse
|
|
||||||
// the `import`s and download any packages that aren't part of the stdlib
|
|
||||||
try {
|
|
||||||
await execa(goBin, ['get'], {
|
|
||||||
env: goEnv,
|
|
||||||
cwd: entrypointDirname,
|
cwd: entrypointDirname,
|
||||||
stdio: 'inherit',
|
|
||||||
});
|
});
|
||||||
|
|
||||||
|
// `go get` will look at `*.go` (note we set `cwd`), parse the `import`s
|
||||||
|
// and download any packages that aren't part of the stdlib
|
||||||
|
try {
|
||||||
|
await go.get();
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
console.log('failed to `go get`');
|
console.log('failed to `go get`');
|
||||||
throw err;
|
throw err;
|
||||||
}
|
}
|
||||||
|
|
||||||
console.log('running go build...');
|
console.log('Running `go build`...');
|
||||||
|
const destPath = join(outDir, 'handler');
|
||||||
try {
|
try {
|
||||||
await execa(
|
const src = [
|
||||||
goBin,
|
join(entrypointDirname, mainGoFileName),
|
||||||
[
|
|
||||||
'build',
|
|
||||||
'-o',
|
|
||||||
path.join(outDir, 'handler'),
|
|
||||||
path.join(entrypointDirname, mainGoFileName),
|
|
||||||
downloadedFiles[entrypoint].fsPath,
|
downloadedFiles[entrypoint].fsPath,
|
||||||
],
|
];
|
||||||
{ env: goEnv, cwd: entrypointDirname, stdio: 'inherit' },
|
await go.build({ src, dest: destPath });
|
||||||
);
|
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
console.log('failed to `go build`');
|
console.log('failed to `go build`');
|
||||||
throw err;
|
throw err;
|
||||||
@@ -133,4 +100,6 @@ exports.build = async ({ files, entrypoint }) => {
|
|||||||
return {
|
return {
|
||||||
[entrypoint]: lambda,
|
[entrypoint]: lambda,
|
||||||
};
|
};
|
||||||
};
|
}
|
||||||
|
|
||||||
|
module.exports = { config, build };
|
||||||
|
|||||||
@@ -1,8 +1,8 @@
|
|||||||
package main
|
package main
|
||||||
|
|
||||||
import (
|
import (
|
||||||
now "../../utils/go/bridge"
|
|
||||||
"net/http"
|
"net/http"
|
||||||
|
now "github.com/zeit/now-builders/utils/go/bridge"
|
||||||
)
|
)
|
||||||
|
|
||||||
func main() {
|
func main() {
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/go",
|
"name": "@now/go",
|
||||||
"version": "0.2.13-canary.0",
|
"version": "0.2.13-canary.1",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
@@ -8,25 +8,19 @@
|
|||||||
"directory": "packages/now-go"
|
"directory": "packages/now-go"
|
||||||
},
|
},
|
||||||
"scripts": {
|
"scripts": {
|
||||||
"test": "best -I test/*.js",
|
"postinstall": "node ./util/install"
|
||||||
"prepublish": "./build.sh"
|
|
||||||
},
|
},
|
||||||
"files": [
|
"files": [
|
||||||
"bin",
|
"*.js",
|
||||||
"download-go-bin.js",
|
"main.go",
|
||||||
"index.js",
|
"util"
|
||||||
"main.go"
|
|
||||||
],
|
],
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
|
"debug": "^4.1.1",
|
||||||
"execa": "^1.0.0",
|
"execa": "^1.0.0",
|
||||||
"fs-extra": "^7.0.0",
|
"fs-extra": "^7.0.0",
|
||||||
"lambda-git": "^0.1.2",
|
|
||||||
"mkdirp-promise": "5.0.1",
|
"mkdirp-promise": "5.0.1",
|
||||||
"node-fetch": "^2.2.1",
|
"node-fetch": "^2.2.1",
|
||||||
"tar": "4.4.6"
|
"tar": "4.4.6"
|
||||||
},
|
|
||||||
"devDependencies": {
|
|
||||||
"@zeit/best": "0.4.3",
|
|
||||||
"rmfr": "2.0.0"
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
18
packages/now-go/util/install.js
Normal file
18
packages/now-go/util/install.js
Normal file
@@ -0,0 +1,18 @@
|
|||||||
|
const { join } = require('path');
|
||||||
|
const { downloadGo } = require('../go-helpers');
|
||||||
|
|
||||||
|
async function main() {
|
||||||
|
// First download the `go` binary for this platform/arch.
|
||||||
|
const go = await downloadGo();
|
||||||
|
|
||||||
|
// Build the `get-exported-function-name` helper program.
|
||||||
|
// `go get` is not necessary because the program has no external deps.
|
||||||
|
const src = join(__dirname, 'get-exported-function-name.go');
|
||||||
|
const dest = join(__dirname, '../get-exported-function-name');
|
||||||
|
await go.build({ src, dest });
|
||||||
|
}
|
||||||
|
|
||||||
|
main().catch((err) => {
|
||||||
|
console.error(err);
|
||||||
|
process.exit(1);
|
||||||
|
});
|
||||||
File diff suppressed because it is too large
Load Diff
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/lambda",
|
"name": "@now/lambda",
|
||||||
"version": "0.4.10-canary.0",
|
"version": "0.4.10-canary.1",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ const {
|
|||||||
testDeployment,
|
testDeployment,
|
||||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||||
|
|
||||||
jest.setTimeout(2 * 60 * 1000);
|
jest.setTimeout(4 * 60 * 1000);
|
||||||
const buildUtilsUrl = '@canary';
|
const buildUtilsUrl = '@canary';
|
||||||
let builderUrl;
|
let builderUrl;
|
||||||
|
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/md",
|
"name": "@now/md",
|
||||||
"version": "0.4.10-canary.0",
|
"version": "0.4.10-canary.1",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ const {
|
|||||||
testDeployment,
|
testDeployment,
|
||||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||||
|
|
||||||
jest.setTimeout(2 * 60 * 1000);
|
jest.setTimeout(4 * 60 * 1000);
|
||||||
const buildUtilsUrl = '@canary';
|
const buildUtilsUrl = '@canary';
|
||||||
let builderUrl;
|
let builderUrl;
|
||||||
|
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/mdx-deck",
|
"name": "@now/mdx-deck",
|
||||||
"version": "0.4.19-canary.0",
|
"version": "0.4.19-canary.1",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ const {
|
|||||||
testDeployment,
|
testDeployment,
|
||||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||||
|
|
||||||
jest.setTimeout(2 * 60 * 1000);
|
jest.setTimeout(4 * 60 * 1000);
|
||||||
const buildUtilsUrl = '@canary';
|
const buildUtilsUrl = '@canary';
|
||||||
let builderUrl;
|
let builderUrl;
|
||||||
|
|
||||||
|
|||||||
@@ -9,15 +9,14 @@ const {
|
|||||||
runPackageJsonScript,
|
runPackageJsonScript,
|
||||||
} = require('@now/build-utils/fs/run-user-scripts.js'); // eslint-disable-line import/no-extraneous-dependencies
|
} = require('@now/build-utils/fs/run-user-scripts.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
const glob = require('@now/build-utils/fs/glob.js'); // eslint-disable-line import/no-extraneous-dependencies
|
const glob = require('@now/build-utils/fs/glob.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
|
const fs = require('fs-extra');
|
||||||
const semver = require('semver');
|
const semver = require('semver');
|
||||||
const nextLegacyVersions = require('./legacy-versions');
|
const nextLegacyVersions = require('./legacy-versions');
|
||||||
const {
|
const {
|
||||||
excludeFiles,
|
excludeFiles,
|
||||||
validateEntrypoint,
|
validateEntrypoint,
|
||||||
includeOnlyEntryDirectory,
|
includeOnlyEntryDirectory,
|
||||||
moveEntryDirectoryToRoot,
|
|
||||||
normalizePackageJson,
|
normalizePackageJson,
|
||||||
excludeStaticDirectory,
|
|
||||||
onlyStaticDirectory,
|
onlyStaticDirectory,
|
||||||
} = require('./utils');
|
} = require('./utils');
|
||||||
|
|
||||||
@@ -33,15 +32,17 @@ const {
|
|||||||
|
|
||||||
/**
|
/**
|
||||||
* Read package.json from files
|
* Read package.json from files
|
||||||
* @param {DownloadedFiles} files
|
* @param {string} entryPath
|
||||||
*/
|
*/
|
||||||
async function readPackageJson(files) {
|
async function readPackageJson(entryPath) {
|
||||||
if (!files['package.json']) {
|
const packagePath = path.join(entryPath, 'package.json');
|
||||||
|
|
||||||
|
try {
|
||||||
|
return JSON.parse(await readFile(packagePath, 'utf8'));
|
||||||
|
} catch (err) {
|
||||||
|
console.log('package.json not found in entry');
|
||||||
return {};
|
return {};
|
||||||
}
|
}
|
||||||
|
|
||||||
const packageJsonPath = files['package.json'].fsPath;
|
|
||||||
return JSON.parse(await readFile(packageJsonPath, 'utf8'));
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
@@ -68,48 +69,17 @@ async function writeNpmRc(workPath, token) {
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
exports.config = {
|
function getNextVersion(packageJson) {
|
||||||
maxLambdaSize: '5mb',
|
|
||||||
};
|
|
||||||
|
|
||||||
/**
|
|
||||||
* @param {BuildParamsType} buildParams
|
|
||||||
* @returns {Promise<Files>}
|
|
||||||
*/
|
|
||||||
exports.build = async ({ files, workPath, entrypoint }) => {
|
|
||||||
validateEntrypoint(entrypoint);
|
|
||||||
|
|
||||||
console.log('downloading user files...');
|
|
||||||
const entryDirectory = path.dirname(entrypoint);
|
|
||||||
const filesOnlyEntryDirectory = includeOnlyEntryDirectory(
|
|
||||||
files,
|
|
||||||
entryDirectory,
|
|
||||||
);
|
|
||||||
const filesWithEntryDirectoryRoot = moveEntryDirectoryToRoot(
|
|
||||||
filesOnlyEntryDirectory,
|
|
||||||
entryDirectory,
|
|
||||||
);
|
|
||||||
const filesWithoutStaticDirectory = excludeStaticDirectory(
|
|
||||||
filesWithEntryDirectoryRoot,
|
|
||||||
);
|
|
||||||
const downloadedFiles = await download(filesWithoutStaticDirectory, workPath);
|
|
||||||
|
|
||||||
const pkg = await readPackageJson(downloadedFiles);
|
|
||||||
|
|
||||||
let nextVersion;
|
let nextVersion;
|
||||||
if (pkg.dependencies && pkg.dependencies.next) {
|
if (packageJson.dependencies && packageJson.dependencies.next) {
|
||||||
nextVersion = pkg.dependencies.next;
|
nextVersion = packageJson.dependencies.next;
|
||||||
} else if (pkg.devDependencies && pkg.devDependencies.next) {
|
} else if (packageJson.devDependencies && packageJson.devDependencies.next) {
|
||||||
nextVersion = pkg.devDependencies.next;
|
nextVersion = packageJson.devDependencies.next;
|
||||||
}
|
}
|
||||||
|
return nextVersion;
|
||||||
|
}
|
||||||
|
|
||||||
if (!nextVersion) {
|
function isLegacyNext(nextVersion) {
|
||||||
throw new Error(
|
|
||||||
'No Next.js version could be detected in "package.json". Make sure `"next"` is installed in "dependencies" or "devDependencies"',
|
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
const isLegacy = (() => {
|
|
||||||
// If version is using the dist-tag instead of a version range
|
// If version is using the dist-tag instead of a version range
|
||||||
if (nextVersion === 'canary' || nextVersion === 'latest') {
|
if (nextVersion === 'canary' || nextVersion === 'latest') {
|
||||||
return false;
|
return false;
|
||||||
@@ -127,19 +97,46 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
|||||||
}
|
}
|
||||||
|
|
||||||
return true;
|
return true;
|
||||||
})();
|
}
|
||||||
|
|
||||||
|
exports.config = {
|
||||||
|
maxLambdaSize: '5mb',
|
||||||
|
};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* @param {BuildParamsType} buildParams
|
||||||
|
* @returns {Promise<Files>}
|
||||||
|
*/
|
||||||
|
exports.build = async ({ files, workPath, entrypoint }) => {
|
||||||
|
validateEntrypoint(entrypoint);
|
||||||
|
|
||||||
|
console.log('downloading user files...');
|
||||||
|
const entryDirectory = path.dirname(entrypoint);
|
||||||
|
await download(files, workPath);
|
||||||
|
const entryPath = path.join(workPath, entryDirectory);
|
||||||
|
|
||||||
|
const pkg = await readPackageJson(entryPath);
|
||||||
|
|
||||||
|
const nextVersion = getNextVersion(pkg);
|
||||||
|
if (!nextVersion) {
|
||||||
|
throw new Error(
|
||||||
|
'No Next.js version could be detected in "package.json". Make sure `"next"` is installed in "dependencies" or "devDependencies"',
|
||||||
|
);
|
||||||
|
}
|
||||||
|
|
||||||
|
const isLegacy = isLegacyNext(nextVersion);
|
||||||
|
|
||||||
console.log(`MODE: ${isLegacy ? 'legacy' : 'serverless'}`);
|
console.log(`MODE: ${isLegacy ? 'legacy' : 'serverless'}`);
|
||||||
|
|
||||||
if (isLegacy) {
|
if (isLegacy) {
|
||||||
try {
|
try {
|
||||||
await unlink(path.join(workPath, 'yarn.lock'));
|
await unlink(path.join(entryPath, 'yarn.lock'));
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
console.log('no yarn.lock removed');
|
console.log('no yarn.lock removed');
|
||||||
}
|
}
|
||||||
|
|
||||||
try {
|
try {
|
||||||
await unlink(path.join(workPath, 'package-lock.json'));
|
await unlink(path.join(entryPath, 'package-lock.json'));
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
console.log('no package-lock.json removed');
|
console.log('no package-lock.json removed');
|
||||||
}
|
}
|
||||||
@@ -151,7 +148,7 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
|||||||
console.log('normalizing package.json');
|
console.log('normalizing package.json');
|
||||||
const packageJson = normalizePackageJson(pkg);
|
const packageJson = normalizePackageJson(pkg);
|
||||||
console.log('normalized package.json result: ', packageJson);
|
console.log('normalized package.json result: ', packageJson);
|
||||||
await writePackageJson(workPath, packageJson);
|
await writePackageJson(entryPath, packageJson);
|
||||||
} else if (!pkg.scripts || !pkg.scripts['now-build']) {
|
} else if (!pkg.scripts || !pkg.scripts['now-build']) {
|
||||||
console.warn(
|
console.warn(
|
||||||
'WARNING: "now-build" script not found. Adding \'"now-build": "next build"\' to "package.json" automatically',
|
'WARNING: "now-build" script not found. Adding \'"now-build": "next build"\' to "package.json" automatically',
|
||||||
@@ -161,38 +158,38 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
|||||||
...(pkg.scripts || {}),
|
...(pkg.scripts || {}),
|
||||||
};
|
};
|
||||||
console.log('normalized package.json result: ', pkg);
|
console.log('normalized package.json result: ', pkg);
|
||||||
await writePackageJson(workPath, pkg);
|
await writePackageJson(entryPath, pkg);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (process.env.NPM_AUTH_TOKEN) {
|
if (process.env.NPM_AUTH_TOKEN) {
|
||||||
console.log('found NPM_AUTH_TOKEN in environment, creating .npmrc');
|
console.log('found NPM_AUTH_TOKEN in environment, creating .npmrc');
|
||||||
await writeNpmRc(workPath, process.env.NPM_AUTH_TOKEN);
|
await writeNpmRc(entryPath, process.env.NPM_AUTH_TOKEN);
|
||||||
}
|
}
|
||||||
|
|
||||||
console.log('installing dependencies...');
|
console.log('installing dependencies...');
|
||||||
await runNpmInstall(workPath, ['--prefer-offline']);
|
await runNpmInstall(entryPath, ['--prefer-offline']);
|
||||||
console.log('running user script...');
|
console.log('running user script...');
|
||||||
await runPackageJsonScript(workPath, 'now-build');
|
await runPackageJsonScript(entryPath, 'now-build');
|
||||||
|
|
||||||
if (isLegacy) {
|
if (isLegacy) {
|
||||||
console.log('running npm install --production...');
|
console.log('running npm install --production...');
|
||||||
await runNpmInstall(workPath, ['--prefer-offline', '--production']);
|
await runNpmInstall(entryPath, ['--prefer-offline', '--production']);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (process.env.NPM_AUTH_TOKEN) {
|
if (process.env.NPM_AUTH_TOKEN) {
|
||||||
await unlink(path.join(workPath, '.npmrc'));
|
await unlink(path.join(entryPath, '.npmrc'));
|
||||||
}
|
}
|
||||||
|
|
||||||
const lambdas = {};
|
const lambdas = {};
|
||||||
|
|
||||||
if (isLegacy) {
|
if (isLegacy) {
|
||||||
const filesAfterBuild = await glob('**', workPath);
|
const filesAfterBuild = await glob('**', entryPath);
|
||||||
|
|
||||||
console.log('preparing lambda files...');
|
console.log('preparing lambda files...');
|
||||||
let buildId;
|
let buildId;
|
||||||
try {
|
try {
|
||||||
buildId = await readFile(
|
buildId = await readFile(
|
||||||
path.join(workPath, '.next', 'BUILD_ID'),
|
path.join(entryPath, '.next', 'BUILD_ID'),
|
||||||
'utf8',
|
'utf8',
|
||||||
);
|
);
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
@@ -201,10 +198,10 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
|||||||
);
|
);
|
||||||
throw new Error('Missing BUILD_ID');
|
throw new Error('Missing BUILD_ID');
|
||||||
}
|
}
|
||||||
const dotNextRootFiles = await glob('.next/*', workPath);
|
const dotNextRootFiles = await glob('.next/*', entryPath);
|
||||||
const dotNextServerRootFiles = await glob('.next/server/*', workPath);
|
const dotNextServerRootFiles = await glob('.next/server/*', entryPath);
|
||||||
const nodeModules = excludeFiles(
|
const nodeModules = excludeFiles(
|
||||||
await glob('node_modules/**', workPath),
|
await glob('node_modules/**', entryPath),
|
||||||
file => file.startsWith('node_modules/.cache'),
|
file => file.startsWith('node_modules/.cache'),
|
||||||
);
|
);
|
||||||
const launcherFiles = {
|
const launcherFiles = {
|
||||||
@@ -221,7 +218,7 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
|||||||
}
|
}
|
||||||
const pages = await glob(
|
const pages = await glob(
|
||||||
'**/*.js',
|
'**/*.js',
|
||||||
path.join(workPath, '.next', 'server', 'static', buildId, 'pages'),
|
path.join(entryPath, '.next', 'server', 'static', buildId, 'pages'),
|
||||||
);
|
);
|
||||||
const launcherPath = path.join(__dirname, 'legacy-launcher.js');
|
const launcherPath = path.join(__dirname, 'legacy-launcher.js');
|
||||||
const launcherData = await readFile(launcherPath, 'utf8');
|
const launcherData = await readFile(launcherPath, 'utf8');
|
||||||
@@ -277,7 +274,7 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
|||||||
};
|
};
|
||||||
const pages = await glob(
|
const pages = await glob(
|
||||||
'**/*.js',
|
'**/*.js',
|
||||||
path.join(workPath, '.next', 'serverless', 'pages'),
|
path.join(entryPath, '.next', 'serverless', 'pages'),
|
||||||
);
|
);
|
||||||
|
|
||||||
const pageKeys = Object.keys(pages);
|
const pageKeys = Object.keys(pages);
|
||||||
@@ -288,6 +285,18 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
|||||||
);
|
);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// An optional assets folder that is placed alongside every page entrypoint
|
||||||
|
const assets = await glob(
|
||||||
|
'assets/**',
|
||||||
|
path.join(entryPath, '.next', 'serverless'),
|
||||||
|
);
|
||||||
|
|
||||||
|
const assetKeys = Object.keys(assets);
|
||||||
|
if (assetKeys.length > 0) {
|
||||||
|
console.log('detected assets to be bundled with lambda:');
|
||||||
|
assetKeys.forEach(assetFile => console.log(`\t${assetFile}`));
|
||||||
|
}
|
||||||
|
|
||||||
await Promise.all(
|
await Promise.all(
|
||||||
pageKeys.map(async (page) => {
|
pageKeys.map(async (page) => {
|
||||||
// These default pages don't have to be handled as they'd always 404
|
// These default pages don't have to be handled as they'd always 404
|
||||||
@@ -301,6 +310,7 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
|||||||
lambdas[path.join(entryDirectory, pathname)] = await createLambda({
|
lambdas[path.join(entryDirectory, pathname)] = await createLambda({
|
||||||
files: {
|
files: {
|
||||||
...launcherFiles,
|
...launcherFiles,
|
||||||
|
...assets,
|
||||||
'page.js': pages[page],
|
'page.js': pages[page],
|
||||||
},
|
},
|
||||||
handler: 'now__launcher.launcher',
|
handler: 'now__launcher.launcher',
|
||||||
@@ -313,17 +323,21 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
|||||||
|
|
||||||
const nextStaticFiles = await glob(
|
const nextStaticFiles = await glob(
|
||||||
'**',
|
'**',
|
||||||
path.join(workPath, '.next', 'static'),
|
path.join(entryPath, '.next', 'static'),
|
||||||
);
|
);
|
||||||
const staticFiles = Object.keys(nextStaticFiles).reduce(
|
const staticFiles = Object.keys(nextStaticFiles).reduce(
|
||||||
(mappedFiles, file) => ({
|
(mappedFiles, file) => ({
|
||||||
...mappedFiles,
|
...mappedFiles,
|
||||||
[path.join(entryDirectory, `_next/static/${file}`)]: nextStaticFiles[file],
|
[path.join(entryDirectory, `_next/static/${file}`)]: nextStaticFiles[
|
||||||
|
file
|
||||||
|
],
|
||||||
}),
|
}),
|
||||||
{},
|
{},
|
||||||
);
|
);
|
||||||
|
|
||||||
const nextStaticDirectory = onlyStaticDirectory(filesWithEntryDirectoryRoot);
|
const nextStaticDirectory = onlyStaticDirectory(
|
||||||
|
includeOnlyEntryDirectory(files, entryDirectory),
|
||||||
|
);
|
||||||
const staticDirectoryFiles = Object.keys(nextStaticDirectory).reduce(
|
const staticDirectoryFiles = Object.keys(nextStaticDirectory).reduce(
|
||||||
(mappedFiles, file) => ({
|
(mappedFiles, file) => ({
|
||||||
...mappedFiles,
|
...mappedFiles,
|
||||||
@@ -334,3 +348,39 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
|||||||
|
|
||||||
return { ...lambdas, ...staticFiles, ...staticDirectoryFiles };
|
return { ...lambdas, ...staticFiles, ...staticDirectoryFiles };
|
||||||
};
|
};
|
||||||
|
|
||||||
|
exports.prepareCache = async ({ cachePath, workPath, entrypoint }) => {
|
||||||
|
console.log('preparing cache ...');
|
||||||
|
|
||||||
|
const entryDirectory = path.dirname(entrypoint);
|
||||||
|
const entryPath = path.join(workPath, entryDirectory);
|
||||||
|
const cacheEntryPath = path.join(cachePath, entryDirectory);
|
||||||
|
|
||||||
|
const pkg = await readPackageJson(entryPath);
|
||||||
|
const nextVersion = getNextVersion(pkg);
|
||||||
|
const isLegacy = isLegacyNext(nextVersion);
|
||||||
|
|
||||||
|
if (isLegacy) {
|
||||||
|
// skip caching legacy mode (swapping deps between all and production can get bug-prone)
|
||||||
|
return {};
|
||||||
|
}
|
||||||
|
|
||||||
|
console.log('clearing old cache ...');
|
||||||
|
fs.removeSync(cacheEntryPath);
|
||||||
|
fs.mkdirpSync(cacheEntryPath);
|
||||||
|
|
||||||
|
console.log('copying build files for cache ...');
|
||||||
|
fs.renameSync(entryPath, cacheEntryPath);
|
||||||
|
|
||||||
|
console.log('producing cache file manifest ...');
|
||||||
|
|
||||||
|
const cacheEntrypoint = path.relative(cachePath, cacheEntryPath);
|
||||||
|
return {
|
||||||
|
...(await glob(
|
||||||
|
path.join(cacheEntrypoint, 'node_modules/{**,!.*,.yarn*}'),
|
||||||
|
cachePath,
|
||||||
|
)),
|
||||||
|
...(await glob(path.join(cacheEntrypoint, 'package-lock.json'), cachePath)),
|
||||||
|
...(await glob(path.join(cacheEntrypoint, 'yarn.lock'), cachePath)),
|
||||||
|
};
|
||||||
|
};
|
||||||
|
|||||||
@@ -4,10 +4,8 @@ const { Server } = require('http');
|
|||||||
const { Bridge } = require('./now__bridge.js');
|
const { Bridge } = require('./now__bridge.js');
|
||||||
const page = require('./page.js');
|
const page = require('./page.js');
|
||||||
|
|
||||||
const bridge = new Bridge();
|
|
||||||
bridge.port = 3000;
|
|
||||||
|
|
||||||
const server = new Server(page.render);
|
const server = new Server(page.render);
|
||||||
server.listen(bridge.port);
|
const bridge = new Bridge(server);
|
||||||
|
bridge.listen();
|
||||||
|
|
||||||
exports.launcher = bridge.launcher;
|
exports.launcher = bridge.launcher;
|
||||||
|
|||||||
@@ -3,9 +3,6 @@ const next = require('next-server');
|
|||||||
const url = require('url');
|
const url = require('url');
|
||||||
const { Bridge } = require('./now__bridge.js');
|
const { Bridge } = require('./now__bridge.js');
|
||||||
|
|
||||||
const bridge = new Bridge();
|
|
||||||
bridge.port = 3000;
|
|
||||||
|
|
||||||
process.env.NODE_ENV = 'production';
|
process.env.NODE_ENV = 'production';
|
||||||
|
|
||||||
const app = next({});
|
const app = next({});
|
||||||
@@ -14,6 +11,8 @@ const server = new Server((req, res) => {
|
|||||||
const parsedUrl = url.parse(req.url, true);
|
const parsedUrl = url.parse(req.url, true);
|
||||||
app.render(req, res, 'PATHNAME_PLACEHOLDER', parsedUrl.query, parsedUrl);
|
app.render(req, res, 'PATHNAME_PLACEHOLDER', parsedUrl.query, parsedUrl);
|
||||||
});
|
});
|
||||||
server.listen(bridge.port);
|
|
||||||
|
const bridge = new Bridge(server);
|
||||||
|
bridge.listen();
|
||||||
|
|
||||||
exports.launcher = bridge.launcher;
|
exports.launcher = bridge.launcher;
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/next",
|
"name": "@now/next",
|
||||||
"version": "0.0.85-canary.1",
|
"version": "0.0.85-canary.8",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
@@ -8,8 +8,9 @@
|
|||||||
"directory": "packages/now-next"
|
"directory": "packages/now-next"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@now/node-bridge": "0.1.4",
|
"@now/node-bridge": "1.0.0-canary.2",
|
||||||
"execa": "^1.0.0",
|
"execa": "^1.0.0",
|
||||||
|
"fs-extra": "^7.0.0",
|
||||||
"fs.promised": "^3.0.0",
|
"fs.promised": "^3.0.0",
|
||||||
"semver": "^5.6.0"
|
"semver": "^5.6.0"
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,5 +1,3 @@
|
|||||||
const rename = require('@now/build-utils/fs/rename.js'); // eslint-disable-line import/no-extraneous-dependencies
|
|
||||||
|
|
||||||
/** @typedef { import('@now/build-utils/file-ref') } FileRef */
|
/** @typedef { import('@now/build-utils/file-ref') } FileRef */
|
||||||
/** @typedef { import('@now/build-utils/file-fs-ref') } FileFsRef */
|
/** @typedef { import('@now/build-utils/file-fs-ref') } FileFsRef */
|
||||||
/** @typedef {{[filePath: string]: FileRef|FileFsRef}} Files */
|
/** @typedef {{[filePath: string]: FileRef|FileFsRef}} Files */
|
||||||
@@ -64,24 +62,6 @@ function includeOnlyEntryDirectory(files, entryDirectory) {
|
|||||||
return excludeFiles(files, matcher);
|
return excludeFiles(files, matcher);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Moves all files under the entry directory to the root directory
|
|
||||||
* @param {Files} files
|
|
||||||
* @param {string} entryDirectory
|
|
||||||
* @returns {Files}
|
|
||||||
*/
|
|
||||||
function moveEntryDirectoryToRoot(files, entryDirectory) {
|
|
||||||
if (entryDirectory === '.') {
|
|
||||||
return files;
|
|
||||||
}
|
|
||||||
|
|
||||||
function delegate(filePath) {
|
|
||||||
return filePath.replace(new RegExp(`^${entryDirectory}/`), '');
|
|
||||||
}
|
|
||||||
|
|
||||||
return rename(files, delegate);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Exclude package manager lockfiles from files
|
* Exclude package manager lockfiles from files
|
||||||
* @param {Files} files
|
* @param {Files} files
|
||||||
@@ -98,19 +78,6 @@ function excludeLockFiles(files) {
|
|||||||
return files;
|
return files;
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* Exclude the static directory from files
|
|
||||||
* @param {Files} files
|
|
||||||
* @returns {Files}
|
|
||||||
*/
|
|
||||||
function excludeStaticDirectory(files) {
|
|
||||||
function matcher(filePath) {
|
|
||||||
return filePath.startsWith('static');
|
|
||||||
}
|
|
||||||
|
|
||||||
return excludeFiles(files, matcher);
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Exclude the static directory from files
|
* Exclude the static directory from files
|
||||||
* @param {Files} files
|
* @param {Files} files
|
||||||
@@ -173,9 +140,7 @@ module.exports = {
|
|||||||
excludeFiles,
|
excludeFiles,
|
||||||
validateEntrypoint,
|
validateEntrypoint,
|
||||||
includeOnlyEntryDirectory,
|
includeOnlyEntryDirectory,
|
||||||
moveEntryDirectoryToRoot,
|
|
||||||
excludeLockFiles,
|
excludeLockFiles,
|
||||||
normalizePackageJson,
|
normalizePackageJson,
|
||||||
excludeStaticDirectory,
|
|
||||||
onlyStaticDirectory,
|
onlyStaticDirectory,
|
||||||
};
|
};
|
||||||
|
|||||||
24
packages/now-node-bridge/.eslintrc
Normal file
24
packages/now-node-bridge/.eslintrc
Normal file
@@ -0,0 +1,24 @@
|
|||||||
|
{
|
||||||
|
"extends": ["prettier", "airbnb-base"],
|
||||||
|
"rules": {
|
||||||
|
"no-console": 0,
|
||||||
|
"import/no-unresolved": 0,
|
||||||
|
"import/no-dynamic-require": 0,
|
||||||
|
"global-require": 0
|
||||||
|
},
|
||||||
|
"overrides": [
|
||||||
|
{
|
||||||
|
"files": ["test/**"],
|
||||||
|
"rules": {
|
||||||
|
"import/no-extraneous-dependencies": 0
|
||||||
|
},
|
||||||
|
"globals": {
|
||||||
|
"describe": true,
|
||||||
|
"it": true,
|
||||||
|
"test": true,
|
||||||
|
"expect": true
|
||||||
|
}
|
||||||
|
}
|
||||||
|
]
|
||||||
|
}
|
||||||
|
|
||||||
1
packages/now-node-bridge/.gitignore
vendored
Normal file
1
packages/now-node-bridge/.gitignore
vendored
Normal file
@@ -0,0 +1 @@
|
|||||||
|
/bridge.*
|
||||||
@@ -1,110 +0,0 @@
|
|||||||
const http = require('http');
|
|
||||||
|
|
||||||
function normalizeEvent(event) {
|
|
||||||
let isApiGateway = true;
|
|
||||||
|
|
||||||
if (event.Action === 'Invoke') {
|
|
||||||
isApiGateway = false;
|
|
||||||
const invokeEvent = JSON.parse(event.body);
|
|
||||||
|
|
||||||
const {
|
|
||||||
method, path, headers, encoding,
|
|
||||||
} = invokeEvent;
|
|
||||||
|
|
||||||
let { body } = invokeEvent;
|
|
||||||
|
|
||||||
if (body) {
|
|
||||||
if (encoding === 'base64') {
|
|
||||||
body = Buffer.from(body, encoding);
|
|
||||||
} else if (encoding === undefined) {
|
|
||||||
body = Buffer.from(body);
|
|
||||||
} else {
|
|
||||||
throw new Error(`Unsupported encoding: ${encoding}`);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
return {
|
|
||||||
isApiGateway, method, path, headers, body,
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
const {
|
|
||||||
httpMethod: method, path, headers, body,
|
|
||||||
} = event;
|
|
||||||
|
|
||||||
return {
|
|
||||||
isApiGateway, method, path, headers, body,
|
|
||||||
};
|
|
||||||
}
|
|
||||||
|
|
||||||
class Bridge {
|
|
||||||
constructor() {
|
|
||||||
this.launcher = this.launcher.bind(this);
|
|
||||||
}
|
|
||||||
|
|
||||||
launcher(event) {
|
|
||||||
// eslint-disable-next-line consistent-return
|
|
||||||
return new Promise((resolve, reject) => {
|
|
||||||
if (this.userError) {
|
|
||||||
console.error('Error while initializing entrypoint:', this.userError);
|
|
||||||
return resolve({ statusCode: 500, body: '' });
|
|
||||||
}
|
|
||||||
|
|
||||||
if (!this.port) {
|
|
||||||
return resolve({ statusCode: 504, body: '' });
|
|
||||||
}
|
|
||||||
|
|
||||||
const {
|
|
||||||
isApiGateway, method, path, headers, body,
|
|
||||||
} = normalizeEvent(event);
|
|
||||||
|
|
||||||
const opts = {
|
|
||||||
hostname: '127.0.0.1',
|
|
||||||
port: this.port,
|
|
||||||
path,
|
|
||||||
method,
|
|
||||||
headers,
|
|
||||||
};
|
|
||||||
|
|
||||||
const req = http.request(opts, (res) => {
|
|
||||||
const response = res;
|
|
||||||
const respBodyChunks = [];
|
|
||||||
response.on('data', chunk => respBodyChunks.push(Buffer.from(chunk)));
|
|
||||||
response.on('error', reject);
|
|
||||||
response.on('end', () => {
|
|
||||||
const bodyBuffer = Buffer.concat(respBodyChunks);
|
|
||||||
delete response.headers.connection;
|
|
||||||
|
|
||||||
if (isApiGateway) {
|
|
||||||
delete response.headers['content-length'];
|
|
||||||
} else
|
|
||||||
if (response.headers['content-length']) {
|
|
||||||
response.headers['content-length'] = bodyBuffer.length;
|
|
||||||
}
|
|
||||||
|
|
||||||
resolve({
|
|
||||||
statusCode: response.statusCode,
|
|
||||||
headers: response.headers,
|
|
||||||
body: bodyBuffer.toString('base64'),
|
|
||||||
encoding: 'base64',
|
|
||||||
});
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
req.on('error', (error) => {
|
|
||||||
setTimeout(() => {
|
|
||||||
// this lets express print the true error of why the connection was closed.
|
|
||||||
// it is probably 'Cannot set headers after they are sent to the client'
|
|
||||||
reject(error);
|
|
||||||
}, 2);
|
|
||||||
});
|
|
||||||
|
|
||||||
if (body) req.write(body);
|
|
||||||
req.end();
|
|
||||||
});
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = {
|
|
||||||
Bridge,
|
|
||||||
};
|
|
||||||
@@ -1,10 +1,26 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/node-bridge",
|
"name": "@now/node-bridge",
|
||||||
"version": "0.1.11-canary.0",
|
"version": "1.0.0-canary.2",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
|
"main": "./index.js",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
"url": "https://github.com/zeit/now-builders.git",
|
"url": "https://github.com/zeit/now-builders.git",
|
||||||
"directory": "packages/now-node-bridge"
|
"directory": "packages/now-node-bridge"
|
||||||
|
},
|
||||||
|
"files": [
|
||||||
|
"bridge.*",
|
||||||
|
"index.js"
|
||||||
|
],
|
||||||
|
"scripts": {
|
||||||
|
"build": "tsc",
|
||||||
|
"test": "npm run build && jest",
|
||||||
|
"prepublish": "npm run build"
|
||||||
|
},
|
||||||
|
"devDependencies": {
|
||||||
|
"@types/aws-lambda": "8.10.19",
|
||||||
|
"@types/node": "11.9.4",
|
||||||
|
"jest": "24.1.0",
|
||||||
|
"typescript": "3.3.3"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
183
packages/now-node-bridge/src/bridge.ts
Normal file
183
packages/now-node-bridge/src/bridge.ts
Normal file
@@ -0,0 +1,183 @@
|
|||||||
|
import { AddressInfo } from 'net';
|
||||||
|
import { APIGatewayProxyEvent } from 'aws-lambda';
|
||||||
|
import {
|
||||||
|
Server,
|
||||||
|
IncomingHttpHeaders,
|
||||||
|
OutgoingHttpHeaders,
|
||||||
|
request
|
||||||
|
} from 'http';
|
||||||
|
|
||||||
|
interface NowProxyEvent {
|
||||||
|
Action: string;
|
||||||
|
body: string;
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface NowProxyRequest {
|
||||||
|
isApiGateway?: boolean;
|
||||||
|
method: string;
|
||||||
|
path: string;
|
||||||
|
headers: IncomingHttpHeaders;
|
||||||
|
body: Buffer;
|
||||||
|
}
|
||||||
|
|
||||||
|
export interface NowProxyResponse {
|
||||||
|
statusCode: number;
|
||||||
|
headers: OutgoingHttpHeaders;
|
||||||
|
body: string;
|
||||||
|
encoding: string;
|
||||||
|
}
|
||||||
|
|
||||||
|
function normalizeNowProxyEvent(event: NowProxyEvent): NowProxyRequest {
|
||||||
|
let bodyBuffer: Buffer | null;
|
||||||
|
const { method, path, headers, encoding, body } = JSON.parse(event.body);
|
||||||
|
|
||||||
|
if (body) {
|
||||||
|
if (encoding === 'base64') {
|
||||||
|
bodyBuffer = Buffer.from(body, encoding);
|
||||||
|
} else if (encoding === undefined) {
|
||||||
|
bodyBuffer = Buffer.from(body);
|
||||||
|
} else {
|
||||||
|
throw new Error(`Unsupported encoding: ${encoding}`);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
bodyBuffer = Buffer.alloc(0);
|
||||||
|
}
|
||||||
|
|
||||||
|
return { isApiGateway: false, method, path, headers, body: bodyBuffer };
|
||||||
|
}
|
||||||
|
|
||||||
|
function normalizeAPIGatewayProxyEvent(
|
||||||
|
event: APIGatewayProxyEvent
|
||||||
|
): NowProxyRequest {
|
||||||
|
let bodyBuffer: Buffer | null;
|
||||||
|
const { httpMethod: method, path, headers, body } = event;
|
||||||
|
|
||||||
|
if (body) {
|
||||||
|
if (event.isBase64Encoded) {
|
||||||
|
bodyBuffer = Buffer.from(body, 'base64');
|
||||||
|
} else {
|
||||||
|
bodyBuffer = Buffer.from(body);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
bodyBuffer = Buffer.alloc(0);
|
||||||
|
}
|
||||||
|
|
||||||
|
return { isApiGateway: true, method, path, headers, body: bodyBuffer };
|
||||||
|
}
|
||||||
|
|
||||||
|
function normalizeEvent(
|
||||||
|
event: NowProxyEvent | APIGatewayProxyEvent
|
||||||
|
): NowProxyRequest {
|
||||||
|
if ('Action' in event) {
|
||||||
|
if (event.Action === 'Invoke') {
|
||||||
|
return normalizeNowProxyEvent(event);
|
||||||
|
} else {
|
||||||
|
throw new Error(`Unexpected event.Action: ${event.Action}`);
|
||||||
|
}
|
||||||
|
} else {
|
||||||
|
return normalizeAPIGatewayProxyEvent(event);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
export class Bridge {
|
||||||
|
private server: Server | null;
|
||||||
|
private listening: Promise<AddressInfo>;
|
||||||
|
private resolveListening: (info: AddressInfo) => void;
|
||||||
|
|
||||||
|
constructor(server?: Server) {
|
||||||
|
this.server = null;
|
||||||
|
if (server) {
|
||||||
|
this.setServer(server);
|
||||||
|
}
|
||||||
|
this.launcher = this.launcher.bind(this);
|
||||||
|
|
||||||
|
// This is just to appease TypeScript strict mode, since it doesn't
|
||||||
|
// understand that the Promise constructor is synchronous
|
||||||
|
this.resolveListening = (info: AddressInfo) => {};
|
||||||
|
|
||||||
|
this.listening = new Promise(resolve => {
|
||||||
|
this.resolveListening = resolve;
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
setServer(server: Server) {
|
||||||
|
this.server = server;
|
||||||
|
server.once('listening', () => {
|
||||||
|
const addr = server.address();
|
||||||
|
if (typeof addr === 'string') {
|
||||||
|
throw new Error(`Unexpected string for \`server.address()\`: ${addr}`);
|
||||||
|
} else if (!addr) {
|
||||||
|
throw new Error('`server.address()` returned `null`');
|
||||||
|
} else {
|
||||||
|
this.resolveListening(addr);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
listen() {
|
||||||
|
if (!this.server) {
|
||||||
|
throw new Error('Server has not been set!');
|
||||||
|
}
|
||||||
|
|
||||||
|
return this.server.listen({
|
||||||
|
host: '127.0.0.1',
|
||||||
|
port: 0
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
async launcher(
|
||||||
|
event: NowProxyEvent | APIGatewayProxyEvent
|
||||||
|
): Promise<NowProxyResponse> {
|
||||||
|
const { port } = await this.listening;
|
||||||
|
|
||||||
|
const { isApiGateway, method, path, headers, body } = normalizeEvent(
|
||||||
|
event
|
||||||
|
);
|
||||||
|
|
||||||
|
const opts = {
|
||||||
|
hostname: '127.0.0.1',
|
||||||
|
port,
|
||||||
|
path,
|
||||||
|
method,
|
||||||
|
headers
|
||||||
|
};
|
||||||
|
|
||||||
|
// eslint-disable-next-line consistent-return
|
||||||
|
return new Promise((resolve, reject) => {
|
||||||
|
const req = request(opts, res => {
|
||||||
|
const response = res;
|
||||||
|
const respBodyChunks: Buffer[] = [];
|
||||||
|
response.on('data', chunk => respBodyChunks.push(Buffer.from(chunk)));
|
||||||
|
response.on('error', reject);
|
||||||
|
response.on('end', () => {
|
||||||
|
const bodyBuffer = Buffer.concat(respBodyChunks);
|
||||||
|
delete response.headers.connection;
|
||||||
|
|
||||||
|
if (isApiGateway) {
|
||||||
|
delete response.headers['content-length'];
|
||||||
|
} else if (response.headers['content-length']) {
|
||||||
|
response.headers['content-length'] = String(bodyBuffer.length);
|
||||||
|
}
|
||||||
|
|
||||||
|
resolve({
|
||||||
|
statusCode: response.statusCode || 200,
|
||||||
|
headers: response.headers,
|
||||||
|
body: bodyBuffer.toString('base64'),
|
||||||
|
encoding: 'base64'
|
||||||
|
});
|
||||||
|
});
|
||||||
|
});
|
||||||
|
|
||||||
|
req.on('error', error => {
|
||||||
|
setTimeout(() => {
|
||||||
|
// this lets express print the true error of why the connection was closed.
|
||||||
|
// it is probably 'Cannot set headers after they are sent to the client'
|
||||||
|
reject(error);
|
||||||
|
}, 2);
|
||||||
|
});
|
||||||
|
|
||||||
|
if (body) req.write(body);
|
||||||
|
req.end();
|
||||||
|
});
|
||||||
|
}
|
||||||
|
}
|
||||||
71
packages/now-node-bridge/test/bridge.test.js
Normal file
71
packages/now-node-bridge/test/bridge.test.js
Normal file
@@ -0,0 +1,71 @@
|
|||||||
|
const assert = require('assert');
|
||||||
|
const { Server } = require('http');
|
||||||
|
const { Bridge } = require('../bridge');
|
||||||
|
|
||||||
|
test('port binding', async () => {
|
||||||
|
const server = new Server();
|
||||||
|
const bridge = new Bridge(server);
|
||||||
|
bridge.listen();
|
||||||
|
|
||||||
|
// Test port binding
|
||||||
|
const info = await bridge.listening;
|
||||||
|
assert.equal(info.address, '127.0.0.1');
|
||||||
|
assert.equal(typeof info.port, 'number');
|
||||||
|
|
||||||
|
server.close();
|
||||||
|
});
|
||||||
|
|
||||||
|
test('`APIGatewayProxyEvent` normalizing', async () => {
|
||||||
|
const server = new Server((req, res) => res.end(
|
||||||
|
JSON.stringify({
|
||||||
|
method: req.method,
|
||||||
|
path: req.url,
|
||||||
|
headers: req.headers,
|
||||||
|
}),
|
||||||
|
));
|
||||||
|
const bridge = new Bridge(server);
|
||||||
|
bridge.listen();
|
||||||
|
const result = await bridge.launcher({
|
||||||
|
httpMethod: 'GET',
|
||||||
|
headers: { foo: 'bar' },
|
||||||
|
path: '/apigateway',
|
||||||
|
body: null,
|
||||||
|
});
|
||||||
|
assert.equal(result.encoding, 'base64');
|
||||||
|
assert.equal(result.statusCode, 200);
|
||||||
|
const body = JSON.parse(Buffer.from(result.body, 'base64').toString());
|
||||||
|
assert.equal(body.method, 'GET');
|
||||||
|
assert.equal(body.path, '/apigateway');
|
||||||
|
assert.equal(body.headers.foo, 'bar');
|
||||||
|
|
||||||
|
server.close();
|
||||||
|
});
|
||||||
|
|
||||||
|
test('`NowProxyEvent` normalizing', async () => {
|
||||||
|
const server = new Server((req, res) => res.end(
|
||||||
|
JSON.stringify({
|
||||||
|
method: req.method,
|
||||||
|
path: req.url,
|
||||||
|
headers: req.headers,
|
||||||
|
}),
|
||||||
|
));
|
||||||
|
const bridge = new Bridge(server);
|
||||||
|
bridge.listen();
|
||||||
|
const result = await bridge.launcher({
|
||||||
|
Action: 'Invoke',
|
||||||
|
body: JSON.stringify({
|
||||||
|
method: 'POST',
|
||||||
|
headers: { foo: 'baz' },
|
||||||
|
path: '/nowproxy',
|
||||||
|
body: 'body=1',
|
||||||
|
}),
|
||||||
|
});
|
||||||
|
assert.equal(result.encoding, 'base64');
|
||||||
|
assert.equal(result.statusCode, 200);
|
||||||
|
const body = JSON.parse(Buffer.from(result.body, 'base64').toString());
|
||||||
|
assert.equal(body.method, 'POST');
|
||||||
|
assert.equal(body.path, '/nowproxy');
|
||||||
|
assert.equal(body.headers.foo, 'baz');
|
||||||
|
|
||||||
|
server.close();
|
||||||
|
});
|
||||||
16
packages/now-node-bridge/tsconfig.json
Normal file
16
packages/now-node-bridge/tsconfig.json
Normal file
@@ -0,0 +1,16 @@
|
|||||||
|
{
|
||||||
|
"compilerOptions": {
|
||||||
|
"target": "es6",
|
||||||
|
"module": "commonjs",
|
||||||
|
"outDir": ".",
|
||||||
|
"strict": true,
|
||||||
|
"sourceMap": true,
|
||||||
|
"declaration": true
|
||||||
|
},
|
||||||
|
"include": [
|
||||||
|
"src/**/*"
|
||||||
|
],
|
||||||
|
"exclude": [
|
||||||
|
"node_modules"
|
||||||
|
]
|
||||||
|
}
|
||||||
@@ -48,7 +48,7 @@ async function downloadInstallAndBundle(
|
|||||||
data: JSON.stringify({
|
data: JSON.stringify({
|
||||||
license: 'UNLICENSED',
|
license: 'UNLICENSED',
|
||||||
dependencies: {
|
dependencies: {
|
||||||
'@zeit/ncc': '0.13.2',
|
'@zeit/ncc': '0.15.2',
|
||||||
},
|
},
|
||||||
}),
|
}),
|
||||||
}),
|
}),
|
||||||
@@ -64,7 +64,7 @@ async function downloadInstallAndBundle(
|
|||||||
async function compile(workNccPath, downloadedFiles, entrypoint) {
|
async function compile(workNccPath, downloadedFiles, entrypoint) {
|
||||||
const input = downloadedFiles[entrypoint].fsPath;
|
const input = downloadedFiles[entrypoint].fsPath;
|
||||||
const ncc = require(path.join(workNccPath, 'node_modules/@zeit/ncc'));
|
const ncc = require(path.join(workNccPath, 'node_modules/@zeit/ncc'));
|
||||||
const { code, assets } = await ncc(input);
|
const { code, assets } = await ncc(input, { sourceMap: true });
|
||||||
|
|
||||||
const preparedFiles = {};
|
const preparedFiles = {};
|
||||||
const blob = new FileBlob({ data: code });
|
const blob = new FileBlob({ data: code });
|
||||||
|
|||||||
@@ -4,22 +4,16 @@ const { Bridge } = require('./bridge.js');
|
|||||||
const bridge = new Bridge();
|
const bridge = new Bridge();
|
||||||
|
|
||||||
const saveListen = Server.prototype.listen;
|
const saveListen = Server.prototype.listen;
|
||||||
Server.prototype.listen = function listen(...args) {
|
Server.prototype.listen = function listen() {
|
||||||
this.on('listening', function listening() {
|
bridge.setServer(this);
|
||||||
bridge.port = this.address().port;
|
Server.prototype.listen = saveListen;
|
||||||
});
|
return bridge.listen();
|
||||||
saveListen.apply(this, args);
|
|
||||||
};
|
};
|
||||||
|
|
||||||
try {
|
if (!process.env.NODE_ENV) {
|
||||||
if (!process.env.NODE_ENV) {
|
|
||||||
process.env.NODE_ENV = 'production';
|
process.env.NODE_ENV = 'production';
|
||||||
}
|
|
||||||
|
|
||||||
// PLACEHOLDER
|
|
||||||
} catch (error) {
|
|
||||||
console.error(error);
|
|
||||||
bridge.userError = error;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
|
// PLACEHOLDER
|
||||||
|
|
||||||
exports.launcher = bridge.launcher;
|
exports.launcher = bridge.launcher;
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/node-server",
|
"name": "@now/node-server",
|
||||||
"version": "0.4.27-canary.2",
|
"version": "0.5.0-canary.3",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
@@ -8,7 +8,7 @@
|
|||||||
"directory": "packages/now-node-server"
|
"directory": "packages/now-node-server"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@now/node-bridge": "^0.1.11-canary.0",
|
"@now/node-bridge": "1.0.0-canary.2",
|
||||||
"fs-extra": "7.0.1"
|
"fs-extra": "7.0.1"
|
||||||
},
|
},
|
||||||
"scripts": {
|
"scripts": {
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ const {
|
|||||||
testDeployment,
|
testDeployment,
|
||||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||||
|
|
||||||
jest.setTimeout(2 * 60 * 1000);
|
jest.setTimeout(4 * 60 * 1000);
|
||||||
const buildUtilsUrl = '@canary';
|
const buildUtilsUrl = '@canary';
|
||||||
let builderUrl;
|
let builderUrl;
|
||||||
|
|
||||||
|
|||||||
2
packages/now-node/.gitignore
vendored
Normal file
2
packages/now-node/.gitignore
vendored
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
/dist
|
||||||
|
/src/bridge.d.ts
|
||||||
@@ -1 +0,0 @@
|
|||||||
/test
|
|
||||||
12
packages/now-node/build.sh
Executable file
12
packages/now-node/build.sh
Executable file
@@ -0,0 +1,12 @@
|
|||||||
|
#!/bin/bash
|
||||||
|
set -euo pipefail
|
||||||
|
|
||||||
|
bridge_entrypoint="$(node -p 'require.resolve("@now/node-bridge")')"
|
||||||
|
bridge_defs="$(dirname "$bridge_entrypoint")/bridge.d.ts"
|
||||||
|
|
||||||
|
if [ ! -e "$bridge_defs" ]; then
|
||||||
|
yarn install
|
||||||
|
fi
|
||||||
|
|
||||||
|
cp -v "$bridge_defs" src
|
||||||
|
tsc
|
||||||
@@ -1,22 +0,0 @@
|
|||||||
const { Server } = require('http');
|
|
||||||
const { Bridge } = require('./bridge.js');
|
|
||||||
|
|
||||||
const bridge = new Bridge();
|
|
||||||
bridge.port = 3000;
|
|
||||||
let listener;
|
|
||||||
|
|
||||||
try {
|
|
||||||
if (!process.env.NODE_ENV) {
|
|
||||||
process.env.NODE_ENV = 'production';
|
|
||||||
}
|
|
||||||
|
|
||||||
// PLACEHOLDER
|
|
||||||
} catch (error) {
|
|
||||||
console.error(error);
|
|
||||||
bridge.userError = error;
|
|
||||||
}
|
|
||||||
|
|
||||||
const server = new Server(listener);
|
|
||||||
server.listen(bridge.port);
|
|
||||||
|
|
||||||
exports.launcher = bridge.launcher;
|
|
||||||
@@ -1,17 +1,27 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/node",
|
"name": "@now/node",
|
||||||
"version": "0.4.29-canary.2",
|
"version": "0.5.0-canary.5",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
|
"main": "./dist/index",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
"url": "https://github.com/zeit/now-builders.git",
|
"url": "https://github.com/zeit/now-builders.git",
|
||||||
"directory": "packages/now-node"
|
"directory": "packages/now-node"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"@now/node-bridge": "^0.1.11-canary.0",
|
"@now/node-bridge": "1.0.0-canary.2",
|
||||||
"fs-extra": "7.0.1"
|
"fs-extra": "7.0.1"
|
||||||
},
|
},
|
||||||
"scripts": {
|
"scripts": {
|
||||||
|
"build": "./build.sh",
|
||||||
"test": "jest"
|
"test": "jest"
|
||||||
|
},
|
||||||
|
"files": [
|
||||||
|
"dist"
|
||||||
|
],
|
||||||
|
"devDependencies": {
|
||||||
|
"@types/node": "11.9.4",
|
||||||
|
"jest": "24.1.0",
|
||||||
|
"typescript": "3.3.3"
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,14 +1,14 @@
|
|||||||
const { createLambda } = require('@now/build-utils/lambda.js'); // eslint-disable-line import/no-extraneous-dependencies
|
import { join, dirname } from 'path';
|
||||||
const download = require('@now/build-utils/fs/download.js'); // eslint-disable-line import/no-extraneous-dependencies
|
import { remove, readFile } from 'fs-extra';
|
||||||
const FileBlob = require('@now/build-utils/file-blob.js'); // eslint-disable-line import/no-extraneous-dependencies
|
import * as glob from '@now/build-utils/fs/glob.js';
|
||||||
const FileFsRef = require('@now/build-utils/file-fs-ref.js'); // eslint-disable-line import/no-extraneous-dependencies
|
import * as download from '@now/build-utils/fs/download.js';
|
||||||
const fs = require('fs-extra');
|
import * as FileBlob from '@now/build-utils/file-blob.js';
|
||||||
const glob = require('@now/build-utils/fs/glob.js'); // eslint-disable-line import/no-extraneous-dependencies
|
import * as FileFsRef from '@now/build-utils/file-fs-ref.js';
|
||||||
const path = require('path');
|
import { createLambda } from '@now/build-utils/lambda.js';
|
||||||
const {
|
import {
|
||||||
runNpmInstall,
|
runNpmInstall,
|
||||||
runPackageJsonScript,
|
runPackageJsonScript
|
||||||
} = require('@now/build-utils/fs/run-user-scripts.js'); // eslint-disable-line import/no-extraneous-dependencies
|
} from '@now/build-utils/fs/run-user-scripts.js';
|
||||||
|
|
||||||
/** @typedef { import('@now/build-utils/file-ref') } FileRef */
|
/** @typedef { import('@now/build-utils/file-ref') } FileRef */
|
||||||
/** @typedef {{[filePath: string]: FileRef}} Files */
|
/** @typedef {{[filePath: string]: FileRef}} Files */
|
||||||
@@ -27,16 +27,16 @@ const {
|
|||||||
*/
|
*/
|
||||||
async function downloadInstallAndBundle(
|
async function downloadInstallAndBundle(
|
||||||
{ files, entrypoint, workPath },
|
{ files, entrypoint, workPath },
|
||||||
{ npmArguments = [] } = {},
|
{ npmArguments = [] } = {}
|
||||||
) {
|
) {
|
||||||
const userPath = path.join(workPath, 'user');
|
const userPath = join(workPath, 'user');
|
||||||
const nccPath = path.join(workPath, 'ncc');
|
const nccPath = join(workPath, 'ncc');
|
||||||
|
|
||||||
console.log('downloading user files...');
|
console.log('downloading user files...');
|
||||||
const downloadedFiles = await download(files, userPath);
|
const downloadedFiles = await download(files, userPath);
|
||||||
|
|
||||||
console.log("installing dependencies for user's code...");
|
console.log("installing dependencies for user's code...");
|
||||||
const entrypointFsDirname = path.join(userPath, path.dirname(entrypoint));
|
const entrypointFsDirname = join(userPath, dirname(entrypoint));
|
||||||
await runNpmInstall(entrypointFsDirname, npmArguments);
|
await runNpmInstall(entrypointFsDirname, npmArguments);
|
||||||
|
|
||||||
console.log('writing ncc package.json...');
|
console.log('writing ncc package.json...');
|
||||||
@@ -46,12 +46,12 @@ async function downloadInstallAndBundle(
|
|||||||
data: JSON.stringify({
|
data: JSON.stringify({
|
||||||
license: 'UNLICENSED',
|
license: 'UNLICENSED',
|
||||||
dependencies: {
|
dependencies: {
|
||||||
'@zeit/ncc': '0.13.2',
|
'@zeit/ncc': '0.15.2',
|
||||||
|
}
|
||||||
|
})
|
||||||
|
})
|
||||||
},
|
},
|
||||||
}),
|
nccPath
|
||||||
}),
|
|
||||||
},
|
|
||||||
nccPath,
|
|
||||||
);
|
);
|
||||||
|
|
||||||
console.log('installing dependencies for ncc...');
|
console.log('installing dependencies for ncc...');
|
||||||
@@ -59,43 +59,41 @@ async function downloadInstallAndBundle(
|
|||||||
return [downloadedFiles, nccPath, entrypointFsDirname];
|
return [downloadedFiles, nccPath, entrypointFsDirname];
|
||||||
}
|
}
|
||||||
|
|
||||||
async function compile(workNccPath, downloadedFiles, entrypoint) {
|
async function compile(workNccPath: string, downloadedFiles, entrypoint: string) {
|
||||||
const input = downloadedFiles[entrypoint].fsPath;
|
const input = downloadedFiles[entrypoint].fsPath;
|
||||||
const ncc = require(path.join(workNccPath, 'node_modules/@zeit/ncc'));
|
const ncc = require(join(workNccPath, 'node_modules/@zeit/ncc'));
|
||||||
const { code, assets } = await ncc(input);
|
const { code, assets } = await ncc(input);
|
||||||
|
|
||||||
const preparedFiles = {};
|
const preparedFiles = {};
|
||||||
const blob = new FileBlob({ data: code });
|
const blob = new FileBlob({ data: code });
|
||||||
// move all user code to 'user' subdirectory
|
// move all user code to 'user' subdirectory
|
||||||
preparedFiles[path.join('user', entrypoint)] = blob;
|
preparedFiles[join('user', entrypoint)] = blob;
|
||||||
// eslint-disable-next-line no-restricted-syntax
|
// eslint-disable-next-line no-restricted-syntax
|
||||||
for (const assetName of Object.keys(assets)) {
|
for (const assetName of Object.keys(assets)) {
|
||||||
const { source: data, permissions: mode } = assets[assetName];
|
const { source: data, permissions: mode } = assets[assetName];
|
||||||
const blob2 = new FileBlob({ data, mode });
|
const blob2 = new FileBlob({ data, mode });
|
||||||
preparedFiles[
|
preparedFiles[join('user', dirname(entrypoint), assetName)] = blob2;
|
||||||
path.join('user', path.dirname(entrypoint), assetName)
|
|
||||||
] = blob2;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
return preparedFiles;
|
return preparedFiles;
|
||||||
}
|
}
|
||||||
|
|
||||||
exports.config = {
|
export const config = {
|
||||||
maxLambdaSize: '5mb',
|
maxLambdaSize: '5mb'
|
||||||
};
|
};
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* @param {BuildParamsType} buildParams
|
* @param {BuildParamsType} buildParams
|
||||||
* @returns {Promise<Files>}
|
* @returns {Promise<Files>}
|
||||||
*/
|
*/
|
||||||
exports.build = async ({ files, entrypoint, workPath }) => {
|
export async function build({ files, entrypoint, workPath }) {
|
||||||
const [
|
const [
|
||||||
downloadedFiles,
|
downloadedFiles,
|
||||||
workNccPath,
|
workNccPath,
|
||||||
entrypointFsDirname,
|
entrypointFsDirname
|
||||||
] = await downloadInstallAndBundle(
|
] = await downloadInstallAndBundle(
|
||||||
{ files, entrypoint, workPath },
|
{ files, entrypoint, workPath },
|
||||||
{ npmArguments: ['--prefer-offline'] },
|
{ npmArguments: ['--prefer-offline'] }
|
||||||
);
|
);
|
||||||
|
|
||||||
console.log('running user script...');
|
console.log('running user script...');
|
||||||
@@ -103,36 +101,34 @@ exports.build = async ({ files, entrypoint, workPath }) => {
|
|||||||
|
|
||||||
console.log('compiling entrypoint with ncc...');
|
console.log('compiling entrypoint with ncc...');
|
||||||
const preparedFiles = await compile(workNccPath, downloadedFiles, entrypoint);
|
const preparedFiles = await compile(workNccPath, downloadedFiles, entrypoint);
|
||||||
const launcherPath = path.join(__dirname, 'launcher.js');
|
const launcherPath = join(__dirname, 'launcher.js');
|
||||||
let launcherData = await fs.readFile(launcherPath, 'utf8');
|
let launcherData = await readFile(launcherPath, 'utf8');
|
||||||
|
|
||||||
launcherData = launcherData.replace(
|
launcherData = launcherData.replace(
|
||||||
'// PLACEHOLDER',
|
'// PLACEHOLDER',
|
||||||
[
|
[
|
||||||
'process.chdir("./user");',
|
'process.chdir("./user");',
|
||||||
`listener = require("./${path.join('user', entrypoint)}");`,
|
`listener = require("./${join('user', entrypoint)}");`,
|
||||||
'if (listener.default) listener = listener.default;',
|
'if (listener.default) listener = listener.default;'
|
||||||
].join(' '),
|
].join(' ')
|
||||||
);
|
);
|
||||||
|
|
||||||
const launcherFiles = {
|
const launcherFiles = {
|
||||||
'launcher.js': new FileBlob({ data: launcherData }),
|
'launcher.js': new FileBlob({ data: launcherData }),
|
||||||
'bridge.js': new FileFsRef({ fsPath: require('@now/node-bridge') }),
|
'bridge.js': new FileFsRef({ fsPath: require('@now/node-bridge') })
|
||||||
};
|
};
|
||||||
|
|
||||||
const lambda = await createLambda({
|
const lambda = await createLambda({
|
||||||
files: { ...preparedFiles, ...launcherFiles },
|
files: { ...preparedFiles, ...launcherFiles },
|
||||||
handler: 'launcher.launcher',
|
handler: 'launcher.launcher',
|
||||||
runtime: 'nodejs8.10',
|
runtime: 'nodejs8.10'
|
||||||
});
|
});
|
||||||
|
|
||||||
return { [entrypoint]: lambda };
|
return { [entrypoint]: lambda };
|
||||||
};
|
}
|
||||||
|
|
||||||
exports.prepareCache = async ({
|
export async function prepareCache({ files, entrypoint, workPath, cachePath }) {
|
||||||
files, entrypoint, workPath, cachePath,
|
await remove(workPath);
|
||||||
}) => {
|
|
||||||
await fs.remove(workPath);
|
|
||||||
await downloadInstallAndBundle({ files, entrypoint, workPath: cachePath });
|
await downloadInstallAndBundle({ files, entrypoint, workPath: cachePath });
|
||||||
|
|
||||||
return {
|
return {
|
||||||
@@ -141,6 +137,6 @@ exports.prepareCache = async ({
|
|||||||
...(await glob('user/yarn.lock', cachePath)),
|
...(await glob('user/yarn.lock', cachePath)),
|
||||||
...(await glob('ncc/node_modules/**', cachePath)),
|
...(await glob('ncc/node_modules/**', cachePath)),
|
||||||
...(await glob('ncc/package-lock.json', cachePath)),
|
...(await glob('ncc/package-lock.json', cachePath)),
|
||||||
...(await glob('ncc/yarn.lock', cachePath)),
|
...(await glob('ncc/yarn.lock', cachePath))
|
||||||
};
|
};
|
||||||
};
|
}
|
||||||
16
packages/now-node/src/launcher.ts
Normal file
16
packages/now-node/src/launcher.ts
Normal file
@@ -0,0 +1,16 @@
|
|||||||
|
import { Server } from 'http';
|
||||||
|
import { Bridge } from './bridge';
|
||||||
|
|
||||||
|
let listener;
|
||||||
|
|
||||||
|
if (!process.env.NODE_ENV) {
|
||||||
|
process.env.NODE_ENV = 'production';
|
||||||
|
}
|
||||||
|
|
||||||
|
// PLACEHOLDER
|
||||||
|
|
||||||
|
const server = new Server(listener);
|
||||||
|
const bridge = new Bridge(server);
|
||||||
|
bridge.listen();
|
||||||
|
|
||||||
|
exports.launcher = bridge.launcher;
|
||||||
@@ -7,7 +7,7 @@ const {
|
|||||||
testDeployment,
|
testDeployment,
|
||||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||||
|
|
||||||
jest.setTimeout(2 * 60 * 1000);
|
jest.setTimeout(4 * 60 * 1000);
|
||||||
const buildUtilsUrl = '@canary';
|
const buildUtilsUrl = '@canary';
|
||||||
let builderUrl;
|
let builderUrl;
|
||||||
|
|
||||||
|
|||||||
15
packages/now-node/tsconfig.json
Normal file
15
packages/now-node/tsconfig.json
Normal file
@@ -0,0 +1,15 @@
|
|||||||
|
{
|
||||||
|
"compilerOptions": {
|
||||||
|
"target": "es6",
|
||||||
|
"module": "commonjs",
|
||||||
|
"outDir": "dist",
|
||||||
|
"sourceMap": false,
|
||||||
|
"declaration": false
|
||||||
|
},
|
||||||
|
"include": [
|
||||||
|
"src/**/*"
|
||||||
|
],
|
||||||
|
"exclude": [
|
||||||
|
"node_modules"
|
||||||
|
]
|
||||||
|
}
|
||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/php",
|
"name": "@now/php",
|
||||||
"version": "0.4.14-canary.0",
|
"version": "0.4.14-canary.1",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
|
|||||||
@@ -7,7 +7,7 @@ const {
|
|||||||
testDeployment,
|
testDeployment,
|
||||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||||
|
|
||||||
jest.setTimeout(2 * 60 * 1000);
|
jest.setTimeout(4 * 60 * 1000);
|
||||||
const buildUtilsUrl = '@canary';
|
const buildUtilsUrl = '@canary';
|
||||||
let builderUrl;
|
let builderUrl;
|
||||||
|
|
||||||
|
|||||||
@@ -40,8 +40,16 @@ exports.build = async ({ files, entrypoint }) => {
|
|||||||
|
|
||||||
await pipInstall(pipPath, srcDir, 'requests');
|
await pipInstall(pipPath, srcDir, 'requests');
|
||||||
|
|
||||||
if (files['requirements.txt']) {
|
const entryDirectory = path.dirname(entrypoint);
|
||||||
console.log('found "requirements.txt"');
|
const requirementsTxt = path.join(entryDirectory, 'requirements.txt');
|
||||||
|
|
||||||
|
if (files[requirementsTxt]) {
|
||||||
|
console.log('found local "requirements.txt"');
|
||||||
|
|
||||||
|
const requirementsTxtPath = files[requirementsTxt].fsPath;
|
||||||
|
await pipInstall(pipPath, srcDir, '-r', requirementsTxtPath);
|
||||||
|
} else if (files['requirements.txt']) {
|
||||||
|
console.log('found global "requirements.txt"');
|
||||||
|
|
||||||
const requirementsTxtPath = files['requirements.txt'].fsPath;
|
const requirementsTxtPath = files['requirements.txt'].fsPath;
|
||||||
await pipInstall(pipPath, srcDir, '-r', requirementsTxtPath);
|
await pipInstall(pipPath, srcDir, '-r', requirementsTxtPath);
|
||||||
|
|||||||
@@ -1,23 +1,35 @@
|
|||||||
|
import base64
|
||||||
from http.server import HTTPServer
|
from http.server import HTTPServer
|
||||||
import json
|
import json
|
||||||
import requests
|
import requests
|
||||||
from __NOW_HANDLER_FILENAME import handler
|
from __NOW_HANDLER_FILENAME import handler
|
||||||
import _thread
|
import _thread
|
||||||
|
|
||||||
|
|
||||||
server = HTTPServer(('', 3000), handler)
|
server = HTTPServer(('', 3000), handler)
|
||||||
|
|
||||||
|
|
||||||
def now_handler(event, context):
|
def now_handler(event, context):
|
||||||
_thread.start_new_thread(server.handle_request, ())
|
_thread.start_new_thread(server.handle_request, ())
|
||||||
|
|
||||||
payload = json.loads(event['body'])
|
payload = json.loads(event['body'])
|
||||||
path = payload['path']
|
path = payload['path']
|
||||||
headers = payload['headers']
|
headers = payload['headers']
|
||||||
method = payload['method']
|
method = payload['method']
|
||||||
|
encoding = payload.get('encoding')
|
||||||
|
body = payload.get('body')
|
||||||
|
|
||||||
res = requests.request(method, 'http://0.0.0.0:3000' + path, headers=headers)
|
if (
|
||||||
|
(body is not None and len(body) > 0) and
|
||||||
|
(encoding is not None and encoding == 'base64')
|
||||||
|
):
|
||||||
|
body = base64.b64decode(body)
|
||||||
|
|
||||||
|
res = requests.request(method, 'http://0.0.0.0:3000' + path,
|
||||||
|
headers=headers, data=body, allow_redirects=False)
|
||||||
|
|
||||||
return {
|
return {
|
||||||
'statusCode': res.status_code,
|
'statusCode': res.status_code,
|
||||||
'headers': dict(res.headers),
|
'headers': dict(res.headers),
|
||||||
'body': res.text
|
'body': res.text,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/python",
|
"name": "@now/python",
|
||||||
"version": "0.0.41-canary.0",
|
"version": "0.0.41-canary.2",
|
||||||
"main": "index.js",
|
"main": "index.js",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"repository": {
|
"repository": {
|
||||||
|
|||||||
2
packages/now-rust/Cargo.lock
generated
2
packages/now-rust/Cargo.lock
generated
@@ -448,7 +448,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
|||||||
|
|
||||||
[[package]]
|
[[package]]
|
||||||
name = "now_lambda"
|
name = "now_lambda"
|
||||||
version = "0.1.0"
|
version = "0.1.2"
|
||||||
dependencies = [
|
dependencies = [
|
||||||
"base64 0.10.1 (registry+https://github.com/rust-lang/crates.io-index)",
|
"base64 0.10.1 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
"http 0.1.15 (registry+https://github.com/rust-lang/crates.io-index)",
|
"http 0.1.15 (registry+https://github.com/rust-lang/crates.io-index)",
|
||||||
|
|||||||
@@ -1,8 +1,18 @@
|
|||||||
[package]
|
[package]
|
||||||
name = "now_lambda"
|
name = "now_lambda"
|
||||||
version = "0.1.0"
|
version = "0.1.2"
|
||||||
authors = ["Antonio Nuno Monteiro <anmonteiro@gmail.com>"]
|
authors = ["Antonio Nuno Monteiro <anmonteiro@gmail.com>"]
|
||||||
edition = "2018"
|
edition = "2018"
|
||||||
|
description = "Rust bindings for Now.sh Lambdas"
|
||||||
|
keywords = ["AWS", "Lambda", "Zeit", "Now", "Rust"]
|
||||||
|
license = "MIT"
|
||||||
|
homepage = "https://github.com/zeit/now-builders"
|
||||||
|
repository = "https://github.com/zeit/now-builders"
|
||||||
|
documentation = "https://docs.rs/now_lambda"
|
||||||
|
include = [
|
||||||
|
"src/*.rs",
|
||||||
|
"Cargo.toml"
|
||||||
|
]
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
serde = "^1"
|
serde = "^1"
|
||||||
|
|||||||
@@ -1,5 +1,6 @@
|
|||||||
const tar = require('tar');
|
const tar = require('tar');
|
||||||
const fetch = require('node-fetch');
|
const fetch = require('node-fetch');
|
||||||
|
const execa = require('execa');
|
||||||
|
|
||||||
const rustUrl = 'https://dmmcy0pwk6bqi.cloudfront.net/rust.tar.gz';
|
const rustUrl = 'https://dmmcy0pwk6bqi.cloudfront.net/rust.tar.gz';
|
||||||
const ccUrl = 'https://dmmcy0pwk6bqi.cloudfront.net/gcc-4.8.5.tgz';
|
const ccUrl = 'https://dmmcy0pwk6bqi.cloudfront.net/gcc-4.8.5.tgz';
|
||||||
@@ -53,10 +54,31 @@ async function downloadGCC() {
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
|
async function installOpenSSL() {
|
||||||
|
console.log('installing openssl-devel...');
|
||||||
|
try {
|
||||||
|
// need to downgrade otherwise yum can't resolve the dependencies given
|
||||||
|
// a later version is already installed in the machine.
|
||||||
|
await execa(
|
||||||
|
'yum',
|
||||||
|
['downgrade', '-y', 'krb5-libs-1.14.1-27.41.amzn1.x86_64'],
|
||||||
|
{
|
||||||
|
stdio: 'inherit',
|
||||||
|
},
|
||||||
|
);
|
||||||
|
await execa('yum', ['install', '-y', 'openssl-devel'], {
|
||||||
|
stdio: 'inherit',
|
||||||
|
});
|
||||||
|
} catch (err) {
|
||||||
|
console.error('failed to `yum install -y openssl-devel`');
|
||||||
|
throw err;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
module.exports = async () => {
|
module.exports = async () => {
|
||||||
await downloadRustToolchain();
|
await downloadRustToolchain();
|
||||||
|
|
||||||
const newEnv = await downloadGCC();
|
const newEnv = await downloadGCC();
|
||||||
|
await installOpenSSL();
|
||||||
|
|
||||||
return newEnv;
|
return newEnv;
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -1,66 +1,72 @@
|
|||||||
const fs = require('fs');
|
const fs = require('fs-extra');
|
||||||
const path = require('path');
|
const path = require('path');
|
||||||
const concat = require('concat-stream');
|
|
||||||
const execa = require('execa');
|
const execa = require('execa');
|
||||||
const toml = require('toml');
|
const toml = require('@iarna/toml');
|
||||||
const rimraf = require('rimraf');
|
|
||||||
const { createLambda } = require('@now/build-utils/lambda.js'); // eslint-disable-line import/no-extraneous-dependencies
|
const { createLambda } = require('@now/build-utils/lambda.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
const download = require('@now/build-utils/fs/download.js'); // eslint-disable-line import/no-extraneous-dependencies
|
const download = require('@now/build-utils/fs/download.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
const glob = require('@now/build-utils/fs/glob.js'); // eslint-disable-line import/no-extraneous-dependencies
|
const glob = require('@now/build-utils/fs/glob.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
const FileFsRef = require('@now/build-utils/file-fs-ref.js'); // eslint-disable-line import/no-extraneous-dependencies
|
const FileFsRef = require('@now/build-utils/file-fs-ref.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
const installRustAndGCC = require('./download-install-rust-toolchain.js');
|
const installRustAndGCC = require('./download-install-rust-toolchain.js');
|
||||||
const inferCargoBinaries = require('./inferCargoBinaries.js');
|
|
||||||
|
|
||||||
exports.config = {
|
exports.config = {
|
||||||
maxLambdaSize: '25mb',
|
maxLambdaSize: '25mb',
|
||||||
};
|
};
|
||||||
|
|
||||||
async function parseTOMLStream(stream) {
|
async function inferCargoBinaries(config) {
|
||||||
return new Promise((resolve) => {
|
|
||||||
stream.pipe(concat(data => resolve(toml.parse(data))));
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
exports.build = async ({ files, entrypoint, workPath }) => {
|
|
||||||
console.log('downloading files');
|
|
||||||
const downloadedFiles = await download(files, workPath);
|
|
||||||
|
|
||||||
const { PATH: toolchainPath, ...otherEnv } = await installRustAndGCC();
|
|
||||||
const { PATH, HOME } = process.env;
|
|
||||||
const rustEnv = {
|
|
||||||
...process.env,
|
|
||||||
...otherEnv,
|
|
||||||
PATH: `${path.join(HOME, '.cargo/bin')}:${toolchainPath}:${PATH}`,
|
|
||||||
};
|
|
||||||
|
|
||||||
let cargoToml;
|
|
||||||
try {
|
try {
|
||||||
cargoToml = await parseTOMLStream(files[entrypoint].toStream());
|
const { stdout: manifestStr } = await execa(
|
||||||
|
'cargo',
|
||||||
|
['read-manifest'],
|
||||||
|
config,
|
||||||
|
);
|
||||||
|
|
||||||
|
const { targets } = JSON.parse(manifestStr);
|
||||||
|
|
||||||
|
return targets
|
||||||
|
.filter(({ kind }) => kind.includes('bin'))
|
||||||
|
.map(({ name }) => name);
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
console.error('Failed to parse TOML from entrypoint:', entrypoint);
|
console.error('failed to run `cargo read-manifest`');
|
||||||
throw err;
|
throw err;
|
||||||
}
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async function parseTOMLStream(stream) {
|
||||||
|
return toml.parse.stream(stream);
|
||||||
|
}
|
||||||
|
|
||||||
|
async function buildWholeProject({
|
||||||
|
entrypoint,
|
||||||
|
downloadedFiles,
|
||||||
|
rustEnv,
|
||||||
|
config,
|
||||||
|
}) {
|
||||||
const entrypointDirname = path.dirname(downloadedFiles[entrypoint].fsPath);
|
const entrypointDirname = path.dirname(downloadedFiles[entrypoint].fsPath);
|
||||||
console.log('running `cargo build --release`...');
|
const { debug } = config;
|
||||||
|
console.log('running `cargo build`...');
|
||||||
try {
|
try {
|
||||||
await execa('cargo', ['build', '--release'], {
|
await execa('cargo', ['build'].concat(debug ? [] : ['--release']), {
|
||||||
env: rustEnv,
|
env: rustEnv,
|
||||||
cwd: entrypointDirname,
|
cwd: entrypointDirname,
|
||||||
stdio: 'inherit',
|
stdio: 'inherit',
|
||||||
});
|
});
|
||||||
} catch (err) {
|
} catch (err) {
|
||||||
console.error('failed to `cargo build --release`');
|
console.error('failed to `cargo build`');
|
||||||
throw err;
|
throw err;
|
||||||
}
|
}
|
||||||
|
|
||||||
const targetPath = path.join(workPath, 'target', 'release');
|
const targetPath = path.join(
|
||||||
const binaries = await inferCargoBinaries(
|
entrypointDirname,
|
||||||
cargoToml,
|
'target',
|
||||||
path.join(workPath, 'src'),
|
debug ? 'debug' : 'release',
|
||||||
);
|
);
|
||||||
|
const binaries = await inferCargoBinaries({
|
||||||
|
env: rustEnv,
|
||||||
|
cwd: entrypointDirname,
|
||||||
|
});
|
||||||
|
|
||||||
const lambdas = {};
|
const lambdas = {};
|
||||||
|
const lambdaPath = path.dirname(entrypoint);
|
||||||
await Promise.all(
|
await Promise.all(
|
||||||
binaries.map(async (binary) => {
|
binaries.map(async (binary) => {
|
||||||
const fsPath = path.join(targetPath, binary);
|
const fsPath = path.join(targetPath, binary);
|
||||||
@@ -72,19 +78,193 @@ exports.build = async ({ files, entrypoint, workPath }) => {
|
|||||||
runtime: 'provided',
|
runtime: 'provided',
|
||||||
});
|
});
|
||||||
|
|
||||||
lambdas[binary] = lambda;
|
lambdas[path.join(lambdaPath, binary)] = lambda;
|
||||||
}),
|
}),
|
||||||
);
|
);
|
||||||
|
|
||||||
return lambdas;
|
return lambdas;
|
||||||
};
|
}
|
||||||
|
|
||||||
exports.prepareCache = async ({ cachePath, workPath }) => {
|
async function cargoLocateProject(config) {
|
||||||
console.log('preparing cache...');
|
try {
|
||||||
rimraf.sync(path.join(cachePath, 'target'));
|
const { stdout: projectDescriptionStr } = await execa(
|
||||||
fs.renameSync(path.join(workPath, 'target'), path.join(cachePath, 'target'));
|
'cargo',
|
||||||
|
['locate-project'],
|
||||||
|
config,
|
||||||
|
);
|
||||||
|
const projectDescription = JSON.parse(projectDescriptionStr);
|
||||||
|
if (projectDescription != null && projectDescription.root != null) {
|
||||||
|
return projectDescription.root;
|
||||||
|
}
|
||||||
|
} catch (e) {
|
||||||
|
if (!/could not find/g.test(e.stderr)) {
|
||||||
|
console.error("Couldn't run `cargo locate-project`");
|
||||||
|
throw e;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
async function buildSingleFile({
|
||||||
|
workPath,
|
||||||
|
entrypoint,
|
||||||
|
downloadedFiles,
|
||||||
|
rustEnv,
|
||||||
|
config,
|
||||||
|
}) {
|
||||||
|
console.log('building single file');
|
||||||
|
const launcherPath = path.join(__dirname, 'launcher.rs');
|
||||||
|
let launcherData = await fs.readFile(launcherPath, 'utf8');
|
||||||
|
|
||||||
|
const entrypointPath = downloadedFiles[entrypoint].fsPath;
|
||||||
|
const entrypointDirname = path.dirname(entrypointPath);
|
||||||
|
launcherData = launcherData.replace(
|
||||||
|
'// PLACEHOLDER',
|
||||||
|
await fs.readFile(path.join(workPath, entrypoint)),
|
||||||
|
);
|
||||||
|
// replace the entrypoint with one that includes the the imports + lambda.start
|
||||||
|
await fs.remove(entrypointPath);
|
||||||
|
await fs.writeFile(entrypointPath, launcherData);
|
||||||
|
|
||||||
|
// Find a Cargo.toml file or TODO: create one
|
||||||
|
const cargoTomlFile = await cargoLocateProject({
|
||||||
|
env: rustEnv,
|
||||||
|
cwd: entrypointDirname,
|
||||||
|
});
|
||||||
|
|
||||||
|
// TODO: we're assuming there's a Cargo.toml file. We need to create one
|
||||||
|
// otherwise
|
||||||
|
let cargoToml;
|
||||||
|
try {
|
||||||
|
cargoToml = await parseTOMLStream(fs.createReadStream(cargoTomlFile));
|
||||||
|
} catch (err) {
|
||||||
|
console.error('Failed to parse TOML from entrypoint:', entrypoint);
|
||||||
|
throw err;
|
||||||
|
}
|
||||||
|
|
||||||
|
const binName = path
|
||||||
|
.basename(entrypointPath)
|
||||||
|
.replace(path.extname(entrypointPath), '');
|
||||||
|
const { package: pkg, dependencies } = cargoToml;
|
||||||
|
// default to latest now_lambda
|
||||||
|
dependencies.now_lambda = '*';
|
||||||
|
const tomlToWrite = toml.stringify({
|
||||||
|
package: pkg,
|
||||||
|
dependencies,
|
||||||
|
bin: [
|
||||||
|
{
|
||||||
|
name: binName,
|
||||||
|
path: entrypointPath,
|
||||||
|
},
|
||||||
|
],
|
||||||
|
});
|
||||||
|
console.log('toml to write:', tomlToWrite);
|
||||||
|
|
||||||
|
// Overwrite the Cargo.toml file with one that includes the `now_lambda`
|
||||||
|
// dependency and our binary. `dependencies` is a map so we don't run the
|
||||||
|
// risk of having 2 `now_lambda`s in there.
|
||||||
|
await fs.writeFile(cargoTomlFile, tomlToWrite);
|
||||||
|
|
||||||
|
const { debug } = config;
|
||||||
|
console.log('running `cargo build`...');
|
||||||
|
try {
|
||||||
|
await execa(
|
||||||
|
'cargo',
|
||||||
|
['build', '--bin', binName].concat(debug ? [] : ['--release']),
|
||||||
|
{
|
||||||
|
env: rustEnv,
|
||||||
|
cwd: entrypointDirname,
|
||||||
|
stdio: 'inherit',
|
||||||
|
},
|
||||||
|
);
|
||||||
|
} catch (err) {
|
||||||
|
console.error('failed to `cargo build`');
|
||||||
|
throw err;
|
||||||
|
}
|
||||||
|
|
||||||
|
const bin = path.join(
|
||||||
|
path.dirname(cargoTomlFile),
|
||||||
|
'target',
|
||||||
|
debug ? 'debug' : 'release',
|
||||||
|
binName,
|
||||||
|
);
|
||||||
|
|
||||||
|
const lambda = await createLambda({
|
||||||
|
files: {
|
||||||
|
bootstrap: new FileFsRef({ mode: 0o755, fsPath: bin }),
|
||||||
|
},
|
||||||
|
handler: 'bootstrap',
|
||||||
|
runtime: 'provided',
|
||||||
|
});
|
||||||
|
|
||||||
return {
|
return {
|
||||||
...(await glob('target/**', path.join(cachePath))),
|
[entrypoint]: lambda,
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
exports.build = async (m) => {
|
||||||
|
const { files, entrypoint, workPath } = m;
|
||||||
|
console.log('downloading files');
|
||||||
|
const downloadedFiles = await download(files, workPath);
|
||||||
|
|
||||||
|
const { PATH: toolchainPath, ...otherEnv } = await installRustAndGCC();
|
||||||
|
const { PATH, HOME } = process.env;
|
||||||
|
const rustEnv = {
|
||||||
|
...process.env,
|
||||||
|
...otherEnv,
|
||||||
|
PATH: `${path.join(HOME, '.cargo/bin')}:${toolchainPath}:${PATH}`,
|
||||||
|
};
|
||||||
|
|
||||||
|
const newM = Object.assign(m, { downloadedFiles, rustEnv });
|
||||||
|
if (path.extname(entrypoint) === '.toml') {
|
||||||
|
return buildWholeProject(newM);
|
||||||
|
}
|
||||||
|
return buildSingleFile(newM);
|
||||||
|
};
|
||||||
|
|
||||||
|
exports.prepareCache = async ({ cachePath, entrypoint, workPath }) => {
|
||||||
|
console.log('preparing cache...');
|
||||||
|
|
||||||
|
let targetFolderDir;
|
||||||
|
if (path.extname(entrypoint) === '.toml') {
|
||||||
|
targetFolderDir = path.dirname(path.join(workPath, entrypoint));
|
||||||
|
} else {
|
||||||
|
const { PATH, HOME } = process.env;
|
||||||
|
const rustEnv = {
|
||||||
|
...process.env,
|
||||||
|
PATH: `${path.join(HOME, '.cargo/bin')}:${PATH}`,
|
||||||
|
};
|
||||||
|
const entrypointDirname = path.dirname(path.join(workPath, entrypoint));
|
||||||
|
const cargoTomlFile = await cargoLocateProject({
|
||||||
|
env: rustEnv,
|
||||||
|
cwd: entrypointDirname,
|
||||||
|
});
|
||||||
|
|
||||||
|
if (cargoTomlFile != null) {
|
||||||
|
targetFolderDir = path.dirname(cargoTomlFile);
|
||||||
|
} else {
|
||||||
|
// `Cargo.toml` doesn't exist, in `build` we put it in the same
|
||||||
|
// path as the entrypoint.
|
||||||
|
targetFolderDir = path.dirname(path.join(workPath, entrypoint));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const cacheEntrypointDirname = path.join(
|
||||||
|
cachePath,
|
||||||
|
path.relative(workPath, targetFolderDir),
|
||||||
|
);
|
||||||
|
|
||||||
|
// Remove the target folder to avoid 'directory already exists' errors
|
||||||
|
fs.removeSync(path.join(cacheEntrypointDirname, 'target'));
|
||||||
|
fs.mkdirpSync(cacheEntrypointDirname);
|
||||||
|
// Move the target folder to the cache location
|
||||||
|
fs.renameSync(
|
||||||
|
path.join(targetFolderDir, 'target'),
|
||||||
|
path.join(cacheEntrypointDirname, 'target'),
|
||||||
|
);
|
||||||
|
|
||||||
|
return {
|
||||||
|
...(await glob('**/**', path.join(cachePath))),
|
||||||
};
|
};
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -1,73 +0,0 @@
|
|||||||
const fs = require('fs');
|
|
||||||
const path = require('path');
|
|
||||||
|
|
||||||
function readdir(dir) {
|
|
||||||
return new Promise((resolve, reject) => {
|
|
||||||
fs.readdir(dir, (err, files) => {
|
|
||||||
if (err != null) {
|
|
||||||
return reject(err);
|
|
||||||
}
|
|
||||||
|
|
||||||
return resolve(files);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
function exists(p) {
|
|
||||||
return new Promise((resolve, reject) => {
|
|
||||||
fs.exists(p, (err, res) => {
|
|
||||||
if (err != null) {
|
|
||||||
return reject(err);
|
|
||||||
}
|
|
||||||
|
|
||||||
return resolve(res);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
function stat(p) {
|
|
||||||
return new Promise((resolve, reject) => {
|
|
||||||
fs.stat(p, (err, stats) => {
|
|
||||||
if (err != null) {
|
|
||||||
return reject(err);
|
|
||||||
}
|
|
||||||
|
|
||||||
return resolve(stats);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
async function inferCargoBinaries(cargoToml, srcDir) {
|
|
||||||
const { package: pkg, bin } = cargoToml;
|
|
||||||
const binaries = [];
|
|
||||||
const hasMain = (await readdir(srcDir)).includes('main.rs');
|
|
||||||
|
|
||||||
if (hasMain) {
|
|
||||||
binaries.push(pkg.name);
|
|
||||||
}
|
|
||||||
|
|
||||||
// From: https://doc.rust-lang.org/cargo/reference/manifest.html#the-project-layout
|
|
||||||
// Do note, however, once you add a [[bin]] section (see below), Cargo will
|
|
||||||
// no longer automatically build files located in src/bin/*.rs. Instead you
|
|
||||||
// must create a [[bin]] section for each file you want to build.
|
|
||||||
if (Array.isArray(bin)) {
|
|
||||||
bin.forEach((binary) => {
|
|
||||||
binaries.push(binary.name);
|
|
||||||
});
|
|
||||||
} else {
|
|
||||||
const binDir = path.join(srcDir, 'bin');
|
|
||||||
const filesInSrcBin = (await exists(binDir)) && (await stat(binDir)).isDirectory()
|
|
||||||
? await readdir(binDir)
|
|
||||||
: [];
|
|
||||||
|
|
||||||
filesInSrcBin.forEach((file) => {
|
|
||||||
if (file.endsWith('.rs')) {
|
|
||||||
binaries.push(file.slice(0, -3));
|
|
||||||
}
|
|
||||||
});
|
|
||||||
}
|
|
||||||
|
|
||||||
return binaries;
|
|
||||||
}
|
|
||||||
|
|
||||||
module.exports = inferCargoBinaries;
|
|
||||||
8
packages/now-rust/launcher.rs
Normal file
8
packages/now-rust/launcher.rs
Normal file
@@ -0,0 +1,8 @@
|
|||||||
|
use now_lambda::lambda;
|
||||||
|
use std::error::Error;
|
||||||
|
|
||||||
|
// PLACEHOLDER
|
||||||
|
|
||||||
|
fn main() -> Result<(), Box<dyn Error>> {
|
||||||
|
Ok(lambda!(handler))
|
||||||
|
}
|
||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/rust",
|
"name": "@now/rust",
|
||||||
"version": "0.0.2-canary.0",
|
"version": "0.0.3-canary.2",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
@@ -10,14 +10,13 @@
|
|||||||
"files": [
|
"files": [
|
||||||
"index.js",
|
"index.js",
|
||||||
"download-install-rust-toolchain.js",
|
"download-install-rust-toolchain.js",
|
||||||
"inferCargoBinaries.js"
|
"launcher.rs"
|
||||||
],
|
],
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"concat-stream": "^2.0.0",
|
"@iarna/toml": "^2.2.1",
|
||||||
"execa": "^1.0.0",
|
"execa": "^1.0.0",
|
||||||
|
"fs-extra": "^7.0.1",
|
||||||
"node-fetch": "^2.3.0",
|
"node-fetch": "^2.3.0",
|
||||||
"rimraf": "^2.6.3",
|
"tar": "^4.4.8"
|
||||||
"tar": "^4.4.8",
|
|
||||||
"toml": "^2.3.3"
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
//! Provides a Now Lambda oriented request and response body entity interface
|
//! Provides a Now Lambda oriented request and response body entity interface
|
||||||
|
|
||||||
use std::{borrow::Cow, ops::Deref};
|
use std::{borrow::Cow, ops::Deref, str};
|
||||||
|
|
||||||
use base64::display::Base64Display;
|
use base64::display::Base64Display;
|
||||||
use serde::ser::{Error as SerError, Serialize, Serializer};
|
use serde::ser::{Error as SerError, Serialize, Serializer};
|
||||||
@@ -76,6 +76,12 @@ impl From<()> for Body {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl From<Body> for () {
|
||||||
|
fn from(_: Body) -> Self {
|
||||||
|
()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl<'a> From<&'a str> for Body {
|
impl<'a> From<&'a str> for Body {
|
||||||
fn from(s: &'a str) -> Self {
|
fn from(s: &'a str) -> Self {
|
||||||
Body::Text(s.into())
|
Body::Text(s.into())
|
||||||
@@ -87,6 +93,15 @@ impl From<String> for Body {
|
|||||||
Body::Text(b)
|
Body::Text(b)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
impl From<Body> for String {
|
||||||
|
fn from(b: Body) -> String {
|
||||||
|
match b {
|
||||||
|
Body::Empty => String::from(""),
|
||||||
|
Body::Text(t) => t,
|
||||||
|
Body::Binary(b) => str::from_utf8(&b).unwrap().to_owned(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl From<Cow<'static, str>> for Body {
|
impl From<Cow<'static, str>> for Body {
|
||||||
#[inline]
|
#[inline]
|
||||||
@@ -98,6 +113,13 @@ impl From<Cow<'static, str>> for Body {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl From<Body> for Cow<'static, str> {
|
||||||
|
#[inline]
|
||||||
|
fn from(b: Body) -> Cow<'static, str> {
|
||||||
|
Cow::Owned(String::from(b))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl From<Cow<'static, [u8]>> for Body {
|
impl From<Cow<'static, [u8]>> for Body {
|
||||||
#[inline]
|
#[inline]
|
||||||
fn from(cow: Cow<'static, [u8]>) -> Body {
|
fn from(cow: Cow<'static, [u8]>) -> Body {
|
||||||
@@ -108,12 +130,29 @@ impl From<Cow<'static, [u8]>> for Body {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl From<Body> for Cow<'static, [u8]> {
|
||||||
|
#[inline]
|
||||||
|
fn from(b: Body) -> Self {
|
||||||
|
Cow::Owned(b.as_ref().to_owned())
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl From<Vec<u8>> for Body {
|
impl From<Vec<u8>> for Body {
|
||||||
fn from(b: Vec<u8>) -> Self {
|
fn from(b: Vec<u8>) -> Self {
|
||||||
Body::Binary(b)
|
Body::Binary(b)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl From<Body> for Vec<u8> {
|
||||||
|
fn from(b: Body) -> Self {
|
||||||
|
match b {
|
||||||
|
Body::Empty => "".as_bytes().to_owned(),
|
||||||
|
Body::Text(t) => t.into_bytes(),
|
||||||
|
Body::Binary(b) => b.to_owned(),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
impl<'a> From<&'a [u8]> for Body {
|
impl<'a> From<&'a [u8]> for Body {
|
||||||
fn from(b: &'a [u8]) -> Self {
|
fn from(b: &'a [u8]) -> Self {
|
||||||
Body::Binary(b.to_vec())
|
Body::Binary(b.to_vec())
|
||||||
|
|||||||
@@ -1,3 +1,4 @@
|
|||||||
|
use http;
|
||||||
use lambda_runtime::error::LambdaErrorExt;
|
use lambda_runtime::error::LambdaErrorExt;
|
||||||
use std::{error::Error, fmt};
|
use std::{error::Error, fmt};
|
||||||
|
|
||||||
@@ -19,13 +20,21 @@ impl fmt::Display for NowError {
|
|||||||
write!(f, "{}", self.msg)
|
write!(f, "{}", self.msg)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
impl Error for NowError {}
|
impl Error for NowError {}
|
||||||
|
|
||||||
impl From<std::num::ParseIntError> for NowError {
|
impl From<std::num::ParseIntError> for NowError {
|
||||||
fn from(i: std::num::ParseIntError) -> Self {
|
fn from(i: std::num::ParseIntError) -> Self {
|
||||||
NowError::new(&format!("{}", i))
|
NowError::new(&format!("{}", i))
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
impl From<http::Error> for NowError {
|
||||||
|
fn from(i: http::Error) -> Self {
|
||||||
|
NowError::new(&format!("{}", i))
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
// the value returned by the error_type function is included as the
|
// the value returned by the error_type function is included as the
|
||||||
// `errorType` in the AWS Lambda response
|
// `errorType` in the AWS Lambda response
|
||||||
impl LambdaErrorExt for NowError {
|
impl LambdaErrorExt for NowError {
|
||||||
|
|||||||
@@ -21,16 +21,16 @@ use crate::{
|
|||||||
pub type Request = http::Request<Body>;
|
pub type Request = http::Request<Body>;
|
||||||
|
|
||||||
/// Functions acting as Now Lambda handlers must conform to this type.
|
/// Functions acting as Now Lambda handlers must conform to this type.
|
||||||
pub trait Handler<R> {
|
pub trait Handler<R, B, E> {
|
||||||
/// Method to execute the handler function
|
/// Method to execute the handler function
|
||||||
fn run(&mut self, event: Request) -> Result<R, NowError>;
|
fn run(&mut self, event: http::Request<B>) -> Result<R, E>;
|
||||||
}
|
}
|
||||||
|
|
||||||
impl<Function, R> Handler<R> for Function
|
impl<Function, R, B, E> Handler<R, B, E> for Function
|
||||||
where
|
where
|
||||||
Function: FnMut(Request) -> Result<R, NowError>,
|
Function: FnMut(http::Request<B>) -> Result<R, E>,
|
||||||
{
|
{
|
||||||
fn run(&mut self, event: Request) -> Result<R, NowError> {
|
fn run(&mut self, event: http::Request<B>) -> Result<R, E> {
|
||||||
(*self)(event)
|
(*self)(event)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
@@ -43,8 +43,10 @@ where
|
|||||||
///
|
///
|
||||||
/// # Panics
|
/// # Panics
|
||||||
/// The function panics if the Lambda environment variables are not set.
|
/// The function panics if the Lambda environment variables are not set.
|
||||||
pub fn start<R>(f: impl Handler<R>, runtime: Option<TokioRuntime>)
|
pub fn start<R, B, E>(f: impl Handler<R, B, E>, runtime: Option<TokioRuntime>)
|
||||||
where
|
where
|
||||||
|
B: From<Body>,
|
||||||
|
E: Into<NowError>,
|
||||||
R: IntoResponse,
|
R: IntoResponse,
|
||||||
{
|
{
|
||||||
// handler requires a mutable ref
|
// handler requires a mutable ref
|
||||||
@@ -56,8 +58,10 @@ where
|
|||||||
match parse_result {
|
match parse_result {
|
||||||
Ok(req) => {
|
Ok(req) => {
|
||||||
debug!("Deserialized Now proxy request successfully");
|
debug!("Deserialized Now proxy request successfully");
|
||||||
func.run(req.into())
|
let request: http::Request<Body> = req.into();
|
||||||
|
func.run(request.map(|b| b.into()))
|
||||||
.map(|resp| NowResponse::from(resp.into_response()))
|
.map(|resp| NowResponse::from(resp.into_response()))
|
||||||
|
.map_err(|e| e.into())
|
||||||
}
|
}
|
||||||
Err(e) => {
|
Err(e) => {
|
||||||
error!("Could not deserialize event body to NowRequest {}", e);
|
error!("Could not deserialize event body to NowRequest {}", e);
|
||||||
|
|||||||
@@ -1,13 +1,8 @@
|
|||||||
use std::{borrow::Cow, fmt, mem};
|
use std::{borrow::Cow, fmt, mem};
|
||||||
|
|
||||||
use http::{self, header::HeaderValue, HeaderMap, Method, Request as HttpRequest};
|
use http::{self, header::HeaderValue, HeaderMap, Method, Request as HttpRequest};
|
||||||
use serde::{
|
use serde::de::{Deserializer, Error as DeError, MapAccess, Visitor};
|
||||||
de::{Error as DeError, MapAccess, Visitor},
|
|
||||||
Deserialize, Deserializer,
|
|
||||||
};
|
|
||||||
use serde_derive::Deserialize;
|
use serde_derive::Deserialize;
|
||||||
#[allow(unused_imports)]
|
|
||||||
use serde_json::Value;
|
|
||||||
|
|
||||||
use crate::body::Body;
|
use crate::body::Body;
|
||||||
|
|
||||||
@@ -92,18 +87,6 @@ where
|
|||||||
deserializer.deserialize_map(HeaderVisitor)
|
deserializer.deserialize_map(HeaderVisitor)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// deserializes (json) null values to their default values
|
|
||||||
// https://github.com/serde-rs/serde/issues/1098
|
|
||||||
#[allow(dead_code)]
|
|
||||||
fn nullable_default<'de, T, D>(deserializer: D) -> Result<T, D::Error>
|
|
||||||
where
|
|
||||||
D: Deserializer<'de>,
|
|
||||||
T: Default + Deserialize<'de>,
|
|
||||||
{
|
|
||||||
let opt = Option::deserialize(deserializer)?;
|
|
||||||
Ok(opt.unwrap_or_else(T::default))
|
|
||||||
}
|
|
||||||
|
|
||||||
impl<'a> From<NowRequest<'a>> for HttpRequest<Body> {
|
impl<'a> From<NowRequest<'a>> for HttpRequest<Body> {
|
||||||
fn from(value: NowRequest<'_>) -> Self {
|
fn from(value: NowRequest<'_>) -> Self {
|
||||||
let NowRequest {
|
let NowRequest {
|
||||||
|
|||||||
@@ -4,10 +4,7 @@ use http::{
|
|||||||
header::{HeaderMap, HeaderValue},
|
header::{HeaderMap, HeaderValue},
|
||||||
Response,
|
Response,
|
||||||
};
|
};
|
||||||
use serde::{
|
use serde::ser::{Error as SerError, SerializeMap, Serializer};
|
||||||
ser::{Error as SerError, SerializeMap},
|
|
||||||
Serializer,
|
|
||||||
};
|
|
||||||
use serde_derive::Serialize;
|
use serde_derive::Serialize;
|
||||||
|
|
||||||
use crate::body::Body;
|
use crate::body::Body;
|
||||||
|
|||||||
@@ -4,10 +4,7 @@ use std::{
|
|||||||
sync::Arc,
|
sync::Arc,
|
||||||
};
|
};
|
||||||
|
|
||||||
use serde::{
|
use serde::de::{Deserialize, Deserializer, MapAccess, Visitor};
|
||||||
de::{MapAccess, Visitor},
|
|
||||||
Deserialize, Deserializer,
|
|
||||||
};
|
|
||||||
|
|
||||||
/// A read-only view into a map of string data
|
/// A read-only view into a map of string data
|
||||||
#[derive(Default, Debug, PartialEq)]
|
#[derive(Default, Debug, PartialEq)]
|
||||||
|
|||||||
@@ -1,154 +0,0 @@
|
|||||||
/* global afterAll, beforeAll, describe, expect, it, jest */
|
|
||||||
const fs = require('fs');
|
|
||||||
const inferCargoBinaries = require('../inferCargoBinaries');
|
|
||||||
|
|
||||||
const { exists, readdir, stat } = fs;
|
|
||||||
const isDir = fs.Stats.prototype.isDirectory;
|
|
||||||
|
|
||||||
beforeAll(() => {
|
|
||||||
fs.exists = jest.fn((p, cb) => cb(null, false));
|
|
||||||
});
|
|
||||||
|
|
||||||
afterAll(() => {
|
|
||||||
fs.readdir = readdir;
|
|
||||||
fs.stat = stat;
|
|
||||||
fs.Stats.prototype.isDirectory = isDir;
|
|
||||||
fs.exists = exists;
|
|
||||||
});
|
|
||||||
|
|
||||||
// src/
|
|
||||||
// |- main.rs
|
|
||||||
describe('one binary, src/main.rs', async () => {
|
|
||||||
beforeAll(() => {
|
|
||||||
fs.readdir = jest.fn((p, cb) => cb(null, ['main.rs']));
|
|
||||||
});
|
|
||||||
|
|
||||||
it('infers only one binary', async () => {
|
|
||||||
const toml = {
|
|
||||||
package: {
|
|
||||||
name: 'foo',
|
|
||||||
},
|
|
||||||
};
|
|
||||||
expect(inferCargoBinaries(toml, '/path/to/src')).resolves.toEqual(['foo']);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
// [[bin]] sections in `Cargo.toml`
|
|
||||||
// `main.rs` -> `package.name`
|
|
||||||
// `bar.rs` -> `bin.name`
|
|
||||||
// src/
|
|
||||||
// |- bar.rs
|
|
||||||
// |- main.rs
|
|
||||||
describe('two binaries, src/main.rs, src/bar.rs', async () => {
|
|
||||||
beforeAll(() => {
|
|
||||||
fs.readdir = jest.fn((p, cb) => cb(null, ['main.rs', 'bar.rs']));
|
|
||||||
});
|
|
||||||
|
|
||||||
it('infers two binaries', async () => {
|
|
||||||
const toml = {
|
|
||||||
package: {
|
|
||||||
name: 'foo',
|
|
||||||
},
|
|
||||||
bin: [{ name: 'bar', path: 'src/bar.rs' }],
|
|
||||||
};
|
|
||||||
expect((await inferCargoBinaries(toml, '/path/to/src')).sort()).toEqual([
|
|
||||||
'bar',
|
|
||||||
'foo',
|
|
||||||
]);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
// no main.rs
|
|
||||||
// src/
|
|
||||||
// |- foo.rs
|
|
||||||
describe('one named binary, no main.rs', async () => {
|
|
||||||
beforeAll(() => {
|
|
||||||
fs.readdir = jest.fn((p, cb) => cb(null, ['bar.rs']));
|
|
||||||
});
|
|
||||||
|
|
||||||
it('infers only one binary', async () => {
|
|
||||||
const toml = {
|
|
||||||
package: {
|
|
||||||
name: 'foo',
|
|
||||||
},
|
|
||||||
bin: [{ name: 'bar', path: 'src/bar.rs' }],
|
|
||||||
};
|
|
||||||
expect((await inferCargoBinaries(toml, '/path/to/src')).sort()).toEqual([
|
|
||||||
'bar',
|
|
||||||
]);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
// `src/bin` folder
|
|
||||||
// src/
|
|
||||||
// |- bin/
|
|
||||||
// | |- bar.rs
|
|
||||||
// | |- baz.rs
|
|
||||||
// |- main.rs
|
|
||||||
describe('multiple binaries in bin/, no [[bin]] section', async () => {
|
|
||||||
beforeAll(() => {
|
|
||||||
fs.readdir = jest.fn((p, cb) => {
|
|
||||||
if (p === '/path/to/src') {
|
|
||||||
return cb(null, ['bin', 'main.rs']);
|
|
||||||
}
|
|
||||||
if (p === '/path/to/src/bin') {
|
|
||||||
return cb(null, ['bar.rs', 'baz.rs']);
|
|
||||||
}
|
|
||||||
|
|
||||||
return cb('some error');
|
|
||||||
});
|
|
||||||
fs.exists = jest.fn((p, cb) => cb(null, p.endsWith('bin')));
|
|
||||||
fs.stat = jest.fn((_, cb) => cb(null, new fs.Stats()));
|
|
||||||
fs.Stats.prototype.isDirectory = jest.fn(() => true);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('infers three binaries', async () => {
|
|
||||||
const toml = {
|
|
||||||
package: {
|
|
||||||
name: 'foo',
|
|
||||||
},
|
|
||||||
};
|
|
||||||
expect((await inferCargoBinaries(toml, '/path/to/src')).sort()).toEqual([
|
|
||||||
'bar',
|
|
||||||
'baz',
|
|
||||||
'foo',
|
|
||||||
]);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
// `src/bin` folder, bin sections ignore baz.rs
|
|
||||||
// src/
|
|
||||||
// |- bin/
|
|
||||||
// | |- bar.rs
|
|
||||||
// | |- baz.rs
|
|
||||||
// |- main.rs
|
|
||||||
describe('src/bin exists but one binary is ignored', async () => {
|
|
||||||
beforeAll(() => {
|
|
||||||
fs.readdir = jest.fn((p, cb) => {
|
|
||||||
if (p === '/path/to/src') {
|
|
||||||
return cb(null, ['bin', 'main.rs']);
|
|
||||||
}
|
|
||||||
if (p === '/path/to/src/bin') {
|
|
||||||
return cb(null, ['bar.rs', 'baz.rs']);
|
|
||||||
}
|
|
||||||
|
|
||||||
return cb('some error');
|
|
||||||
});
|
|
||||||
fs.exists = jest.fn((p, cb) => cb(null, p.endsWith('bin')));
|
|
||||||
fs.stat = jest.fn((_, cb) => cb(null, new fs.Stats()));
|
|
||||||
fs.Stats.prototype.isDirectory = jest.fn(() => true);
|
|
||||||
});
|
|
||||||
|
|
||||||
it('infers only one binary', async () => {
|
|
||||||
const toml = {
|
|
||||||
package: {
|
|
||||||
name: 'foo',
|
|
||||||
},
|
|
||||||
bin: [{ name: 'bar', path: 'src/bar.rs' }],
|
|
||||||
};
|
|
||||||
expect((await inferCargoBinaries(toml, '/path/to/src')).sort()).toEqual([
|
|
||||||
'bar',
|
|
||||||
'foo',
|
|
||||||
]);
|
|
||||||
});
|
|
||||||
});
|
|
||||||
@@ -1,12 +1,22 @@
|
|||||||
const download = require('@now/build-utils/fs/download.js'); // eslint-disable-line import/no-extraneous-dependencies
|
const download = require('@now/build-utils/fs/download.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
const glob = require('@now/build-utils/fs/glob.js'); // eslint-disable-line import/no-extraneous-dependencies
|
const glob = require('@now/build-utils/fs/glob.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
const path = require('path');
|
const path = require('path');
|
||||||
|
const { existsSync } = require('fs');
|
||||||
const {
|
const {
|
||||||
runNpmInstall,
|
runNpmInstall,
|
||||||
runPackageJsonScript,
|
runPackageJsonScript,
|
||||||
runShellScript,
|
runShellScript,
|
||||||
} = require('@now/build-utils/fs/run-user-scripts.js'); // eslint-disable-line import/no-extraneous-dependencies
|
} = require('@now/build-utils/fs/run-user-scripts.js'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
|
|
||||||
|
function validateDistDir(distDir) {
|
||||||
|
const distDirName = path.basename(distDir);
|
||||||
|
if (!existsSync(distDir)) {
|
||||||
|
const message = `Build was unable to create the distDir: ${distDirName}.`
|
||||||
|
+ '\nMake sure you mentioned the correct dist directory: https://zeit.co/docs/v2/deployments/official-builders/static-build-now-static-build/#configuring-the-dist-directory';
|
||||||
|
throw new Error(message);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
exports.build = async ({
|
exports.build = async ({
|
||||||
files, entrypoint, workPath, config,
|
files, entrypoint, workPath, config,
|
||||||
}) => {
|
}) => {
|
||||||
@@ -24,6 +34,7 @@ exports.build = async ({
|
|||||||
if (path.basename(entrypoint) === 'package.json') {
|
if (path.basename(entrypoint) === 'package.json') {
|
||||||
await runNpmInstall(entrypointFsDirname, ['--prefer-offline']);
|
await runNpmInstall(entrypointFsDirname, ['--prefer-offline']);
|
||||||
if (await runPackageJsonScript(entrypointFsDirname, 'now-build')) {
|
if (await runPackageJsonScript(entrypointFsDirname, 'now-build')) {
|
||||||
|
validateDistDir(distPath);
|
||||||
return glob('**', distPath, mountpoint);
|
return glob('**', distPath, mountpoint);
|
||||||
}
|
}
|
||||||
throw new Error(`An error running "now-build" script in "${entrypoint}"`);
|
throw new Error(`An error running "now-build" script in "${entrypoint}"`);
|
||||||
@@ -31,6 +42,7 @@ exports.build = async ({
|
|||||||
|
|
||||||
if (path.extname(entrypoint) === '.sh') {
|
if (path.extname(entrypoint) === '.sh') {
|
||||||
await runShellScript(path.join(workPath, entrypoint));
|
await runShellScript(path.join(workPath, entrypoint));
|
||||||
|
validateDistDir(distPath);
|
||||||
return glob('**', distPath, mountpoint);
|
return glob('**', distPath, mountpoint);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -1,6 +1,6 @@
|
|||||||
{
|
{
|
||||||
"name": "@now/static-build",
|
"name": "@now/static-build",
|
||||||
"version": "0.4.18-canary.0",
|
"version": "0.4.19-canary.1",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
|
|||||||
6
packages/now-static-build/test/fixtures/04-wrong-dist-dir/now.json
vendored
Normal file
6
packages/now-static-build/test/fixtures/04-wrong-dist-dir/now.json
vendored
Normal file
@@ -0,0 +1,6 @@
|
|||||||
|
{
|
||||||
|
"version": 2,
|
||||||
|
"builds": [
|
||||||
|
{ "src": "package.json", "use": "@now/static-build", "config": {"distDir": "out"} }
|
||||||
|
]
|
||||||
|
}
|
||||||
8
packages/now-static-build/test/fixtures/04-wrong-dist-dir/package.json
vendored
Normal file
8
packages/now-static-build/test/fixtures/04-wrong-dist-dir/package.json
vendored
Normal file
@@ -0,0 +1,8 @@
|
|||||||
|
{
|
||||||
|
"dependencies": {
|
||||||
|
"cowsay": "^1.3.1"
|
||||||
|
},
|
||||||
|
"scripts": {
|
||||||
|
"now-build": "mkdir dist && cowsay cow:RANDOMNESS_PLACEHOLDER > dist/index.txt"
|
||||||
|
}
|
||||||
|
}
|
||||||
8
packages/now-static-build/test/fixtures/04-wrong-dist-dir/subdirectory/package.json
vendored
Normal file
8
packages/now-static-build/test/fixtures/04-wrong-dist-dir/subdirectory/package.json
vendored
Normal file
@@ -0,0 +1,8 @@
|
|||||||
|
{
|
||||||
|
"dependencies": {
|
||||||
|
"yodasay": "^1.1.6"
|
||||||
|
},
|
||||||
|
"scripts": {
|
||||||
|
"now-build": "mkdir dist && yodasay yoda:RANDOMNESS_PLACEHOLDER > dist/index.txt"
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -7,7 +7,7 @@ const {
|
|||||||
testDeployment,
|
testDeployment,
|
||||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||||
|
|
||||||
jest.setTimeout(2 * 60 * 1000);
|
jest.setTimeout(4 * 60 * 1000);
|
||||||
const buildUtilsUrl = '@canary';
|
const buildUtilsUrl = '@canary';
|
||||||
let builderUrl;
|
let builderUrl;
|
||||||
|
|
||||||
@@ -21,6 +21,21 @@ const fixturesPath = path.resolve(__dirname, 'fixtures');
|
|||||||
|
|
||||||
// eslint-disable-next-line no-restricted-syntax
|
// eslint-disable-next-line no-restricted-syntax
|
||||||
for (const fixture of fs.readdirSync(fixturesPath)) {
|
for (const fixture of fs.readdirSync(fixturesPath)) {
|
||||||
|
if (fixture === '04-wrong-dist-dir') {
|
||||||
|
// eslint-disable-next-line no-loop-func
|
||||||
|
it(`should not build ${fixture}`, async () => {
|
||||||
|
try {
|
||||||
|
await testDeployment(
|
||||||
|
{ builderUrl, buildUtilsUrl },
|
||||||
|
path.join(fixturesPath, fixture),
|
||||||
|
);
|
||||||
|
} catch (err) {
|
||||||
|
expect(err.message).toMatch(/is ERROR/);
|
||||||
|
}
|
||||||
|
});
|
||||||
|
continue; //eslint-disable-line
|
||||||
|
}
|
||||||
|
|
||||||
// eslint-disable-next-line no-loop-func
|
// eslint-disable-next-line no-loop-func
|
||||||
it(`should build ${fixture}`, async () => {
|
it(`should build ${fixture}`, async () => {
|
||||||
await expect(
|
await expect(
|
||||||
|
|||||||
1
test/integration/now-next/monorepo/shared/hello.js
Normal file
1
test/integration/now-next/monorepo/shared/hello.js
Normal file
@@ -0,0 +1 @@
|
|||||||
|
module.exports = () => 'Hello!';
|
||||||
@@ -1,9 +1,9 @@
|
|||||||
{
|
{
|
||||||
"name": "monorepo",
|
"name": "monorepo",
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"next": "^8.0.0-canary.2",
|
"next": "^8.0.0",
|
||||||
"react": "^16.7.0",
|
"react": "^16.8.0",
|
||||||
"react-dom": "^16.7.0"
|
"react-dom": "^16.8.0"
|
||||||
},
|
},
|
||||||
"scripts": {
|
"scripts": {
|
||||||
"now-build": "next build"
|
"now-build": "next build"
|
||||||
|
|||||||
@@ -1 +1,3 @@
|
|||||||
export default () => 'Index page';
|
import hello from '../../shared/hello';
|
||||||
|
|
||||||
|
export default () => `${hello()} Welcome to the index page`;
|
||||||
|
|||||||
@@ -18,8 +18,9 @@ async function nowDeploy (bodies, randomness) {
|
|||||||
|
|
||||||
const nowDeployPayload = {
|
const nowDeployPayload = {
|
||||||
version: 2,
|
version: 2,
|
||||||
env: Object.assign({}, nowJson.env, { RANDOMNESS_ENV_VAR: randomness }),
|
public: true,
|
||||||
build: { env: { RANDOMNESS_BUILD_ENV_VAR: randomness } },
|
env: { ...nowJson.env, RANDOMNESS_ENV_VAR: randomness },
|
||||||
|
build: { env: { ...(nowJson.build || {}).env, RANDOMNESS_BUILD_ENV_VAR: randomness } },
|
||||||
name: 'test',
|
name: 'test',
|
||||||
files,
|
files,
|
||||||
builds: nowJson.builds,
|
builds: nowJson.builds,
|
||||||
|
|||||||
@@ -2,11 +2,7 @@ const {
|
|||||||
excludeFiles,
|
excludeFiles,
|
||||||
validateEntrypoint,
|
validateEntrypoint,
|
||||||
includeOnlyEntryDirectory,
|
includeOnlyEntryDirectory,
|
||||||
moveEntryDirectoryToRoot,
|
|
||||||
excludeLockFiles,
|
|
||||||
normalizePackageJson,
|
normalizePackageJson,
|
||||||
excludeStaticDirectory,
|
|
||||||
onlyStaticDirectory,
|
|
||||||
} = require('@now/next/utils');
|
} = require('@now/next/utils');
|
||||||
const FileRef = require('@now/build-utils/file-ref'); // eslint-disable-line import/no-extraneous-dependencies
|
const FileRef = require('@now/build-utils/file-ref'); // eslint-disable-line import/no-extraneous-dependencies
|
||||||
|
|
||||||
@@ -46,7 +42,7 @@ describe('validateEntrypoint', () => {
|
|||||||
});
|
});
|
||||||
|
|
||||||
describe('includeOnlyEntryDirectory', () => {
|
describe('includeOnlyEntryDirectory', () => {
|
||||||
it('should exclude files outside entry directory', () => {
|
it('should include files outside entry directory', () => {
|
||||||
const entryDirectory = 'frontend';
|
const entryDirectory = 'frontend';
|
||||||
const files = {
|
const files = {
|
||||||
'frontend/pages/index.js': new FileRef({ digest: 'index' }),
|
'frontend/pages/index.js': new FileRef({ digest: 'index' }),
|
||||||
@@ -58,149 +54,6 @@ describe('includeOnlyEntryDirectory', () => {
|
|||||||
expect(result['package.json']).toBeUndefined();
|
expect(result['package.json']).toBeUndefined();
|
||||||
expect(result['package-lock.json']).toBeUndefined();
|
expect(result['package-lock.json']).toBeUndefined();
|
||||||
});
|
});
|
||||||
|
|
||||||
it('should handle entry directory being dot', () => {
|
|
||||||
const entryDirectory = '.';
|
|
||||||
const files = {
|
|
||||||
'frontend/pages/index.js': new FileRef({ digest: 'index' }),
|
|
||||||
'package.json': new FileRef({ digest: 'package' }),
|
|
||||||
'package-lock.json': new FileRef({ digest: 'package-lock' }),
|
|
||||||
};
|
|
||||||
const result = includeOnlyEntryDirectory(files, entryDirectory);
|
|
||||||
expect(result['frontend/pages/index.js']).toBeDefined();
|
|
||||||
expect(result['package.json']).toBeDefined();
|
|
||||||
expect(result['package-lock.json']).toBeDefined();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('moveEntryDirectoryToRoot', () => {
|
|
||||||
it('should move entrydirectory files to the root', () => {
|
|
||||||
const entryDirectory = 'frontend';
|
|
||||||
const files = {
|
|
||||||
'frontend/pages/index.js': new FileRef({ digest: 'index' }),
|
|
||||||
};
|
|
||||||
const result = moveEntryDirectoryToRoot(files, entryDirectory);
|
|
||||||
expect(result['pages/index.js']).toBeDefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should work with deep nested subdirectories', () => {
|
|
||||||
const entryDirectory = 'frontend/my/app';
|
|
||||||
const files = {
|
|
||||||
'frontend/my/app/pages/index.js': new FileRef({ digest: 'index' }),
|
|
||||||
};
|
|
||||||
const result = moveEntryDirectoryToRoot(files, entryDirectory);
|
|
||||||
expect(result['pages/index.js']).toBeDefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should do nothing when entry directory is dot', () => {
|
|
||||||
const entryDirectory = '.';
|
|
||||||
const files = {
|
|
||||||
'frontend/pages/index.js': new FileRef({ digest: 'index' }),
|
|
||||||
};
|
|
||||||
const result = moveEntryDirectoryToRoot(files, entryDirectory);
|
|
||||||
expect(result['frontend/pages/index.js']).toBeDefined();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('excludeLockFiles', () => {
|
|
||||||
it('should remove package-lock.json', () => {
|
|
||||||
const files = {
|
|
||||||
'frontend/pages/index.js': new FileRef({ digest: 'index' }),
|
|
||||||
'package.json': new FileRef({ digest: 'package' }),
|
|
||||||
'package-lock.json': new FileRef({ digest: 'package-lock' }),
|
|
||||||
};
|
|
||||||
const result = excludeLockFiles(files);
|
|
||||||
expect(result['frontend/pages/index.js']).toBeDefined();
|
|
||||||
expect(result['package-lock.json']).toBeUndefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should remove yarn.lock', () => {
|
|
||||||
const files = {
|
|
||||||
'frontend/pages/index.js': new FileRef({ digest: 'index' }),
|
|
||||||
'package.json': new FileRef({ digest: 'package' }),
|
|
||||||
'yarn.lock': new FileRef({ digest: 'yarn-lock' }),
|
|
||||||
};
|
|
||||||
const result = excludeLockFiles(files);
|
|
||||||
expect(result['frontend/pages/index.js']).toBeDefined();
|
|
||||||
expect(result['yarn.lock']).toBeUndefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should remove both package-lock.json and yarn.lock', () => {
|
|
||||||
const files = {
|
|
||||||
'frontend/pages/index.js': new FileRef({ digest: 'index' }),
|
|
||||||
'package.json': new FileRef({ digest: 'package' }),
|
|
||||||
'yarn.lock': new FileRef({ digest: 'yarn-lock' }),
|
|
||||||
'package-lock.json': new FileRef({ digest: 'package-lock' }),
|
|
||||||
};
|
|
||||||
const result = excludeLockFiles(files);
|
|
||||||
expect(result['frontend/pages/index.js']).toBeDefined();
|
|
||||||
expect(result['yarn.lock']).toBeUndefined();
|
|
||||||
expect(result['package-lock.json']).toBeUndefined();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('excludeStaticDirectory', () => {
|
|
||||||
it('should remove the /static directory files', () => {
|
|
||||||
const files = {
|
|
||||||
'frontend/pages/index.js': new FileRef({ digest: 'index' }),
|
|
||||||
'package.json': new FileRef({ digest: 'package' }),
|
|
||||||
'yarn.lock': new FileRef({ digest: 'yarn-lock' }),
|
|
||||||
'package-lock.json': new FileRef({ digest: 'package-lock' }),
|
|
||||||
'static/image.png': new FileRef({ digest: 'image' }),
|
|
||||||
};
|
|
||||||
const result = excludeStaticDirectory(files);
|
|
||||||
expect(result['frontend/pages/index.js']).toBeDefined();
|
|
||||||
expect(result['yarn.lock']).toBeDefined();
|
|
||||||
expect(result['package-lock.json']).toBeDefined();
|
|
||||||
expect(result['static/image.png']).toBeUndefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should remove the nested /static directory files', () => {
|
|
||||||
const files = {
|
|
||||||
'frontend/pages/index.js': new FileRef({ digest: 'index' }),
|
|
||||||
'package.json': new FileRef({ digest: 'package' }),
|
|
||||||
'yarn.lock': new FileRef({ digest: 'yarn-lock' }),
|
|
||||||
'package-lock.json': new FileRef({ digest: 'package-lock' }),
|
|
||||||
'static/images/png/image.png': new FileRef({ digest: 'image' }),
|
|
||||||
};
|
|
||||||
const result = excludeStaticDirectory(files);
|
|
||||||
expect(result['frontend/pages/index.js']).toBeDefined();
|
|
||||||
expect(result['yarn.lock']).toBeDefined();
|
|
||||||
expect(result['package-lock.json']).toBeDefined();
|
|
||||||
expect(result['static/images/png/image.png']).toBeUndefined();
|
|
||||||
});
|
|
||||||
});
|
|
||||||
|
|
||||||
describe('onlyStaticDirectory', () => {
|
|
||||||
it('should keep only /static directory files', () => {
|
|
||||||
const files = {
|
|
||||||
'frontend/pages/index.js': new FileRef({ digest: 'index' }),
|
|
||||||
'package.json': new FileRef({ digest: 'package' }),
|
|
||||||
'yarn.lock': new FileRef({ digest: 'yarn-lock' }),
|
|
||||||
'package-lock.json': new FileRef({ digest: 'package-lock' }),
|
|
||||||
'static/image.png': new FileRef({ digest: 'image' }),
|
|
||||||
};
|
|
||||||
const result = onlyStaticDirectory(files);
|
|
||||||
expect(result['frontend/pages/index.js']).toBeUndefined();
|
|
||||||
expect(result['yarn.lock']).toBeUndefined();
|
|
||||||
expect(result['package-lock.json']).toBeUndefined();
|
|
||||||
expect(result['static/image.png']).toBeDefined();
|
|
||||||
});
|
|
||||||
|
|
||||||
it('should keep nested /static directory files', () => {
|
|
||||||
const files = {
|
|
||||||
'frontend/pages/index.js': new FileRef({ digest: 'index' }),
|
|
||||||
'package.json': new FileRef({ digest: 'package' }),
|
|
||||||
'yarn.lock': new FileRef({ digest: 'yarn-lock' }),
|
|
||||||
'package-lock.json': new FileRef({ digest: 'package-lock' }),
|
|
||||||
'static/images/png/image.png': new FileRef({ digest: 'image' }),
|
|
||||||
};
|
|
||||||
const result = onlyStaticDirectory(files);
|
|
||||||
expect(result['frontend/pages/index.js']).toBeUndefined();
|
|
||||||
expect(result['yarn.lock']).toBeUndefined();
|
|
||||||
expect(result['package-lock.json']).toBeUndefined();
|
|
||||||
expect(result['static/images/png/image.png']).toBeDefined();
|
|
||||||
});
|
|
||||||
});
|
});
|
||||||
|
|
||||||
describe('normalizePackageJson', () => {
|
describe('normalizePackageJson', () => {
|
||||||
|
|||||||
Reference in New Issue
Block a user