mirror of
https://github.com/LukeHagar/vercel.git
synced 2025-12-26 11:49:15 +00:00
Compare commits
77 Commits
@now/node@
...
@now/node@
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
fb2029c464 | ||
|
|
3b15755054 | ||
|
|
4f65cc3aa8 | ||
|
|
9936e35280 | ||
|
|
a04fd242b8 | ||
|
|
17bc6174a9 | ||
|
|
a7c2d9648a | ||
|
|
faa5ea9e21 | ||
|
|
c52f30c898 | ||
|
|
d675edf1dc | ||
|
|
f85c4f496f | ||
|
|
d52d7904c2 | ||
|
|
79232024bd | ||
|
|
660b787bc3 | ||
|
|
2dbf983ddb | ||
|
|
0866ba9391 | ||
|
|
d259a722a0 | ||
|
|
bf77c51f64 | ||
|
|
062b78130c | ||
|
|
fa70bc50cb | ||
|
|
08e22b35d1 | ||
|
|
9d8f3315a1 | ||
|
|
a737a99a9d | ||
|
|
ee92d92c9f | ||
|
|
34d3ebd289 | ||
|
|
785f187e5d | ||
|
|
44449f474e | ||
|
|
f44dae7f39 | ||
|
|
06f973f641 | ||
|
|
47bb47804e | ||
|
|
5df692979a | ||
|
|
cd02d5390f | ||
|
|
c93fd416c4 | ||
|
|
431db7a62d | ||
|
|
6f86c70313 | ||
|
|
0923fc9200 | ||
|
|
787675f462 | ||
|
|
977615720a | ||
|
|
1a4e64cd27 | ||
|
|
c9fc255002 | ||
|
|
e83d4d4249 | ||
|
|
d2ca763079 | ||
|
|
2a95388f89 | ||
|
|
be9fedfdc4 | ||
|
|
f0dee65f69 | ||
|
|
5514753c07 | ||
|
|
7028556919 | ||
|
|
a1f24853fc | ||
|
|
9804e82f8f | ||
|
|
e96596634b | ||
|
|
495c193a40 | ||
|
|
9d77d89513 | ||
|
|
bcffe0db91 | ||
|
|
228892fe9f | ||
|
|
c7d434c81d | ||
|
|
66bf4aa25a | ||
|
|
f328a53a3f | ||
|
|
168b4cf2bb | ||
|
|
a91e3da7d8 | ||
|
|
756c452203 | ||
|
|
80db1937f8 | ||
|
|
d9df4b0929 | ||
|
|
08c52ebbf2 | ||
|
|
fe8ead391a | ||
|
|
5d45f2df31 | ||
|
|
1385c1ca48 | ||
|
|
4c63009123 | ||
|
|
389c85cef5 | ||
|
|
227c239bbc | ||
|
|
b93736132a | ||
|
|
26849422ff | ||
|
|
31dd1ca491 | ||
|
|
04a1ce89b3 | ||
|
|
3e247fd16d | ||
|
|
9e127421ad | ||
|
|
337d0cb1ed | ||
|
|
ad3cdf46f4 |
@@ -19,3 +19,9 @@ For the canary channel use:
|
||||
```
|
||||
yarn publish-canary
|
||||
```
|
||||
|
||||
CircleCI will take care of publishing the updated packages to npm from there.
|
||||
|
||||
If for some reason CircleCI fails to publish the npm package, you may do so
|
||||
manually by running `npm publish` from the package directory. Make sure to
|
||||
include the `--tag canary` parameter if you are publishing a canary release!
|
||||
|
||||
@@ -1,5 +1,6 @@
|
||||
{
|
||||
"name": "now-builders",
|
||||
"version": "0.0.0",
|
||||
"private": true,
|
||||
"license": "MIT",
|
||||
"dependencies": {
|
||||
@@ -30,6 +31,8 @@
|
||||
"@types/glob": "^7.1.1",
|
||||
"@types/multistream": "^2.1.1",
|
||||
"@types/node": "^10.12.8",
|
||||
"async-retry": "1.2.3",
|
||||
"buffer-replace": "^1.0.0",
|
||||
"eslint": "^5.9.0",
|
||||
"eslint-config-airbnb-base": "^13.1.0",
|
||||
"eslint-config-prettier": "^3.1.0",
|
||||
|
||||
2
packages/now-bash/.gitignore
vendored
Normal file
2
packages/now-bash/.gitignore
vendored
Normal file
@@ -0,0 +1,2 @@
|
||||
node_modules
|
||||
handler
|
||||
14
packages/now-bash/bootstrap
Executable file
14
packages/now-bash/bootstrap
Executable file
@@ -0,0 +1,14 @@
|
||||
#!/bin/bash
|
||||
set -euo pipefail
|
||||
cd "$LAMBDA_TASK_ROOT"
|
||||
|
||||
# Configure `import`
|
||||
export IMPORT_CACHE="$LAMBDA_TASK_ROOT/.import-cache"
|
||||
export PATH="$IMPORT_CACHE/bin:$PATH"
|
||||
|
||||
# Load `import` and runtime
|
||||
. "$(which import)"
|
||||
. "$IMPORT_CACHE/runtime.sh"
|
||||
|
||||
# Load user code and process events in a loop forever
|
||||
_lambda_runtime_init
|
||||
49
packages/now-bash/builder.sh
Executable file
49
packages/now-bash/builder.sh
Executable file
@@ -0,0 +1,49 @@
|
||||
#!/bin/bash
|
||||
set -euo pipefail
|
||||
|
||||
# `import` debug logs are always enabled during build
|
||||
export IMPORT_DEBUG=1
|
||||
|
||||
# Install `import`
|
||||
IMPORT_BIN="$IMPORT_CACHE/bin/import"
|
||||
mkdir -p "$(dirname "$IMPORT_BIN")"
|
||||
curl -sfLS https://import.pw > "$IMPORT_BIN"
|
||||
chmod +x "$IMPORT_BIN"
|
||||
|
||||
# For now only the entrypoint file is copied into the lambda
|
||||
mkdir -p "$(dirname "$ENTRYPOINT")"
|
||||
cp "$SRC/$ENTRYPOINT" "$ENTRYPOINT"
|
||||
|
||||
# Copy in the runtime
|
||||
cp "$BUILDER/runtime.sh" "$IMPORT_CACHE"
|
||||
cp "$BUILDER/bootstrap" .
|
||||
|
||||
# Load `import`
|
||||
. "$(which import)"
|
||||
|
||||
# Cache runtime and user dependencies
|
||||
echo "Caching imports in \"$ENTRYPOINT\"…"
|
||||
. "$IMPORT_CACHE/runtime.sh"
|
||||
. "$ENTRYPOINT"
|
||||
echo "Done caching imports"
|
||||
|
||||
# Run user build script
|
||||
if declare -f build > /dev/null; then
|
||||
echo "Running \`build\` function in \"$ENTRYPOINT\"…"
|
||||
build "$@"
|
||||
fi
|
||||
|
||||
# Ensure the entrypoint defined a `handler` function
|
||||
if ! declare -f handler > /dev/null; then
|
||||
echo "ERROR: A \`handler\` function must be defined in \"$ENTRYPOINT\"!" >&2
|
||||
exit 1
|
||||
fi
|
||||
|
||||
# Show a tree of the final lambda build
|
||||
show_tree() {
|
||||
import "static-binaries@0.0.6"
|
||||
static_binaries tree
|
||||
echo "Final lambda file tree:"
|
||||
tree -a .
|
||||
}
|
||||
IMPORT_DEBUG= IMPORT_CACHE="$(mktemp -d)" show_tree
|
||||
55
packages/now-bash/index.js
Normal file
55
packages/now-bash/index.js
Normal file
@@ -0,0 +1,55 @@
|
||||
const execa = require('execa');
|
||||
const { join } = require('path');
|
||||
const snakeCase = require('snake-case');
|
||||
const glob = require('@now/build-utils/fs/glob');
|
||||
const download = require('@now/build-utils/fs/download');
|
||||
const { createLambda } = require('@now/build-utils/lambda');
|
||||
const getWritableDirectory = require('@now/build-utils/fs/get-writable-directory');
|
||||
|
||||
exports.config = {
|
||||
maxLambdaSize: '10mb',
|
||||
};
|
||||
|
||||
exports.analyze = ({ files, entrypoint }) => files[entrypoint].digest;
|
||||
|
||||
exports.build = async ({ files, entrypoint, config }) => {
|
||||
const srcDir = await getWritableDirectory();
|
||||
const workDir = await getWritableDirectory();
|
||||
|
||||
console.log('downloading files...');
|
||||
await download(files, srcDir);
|
||||
|
||||
const configEnv = Object.keys(config).reduce((o, v) => {
|
||||
o[`IMPORT_${snakeCase(v).toUpperCase()}`] = config[v]; // eslint-disable-line no-param-reassign
|
||||
return o;
|
||||
}, {});
|
||||
|
||||
const IMPORT_CACHE = `${workDir}/.import-cache`;
|
||||
const env = Object.assign({}, process.env, configEnv, {
|
||||
PATH: `${IMPORT_CACHE}/bin:${process.env.PATH}`,
|
||||
IMPORT_CACHE,
|
||||
SRC: srcDir,
|
||||
BUILDER: __dirname,
|
||||
ENTRYPOINT: entrypoint,
|
||||
});
|
||||
|
||||
const builderPath = join(__dirname, 'builder.sh');
|
||||
|
||||
await execa(builderPath, [entrypoint], {
|
||||
env,
|
||||
cwd: workDir,
|
||||
});
|
||||
|
||||
const lambda = await createLambda({
|
||||
files: await glob('**', workDir),
|
||||
handler: entrypoint, // not actually used in `bootstrap`
|
||||
runtime: 'provided',
|
||||
environment: Object.assign({}, configEnv, {
|
||||
SCRIPT_FILENAME: entrypoint,
|
||||
}),
|
||||
});
|
||||
|
||||
return {
|
||||
[entrypoint]: lambda,
|
||||
};
|
||||
};
|
||||
22
packages/now-bash/package.json
Normal file
22
packages/now-bash/package.json
Normal file
@@ -0,0 +1,22 @@
|
||||
{
|
||||
"name": "@now/bash",
|
||||
"version": "0.0.4-canary.0",
|
||||
"description": "Now 2.0 builder for HTTP endpoints written in Bash",
|
||||
"main": "index.js",
|
||||
"author": "Nathan Rajlich <nate@zeit.co>",
|
||||
"license": "MIT",
|
||||
"files": [
|
||||
"builder.sh",
|
||||
"runtime.sh",
|
||||
"bootstrap",
|
||||
"index.js",
|
||||
"package.json"
|
||||
],
|
||||
"dependencies": {
|
||||
"execa": "^1.0.0",
|
||||
"snake-case": "^2.1.0"
|
||||
},
|
||||
"peerDependencies": {
|
||||
"@now/build-utils": ">=0.0.1"
|
||||
}
|
||||
}
|
||||
115
packages/now-bash/runtime.sh
Normal file
115
packages/now-bash/runtime.sh
Normal file
@@ -0,0 +1,115 @@
|
||||
import "static-binaries@0.0.6"
|
||||
static_binaries jq
|
||||
|
||||
# These get reset upon each request
|
||||
_STATUS_CODE="$(mktemp)"
|
||||
_HEADERS="$(mktemp)"
|
||||
|
||||
_lambda_runtime_api() {
|
||||
local endpoint="$1"
|
||||
shift
|
||||
curl -sfLS "http://$AWS_LAMBDA_RUNTIME_API/2018-06-01/runtime/$endpoint" "$@"
|
||||
}
|
||||
|
||||
_lambda_runtime_init() {
|
||||
# Initialize user code
|
||||
. "$SCRIPT_FILENAME" || {
|
||||
local exit_code="$?"
|
||||
local error
|
||||
error='{"exitCode":'"$exit_code"'}'
|
||||
_lambda_runtime_api "init/error" -X POST -d "$error"
|
||||
exit "$EXIT_CODE"
|
||||
}
|
||||
|
||||
# Process events
|
||||
while true; do _lambda_runtime_next; done
|
||||
}
|
||||
|
||||
_lambda_runtime_next() {
|
||||
echo 200 > "$_STATUS_CODE"
|
||||
echo '{"content-type":"text/plain; charset=utf8"}' > "$_HEADERS"
|
||||
|
||||
local headers
|
||||
headers="$(mktemp)"
|
||||
|
||||
# Get an event
|
||||
local event
|
||||
event="$(mktemp)"
|
||||
_lambda_runtime_api invocation/next -D "$headers" | jq -r '.body' > "$event"
|
||||
|
||||
local request_id
|
||||
request_id="$(grep -Fi Lambda-Runtime-Aws-Request-Id "$headers" | tr -d '[:space:]' | cut -d: -f2)"
|
||||
echo "Request-Id: $request_id" >&2
|
||||
rm -f "$headers"
|
||||
|
||||
# Execute the handler function from the script
|
||||
local body
|
||||
body="$(mktemp)"
|
||||
|
||||
local exit_code=0
|
||||
REQUEST="$event"
|
||||
|
||||
# Stdin of the `handler` function is the HTTP request body.
|
||||
# Need to use a fifo here instead of bash <() because Lambda
|
||||
# errors with "/dev/fd/63 not found" for some reason :/
|
||||
local stdin
|
||||
stdin="$(mktemp --dry-run)"
|
||||
mkfifo "$stdin"
|
||||
_lambda_runtime_body "$event" > "$stdin" &
|
||||
|
||||
handler "$event" < "$stdin" > "$body" || exit_code="$?"
|
||||
rm -f "$event" "$stdin"
|
||||
|
||||
if [ "$exit_code" -eq 0 ]; then
|
||||
# Send the response
|
||||
local response
|
||||
response="$(jq -cnMr \
|
||||
--arg statusCode "$(cat "$_STATUS_CODE")" \
|
||||
--argjson headers "$(cat "$_HEADERS")" \
|
||||
--arg body "$(base64 --wrap=0 < "$body")" \
|
||||
'{statusCode:$statusCode|tonumber, headers:$headers, encoding:"base64", body:$body}')"
|
||||
rm -f "$body" "$_HEADERS"
|
||||
_lambda_runtime_api "invocation/$request_id/response" -X POST -d "$response"
|
||||
else
|
||||
local error
|
||||
error='{"exitCode":'"$exit_code"'}'
|
||||
_lambda_runtime_api "invocation/$request_id/error" -X POST -d "$error"
|
||||
fi
|
||||
}
|
||||
|
||||
_lambda_runtime_body() {
|
||||
if [ "$(jq -r '.body | type' < "$1")" = "string" ]; then
|
||||
if [ "$(jq -r '.encoding' < "$1")" = "base64" ]; then
|
||||
jq -r '.body' < "$1" | base64 -d
|
||||
else
|
||||
# assume plain-text body
|
||||
jq -r '.body' < "$1"
|
||||
fi
|
||||
fi
|
||||
}
|
||||
|
||||
|
||||
# Set the response status code.
|
||||
http_response_code() {
|
||||
echo "$1" > "$_STATUS_CODE"
|
||||
}
|
||||
|
||||
# Sets a response header.
|
||||
# Overrides existing header if it has already been set.
|
||||
http_response_header() {
|
||||
local name="$1"
|
||||
local value="$2"
|
||||
local tmp
|
||||
tmp="$(mktemp)"
|
||||
jq --arg name "$name" --arg value "$value" '.[$name] = $value' < "$_HEADERS" > "$tmp"
|
||||
mv -f "$tmp" "$_HEADERS"
|
||||
}
|
||||
|
||||
http_response_redirect() {
|
||||
http_response_code "${2:-302}"
|
||||
http_response_header "location" "$1"
|
||||
}
|
||||
|
||||
http_response_json() {
|
||||
http_response_header "content-type" "application/json; charset=utf8"
|
||||
}
|
||||
146
packages/now-bash/yarn.lock
Normal file
146
packages/now-bash/yarn.lock
Normal file
@@ -0,0 +1,146 @@
|
||||
# THIS IS AN AUTOGENERATED FILE. DO NOT EDIT THIS FILE DIRECTLY.
|
||||
# yarn lockfile v1
|
||||
|
||||
|
||||
cross-spawn@^6.0.0:
|
||||
version "6.0.5"
|
||||
resolved "https://registry.yarnpkg.com/cross-spawn/-/cross-spawn-6.0.5.tgz#4a5ec7c64dfae22c3a14124dbacdee846d80cbc4"
|
||||
integrity sha512-eTVLrBSt7fjbDygz805pMnstIs2VTBNkRm0qxZd+M7A5XDdxVRWO5MxGBXZhjY4cqLYLdtrGqRf8mBPmzwSpWQ==
|
||||
dependencies:
|
||||
nice-try "^1.0.4"
|
||||
path-key "^2.0.1"
|
||||
semver "^5.5.0"
|
||||
shebang-command "^1.2.0"
|
||||
which "^1.2.9"
|
||||
|
||||
end-of-stream@^1.1.0:
|
||||
version "1.4.1"
|
||||
resolved "https://registry.yarnpkg.com/end-of-stream/-/end-of-stream-1.4.1.tgz#ed29634d19baba463b6ce6b80a37213eab71ec43"
|
||||
integrity sha512-1MkrZNvWTKCaigbn+W15elq2BB/L22nqrSY5DKlo3X6+vclJm8Bb5djXJBmEX6fS3+zCh/F4VBK5Z2KxJt4s2Q==
|
||||
dependencies:
|
||||
once "^1.4.0"
|
||||
|
||||
execa@^1.0.0:
|
||||
version "1.0.0"
|
||||
resolved "https://registry.yarnpkg.com/execa/-/execa-1.0.0.tgz#c6236a5bb4df6d6f15e88e7f017798216749ddd8"
|
||||
integrity sha512-adbxcyWV46qiHyvSp50TKt05tB4tK3HcmF7/nxfAdhnox83seTDbwnaqKO4sXRy7roHAIFqJP/Rw/AuEbX61LA==
|
||||
dependencies:
|
||||
cross-spawn "^6.0.0"
|
||||
get-stream "^4.0.0"
|
||||
is-stream "^1.1.0"
|
||||
npm-run-path "^2.0.0"
|
||||
p-finally "^1.0.0"
|
||||
signal-exit "^3.0.0"
|
||||
strip-eof "^1.0.0"
|
||||
|
||||
get-stream@^4.0.0:
|
||||
version "4.1.0"
|
||||
resolved "https://registry.yarnpkg.com/get-stream/-/get-stream-4.1.0.tgz#c1b255575f3dc21d59bfc79cd3d2b46b1c3a54b5"
|
||||
integrity sha512-GMat4EJ5161kIy2HevLlr4luNjBgvmj413KaQA7jt4V8B4RDsfpHk7WQ9GVqfYyyx8OS/L66Kox+rJRNklLK7w==
|
||||
dependencies:
|
||||
pump "^3.0.0"
|
||||
|
||||
is-stream@^1.1.0:
|
||||
version "1.1.0"
|
||||
resolved "https://registry.yarnpkg.com/is-stream/-/is-stream-1.1.0.tgz#12d4a3dd4e68e0b79ceb8dbc84173ae80d91ca44"
|
||||
integrity sha1-EtSj3U5o4Lec6428hBc66A2RykQ=
|
||||
|
||||
isexe@^2.0.0:
|
||||
version "2.0.0"
|
||||
resolved "https://registry.yarnpkg.com/isexe/-/isexe-2.0.0.tgz#e8fbf374dc556ff8947a10dcb0572d633f2cfa10"
|
||||
integrity sha1-6PvzdNxVb/iUehDcsFctYz8s+hA=
|
||||
|
||||
lower-case@^1.1.1:
|
||||
version "1.1.4"
|
||||
resolved "https://registry.yarnpkg.com/lower-case/-/lower-case-1.1.4.tgz#9a2cabd1b9e8e0ae993a4bf7d5875c39c42e8eac"
|
||||
integrity sha1-miyr0bno4K6ZOkv31YdcOcQujqw=
|
||||
|
||||
nice-try@^1.0.4:
|
||||
version "1.0.5"
|
||||
resolved "https://registry.yarnpkg.com/nice-try/-/nice-try-1.0.5.tgz#a3378a7696ce7d223e88fc9b764bd7ef1089e366"
|
||||
integrity sha512-1nh45deeb5olNY7eX82BkPO7SSxR5SSYJiPTrTdFUVYwAl8CKMA5N9PjTYkHiRjisVcxcQ1HXdLhx2qxxJzLNQ==
|
||||
|
||||
no-case@^2.2.0:
|
||||
version "2.3.2"
|
||||
resolved "https://registry.yarnpkg.com/no-case/-/no-case-2.3.2.tgz#60b813396be39b3f1288a4c1ed5d1e7d28b464ac"
|
||||
integrity sha512-rmTZ9kz+f3rCvK2TD1Ue/oZlns7OGoIWP4fc3llxxRXlOkHKoWPPWJOfFYpITabSow43QJbRIoHQXtt10VldyQ==
|
||||
dependencies:
|
||||
lower-case "^1.1.1"
|
||||
|
||||
npm-run-path@^2.0.0:
|
||||
version "2.0.2"
|
||||
resolved "https://registry.yarnpkg.com/npm-run-path/-/npm-run-path-2.0.2.tgz#35a9232dfa35d7067b4cb2ddf2357b1871536c5f"
|
||||
integrity sha1-NakjLfo11wZ7TLLd8jV7GHFTbF8=
|
||||
dependencies:
|
||||
path-key "^2.0.0"
|
||||
|
||||
once@^1.3.1, once@^1.4.0:
|
||||
version "1.4.0"
|
||||
resolved "https://registry.yarnpkg.com/once/-/once-1.4.0.tgz#583b1aa775961d4b113ac17d9c50baef9dd76bd1"
|
||||
integrity sha1-WDsap3WWHUsROsF9nFC6753Xa9E=
|
||||
dependencies:
|
||||
wrappy "1"
|
||||
|
||||
p-finally@^1.0.0:
|
||||
version "1.0.0"
|
||||
resolved "https://registry.yarnpkg.com/p-finally/-/p-finally-1.0.0.tgz#3fbcfb15b899a44123b34b6dcc18b724336a2cae"
|
||||
integrity sha1-P7z7FbiZpEEjs0ttzBi3JDNqLK4=
|
||||
|
||||
path-key@^2.0.0, path-key@^2.0.1:
|
||||
version "2.0.1"
|
||||
resolved "https://registry.yarnpkg.com/path-key/-/path-key-2.0.1.tgz#411cadb574c5a140d3a4b1910d40d80cc9f40b40"
|
||||
integrity sha1-QRyttXTFoUDTpLGRDUDYDMn0C0A=
|
||||
|
||||
pump@^3.0.0:
|
||||
version "3.0.0"
|
||||
resolved "https://registry.yarnpkg.com/pump/-/pump-3.0.0.tgz#b4a2116815bde2f4e1ea602354e8c75565107a64"
|
||||
integrity sha512-LwZy+p3SFs1Pytd/jYct4wpv49HiYCqd9Rlc5ZVdk0V+8Yzv6jR5Blk3TRmPL1ft69TxP0IMZGJ+WPFU2BFhww==
|
||||
dependencies:
|
||||
end-of-stream "^1.1.0"
|
||||
once "^1.3.1"
|
||||
|
||||
semver@^5.5.0:
|
||||
version "5.6.0"
|
||||
resolved "https://registry.yarnpkg.com/semver/-/semver-5.6.0.tgz#7e74256fbaa49c75aa7c7a205cc22799cac80004"
|
||||
integrity sha512-RS9R6R35NYgQn++fkDWaOmqGoj4Ek9gGs+DPxNUZKuwE183xjJroKvyo1IzVFeXvUrvmALy6FWD5xrdJT25gMg==
|
||||
|
||||
shebang-command@^1.2.0:
|
||||
version "1.2.0"
|
||||
resolved "https://registry.yarnpkg.com/shebang-command/-/shebang-command-1.2.0.tgz#44aac65b695b03398968c39f363fee5deafdf1ea"
|
||||
integrity sha1-RKrGW2lbAzmJaMOfNj/uXer98eo=
|
||||
dependencies:
|
||||
shebang-regex "^1.0.0"
|
||||
|
||||
shebang-regex@^1.0.0:
|
||||
version "1.0.0"
|
||||
resolved "https://registry.yarnpkg.com/shebang-regex/-/shebang-regex-1.0.0.tgz#da42f49740c0b42db2ca9728571cb190c98efea3"
|
||||
integrity sha1-2kL0l0DAtC2yypcoVxyxkMmO/qM=
|
||||
|
||||
signal-exit@^3.0.0:
|
||||
version "3.0.2"
|
||||
resolved "https://registry.yarnpkg.com/signal-exit/-/signal-exit-3.0.2.tgz#b5fdc08f1287ea1178628e415e25132b73646c6d"
|
||||
integrity sha1-tf3AjxKH6hF4Yo5BXiUTK3NkbG0=
|
||||
|
||||
snake-case@^2.1.0:
|
||||
version "2.1.0"
|
||||
resolved "https://registry.yarnpkg.com/snake-case/-/snake-case-2.1.0.tgz#41bdb1b73f30ec66a04d4e2cad1b76387d4d6d9f"
|
||||
integrity sha1-Qb2xtz8w7GagTU4srRt2OH1NbZ8=
|
||||
dependencies:
|
||||
no-case "^2.2.0"
|
||||
|
||||
strip-eof@^1.0.0:
|
||||
version "1.0.0"
|
||||
resolved "https://registry.yarnpkg.com/strip-eof/-/strip-eof-1.0.0.tgz#bb43ff5598a6eb05d89b59fcd129c983313606bf"
|
||||
integrity sha1-u0P/VZim6wXYm1n80SnJgzE2Br8=
|
||||
|
||||
which@^1.2.9:
|
||||
version "1.3.1"
|
||||
resolved "https://registry.yarnpkg.com/which/-/which-1.3.1.tgz#a45043d54f5805316da8d62f9f50918d3da70b0a"
|
||||
integrity sha512-HxJdYWq1MTIQbJ3nw0cqssHoTNU267KlrDuGZ1WYlxDStUtKUhOaJmh112/TZmHxxUfuJqPXSOm7tDyas0OSIQ==
|
||||
dependencies:
|
||||
isexe "^2.0.0"
|
||||
|
||||
wrappy@1:
|
||||
version "1.0.2"
|
||||
resolved "https://registry.yarnpkg.com/wrappy/-/wrappy-1.0.2.tgz#b5243d8f3ec1aa35f1364605bc0d1036e30ab69f"
|
||||
integrity sha1-tSQ9jz7BqjXxNkYFvA0QNuMKtp8=
|
||||
83
packages/now-build-utils/fs/bootstrap-yarn.js
vendored
Normal file
83
packages/now-build-utils/fs/bootstrap-yarn.js
vendored
Normal file
@@ -0,0 +1,83 @@
|
||||
const MemoryFileSystem = require('memory-fs');
|
||||
const fs = require('fs');
|
||||
const path = require('path');
|
||||
const { spawnSync } = require('child_process');
|
||||
|
||||
const yarnPath = spawnSync('which', ['yarn'])
|
||||
.stdout.toString()
|
||||
.trim();
|
||||
|
||||
const cachePath = spawnSync(yarnPath, ['cache', 'dir'])
|
||||
.stdout.toString()
|
||||
.trim();
|
||||
|
||||
spawnSync(yarnPath, ['cache', 'clean']);
|
||||
const vfs = new MemoryFileSystem();
|
||||
|
||||
function isOutsideCachePath(filename) {
|
||||
const relative = path.relative(cachePath, filename);
|
||||
return relative.startsWith('..');
|
||||
}
|
||||
|
||||
const saveCreateWriteStream = fs.createWriteStream;
|
||||
fs.createWriteStream = (...args) => {
|
||||
const filename = args[0];
|
||||
if (isOutsideCachePath(filename)) {
|
||||
return saveCreateWriteStream.call(fs, ...args);
|
||||
}
|
||||
|
||||
vfs.mkdirpSync(path.dirname(filename));
|
||||
fs.writeFileSync(filename, Buffer.alloc(0));
|
||||
const stream = vfs.createWriteStream(...args);
|
||||
|
||||
stream.on('finish', () => {
|
||||
setTimeout(() => {
|
||||
stream.emit('close');
|
||||
});
|
||||
});
|
||||
|
||||
return stream;
|
||||
};
|
||||
|
||||
const saveReadFile = fs.readFile;
|
||||
fs.readFile = (...args) => {
|
||||
const filename = args[0];
|
||||
if (isOutsideCachePath(filename)) {
|
||||
return saveReadFile.call(fs, ...args);
|
||||
}
|
||||
|
||||
const callback = args[args.length - 1];
|
||||
return vfs.readFile(...args.slice(0, -1), (error, result) => {
|
||||
if (error) {
|
||||
saveReadFile.call(fs, ...args);
|
||||
return;
|
||||
}
|
||||
|
||||
callback(error, result);
|
||||
});
|
||||
};
|
||||
|
||||
const saveCopyFile = fs.copyFile;
|
||||
fs.copyFile = (...args) => {
|
||||
const src = args[0];
|
||||
if (isOutsideCachePath(src)) {
|
||||
return saveCopyFile.call(fs, ...args);
|
||||
}
|
||||
|
||||
const dest = args[1];
|
||||
const callback = args[args.length - 1];
|
||||
const buffer = vfs.readFileSync(src);
|
||||
return fs.writeFile(dest, buffer, callback);
|
||||
};
|
||||
|
||||
const saveWriteFile = fs.writeFile;
|
||||
fs.writeFile = (...args) => {
|
||||
const filename = args[0];
|
||||
if (isOutsideCachePath(filename)) {
|
||||
return saveWriteFile.call(fs, ...args);
|
||||
}
|
||||
|
||||
return vfs.writeFile(...args);
|
||||
};
|
||||
|
||||
require(yarnPath);
|
||||
@@ -35,7 +35,7 @@ async function scanParentDirs(destPath, scriptName) {
|
||||
// eslint-disable-next-line no-await-in-loop
|
||||
const packageJson = JSON.parse(await fs.readFile(packageJsonPath));
|
||||
hasScript = Boolean(
|
||||
packageJson.scripts && packageJson.scripts[scriptName],
|
||||
packageJson.scripts && scriptName && packageJson.scripts[scriptName],
|
||||
);
|
||||
// eslint-disable-next-line no-await-in-loop
|
||||
hasPackageLockJson = await fs.exists(
|
||||
@@ -63,6 +63,15 @@ async function runNpmInstall(destPath, args = []) {
|
||||
commandArgs = args.filter(a => a !== '--prefer-offline');
|
||||
await spawnAsync('npm', ['install'].concat(commandArgs), destPath);
|
||||
await spawnAsync('npm', ['cache', 'clean', '--force'], destPath);
|
||||
} else if (process.env.AWS_EXECUTION_ENV) {
|
||||
console.log('using memory-fs for yarn cache');
|
||||
await spawnAsync(
|
||||
'node',
|
||||
[path.join(__dirname, 'bootstrap-yarn.js'), '--cwd', destPath].concat(
|
||||
commandArgs,
|
||||
),
|
||||
destPath,
|
||||
);
|
||||
} else {
|
||||
await spawnAsync('yarn', ['--cwd', destPath].concat(commandArgs), destPath);
|
||||
await spawnAsync('yarn', ['cache', 'clean'], destPath);
|
||||
|
||||
@@ -1,4 +1,5 @@
|
||||
const assert = require('assert');
|
||||
const Sema = require('async-sema');
|
||||
const { ZipFile } = require('yazl');
|
||||
const streamToBuffer = require('./fs/stream-to-buffer.js');
|
||||
|
||||
@@ -14,6 +15,7 @@ class Lambda {
|
||||
}
|
||||
}
|
||||
|
||||
const sema = new Sema(10);
|
||||
const mtime = new Date(1540000000000);
|
||||
|
||||
async function createLambda({
|
||||
@@ -23,24 +25,33 @@ async function createLambda({
|
||||
assert(typeof handler === 'string', '"handler" is not a string');
|
||||
assert(typeof runtime === 'string', '"runtime" is not a string');
|
||||
assert(typeof environment === 'object', '"environment" is not an object');
|
||||
const zipFile = new ZipFile();
|
||||
|
||||
Object.keys(files)
|
||||
.sort()
|
||||
.forEach((name) => {
|
||||
const file = files[name];
|
||||
const stream = file.toStream();
|
||||
zipFile.addReadStream(stream, name, { mode: file.mode, mtime });
|
||||
await sema.acquire();
|
||||
try {
|
||||
const zipFile = new ZipFile();
|
||||
const zipBuffer = await new Promise((resolve, reject) => {
|
||||
Object.keys(files)
|
||||
.sort()
|
||||
.forEach((name) => {
|
||||
const file = files[name];
|
||||
const stream = file.toStream();
|
||||
stream.on('error', reject);
|
||||
zipFile.addReadStream(stream, name, { mode: file.mode, mtime });
|
||||
});
|
||||
|
||||
zipFile.end();
|
||||
streamToBuffer(zipFile.outputStream).then(resolve).catch(reject);
|
||||
});
|
||||
|
||||
zipFile.end();
|
||||
const zipBuffer = await streamToBuffer(zipFile.outputStream);
|
||||
return new Lambda({
|
||||
zipBuffer,
|
||||
handler,
|
||||
runtime,
|
||||
environment,
|
||||
});
|
||||
return new Lambda({
|
||||
zipBuffer,
|
||||
handler,
|
||||
runtime,
|
||||
environment,
|
||||
});
|
||||
} finally {
|
||||
sema.release();
|
||||
}
|
||||
}
|
||||
|
||||
module.exports = {
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@now/build-utils",
|
||||
"version": "0.4.30",
|
||||
"version": "0.4.32-canary.1",
|
||||
"dependencies": {
|
||||
"async-retry": "1.2.3",
|
||||
"async-sema": "2.1.4",
|
||||
@@ -8,6 +8,7 @@
|
||||
"fs-extra": "7.0.0",
|
||||
"glob": "7.1.3",
|
||||
"into-stream": "4.0.0",
|
||||
"memory-fs": "0.4.1",
|
||||
"multistream": "2.1.1",
|
||||
"node-fetch": "2.2.0",
|
||||
"yazl": "2.4.3"
|
||||
|
||||
@@ -8,6 +8,7 @@ const {
|
||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||
|
||||
jest.setTimeout(2 * 60 * 1000);
|
||||
const builderUrl = '@canary';
|
||||
let buildUtilsUrl;
|
||||
|
||||
beforeAll(async () => {
|
||||
@@ -25,8 +26,11 @@ for (const fixture of fs.readdirSync(fixturesPath)) {
|
||||
// eslint-disable-next-line no-loop-func
|
||||
it(`should build ${fixture}`, async () => {
|
||||
await expect(
|
||||
testDeployment({ buildUtilsUrl }, path.join(fixturesPath, fixture)),
|
||||
).resolves.toBe(undefined);
|
||||
testDeployment(
|
||||
{ builderUrl, buildUtilsUrl },
|
||||
path.join(fixturesPath, fixture),
|
||||
),
|
||||
).resolves.toBeDefined();
|
||||
});
|
||||
}
|
||||
|
||||
@@ -43,11 +47,17 @@ for (const builder of buildersToTestWith) {
|
||||
|
||||
// eslint-disable-next-line no-restricted-syntax
|
||||
for (const fixture of fs.readdirSync(fixturesPath2)) {
|
||||
// eslint-disable-next-line no-loop-func
|
||||
it(`should build ${builder}/${fixture}`, async () => {
|
||||
await expect(
|
||||
testDeployment({ buildUtilsUrl }, path.join(fixturesPath2, fixture)),
|
||||
).resolves.toBe(undefined);
|
||||
});
|
||||
// don't run all foreign fixtures, just some
|
||||
if (['01-cowsay', '03-env-vars'].includes(fixture)) {
|
||||
// eslint-disable-next-line no-loop-func
|
||||
it(`should build ${builder}/${fixture}`, async () => {
|
||||
await expect(
|
||||
testDeployment(
|
||||
{ builderUrl, buildUtilsUrl },
|
||||
path.join(fixturesPath2, fixture),
|
||||
),
|
||||
).resolves.toBeDefined();
|
||||
});
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,152 +1,36 @@
|
||||
package main
|
||||
|
||||
import (
|
||||
"os"
|
||||
"fmt"
|
||||
"net"
|
||||
"strings"
|
||||
"io/ioutil"
|
||||
now "../../utils/go/bridge"
|
||||
"net/http"
|
||||
"net/http/cgi"
|
||||
"os"
|
||||
"path/filepath"
|
||||
"encoding/json"
|
||||
b64 "encoding/base64"
|
||||
"github.com/aws/aws-lambda-go/events"
|
||||
"github.com/aws/aws-lambda-go/lambda"
|
||||
)
|
||||
|
||||
type Request struct {
|
||||
Host string `json:"host"`
|
||||
Path string `json:"path"`
|
||||
Method string `json:"method"`
|
||||
Headers map[string]string `json:"headers"`
|
||||
Encoding string `json:"encoding,omitempty"`
|
||||
Body string `json:"body"`
|
||||
}
|
||||
|
||||
type Response struct {
|
||||
StatusCode int `json:"statusCode"`
|
||||
Headers map[string]string `json:"headers"`
|
||||
Encoding string `json:"encoding,omitemtpy"`
|
||||
Body string `json:"body"`
|
||||
}
|
||||
|
||||
type ResponseError struct {
|
||||
Code string `json:"code"`
|
||||
Message string `json:"message"`
|
||||
}
|
||||
|
||||
type ResponseErrorWrapper struct {
|
||||
Error ResponseError `json:"error"`
|
||||
}
|
||||
|
||||
type CgiHandler struct {
|
||||
http.Handler
|
||||
Dir string
|
||||
Script string
|
||||
}
|
||||
|
||||
func createErrorResponse(message string, code string, statusCode int) (Response, error) {
|
||||
obj := ResponseErrorWrapper{
|
||||
Error: ResponseError{
|
||||
Code: code,
|
||||
Message: message,
|
||||
},
|
||||
}
|
||||
|
||||
body, _ := json.Marshal(obj)
|
||||
|
||||
return Response{
|
||||
StatusCode: statusCode,
|
||||
Headers: map[string]string{
|
||||
"Content-Type": "application/json",
|
||||
},
|
||||
Body: string(body),
|
||||
}, nil
|
||||
}
|
||||
|
||||
func (h *CgiHandler) ServeHTTP(w http.ResponseWriter, r *http.Request) {
|
||||
cgih := cgi.Handler{
|
||||
handler := cgi.Handler{
|
||||
Path: h.Script,
|
||||
Root: "/" + h.Script,
|
||||
Dir: h.Dir,
|
||||
Env: []string{"SERVER_PORT=443", "HTTPS=on", "SERVER_SOFTWARE=@now/cgi"},
|
||||
Dir: h.Dir,
|
||||
Env: []string{
|
||||
"HTTPS=on",
|
||||
"SERVER_PORT=443",
|
||||
"SERVER_SOFTWARE=@now/cgi",
|
||||
},
|
||||
}
|
||||
cgih.ServeHTTP(w, r)
|
||||
handler.ServeHTTP(w, r)
|
||||
}
|
||||
|
||||
func main() {
|
||||
l, err := net.Listen("tcp", ":0")
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
|
||||
workdir, _ := filepath.Abs(".")
|
||||
script := os.Getenv("SCRIPT_FILENAME")
|
||||
h := &CgiHandler{nil, workdir, script}
|
||||
|
||||
http.Handle("/", h)
|
||||
go http.Serve(l, nil)
|
||||
|
||||
handler := func(_req events.APIGatewayProxyRequest) (Response, error) {
|
||||
var req Request
|
||||
|
||||
err := json.Unmarshal([]byte(_req.Body), &req)
|
||||
|
||||
if err != nil {
|
||||
fmt.Println(err)
|
||||
return createErrorResponse("Invalid payload", "bad_request", 400)
|
||||
}
|
||||
|
||||
if req.Encoding == "base64" {
|
||||
decoded, _ := b64.StdEncoding.DecodeString(req.Body)
|
||||
req.Body = string(decoded)
|
||||
}
|
||||
|
||||
url := "http://" + l.Addr().String() + req.Path
|
||||
|
||||
internalReq, err := http.NewRequest(req.Method, url, strings.NewReader(req.Body))
|
||||
if err != nil {
|
||||
fmt.Println(err)
|
||||
return createErrorResponse("Bad gateway internal req failed", "bad_gateway", 502)
|
||||
}
|
||||
|
||||
for k, v := range req.Headers {
|
||||
internalReq.Header.Add(k, v)
|
||||
if strings.ToLower(k) == "host" {
|
||||
internalReq.Host = v
|
||||
}
|
||||
}
|
||||
|
||||
client := &http.Client{}
|
||||
internalRes, err := client.Do(internalReq)
|
||||
if err != nil {
|
||||
fmt.Println(err)
|
||||
return createErrorResponse("Bad gateway internal req Do failed", "bad_gateway", 502)
|
||||
}
|
||||
defer internalRes.Body.Close()
|
||||
|
||||
resHeaders := make(map[string]string, len(internalRes.Header))
|
||||
for k, v := range internalRes.Header {
|
||||
// FIXME: support multiple values via concatenating with ','
|
||||
// see RFC 7230, section 3.2.2
|
||||
resHeaders[k] = v[0]
|
||||
}
|
||||
|
||||
bodyBytes, err := ioutil.ReadAll(internalRes.Body)
|
||||
if err != nil {
|
||||
return createErrorResponse("Bad gateway ReadAll bytes from response failed", "bad_gateway", 502)
|
||||
}
|
||||
|
||||
resBody := b64.StdEncoding.EncodeToString(bodyBytes)
|
||||
|
||||
return Response{
|
||||
StatusCode: internalRes.StatusCode,
|
||||
Headers: resHeaders,
|
||||
Encoding: "base64",
|
||||
Body: resBody,
|
||||
}, nil
|
||||
}
|
||||
|
||||
lambda.Start(handler)
|
||||
handler := &CgiHandler{nil, workdir, script}
|
||||
now.Start(handler)
|
||||
}
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@now/cgi",
|
||||
"version": "0.0.13",
|
||||
"version": "0.0.14",
|
||||
"scripts": {
|
||||
"test": "best -I test/*.js",
|
||||
"prepublish": "./build.sh"
|
||||
|
||||
1
packages/now-go/.gitignore
vendored
1
packages/now-go/.gitignore
vendored
@@ -1,3 +1,4 @@
|
||||
node_modules
|
||||
*.log
|
||||
launcher
|
||||
bin
|
||||
|
||||
@@ -9,7 +9,7 @@ const downloadGit = require('lambda-git');
|
||||
const glob = require('@now/build-utils/fs/glob.js');
|
||||
const downloadGoBin = require('./download-go-bin');
|
||||
|
||||
// creates a `$GOPATH` direcotry tree, as per
|
||||
// creates a `$GOPATH` directory tree, as per
|
||||
// `go help gopath`'s instructions.
|
||||
// without this, Go won't recognize the `$GOPATH`
|
||||
async function createGoPathTree(goPath) {
|
||||
|
||||
@@ -1,143 +1,10 @@
|
||||
package main
|
||||
|
||||
import (
|
||||
b64 "encoding/base64"
|
||||
"encoding/json"
|
||||
"fmt"
|
||||
"github.com/aws/aws-lambda-go/events"
|
||||
"github.com/aws/aws-lambda-go/lambda"
|
||||
"io/ioutil"
|
||||
"net"
|
||||
now "../../utils/go/bridge"
|
||||
"net/http"
|
||||
"strings"
|
||||
)
|
||||
|
||||
type Request struct {
|
||||
Host string `json:"host"`
|
||||
Path string `json:"path"`
|
||||
Method string `json`
|
||||
Headers map[string]string `json:"headers"`
|
||||
Encoding string `json"encoding,omitempty"`
|
||||
Body string `json:"body"`
|
||||
}
|
||||
|
||||
type Response struct {
|
||||
StatusCode int `json:"statusCode"`
|
||||
Headers map[string]string `json:"headers"`
|
||||
Encoding string `json:"encoding,omitemtpy"`
|
||||
Body string `json:"body"`
|
||||
}
|
||||
|
||||
type ResponseError struct {
|
||||
Code string `json:"code"`
|
||||
Message string `json:"message"`
|
||||
}
|
||||
|
||||
type ResponseErrorWrapper struct {
|
||||
Error ResponseError `json:"error"`
|
||||
}
|
||||
|
||||
func createErrorResponse(message string, code string, statusCode int) (Response, error) {
|
||||
obj := ResponseErrorWrapper{
|
||||
Error: ResponseError{
|
||||
Code: code,
|
||||
Message: message,
|
||||
},
|
||||
}
|
||||
|
||||
body, _ := json.Marshal(obj)
|
||||
|
||||
return Response{
|
||||
StatusCode: statusCode,
|
||||
Headers: map[string]string{
|
||||
"Content-Type": "application/json",
|
||||
},
|
||||
Body: string(body),
|
||||
}, nil
|
||||
}
|
||||
|
||||
func main() {
|
||||
l, err := net.Listen("tcp", ":0")
|
||||
if err != nil {
|
||||
panic(err)
|
||||
}
|
||||
|
||||
http.HandleFunc("/", __NOW_HANDLER_FUNC_NAME)
|
||||
go http.Serve(l, nil)
|
||||
|
||||
handler := func(_req events.APIGatewayProxyRequest) (Response, error) {
|
||||
var req Request
|
||||
|
||||
err := json.Unmarshal([]byte(_req.Body), &req)
|
||||
|
||||
if err != nil {
|
||||
fmt.Println(err)
|
||||
return createErrorResponse("Invalid payload", "bad_request", 400)
|
||||
}
|
||||
|
||||
if req.Encoding == "base64" {
|
||||
decoded, _ := b64.StdEncoding.DecodeString(req.Body)
|
||||
req.Body = string(decoded)
|
||||
}
|
||||
|
||||
url := "http://" + l.Addr().String() + req.Path
|
||||
|
||||
internalReq, err := http.NewRequest(req.Method, url, strings.NewReader(req.Body))
|
||||
if err != nil {
|
||||
fmt.Println(err)
|
||||
return createErrorResponse("Bad gateway", "bad_gateway", 502)
|
||||
}
|
||||
|
||||
for k, v := range req.Headers {
|
||||
internalReq.Header.Add(k, v)
|
||||
if strings.ToLower(k) == "host" {
|
||||
// we need to set `Host` in the request
|
||||
// because Go likes to ignore the `Host` header
|
||||
// see https://github.com/golang/go/issues/7682
|
||||
internalReq.Host = v
|
||||
}
|
||||
}
|
||||
|
||||
client := &http.Client{}
|
||||
internalRes, err := client.Do(internalReq)
|
||||
if err != nil {
|
||||
fmt.Println(err)
|
||||
return createErrorResponse("Bad gateway", "bad_gateway", 502)
|
||||
}
|
||||
defer internalRes.Body.Close()
|
||||
|
||||
resHeaders := make(map[string]string, len(internalRes.Header))
|
||||
var resEncoding string
|
||||
for k, v := range internalRes.Header {
|
||||
// FIXME: support multiple values via concatenating with ','
|
||||
// see RFC 7230, section 3.2.2
|
||||
if strings.ToLower(k) == "x-now-response-encoding" {
|
||||
// we don't want to send this header down
|
||||
resEncoding = v[0]
|
||||
} else {
|
||||
resHeaders[k] = v[0]
|
||||
}
|
||||
}
|
||||
|
||||
bodyBytes, err := ioutil.ReadAll(internalRes.Body)
|
||||
if err != nil {
|
||||
return createErrorResponse("Bad gateway", "bad_gateway", 502)
|
||||
}
|
||||
|
||||
var resBody string
|
||||
if resEncoding == "base64" {
|
||||
resBody = b64.StdEncoding.EncodeToString(bodyBytes)
|
||||
} else {
|
||||
resBody = string(bodyBytes)
|
||||
}
|
||||
|
||||
return Response{
|
||||
StatusCode: internalRes.StatusCode,
|
||||
Headers: resHeaders,
|
||||
Encoding: resEncoding,
|
||||
Body: resBody,
|
||||
}, nil
|
||||
}
|
||||
|
||||
lambda.Start(handler)
|
||||
now.Start(http.HandlerFunc(__NOW_HANDLER_FUNC_NAME))
|
||||
}
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@now/go",
|
||||
"version": "0.2.10",
|
||||
"version": "0.2.11",
|
||||
"scripts": {
|
||||
"test": "best -I test/*.js",
|
||||
"prepublish": "./build.sh"
|
||||
|
||||
1
packages/now-lambda/.npmignore
Normal file
1
packages/now-lambda/.npmignore
Normal file
@@ -0,0 +1 @@
|
||||
/test
|
||||
@@ -1,7 +1,10 @@
|
||||
{
|
||||
"name": "@now/lambda",
|
||||
"version": "0.4.7",
|
||||
"version": "0.4.8",
|
||||
"peerDependencies": {
|
||||
"@now/build-utils": ">=0.0.1"
|
||||
},
|
||||
"scripts": {
|
||||
"test": "jest"
|
||||
}
|
||||
}
|
||||
|
||||
BIN
packages/now-lambda/test/fixtures/01-cowsay/index.zip
vendored
Normal file
BIN
packages/now-lambda/test/fixtures/01-cowsay/index.zip
vendored
Normal file
Binary file not shown.
11
packages/now-lambda/test/fixtures/01-cowsay/now.json
vendored
Normal file
11
packages/now-lambda/test/fixtures/01-cowsay/now.json
vendored
Normal file
@@ -0,0 +1,11 @@
|
||||
{
|
||||
"version": 2,
|
||||
"builds": [
|
||||
{ "src": "index.zip", "use": "@now/lambda", "config": { "handler": "index.handler", "runtime": "nodejs8.10" } },
|
||||
{ "src": "subdirectory/index.zip", "use": "@now/lambda", "config": { "handler": "index.handler", "runtime": "nodejs8.10" } }
|
||||
],
|
||||
"probes": [
|
||||
{ "path": "/", "mustContain": "cow:NO_REPLACE_TO_AVOID_CRC_MISMATCH" },
|
||||
{ "path": "/subdirectory/", "mustContain": "yoda:NO_REPLACE_TO_AVOID_CRC_MISMATCH" }
|
||||
]
|
||||
}
|
||||
BIN
packages/now-lambda/test/fixtures/01-cowsay/subdirectory/index.zip
vendored
Normal file
BIN
packages/now-lambda/test/fixtures/01-cowsay/subdirectory/index.zip
vendored
Normal file
Binary file not shown.
33
packages/now-lambda/test/test.js
Normal file
33
packages/now-lambda/test/test.js
Normal file
@@ -0,0 +1,33 @@
|
||||
/* global beforeAll, expect, it, jest */
|
||||
const fs = require('fs');
|
||||
const path = require('path');
|
||||
|
||||
const {
|
||||
packAndDeploy,
|
||||
testDeployment,
|
||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||
|
||||
jest.setTimeout(2 * 60 * 1000);
|
||||
const buildUtilsUrl = '@canary';
|
||||
let builderUrl;
|
||||
|
||||
beforeAll(async () => {
|
||||
const builderPath = path.resolve(__dirname, '..');
|
||||
builderUrl = await packAndDeploy(builderPath);
|
||||
console.log('builderUrl', builderUrl);
|
||||
});
|
||||
|
||||
const fixturesPath = path.resolve(__dirname, 'fixtures');
|
||||
|
||||
// eslint-disable-next-line no-restricted-syntax
|
||||
for (const fixture of fs.readdirSync(fixturesPath)) {
|
||||
// eslint-disable-next-line no-loop-func
|
||||
it(`should build ${fixture}`, async () => {
|
||||
await expect(
|
||||
testDeployment(
|
||||
{ builderUrl, buildUtilsUrl },
|
||||
path.join(fixturesPath, fixture),
|
||||
),
|
||||
).resolves.toBeDefined();
|
||||
});
|
||||
}
|
||||
1
packages/now-md/.npmignore
Normal file
1
packages/now-md/.npmignore
Normal file
@@ -0,0 +1 @@
|
||||
/test
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@now/md",
|
||||
"version": "0.4.7",
|
||||
"version": "0.4.8",
|
||||
"dependencies": {
|
||||
"rehype-document": "^2.2.0",
|
||||
"rehype-format": "^2.3.0",
|
||||
@@ -12,5 +12,8 @@
|
||||
},
|
||||
"peerDependencies": {
|
||||
"@now/build-utils": ">=0.0.1"
|
||||
},
|
||||
"scripts": {
|
||||
"test": "jest"
|
||||
}
|
||||
}
|
||||
|
||||
5
packages/now-md/test/fixtures/01-cowsay/index.md
vendored
Normal file
5
packages/now-md/test/fixtures/01-cowsay/index.md
vendored
Normal file
@@ -0,0 +1,5 @@
|
||||
# Testing Markdown
|
||||
|
||||
cow:RANDOMNESS_PLACEHOLDER
|
||||
|
||||
[Wow a link!](https://zeit.co)
|
||||
11
packages/now-md/test/fixtures/01-cowsay/now.json
vendored
Normal file
11
packages/now-md/test/fixtures/01-cowsay/now.json
vendored
Normal file
@@ -0,0 +1,11 @@
|
||||
{
|
||||
"version": 2,
|
||||
"builds": [
|
||||
{ "src": "index.md", "use": "@now/md" },
|
||||
{ "src": "subdirectory/index.md", "use": "@now/md" }
|
||||
],
|
||||
"probes": [
|
||||
{ "path": "/", "mustContain": "cow:RANDOMNESS_PLACEHOLDER" },
|
||||
{ "path": "/subdirectory/", "mustContain": "yoda:RANDOMNESS_PLACEHOLDER" }
|
||||
]
|
||||
}
|
||||
5
packages/now-md/test/fixtures/01-cowsay/subdirectory/index.md
vendored
Normal file
5
packages/now-md/test/fixtures/01-cowsay/subdirectory/index.md
vendored
Normal file
@@ -0,0 +1,5 @@
|
||||
# Testing Markdown
|
||||
|
||||
yoda:RANDOMNESS_PLACEHOLDER
|
||||
|
||||
[Wow a link!](https://zeit.co)
|
||||
33
packages/now-md/test/test.js
Normal file
33
packages/now-md/test/test.js
Normal file
@@ -0,0 +1,33 @@
|
||||
/* global beforeAll, expect, it, jest */
|
||||
const fs = require('fs');
|
||||
const path = require('path');
|
||||
|
||||
const {
|
||||
packAndDeploy,
|
||||
testDeployment,
|
||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||
|
||||
jest.setTimeout(2 * 60 * 1000);
|
||||
const buildUtilsUrl = '@canary';
|
||||
let builderUrl;
|
||||
|
||||
beforeAll(async () => {
|
||||
const builderPath = path.resolve(__dirname, '..');
|
||||
builderUrl = await packAndDeploy(builderPath);
|
||||
console.log('builderUrl', builderUrl);
|
||||
});
|
||||
|
||||
const fixturesPath = path.resolve(__dirname, 'fixtures');
|
||||
|
||||
// eslint-disable-next-line no-restricted-syntax
|
||||
for (const fixture of fs.readdirSync(fixturesPath)) {
|
||||
// eslint-disable-next-line no-loop-func
|
||||
it(`should build ${fixture}`, async () => {
|
||||
await expect(
|
||||
testDeployment(
|
||||
{ builderUrl, buildUtilsUrl },
|
||||
path.join(fixturesPath, fixture),
|
||||
),
|
||||
).resolves.toBeDefined();
|
||||
});
|
||||
}
|
||||
1
packages/now-mdx-deck/.npmignore
Normal file
1
packages/now-mdx-deck/.npmignore
Normal file
@@ -0,0 +1 @@
|
||||
/test
|
||||
@@ -6,15 +6,13 @@ const glob = require('@now/build-utils/fs/glob.js');
|
||||
const path = require('path');
|
||||
const { runNpmInstall } = require('@now/build-utils/fs/run-user-scripts.js');
|
||||
|
||||
exports.analyze = ({ files, entrypoint }) => files[entrypoint].digest;
|
||||
|
||||
const writeFile = promisify(fs.writeFile);
|
||||
|
||||
exports.build = async ({ files, entrypoint, workPath }) => {
|
||||
console.log('downloading user files...');
|
||||
const downloadedFiles = await download(files, workPath);
|
||||
console.log('writing package.json...');
|
||||
const packageJson = { dependencies: { 'mdx-deck': '1.7.7' } };
|
||||
const packageJson = { dependencies: { 'mdx-deck': '1.7.15' } };
|
||||
const packageJsonPath = path.join(workPath, 'package.json');
|
||||
await writeFile(packageJsonPath, JSON.stringify(packageJson));
|
||||
console.log('running npm install...');
|
||||
@@ -47,7 +45,7 @@ exports.build = async ({ files, entrypoint, workPath }) => {
|
||||
|
||||
exports.prepareCache = async ({ cachePath }) => {
|
||||
console.log('writing package.json...');
|
||||
const packageJson = { dependencies: { 'mdx-deck': '1.7.7' } };
|
||||
const packageJson = { dependencies: { 'mdx-deck': '1.7.15' } };
|
||||
const packageJsonPath = path.join(cachePath, 'package.json');
|
||||
await writeFile(packageJsonPath, JSON.stringify(packageJson));
|
||||
console.log('running npm install...');
|
||||
|
||||
@@ -1,7 +1,10 @@
|
||||
{
|
||||
"name": "@now/mdx-deck",
|
||||
"version": "0.4.16",
|
||||
"version": "0.4.18-canary.0",
|
||||
"peerDependencies": {
|
||||
"@now/build-utils": ">=0.0.1"
|
||||
},
|
||||
"scripts": {
|
||||
"test": "jest"
|
||||
}
|
||||
}
|
||||
|
||||
5
packages/now-mdx-deck/test/fixtures/01-cowsay/index.mdx
vendored
Normal file
5
packages/now-mdx-deck/test/fixtures/01-cowsay/index.mdx
vendored
Normal file
@@ -0,0 +1,5 @@
|
||||
# Testing Markdown
|
||||
---
|
||||
cow:RANDOMNESS_PLACEHOLDER
|
||||
---
|
||||
[Wow a link!](https://zeit.co)
|
||||
11
packages/now-mdx-deck/test/fixtures/01-cowsay/now.json
vendored
Normal file
11
packages/now-mdx-deck/test/fixtures/01-cowsay/now.json
vendored
Normal file
@@ -0,0 +1,11 @@
|
||||
{
|
||||
"version": 2,
|
||||
"builds": [
|
||||
{ "src": "index.mdx", "use": "@now/mdx-deck" },
|
||||
{ "src": "subdirectory/index.mdx", "use": "@now/mdx-deck" }
|
||||
],
|
||||
"probes": [
|
||||
{ "path": "/", "mustContain": "cow:RANDOMNESS_PLACEHOLDER" },
|
||||
{ "path": "/subdirectory/", "mustContain": "yoda:RANDOMNESS_PLACEHOLDER" }
|
||||
]
|
||||
}
|
||||
5
packages/now-mdx-deck/test/fixtures/01-cowsay/subdirectory/index.mdx
vendored
Normal file
5
packages/now-mdx-deck/test/fixtures/01-cowsay/subdirectory/index.mdx
vendored
Normal file
@@ -0,0 +1,5 @@
|
||||
# Testing Markdown
|
||||
---
|
||||
yoda:RANDOMNESS_PLACEHOLDER
|
||||
---
|
||||
[Wow a link!](https://zeit.co)
|
||||
33
packages/now-mdx-deck/test/test.js
Normal file
33
packages/now-mdx-deck/test/test.js
Normal file
@@ -0,0 +1,33 @@
|
||||
/* global beforeAll, expect, it, jest */
|
||||
const fs = require('fs');
|
||||
const path = require('path');
|
||||
|
||||
const {
|
||||
packAndDeploy,
|
||||
testDeployment,
|
||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||
|
||||
jest.setTimeout(2 * 60 * 1000);
|
||||
const buildUtilsUrl = '@canary';
|
||||
let builderUrl;
|
||||
|
||||
beforeAll(async () => {
|
||||
const builderPath = path.resolve(__dirname, '..');
|
||||
builderUrl = await packAndDeploy(builderPath);
|
||||
console.log('builderUrl', builderUrl);
|
||||
});
|
||||
|
||||
const fixturesPath = path.resolve(__dirname, 'fixtures');
|
||||
|
||||
// eslint-disable-next-line no-restricted-syntax
|
||||
for (const fixture of fs.readdirSync(fixturesPath)) {
|
||||
// eslint-disable-next-line no-loop-func
|
||||
it(`should build ${fixture}`, async () => {
|
||||
await expect(
|
||||
testDeployment(
|
||||
{ builderUrl, buildUtilsUrl },
|
||||
path.join(fixturesPath, fixture),
|
||||
),
|
||||
).resolves.toBeDefined();
|
||||
});
|
||||
}
|
||||
@@ -17,6 +17,7 @@ const {
|
||||
excludeLockFiles,
|
||||
normalizePackageJson,
|
||||
excludeStaticDirectory,
|
||||
onlyStaticDirectory,
|
||||
} = require('./utils');
|
||||
|
||||
/** @typedef { import('@now/build-utils/file-ref').Files } Files */
|
||||
@@ -91,7 +92,7 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
||||
const filesWithoutStaticDirectory = excludeStaticDirectory(
|
||||
filesWithoutLockfiles,
|
||||
);
|
||||
let downloadedFiles = await download(filesWithoutStaticDirectory, workPath);
|
||||
const downloadedFiles = await download(filesWithoutStaticDirectory, workPath);
|
||||
|
||||
console.log('normalizing package.json');
|
||||
const packageJson = normalizePackageJson(
|
||||
@@ -105,8 +106,6 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
||||
await writeNpmRc(workPath, process.env.NPM_AUTH_TOKEN);
|
||||
}
|
||||
|
||||
downloadedFiles = await glob('**', workPath);
|
||||
|
||||
console.log('running npm install...');
|
||||
await runNpmInstall(workPath, ['--prefer-offline']);
|
||||
console.log('running user script...');
|
||||
@@ -117,7 +116,7 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
||||
await unlink(path.join(workPath, '.npmrc'));
|
||||
}
|
||||
|
||||
downloadedFiles = await glob('**', workPath);
|
||||
const filesAfterBuild = await glob('**', workPath);
|
||||
|
||||
console.log('preparing lambda files...');
|
||||
let buildId;
|
||||
@@ -144,8 +143,8 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
||||
...dotNextServerRootFiles,
|
||||
...launcherFiles,
|
||||
};
|
||||
if (downloadedFiles['next.config.js']) {
|
||||
nextFiles['next.config.js'] = downloadedFiles['next.config.js'];
|
||||
if (filesAfterBuild['next.config.js']) {
|
||||
nextFiles['next.config.js'] = filesAfterBuild['next.config.js'];
|
||||
}
|
||||
const pages = await glob(
|
||||
'**/*.js',
|
||||
@@ -169,16 +168,16 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
||||
);
|
||||
|
||||
const pageFiles = {
|
||||
[`.next/server/static/${buildId}/pages/_document.js`]: downloadedFiles[
|
||||
[`.next/server/static/${buildId}/pages/_document.js`]: filesAfterBuild[
|
||||
`.next/server/static/${buildId}/pages/_document.js`
|
||||
],
|
||||
[`.next/server/static/${buildId}/pages/_app.js`]: downloadedFiles[
|
||||
[`.next/server/static/${buildId}/pages/_app.js`]: filesAfterBuild[
|
||||
`.next/server/static/${buildId}/pages/_app.js`
|
||||
],
|
||||
[`.next/server/static/${buildId}/pages/_error.js`]: downloadedFiles[
|
||||
[`.next/server/static/${buildId}/pages/_error.js`]: filesAfterBuild[
|
||||
`.next/server/static/${buildId}/pages/_error.js`
|
||||
],
|
||||
[`.next/server/static/${buildId}/pages/${page}`]: downloadedFiles[
|
||||
[`.next/server/static/${buildId}/pages/${page}`]: filesAfterBuild[
|
||||
`.next/server/static/${buildId}/pages/${page}`
|
||||
],
|
||||
};
|
||||
@@ -209,12 +208,36 @@ exports.build = async ({ files, workPath, entrypoint }) => {
|
||||
{},
|
||||
);
|
||||
|
||||
return { ...lambdas, ...staticFiles };
|
||||
const nextStaticDirectory = onlyStaticDirectory(filesWithoutLockfiles);
|
||||
const staticDirectoryFiles = Object.keys(nextStaticDirectory).reduce(
|
||||
(mappedFiles, file) => ({
|
||||
...mappedFiles,
|
||||
[path.join(entryDirectory, file)]: nextStaticDirectory[file],
|
||||
}),
|
||||
{},
|
||||
);
|
||||
|
||||
return { ...lambdas, ...staticFiles, ...staticDirectoryFiles };
|
||||
};
|
||||
|
||||
exports.prepareCache = async ({ files, cachePath, workPath }) => {
|
||||
exports.prepareCache = async ({
|
||||
files, entrypoint, cachePath, workPath,
|
||||
}) => {
|
||||
console.log('downloading user files...');
|
||||
await download(files, cachePath);
|
||||
const entryDirectory = path.dirname(entrypoint);
|
||||
const filesOnlyEntryDirectory = includeOnlyEntryDirectory(
|
||||
files,
|
||||
entryDirectory,
|
||||
);
|
||||
const filesWithEntryDirectoryRoot = moveEntryDirectoryToRoot(
|
||||
filesOnlyEntryDirectory,
|
||||
entryDirectory,
|
||||
);
|
||||
const filesWithoutLockfiles = excludeLockFiles(filesWithEntryDirectoryRoot);
|
||||
const filesWithoutStaticDirectory = excludeStaticDirectory(
|
||||
filesWithoutLockfiles,
|
||||
);
|
||||
await download(filesWithoutStaticDirectory, workPath);
|
||||
await download(await glob('.next/**', workPath), cachePath);
|
||||
await download(await glob('node_modules/**', workPath), cachePath);
|
||||
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@now/next",
|
||||
"version": "0.0.80",
|
||||
"version": "0.0.82-canary.0",
|
||||
"dependencies": {
|
||||
"@now/node-bridge": "0.1.4",
|
||||
"execa": "^1.0.0",
|
||||
|
||||
@@ -111,6 +111,19 @@ function excludeStaticDirectory(files) {
|
||||
return excludeFiles(files, matcher);
|
||||
}
|
||||
|
||||
/**
|
||||
* Exclude the static directory from files
|
||||
* @param {Files} files
|
||||
* @returns {Files}
|
||||
*/
|
||||
function onlyStaticDirectory(files) {
|
||||
function matcher(filePath) {
|
||||
return !filePath.startsWith('static');
|
||||
}
|
||||
|
||||
return excludeFiles(files, matcher);
|
||||
}
|
||||
|
||||
/**
|
||||
* Enforce specific package.json configuration for smallest possible lambda
|
||||
* @param {{dependencies?: any, devDependencies?: any, scripts?: any}} defaultPackageJson
|
||||
@@ -151,7 +164,7 @@ function normalizePackageJson(defaultPackageJson = {}) {
|
||||
},
|
||||
scripts: {
|
||||
...defaultPackageJson.scripts,
|
||||
'now-build': 'next build --lambdas',
|
||||
'now-build': 'NODE_OPTIONS=--max_old_space_size=3000 next build --lambdas',
|
||||
},
|
||||
};
|
||||
}
|
||||
@@ -164,4 +177,5 @@ module.exports = {
|
||||
excludeLockFiles,
|
||||
normalizePackageJson,
|
||||
excludeStaticDirectory,
|
||||
onlyStaticDirectory,
|
||||
};
|
||||
|
||||
@@ -47,7 +47,7 @@ async function downloadInstallAndBundle(
|
||||
'package.json': new FileBlob({
|
||||
data: JSON.stringify({
|
||||
dependencies: {
|
||||
'@zeit/ncc': '0.1.12',
|
||||
'@zeit/ncc': '0.5.3',
|
||||
},
|
||||
}),
|
||||
}),
|
||||
|
||||
@@ -12,7 +12,10 @@ Server.prototype.listen = function listen(...args) {
|
||||
};
|
||||
|
||||
try {
|
||||
process.env.NODE_ENV = 'production';
|
||||
if (!process.env.NODE_ENV) {
|
||||
process.env.NODE_ENV = 'production';
|
||||
}
|
||||
|
||||
// PLACEHOLDER
|
||||
} catch (error) {
|
||||
console.error(error);
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@now/node-server",
|
||||
"version": "0.4.24",
|
||||
"version": "0.4.26-canary.2",
|
||||
"dependencies": {
|
||||
"@now/node-bridge": "^0.1.9",
|
||||
"fs-extra": "7.0.1"
|
||||
|
||||
@@ -6,6 +6,6 @@
|
||||
],
|
||||
"probes": [
|
||||
{ "path": "/", "mustContain": "cow:RANDOMNESS_PLACEHOLDER" },
|
||||
{ "path": "/subdirectory", "mustContain": "yoda:RANDOMNESS_PLACEHOLDER" }
|
||||
{ "path": "/subdirectory/", "mustContain": "yoda:RANDOMNESS_PLACEHOLDER" }
|
||||
]
|
||||
}
|
||||
|
||||
15
packages/now-node-server/test/fixtures/08-assets/index.js
vendored
Normal file
15
packages/now-node-server/test/fixtures/08-assets/index.js
vendored
Normal file
@@ -0,0 +1,15 @@
|
||||
const fs = require('fs');
|
||||
const http = require('http');
|
||||
const path = require('path');
|
||||
|
||||
const server = http.createServer((req, resp) => {
|
||||
const asset1 = fs.readFileSync(
|
||||
path.join(__dirname, 'subdirectory1/asset.txt'),
|
||||
);
|
||||
const asset2 = fs.readFileSync(
|
||||
path.join(__dirname, 'subdirectory2/asset.txt'),
|
||||
);
|
||||
resp.end(`${asset1},${asset2}`);
|
||||
});
|
||||
|
||||
server.listen();
|
||||
9
packages/now-node-server/test/fixtures/08-assets/now.json
vendored
Normal file
9
packages/now-node-server/test/fixtures/08-assets/now.json
vendored
Normal file
@@ -0,0 +1,9 @@
|
||||
{
|
||||
"version": 2,
|
||||
"builds": [
|
||||
{ "src": "index.js", "use": "@now/node-server" }
|
||||
],
|
||||
"probes": [
|
||||
{ "path": "/", "mustContain": "asset1:RANDOMNESS_PLACEHOLDER,asset2:RANDOMNESS_PLACEHOLDER" }
|
||||
]
|
||||
}
|
||||
1
packages/now-node-server/test/fixtures/08-assets/subdirectory1/asset.txt
vendored
Normal file
1
packages/now-node-server/test/fixtures/08-assets/subdirectory1/asset.txt
vendored
Normal file
@@ -0,0 +1 @@
|
||||
asset1:RANDOMNESS_PLACEHOLDER
|
||||
1
packages/now-node-server/test/fixtures/08-assets/subdirectory2/asset.txt
vendored
Normal file
1
packages/now-node-server/test/fixtures/08-assets/subdirectory2/asset.txt
vendored
Normal file
@@ -0,0 +1 @@
|
||||
asset2:RANDOMNESS_PLACEHOLDER
|
||||
@@ -5,7 +5,7 @@
|
||||
{ "src": "without-bundle/index.js", "use": "@now/node-server", "config": { "bundle": false } }
|
||||
],
|
||||
"probes": [
|
||||
{ "path": "/with-bundle", "mustContain": "RANDOMNESS_PLACEHOLDER" },
|
||||
{ "path": "/without-bundle", "mustContain": "RANDOMNESS_PLACEHOLDER" }
|
||||
{ "path": "/with-bundle", "mustContain": "RANDOMNESS_PLACEHOLDER:with-bundle" },
|
||||
{ "path": "/without-bundle", "mustContain": "RANDOMNESS_PLACEHOLDER:without-bundle" }
|
||||
]
|
||||
}
|
||||
|
||||
@@ -2,7 +2,7 @@ const http = require('http');
|
||||
const isBundled = require('./is-bundled.js');
|
||||
|
||||
const server = http.createServer((req, resp) => {
|
||||
resp.end(isBundled() ? 'RANDOMNESS_PLACEHOLDER' : '');
|
||||
resp.end(isBundled() ? 'RANDOMNESS_PLACEHOLDER:with-bundle' : '');
|
||||
});
|
||||
|
||||
server.listen();
|
||||
|
||||
@@ -2,7 +2,7 @@ const http = require('http');
|
||||
const isBundled = require('./is-bundled.js');
|
||||
|
||||
const server = http.createServer((req, resp) => {
|
||||
resp.end(isBundled() ? '' : 'RANDOMNESS_PLACEHOLDER');
|
||||
resp.end(isBundled() ? '' : 'RANDOMNESS_PLACEHOLDER:without-bundle');
|
||||
});
|
||||
|
||||
server.listen();
|
||||
|
||||
File diff suppressed because it is too large
Load Diff
@@ -8,6 +8,7 @@ const {
|
||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||
|
||||
jest.setTimeout(2 * 60 * 1000);
|
||||
const buildUtilsUrl = '@canary';
|
||||
let builderUrl;
|
||||
|
||||
beforeAll(async () => {
|
||||
@@ -23,7 +24,10 @@ for (const fixture of fs.readdirSync(fixturesPath)) {
|
||||
// eslint-disable-next-line no-loop-func
|
||||
it(`should build ${fixture}`, async () => {
|
||||
await expect(
|
||||
testDeployment({ builderUrl }, path.join(fixturesPath, fixture)),
|
||||
).resolves.toBe(undefined);
|
||||
testDeployment(
|
||||
{ builderUrl, buildUtilsUrl },
|
||||
path.join(fixturesPath, fixture),
|
||||
),
|
||||
).resolves.toBeDefined();
|
||||
});
|
||||
}
|
||||
|
||||
@@ -45,7 +45,7 @@ async function downloadInstallAndBundle(
|
||||
'package.json': new FileBlob({
|
||||
data: JSON.stringify({
|
||||
dependencies: {
|
||||
'@zeit/ncc': '0.1.12',
|
||||
'@zeit/ncc': '0.5.3',
|
||||
},
|
||||
}),
|
||||
}),
|
||||
@@ -109,6 +109,7 @@ exports.build = async ({ files, entrypoint, workPath }) => {
|
||||
[
|
||||
'process.chdir("./user");',
|
||||
`listener = require("./${path.join('user', entrypoint)}");`,
|
||||
'if (listener.default) listener = listener.default;',
|
||||
].join(' '),
|
||||
);
|
||||
|
||||
|
||||
@@ -6,7 +6,10 @@ bridge.port = 3000;
|
||||
let listener;
|
||||
|
||||
try {
|
||||
process.env.NODE_ENV = 'production';
|
||||
if (!process.env.NODE_ENV) {
|
||||
process.env.NODE_ENV = 'production';
|
||||
}
|
||||
|
||||
// PLACEHOLDER
|
||||
} catch (error) {
|
||||
console.error(error);
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@now/node",
|
||||
"version": "0.4.26",
|
||||
"version": "0.4.28-canary.1",
|
||||
"dependencies": {
|
||||
"@now/node-bridge": "^0.1.9",
|
||||
"fs-extra": "7.0.1"
|
||||
|
||||
@@ -6,6 +6,6 @@
|
||||
],
|
||||
"probes": [
|
||||
{ "path": "/", "mustContain": "cow:RANDOMNESS_PLACEHOLDER" },
|
||||
{ "path": "/subdirectory", "mustContain": "yoda:RANDOMNESS_PLACEHOLDER" }
|
||||
{ "path": "/subdirectory/", "mustContain": "yoda:RANDOMNESS_PLACEHOLDER" }
|
||||
]
|
||||
}
|
||||
|
||||
12
packages/now-node/test/fixtures/08-assets/index.js
vendored
Normal file
12
packages/now-node/test/fixtures/08-assets/index.js
vendored
Normal file
@@ -0,0 +1,12 @@
|
||||
const fs = require('fs');
|
||||
const path = require('path');
|
||||
|
||||
module.exports = (req, resp) => {
|
||||
const asset1 = fs.readFileSync(
|
||||
path.join(__dirname, 'subdirectory1/asset.txt'),
|
||||
);
|
||||
const asset2 = fs.readFileSync(
|
||||
path.join(__dirname, 'subdirectory2/asset.txt'),
|
||||
);
|
||||
resp.end(`${asset1},${asset2}`);
|
||||
};
|
||||
9
packages/now-node/test/fixtures/08-assets/now.json
vendored
Normal file
9
packages/now-node/test/fixtures/08-assets/now.json
vendored
Normal file
@@ -0,0 +1,9 @@
|
||||
{
|
||||
"version": 2,
|
||||
"builds": [
|
||||
{ "src": "index.js", "use": "@now/node" }
|
||||
],
|
||||
"probes": [
|
||||
{ "path": "/", "mustContain": "asset1:RANDOMNESS_PLACEHOLDER,asset2:RANDOMNESS_PLACEHOLDER" }
|
||||
]
|
||||
}
|
||||
1
packages/now-node/test/fixtures/08-assets/subdirectory1/asset.txt
vendored
Normal file
1
packages/now-node/test/fixtures/08-assets/subdirectory1/asset.txt
vendored
Normal file
@@ -0,0 +1 @@
|
||||
asset1:RANDOMNESS_PLACEHOLDER
|
||||
1
packages/now-node/test/fixtures/08-assets/subdirectory2/asset.txt
vendored
Normal file
1
packages/now-node/test/fixtures/08-assets/subdirectory2/asset.txt
vendored
Normal file
@@ -0,0 +1 @@
|
||||
asset2:RANDOMNESS_PLACEHOLDER
|
||||
@@ -8,6 +8,7 @@ const {
|
||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||
|
||||
jest.setTimeout(2 * 60 * 1000);
|
||||
const buildUtilsUrl = '@canary';
|
||||
let builderUrl;
|
||||
|
||||
beforeAll(async () => {
|
||||
@@ -23,7 +24,10 @@ for (const fixture of fs.readdirSync(fixturesPath)) {
|
||||
// eslint-disable-next-line no-loop-func
|
||||
it(`should build ${fixture}`, async () => {
|
||||
await expect(
|
||||
testDeployment({ builderUrl }, path.join(fixturesPath, fixture)),
|
||||
).resolves.toBe(undefined);
|
||||
testDeployment(
|
||||
{ builderUrl, buildUtilsUrl },
|
||||
path.join(fixturesPath, fixture),
|
||||
),
|
||||
).resolves.toBeDefined();
|
||||
});
|
||||
}
|
||||
|
||||
BIN
packages/now-php/dist/launcher
vendored
BIN
packages/now-php/dist/launcher
vendored
Binary file not shown.
BIN
packages/now-php/dist/libphp7-7.1.so
vendored
BIN
packages/now-php/dist/libphp7-7.1.so
vendored
Binary file not shown.
BIN
packages/now-php/dist/modules/curl.so
vendored
BIN
packages/now-php/dist/modules/curl.so
vendored
Binary file not shown.
BIN
packages/now-php/dist/modules/json.so
vendored
BIN
packages/now-php/dist/modules/json.so
vendored
Binary file not shown.
@@ -12,10 +12,11 @@ RUN go get -v github.com/aws/aws-lambda-go/events
|
||||
RUN go get -v github.com/deuill/go-php
|
||||
|
||||
WORKDIR /root/go/app
|
||||
COPY ./utils/bridge.go /root/go/app/utils/bridge.go
|
||||
COPY ./launcher.go /root/go/app/launcher.go
|
||||
COPY ./php.ini /root/go/app/php.ini
|
||||
COPY ./test.go /root/go/app/test.go
|
||||
COPY ./test.php /root/go/app/public/test.php
|
||||
COPY ./test.php /root/go/app/test.php
|
||||
COPY ./test.sh /root/go/app/test.sh
|
||||
|
||||
RUN GOOS=linux GOARCH=amd64 CGO_ENABLED=1 go build -o launcher launcher.go
|
||||
|
||||
@@ -1,12 +1,15 @@
|
||||
rm -rf ../dist
|
||||
mkdir -p ../dist/modules
|
||||
docker rmi go-php-builder --force
|
||||
docker build . -t go-php-builder
|
||||
docker run go-php-builder
|
||||
docker run go-php-builder /bin/cat /root/go/app/launcher > ../dist/launcher
|
||||
docker run go-php-builder /bin/cat /root/go/app/php.ini > ../dist/php.ini
|
||||
docker run go-php-builder /bin/cat /usr/lib64/libphp7-7.1.so > ../dist/libphp7-7.1.so
|
||||
docker run go-php-builder /bin/cat /usr/lib64/php/modules/curl.so > ../dist/modules/curl.so
|
||||
docker run go-php-builder /bin/cat /usr/lib64/php/modules/json.so > ../dist/modules/json.so
|
||||
docker run go-php-builder /bin/cat /usr/lib64/php/modules/mbstring.so > ../dist/modules/mbstring.so
|
||||
mkdir ./utils
|
||||
cp ../../../utils/go/bridge/bridge.go ./utils/bridge.go
|
||||
docker rmi now-php-docker-image --force
|
||||
docker build . -t now-php-docker-image
|
||||
docker run now-php-docker-image
|
||||
docker run now-php-docker-image /bin/cat /root/go/app/launcher > ../dist/launcher
|
||||
docker run now-php-docker-image /bin/cat /root/go/app/php.ini > ../dist/php.ini
|
||||
docker run now-php-docker-image /bin/cat /usr/lib64/libphp7-7.1.so > ../dist/libphp7-7.1.so
|
||||
docker run now-php-docker-image /bin/cat /usr/lib64/php/modules/curl.so > ../dist/modules/curl.so
|
||||
docker run now-php-docker-image /bin/cat /usr/lib64/php/modules/json.so > ../dist/modules/json.so
|
||||
docker run now-php-docker-image /bin/cat /usr/lib64/php/modules/mbstring.so > ../dist/modules/mbstring.so
|
||||
chmod +x ../dist/launcher
|
||||
rm -rf ./utils
|
||||
|
||||
@@ -1,72 +1,95 @@
|
||||
package main
|
||||
|
||||
import (
|
||||
now "./utils"
|
||||
"bytes"
|
||||
"context"
|
||||
"encoding/base64"
|
||||
"encoding/json"
|
||||
php "github.com/deuill/go-php"
|
||||
"io/ioutil"
|
||||
"net/http"
|
||||
"net/url"
|
||||
"os"
|
||||
"path"
|
||||
"path/filepath"
|
||||
"strings"
|
||||
|
||||
"github.com/aws/aws-lambda-go/events"
|
||||
"github.com/aws/aws-lambda-go/lambda"
|
||||
php "github.com/deuill/go-php"
|
||||
)
|
||||
|
||||
type Request struct {
|
||||
Host string `json:"host"`
|
||||
Path string `json:"path"`
|
||||
Method string `json:"method"`
|
||||
Headers map[string]string `json:"headers"`
|
||||
Encoding string `json:"encoding,omitempty"`
|
||||
Body string `json:"body"`
|
||||
type PhpHandler struct {
|
||||
http.Handler
|
||||
ScriptFull string
|
||||
}
|
||||
|
||||
var phpScript = ""
|
||||
var phpScriptFull = ""
|
||||
|
||||
func handler(ctx context.Context, event events.APIGatewayProxyRequest) (events.APIGatewayProxyResponse, error) {
|
||||
func (h *PhpHandler) ServeHTTP(w http.ResponseWriter, r *http.Request) {
|
||||
engine, _ := php.New()
|
||||
context, _ := engine.NewContext()
|
||||
|
||||
var req Request
|
||||
json.Unmarshal([]byte(event.Body), &req)
|
||||
|
||||
var body string
|
||||
if req.Encoding == "base64" {
|
||||
decoded, _ := base64.StdEncoding.DecodeString(req.Body)
|
||||
body = string(decoded)
|
||||
} else {
|
||||
body = string(req.Body)
|
||||
}
|
||||
|
||||
var bodyReader = strings.NewReader(body)
|
||||
var httpReq, _ = http.NewRequest(req.Method, req.Path, bodyReader)
|
||||
|
||||
for k, v := range req.Headers {
|
||||
httpReq.Header.Add(k, v)
|
||||
}
|
||||
|
||||
var query = httpReq.URL.Query()
|
||||
getMap := make(map[string]string)
|
||||
for k, v := range query {
|
||||
for _, s := range v {
|
||||
getMap[k] = s
|
||||
for k, v := range r.URL.Query() {
|
||||
if strings.HasSuffix(k, "[]") {
|
||||
sb := ""
|
||||
for _, s := range v {
|
||||
if sb != "" {
|
||||
sb += ","
|
||||
}
|
||||
sb += "'" + s + "'"
|
||||
}
|
||||
k = strings.TrimSuffix(k, "[]")
|
||||
context.Eval("$_GET['" + k + "']=Array(" + sb + ");")
|
||||
context.Eval("$_REQUEST['" + k + "']=Array(" + sb + ");")
|
||||
} else {
|
||||
s := v[len(v) - 1]
|
||||
context.Eval("$_GET['" + k + "']='" + s + "';") // TODO escape quotes
|
||||
context.Eval("$_REQUEST['" + k + "']='" + s + "';") // TODO escape quotes
|
||||
}
|
||||
}
|
||||
context.Bind("_GET", getMap)
|
||||
|
||||
httpReq.ParseForm()
|
||||
postMap := make(map[string]string)
|
||||
for k, v := range httpReq.PostForm {
|
||||
for _, s := range v {
|
||||
postMap[k] = s
|
||||
body, _ := ioutil.ReadAll(r.Body)
|
||||
context.Eval("$HTTP_RAW_POST_DATA='" + string(body) + "';") // TODO escape-unescape
|
||||
r.Body = ioutil.NopCloser(bytes.NewBuffer(body))
|
||||
r.ParseForm()
|
||||
for k, v := range r.PostForm {
|
||||
if strings.HasSuffix(k, "[]") {
|
||||
sb := ""
|
||||
for _, s := range v {
|
||||
if sb != "" {
|
||||
sb += ","
|
||||
}
|
||||
sb += "'" + s + "'"
|
||||
}
|
||||
k = strings.TrimSuffix(k, "[]")
|
||||
context.Eval("$_POST['" + k + "']=Array(" + sb + ");")
|
||||
context.Eval("$_REQUEST['" + k + "']=Array(" + sb + ");")
|
||||
} else {
|
||||
s := v[len(v) - 1]
|
||||
context.Eval("$_POST['" + k + "']='" + s + "';") // TODO escape quotes
|
||||
context.Eval("$_REQUEST['" + k + "']='" + s + "';") // TODO escape quotes
|
||||
}
|
||||
}
|
||||
|
||||
cookies := r.Cookies()
|
||||
cookieMap := make(map[string]string)
|
||||
for _, c := range cookies {
|
||||
k, _ := url.QueryUnescape(c.Name)
|
||||
v, _ := url.QueryUnescape(c.Value)
|
||||
s := "'" + v + "'" // TODO escape quotes
|
||||
if strings.HasSuffix(k, "[]") {
|
||||
if value, exists := cookieMap[k]; exists {
|
||||
cookieMap[k] = value + "," + s
|
||||
} else {
|
||||
cookieMap[k] = s
|
||||
}
|
||||
} else {
|
||||
if _, exists := cookieMap[k]; !exists {
|
||||
cookieMap[k] = s
|
||||
}
|
||||
}
|
||||
}
|
||||
for k, v := range cookieMap {
|
||||
if strings.HasSuffix(k, "[]") {
|
||||
k = strings.TrimSuffix(k, "[]")
|
||||
context.Eval("$_COOKIE['" + k + "']=Array(" + v + ");")
|
||||
} else {
|
||||
context.Eval("$_COOKIE['" + k + "']=" + v + ";")
|
||||
}
|
||||
}
|
||||
context.Bind("_POST", postMap)
|
||||
|
||||
envMap := make(map[string]string)
|
||||
for _, e := range os.Environ() {
|
||||
@@ -75,29 +98,46 @@ func handler(ctx context.Context, event events.APIGatewayProxyRequest) (events.A
|
||||
}
|
||||
context.Bind("_ENV", envMap)
|
||||
|
||||
context.Eval("$_SERVER[\"SERVER_NAME\"]=\"" + req.Host + "\";")
|
||||
context.Eval("$_SERVER[\"SERVER_PORT\"]=\"443\";")
|
||||
context.Eval("$_SERVER[\"HTTPS\"]=\"on\";")
|
||||
|
||||
var stdout bytes.Buffer
|
||||
context.Output = &stdout
|
||||
context.Exec(phpScriptFull)
|
||||
|
||||
headers := make(map[string]string)
|
||||
headers["content-type"] = "text/html"
|
||||
for k, v := range context.Header {
|
||||
for k, v := range r.Header {
|
||||
for _, s := range v {
|
||||
headers[k] = s
|
||||
h := "HTTP_" + strings.ToUpper(strings.Replace(k, "-", "_", -1))
|
||||
context.Eval("$_SERVER['" + h + "']='" + s + "';")
|
||||
}
|
||||
}
|
||||
|
||||
context.Eval("$_SERVER['SCRIPT_FILENAME']='" + h.ScriptFull + "';")
|
||||
context.Eval("$_SERVER['REQUEST_METHOD']='" + r.Method + "';")
|
||||
context.Eval("$_SERVER['REQUEST_URI']='" + r.URL.RequestURI() + "';") // TODO must be unescaped to align with php
|
||||
context.Eval("$_SERVER['SERVER_PROTOCOL']='" + r.Proto + "';");
|
||||
context.Eval("$_SERVER['SERVER_NAME']='" + r.Host + "';")
|
||||
context.Eval("$_SERVER['SERVER_PORT']='443';")
|
||||
context.Eval("$_SERVER['HTTPS']='on';")
|
||||
context.Eval("http_response_code(200);")
|
||||
|
||||
var stdout bytes.Buffer
|
||||
context.Output = &stdout
|
||||
context.Exec(h.ScriptFull)
|
||||
|
||||
statusCodeVal, _ := context.Eval("return http_response_code();")
|
||||
w.WriteHeader(int(statusCodeVal.Int()))
|
||||
|
||||
headers := w.Header()
|
||||
headers.Add("content-type", "text/html")
|
||||
for k, v := range context.Header {
|
||||
for _, s := range v {
|
||||
headers.Add(k, s)
|
||||
}
|
||||
}
|
||||
|
||||
w.Write(stdout.Bytes())
|
||||
engine.Destroy()
|
||||
return events.APIGatewayProxyResponse{StatusCode: 200, Headers: headers, Body: stdout.String()}, nil
|
||||
}
|
||||
|
||||
func main() {
|
||||
ex, _ := os.Executable()
|
||||
phpScript = os.Getenv("NOW_PHP_SCRIPT")
|
||||
phpScriptFull = path.Join(filepath.Dir(ex), phpScript)
|
||||
lambda.Start(handler)
|
||||
handler := &PhpHandler{
|
||||
nil,
|
||||
path.Join(filepath.Dir(ex), os.Getenv("NOW_PHP_SCRIPT")),
|
||||
}
|
||||
now.Start(handler)
|
||||
}
|
||||
|
||||
@@ -1,54 +1,17 @@
|
||||
package main
|
||||
|
||||
import (
|
||||
"bytes"
|
||||
"fmt"
|
||||
"net/http"
|
||||
"os"
|
||||
"path"
|
||||
"path/filepath"
|
||||
"strings"
|
||||
php "github.com/deuill/go-php"
|
||||
"bytes"
|
||||
"fmt"
|
||||
php "github.com/deuill/go-php"
|
||||
)
|
||||
|
||||
var public = ""
|
||||
|
||||
func handler() {
|
||||
engine, _ := php.New()
|
||||
context, _ := engine.NewContext()
|
||||
|
||||
var bodyReader = strings.NewReader("Message=from test.go")
|
||||
var httpReq, _ = http.NewRequest("POST", "/dummy/path", bodyReader)
|
||||
httpReq.Header.Add("Content-Type", "application/x-www-form-urlencoded")
|
||||
httpReq.ParseForm()
|
||||
|
||||
postMap := make(map[string]string)
|
||||
for k, v := range httpReq.PostForm {
|
||||
for _, s := range v {
|
||||
postMap[k] = s
|
||||
}
|
||||
}
|
||||
|
||||
context.Bind("_POST", postMap)
|
||||
var stdout bytes.Buffer
|
||||
context.Output = &stdout
|
||||
context.Exec(path.Join(public, "test.php"))
|
||||
|
||||
for k, v := range context.Header {
|
||||
// see https://golang.org/src/net/http/header.go function writeSubset
|
||||
for _, s := range v {
|
||||
fmt.Printf("%s: %s\n", k, s)
|
||||
}
|
||||
}
|
||||
|
||||
fmt.Printf("\n")
|
||||
fmt.Println(stdout.String())
|
||||
engine.Destroy()
|
||||
}
|
||||
|
||||
func main() {
|
||||
ex, _ := os.Executable()
|
||||
public = path.Join(filepath.Dir(ex), "public")
|
||||
fmt.Printf("public %s\n", path.Join(filepath.Dir(ex), "public"))
|
||||
handler()
|
||||
engine, _ := php.New()
|
||||
context, _ := engine.NewContext()
|
||||
var stdout bytes.Buffer
|
||||
context.Output = &stdout
|
||||
context.Exec("test.php")
|
||||
fmt.Println(stdout.String())
|
||||
engine.Destroy()
|
||||
}
|
||||
|
||||
@@ -1,5 +1,2 @@
|
||||
<?php
|
||||
header('X-See-You: Tomorrow');
|
||||
header('Content-Type: text/html; charset=UTF-8');
|
||||
phpinfo();
|
||||
?>
|
||||
phpinfo();
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@now/php",
|
||||
"version": "0.4.11",
|
||||
"version": "0.4.13-canary.1",
|
||||
"peerDependencies": {
|
||||
"@now/build-utils": ">=0.0.1"
|
||||
},
|
||||
|
||||
@@ -1,3 +1,2 @@
|
||||
<?php
|
||||
print "cow:RANDOMNESS_PLACEHOLDER";
|
||||
?>
|
||||
print('cow:RANDOMNESS_PLACEHOLDER');
|
||||
|
||||
@@ -6,6 +6,6 @@
|
||||
],
|
||||
"probes": [
|
||||
{ "path": "/", "mustContain": "cow:RANDOMNESS_PLACEHOLDER" },
|
||||
{ "path": "/subdirectory", "mustContain": "yoda:RANDOMNESS_PLACEHOLDER" }
|
||||
{ "path": "/subdirectory/", "mustContain": "yoda:RANDOMNESS_PLACEHOLDER" }
|
||||
]
|
||||
}
|
||||
|
||||
@@ -1,3 +1,2 @@
|
||||
<?php
|
||||
print "yoda:RANDOMNESS_PLACEHOLDER";
|
||||
?>
|
||||
print('yoda:RANDOMNESS_PLACEHOLDER');
|
||||
|
||||
@@ -1,3 +1,2 @@
|
||||
<?php
|
||||
print $_ENV["RANDOMNESS_ENV_VAR"] . ":env";
|
||||
?>
|
||||
print($_ENV['RANDOMNESS_ENV_VAR'] . ':env');
|
||||
|
||||
28
packages/now-php/test/fixtures/11-globals/index.php
vendored
Normal file
28
packages/now-php/test/fixtures/11-globals/index.php
vendored
Normal file
@@ -0,0 +1,28 @@
|
||||
<?php
|
||||
header('Content-Type: text/plain');
|
||||
print($_SERVER['SCRIPT_FILENAME'] . PHP_EOL);
|
||||
print($_SERVER['REQUEST_METHOD'] . PHP_EOL);
|
||||
print($_SERVER['REQUEST_URI'] . PHP_EOL);
|
||||
print($_SERVER['HTTP_HOST'] . PHP_EOL);
|
||||
print($_SERVER['HTTP_X_SOME_HEADER'] . PHP_EOL);
|
||||
print($_SERVER['SERVER_PROTOCOL'] . PHP_EOL);
|
||||
print($_SERVER['SERVER_NAME'] . PHP_EOL);
|
||||
print($_SERVER['SERVER_PORT'] . PHP_EOL);
|
||||
print($_SERVER['HTTPS'] . PHP_EOL);
|
||||
|
||||
print($_GET['get1'] . PHP_EOL);
|
||||
var_dump($_GET['get2']);
|
||||
print($_POST['post1'] . PHP_EOL);
|
||||
var_dump($_POST['post2']);
|
||||
print($_COOKIE['cookie1'] . PHP_EOL);
|
||||
var_dump($_COOKIE['cookie2']);
|
||||
|
||||
print($_REQUEST['get1'] . PHP_EOL);
|
||||
var_dump($_REQUEST['get2']);
|
||||
print($_REQUEST['post1'] . PHP_EOL);
|
||||
var_dump($_REQUEST['post2']);
|
||||
print($_REQUEST['cookie1'] . PHP_EOL);
|
||||
var_dump($_REQUEST['cookie2']);
|
||||
|
||||
print($HTTP_RAW_POST_DATA . PHP_EOL);
|
||||
print('end' . PHP_EOL);
|
||||
6
packages/now-php/test/fixtures/11-globals/now.json
vendored
Normal file
6
packages/now-php/test/fixtures/11-globals/now.json
vendored
Normal file
@@ -0,0 +1,6 @@
|
||||
{
|
||||
"version": 2,
|
||||
"builds": [
|
||||
{ "src": "index.php", "use": "@now/php" }
|
||||
]
|
||||
}
|
||||
152
packages/now-php/test/fixtures/11-globals/probe.js
vendored
Normal file
152
packages/now-php/test/fixtures/11-globals/probe.js
vendored
Normal file
@@ -0,0 +1,152 @@
|
||||
const assert = require('assert');
|
||||
|
||||
async function test1({ deploymentUrl, fetch }) {
|
||||
const resp = await fetch(
|
||||
`https://${deploymentUrl}/index.php?get1=foo&get1=bar&get2[]=bim&get2[]=bom`, {
|
||||
headers: {
|
||||
'X-Some-Header': 'x-some-header-value',
|
||||
},
|
||||
},
|
||||
);
|
||||
assert(resp.status === 200);
|
||||
const text = await resp.text();
|
||||
const lines = text.trim().split('\n');
|
||||
|
||||
assert.deepEqual(lines, [
|
||||
'/var/task/user/index.php',
|
||||
'GET',
|
||||
'/index.php?get1=foo&get1=bar&get2%5B%5D=bim&get2%5B%5D=bom', // TODO fake news, must be unescaped
|
||||
deploymentUrl, // example 'test-19phw91ph.now.sh'
|
||||
'x-some-header-value',
|
||||
'HTTP/1.1',
|
||||
deploymentUrl, // example 'test-19phw91ph.now.sh'
|
||||
'443',
|
||||
'on',
|
||||
'bar',
|
||||
'array(2) {',
|
||||
' [0]=>',
|
||||
' string(3) "bim"',
|
||||
' [1]=>',
|
||||
' string(3) "bom"',
|
||||
'}',
|
||||
'',
|
||||
'NULL',
|
||||
'',
|
||||
'NULL',
|
||||
'bar',
|
||||
'array(2) {',
|
||||
' [0]=>',
|
||||
' string(3) "bim"',
|
||||
' [1]=>',
|
||||
' string(3) "bom"',
|
||||
'}',
|
||||
'',
|
||||
'NULL',
|
||||
'',
|
||||
'NULL',
|
||||
'',
|
||||
'end',
|
||||
]);
|
||||
}
|
||||
|
||||
async function test2({ deploymentUrl, fetch }) {
|
||||
const resp = await fetch(
|
||||
`https://${deploymentUrl}/index.php`, {
|
||||
method: 'POST',
|
||||
body: 'post1=baz&post1=bat&post2[]=pim&post2[]=pom',
|
||||
headers: {
|
||||
'Content-Type': 'application/x-www-form-urlencoded',
|
||||
},
|
||||
},
|
||||
);
|
||||
assert(resp.status === 200);
|
||||
const text = await resp.text();
|
||||
const lines = text.trim().split('\n');
|
||||
|
||||
assert.deepEqual(lines, [
|
||||
'/var/task/user/index.php',
|
||||
'POST',
|
||||
'/index.php',
|
||||
deploymentUrl, // example 'test-19phw91ph.now.sh'
|
||||
'',
|
||||
'HTTP/1.1',
|
||||
deploymentUrl, // example 'test-19phw91ph.now.sh'
|
||||
'443',
|
||||
'on',
|
||||
'',
|
||||
'NULL',
|
||||
'bat',
|
||||
'array(2) {',
|
||||
' [0]=>',
|
||||
' string(3) "pim"',
|
||||
' [1]=>',
|
||||
' string(3) "pom"',
|
||||
'}',
|
||||
'',
|
||||
'NULL',
|
||||
'',
|
||||
'NULL',
|
||||
'bat',
|
||||
'array(2) {',
|
||||
' [0]=>',
|
||||
' string(3) "pim"',
|
||||
' [1]=>',
|
||||
' string(3) "pom"',
|
||||
'}',
|
||||
'',
|
||||
'NULL',
|
||||
'post1=baz&post1=bat&post2[]=pim&post2[]=pom',
|
||||
'end',
|
||||
]);
|
||||
}
|
||||
|
||||
async function test3({ deploymentUrl, fetch }) {
|
||||
const resp = await fetch(
|
||||
`https://${deploymentUrl}/index.php`, {
|
||||
method: 'GET',
|
||||
headers: {
|
||||
Cookie: `cookie1=foo; cookie1=${escape('bar|bar')}; ${escape('cookie2[]')}=dim; ${escape('cookie2[]')}=${escape('dom|dom')}`,
|
||||
},
|
||||
},
|
||||
);
|
||||
assert(resp.status === 200);
|
||||
const text = await resp.text();
|
||||
const lines = text.trim().split('\n');
|
||||
|
||||
assert.deepEqual(lines, [
|
||||
'/var/task/user/index.php',
|
||||
'GET',
|
||||
'/index.php',
|
||||
deploymentUrl, // example 'test-19phw91ph.now.sh'
|
||||
'',
|
||||
'HTTP/1.1',
|
||||
deploymentUrl, // example 'test-19phw91ph.now.sh'
|
||||
'443',
|
||||
'on',
|
||||
'',
|
||||
'NULL',
|
||||
'',
|
||||
'NULL',
|
||||
'foo',
|
||||
'array(2) {',
|
||||
' [0]=>',
|
||||
' string(3) "dim"',
|
||||
' [1]=>',
|
||||
' string(7) "dom|dom"',
|
||||
'}',
|
||||
'',
|
||||
'NULL',
|
||||
'',
|
||||
'NULL',
|
||||
'',
|
||||
'NULL',
|
||||
'',
|
||||
'end',
|
||||
]);
|
||||
}
|
||||
|
||||
module.exports = async (opts) => {
|
||||
await test1(opts);
|
||||
await test2(opts);
|
||||
await test3(opts);
|
||||
};
|
||||
5
packages/now-php/test/fixtures/12-setcookie/index.php
vendored
Normal file
5
packages/now-php/test/fixtures/12-setcookie/index.php
vendored
Normal file
@@ -0,0 +1,5 @@
|
||||
<?php
|
||||
header('Content-Type: text/plain');
|
||||
header('Content-Type: text/plain; charset=UTF-16');
|
||||
setcookie('cookie1', 'cookie1value');
|
||||
setcookie('cookie2', 'cookie2value');
|
||||
6
packages/now-php/test/fixtures/12-setcookie/now.json
vendored
Normal file
6
packages/now-php/test/fixtures/12-setcookie/now.json
vendored
Normal file
@@ -0,0 +1,6 @@
|
||||
{
|
||||
"version": 2,
|
||||
"builds": [
|
||||
{ "src": "index.php", "use": "@now/php" }
|
||||
]
|
||||
}
|
||||
9
packages/now-php/test/fixtures/12-setcookie/probe.js
vendored
Normal file
9
packages/now-php/test/fixtures/12-setcookie/probe.js
vendored
Normal file
@@ -0,0 +1,9 @@
|
||||
const assert = require('assert');
|
||||
|
||||
module.exports = async ({ deploymentUrl, fetch }) => {
|
||||
const resp = await fetch(`https://${deploymentUrl}/index.php`);
|
||||
assert(resp.status === 200);
|
||||
assert.equal(resp.headers.get('content-type'), 'text/plain; charset=UTF-16');
|
||||
assert(resp.headers.get('set-cookie').includes('cookie1=cookie1value'));
|
||||
assert(resp.headers.get('set-cookie').includes('cookie2=cookie2value'));
|
||||
};
|
||||
10
packages/now-php/test/fixtures/13-function/index.php
vendored
Normal file
10
packages/now-php/test/fixtures/13-function/index.php
vendored
Normal file
@@ -0,0 +1,10 @@
|
||||
<?php
|
||||
|
||||
// regression test for go-php engine reusage. on failure prints
|
||||
// Fatal error: Cannot redeclare some_function() (previously declared in /var/task/user/index.php:7)
|
||||
|
||||
function some_function() {
|
||||
print("paskantamasaari");
|
||||
}
|
||||
|
||||
some_function();
|
||||
6
packages/now-php/test/fixtures/13-function/now.json
vendored
Normal file
6
packages/now-php/test/fixtures/13-function/now.json
vendored
Normal file
@@ -0,0 +1,6 @@
|
||||
{
|
||||
"version": 2,
|
||||
"builds": [
|
||||
{ "src": "index.php", "use": "@now/php" }
|
||||
]
|
||||
}
|
||||
8
packages/now-php/test/fixtures/13-function/probe.js
vendored
Normal file
8
packages/now-php/test/fixtures/13-function/probe.js
vendored
Normal file
@@ -0,0 +1,8 @@
|
||||
const assert = require('assert');
|
||||
|
||||
module.exports = async ({ deploymentUrl, fetch }) => {
|
||||
const resp1 = await fetch(`https://${deploymentUrl}/index.php`);
|
||||
assert.equal(await resp1.text(), 'paskantamasaari');
|
||||
const resp2 = await fetch(`https://${deploymentUrl}/index.php`);
|
||||
assert.equal(await resp2.text(), 'paskantamasaari');
|
||||
};
|
||||
@@ -8,6 +8,7 @@ const {
|
||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||
|
||||
jest.setTimeout(2 * 60 * 1000);
|
||||
const buildUtilsUrl = '@canary';
|
||||
let builderUrl;
|
||||
|
||||
beforeAll(async () => {
|
||||
@@ -23,7 +24,10 @@ for (const fixture of fs.readdirSync(fixturesPath)) {
|
||||
// eslint-disable-next-line no-loop-func
|
||||
it(`should build ${fixture}`, async () => {
|
||||
await expect(
|
||||
testDeployment({ builderUrl }, path.join(fixturesPath, fixture)),
|
||||
).resolves.toBe(undefined);
|
||||
testDeployment(
|
||||
{ builderUrl, buildUtilsUrl },
|
||||
path.join(fixturesPath, fixture),
|
||||
),
|
||||
).resolves.toBeDefined();
|
||||
});
|
||||
}
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"name": "@now/static-build",
|
||||
"version": "0.4.15",
|
||||
"version": "0.4.16",
|
||||
"peerDependencies": {
|
||||
"@now/build-utils": ">=0.0.1"
|
||||
},
|
||||
|
||||
@@ -6,6 +6,6 @@
|
||||
],
|
||||
"probes": [
|
||||
{ "path": "/", "mustContain": "cow:RANDOMNESS_PLACEHOLDER" },
|
||||
{ "path": "/subdirectory", "mustContain": "yoda:RANDOMNESS_PLACEHOLDER" }
|
||||
{ "path": "/subdirectory/", "mustContain": "yoda:RANDOMNESS_PLACEHOLDER" }
|
||||
]
|
||||
}
|
||||
|
||||
@@ -6,6 +6,6 @@
|
||||
],
|
||||
"probes": [
|
||||
{ "path": "/", "mustContain": "cow:RANDOMNESS_PLACEHOLDER" },
|
||||
{ "path": "/subdirectory", "mustContain": "yoda:RANDOMNESS_PLACEHOLDER" }
|
||||
{ "path": "/subdirectory/", "mustContain": "yoda:RANDOMNESS_PLACEHOLDER" }
|
||||
]
|
||||
}
|
||||
|
||||
@@ -8,6 +8,7 @@ const {
|
||||
} = require('../../../test/lib/deployment/test-deployment.js');
|
||||
|
||||
jest.setTimeout(2 * 60 * 1000);
|
||||
const buildUtilsUrl = '@canary';
|
||||
let builderUrl;
|
||||
|
||||
beforeAll(async () => {
|
||||
@@ -23,7 +24,10 @@ for (const fixture of fs.readdirSync(fixturesPath)) {
|
||||
// eslint-disable-next-line no-loop-func
|
||||
it(`should build ${fixture}`, async () => {
|
||||
await expect(
|
||||
testDeployment({ builderUrl }, path.join(fixturesPath, fixture)),
|
||||
).resolves.toBe(undefined);
|
||||
testDeployment(
|
||||
{ builderUrl, buildUtilsUrl },
|
||||
path.join(fixturesPath, fixture),
|
||||
),
|
||||
).resolves.toBeDefined();
|
||||
});
|
||||
}
|
||||
|
||||
@@ -41,3 +41,14 @@ it('Should throw when package.json or next.config.js is not the "src"', async ()
|
||||
expect(err.message).toMatch(/package\.json/);
|
||||
}
|
||||
});
|
||||
|
||||
it(
|
||||
'Should build the static-files test',
|
||||
async () => {
|
||||
const { buildResult } = await runBuildLambda(
|
||||
path.join(__dirname, 'static-files'),
|
||||
);
|
||||
expect(buildResult['static/test.txt']).toBeDefined();
|
||||
},
|
||||
FOUR_MINUTES,
|
||||
);
|
||||
|
||||
1
test/integration/now-next/static-files/next.config.js
Normal file
1
test/integration/now-next/static-files/next.config.js
Normal file
@@ -0,0 +1 @@
|
||||
module.exports = {};
|
||||
6
test/integration/now-next/static-files/now.json
Normal file
6
test/integration/now-next/static-files/now.json
Normal file
@@ -0,0 +1,6 @@
|
||||
{
|
||||
"version": 2,
|
||||
"builds": [
|
||||
{"src": "next.config.js", "use": "@now/next"}
|
||||
]
|
||||
}
|
||||
1
test/integration/now-next/static-files/pages/index.js
Normal file
1
test/integration/now-next/static-files/pages/index.js
Normal file
@@ -0,0 +1 @@
|
||||
export default () => 'Index page';
|
||||
1
test/integration/now-next/static-files/static/test.txt
Normal file
1
test/integration/now-next/static-files/static/test.txt
Normal file
@@ -0,0 +1 @@
|
||||
hello world
|
||||
@@ -19,6 +19,7 @@
|
||||
"no-restricted-syntax": "off",
|
||||
"no-return-await": "off",
|
||||
"no-use-before-define": "off",
|
||||
"prefer-destructuring": "off",
|
||||
"space-before-function-paren": [
|
||||
"error",
|
||||
"always"
|
||||
|
||||
23
test/lib/deployment/fetch-retry.js
Normal file
23
test/lib/deployment/fetch-retry.js
Normal file
@@ -0,0 +1,23 @@
|
||||
const fetch = require('node-fetch');
|
||||
const retryBailByDefault = require('./retry-bail-by-default.js');
|
||||
|
||||
async function fetchRetry (...args) {
|
||||
return await retryBailByDefault(async (canRetry) => {
|
||||
try {
|
||||
return await fetch(...args);
|
||||
} catch (error) {
|
||||
if (error.code === 'ENOTFOUND') {
|
||||
// getaddrinfo ENOTFOUND api.zeit.co like some transient dns issue
|
||||
throw canRetry(error);
|
||||
} else
|
||||
if (error.code === 'ETIMEDOUT') {
|
||||
// request to https://api-gru1.zeit.co/v3/now/deployments/dpl_FBWWhpQomjgwjJLu396snLrGZYCm failed, reason:
|
||||
// connect ETIMEDOUT 18.228.143.224:443
|
||||
throw canRetry(error);
|
||||
}
|
||||
throw error;
|
||||
}
|
||||
}, { factor: 1, retries: 3 });
|
||||
}
|
||||
|
||||
module.exports = fetchRetry;
|
||||
@@ -1,11 +1,8 @@
|
||||
const assert = require('assert');
|
||||
const { createHash } = require('crypto');
|
||||
const fetch = require('node-fetch');
|
||||
const fs = require('fs-extra');
|
||||
const { homedir } = require('os');
|
||||
const path = require('path');
|
||||
|
||||
const API_URL = 'https://api.zeit.co';
|
||||
const fetch = require('./fetch-retry.js');
|
||||
|
||||
async function nowDeploy (bodies, randomness) {
|
||||
const files = Object.keys(bodies)
|
||||
@@ -21,7 +18,7 @@ async function nowDeploy (bodies, randomness) {
|
||||
|
||||
const nowDeployPayload = {
|
||||
version: 2,
|
||||
env: { RANDOMNESS_ENV_VAR: randomness },
|
||||
env: Object.assign({}, nowJson.env, { RANDOMNESS_ENV_VAR: randomness }),
|
||||
build: { env: { RANDOMNESS_BUILD_ENV_VAR: randomness } },
|
||||
name: 'test',
|
||||
files,
|
||||
@@ -30,12 +27,13 @@ async function nowDeploy (bodies, randomness) {
|
||||
meta: {}
|
||||
};
|
||||
|
||||
console.log(`posting ${files.length} files`);
|
||||
|
||||
for (const { file: filename } of files) {
|
||||
const json = await filePost(
|
||||
await filePost(
|
||||
bodies[filename],
|
||||
digestOfFile(bodies[filename])
|
||||
);
|
||||
if (json.error) throw new Error(json.error.message);
|
||||
}
|
||||
|
||||
let deploymentId;
|
||||
@@ -48,6 +46,8 @@ async function nowDeploy (bodies, randomness) {
|
||||
deploymentUrl = json.url;
|
||||
}
|
||||
|
||||
console.log('id', deploymentId);
|
||||
|
||||
for (let i = 0; i < 500; i += 1) {
|
||||
const { state } = await deploymentGet(deploymentId);
|
||||
if (state === 'ERROR') throw new Error(`State of ${deploymentUrl} is ${state}`);
|
||||
@@ -55,7 +55,7 @@ async function nowDeploy (bodies, randomness) {
|
||||
await new Promise((r) => setTimeout(r, 1000));
|
||||
}
|
||||
|
||||
return deploymentUrl;
|
||||
return { deploymentId, deploymentUrl };
|
||||
}
|
||||
|
||||
function digestOfFile (body) {
|
||||
@@ -74,56 +74,77 @@ async function filePost (body, digest) {
|
||||
'x-now-size': body.length
|
||||
};
|
||||
|
||||
const resp = await fetchWithAuth(`${API_URL}/v2/now/files`, {
|
||||
const resp = await fetchWithAuth('/v2/now/files', {
|
||||
method: 'POST',
|
||||
headers,
|
||||
body
|
||||
});
|
||||
const json = await resp.json();
|
||||
|
||||
return await resp.json();
|
||||
if (json.error) {
|
||||
console.log('headers', resp.headers);
|
||||
throw new Error(json.error.message);
|
||||
}
|
||||
return json;
|
||||
}
|
||||
|
||||
async function deploymentPost (payload) {
|
||||
const resp = await fetchWithAuth(`${API_URL}/v6/now/deployments?forceNew=1`, {
|
||||
const resp = await fetchWithAuth('/v6/now/deployments?forceNew=1', {
|
||||
method: 'POST',
|
||||
body: JSON.stringify(payload)
|
||||
});
|
||||
const json = await resp.json();
|
||||
if (json.error) throw new Error(json.error.message);
|
||||
|
||||
if (json.error) {
|
||||
console.log('headers', resp.headers);
|
||||
throw new Error(json.error.message);
|
||||
}
|
||||
return json;
|
||||
}
|
||||
|
||||
async function deploymentGet (deploymentId) {
|
||||
const resp = await fetchWithAuth(
|
||||
`${API_URL}/v3/now/deployments/${deploymentId}`
|
||||
);
|
||||
const resp = await fetchWithAuth(`/v3/now/deployments/${deploymentId}`);
|
||||
return await resp.json();
|
||||
}
|
||||
|
||||
async function fetchWithAuth (url, opts = {}) {
|
||||
if (!opts.headers) opts.headers = {};
|
||||
const authJsonPath = path.join(homedir(), '.now/auth.json');
|
||||
if (!(await fs.exists(authJsonPath))) {
|
||||
await fs.mkdirp(path.dirname(authJsonPath));
|
||||
await fs.writeFile(
|
||||
authJsonPath,
|
||||
JSON.stringify({
|
||||
token: process.env.NOW_AUTH_TOKEN
|
||||
})
|
||||
);
|
||||
|
||||
if (!opts.headers.Authorization) {
|
||||
let token;
|
||||
if (process.env.NOW_AUTH_TOKENS) {
|
||||
const tokens = process.env.NOW_AUTH_TOKENS.split(',');
|
||||
if (process.env.CIRCLE_BUILD_NUM) {
|
||||
token = tokens[Number(process.env.CIRCLE_BUILD_NUM) % tokens.length];
|
||||
} else {
|
||||
token = tokens[Math.floor(Math.random() * tokens.length)];
|
||||
}
|
||||
} else {
|
||||
const authJsonPath = path.join(homedir(), '.now/auth.json');
|
||||
token = require(authJsonPath).token;
|
||||
}
|
||||
|
||||
opts.headers.Authorization = `Bearer ${token}`;
|
||||
}
|
||||
|
||||
const { token } = require(authJsonPath);
|
||||
opts.headers.Authorization = `Bearer ${token}`;
|
||||
return await fetchApiWithChecks(url, opts);
|
||||
return await fetchApi(url, opts);
|
||||
}
|
||||
|
||||
async function fetchApiWithChecks (url, opts = {}) {
|
||||
// const { method = 'GET', body } = opts;
|
||||
// console.log('fetch', method, url);
|
||||
// if (body) console.log(encodeURIComponent(body).slice(0, 80));
|
||||
const resp = await fetch(url, opts);
|
||||
return resp;
|
||||
async function fetchApi (url, opts = {}) {
|
||||
const apiHost = process.env.API_HOST || 'api.zeit.co';
|
||||
const urlWithHost = `https://${apiHost}${url}`;
|
||||
const { method = 'GET', body } = opts;
|
||||
|
||||
if (process.env.VERBOSE) {
|
||||
console.log('fetch', method, url);
|
||||
if (body) console.log(encodeURIComponent(body).slice(0, 80));
|
||||
}
|
||||
|
||||
return await fetch(urlWithHost, opts);
|
||||
}
|
||||
|
||||
module.exports = nowDeploy;
|
||||
module.exports = {
|
||||
fetchApi,
|
||||
fetchWithAuth,
|
||||
nowDeploy
|
||||
};
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user