60af1969c5
Noticed a couple of directories had unit tests in `dist` which are being run and reported during the Turbopack runs. These are generally quite small so I'm not expecting this to have much effect on the package size or CI runs, just making sure the data is correct. <!-- Thanks for opening a PR! Your contribution is much appreciated. To make sure your PR is handled as smoothly as possible we request that you follow the checklist sections below. Choose the right checklist for the change(s) that you're making: ## For Contributors ### Improving Documentation - Run `pnpm prettier-fix` to fix formatting issues before opening the PR. - Read the Docs Contribution Guide to ensure your contribution follows the docs guidelines: https://nextjs.org/docs/community/contribution-guide ### Adding or Updating Examples - The "examples guidelines" are followed from our contributing doc https://github.com/vercel/next.js/blob/canary/contributing/examples/adding-examples.md - Make sure the linting passes by running `pnpm build && pnpm lint`. See https://github.com/vercel/next.js/blob/canary/contributing/repository/linting.md ### Fixing a bug - Related issues linked using `fixes #number` - Tests added. See: https://github.com/vercel/next.js/blob/canary/contributing/core/testing.md#writing-tests-for-nextjs - Errors have a helpful link attached, see https://github.com/vercel/next.js/blob/canary/contributing.md ### Adding a feature - Implements an existing feature request or RFC. Make sure the feature request has been accepted for implementation before opening a PR. (A discussion must be opened, see https://github.com/vercel/next.js/discussions/new?category=ideas) - Related issues/discussions are linked using `fixes #number` - e2e tests added (https://github.com/vercel/next.js/blob/canary/contributing/core/testing.md#writing-tests-for-nextjs) - Documentation added - Telemetry added. In case of a feature if it's used or not. - Errors have a helpful link attached, see https://github.com/vercel/next.js/blob/canary/contributing.md ## For Maintainers - Minimal description (aim for explaining to someone not on the team to understand the PR) - When linking to a Slack thread, you might want to share details of the conclusion - Link both the Linear (Fixes NEXT-xxx) and the GitHub issues - Add review comments if necessary to explain to the reviewer the logic behind a change ### What? ### Why? ### How? Closes NEXT- Fixes # -->
139 lines
4.5 KiB
JavaScript
139 lines
4.5 KiB
JavaScript
const fetch = require('node-fetch')
|
|
const fs = require('fs')
|
|
const prettier = require('prettier')
|
|
|
|
async function format(text) {
|
|
const options = await prettier.resolveConfig(__filename)
|
|
return prettier.format(text, { ...options, parser: 'json' })
|
|
}
|
|
|
|
const override = process.argv.includes('--override')
|
|
|
|
// TODO: Switch to nextjs-integration-test-data branch once https://github.com/vercel/turbo/pull/5999 is merged.
|
|
const RESULT_URL =
|
|
'https://raw.githubusercontent.com/vercel/turbo/nextjs-integration-test-data/test-results/main/nextjs-test-results.json'
|
|
const PASSING_JSON_PATH = `${__dirname}/turbopack-tests-manifest.json`
|
|
const WORKING_PATH = '/home/runner/work/turbo/turbo/'
|
|
|
|
const INITIALIZING_TEST_CASES = [
|
|
'compile successfully',
|
|
'should build successfully',
|
|
]
|
|
|
|
const SKIPPED_TEST_SUITES = new Set([
|
|
'test/integration/router-rerender/test/index.test.js',
|
|
'test/e2e/basepath.test.ts',
|
|
'test/development/acceptance-app/ReactRefreshRequire.test.ts',
|
|
'test/integration/dynamic-routing/test/middleware.test.js',
|
|
'test/integration/css/test/css-modules.test.js',
|
|
'test/development/acceptance/ReactRefreshRequire.test.ts',
|
|
'test/integration/custom-routes/test/index.test.js',
|
|
'test/integration/absolute-assetprefix/test/index.test.js',
|
|
'test/e2e/middleware-rewrites/test/index.test.ts',
|
|
])
|
|
|
|
async function updatePassingTests() {
|
|
const passing = { __proto__: null }
|
|
const res = await fetch(RESULT_URL)
|
|
const results = await res.json()
|
|
|
|
for (const result of results.result) {
|
|
const runtimeError = result.data.numRuntimeErrorTestSuites > 0
|
|
for (const testResult of result.data.testResults) {
|
|
const filepath = stripWorkingPath(testResult.name)
|
|
|
|
if (SKIPPED_TEST_SUITES.has(filepath)) continue
|
|
const fileResults = (passing[filepath] ??= {
|
|
passed: [],
|
|
failed: [],
|
|
pending: [],
|
|
runtimeError,
|
|
})
|
|
|
|
let initializationFailed = false
|
|
for (const testCase of testResult.assertionResults) {
|
|
let { fullName, status } = testCase
|
|
if (
|
|
status === 'failed' &&
|
|
INITIALIZING_TEST_CASES.some((name) => fullName.includes(name))
|
|
) {
|
|
initializationFailed = true
|
|
} else if (initializationFailed) {
|
|
status = 'failed'
|
|
}
|
|
const statusArray = fileResults[status]
|
|
if (!statusArray) {
|
|
throw new Error(`unexpected status "${status}"`)
|
|
}
|
|
statusArray.push(fullName)
|
|
}
|
|
}
|
|
}
|
|
|
|
for (const info of Object.values(passing)) {
|
|
info.failed = [...new Set(info.failed)].sort()
|
|
info.pending = [...new Set(info.pending)].sort()
|
|
info.passed = [
|
|
...new Set(info.passed.filter((name) => !info.failed.includes(name))),
|
|
].sort()
|
|
}
|
|
|
|
if (!override) {
|
|
const oldPassingData = JSON.parse(
|
|
fs.readFileSync(PASSING_JSON_PATH, 'utf8')
|
|
)
|
|
|
|
for (const file of Object.keys(oldPassingData)) {
|
|
const newData = passing[file]
|
|
const oldData = oldPassingData[file]
|
|
if (!newData) continue
|
|
// We only want to keep test cases from the old data that are still exiting
|
|
oldData.passed = oldData.passed.filter(
|
|
(name) => newData.failed.includes(name) || newData.passed.includes(name)
|
|
)
|
|
// Grab test cases that passed before, but fail now
|
|
const shouldPass = new Set(
|
|
oldData.passed.filter((name) => newData.failed.includes(name))
|
|
)
|
|
if (shouldPass.size > 0) {
|
|
const list = JSON.stringify([...shouldPass], 0, 2)
|
|
console.log(
|
|
`${file} has ${shouldPass.size} test(s) that should pass but failed: ${list}`
|
|
)
|
|
}
|
|
// Merge the old passing tests with the new ones
|
|
newData.passed = [
|
|
...new Set([...oldData.passed, ...newData.passed]),
|
|
].sort()
|
|
// but remove them also from the failed list
|
|
newData.failed = newData.failed
|
|
.filter((name) => !shouldPass.has(name))
|
|
.sort()
|
|
}
|
|
}
|
|
|
|
// JS keys are ordered, this ensures the tests are written in a consistent order
|
|
// https://stackoverflow.com/questions/5467129/sort-javascript-object-by-key
|
|
const ordered = Object.keys(passing)
|
|
.sort()
|
|
.reduce((obj, key) => {
|
|
obj[key] = passing[key]
|
|
return obj
|
|
}, {})
|
|
|
|
fs.writeFileSync(
|
|
PASSING_JSON_PATH,
|
|
await format(JSON.stringify(ordered, null, 2))
|
|
)
|
|
}
|
|
|
|
function stripWorkingPath(path) {
|
|
if (!path.startsWith(WORKING_PATH)) {
|
|
throw new Error(
|
|
`found unexpected working path in "${path}", expected it to begin with ${WORKING_PATH}`
|
|
)
|
|
}
|
|
return path.slice(WORKING_PATH.length)
|
|
}
|
|
|
|
updatePassingTests()
|