Compare commits
4 Commits
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
495b52661e | ||
|
|
9a6aa52a40 | ||
|
|
c106903f01 | ||
|
|
d40cda150c |
4
.github/workflows/e2e-tests.yml
vendored
4
.github/workflows/e2e-tests.yml
vendored
@@ -26,7 +26,7 @@ jobs:
|
||||
|
||||
- name: Start containers
|
||||
working-directory: extensions/ql-vscode/test/e2e
|
||||
run: docker compose -f "docker-compose.yml" up -d --build
|
||||
run: docker-compose -f "docker-compose.yml" up -d --build
|
||||
|
||||
- name: Install Playwright Browsers
|
||||
working-directory: extensions/ql-vscode
|
||||
@@ -43,4 +43,4 @@ jobs:
|
||||
- name: Stop containers
|
||||
working-directory: extensions/ql-vscode/test/e2e
|
||||
if: always()
|
||||
run: docker compose -f "docker-compose.yml" down -v
|
||||
run: docker-compose -f "docker-compose.yml" down -v
|
||||
|
||||
22
.github/workflows/main.yml
vendored
22
.github/workflows/main.yml
vendored
@@ -79,6 +79,13 @@ jobs:
|
||||
run: |
|
||||
npm run check-types
|
||||
|
||||
- name: Lint
|
||||
working-directory: extensions/ql-vscode
|
||||
env:
|
||||
NODE_OPTIONS: '--max-old-space-size=4096'
|
||||
run: |
|
||||
npm run lint
|
||||
|
||||
- name: Lint Markdown
|
||||
working-directory: extensions/ql-vscode
|
||||
run: |
|
||||
@@ -94,21 +101,6 @@ jobs:
|
||||
run: |
|
||||
npm run find-deadcode
|
||||
|
||||
- name: Lint
|
||||
if: "${{ !cancelled() }}"
|
||||
working-directory: extensions/ql-vscode
|
||||
env:
|
||||
NODE_OPTIONS: '--max-old-space-size=4096'
|
||||
run: |
|
||||
npm run lint-ci
|
||||
|
||||
- name: Upload ESLint results to Code Scanning
|
||||
if: "${{ !cancelled() && !startsWith(github.head_ref, 'dependabot/')}}"
|
||||
uses: github/codeql-action/upload-sarif@main
|
||||
with:
|
||||
sarif_file: extensions/ql-vscode/build/eslint.sarif
|
||||
category: eslint
|
||||
|
||||
generated:
|
||||
name: Check generated code
|
||||
runs-on: ubuntu-latest
|
||||
|
||||
14
.github/workflows/release.yml
vendored
14
.github/workflows/release.yml
vendored
@@ -127,13 +127,6 @@ jobs:
|
||||
env:
|
||||
VSCE_TOKEN: ${{ secrets.VSCE_TOKEN }}
|
||||
steps:
|
||||
- name: Checkout
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version-file: extensions/ql-vscode/.nvmrc
|
||||
|
||||
- name: Download artifact
|
||||
uses: actions/download-artifact@v4
|
||||
with:
|
||||
@@ -151,13 +144,6 @@ jobs:
|
||||
env:
|
||||
OPEN_VSX_TOKEN: ${{ secrets.OPEN_VSX_TOKEN }}
|
||||
steps:
|
||||
- name: Checkout
|
||||
uses: actions/checkout@v4
|
||||
|
||||
- uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version-file: extensions/ql-vscode/.nvmrc
|
||||
|
||||
- name: Download artifact
|
||||
uses: actions/download-artifact@v4
|
||||
with:
|
||||
|
||||
@@ -1 +1,4 @@
|
||||
#!/usr/bin/env sh
|
||||
. "$(dirname -- "$0")/_/husky.sh"
|
||||
|
||||
cd extensions/ql-vscode && npm run format-staged
|
||||
|
||||
@@ -1 +1,4 @@
|
||||
#!/usr/bin/env sh
|
||||
. "$(dirname -- "$0")/_/husky.sh"
|
||||
|
||||
cd extensions/ql-vscode && ./scripts/forbid-test-only
|
||||
|
||||
6
.vscode/settings.json
vendored
6
.vscode/settings.json
vendored
@@ -111,4 +111,10 @@
|
||||
"editor.defaultFormatter": "esbenp.prettier-vscode",
|
||||
"editor.formatOnSave": true,
|
||||
},
|
||||
"github.copilot.advanced": {
|
||||
|
||||
|
||||
},
|
||||
"codeQL.variantAnalysis.enableGhecDr": true,
|
||||
"github-enterprise.uri": "http://server:8080/"
|
||||
}
|
||||
|
||||
@@ -3,14 +3,5 @@ node_modules/
|
||||
out/
|
||||
build/
|
||||
|
||||
# Ignore js files
|
||||
.eslintrc.js
|
||||
jest.config.js
|
||||
test/vscode-tests/activated-extension/jest-runner-vscode.config.js
|
||||
test/vscode-tests/cli-integration/jest-runner-vscode.config.js
|
||||
test/vscode-tests/jest-runner-vscode.config.base.js
|
||||
test/vscode-tests/minimal-workspace/jest-runner-vscode.config.js
|
||||
test/vscode-tests/no-workspace/jest-runner-vscode.config.js
|
||||
|
||||
# Include the Storybook config
|
||||
!.storybook
|
||||
|
||||
@@ -45,7 +45,7 @@ const baseConfig = {
|
||||
"@typescript-eslint/no-invalid-this": "off",
|
||||
"@typescript-eslint/no-shadow": "off",
|
||||
"prefer-const": ["warn", { destructuring: "all" }],
|
||||
"@typescript-eslint/only-throw-error": "error",
|
||||
"@typescript-eslint/no-throw-literal": "error",
|
||||
"@typescript-eslint/consistent-type-imports": "error",
|
||||
"import/consistent-type-specifier-style": ["error", "prefer-top-level"],
|
||||
curly: ["error", "all"],
|
||||
@@ -133,7 +133,18 @@ module.exports = {
|
||||
...baseConfig.rules,
|
||||
// We want to allow mocking of functions in modules, so we need to allow namespace imports.
|
||||
"import/no-namespace": "off",
|
||||
"@typescript-eslint/no-unsafe-function-type": "off",
|
||||
"@typescript-eslint/ban-types": [
|
||||
"error",
|
||||
{
|
||||
// For a full list of the default banned types, see:
|
||||
// https://github.com/typescript-eslint/typescript-eslint/blob/master/packages/eslint-plugin/docs/rules/ban-types.md
|
||||
extendDefaults: true,
|
||||
types: {
|
||||
// Don't complain about the `Function` type in test files. (Default is `true`.)
|
||||
Function: false,
|
||||
},
|
||||
},
|
||||
],
|
||||
},
|
||||
},
|
||||
{
|
||||
|
||||
@@ -1 +1 @@
|
||||
v20.16.0
|
||||
v20.9.0
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
"compilerOptions": {
|
||||
"module": "esnext",
|
||||
"moduleResolution": "node",
|
||||
"target": "es2021",
|
||||
"target": "es6",
|
||||
"outDir": "out",
|
||||
"lib": ["ES2021", "dom"],
|
||||
"jsx": "react",
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
import * as React from "react";
|
||||
import { addons } from "@storybook/manager-api";
|
||||
import { Addon_TypesEnum } from "storybook/internal/types";
|
||||
import { Addon_TypesEnum } from "@storybook/types";
|
||||
import { ThemeSelector } from "./ThemeSelector";
|
||||
|
||||
const ADDON_ID = "vscode-theme-addon";
|
||||
|
||||
@@ -1,27 +1,6 @@
|
||||
# CodeQL for Visual Studio Code: Changelog
|
||||
|
||||
## 1.16.1 - 6 November 2024
|
||||
|
||||
- Support result columns of type `QlBuiltins::BigInt` in quick evaluations. [#3647](https://github.com/github/vscode-codeql/pull/3647)
|
||||
- Fix a bug where the CodeQL CLI would be re-downloaded if you switched to a different filesystem (for example Codespaces or a remote SSH host). [#3762](https://github.com/github/vscode-codeql/pull/3762)
|
||||
- Clean up old extension-managed CodeQL CLI distributions. [#3763](https://github.com/github/vscode-codeql/pull/3763)
|
||||
- Only compare the source and sink of a path when comparing alerts of local queries. [#3772](https://github.com/github/vscode-codeql/pull/3772)
|
||||
|
||||
## 1.16.0 - 10 October 2024
|
||||
|
||||
- Increase the required version of VS Code to 1.90.0. [#3737](https://github.com/github/vscode-codeql/pull/3737)
|
||||
- Fix a bug where some variant analysis results failed to download. [#3750](https://github.com/github/vscode-codeql/pull/3750)
|
||||
|
||||
## 1.15.0 - 26 September 2024
|
||||
|
||||
- Update results view to display the length of the shortest path for path queries. [#3687](https://github.com/github/vscode-codeql/pull/3687)
|
||||
- Remove support for CodeQL CLI versions older than 2.16.6. [#3728](https://github.com/github/vscode-codeql/pull/3728)
|
||||
|
||||
## 1.14.0 - 7 August 2024
|
||||
|
||||
- Add Python support to the CodeQL Model Editor. [#3676](https://github.com/github/vscode-codeql/pull/3676)
|
||||
- Update variant analysis view to display the length of the shortest path for path queries. [#3671](https://github.com/github/vscode-codeql/pull/3671)
|
||||
- Remove support for CodeQL CLI versions older than 2.15.5. [#3681](https://github.com/github/vscode-codeql/pull/3681)
|
||||
## [UNRELEASED]
|
||||
|
||||
## 1.13.1 - 29 May 2024
|
||||
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
import { src, dest } from "gulp";
|
||||
// eslint-disable-next-line @typescript-eslint/no-require-imports,import/no-commonjs
|
||||
// eslint-disable-next-line @typescript-eslint/no-var-requires,import/no-commonjs
|
||||
const replace = require("gulp-replace");
|
||||
|
||||
/** Inject the application insights key into the telemetry file */
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
{
|
||||
"chromiumVersion": "122",
|
||||
"electronVersion": "29.4.0"
|
||||
"chromiumVersion": "114",
|
||||
"electronVersion": "25.8.0"
|
||||
}
|
||||
|
||||
@@ -77,8 +77,5 @@ export function copyWasmFiles() {
|
||||
// to configure the path to the WASM file. So, source-map will always load the file from `__dirname/mappings.wasm`.
|
||||
// In version 0.8.0, it may be possible to do this properly by calling SourceMapConsumer.initialize by
|
||||
// using the "browser" field in source-map's package.json to load the WASM file from a given file path.
|
||||
return src("node_modules/source-map/lib/mappings.wasm", {
|
||||
// WASM is a binary format, so don't try to re-encode it as text.
|
||||
encoding: false,
|
||||
}).pipe(dest("out"));
|
||||
return src("node_modules/source-map/lib/mappings.wasm").pipe(dest("out"));
|
||||
}
|
||||
|
||||
11367
extensions/ql-vscode/package-lock.json
generated
11367
extensions/ql-vscode/package-lock.json
generated
File diff suppressed because it is too large
Load Diff
@@ -4,7 +4,7 @@
|
||||
"description": "CodeQL for Visual Studio Code",
|
||||
"author": "GitHub",
|
||||
"private": true,
|
||||
"version": "1.16.1",
|
||||
"version": "1.13.2",
|
||||
"publisher": "GitHub",
|
||||
"license": "MIT",
|
||||
"icon": "media/VS-marketplace-CodeQL-icon.png",
|
||||
@@ -13,8 +13,8 @@
|
||||
"url": "https://github.com/github/vscode-codeql"
|
||||
},
|
||||
"engines": {
|
||||
"vscode": "^1.90.0",
|
||||
"node": "^20.16.0",
|
||||
"vscode": "^1.82.0",
|
||||
"node": "^20.9.0",
|
||||
"npm": ">=7.20.6"
|
||||
},
|
||||
"categories": [
|
||||
@@ -339,13 +339,6 @@
|
||||
"title": "Variant analysis",
|
||||
"order": 5,
|
||||
"properties": {
|
||||
"codeQL.variantAnalysis.controllerRepo": {
|
||||
"type": "string",
|
||||
"default": "",
|
||||
"pattern": "^$|^(?:[a-zA-Z0-9]+-)*[a-zA-Z0-9]+/[a-zA-Z0-9-_]+$",
|
||||
"patternErrorMessage": "Please enter a valid GitHub repository",
|
||||
"markdownDescription": "[For internal use only] The name of the GitHub repository in which the GitHub Actions workflow is run when using the \"Run Variant Analysis\" command. The repository should be of the form `<owner>/<repo>`)."
|
||||
},
|
||||
"codeQL.variantAnalysis.defaultResultsFilter": {
|
||||
"type": "string",
|
||||
"default": "all",
|
||||
@@ -1790,7 +1783,8 @@
|
||||
"when": "false"
|
||||
},
|
||||
{
|
||||
"command": "codeQL.trimCache"
|
||||
"command": "codeQL.trimCache",
|
||||
"when": "codeql.supportsTrimCache"
|
||||
}
|
||||
],
|
||||
"editor/context": [
|
||||
@@ -1930,11 +1924,6 @@
|
||||
{
|
||||
"view": "codeQLEvalLogViewer",
|
||||
"contents": "Run the 'Show Evaluator Log (UI)' command on a CodeQL query run in the Query History view."
|
||||
},
|
||||
{
|
||||
"view": "codeQLVariantAnalysisRepositories",
|
||||
"contents": "Set up a controller repository to start using variant analysis. [Learn more](https://codeql.github.com/docs/codeql-for-visual-studio-code/running-codeql-queries-at-scale-with-mrva#controller-repository) about controller repositories. \n[Set up controller repository](command:codeQLVariantAnalysisRepositories.setupControllerRepository)",
|
||||
"when": "!config.codeQL.variantAnalysis.controllerRepo"
|
||||
}
|
||||
]
|
||||
},
|
||||
@@ -1953,7 +1942,6 @@
|
||||
"update-vscode": "node ./node_modules/vscode/bin/install",
|
||||
"format": "prettier --write **/*.{ts,tsx} && eslint . --ext .ts,.tsx --fix",
|
||||
"lint": "eslint . --ext .js,.ts,.tsx --max-warnings=0",
|
||||
"lint-ci": "SARIF_ESLINT_IGNORE_SUPPRESSED=true eslint . --ext .js,.ts,.tsx --max-warnings=0 --format @microsoft/eslint-formatter-sarif --output-file=build/eslint.sarif",
|
||||
"lint:markdown": "markdownlint-cli2 \"../../**/*.{md,mdx}\" \"!**/node_modules/**\" \"!**/.vscode-test/**\" \"!**/build/cli/v*/**\"",
|
||||
"find-deadcode": "vite-node scripts/find-deadcode.ts",
|
||||
"format-staged": "lint-staged",
|
||||
@@ -1963,18 +1951,18 @@
|
||||
"generate": "npm-run-all -p generate:*",
|
||||
"generate:schemas": "vite-node scripts/generate-schemas.ts",
|
||||
"generate:chromium-version": "vite-node scripts/generate-chromium-version.ts",
|
||||
"check-types": "find . -type f -name \"tsconfig.json\" -not -path \"./node_modules/*\" -not -path \"*/.vscode-test/*\" | sed -r 's|/[^/]+$||' | sort | uniq | xargs -I {} sh -c \"echo Checking types in {} && cd {} && npx tsc --noEmit\"",
|
||||
"check-types": "find . -type f -name \"tsconfig.json\" -not -path \"./node_modules/*\" | sed -r 's|/[^/]+$||' | sort | uniq | xargs -I {} sh -c \"echo Checking types in {} && cd {} && npx tsc --noEmit\"",
|
||||
"postinstall": "patch-package",
|
||||
"prepare": "cd ../.. && husky"
|
||||
"prepare": "cd ../.. && husky install"
|
||||
},
|
||||
"dependencies": {
|
||||
"@floating-ui/react": "^0.26.24",
|
||||
"@octokit/plugin-retry": "^7.1.2",
|
||||
"@octokit/plugin-throttling": "^9.3.2",
|
||||
"@octokit/rest": "^21.0.2",
|
||||
"@vscode/codicons": "^0.0.36",
|
||||
"@floating-ui/react": "^0.26.12",
|
||||
"@octokit/plugin-retry": "^6.0.1",
|
||||
"@octokit/plugin-throttling": "^8.0.0",
|
||||
"@octokit/rest": "^20.0.2",
|
||||
"@vscode/codicons": "^0.0.35",
|
||||
"@vscode/debugadapter": "^1.59.0",
|
||||
"@vscode/debugprotocol": "^1.68.0",
|
||||
"@vscode/debugprotocol": "^1.65.0",
|
||||
"@vscode/webview-ui-toolkit": "^1.0.1",
|
||||
"ajv": "^8.11.0",
|
||||
"child-process-promise": "^2.2.1",
|
||||
@@ -1985,8 +1973,8 @@
|
||||
"js-yaml": "^4.1.0",
|
||||
"msw": "^2.2.13",
|
||||
"nanoid": "^5.0.7",
|
||||
"node-fetch": "^2.6.7",
|
||||
"p-queue": "^8.0.1",
|
||||
"proper-lockfile": "^4.1.2",
|
||||
"react": "^18.3.1",
|
||||
"react-dom": "^18.3.1",
|
||||
"semver": "^7.6.2",
|
||||
@@ -1998,37 +1986,36 @@
|
||||
"tmp-promise": "^3.0.2",
|
||||
"tree-kill": "^1.2.2",
|
||||
"vscode-extension-telemetry": "^0.1.6",
|
||||
"vscode-jsonrpc": "^8.2.1",
|
||||
"vscode-jsonrpc": "^8.0.2",
|
||||
"vscode-languageclient": "^8.0.2",
|
||||
"yauzl": "^2.10.0",
|
||||
"zip-a-folder": "^3.1.6"
|
||||
},
|
||||
"devDependencies": {
|
||||
"@babel/core": "^7.24.6",
|
||||
"@babel/plugin-transform-modules-commonjs": "^7.24.7",
|
||||
"@babel/plugin-transform-modules-commonjs": "^7.18.6",
|
||||
"@babel/preset-env": "^7.24.4",
|
||||
"@babel/preset-react": "^7.25.9",
|
||||
"@babel/preset-react": "^7.24.1",
|
||||
"@babel/preset-typescript": "^7.21.4",
|
||||
"@faker-js/faker": "^9.0.3",
|
||||
"@github/markdownlint-github": "^0.6.3",
|
||||
"@microsoft/eslint-formatter-sarif": "^3.1.0",
|
||||
"@faker-js/faker": "^8.4.1",
|
||||
"@github/markdownlint-github": "^0.6.2",
|
||||
"@playwright/test": "^1.40.1",
|
||||
"@storybook/addon-a11y": "^8.4.0",
|
||||
"@storybook/addon-actions": "^8.4.0",
|
||||
"@storybook/addon-essentials": "^8.4.0",
|
||||
"@storybook/addon-interactions": "^8.4.0",
|
||||
"@storybook/addon-links": "^8.4.0",
|
||||
"@storybook/addon-a11y": "^8.1.10",
|
||||
"@storybook/addon-actions": "^8.1.10",
|
||||
"@storybook/addon-essentials": "^8.1.10",
|
||||
"@storybook/addon-interactions": "^8.1.10",
|
||||
"@storybook/addon-links": "^8.1.10",
|
||||
"@storybook/blocks": "^8.0.2",
|
||||
"@storybook/components": "^8.4.0",
|
||||
"@storybook/csf": "^0.1.11",
|
||||
"@storybook/icons": "^1.2.12",
|
||||
"@storybook/manager-api": "^8.4.0",
|
||||
"@storybook/react": "^8.4.0",
|
||||
"@storybook/react-vite": "^8.4.0",
|
||||
"@storybook/theming": "^8.2.4",
|
||||
"@testing-library/dom": "^10.4.0",
|
||||
"@testing-library/jest-dom": "^6.6.2",
|
||||
"@testing-library/react": "^16.0.1",
|
||||
"@storybook/components": "^8.0.2",
|
||||
"@storybook/csf": "^0.1.8",
|
||||
"@storybook/icons": "^1.2.9",
|
||||
"@storybook/manager-api": "^8.1.10",
|
||||
"@storybook/react": "^8.1.10",
|
||||
"@storybook/react-vite": "^8.1.10",
|
||||
"@storybook/theming": "^8.1.10",
|
||||
"@testing-library/dom": "^10.1.0",
|
||||
"@testing-library/jest-dom": "^6.4.6",
|
||||
"@testing-library/react": "^16.0.0",
|
||||
"@testing-library/user-event": "^14.5.2",
|
||||
"@types/child-process-promise": "^2.2.1",
|
||||
"@types/d3": "^7.4.0",
|
||||
@@ -2039,8 +2026,9 @@
|
||||
"@types/gulp-replace": "^1.1.0",
|
||||
"@types/jest": "^29.5.12",
|
||||
"@types/js-yaml": "^4.0.6",
|
||||
"@types/node": "20.16.*",
|
||||
"@types/proper-lockfile": "^4.1.4",
|
||||
"@types/nanoid": "^3.0.0",
|
||||
"@types/node": "20.9.*",
|
||||
"@types/node-fetch": "^2.5.2",
|
||||
"@types/react": "^18.3.1",
|
||||
"@types/react-dom": "^18.3.0",
|
||||
"@types/sarif": "^2.1.2",
|
||||
@@ -2050,10 +2038,10 @@
|
||||
"@types/tar-stream": "^3.1.3",
|
||||
"@types/through2": "^2.0.36",
|
||||
"@types/tmp": "^0.2.6",
|
||||
"@types/vscode": "1.90.0",
|
||||
"@types/vscode": "^1.82.0",
|
||||
"@types/yauzl": "^2.10.3",
|
||||
"@typescript-eslint/eslint-plugin": "^8.12.2",
|
||||
"@typescript-eslint/parser": "^8.12.2",
|
||||
"@typescript-eslint/eslint-plugin": "^7.5.0",
|
||||
"@typescript-eslint/parser": "^7.5.0",
|
||||
"@vscode/test-electron": "^2.3.9",
|
||||
"@vscode/vsce": "^2.24.0",
|
||||
"ansi-colors": "^4.1.1",
|
||||
@@ -2064,40 +2052,40 @@
|
||||
"eslint": "^8.56.0",
|
||||
"eslint-config-prettier": "^9.0.0",
|
||||
"eslint-import-resolver-typescript": "^3.6.1",
|
||||
"eslint-plugin-deprecation": "^3.0.0",
|
||||
"eslint-plugin-deprecation": "^2.0.0",
|
||||
"eslint-plugin-etc": "^2.0.2",
|
||||
"eslint-plugin-github": "^5.0.1",
|
||||
"eslint-plugin-import": "^2.31.0",
|
||||
"eslint-plugin-jest-dom": "^5.4.0",
|
||||
"eslint-plugin-github": "^4.10.2",
|
||||
"eslint-plugin-import": "^2.29.1",
|
||||
"eslint-plugin-jest-dom": "^5.2.0",
|
||||
"eslint-plugin-prettier": "^5.1.3",
|
||||
"eslint-plugin-react": "^7.37.1",
|
||||
"eslint-plugin-react-hooks": "^4.6.2",
|
||||
"eslint-plugin-react": "^7.34.1",
|
||||
"eslint-plugin-react-hooks": "^4.6.0",
|
||||
"eslint-plugin-storybook": "^0.8.0",
|
||||
"glob": "^11.0.0",
|
||||
"gulp": "^5.0.0",
|
||||
"gulp-esbuild": "^0.12.1",
|
||||
"glob": "^10.0.0",
|
||||
"gulp": "^4.0.2",
|
||||
"gulp-esbuild": "^0.12.0",
|
||||
"gulp-replace": "^1.1.3",
|
||||
"gulp-typescript": "^5.0.1",
|
||||
"husky": "^9.1.5",
|
||||
"husky": "^9.1.7",
|
||||
"jest": "^29.0.3",
|
||||
"jest-environment-jsdom": "^29.0.3",
|
||||
"jest-runner-vscode": "^3.0.1",
|
||||
"lint-staged": "^15.2.10",
|
||||
"lint-staged": "^15.2.2",
|
||||
"markdownlint-cli2": "^0.13.0",
|
||||
"markdownlint-cli2-formatter-pretty": "^0.0.6",
|
||||
"npm-run-all": "^4.1.5",
|
||||
"patch-package": "^8.0.0",
|
||||
"prettier": "^3.2.5",
|
||||
"storybook": "^8.4.0",
|
||||
"storybook": "^8.1.10",
|
||||
"tar-stream": "^3.1.7",
|
||||
"through2": "^4.0.2",
|
||||
"ts-jest": "^29.2.5",
|
||||
"ts-json-schema-generator": "^2.3.0",
|
||||
"ts-jest": "^29.1.4",
|
||||
"ts-json-schema-generator": "^2.1.1",
|
||||
"ts-node": "^10.9.2",
|
||||
"ts-unused-exports": "^10.1.0",
|
||||
"typescript": "^5.6.2",
|
||||
"vite": "^5.4.6",
|
||||
"vite-node": "^2.0.5"
|
||||
"typescript": "^5.0.2",
|
||||
"vite": "^5.2.11",
|
||||
"vite-node": "^1.5.3"
|
||||
},
|
||||
"lint-staged": {
|
||||
"./**/*.{json,css,scss}": [
|
||||
|
||||
@@ -1,3 +1,29 @@
|
||||
diff --git a/node_modules/jest-runner-vscode/dist/child/environment.js b/node_modules/jest-runner-vscode/dist/child/environment.js
|
||||
index 1ac28d5..f91f216 100644
|
||||
--- a/node_modules/jest-runner-vscode/dist/child/environment.js
|
||||
+++ b/node_modules/jest-runner-vscode/dist/child/environment.js
|
||||
@@ -10,6 +10,21 @@ const wrap_io_1 = __importDefault(require("./wrap-io"));
|
||||
const load_pnp_1 = __importDefault(require("./load-pnp"));
|
||||
const ipc = new ipc_client_1.default('env');
|
||||
class VSCodeEnvironment extends jest_environment_node_1.default {
|
||||
+ constructor(config, context) {
|
||||
+ super(config, context);
|
||||
+ // The _VSCODE_NODE_MODULES is a proxy which will require a module if any property
|
||||
+ // on it is accessed. This is a workaround for the fact that jest will call
|
||||
+ // _isMockFunction on the module, which will cause that function to be required.
|
||||
+ this.global._VSCODE_NODE_MODULES = new Proxy(this.global._VSCODE_NODE_MODULES, {
|
||||
+ get(target, prop) {
|
||||
+ if (prop === '_isMockFunction') {
|
||||
+ return undefined;
|
||||
+ }
|
||||
+ return target[prop];
|
||||
+ },
|
||||
+ });
|
||||
+ }
|
||||
+
|
||||
async setup() {
|
||||
await super.setup();
|
||||
await (0, load_pnp_1.default)();
|
||||
diff --git a/node_modules/jest-runner-vscode/dist/child/runner.js b/node_modules/jest-runner-vscode/dist/child/runner.js
|
||||
index 0663c5c..bdf4a8b 100644
|
||||
--- a/node_modules/jest-runner-vscode/dist/child/runner.js
|
||||
|
||||
@@ -41,7 +41,6 @@ async function extractSourceMap() {
|
||||
const releaseAssetsDirectory = resolve(
|
||||
__dirname,
|
||||
"..",
|
||||
"artifacts",
|
||||
"release-assets",
|
||||
versionNumber,
|
||||
);
|
||||
@@ -65,9 +64,7 @@ async function extractSourceMap() {
|
||||
]);
|
||||
|
||||
const sourcemapAsset = release.assets.find(
|
||||
(asset) =>
|
||||
asset.label === `vscode-codeql-sourcemaps-${versionNumber}.zip` ||
|
||||
asset.name === "vscode-codeql-sourcemaps.zip",
|
||||
(asset) => asset.name === `vscode-codeql-sourcemaps-${versionNumber}.zip`,
|
||||
);
|
||||
|
||||
if (sourcemapAsset) {
|
||||
@@ -216,7 +213,9 @@ extractSourceMap().catch((e: unknown) => {
|
||||
function runGh(args: readonly string[]): string {
|
||||
const gh = spawnSync("gh", args);
|
||||
if (gh.status !== 0) {
|
||||
throw new Error(`Failed to run gh ${args.join(" ")}: ${gh.stderr}`);
|
||||
throw new Error(
|
||||
`Failed to get the source map for ${versionNumber}: ${gh.stderr}`,
|
||||
);
|
||||
}
|
||||
return gh.stdout.toString("utf-8");
|
||||
}
|
||||
@@ -228,7 +227,6 @@ function runGhJSON<T>(args: readonly string[]): T {
|
||||
type ReleaseAsset = {
|
||||
id: string;
|
||||
name: string;
|
||||
label: string;
|
||||
};
|
||||
|
||||
type Release = {
|
||||
|
||||
@@ -12,7 +12,6 @@ interface VersionResult {
|
||||
export interface CliFeatures {
|
||||
featuresInVersionResult?: boolean;
|
||||
mrvaPackCreate?: boolean;
|
||||
generateSummarySymbolMap?: boolean;
|
||||
}
|
||||
|
||||
export interface VersionAndFeatures {
|
||||
|
||||
@@ -3,7 +3,7 @@ import { spawn } from "child-process-promise";
|
||||
import type { ChildProcessWithoutNullStreams } from "child_process";
|
||||
import { spawn as spawnChildProcess } from "child_process";
|
||||
import { readFile } from "fs-extra";
|
||||
import { delimiter, join } from "path";
|
||||
import { delimiter, dirname, join } from "path";
|
||||
import type { Log } from "sarif";
|
||||
import { SemVer } from "semver";
|
||||
import type { Readable } from "stream";
|
||||
@@ -91,6 +91,15 @@ export type QlpacksInfo = { [name: string]: string[] };
|
||||
*/
|
||||
type LanguagesInfo = { [name: string]: string[] };
|
||||
|
||||
/** Information about an ML model, as resolved by `codeql resolve ml-models`. */
|
||||
type MlModelInfo = {
|
||||
checksum: string;
|
||||
path: string;
|
||||
};
|
||||
|
||||
/** The expected output of `codeql resolve ml-models`. */
|
||||
type MlModelsInfo = { models: MlModelInfo[] };
|
||||
|
||||
/** Information about a data extension predicate, as resolved by `codeql resolve extensions`. */
|
||||
type DataExtensionResult = {
|
||||
predicate: string;
|
||||
@@ -100,6 +109,7 @@ type DataExtensionResult = {
|
||||
|
||||
/** The expected output of `codeql resolve extensions`. */
|
||||
type ResolveExtensionsResult = {
|
||||
models: MlModelInfo[];
|
||||
data: {
|
||||
[path: string]: DataExtensionResult[];
|
||||
};
|
||||
@@ -933,7 +943,7 @@ export class CodeQLCliServer implements Disposable {
|
||||
if (line.startsWith("Enter value for --github-auth-stdin")) {
|
||||
try {
|
||||
return await this.app.credentials.getAccessToken();
|
||||
} catch {
|
||||
} catch (e) {
|
||||
// If the user cancels the authentication prompt, we still need to give a value to the CLI.
|
||||
// By giving a potentially invalid value, the user will just get a 401/403 when they try to access a
|
||||
// private package and the access token is invalid.
|
||||
@@ -1087,6 +1097,24 @@ export class CodeQLCliServer implements Disposable {
|
||||
);
|
||||
}
|
||||
|
||||
/** Resolves the ML models that should be available when evaluating a query. */
|
||||
async resolveMlModels(
|
||||
additionalPacks: string[],
|
||||
queryPath: string,
|
||||
): Promise<MlModelsInfo> {
|
||||
const args =
|
||||
// use the dirname of the path so that we can handle query libraries
|
||||
[...this.getAdditionalPacksArg(additionalPacks), dirname(queryPath)];
|
||||
return await this.runJsonCodeQlCliCommand<MlModelsInfo>(
|
||||
["resolve", "ml-models"],
|
||||
args,
|
||||
"Resolving ML models",
|
||||
{
|
||||
addFormat: false,
|
||||
},
|
||||
);
|
||||
}
|
||||
|
||||
/**
|
||||
* Gets the RAM setting for the query server.
|
||||
* @param queryMemoryMb The maximum amount of RAM to use, in MB.
|
||||
@@ -1183,15 +1211,10 @@ export class CodeQLCliServer implements Disposable {
|
||||
outputPath: string,
|
||||
endSummaryPath: string,
|
||||
): Promise<string> {
|
||||
const supportsGenerateSummarySymbolMap =
|
||||
await this.cliConstraints.supportsGenerateSummarySymbolMap();
|
||||
const subcommandArgs = [
|
||||
"--format=text",
|
||||
`--end-summary=${endSummaryPath}`,
|
||||
"--sourcemap",
|
||||
...(supportsGenerateSummarySymbolMap
|
||||
? ["--summary-symbol-map", "--minify-output"]
|
||||
: []),
|
||||
inputPath,
|
||||
outputPath,
|
||||
];
|
||||
@@ -1727,6 +1750,14 @@ export class CodeQLCliServer implements Disposable {
|
||||
this._versionChangedListeners.forEach((listener) =>
|
||||
listener(newVersionAndFeatures),
|
||||
);
|
||||
// this._version is only undefined upon config change, so we reset CLI-based context key only when necessary.
|
||||
await this.app.commands.execute(
|
||||
"setContext",
|
||||
"codeql.supportsTrimCache",
|
||||
newVersionAndFeatures.version.compare(
|
||||
CliVersionConstraint.CLI_VERSION_WITH_TRIM_CACHE,
|
||||
) >= 0,
|
||||
);
|
||||
} catch (e) {
|
||||
this._versionChangedListeners.forEach((listener) =>
|
||||
listener(undefined),
|
||||
@@ -1881,17 +1912,45 @@ function shouldDebugCliServer() {
|
||||
export class CliVersionConstraint {
|
||||
// The oldest version of the CLI that we support. This is used to determine
|
||||
// whether to show a warning about the CLI being too old on startup.
|
||||
public static OLDEST_SUPPORTED_CLI_VERSION = new SemVer("2.16.6");
|
||||
public static OLDEST_SUPPORTED_CLI_VERSION = new SemVer("2.14.6");
|
||||
|
||||
/**
|
||||
* CLI version where the query server supports the `evaluation/trimCache` method
|
||||
* with `codeql database cleanup --mode=trim` semantics.
|
||||
*/
|
||||
public static CLI_VERSION_WITH_TRIM_CACHE = new SemVer("2.15.1");
|
||||
|
||||
public static CLI_VERSION_WITHOUT_MRVA_EXTENSIBLE_PREDICATE_HACK = new SemVer(
|
||||
"2.16.1",
|
||||
);
|
||||
|
||||
/**
|
||||
* CLI version where there is support for multiple queries on the pack create command.
|
||||
*/
|
||||
public static CLI_VERSION_WITH_MULTI_QUERY_PACK_CREATE = new SemVer("2.16.1");
|
||||
|
||||
constructor(private readonly cli: CodeQLCliServer) {
|
||||
/**/
|
||||
}
|
||||
|
||||
private async isVersionAtLeast(v: SemVer) {
|
||||
return (await this.cli.getVersion()).compare(v) >= 0;
|
||||
}
|
||||
|
||||
async preservesExtensiblePredicatesInMrvaPack() {
|
||||
// Negated, because we _stopped_ preserving these in 2.16.1.
|
||||
return !(await this.isVersionAtLeast(
|
||||
CliVersionConstraint.CLI_VERSION_WITHOUT_MRVA_EXTENSIBLE_PREDICATE_HACK,
|
||||
));
|
||||
}
|
||||
|
||||
async supportsPackCreateWithMultipleQueries() {
|
||||
return this.isVersionAtLeast(
|
||||
CliVersionConstraint.CLI_VERSION_WITH_MULTI_QUERY_PACK_CREATE,
|
||||
);
|
||||
}
|
||||
|
||||
async supportsMrvaPackCreate(): Promise<boolean> {
|
||||
return (await this.cli.getFeatures()).mrvaPackCreate === true;
|
||||
}
|
||||
|
||||
async supportsGenerateSummarySymbolMap(): Promise<boolean> {
|
||||
return (await this.cli.getFeatures()).generateSummarySymbolMap === true;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,12 +1,5 @@
|
||||
import type { WriteStream } from "fs";
|
||||
import {
|
||||
createWriteStream,
|
||||
mkdtemp,
|
||||
outputJson,
|
||||
pathExists,
|
||||
readJson,
|
||||
remove,
|
||||
} from "fs-extra";
|
||||
import { createWriteStream, mkdtemp, pathExists, remove } from "fs-extra";
|
||||
import { tmpdir } from "os";
|
||||
import { delimiter, dirname, join } from "path";
|
||||
import { Range, satisfies } from "semver";
|
||||
@@ -26,9 +19,7 @@ import {
|
||||
InvocationRateLimiter,
|
||||
InvocationRateLimiterResultKind,
|
||||
} from "../common/invocation-rate-limiter";
|
||||
import type { NotificationLogger } from "../common/logging";
|
||||
import {
|
||||
showAndLogExceptionWithTelemetry,
|
||||
showAndLogErrorMessage,
|
||||
showAndLogWarningMessage,
|
||||
} from "../common/logging";
|
||||
@@ -37,12 +28,7 @@ import { reportUnzipProgress } from "../common/vscode/unzip-progress";
|
||||
import type { Release } from "./distribution/release";
|
||||
import { ReleasesApiConsumer } from "./distribution/releases-api-consumer";
|
||||
import { createTimeoutSignal } from "../common/fetch-stream";
|
||||
import { withDistributionUpdateLock } from "./lock";
|
||||
import { asError, getErrorMessage } from "../common/helpers-pure";
|
||||
import { isIOError } from "../common/files";
|
||||
import { telemetryListener } from "../common/vscode/telemetry";
|
||||
import { redactableError } from "../common/errors";
|
||||
import { ExtensionManagedDistributionCleaner } from "./distribution/cleaner";
|
||||
import { AbortError } from "node-fetch";
|
||||
|
||||
/**
|
||||
* distribution.ts
|
||||
@@ -68,11 +54,6 @@ const NIGHTLY_DISTRIBUTION_REPOSITORY_NWO = "dsp-testing/codeql-cli-nightlies";
|
||||
*/
|
||||
export const DEFAULT_DISTRIBUTION_VERSION_RANGE: Range = new Range("2.x");
|
||||
|
||||
export interface DistributionState {
|
||||
folderIndex: number;
|
||||
release: Release | null;
|
||||
}
|
||||
|
||||
export interface DistributionProvider {
|
||||
getCodeQlPathWithoutVersionCheck(): Promise<string | undefined>;
|
||||
onDidChangeDistribution?: Event<void>;
|
||||
@@ -84,7 +65,6 @@ export class DistributionManager implements DistributionProvider {
|
||||
public readonly config: DistributionConfig,
|
||||
private readonly versionRange: Range,
|
||||
extensionContext: ExtensionContext,
|
||||
logger: NotificationLogger,
|
||||
) {
|
||||
this._onDidChangeDistribution = config.onDidChangeConfiguration;
|
||||
this.extensionSpecificDistributionManager =
|
||||
@@ -92,7 +72,6 @@ export class DistributionManager implements DistributionProvider {
|
||||
config,
|
||||
versionRange,
|
||||
extensionContext,
|
||||
logger,
|
||||
);
|
||||
this.updateCheckRateLimiter = new InvocationRateLimiter(
|
||||
extensionContext.globalState,
|
||||
@@ -100,16 +79,6 @@ export class DistributionManager implements DistributionProvider {
|
||||
() =>
|
||||
this.extensionSpecificDistributionManager.checkForUpdatesToDistribution(),
|
||||
);
|
||||
this.extensionManagedDistributionCleaner =
|
||||
new ExtensionManagedDistributionCleaner(
|
||||
extensionContext,
|
||||
logger,
|
||||
this.extensionSpecificDistributionManager,
|
||||
);
|
||||
}
|
||||
|
||||
public async initialize(): Promise<void> {
|
||||
await this.extensionSpecificDistributionManager.initialize();
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -287,10 +256,6 @@ export class DistributionManager implements DistributionProvider {
|
||||
);
|
||||
}
|
||||
|
||||
public startCleanup() {
|
||||
this.extensionManagedDistributionCleaner.start();
|
||||
}
|
||||
|
||||
public get onDidChangeDistribution(): Event<void> | undefined {
|
||||
return this._onDidChangeDistribution;
|
||||
}
|
||||
@@ -312,63 +277,18 @@ export class DistributionManager implements DistributionProvider {
|
||||
|
||||
private readonly extensionSpecificDistributionManager: ExtensionSpecificDistributionManager;
|
||||
private readonly updateCheckRateLimiter: InvocationRateLimiter<DistributionUpdateCheckResult>;
|
||||
private readonly extensionManagedDistributionCleaner: ExtensionManagedDistributionCleaner;
|
||||
private readonly _onDidChangeDistribution: Event<void> | undefined;
|
||||
}
|
||||
|
||||
class ExtensionSpecificDistributionManager {
|
||||
private distributionState: DistributionState | undefined;
|
||||
|
||||
constructor(
|
||||
private readonly config: DistributionConfig,
|
||||
private readonly versionRange: Range,
|
||||
private readonly extensionContext: ExtensionContext,
|
||||
private readonly logger: NotificationLogger,
|
||||
) {
|
||||
/**/
|
||||
}
|
||||
|
||||
public async initialize() {
|
||||
await this.ensureDistributionStateExists();
|
||||
}
|
||||
|
||||
private async ensureDistributionStateExists() {
|
||||
const distributionStatePath = this.getDistributionStatePath();
|
||||
try {
|
||||
this.distributionState = await readJson(distributionStatePath);
|
||||
} catch (e: unknown) {
|
||||
if (isIOError(e) && e.code === "ENOENT") {
|
||||
// If the file doesn't exist, that just means we need to create it
|
||||
|
||||
this.distributionState = {
|
||||
folderIndex:
|
||||
this.extensionContext.globalState.get(
|
||||
"distributionFolderIndex",
|
||||
0,
|
||||
) ?? 0,
|
||||
release: (this.extensionContext.globalState.get(
|
||||
"distributionRelease",
|
||||
) ?? null) as Release | null,
|
||||
};
|
||||
|
||||
// This may result in a race condition, but when this happens both processes should write the same file.
|
||||
await outputJson(distributionStatePath, this.distributionState);
|
||||
} else {
|
||||
void showAndLogExceptionWithTelemetry(
|
||||
this.logger,
|
||||
telemetryListener,
|
||||
redactableError(
|
||||
asError(e),
|
||||
)`Failed to read distribution state from ${distributionStatePath}: ${getErrorMessage(e)}`,
|
||||
);
|
||||
this.distributionState = {
|
||||
folderIndex: 0,
|
||||
release: null,
|
||||
};
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
public async getCodeQlPathWithoutVersionCheck(): Promise<string | undefined> {
|
||||
if (this.getInstalledRelease() !== undefined) {
|
||||
// An extension specific distribution has been installed.
|
||||
@@ -431,21 +351,9 @@ class ExtensionSpecificDistributionManager {
|
||||
release: Release,
|
||||
progressCallback?: ProgressCallback,
|
||||
): Promise<void> {
|
||||
if (!this.distributionState) {
|
||||
await this.ensureDistributionStateExists();
|
||||
}
|
||||
|
||||
const distributionStatePath = this.getDistributionStatePath();
|
||||
|
||||
await withDistributionUpdateLock(
|
||||
// .lock will be appended to this filename
|
||||
distributionStatePath,
|
||||
async () => {
|
||||
await this.downloadDistribution(release, progressCallback);
|
||||
// Store the installed release within the global extension state.
|
||||
await this.storeInstalledRelease(release);
|
||||
},
|
||||
);
|
||||
await this.downloadDistribution(release, progressCallback);
|
||||
// Store the installed release within the global extension state.
|
||||
await this.storeInstalledRelease(release);
|
||||
}
|
||||
|
||||
private async downloadDistribution(
|
||||
@@ -496,11 +404,6 @@ class ExtensionSpecificDistributionManager {
|
||||
signal,
|
||||
);
|
||||
|
||||
const body = assetStream.body;
|
||||
if (!body) {
|
||||
throw new Error("No body in asset stream");
|
||||
}
|
||||
|
||||
const archivePath = join(tmpDirectory, "distributionDownload.zip");
|
||||
archiveFile = createWriteStream(archivePath);
|
||||
|
||||
@@ -508,40 +411,27 @@ class ExtensionSpecificDistributionManager {
|
||||
const totalNumBytes = contentLength
|
||||
? parseInt(contentLength, 10)
|
||||
: undefined;
|
||||
|
||||
const reportProgress = reportStreamProgress(
|
||||
reportStreamProgress(
|
||||
assetStream.body,
|
||||
`Downloading CodeQL CLI ${release.name}…`,
|
||||
totalNumBytes,
|
||||
progressCallback,
|
||||
);
|
||||
|
||||
const reader = body.getReader();
|
||||
for (;;) {
|
||||
const { done, value } = await reader.read();
|
||||
if (done) {
|
||||
break;
|
||||
}
|
||||
|
||||
onData();
|
||||
reportProgress(value?.length ?? 0);
|
||||
|
||||
await new Promise((resolve, reject) => {
|
||||
archiveFile?.write(value, (err) => {
|
||||
if (err) {
|
||||
reject(err);
|
||||
}
|
||||
resolve(undefined);
|
||||
});
|
||||
});
|
||||
}
|
||||
assetStream.body.on("data", onData);
|
||||
|
||||
await new Promise((resolve, reject) => {
|
||||
archiveFile?.close((err) => {
|
||||
if (err) {
|
||||
reject(err);
|
||||
}
|
||||
resolve(undefined);
|
||||
});
|
||||
if (!archiveFile) {
|
||||
throw new Error("Invariant violation: archiveFile not set");
|
||||
}
|
||||
|
||||
assetStream.body
|
||||
.pipe(archiveFile)
|
||||
.on("finish", resolve)
|
||||
.on("error", reject);
|
||||
|
||||
// If an error occurs on the body, we also want to reject the promise (e.g. during a timeout error).
|
||||
assetStream.body.on("error", reject);
|
||||
});
|
||||
|
||||
disposeTimeout();
|
||||
@@ -562,8 +452,8 @@ class ExtensionSpecificDistributionManager {
|
||||
: undefined,
|
||||
);
|
||||
} catch (e) {
|
||||
if (e instanceof DOMException && e.name === "AbortError") {
|
||||
const thrownError = new Error("The download timed out.");
|
||||
if (e instanceof AbortError) {
|
||||
const thrownError = new AbortError("The download timed out.");
|
||||
thrownError.stack = e.stack;
|
||||
throw thrownError;
|
||||
}
|
||||
@@ -657,19 +547,23 @@ class ExtensionSpecificDistributionManager {
|
||||
}
|
||||
|
||||
private async bumpDistributionFolderIndex(): Promise<void> {
|
||||
await this.updateState((oldState) => {
|
||||
return {
|
||||
...oldState,
|
||||
folderIndex: (oldState.folderIndex ?? 0) + 1,
|
||||
};
|
||||
});
|
||||
const index = this.extensionContext.globalState.get(
|
||||
ExtensionSpecificDistributionManager._currentDistributionFolderIndexStateKey,
|
||||
0,
|
||||
);
|
||||
await this.extensionContext.globalState.update(
|
||||
ExtensionSpecificDistributionManager._currentDistributionFolderIndexStateKey,
|
||||
index + 1,
|
||||
);
|
||||
}
|
||||
|
||||
private getDistributionStoragePath(): string {
|
||||
const distributionState = this.getDistributionState();
|
||||
|
||||
// Use an empty string for the initial distribution for backwards compatibility.
|
||||
const distributionFolderIndex = distributionState.folderIndex || "";
|
||||
const distributionFolderIndex =
|
||||
this.extensionContext.globalState.get(
|
||||
ExtensionSpecificDistributionManager._currentDistributionFolderIndexStateKey,
|
||||
0,
|
||||
) || "";
|
||||
return join(
|
||||
this.extensionContext.globalStorageUri.fsPath,
|
||||
ExtensionSpecificDistributionManager._currentDistributionFolderBaseName +
|
||||
@@ -684,65 +578,26 @@ class ExtensionSpecificDistributionManager {
|
||||
);
|
||||
}
|
||||
|
||||
private getDistributionStatePath(): string {
|
||||
return join(
|
||||
this.extensionContext.globalStorageUri.fsPath,
|
||||
ExtensionSpecificDistributionManager._distributionStateFilename,
|
||||
);
|
||||
}
|
||||
|
||||
private getInstalledRelease(): Release | undefined {
|
||||
return this.getDistributionState().release ?? undefined;
|
||||
return this.extensionContext.globalState.get(
|
||||
ExtensionSpecificDistributionManager._installedReleaseStateKey,
|
||||
);
|
||||
}
|
||||
|
||||
private async storeInstalledRelease(
|
||||
release: Release | undefined,
|
||||
): Promise<void> {
|
||||
await this.updateState((oldState) => ({
|
||||
...oldState,
|
||||
release: release ?? null,
|
||||
}));
|
||||
}
|
||||
|
||||
private getDistributionState(): DistributionState {
|
||||
const distributionState = this.distributionState;
|
||||
if (distributionState === undefined) {
|
||||
throw new Error(
|
||||
"Invariant violation: distribution state not initialized",
|
||||
);
|
||||
}
|
||||
return distributionState;
|
||||
}
|
||||
|
||||
private async updateState(
|
||||
f: (oldState: DistributionState) => DistributionState,
|
||||
) {
|
||||
const oldState = this.distributionState;
|
||||
if (oldState === undefined) {
|
||||
throw new Error(
|
||||
"Invariant violation: distribution state not initialized",
|
||||
);
|
||||
}
|
||||
const newState = f(oldState);
|
||||
this.distributionState = newState;
|
||||
|
||||
const distributionStatePath = this.getDistributionStatePath();
|
||||
await outputJson(distributionStatePath, newState);
|
||||
}
|
||||
|
||||
public get folderIndex() {
|
||||
const distributionState = this.getDistributionState();
|
||||
|
||||
return distributionState.folderIndex;
|
||||
}
|
||||
|
||||
public get distributionFolderPrefix() {
|
||||
return ExtensionSpecificDistributionManager._currentDistributionFolderBaseName;
|
||||
await this.extensionContext.globalState.update(
|
||||
ExtensionSpecificDistributionManager._installedReleaseStateKey,
|
||||
release,
|
||||
);
|
||||
}
|
||||
|
||||
private static readonly _currentDistributionFolderBaseName = "distribution";
|
||||
private static readonly _currentDistributionFolderIndexStateKey =
|
||||
"distributionFolderIndex";
|
||||
private static readonly _installedReleaseStateKey = "distributionRelease";
|
||||
private static readonly _codeQlExtractedFolderName = "codeql";
|
||||
private static readonly _distributionStateFilename = "distribution.json";
|
||||
}
|
||||
|
||||
/*
|
||||
|
||||
@@ -1,127 +0,0 @@
|
||||
import type { ExtensionContext } from "vscode";
|
||||
import { getDirectoryNamesInsidePath, isIOError } from "../../common/files";
|
||||
import { sleep } from "../../common/time";
|
||||
import type { BaseLogger } from "../../common/logging";
|
||||
import { join } from "path";
|
||||
import { getErrorMessage } from "../../common/helpers-pure";
|
||||
import { pathExists, remove } from "fs-extra";
|
||||
|
||||
interface ExtensionManagedDistributionManager {
|
||||
folderIndex: number;
|
||||
distributionFolderPrefix: string;
|
||||
}
|
||||
|
||||
interface DistributionDirectory {
|
||||
directoryName: string;
|
||||
folderIndex: number;
|
||||
}
|
||||
|
||||
/**
|
||||
* This class is responsible for cleaning up old distributions that are no longer needed. In normal operation, this
|
||||
* should not be necessary as the old distribution is deleted when the distribution is updated. However, in some cases
|
||||
* the extension may leave behind old distribution which can result in a significant amount of space (> 100 GB) being
|
||||
* taking up by unused distributions.
|
||||
*/
|
||||
export class ExtensionManagedDistributionCleaner {
|
||||
constructor(
|
||||
private readonly extensionContext: ExtensionContext,
|
||||
private readonly logger: BaseLogger,
|
||||
private readonly manager: ExtensionManagedDistributionManager,
|
||||
) {}
|
||||
|
||||
public start() {
|
||||
// Intentionally starting this without waiting for it
|
||||
void this.cleanup().catch((e: unknown) => {
|
||||
void this.logger.log(
|
||||
`Failed to clean up old versions of the CLI: ${getErrorMessage(e)}`,
|
||||
);
|
||||
});
|
||||
}
|
||||
|
||||
public async cleanup() {
|
||||
if (!(await pathExists(this.extensionContext.globalStorageUri.fsPath))) {
|
||||
return;
|
||||
}
|
||||
|
||||
const currentFolderIndex = this.manager.folderIndex;
|
||||
|
||||
const distributionDirectoryRegex = new RegExp(
|
||||
`^${this.manager.distributionFolderPrefix}(\\d+)$`,
|
||||
);
|
||||
|
||||
const existingDirectories = await getDirectoryNamesInsidePath(
|
||||
this.extensionContext.globalStorageUri.fsPath,
|
||||
);
|
||||
const distributionDirectories = existingDirectories
|
||||
.map((dir): DistributionDirectory | null => {
|
||||
const match = dir.match(distributionDirectoryRegex);
|
||||
if (!match) {
|
||||
// When the folderIndex is 0, the distributionFolderPrefix is used as the directory name
|
||||
if (dir === this.manager.distributionFolderPrefix) {
|
||||
return {
|
||||
directoryName: dir,
|
||||
folderIndex: 0,
|
||||
};
|
||||
}
|
||||
|
||||
return null;
|
||||
}
|
||||
|
||||
return {
|
||||
directoryName: dir,
|
||||
folderIndex: parseInt(match[1]),
|
||||
};
|
||||
})
|
||||
.filter((dir) => dir !== null);
|
||||
|
||||
// Clean up all directories that are older than the current one
|
||||
const cleanableDirectories = distributionDirectories.filter(
|
||||
(dir) => dir.folderIndex < currentFolderIndex,
|
||||
);
|
||||
|
||||
if (cleanableDirectories.length === 0) {
|
||||
return;
|
||||
}
|
||||
|
||||
// Shuffle the array so that multiple VS Code processes don't all try to clean up the same directory at the same time
|
||||
for (let i = cleanableDirectories.length - 1; i > 0; i--) {
|
||||
const j = Math.floor(Math.random() * (i + 1));
|
||||
[cleanableDirectories[i], cleanableDirectories[j]] = [
|
||||
cleanableDirectories[j],
|
||||
cleanableDirectories[i],
|
||||
];
|
||||
}
|
||||
|
||||
void this.logger.log(
|
||||
`Cleaning up ${cleanableDirectories.length} old versions of the CLI.`,
|
||||
);
|
||||
|
||||
for (const cleanableDirectory of cleanableDirectories) {
|
||||
// Wait 10 seconds between each cleanup to avoid overloading the system (even though the remove call should be async)
|
||||
await sleep(10_000);
|
||||
|
||||
const path = join(
|
||||
this.extensionContext.globalStorageUri.fsPath,
|
||||
cleanableDirectory.directoryName,
|
||||
);
|
||||
|
||||
// Delete this directory
|
||||
try {
|
||||
await remove(path);
|
||||
} catch (e) {
|
||||
if (isIOError(e) && e.code === "ENOENT") {
|
||||
// If the directory doesn't exist, that's fine
|
||||
continue;
|
||||
}
|
||||
|
||||
void this.logger.log(
|
||||
`Tried to clean up an old version of the CLI at ${path} but encountered an error: ${getErrorMessage(e)}.`,
|
||||
);
|
||||
}
|
||||
}
|
||||
|
||||
void this.logger.log(
|
||||
`Cleaned up ${cleanableDirectories.length} old versions of the CLI.`,
|
||||
);
|
||||
}
|
||||
}
|
||||
@@ -1,3 +1,5 @@
|
||||
import type { Response } from "node-fetch";
|
||||
import { default as fetch } from "node-fetch";
|
||||
import type { Range } from "semver";
|
||||
import { compare, parse, satisfies } from "semver";
|
||||
import { URL } from "url";
|
||||
@@ -32,9 +34,9 @@ export class ReleasesApiConsumer {
|
||||
additionalCompatibilityCheck?: (release: GithubRelease) => boolean,
|
||||
): Promise<Release> {
|
||||
const apiPath = `/repos/${this.repositoryNwo}/releases`;
|
||||
const allReleases = (await (
|
||||
const allReleases: GithubRelease[] = await (
|
||||
await this.makeApiCall(apiPath)
|
||||
).json()) as GithubRelease[];
|
||||
).json();
|
||||
const compatibleReleases = allReleases.filter((release) => {
|
||||
if (release.prerelease && !includePrerelease) {
|
||||
return false;
|
||||
|
||||
@@ -1,22 +0,0 @@
|
||||
import { lock } from "proper-lockfile";
|
||||
|
||||
export async function withDistributionUpdateLock(
|
||||
lockFile: string,
|
||||
f: () => Promise<void>,
|
||||
) {
|
||||
const release = await lock(lockFile, {
|
||||
stale: 60_000, // 1 minute. We can take the lock longer than this because that's based on the update interval.
|
||||
update: 10_000, // 10 seconds
|
||||
retries: {
|
||||
minTimeout: 10_000,
|
||||
maxTimeout: 60_000,
|
||||
retries: 100,
|
||||
},
|
||||
});
|
||||
|
||||
try {
|
||||
await f();
|
||||
} finally {
|
||||
await release();
|
||||
}
|
||||
}
|
||||
@@ -36,7 +36,7 @@ export async function findLanguage(
|
||||
void extLogger.log(
|
||||
"Query language is unsupported. Select language manually.",
|
||||
);
|
||||
} catch {
|
||||
} catch (e) {
|
||||
void extLogger.log(
|
||||
"Could not autodetect query language. Select language manually.",
|
||||
);
|
||||
|
||||
@@ -11,7 +11,6 @@ export namespace BqrsColumnKindCode {
|
||||
export const BOOLEAN = "b";
|
||||
export const DATE = "d";
|
||||
export const ENTITY = "e";
|
||||
export const BIGINT = "z";
|
||||
}
|
||||
|
||||
export type BqrsColumnKind =
|
||||
@@ -20,8 +19,7 @@ export type BqrsColumnKind =
|
||||
| typeof BqrsColumnKindCode.STRING
|
||||
| typeof BqrsColumnKindCode.BOOLEAN
|
||||
| typeof BqrsColumnKindCode.DATE
|
||||
| typeof BqrsColumnKindCode.ENTITY
|
||||
| typeof BqrsColumnKindCode.BIGINT;
|
||||
| typeof BqrsColumnKindCode.ENTITY;
|
||||
|
||||
export interface BqrsSchemaColumn {
|
||||
name?: string;
|
||||
@@ -81,8 +79,7 @@ export type BqrsKind =
|
||||
| "Integer"
|
||||
| "Boolean"
|
||||
| "Date"
|
||||
| "Entity"
|
||||
| "BigInt";
|
||||
| "Entity";
|
||||
|
||||
interface BqrsColumn {
|
||||
name?: string;
|
||||
|
||||
@@ -76,8 +76,6 @@ function mapColumnKind(kind: BqrsColumnKind): ColumnKind {
|
||||
return ColumnKind.Date;
|
||||
case BqrsColumnKindCode.ENTITY:
|
||||
return ColumnKind.Entity;
|
||||
case BqrsColumnKindCode.BIGINT:
|
||||
return ColumnKind.BigInt;
|
||||
default:
|
||||
assertNever(kind);
|
||||
}
|
||||
|
||||
@@ -290,7 +290,6 @@ export type DatabasePanelCommands = {
|
||||
"codeQLVariantAnalysisRepositories.openConfigFile": () => Promise<void>;
|
||||
"codeQLVariantAnalysisRepositories.addNewDatabase": () => Promise<void>;
|
||||
"codeQLVariantAnalysisRepositories.addNewList": () => Promise<void>;
|
||||
"codeQLVariantAnalysisRepositories.setupControllerRepository": () => Promise<void>;
|
||||
|
||||
"codeQLVariantAnalysisRepositories.setSelectedItem": TreeViewContextSingleSelectionCommandFunction<DbTreeViewItem>;
|
||||
"codeQLVariantAnalysisRepositories.setSelectedItemContextMenu": TreeViewContextSingleSelectionCommandFunction<DbTreeViewItem>;
|
||||
|
||||
@@ -78,7 +78,7 @@ function getNwoOrOwnerFromGitHubUrl(
|
||||
}
|
||||
const nwo = `${paths[0]}/${paths[1]}`;
|
||||
return paths[1] ? nwo : undefined;
|
||||
} catch {
|
||||
} catch (e) {
|
||||
// Ignore the error here, since we catch failures at a higher level.
|
||||
return;
|
||||
}
|
||||
|
||||
@@ -147,21 +147,6 @@ interface SetStateMsg {
|
||||
parsedResultSets: ParsedResultSets;
|
||||
}
|
||||
|
||||
export interface UserSettings {
|
||||
/** Whether to display links to the dataflow models that generated particular nodes in a flow path. */
|
||||
shouldShowProvenance: boolean;
|
||||
}
|
||||
|
||||
export const DEFAULT_USER_SETTINGS: UserSettings = {
|
||||
shouldShowProvenance: false,
|
||||
};
|
||||
|
||||
/** Message indicating that the user's configuration settings have changed. */
|
||||
interface SetUserSettingsMsg {
|
||||
t: "setUserSettings";
|
||||
userSettings: UserSettings;
|
||||
}
|
||||
|
||||
/**
|
||||
* Message indicating that the results view should display interpreted
|
||||
* results.
|
||||
@@ -206,7 +191,6 @@ interface UntoggleShowProblemsMsg {
|
||||
export type IntoResultsViewMsg =
|
||||
| ResultsUpdatingMsg
|
||||
| SetStateMsg
|
||||
| SetUserSettingsMsg
|
||||
| ShowInterpretedPageMsg
|
||||
| NavigateMsg
|
||||
| UntoggleShowProblemsMsg;
|
||||
@@ -224,15 +208,13 @@ export type FromResultsViewMsg =
|
||||
| OpenFileMsg;
|
||||
|
||||
/**
|
||||
* Message from the results view to open a source
|
||||
* Message from the results view to open a database source
|
||||
* file at the provided location.
|
||||
*/
|
||||
interface ViewSourceFileMsg {
|
||||
t: "viewSourceFile";
|
||||
loc: UrlValueResolvable;
|
||||
/** URI of the database whose source archive contains the file, or `undefined` to open a file from
|
||||
* the local disk. The latter case is used for opening links to data extension model files. */
|
||||
databaseUri: string | undefined;
|
||||
databaseUri: string;
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -359,11 +341,7 @@ interface ChangeCompareMessage {
|
||||
|
||||
export type ToCompareViewMessage =
|
||||
| SetComparisonQueryInfoMessage
|
||||
| SetComparisonsMessage
|
||||
| StreamingComparisonSetupMessage
|
||||
| StreamingComparisonAddResultsMessage
|
||||
| StreamingComparisonCompleteMessage
|
||||
| SetUserSettingsMsg;
|
||||
| SetComparisonsMessage;
|
||||
|
||||
/**
|
||||
* Message to the compare view that sets the metadata of the compared queries.
|
||||
@@ -422,28 +400,6 @@ export type InterpretedQueryCompareResult = {
|
||||
to: Result[];
|
||||
};
|
||||
|
||||
export interface StreamingComparisonSetupMessage {
|
||||
readonly t: "streamingComparisonSetup";
|
||||
// The id of this streaming comparison
|
||||
readonly id: string;
|
||||
readonly currentResultSetName: string;
|
||||
readonly message: string | undefined;
|
||||
// The from and to fields will only contain a chunk of the results
|
||||
readonly result: QueryCompareResult;
|
||||
}
|
||||
|
||||
interface StreamingComparisonAddResultsMessage {
|
||||
readonly t: "streamingComparisonAddResults";
|
||||
readonly id: string;
|
||||
// The from and to fields will only contain a chunk of the results
|
||||
readonly result: QueryCompareResult;
|
||||
}
|
||||
|
||||
interface StreamingComparisonCompleteMessage {
|
||||
readonly t: "streamingComparisonComplete";
|
||||
readonly id: string;
|
||||
}
|
||||
|
||||
/**
|
||||
* Extract the name of the default result. Prefer returning
|
||||
* 'alerts', or '#select'. Otherwise return the first in the list.
|
||||
@@ -585,6 +541,16 @@ interface SetModifiedMethodsMessage {
|
||||
methodSignatures: string[];
|
||||
}
|
||||
|
||||
interface SetInProgressMethodsMessage {
|
||||
t: "setInProgressMethods";
|
||||
methods: string[];
|
||||
}
|
||||
|
||||
interface SetProcessedByAutoModelMethodsMessage {
|
||||
t: "setProcessedByAutoModelMethods";
|
||||
methods: string[];
|
||||
}
|
||||
|
||||
interface SwitchModeMessage {
|
||||
t: "switchMode";
|
||||
mode: Mode;
|
||||
@@ -616,6 +582,17 @@ interface GenerateMethodMessage {
|
||||
t: "generateMethod";
|
||||
}
|
||||
|
||||
interface GenerateMethodsFromLlmMessage {
|
||||
t: "generateMethodsFromLlm";
|
||||
packageName: string;
|
||||
methodSignatures: string[];
|
||||
}
|
||||
|
||||
interface StopGeneratingMethodsFromLlmMessage {
|
||||
t: "stopGeneratingMethodsFromLlm";
|
||||
packageName: string;
|
||||
}
|
||||
|
||||
interface StartModelEvaluationMessage {
|
||||
t: "startModelEvaluation";
|
||||
}
|
||||
@@ -653,6 +630,16 @@ interface SetInModelingModeMessage {
|
||||
inModelingMode: boolean;
|
||||
}
|
||||
|
||||
interface SetInProgressMessage {
|
||||
t: "setInProgress";
|
||||
inProgress: boolean;
|
||||
}
|
||||
|
||||
interface SetProcessedByAutoModelMessage {
|
||||
t: "setProcessedByAutoModel";
|
||||
processedByAutoModel: boolean;
|
||||
}
|
||||
|
||||
interface RevealMethodMessage {
|
||||
t: "revealMethod";
|
||||
methodSignature: string;
|
||||
@@ -673,6 +660,8 @@ export type ToModelEditorMessage =
|
||||
| SetMethodsMessage
|
||||
| SetModeledAndModifiedMethodsMessage
|
||||
| SetModifiedMethodsMessage
|
||||
| SetInProgressMethodsMessage
|
||||
| SetProcessedByAutoModelMethodsMessage
|
||||
| RevealMethodMessage
|
||||
| SetAccessPathSuggestionsMessage
|
||||
| SetModelEvaluationRunMessage;
|
||||
@@ -686,6 +675,8 @@ export type FromModelEditorMessage =
|
||||
| JumpToMethodMessage
|
||||
| SaveModeledMethods
|
||||
| GenerateMethodMessage
|
||||
| GenerateMethodsFromLlmMessage
|
||||
| StopGeneratingMethodsFromLlmMessage
|
||||
| ModelDependencyMessage
|
||||
| HideModeledMethodsMessage
|
||||
| SetMultipleModeledMethodsMessage
|
||||
@@ -728,6 +719,8 @@ interface SetSelectedMethodMessage {
|
||||
method: Method;
|
||||
modeledMethods: ModeledMethod[];
|
||||
isModified: boolean;
|
||||
isInProgress: boolean;
|
||||
processedByAutoModel: boolean;
|
||||
}
|
||||
|
||||
export type ToMethodModelingMessage =
|
||||
@@ -736,7 +729,9 @@ export type ToMethodModelingMessage =
|
||||
| SetMethodModifiedMessage
|
||||
| SetNoMethodSelectedMessage
|
||||
| SetSelectedMethodMessage
|
||||
| SetInModelingModeMessage;
|
||||
| SetInModelingModeMessage
|
||||
| SetInProgressMessage
|
||||
| SetProcessedByAutoModelMessage;
|
||||
|
||||
interface SetModelAlertsViewStateMessage {
|
||||
t: "setModelAlertsViewState";
|
||||
|
||||
@@ -14,6 +14,7 @@ export enum RequestKind {
|
||||
GetVariantAnalysisRepo = "getVariantAnalysisRepo",
|
||||
GetVariantAnalysisRepoResult = "getVariantAnalysisRepoResult",
|
||||
CodeSearch = "codeSearch",
|
||||
AutoModel = "autoModel",
|
||||
}
|
||||
|
||||
export interface BasicErrorResponse {
|
||||
@@ -68,7 +69,7 @@ export interface GetVariantAnalysisRepoResultRequest {
|
||||
};
|
||||
response: {
|
||||
status: number;
|
||||
body?: ArrayBuffer | string;
|
||||
body?: Buffer | string;
|
||||
contentType: string;
|
||||
};
|
||||
}
|
||||
@@ -91,13 +92,31 @@ interface CodeSearchRequest {
|
||||
};
|
||||
}
|
||||
|
||||
export interface AutoModelResponse {
|
||||
models: string;
|
||||
}
|
||||
|
||||
interface AutoModelRequest {
|
||||
request: {
|
||||
kind: RequestKind.AutoModel;
|
||||
body?: {
|
||||
candidates: string;
|
||||
};
|
||||
};
|
||||
response: {
|
||||
status: number;
|
||||
body?: AutoModelResponse | BasicErrorResponse;
|
||||
};
|
||||
}
|
||||
|
||||
export type GitHubApiRequest =
|
||||
| GetRepoRequest
|
||||
| SubmitVariantAnalysisRequest
|
||||
| GetVariantAnalysisRequest
|
||||
| GetVariantAnalysisRepoRequest
|
||||
| GetVariantAnalysisRepoResultRequest
|
||||
| CodeSearchRequest;
|
||||
| CodeSearchRequest
|
||||
| AutoModelRequest;
|
||||
|
||||
export const isGetRepoRequest = (
|
||||
request: GitHubApiRequest,
|
||||
@@ -127,3 +146,8 @@ export const isCodeSearchRequest = (
|
||||
request: GitHubApiRequest,
|
||||
): request is CodeSearchRequest =>
|
||||
request.request.kind === RequestKind.CodeSearch;
|
||||
|
||||
export const isAutoModelRequest = (
|
||||
request: GitHubApiRequest,
|
||||
): request is AutoModelRequest =>
|
||||
request.request.kind === RequestKind.AutoModel;
|
||||
|
||||
@@ -2,7 +2,6 @@ import { join, resolve } from "path";
|
||||
import { pathExists } from "fs-extra";
|
||||
import type { SetupServer } from "msw/node";
|
||||
import { setupServer } from "msw/node";
|
||||
import type { UnhandledRequestStrategy } from "msw/lib/core/utils/request/onUnhandledRequest";
|
||||
|
||||
import { DisposableObject } from "../disposable-object";
|
||||
|
||||
@@ -27,14 +26,12 @@ export class MockGitHubApiServer extends DisposableObject {
|
||||
this.recorder = this.push(new Recorder(this.server));
|
||||
}
|
||||
|
||||
public startServer(
|
||||
onUnhandledRequest: UnhandledRequestStrategy = "bypass",
|
||||
): void {
|
||||
public startServer(): void {
|
||||
if (this._isListening) {
|
||||
return;
|
||||
}
|
||||
|
||||
this.server.listen({ onUnhandledRequest });
|
||||
this.server.listen({ onUnhandledRequest: "bypass" });
|
||||
this._isListening = true;
|
||||
}
|
||||
|
||||
@@ -57,7 +54,8 @@ export class MockGitHubApiServer extends DisposableObject {
|
||||
const scenarioPath = join(scenariosPath, scenarioName);
|
||||
|
||||
const handlers = await createRequestHandlers(scenarioPath);
|
||||
this.server.resetHandlers(...handlers);
|
||||
this.server.resetHandlers();
|
||||
this.server.use(...handlers);
|
||||
}
|
||||
|
||||
public async saveScenario(
|
||||
|
||||
@@ -1,12 +1,14 @@
|
||||
import { ensureDir, writeFile } from "fs-extra";
|
||||
import { join } from "path";
|
||||
|
||||
import fetch from "node-fetch";
|
||||
import type { SetupServer } from "msw/node";
|
||||
|
||||
import { DisposableObject } from "../disposable-object";
|
||||
import { gzipDecode } from "../zlib";
|
||||
|
||||
import type {
|
||||
AutoModelResponse,
|
||||
BasicErrorResponse,
|
||||
CodeSearchResponse,
|
||||
GetVariantAnalysisRepoResultRequest,
|
||||
@@ -89,14 +91,7 @@ export class Recorder extends DisposableObject {
|
||||
|
||||
let bodyFileLink = undefined;
|
||||
if (writtenRequest.response.body) {
|
||||
if (typeof writtenRequest.response.body === "string") {
|
||||
await writeFile(bodyFilePath, writtenRequest.response.body);
|
||||
} else {
|
||||
await writeFile(
|
||||
bodyFilePath,
|
||||
Buffer.from(writtenRequest.response.body),
|
||||
);
|
||||
}
|
||||
await writeFile(bodyFilePath, writtenRequest.response.body);
|
||||
bodyFileLink = `file:${bodyFileName}`;
|
||||
}
|
||||
|
||||
@@ -231,7 +226,7 @@ async function createGitHubApiRequest(
|
||||
"x-vscode-codeql-msw-bypass": "true",
|
||||
},
|
||||
});
|
||||
const responseBuffer = await response.arrayBuffer();
|
||||
const responseBuffer = await response.buffer();
|
||||
|
||||
return {
|
||||
request: {
|
||||
@@ -263,6 +258,23 @@ async function createGitHubApiRequest(
|
||||
};
|
||||
}
|
||||
|
||||
const autoModelMatch = url.match(
|
||||
/\/repos\/github\/codeql\/code-scanning\/codeql\/auto-model/,
|
||||
);
|
||||
if (autoModelMatch) {
|
||||
return {
|
||||
request: {
|
||||
kind: RequestKind.AutoModel,
|
||||
},
|
||||
response: {
|
||||
status,
|
||||
body: await jsonResponseBody<
|
||||
BasicErrorResponse | AutoModelResponse | undefined
|
||||
>(response),
|
||||
},
|
||||
};
|
||||
}
|
||||
|
||||
return undefined;
|
||||
}
|
||||
|
||||
|
||||
@@ -4,6 +4,7 @@ import type { RequestHandler } from "msw";
|
||||
import { http } from "msw";
|
||||
import type { GitHubApiRequest } from "./gh-api-request";
|
||||
import {
|
||||
isAutoModelRequest,
|
||||
isCodeSearchRequest,
|
||||
isGetRepoRequest,
|
||||
isGetVariantAnalysisRepoRequest,
|
||||
@@ -40,6 +41,7 @@ export async function createRequestHandlers(
|
||||
createGetVariantAnalysisRepoRequestHandler(requests),
|
||||
createGetVariantAnalysisRepoResultRequestHandler(requests),
|
||||
createCodeSearchRequestHandler(requests),
|
||||
createAutoModelRequestHandler(requests),
|
||||
];
|
||||
|
||||
return handlers;
|
||||
@@ -228,3 +230,29 @@ function createCodeSearchRequestHandler(
|
||||
});
|
||||
});
|
||||
}
|
||||
|
||||
function createAutoModelRequestHandler(
|
||||
requests: GitHubApiRequest[],
|
||||
): RequestHandler {
|
||||
const autoModelRequests = requests.filter(isAutoModelRequest);
|
||||
let requestIndex = 0;
|
||||
|
||||
// During automodeling there can be multiple API requests for each batch
|
||||
// of candidates we want to model. We need to return different responses for each request,
|
||||
// so keep an index of the request and return the appropriate response.
|
||||
return http.post(
|
||||
`${baseUrl}/repos/github/codeql/code-scanning/codeql/auto-model`,
|
||||
() => {
|
||||
const request = autoModelRequests[requestIndex];
|
||||
|
||||
if (requestIndex < autoModelRequests.length - 1) {
|
||||
// If there are more requests to come, increment the index.
|
||||
requestIndex++;
|
||||
}
|
||||
|
||||
return jsonResponse(request.response.body, {
|
||||
status: request.response.status,
|
||||
});
|
||||
},
|
||||
);
|
||||
}
|
||||
|
||||
@@ -0,0 +1,11 @@
|
||||
{
|
||||
"request": {
|
||||
"kind": "autoModel"
|
||||
},
|
||||
"response": {
|
||||
"status": 200,
|
||||
"body": {
|
||||
"models": "extensions:\n- addsTo: {extensible: sinkModel, pack: codeql/java-all}\n data:\n - [javax.servlet.http, HttpServletResponse, true, sendRedirect, (String), '', 'Argument[this]',\n request-forgery, ai-generated]\n - [javax.servlet.http, HttpServletResponse, true, sendRedirect, (String), '', 'Argument[0]',\n request-forgery, ai-generated]\n"
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -0,0 +1,11 @@
|
||||
{
|
||||
"request": {
|
||||
"kind": "autoModel"
|
||||
},
|
||||
"response": {
|
||||
"status": 200,
|
||||
"body": {
|
||||
"models": "extensions:\n- addsTo: {extensible: sinkModel, pack: codeql/java-all}\n data:\n - [javax.servlet, MultipartConfigElement, true, MultipartConfigElement, (String),\n '', 'Argument[0]', request-forgery, ai-generated]\n"
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -0,0 +1 @@
|
||||
This scenario is best when modeling the `javax.servlet-api` package.
|
||||
@@ -1,13 +1,10 @@
|
||||
import { Octokit } from "@octokit/rest";
|
||||
import { retry } from "@octokit/plugin-retry";
|
||||
import fetch from "node-fetch";
|
||||
|
||||
export const AppOctokit = Octokit.defaults({
|
||||
request: {
|
||||
// MSW replaces the global fetch object, so we can't just pass a reference to the
|
||||
// fetch object at initialization time. Instead, we pass a function that will
|
||||
// always call the global fetch object.
|
||||
fetch: (input: string | URL | Request, init?: RequestInit) =>
|
||||
fetch(input, init),
|
||||
fetch,
|
||||
},
|
||||
retry,
|
||||
});
|
||||
|
||||
@@ -5,7 +5,6 @@ export enum ColumnKind {
|
||||
Boolean = "boolean",
|
||||
Date = "date",
|
||||
Entity = "entity",
|
||||
BigInt = "bigint",
|
||||
}
|
||||
|
||||
export type Column = {
|
||||
@@ -62,11 +61,6 @@ type CellValueNumber = {
|
||||
value: number;
|
||||
};
|
||||
|
||||
type CellValueBigInt = {
|
||||
type: "number";
|
||||
value: number;
|
||||
};
|
||||
|
||||
type CellValueString = {
|
||||
type: "string";
|
||||
value: string;
|
||||
@@ -81,8 +75,7 @@ export type CellValue =
|
||||
| CellValueEntity
|
||||
| CellValueNumber
|
||||
| CellValueString
|
||||
| CellValueBoolean
|
||||
| CellValueBigInt;
|
||||
| CellValueBoolean;
|
||||
|
||||
export type Row = CellValue[];
|
||||
|
||||
|
||||
@@ -1,14 +1,14 @@
|
||||
export type DeepReadonly<T> =
|
||||
T extends Array<infer R>
|
||||
? DeepReadonlyArray<R>
|
||||
: // eslint-disable-next-line @typescript-eslint/no-unsafe-function-type
|
||||
: // eslint-disable-next-line @typescript-eslint/ban-types
|
||||
T extends Function
|
||||
? T
|
||||
: T extends object
|
||||
? DeepReadonlyObject<T>
|
||||
: T;
|
||||
|
||||
type DeepReadonlyArray<T> = ReadonlyArray<DeepReadonly<T>>;
|
||||
interface DeepReadonlyArray<T> extends ReadonlyArray<DeepReadonly<T>> {}
|
||||
|
||||
type DeepReadonlyObject<T> = {
|
||||
readonly [P in keyof T]: DeepReadonly<T[P]>;
|
||||
|
||||
@@ -1,20 +1,18 @@
|
||||
import type { Log } from "sarif";
|
||||
import type { Log, Tool } from "sarif";
|
||||
import { createReadStream } from "fs-extra";
|
||||
import { connectTo } from "stream-json/Assembler";
|
||||
import { getErrorMessage } from "./helpers-pure";
|
||||
import { withParser } from "stream-json/filters/Ignore";
|
||||
import { withParser } from "stream-json/filters/Pick";
|
||||
|
||||
const DUMMY_TOOL: Tool = { driver: { name: "" } };
|
||||
|
||||
export async function sarifParser(
|
||||
interpretedResultsPath: string,
|
||||
): Promise<Log> {
|
||||
try {
|
||||
// Parse the SARIF file into token streams, filtering out some of the larger subtrees that we
|
||||
// don't need.
|
||||
// Parse the SARIF file into token streams, filtering out only the results array.
|
||||
const pipeline = createReadStream(interpretedResultsPath).pipe(
|
||||
withParser({
|
||||
// We don't need to run's `artifacts` property, nor the driver's `notifications` property.
|
||||
filter: /^runs\.\d+\.(artifacts|tool\.driver\.notifications)/,
|
||||
}),
|
||||
withParser({ filter: "runs.0.results" }),
|
||||
);
|
||||
|
||||
// Creates JavaScript objects from the token stream
|
||||
@@ -40,17 +38,15 @@ export async function sarifParser(
|
||||
});
|
||||
|
||||
asm.on("done", (asm) => {
|
||||
const log = asm.current;
|
||||
|
||||
// Do some trivial validation. This isn't a full validation of the SARIF file, but it's at
|
||||
// least enough to ensure that we're not trying to parse complete garbage later.
|
||||
if (log.runs === undefined || log.runs.length < 1) {
|
||||
reject(
|
||||
new Error(
|
||||
"Invalid SARIF file: expecting at least one run with result.",
|
||||
),
|
||||
);
|
||||
}
|
||||
const log: Log = {
|
||||
version: "2.1.0",
|
||||
runs: [
|
||||
{
|
||||
tool: DUMMY_TOOL,
|
||||
results: asm.current ?? [],
|
||||
},
|
||||
],
|
||||
};
|
||||
|
||||
resolve(log);
|
||||
alreadyDone = true;
|
||||
|
||||
@@ -97,15 +97,17 @@ export function withProgress<R>(
|
||||
* Displays a progress monitor that indicates how much progess has been made
|
||||
* reading from a stream.
|
||||
*
|
||||
* @param readable The stream to read progress from
|
||||
* @param messagePrefix A prefix for displaying the message
|
||||
* @param totalNumBytes Total number of bytes in this stream
|
||||
* @param progress The progress callback used to set messages
|
||||
*/
|
||||
export function reportStreamProgress(
|
||||
readable: NodeJS.ReadableStream,
|
||||
messagePrefix: string,
|
||||
totalNumBytes?: number,
|
||||
progress?: ProgressCallback,
|
||||
): (bytesRead: number) => void {
|
||||
) {
|
||||
if (progress && totalNumBytes) {
|
||||
let numBytesDownloaded = 0;
|
||||
const updateProgress = () => {
|
||||
@@ -121,10 +123,10 @@ export function reportStreamProgress(
|
||||
// Display the progress straight away rather than waiting for the first chunk.
|
||||
updateProgress();
|
||||
|
||||
return (bytesRead: number) => {
|
||||
numBytesDownloaded += bytesRead;
|
||||
readable.on("data", (data) => {
|
||||
numBytesDownloaded += data.length;
|
||||
updateProgress();
|
||||
};
|
||||
});
|
||||
} else if (progress) {
|
||||
progress({
|
||||
step: 1,
|
||||
@@ -132,6 +134,4 @@ export function reportStreamProgress(
|
||||
message: `${messagePrefix} (Size unknown)`,
|
||||
});
|
||||
}
|
||||
|
||||
return () => {};
|
||||
}
|
||||
|
||||
@@ -1,5 +1,10 @@
|
||||
import { promisify } from "util";
|
||||
import { gunzip } from "zlib";
|
||||
import { gzip, gunzip } from "zlib";
|
||||
|
||||
/**
|
||||
* Promisified version of zlib.gzip
|
||||
*/
|
||||
export const gzipEncode = promisify(gzip);
|
||||
|
||||
/**
|
||||
* Promisified version of zlib.gunzip
|
||||
|
||||
@@ -33,8 +33,6 @@ import {
|
||||
getResultSetNames,
|
||||
} from "./result-set-names";
|
||||
import { compareInterpretedResults } from "./interpreted-results";
|
||||
import { isCanary } from "../config";
|
||||
import { nanoid } from "nanoid";
|
||||
|
||||
interface ComparePair {
|
||||
from: CompletedLocalQueryInfo;
|
||||
@@ -118,13 +116,6 @@ export class CompareView extends AbstractWebview<
|
||||
panel.reveal(undefined, true);
|
||||
await this.waitForPanelLoaded();
|
||||
|
||||
await this.postMessage({
|
||||
t: "setUserSettings",
|
||||
userSettings: {
|
||||
shouldShowProvenance: isCanary(),
|
||||
},
|
||||
});
|
||||
|
||||
await this.postMessage({
|
||||
t: "setComparisonQueryInfo",
|
||||
stats: {
|
||||
@@ -184,97 +175,13 @@ export class CompareView extends AbstractWebview<
|
||||
message = getErrorMessage(e);
|
||||
}
|
||||
|
||||
await this.streamResults(result, currentResultSetDisplayName, message);
|
||||
}
|
||||
}
|
||||
|
||||
private async streamResults(
|
||||
result: QueryCompareResult | undefined,
|
||||
currentResultSetName: string,
|
||||
message: string | undefined,
|
||||
) {
|
||||
// Since there is a string limit of 1GB in Node.js, the comparison is send as a JSON.stringified string to the webview
|
||||
// and some comparisons may be larger than that, we sometimes need to stream results. This uses a heuristic of 2,000 results
|
||||
// to determine if we should stream results.
|
||||
|
||||
if (!this.shouldStreamResults(result)) {
|
||||
await this.postMessage({
|
||||
t: "setComparisons",
|
||||
result,
|
||||
currentResultSetName,
|
||||
currentResultSetName: currentResultSetDisplayName,
|
||||
message,
|
||||
});
|
||||
return;
|
||||
}
|
||||
|
||||
const id = nanoid();
|
||||
|
||||
// Streaming itself is implemented like this:
|
||||
// - 1 setup message which contains the first 1,000 results
|
||||
// - n "add results" messages which contain 1,000 results each
|
||||
// - 1 complete message which just tells the webview that we're done
|
||||
|
||||
await this.postMessage({
|
||||
t: "streamingComparisonSetup",
|
||||
id,
|
||||
result: this.chunkResults(result, 0, 1000),
|
||||
currentResultSetName,
|
||||
message,
|
||||
});
|
||||
|
||||
const { from, to } = result;
|
||||
|
||||
const maxResults = Math.max(from.length, to.length);
|
||||
for (let i = 1000; i < maxResults; i += 1000) {
|
||||
const chunk = this.chunkResults(result, i, i + 1000);
|
||||
|
||||
await this.postMessage({
|
||||
t: "streamingComparisonAddResults",
|
||||
id,
|
||||
result: chunk,
|
||||
});
|
||||
}
|
||||
|
||||
await this.postMessage({
|
||||
t: "streamingComparisonComplete",
|
||||
id,
|
||||
});
|
||||
}
|
||||
|
||||
private shouldStreamResults(
|
||||
result: QueryCompareResult | undefined,
|
||||
): result is QueryCompareResult {
|
||||
if (result === undefined) {
|
||||
return false;
|
||||
}
|
||||
|
||||
// We probably won't run into limits if we have less than 2,000 total results
|
||||
const totalResults = result.from.length + result.to.length;
|
||||
return totalResults > 2000;
|
||||
}
|
||||
|
||||
private chunkResults(
|
||||
result: QueryCompareResult,
|
||||
start: number,
|
||||
end: number,
|
||||
): QueryCompareResult {
|
||||
if (result.kind === "raw") {
|
||||
return {
|
||||
...result,
|
||||
from: result.from.slice(start, end),
|
||||
to: result.to.slice(start, end),
|
||||
};
|
||||
}
|
||||
|
||||
if (result.kind === "interpreted") {
|
||||
return {
|
||||
...result,
|
||||
from: result.from.slice(start, end),
|
||||
to: result.to.slice(start, end),
|
||||
};
|
||||
}
|
||||
|
||||
assertNever(result);
|
||||
}
|
||||
|
||||
protected getPanelConfig(): WebviewPanelConfig {
|
||||
|
||||
@@ -1,4 +1,4 @@
|
||||
import type { Location, Result, ThreadFlowLocation } from "sarif";
|
||||
import type { Location, Result } from "sarif";
|
||||
|
||||
function toCanonicalLocation(location: Location): Location {
|
||||
if (location.physicalLocation?.artifactLocation?.index !== undefined) {
|
||||
@@ -25,19 +25,6 @@ function toCanonicalLocation(location: Location): Location {
|
||||
return location;
|
||||
}
|
||||
|
||||
function toCanonicalThreadFlowLocation(
|
||||
threadFlowLocation: ThreadFlowLocation,
|
||||
): ThreadFlowLocation {
|
||||
if (threadFlowLocation.location) {
|
||||
return {
|
||||
...threadFlowLocation,
|
||||
location: toCanonicalLocation(threadFlowLocation.location),
|
||||
};
|
||||
}
|
||||
|
||||
return threadFlowLocation;
|
||||
}
|
||||
|
||||
function toCanonicalResult(result: Result): Result {
|
||||
const canonicalResult = {
|
||||
...result,
|
||||
@@ -53,30 +40,37 @@ function toCanonicalResult(result: Result): Result {
|
||||
canonicalResult.relatedLocations.map(toCanonicalLocation);
|
||||
}
|
||||
|
||||
if (canonicalResult.codeFlows && canonicalResult.codeFlows.length > 0) {
|
||||
// If there are codeFlows, we don't want to compare the full codeFlows. Instead, we just want to compare the
|
||||
// source and the sink (i.e. the first and last item). CodeQL should guarantee that the first and last threadFlow
|
||||
// of every codeFlow is the same (i.e. every codeFlow has the same source and sink). Therefore, we just compare the
|
||||
// first codeFlow and ignore the other codeFlows completely.
|
||||
// If the codeFlow has a length of 1, this doesn't change the result.
|
||||
if (canonicalResult.codeFlows) {
|
||||
canonicalResult.codeFlows = canonicalResult.codeFlows.map((codeFlow) => {
|
||||
if (codeFlow.threadFlows) {
|
||||
return {
|
||||
...codeFlow,
|
||||
threadFlows: codeFlow.threadFlows.map((threadFlow) => {
|
||||
if (threadFlow.locations) {
|
||||
return {
|
||||
...threadFlow,
|
||||
locations: threadFlow.locations.map((threadFlowLocation) => {
|
||||
if (threadFlowLocation.location) {
|
||||
return {
|
||||
...threadFlowLocation,
|
||||
location: toCanonicalLocation(
|
||||
threadFlowLocation.location,
|
||||
),
|
||||
};
|
||||
}
|
||||
|
||||
const source = {
|
||||
...canonicalResult.codeFlows[0].threadFlows[0],
|
||||
};
|
||||
const sink = {
|
||||
...canonicalResult.codeFlows[0].threadFlows[
|
||||
canonicalResult.codeFlows[0].threadFlows.length - 1
|
||||
],
|
||||
};
|
||||
source.locations = source.locations.map(toCanonicalThreadFlowLocation);
|
||||
sink.locations = sink.locations.map(toCanonicalThreadFlowLocation);
|
||||
return threadFlowLocation;
|
||||
}),
|
||||
};
|
||||
}
|
||||
|
||||
canonicalResult.codeFlows = [
|
||||
{
|
||||
...canonicalResult.codeFlows[0],
|
||||
threadFlows: [source, sink],
|
||||
},
|
||||
];
|
||||
return threadFlow;
|
||||
}),
|
||||
};
|
||||
}
|
||||
|
||||
return codeFlow;
|
||||
});
|
||||
}
|
||||
|
||||
return canonicalResult;
|
||||
@@ -85,9 +79,11 @@ function toCanonicalResult(result: Result): Result {
|
||||
/**
|
||||
* Compare the alerts of two queries. Use deep equality to determine if
|
||||
* results have been added or removed across two invocations of a query.
|
||||
* It first canonicalizes the results to ensure that when small changes
|
||||
* to the query are made, the results are still considered the same. This
|
||||
* includes the removal of all paths except for the source and sink.
|
||||
*
|
||||
* Assumptions:
|
||||
*
|
||||
* 1. Queries have the same sort order
|
||||
* 2. Results are not changed or re-ordered, they are only added or removed
|
||||
*
|
||||
* @param fromResults the source query
|
||||
* @param toResults the target query
|
||||
@@ -108,30 +104,19 @@ export function sarifDiff(fromResults: Result[], toResults: Result[]) {
|
||||
const canonicalFromResults = fromResults.map(toCanonicalResult);
|
||||
const canonicalToResults = toResults.map(toCanonicalResult);
|
||||
|
||||
const diffResults = {
|
||||
const results = {
|
||||
from: arrayDiff(canonicalFromResults, canonicalToResults),
|
||||
to: arrayDiff(canonicalToResults, canonicalFromResults),
|
||||
};
|
||||
|
||||
if (
|
||||
fromResults.length === diffResults.from.length &&
|
||||
toResults.length === diffResults.to.length
|
||||
fromResults.length === results.from.length &&
|
||||
toResults.length === results.to.length
|
||||
) {
|
||||
throw new Error("CodeQL Compare: No overlap between the selected queries.");
|
||||
}
|
||||
|
||||
// We don't want to return the canonical results, we want to return the original results.
|
||||
// We can retrieve this by finding the index of the canonical result in the canonical results
|
||||
// and then using that index to find the original result. This is possible because we know that
|
||||
// we did a 1-to-1 map between the canonical results and the original results.
|
||||
return {
|
||||
from: diffResults.from.map(
|
||||
(result) => fromResults[canonicalFromResults.indexOf(result)],
|
||||
),
|
||||
to: diffResults.to.map(
|
||||
(result) => toResults[canonicalToResults.indexOf(result)],
|
||||
),
|
||||
};
|
||||
return results;
|
||||
}
|
||||
|
||||
function arrayDiff<T>(source: readonly T[], toRemove: readonly T[]): T[] {
|
||||
|
||||
@@ -112,7 +112,9 @@ export function hasEnterpriseUri(): boolean {
|
||||
* Does the uri look like GHEC-DR?
|
||||
*/
|
||||
function isGhecDrUri(uri: Uri | undefined): boolean {
|
||||
return uri !== undefined && uri.authority.toLowerCase().endsWith(".ghe.com");
|
||||
return (
|
||||
uri !== undefined && !uri.authority.toLowerCase().endsWith("github.com")
|
||||
);
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -591,27 +593,7 @@ export const NO_CACHE_CONTEXTUAL_QUERIES = new Setting(
|
||||
// Settings for variant analysis
|
||||
const VARIANT_ANALYSIS_SETTING = new Setting("variantAnalysis", ROOT_SETTING);
|
||||
|
||||
/**
|
||||
* The name of the "controller" repository that you want to use with the "Run Variant Analysis" command.
|
||||
* Note: This command is only available for internal users.
|
||||
*
|
||||
* This setting should be a GitHub repository of the form `<owner>/<repo>`.
|
||||
*/
|
||||
const REMOTE_CONTROLLER_REPO = new Setting(
|
||||
"controllerRepo",
|
||||
VARIANT_ANALYSIS_SETTING,
|
||||
);
|
||||
|
||||
export function getRemoteControllerRepo(): string | undefined {
|
||||
return REMOTE_CONTROLLER_REPO.getValue<string>() || undefined;
|
||||
}
|
||||
|
||||
export async function setRemoteControllerRepo(repo: string | undefined) {
|
||||
await REMOTE_CONTROLLER_REPO.updateValue(repo, ConfigurationTarget.Global);
|
||||
}
|
||||
|
||||
export interface VariantAnalysisConfig {
|
||||
controllerRepo: string | undefined;
|
||||
showSystemDefinedRepositoryLists: boolean;
|
||||
/**
|
||||
* This uses a URL instead of a URI because the URL class is available in
|
||||
@@ -632,10 +614,6 @@ export class VariantAnalysisConfigListener
|
||||
);
|
||||
}
|
||||
|
||||
public get controllerRepo(): string | undefined {
|
||||
return getRemoteControllerRepo();
|
||||
}
|
||||
|
||||
public get showSystemDefinedRepositoryLists(): boolean {
|
||||
return !hasEnterpriseUri();
|
||||
}
|
||||
@@ -828,9 +806,19 @@ export async function setAutogenerateQlPacks(choice: AutogenerateQLPacks) {
|
||||
|
||||
const MODEL_SETTING = new Setting("model", ROOT_SETTING);
|
||||
const FLOW_GENERATION = new Setting("flowGeneration", MODEL_SETTING);
|
||||
const LLM_GENERATION = new Setting("llmGeneration", MODEL_SETTING);
|
||||
const LLM_GENERATION_BATCH_SIZE = new Setting(
|
||||
"llmGenerationBatchSize",
|
||||
MODEL_SETTING,
|
||||
);
|
||||
const LLM_GENERATION_DEV_ENDPOINT = new Setting(
|
||||
"llmGenerationDevEndpoint",
|
||||
MODEL_SETTING,
|
||||
);
|
||||
const MODEL_EVALUATION = new Setting("evaluation", MODEL_SETTING);
|
||||
const MODEL_PACK_LOCATION = new Setting("packLocation", MODEL_SETTING);
|
||||
const MODEL_PACK_NAME = new Setting("packName", MODEL_SETTING);
|
||||
const ENABLE_PYTHON = new Setting("enablePython", MODEL_SETTING);
|
||||
|
||||
export type ModelConfigPackVariables = {
|
||||
database: string;
|
||||
@@ -841,11 +829,13 @@ export type ModelConfigPackVariables = {
|
||||
|
||||
export interface ModelConfig {
|
||||
flowGeneration: boolean;
|
||||
llmGeneration: boolean;
|
||||
getPackLocation(
|
||||
languageId: string,
|
||||
variables: ModelConfigPackVariables,
|
||||
): string;
|
||||
getPackName(languageId: string, variables: ModelConfigPackVariables): string;
|
||||
enablePython: boolean;
|
||||
}
|
||||
|
||||
export class ModelConfigListener extends ConfigListener implements ModelConfig {
|
||||
@@ -860,6 +850,26 @@ export class ModelConfigListener extends ConfigListener implements ModelConfig {
|
||||
return !!FLOW_GENERATION.getValue<boolean>();
|
||||
}
|
||||
|
||||
public get llmGeneration(): boolean {
|
||||
return !!LLM_GENERATION.getValue<boolean>() && !hasEnterpriseUri();
|
||||
}
|
||||
|
||||
/**
|
||||
* Limits the number of candidates we send to the model in each request to avoid long requests.
|
||||
* Note that the model may return fewer than this number of candidates.
|
||||
*/
|
||||
public get llmGenerationBatchSize(): number {
|
||||
return LLM_GENERATION_BATCH_SIZE.getValue<number | null>() || 5;
|
||||
}
|
||||
|
||||
/**
|
||||
* The URL of the endpoint to use for LLM generation. This should only be set
|
||||
* if you want to test against a dev server.
|
||||
*/
|
||||
public get llmGenerationDevEndpoint(): string | undefined {
|
||||
return LLM_GENERATION_DEV_ENDPOINT.getValue<string | undefined>();
|
||||
}
|
||||
|
||||
public get modelEvaluation(): boolean {
|
||||
return !!MODEL_EVALUATION.getValue<boolean>();
|
||||
}
|
||||
@@ -887,6 +897,10 @@ export class ModelConfigListener extends ConfigListener implements ModelConfig {
|
||||
variables,
|
||||
);
|
||||
}
|
||||
|
||||
public get enablePython(): boolean {
|
||||
return !!ENABLE_PYTHON.getValue<boolean>();
|
||||
}
|
||||
}
|
||||
|
||||
const GITHUB_DATABASE_SETTING = new Setting("githubDatabase", ROOT_SETTING);
|
||||
|
||||
@@ -298,7 +298,7 @@ export class DbConfigStore extends DisposableObject {
|
||||
let newConfig: DbConfig | undefined = undefined;
|
||||
try {
|
||||
newConfig = await readJSON(this.configPath);
|
||||
} catch {
|
||||
} catch (e) {
|
||||
this.configErrors = [
|
||||
{
|
||||
kind: DbConfigValidationErrorKind.InvalidJson,
|
||||
@@ -332,7 +332,7 @@ export class DbConfigStore extends DisposableObject {
|
||||
let newConfig: DbConfig | undefined = undefined;
|
||||
try {
|
||||
newConfig = readJSONSync(this.configPath);
|
||||
} catch {
|
||||
} catch (e) {
|
||||
this.configErrors = [
|
||||
{
|
||||
kind: DbConfigValidationErrorKind.InvalidJson,
|
||||
|
||||
@@ -1,3 +1,5 @@
|
||||
import type { Response } from "node-fetch";
|
||||
import fetch, { AbortError } from "node-fetch";
|
||||
import type { InputBoxOptions } from "vscode";
|
||||
import { Uri, window } from "vscode";
|
||||
import type { CodeQLCliServer } from "../codeql-cli/cli";
|
||||
@@ -452,7 +454,7 @@ export class DatabaseFetcher {
|
||||
let uri;
|
||||
try {
|
||||
uri = Uri.parse(databaseUrl, true);
|
||||
} catch {
|
||||
} catch (e) {
|
||||
throw new Error(`Invalid url: ${databaseUrl}`);
|
||||
}
|
||||
|
||||
@@ -534,8 +536,8 @@ export class DatabaseFetcher {
|
||||
} catch (e) {
|
||||
disposeTimeout();
|
||||
|
||||
if (e instanceof DOMException && e.name === "AbortError") {
|
||||
const thrownError = new Error("The request timed out.");
|
||||
if (e instanceof AbortError) {
|
||||
const thrownError = new AbortError("The request timed out.");
|
||||
thrownError.stack = e.stack;
|
||||
throw thrownError;
|
||||
}
|
||||
@@ -543,52 +545,30 @@ export class DatabaseFetcher {
|
||||
throw e;
|
||||
}
|
||||
|
||||
const body = response.body;
|
||||
if (!body) {
|
||||
throw new Error("No response body found");
|
||||
}
|
||||
|
||||
const archiveFileStream = createWriteStream(archivePath);
|
||||
|
||||
const contentLength = response.headers.get("content-length");
|
||||
const totalNumBytes = contentLength
|
||||
? parseInt(contentLength, 10)
|
||||
: undefined;
|
||||
|
||||
const reportProgress = reportStreamProgress(
|
||||
reportStreamProgress(
|
||||
response.body,
|
||||
"Downloading database",
|
||||
totalNumBytes,
|
||||
progress,
|
||||
);
|
||||
|
||||
response.body.on("data", onData);
|
||||
|
||||
try {
|
||||
const reader = body.getReader();
|
||||
for (;;) {
|
||||
const { done, value } = await reader.read();
|
||||
if (done) {
|
||||
break;
|
||||
}
|
||||
|
||||
onData();
|
||||
reportProgress(value?.length ?? 0);
|
||||
|
||||
await new Promise((resolve, reject) => {
|
||||
archiveFileStream.write(value, (err) => {
|
||||
if (err) {
|
||||
reject(err);
|
||||
}
|
||||
resolve(undefined);
|
||||
});
|
||||
});
|
||||
}
|
||||
|
||||
await new Promise((resolve, reject) => {
|
||||
archiveFileStream.close((err) => {
|
||||
if (err) {
|
||||
reject(err);
|
||||
}
|
||||
resolve(undefined);
|
||||
});
|
||||
response.body
|
||||
.pipe(archiveFileStream)
|
||||
.on("finish", resolve)
|
||||
.on("error", reject);
|
||||
|
||||
// If an error occurs on the body, we also want to reject the promise (e.g. during a timeout error).
|
||||
response.body.on("error", reject);
|
||||
});
|
||||
} catch (e) {
|
||||
// Close and remove the file if an error occurs
|
||||
@@ -596,8 +576,8 @@ export class DatabaseFetcher {
|
||||
void remove(archivePath);
|
||||
});
|
||||
|
||||
if (e instanceof DOMException && e.name === "AbortError") {
|
||||
const thrownError = new Error("The download timed out.");
|
||||
if (e instanceof AbortError) {
|
||||
const thrownError = new AbortError("The download timed out.");
|
||||
thrownError.stack = e.stack;
|
||||
throw thrownError;
|
||||
}
|
||||
@@ -632,7 +612,7 @@ export class DatabaseFetcher {
|
||||
const obj = JSON.parse(text);
|
||||
msg =
|
||||
obj.error || obj.message || obj.reason || JSON.stringify(obj, null, 2);
|
||||
} catch {
|
||||
} catch (e) {
|
||||
msg = text;
|
||||
}
|
||||
throw new Error(`${errorMessage}.\n\nReason: ${msg}`);
|
||||
|
||||
@@ -231,7 +231,7 @@ export class DatabaseManager extends DisposableObject {
|
||||
let originStat;
|
||||
try {
|
||||
originStat = await stat(originDbYml);
|
||||
} catch {
|
||||
} catch (e) {
|
||||
// if there is an error here, assume that the origin database
|
||||
// is no longer available. Safely ignore and do not try to re-import.
|
||||
return false;
|
||||
@@ -240,7 +240,7 @@ export class DatabaseManager extends DisposableObject {
|
||||
try {
|
||||
const importedStat = await stat(importedDbYml);
|
||||
return originStat.mtimeMs > importedStat.mtimeMs;
|
||||
} catch {
|
||||
} catch (e) {
|
||||
// If either of the files does not exist, we assume the origin is newer.
|
||||
// This shouldn't happen unless the user manually deleted one of the files.
|
||||
return true;
|
||||
|
||||
@@ -37,12 +37,11 @@ export const shownLocationLineDecoration =
|
||||
/**
|
||||
* Resolves the specified CodeQL location to a URI into the source archive.
|
||||
* @param loc CodeQL location to resolve. Must have a non-empty value for `loc.file`.
|
||||
* @param databaseItem Database in which to resolve the file location, or `undefined` to resolve
|
||||
* from the local file system.
|
||||
* @param databaseItem Database in which to resolve the file location.
|
||||
*/
|
||||
function resolveFivePartLocation(
|
||||
loc: UrlValueLineColumnLocation,
|
||||
databaseItem: DatabaseItem | undefined,
|
||||
databaseItem: DatabaseItem,
|
||||
): Location {
|
||||
// `Range` is a half-open interval, and is zero-based. CodeQL locations are closed intervals, and
|
||||
// are one-based. Adjust accordingly.
|
||||
@@ -53,10 +52,7 @@ function resolveFivePartLocation(
|
||||
Math.max(1, loc.endColumn),
|
||||
);
|
||||
|
||||
return new Location(
|
||||
databaseItem?.resolveSourceFile(loc.uri) ?? Uri.parse(loc.uri),
|
||||
range,
|
||||
);
|
||||
return new Location(databaseItem.resolveSourceFile(loc.uri), range);
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -66,26 +62,22 @@ function resolveFivePartLocation(
|
||||
*/
|
||||
function resolveWholeFileLocation(
|
||||
loc: UrlValueWholeFileLocation,
|
||||
databaseItem: DatabaseItem | undefined,
|
||||
databaseItem: DatabaseItem,
|
||||
): Location {
|
||||
// A location corresponding to the start of the file.
|
||||
const range = new Range(0, 0, 0, 0);
|
||||
return new Location(
|
||||
databaseItem?.resolveSourceFile(loc.uri) ?? Uri.parse(loc.uri),
|
||||
range,
|
||||
);
|
||||
return new Location(databaseItem.resolveSourceFile(loc.uri), range);
|
||||
}
|
||||
|
||||
/**
|
||||
* Try to resolve the specified CodeQL location to a URI into the source archive. If no exact location
|
||||
* can be resolved, returns `undefined`.
|
||||
* @param loc CodeQL location to resolve
|
||||
* @param databaseItem Database in which to resolve the file location, or `undefined` to resolve
|
||||
* from the local file system.
|
||||
* @param databaseItem Database in which to resolve the file location.
|
||||
*/
|
||||
export function tryResolveLocation(
|
||||
loc: UrlValueResolvable | undefined,
|
||||
databaseItem: DatabaseItem | undefined,
|
||||
databaseItem: DatabaseItem,
|
||||
): Location | undefined {
|
||||
if (!loc) {
|
||||
return;
|
||||
@@ -103,7 +95,7 @@ export function tryResolveLocation(
|
||||
|
||||
export async function showResolvableLocation(
|
||||
loc: UrlValueResolvable,
|
||||
databaseItem: DatabaseItem | undefined,
|
||||
databaseItem: DatabaseItem,
|
||||
logger: Logger,
|
||||
): Promise<void> {
|
||||
try {
|
||||
@@ -159,14 +151,13 @@ export async function showLocation(location?: Location) {
|
||||
}
|
||||
|
||||
export async function jumpToLocation(
|
||||
databaseUri: string | undefined,
|
||||
databaseUri: string,
|
||||
loc: UrlValueResolvable,
|
||||
databaseManager: DatabaseManager,
|
||||
logger: Logger,
|
||||
) {
|
||||
const databaseItem =
|
||||
databaseUri !== undefined
|
||||
? databaseManager.findDatabaseItem(Uri.parse(databaseUri))
|
||||
: undefined;
|
||||
await showResolvableLocation(loc, databaseItem, logger);
|
||||
const databaseItem = databaseManager.findDatabaseItem(Uri.parse(databaseUri));
|
||||
if (databaseItem !== undefined) {
|
||||
await showResolvableLocation(loc, databaseItem, logger);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -18,8 +18,6 @@ import type { DbManager } from "../db-manager";
|
||||
import { DbTreeDataProvider } from "./db-tree-data-provider";
|
||||
import type { DbTreeViewItem } from "./db-tree-view-item";
|
||||
import { getGitHubUrl } from "./db-tree-view-item-action";
|
||||
import { getControllerRepo } from "../../variant-analysis/run-remote-query";
|
||||
import { getErrorMessage } from "../../common/helpers-pure";
|
||||
import type { DatabasePanelCommands } from "../../common/commands";
|
||||
import type { App } from "../../common/app";
|
||||
import { QueryLanguage } from "../../common/query-language";
|
||||
@@ -74,9 +72,6 @@ export class DbPanel extends DisposableObject {
|
||||
this.addNewRemoteDatabase.bind(this),
|
||||
"codeQLVariantAnalysisRepositories.addNewList":
|
||||
this.addNewList.bind(this),
|
||||
"codeQLVariantAnalysisRepositories.setupControllerRepository":
|
||||
this.setupControllerRepository.bind(this),
|
||||
|
||||
"codeQLVariantAnalysisRepositories.setSelectedItem":
|
||||
this.setSelectedItem.bind(this),
|
||||
"codeQLVariantAnalysisRepositories.setSelectedItemContextMenu":
|
||||
@@ -427,22 +422,4 @@ export class DbPanel extends DisposableObject {
|
||||
|
||||
await this.app.commands.execute("vscode.open", Uri.parse(githubUrl));
|
||||
}
|
||||
|
||||
private async setupControllerRepository(): Promise<void> {
|
||||
try {
|
||||
// This will also validate that the controller repository is valid
|
||||
await getControllerRepo(this.app.credentials);
|
||||
} catch (e: unknown) {
|
||||
if (e instanceof UserCancellationException) {
|
||||
return;
|
||||
}
|
||||
|
||||
void showAndLogErrorMessage(
|
||||
this.app.logger,
|
||||
`An error occurred while setting up the controller repository: ${getErrorMessage(
|
||||
e,
|
||||
)}`,
|
||||
);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
@@ -83,11 +83,6 @@ export class DbTreeDataProvider
|
||||
}
|
||||
|
||||
private createTree(): DbTreeViewItem[] {
|
||||
// Returning an empty tree here will show the welcome view
|
||||
if (!this.variantAnalysisConfig.controllerRepo) {
|
||||
return [];
|
||||
}
|
||||
|
||||
const dbItemsResult = this.dbManager.getDbItems();
|
||||
|
||||
if (dbItemsResult.isFailure) {
|
||||
|
||||
@@ -96,7 +96,7 @@ export type Response = DebugProtocol.Response & { type: "response" };
|
||||
export type InitializeResponse = DebugProtocol.InitializeResponse &
|
||||
Response & { command: "initialize" };
|
||||
|
||||
export type QuickEvalResponse = Response;
|
||||
export interface QuickEvalResponse extends Response {}
|
||||
|
||||
export type AnyResponse = InitializeResponse | QuickEvalResponse;
|
||||
|
||||
|
||||
@@ -300,12 +300,12 @@ const shouldUpdateOnNextActivationKey = "shouldUpdateOnNextActivation";
|
||||
|
||||
const codeQlVersionRange = DEFAULT_DISTRIBUTION_VERSION_RANGE;
|
||||
|
||||
// This is the minimum version of vscode that we _want_ to support. We want to update to Node 20, but that
|
||||
// requires 1.90 or later. If we change the minimum version in the package.json, then anyone on an older version of vscode will
|
||||
// This is the minimum version of vscode that we _want_ to support. We want to update to Node 18, but that
|
||||
// requires 1.82 or later. If we change the minimum version in the package.json, then anyone on an older version of vscode will
|
||||
// silently be unable to upgrade. So, the solution is to first bump the minimum version here and release. Then
|
||||
// bump the version in the package.json and release again. This way, anyone on an older version of vscode will get a warning
|
||||
// before silently being refused to upgrade.
|
||||
const MIN_VERSION = "1.90.0";
|
||||
const MIN_VERSION = "1.82.0";
|
||||
|
||||
function sendConfigTelemetryData() {
|
||||
const config: Record<string, string> = {};
|
||||
@@ -362,9 +362,7 @@ export async function activate(
|
||||
distributionConfigListener,
|
||||
codeQlVersionRange,
|
||||
ctx,
|
||||
app.logger,
|
||||
);
|
||||
await distributionManager.initialize();
|
||||
|
||||
registerErrorStubs([checkForUpdatesCommand], (command) => async () => {
|
||||
void showAndLogErrorMessage(
|
||||
@@ -1125,8 +1123,6 @@ async function activateWithInstalledDistribution(
|
||||
void extLogger.log("Reading query history");
|
||||
await qhm.readQueryHistory();
|
||||
|
||||
distributionManager.startCleanup();
|
||||
|
||||
void extLogger.log("Successfully finished extension initialization.");
|
||||
|
||||
return {
|
||||
|
||||
@@ -31,7 +31,7 @@ export function fileRangeFromURI(
|
||||
return new Location(db.resolveSourceFile(uri.uri), range);
|
||||
}
|
||||
return undefined;
|
||||
} catch {
|
||||
} catch (e) {
|
||||
return undefined;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -28,7 +28,7 @@ export async function resolveContextualQlPacksForDatabase(
|
||||
): Promise<QlPacksForLanguage> {
|
||||
try {
|
||||
return await qlpackOfDatabase(cli, databaseItem);
|
||||
} catch {
|
||||
} catch (e) {
|
||||
// If we can't find the qlpacks for the database, use the defaults instead
|
||||
}
|
||||
|
||||
|
||||
@@ -12,7 +12,7 @@ import { languages, IndentAction } from "vscode";
|
||||
* See https://github.com/microsoft/vscode/blob/master/src/vs/editor/test/common/modes/supports/javascriptOnEnterRules.ts
|
||||
*/
|
||||
export function install() {
|
||||
// eslint-disable-next-line @typescript-eslint/no-require-imports
|
||||
// eslint-disable-next-line @typescript-eslint/no-var-requires
|
||||
const langConfig = require("../../language-configuration.json");
|
||||
// setLanguageConfiguration requires a regexp for the wordpattern, not a string
|
||||
langConfig.wordPattern = new RegExp(langConfig.wordPattern);
|
||||
|
||||
@@ -537,14 +537,6 @@ export class ResultsView extends AbstractWebview<
|
||||
resultSetNames,
|
||||
};
|
||||
|
||||
await this.postMessage({
|
||||
t: "setUserSettings",
|
||||
userSettings: {
|
||||
// Only show provenance info in canary mode for now.
|
||||
shouldShowProvenance: isCanary(),
|
||||
},
|
||||
});
|
||||
|
||||
await this.postMessage({
|
||||
t: "setState",
|
||||
interpretation: interpretationPage,
|
||||
|
||||
65
extensions/ql-vscode/src/model-editor/auto-model-api.ts
Normal file
65
extensions/ql-vscode/src/model-editor/auto-model-api.ts
Normal file
@@ -0,0 +1,65 @@
|
||||
import type { Credentials } from "../common/authentication";
|
||||
import type { OctokitResponse } from "@octokit/types";
|
||||
import fetch from "node-fetch";
|
||||
import type { ModelConfigListener } from "../config";
|
||||
|
||||
export enum AutomodelMode {
|
||||
Unspecified = "AUTOMODEL_MODE_UNSPECIFIED",
|
||||
Framework = "AUTOMODEL_MODE_FRAMEWORK",
|
||||
Application = "AUTOMODEL_MODE_APPLICATION",
|
||||
}
|
||||
|
||||
export interface ModelRequest {
|
||||
mode: AutomodelMode;
|
||||
// Base64-encoded GZIP-compressed SARIF log
|
||||
candidates: string;
|
||||
}
|
||||
|
||||
export interface ModelResponse {
|
||||
models: string;
|
||||
}
|
||||
|
||||
export async function autoModel(
|
||||
credentials: Credentials,
|
||||
request: ModelRequest,
|
||||
modelingConfig: ModelConfigListener,
|
||||
): Promise<ModelResponse> {
|
||||
const devEndpoint = modelingConfig.llmGenerationDevEndpoint;
|
||||
if (devEndpoint) {
|
||||
return callAutoModelDevEndpoint(devEndpoint, request);
|
||||
} else {
|
||||
const octokit = await credentials.getOctokit();
|
||||
|
||||
const response: OctokitResponse<ModelResponse> = await octokit.request(
|
||||
"POST /repos/github/codeql/code-scanning/codeql/auto-model",
|
||||
{
|
||||
data: request,
|
||||
},
|
||||
);
|
||||
|
||||
return response.data;
|
||||
}
|
||||
}
|
||||
|
||||
async function callAutoModelDevEndpoint(
|
||||
endpoint: string,
|
||||
request: ModelRequest,
|
||||
): Promise<ModelResponse> {
|
||||
const json = JSON.stringify(request);
|
||||
const response = await fetch(endpoint, {
|
||||
method: "POST",
|
||||
headers: {
|
||||
"Content-Type": "application/json",
|
||||
},
|
||||
body: json,
|
||||
});
|
||||
|
||||
if (!response.ok) {
|
||||
throw new Error(
|
||||
`Error calling auto-model API: ${response.status} ${response.statusText}`,
|
||||
);
|
||||
}
|
||||
|
||||
const data = await response.json();
|
||||
return data as ModelResponse;
|
||||
}
|
||||
@@ -0,0 +1,233 @@
|
||||
import type { CodeQLCliServer, SourceInfo } from "../codeql-cli/cli";
|
||||
import type { CoreCompletedQuery, QueryRunner } from "../query-server";
|
||||
import type { DatabaseItem } from "../databases/local-databases";
|
||||
import type { ProgressCallback } from "../common/vscode/progress";
|
||||
import type { Log } from "sarif";
|
||||
import type { Mode } from "./shared/mode";
|
||||
import { getOnDiskWorkspaceFolders } from "../common/vscode/workspace-folders";
|
||||
import { interpretResultsSarif } from "../query-results";
|
||||
import { join } from "path";
|
||||
import { dir } from "tmp-promise";
|
||||
import { writeFile, outputFile } from "fs-extra";
|
||||
import { dump as dumpYaml } from "js-yaml";
|
||||
import type { MethodSignature } from "./method";
|
||||
import { runQuery } from "../local-queries/run-query";
|
||||
import type { QueryMetadata } from "../common/interface-types";
|
||||
import type { CancellationTokenSource } from "vscode";
|
||||
import { resolveQueries } from "../local-queries";
|
||||
import { modeTag } from "./mode-tag";
|
||||
|
||||
type AutoModelQueriesOptions = {
|
||||
mode: Mode;
|
||||
candidateMethods: MethodSignature[];
|
||||
cliServer: CodeQLCliServer;
|
||||
queryRunner: QueryRunner;
|
||||
databaseItem: DatabaseItem;
|
||||
queryStorageDir: string;
|
||||
|
||||
progress: ProgressCallback;
|
||||
cancellationTokenSource: CancellationTokenSource;
|
||||
};
|
||||
|
||||
export type AutoModelQueriesResult = {
|
||||
candidates: Log;
|
||||
};
|
||||
|
||||
export async function runAutoModelQueries({
|
||||
mode,
|
||||
candidateMethods,
|
||||
cliServer,
|
||||
queryRunner,
|
||||
databaseItem,
|
||||
queryStorageDir,
|
||||
progress,
|
||||
cancellationTokenSource,
|
||||
}: AutoModelQueriesOptions): Promise<AutoModelQueriesResult | undefined> {
|
||||
// First, resolve the query that we want to run.
|
||||
const queryPath = await resolveAutomodelQuery(
|
||||
cliServer,
|
||||
databaseItem,
|
||||
"candidates",
|
||||
mode,
|
||||
);
|
||||
|
||||
// Generate a pack containing the candidate filters
|
||||
const { packDir: filterPackDir, cleanup: cleanupFilterPack } =
|
||||
await generateCandidateFilterPack(databaseItem.language, candidateMethods);
|
||||
|
||||
const additionalPacks = [...getOnDiskWorkspaceFolders(), filterPackDir];
|
||||
const extensionPacks = Object.keys(
|
||||
await cliServer.resolveQlpacks(additionalPacks, true),
|
||||
);
|
||||
|
||||
// Run the actual query
|
||||
const completedQuery = await runQuery({
|
||||
queryRunner,
|
||||
databaseItem,
|
||||
queryPath,
|
||||
queryStorageDir,
|
||||
additionalPacks,
|
||||
extensionPacks,
|
||||
progress,
|
||||
token: cancellationTokenSource.token,
|
||||
});
|
||||
|
||||
await cleanupFilterPack();
|
||||
|
||||
if (!completedQuery) {
|
||||
return undefined;
|
||||
}
|
||||
|
||||
// Get metadata for the query. This is required to interpret the results. We already know the kind is problem
|
||||
// (because of the constraint in resolveQueries), so we don't need any more checks on the metadata.
|
||||
const metadata = await cliServer.resolveMetadata(queryPath);
|
||||
|
||||
// CodeQL needs to have access to the database to be able to retrieve the
|
||||
// snippets from it. The source location prefix is used to determine the
|
||||
// base path of the database.
|
||||
const sourceLocationPrefix =
|
||||
await databaseItem.getSourceLocationPrefix(cliServer);
|
||||
const sourceArchiveUri = databaseItem.sourceArchive;
|
||||
const sourceInfo =
|
||||
sourceArchiveUri === undefined
|
||||
? undefined
|
||||
: {
|
||||
sourceArchive: sourceArchiveUri.fsPath,
|
||||
sourceLocationPrefix,
|
||||
};
|
||||
|
||||
const candidates = await interpretAutomodelResults(
|
||||
cliServer,
|
||||
completedQuery,
|
||||
metadata,
|
||||
sourceInfo,
|
||||
);
|
||||
|
||||
return {
|
||||
candidates,
|
||||
};
|
||||
}
|
||||
|
||||
async function resolveAutomodelQuery(
|
||||
cliServer: CodeQLCliServer,
|
||||
databaseItem: DatabaseItem,
|
||||
queryTag: string,
|
||||
mode: Mode,
|
||||
): Promise<string> {
|
||||
const packsToSearch = [`codeql/${databaseItem.language}-automodel-queries`];
|
||||
|
||||
// First, resolve the query that we want to run.
|
||||
// All queries are tagged like this:
|
||||
// internal extract automodel <mode> <queryTag>
|
||||
// Example: internal extract automodel framework-mode candidates
|
||||
const queries = await resolveQueries(
|
||||
cliServer,
|
||||
packsToSearch,
|
||||
`Extract automodel ${queryTag}`,
|
||||
{
|
||||
kind: "problem",
|
||||
"tags contain all": ["automodel", modeTag(mode), ...queryTag.split(" ")],
|
||||
},
|
||||
);
|
||||
if (queries.length > 1) {
|
||||
throw new Error(
|
||||
`Found multiple auto model queries for ${mode} ${queryTag}. Can't continue`,
|
||||
);
|
||||
}
|
||||
if (queries.length === 0) {
|
||||
throw new Error(
|
||||
`Did not found any auto model queries for ${mode} ${queryTag}. Can't continue`,
|
||||
);
|
||||
}
|
||||
|
||||
return queries[0];
|
||||
}
|
||||
|
||||
type CandidateFilterPackResult = {
|
||||
packDir: string;
|
||||
cleanup: () => Promise<void>;
|
||||
};
|
||||
|
||||
/**
|
||||
* generateCandidateFilterPack will create a temporary extension pack.
|
||||
* This pack will contain a filter that will restrict the automodel queries
|
||||
* to the specified candidate methods only.
|
||||
* This is done using the `extensible` predicate "automodelCandidateFilter".
|
||||
* @param language
|
||||
* @param candidateMethods
|
||||
* @returns
|
||||
*/
|
||||
export async function generateCandidateFilterPack(
|
||||
language: string,
|
||||
candidateMethods: MethodSignature[],
|
||||
): Promise<CandidateFilterPackResult> {
|
||||
// Pack resides in a temporary directory, to not pollute the workspace.
|
||||
const { path: packDir, cleanup } = await dir({ unsafeCleanup: true });
|
||||
|
||||
const syntheticConfigPack = {
|
||||
name: "codeql/automodel-filter",
|
||||
version: "0.0.0",
|
||||
library: true,
|
||||
extensionTargets: {
|
||||
[`codeql/${language}-automodel-queries`]: "*",
|
||||
},
|
||||
dataExtensions: ["filter.yml"],
|
||||
};
|
||||
|
||||
const qlpackFile = join(packDir, "codeql-pack.yml");
|
||||
await outputFile(qlpackFile, dumpYaml(syntheticConfigPack), "utf8");
|
||||
|
||||
// The predicate has the following defintion:
|
||||
// extensible predicate automodelCandidateFilter(string package, string type, string name, string signature)
|
||||
const dataRows = candidateMethods.map((method) => [
|
||||
method.packageName,
|
||||
method.typeName,
|
||||
method.methodName,
|
||||
method.methodParameters,
|
||||
]);
|
||||
|
||||
const filter = {
|
||||
extensions: [
|
||||
{
|
||||
addsTo: {
|
||||
pack: `codeql/${language}-automodel-queries`,
|
||||
extensible: "automodelCandidateFilter",
|
||||
},
|
||||
data: dataRows,
|
||||
},
|
||||
],
|
||||
};
|
||||
|
||||
const filterFile = join(packDir, "filter.yml");
|
||||
await writeFile(filterFile, dumpYaml(filter), "utf8");
|
||||
|
||||
return {
|
||||
packDir,
|
||||
cleanup,
|
||||
};
|
||||
}
|
||||
|
||||
async function interpretAutomodelResults(
|
||||
cliServer: CodeQLCliServer,
|
||||
completedQuery: CoreCompletedQuery,
|
||||
metadata: QueryMetadata,
|
||||
sourceInfo: SourceInfo | undefined,
|
||||
): Promise<Log> {
|
||||
const interpretedResultsPath = join(
|
||||
completedQuery.outputDir.querySaveDir,
|
||||
"results.sarif",
|
||||
);
|
||||
|
||||
const { ...sarif } = await interpretResultsSarif(
|
||||
cliServer,
|
||||
metadata,
|
||||
{
|
||||
resultsPath: completedQuery.outputDir.bqrsPath,
|
||||
interpretedResultsPath,
|
||||
},
|
||||
sourceInfo,
|
||||
["--sarif-add-snippets"],
|
||||
);
|
||||
|
||||
return sarif;
|
||||
}
|
||||
41
extensions/ql-vscode/src/model-editor/auto-model.ts
Normal file
41
extensions/ql-vscode/src/model-editor/auto-model.ts
Normal file
@@ -0,0 +1,41 @@
|
||||
import type { ModelRequest } from "./auto-model-api";
|
||||
import { AutomodelMode } from "./auto-model-api";
|
||||
import { Mode } from "./shared/mode";
|
||||
import type { AutoModelQueriesResult } from "./auto-model-codeml-queries";
|
||||
import { assertNever } from "../common/helpers-pure";
|
||||
import type { Log } from "sarif";
|
||||
import { gzipEncode } from "../common/zlib";
|
||||
|
||||
/**
|
||||
* Encode a SARIF log to the format expected by the server: JSON, GZIP-compressed, base64-encoded
|
||||
* @param log SARIF log to encode
|
||||
* @returns base64-encoded GZIP-compressed SARIF log
|
||||
*/
|
||||
export async function encodeSarif(log: Log): Promise<string> {
|
||||
const json = JSON.stringify(log);
|
||||
const buffer = Buffer.from(json, "utf-8");
|
||||
const compressed = await gzipEncode(buffer);
|
||||
return compressed.toString("base64");
|
||||
}
|
||||
|
||||
export async function createAutoModelRequest(
|
||||
mode: Mode,
|
||||
result: AutoModelQueriesResult,
|
||||
): Promise<ModelRequest> {
|
||||
let requestMode: AutomodelMode;
|
||||
switch (mode) {
|
||||
case Mode.Application:
|
||||
requestMode = AutomodelMode.Application;
|
||||
break;
|
||||
case Mode.Framework:
|
||||
requestMode = AutomodelMode.Framework;
|
||||
break;
|
||||
default:
|
||||
assertNever(mode);
|
||||
}
|
||||
|
||||
return {
|
||||
mode: requestMode,
|
||||
candidates: await encodeSarif(result.candidates),
|
||||
};
|
||||
}
|
||||
249
extensions/ql-vscode/src/model-editor/auto-modeler.ts
Normal file
249
extensions/ql-vscode/src/model-editor/auto-modeler.ts
Normal file
@@ -0,0 +1,249 @@
|
||||
import type { Method, MethodSignature } from "./method";
|
||||
import type { ModeledMethod } from "./modeled-method";
|
||||
import { load as loadYaml } from "js-yaml";
|
||||
import type { ProgressCallback } from "../common/vscode/progress";
|
||||
import { withProgress } from "../common/vscode/progress";
|
||||
import { createAutoModelRequest } from "./auto-model";
|
||||
import { getCandidates } from "./shared/auto-model-candidates";
|
||||
import { runAutoModelQueries } from "./auto-model-codeml-queries";
|
||||
import { loadDataExtensionYaml } from "./yaml";
|
||||
import type { ModelRequest, ModelResponse } from "./auto-model-api";
|
||||
import { autoModel } from "./auto-model-api";
|
||||
import { RequestError } from "@octokit/request-error";
|
||||
import { showAndLogExceptionWithTelemetry } from "../common/logging";
|
||||
import { redactableError } from "../common/errors";
|
||||
import type { App } from "../common/app";
|
||||
import type { CodeQLCliServer } from "../codeql-cli/cli";
|
||||
import type { QueryRunner } from "../query-server";
|
||||
import type { DatabaseItem } from "../databases/local-databases";
|
||||
import type { Mode } from "./shared/mode";
|
||||
import { CancellationTokenSource } from "vscode";
|
||||
import type { ModelingStore } from "./modeling-store";
|
||||
import type { ModelConfigListener } from "../config";
|
||||
import type { QueryLanguage } from "../common/query-language";
|
||||
|
||||
/**
|
||||
* The auto-modeler holds state around auto-modeling jobs and allows
|
||||
* starting and stopping them.
|
||||
*/
|
||||
export class AutoModeler {
|
||||
// Keep track of auto-modeling jobs that are in progress
|
||||
// so that we can stop them.
|
||||
private readonly jobs: Map<string, CancellationTokenSource>;
|
||||
|
||||
constructor(
|
||||
private readonly app: App,
|
||||
private readonly cliServer: CodeQLCliServer,
|
||||
private readonly queryRunner: QueryRunner,
|
||||
private readonly modelConfig: ModelConfigListener,
|
||||
private readonly modelingStore: ModelingStore,
|
||||
private readonly queryStorageDir: string,
|
||||
private readonly databaseItem: DatabaseItem,
|
||||
private readonly language: QueryLanguage,
|
||||
private readonly addModeledMethods: (
|
||||
modeledMethods: Record<string, ModeledMethod[]>,
|
||||
) => Promise<void>,
|
||||
) {
|
||||
this.jobs = new Map<string, CancellationTokenSource>();
|
||||
}
|
||||
|
||||
/**
|
||||
* Models the given package's external API usages, except
|
||||
* the ones that are already modeled.
|
||||
* @param packageName The name of the package to model.
|
||||
* @param methods The methods.
|
||||
* @param modeledMethods The currently modeled methods.
|
||||
* @param mode The mode we are modeling in.
|
||||
*/
|
||||
public async startModeling(
|
||||
packageName: string,
|
||||
methods: readonly Method[],
|
||||
modeledMethods: Record<string, readonly ModeledMethod[]>,
|
||||
processedByAutoModelMethods: Set<string>,
|
||||
mode: Mode,
|
||||
): Promise<void> {
|
||||
if (this.jobs.has(packageName)) {
|
||||
return;
|
||||
}
|
||||
|
||||
const cancellationTokenSource = new CancellationTokenSource();
|
||||
this.jobs.set(packageName, cancellationTokenSource);
|
||||
|
||||
try {
|
||||
await this.modelPackage(
|
||||
packageName,
|
||||
methods,
|
||||
modeledMethods,
|
||||
processedByAutoModelMethods,
|
||||
mode,
|
||||
cancellationTokenSource,
|
||||
);
|
||||
} finally {
|
||||
this.jobs.delete(packageName);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Stops modeling the given package.
|
||||
* @param packageName The name of the package to stop modeling.
|
||||
*/
|
||||
public async stopModeling(packageName: string): Promise<void> {
|
||||
void this.app.logger.log(`Stopping modeling for package ${packageName}`);
|
||||
const cancellationTokenSource = this.jobs.get(packageName);
|
||||
if (cancellationTokenSource) {
|
||||
cancellationTokenSource.cancel();
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* Stops all in-progress modeling jobs.
|
||||
*/
|
||||
public async stopAllModeling(): Promise<void> {
|
||||
for (const cancellationTokenSource of this.jobs.values()) {
|
||||
cancellationTokenSource.cancel();
|
||||
}
|
||||
}
|
||||
|
||||
private async modelPackage(
|
||||
packageName: string,
|
||||
methods: readonly Method[],
|
||||
modeledMethods: Record<string, readonly ModeledMethod[]>,
|
||||
processedByAutoModelMethods: Set<string>,
|
||||
mode: Mode,
|
||||
cancellationTokenSource: CancellationTokenSource,
|
||||
): Promise<void> {
|
||||
void this.app.logger.log(`Modeling package ${packageName}`);
|
||||
|
||||
const candidateBatchSize = this.modelConfig.llmGenerationBatchSize;
|
||||
|
||||
await withProgress(async (progress) => {
|
||||
// Fetch the candidates to send to the model
|
||||
const allCandidateMethods = getCandidates(
|
||||
mode,
|
||||
methods,
|
||||
modeledMethods,
|
||||
processedByAutoModelMethods,
|
||||
);
|
||||
|
||||
// If there are no candidates, there is nothing to model and we just return
|
||||
if (allCandidateMethods.length === 0) {
|
||||
void this.app.logger.log("No candidates to model. Stopping.");
|
||||
return;
|
||||
}
|
||||
|
||||
// Find number of slices to make
|
||||
const batchNumber = Math.ceil(
|
||||
allCandidateMethods.length / candidateBatchSize,
|
||||
);
|
||||
try {
|
||||
for (let i = 0; i < batchNumber; i++) {
|
||||
// Check if we should stop
|
||||
if (cancellationTokenSource.token.isCancellationRequested) {
|
||||
break;
|
||||
}
|
||||
|
||||
const start = i * candidateBatchSize;
|
||||
const end = start + candidateBatchSize;
|
||||
const candidatesToProcess = allCandidateMethods.slice(start, end);
|
||||
const candidateSignatures = candidatesToProcess.map(
|
||||
(c) => c.signature,
|
||||
);
|
||||
|
||||
// Let the UI know which candidates we are modeling
|
||||
this.modelingStore.addInProgressMethods(
|
||||
this.databaseItem,
|
||||
candidateSignatures,
|
||||
);
|
||||
|
||||
// Kick off the process to model the slice of candidates
|
||||
await this.modelCandidates(
|
||||
candidatesToProcess,
|
||||
mode,
|
||||
progress,
|
||||
cancellationTokenSource,
|
||||
);
|
||||
|
||||
// Let the UI know which candidates we are done modeling
|
||||
this.modelingStore.removeInProgressMethods(
|
||||
this.databaseItem,
|
||||
candidateSignatures,
|
||||
);
|
||||
|
||||
// Let the UI know which methods have been sent to the LLM
|
||||
this.modelingStore.addProcessedByAutoModelMethods(
|
||||
this.databaseItem,
|
||||
candidateSignatures,
|
||||
);
|
||||
}
|
||||
} finally {
|
||||
// Clear out in progress methods in case anything went wrong
|
||||
this.modelingStore.removeInProgressMethods(
|
||||
this.databaseItem,
|
||||
allCandidateMethods.map((c) => c.signature),
|
||||
);
|
||||
}
|
||||
});
|
||||
}
|
||||
|
||||
private async modelCandidates(
|
||||
candidateMethods: MethodSignature[],
|
||||
mode: Mode,
|
||||
progress: ProgressCallback,
|
||||
cancellationTokenSource: CancellationTokenSource,
|
||||
): Promise<void> {
|
||||
void this.app.logger.log("Executing auto-model queries");
|
||||
|
||||
const usages = await runAutoModelQueries({
|
||||
mode,
|
||||
candidateMethods,
|
||||
cliServer: this.cliServer,
|
||||
queryRunner: this.queryRunner,
|
||||
queryStorageDir: this.queryStorageDir,
|
||||
databaseItem: this.databaseItem,
|
||||
progress: (update) => progress({ ...update }),
|
||||
cancellationTokenSource,
|
||||
});
|
||||
if (!usages) {
|
||||
return;
|
||||
}
|
||||
|
||||
const request = await createAutoModelRequest(mode, usages);
|
||||
|
||||
void this.app.logger.log("Calling auto-model API");
|
||||
|
||||
const response = await this.callAutoModelApi(request);
|
||||
if (!response) {
|
||||
return;
|
||||
}
|
||||
|
||||
const models = loadYaml(response.models, {
|
||||
filename: "auto-model.yml",
|
||||
});
|
||||
|
||||
const loadedMethods = loadDataExtensionYaml(models, this.language);
|
||||
if (!loadedMethods) {
|
||||
return;
|
||||
}
|
||||
|
||||
await this.addModeledMethods(loadedMethods);
|
||||
}
|
||||
|
||||
private async callAutoModelApi(
|
||||
request: ModelRequest,
|
||||
): Promise<ModelResponse | null> {
|
||||
try {
|
||||
return await autoModel(this.app.credentials, request, this.modelConfig);
|
||||
} catch (e) {
|
||||
if (e instanceof RequestError && e.status === 429) {
|
||||
void showAndLogExceptionWithTelemetry(
|
||||
this.app.logger,
|
||||
this.app.telemetry,
|
||||
redactableError`Rate limit hit, please try again soon.`,
|
||||
);
|
||||
return null;
|
||||
} else {
|
||||
throw e;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -33,7 +33,6 @@ export function decodeBqrsToMethods(
|
||||
let libraryVersion: string | undefined;
|
||||
let type: ModeledMethodType;
|
||||
let classification: CallClassification;
|
||||
let endpointKindColumn: string | BqrsEntityValue | undefined;
|
||||
let endpointType: EndpointType | undefined = undefined;
|
||||
|
||||
if (mode === Mode.Application) {
|
||||
@@ -48,7 +47,6 @@ export function decodeBqrsToMethods(
|
||||
libraryVersion,
|
||||
type,
|
||||
classification,
|
||||
endpointKindColumn,
|
||||
] = tuple as ApplicationModeTuple;
|
||||
} else {
|
||||
[
|
||||
@@ -60,7 +58,6 @@ export function decodeBqrsToMethods(
|
||||
supported,
|
||||
library,
|
||||
type,
|
||||
endpointKindColumn,
|
||||
] = tuple as FrameworkModeTuple;
|
||||
|
||||
classification = CallClassification.Unknown;
|
||||
@@ -71,18 +68,13 @@ export function decodeBqrsToMethods(
|
||||
}
|
||||
|
||||
if (definition.endpointTypeForEndpoint) {
|
||||
endpointType = definition.endpointTypeForEndpoint(
|
||||
{
|
||||
endpointType,
|
||||
packageName,
|
||||
typeName,
|
||||
methodName,
|
||||
methodParameters,
|
||||
},
|
||||
typeof endpointKindColumn === "object"
|
||||
? endpointKindColumn.label
|
||||
: endpointKindColumn,
|
||||
);
|
||||
endpointType = definition.endpointTypeForEndpoint({
|
||||
endpointType,
|
||||
packageName,
|
||||
typeName,
|
||||
methodName,
|
||||
methodParameters,
|
||||
});
|
||||
}
|
||||
|
||||
if (endpointType === undefined) {
|
||||
|
||||
@@ -133,7 +133,7 @@ async function findGitFolder(
|
||||
const stat = await workspace.fs.stat(gitFolder);
|
||||
// Check whether it's a directory
|
||||
return (stat.type & FileType.Directory) !== 0;
|
||||
} catch {
|
||||
} catch (e) {
|
||||
return false;
|
||||
}
|
||||
}),
|
||||
|
||||
@@ -174,14 +174,11 @@ export type ModelsAsDataLanguage = {
|
||||
* be determined by heuristics.
|
||||
* @param method The method to get the endpoint type for. The endpoint type can be undefined if the
|
||||
* query does not return an endpoint type.
|
||||
* @param endpointKind An optional column that may be provided by the query to help determine the
|
||||
* endpoint type.
|
||||
*/
|
||||
endpointTypeForEndpoint?: (
|
||||
method: Omit<MethodDefinition, "endpointType"> & {
|
||||
endpointType: EndpointType | undefined;
|
||||
},
|
||||
endpointKind: string | undefined,
|
||||
) => EndpointType | undefined;
|
||||
predicates: ModelsAsDataLanguagePredicates;
|
||||
modelGeneration?: ModelsAsDataLanguageModelGeneration;
|
||||
|
||||
@@ -4,26 +4,7 @@ import { EndpointType } from "../../method";
|
||||
|
||||
const memberTokenRegex = /^Member\[(.+)]$/;
|
||||
|
||||
// In Python, the type can contain both the package name and the type name.
|
||||
export function parsePythonType(type: string) {
|
||||
// The first part is always the package name. All remaining parts are the type
|
||||
// name.
|
||||
|
||||
const parts = type.split(".");
|
||||
const packageName = parts.shift() ?? "";
|
||||
|
||||
return {
|
||||
packageName,
|
||||
typeName: parts.join("."),
|
||||
};
|
||||
}
|
||||
|
||||
// The type name can also be specified in the type, so this will combine
|
||||
// the already parsed type name and the type name from the access path.
|
||||
export function parsePythonAccessPath(
|
||||
path: string,
|
||||
shortTypeName: string,
|
||||
): {
|
||||
export function parsePythonAccessPath(path: string): {
|
||||
typeName: string;
|
||||
methodName: string;
|
||||
endpointType: EndpointType;
|
||||
@@ -32,12 +13,8 @@ export function parsePythonAccessPath(
|
||||
const tokens = parseAccessPathTokens(path);
|
||||
|
||||
if (tokens.length === 0) {
|
||||
const typeName = shortTypeName.endsWith("!")
|
||||
? shortTypeName.slice(0, -1)
|
||||
: shortTypeName;
|
||||
|
||||
return {
|
||||
typeName,
|
||||
typeName: "",
|
||||
methodName: "",
|
||||
endpointType: EndpointType.Method,
|
||||
path: "",
|
||||
@@ -46,10 +23,6 @@ export function parsePythonAccessPath(
|
||||
|
||||
const typeParts = [];
|
||||
let endpointType = EndpointType.Function;
|
||||
// If a short type name was given and it doesn't end in a `!`, then this refers to a method.
|
||||
if (shortTypeName !== "" && !shortTypeName.endsWith("!")) {
|
||||
endpointType = EndpointType.Method;
|
||||
}
|
||||
|
||||
let remainingTokens: typeof tokens = [];
|
||||
|
||||
@@ -59,7 +32,6 @@ export function parsePythonAccessPath(
|
||||
if (memberMatch) {
|
||||
typeParts.push(memberMatch[1]);
|
||||
} else if (token.text === "Instance") {
|
||||
// Alternative way of specifying that this refers to a method.
|
||||
endpointType = EndpointType.Method;
|
||||
} else {
|
||||
remainingTokens = tokens.slice(i);
|
||||
@@ -68,22 +40,9 @@ export function parsePythonAccessPath(
|
||||
}
|
||||
|
||||
const methodName = typeParts.pop() ?? "";
|
||||
let typeName = typeParts.join(".");
|
||||
const typeName = typeParts.join(".");
|
||||
const remainingPath = remainingTokens.map((token) => token.text).join(".");
|
||||
|
||||
if (shortTypeName !== "") {
|
||||
if (shortTypeName.endsWith("!")) {
|
||||
// The actual type name is the name without the `!`.
|
||||
shortTypeName = shortTypeName.slice(0, -1);
|
||||
}
|
||||
|
||||
if (typeName !== "") {
|
||||
typeName = `${shortTypeName}.${typeName}`;
|
||||
} else {
|
||||
typeName = shortTypeName;
|
||||
}
|
||||
}
|
||||
|
||||
return {
|
||||
methodName,
|
||||
typeName,
|
||||
@@ -92,59 +51,53 @@ export function parsePythonAccessPath(
|
||||
};
|
||||
}
|
||||
|
||||
export function parsePythonTypeAndPath(
|
||||
type: string,
|
||||
path: string,
|
||||
): {
|
||||
packageName: string;
|
||||
typeName: string;
|
||||
methodName: string;
|
||||
endpointType: EndpointType;
|
||||
path: string;
|
||||
} {
|
||||
const { packageName, typeName: shortTypeName } = parsePythonType(type);
|
||||
const {
|
||||
typeName,
|
||||
methodName,
|
||||
endpointType,
|
||||
path: remainingPath,
|
||||
} = parsePythonAccessPath(path, shortTypeName);
|
||||
|
||||
return {
|
||||
packageName,
|
||||
typeName,
|
||||
methodName,
|
||||
endpointType,
|
||||
path: remainingPath,
|
||||
};
|
||||
}
|
||||
|
||||
export function pythonMethodSignature(typeName: string, methodName: string) {
|
||||
return `${typeName}#${methodName}`;
|
||||
}
|
||||
|
||||
export function pythonType(
|
||||
packageName: string,
|
||||
typeName: string,
|
||||
endpointType: EndpointType,
|
||||
) {
|
||||
if (typeName !== "" && packageName !== "") {
|
||||
return `${packageName}.${typeName}${endpointType === EndpointType.Function ? "!" : ""}`;
|
||||
}
|
||||
|
||||
return `${packageName}${typeName}`;
|
||||
}
|
||||
|
||||
export function pythonMethodPath(methodName: string) {
|
||||
if (methodName === "") {
|
||||
function pythonTypePath(typeName: string) {
|
||||
if (typeName === "") {
|
||||
return "";
|
||||
}
|
||||
|
||||
return `Member[${methodName}]`;
|
||||
return typeName
|
||||
.split(".")
|
||||
.map((part) => `Member[${part}]`)
|
||||
.join(".");
|
||||
}
|
||||
|
||||
export function pythonPath(methodName: string, path: string) {
|
||||
const methodPath = pythonMethodPath(methodName);
|
||||
export function pythonMethodPath(
|
||||
typeName: string,
|
||||
methodName: string,
|
||||
endpointType: EndpointType,
|
||||
) {
|
||||
if (methodName === "") {
|
||||
return pythonTypePath(typeName);
|
||||
}
|
||||
|
||||
const typePath = pythonTypePath(typeName);
|
||||
|
||||
let result = typePath;
|
||||
if (typePath !== "" && endpointType === EndpointType.Method) {
|
||||
result += ".Instance";
|
||||
}
|
||||
|
||||
if (result !== "") {
|
||||
result += ".";
|
||||
}
|
||||
|
||||
result += `Member[${methodName}]`;
|
||||
|
||||
return result;
|
||||
}
|
||||
|
||||
export function pythonPath(
|
||||
typeName: string,
|
||||
methodName: string,
|
||||
endpointType: EndpointType,
|
||||
path: string,
|
||||
) {
|
||||
const methodPath = pythonMethodPath(typeName, methodName, endpointType);
|
||||
if (methodPath === "") {
|
||||
return path;
|
||||
}
|
||||
@@ -158,24 +111,7 @@ export function pythonPath(methodName: string, path: string) {
|
||||
|
||||
export function pythonEndpointType(
|
||||
method: Omit<MethodDefinition, "endpointType">,
|
||||
endpointKind: string | undefined,
|
||||
): EndpointType {
|
||||
switch (endpointKind) {
|
||||
case "Function":
|
||||
return EndpointType.Function;
|
||||
case "InstanceMethod":
|
||||
return EndpointType.Method;
|
||||
case "ClassMethod":
|
||||
return EndpointType.ClassMethod;
|
||||
case "StaticMethod":
|
||||
return EndpointType.StaticMethod;
|
||||
case "InitMethod":
|
||||
return EndpointType.Constructor;
|
||||
case "Class":
|
||||
return EndpointType.Class;
|
||||
}
|
||||
|
||||
// Legacy behavior for when the kind column is missing.
|
||||
if (
|
||||
method.methodParameters.startsWith("(self,") ||
|
||||
method.methodParameters === "(self)"
|
||||
@@ -184,12 +120,3 @@ export function pythonEndpointType(
|
||||
}
|
||||
return EndpointType.Function;
|
||||
}
|
||||
|
||||
export function hasPythonSelfArgument(endpointType: EndpointType): boolean {
|
||||
// Instance methods and class methods both use `Argument[self]` for the first parameter. The first
|
||||
// parameter after self is called `Argument[0]`.
|
||||
return (
|
||||
endpointType === EndpointType.Method ||
|
||||
endpointType === EndpointType.ClassMethod
|
||||
);
|
||||
}
|
||||
|
||||
@@ -4,48 +4,44 @@ import { Mode } from "../../shared/mode";
|
||||
import type { MethodArgument } from "../../method";
|
||||
import { EndpointType, getArgumentsList } from "../../method";
|
||||
import {
|
||||
hasPythonSelfArgument,
|
||||
parsePythonTypeAndPath,
|
||||
parsePythonAccessPath,
|
||||
pythonEndpointType,
|
||||
pythonMethodPath,
|
||||
pythonMethodSignature,
|
||||
pythonPath,
|
||||
pythonType,
|
||||
} from "./access-paths";
|
||||
|
||||
export const python: ModelsAsDataLanguage = {
|
||||
availableModes: [Mode.Framework],
|
||||
createMethodSignature: ({ typeName, methodName }) =>
|
||||
`${typeName}#${methodName}`,
|
||||
endpointTypeForEndpoint: (method, endpointKind) =>
|
||||
pythonEndpointType(method, endpointKind),
|
||||
endpointTypeForEndpoint: (method) => pythonEndpointType(method),
|
||||
predicates: {
|
||||
source: {
|
||||
extensiblePredicate: sharedExtensiblePredicates.source,
|
||||
supportedKinds: sharedKinds.source,
|
||||
supportedEndpointTypes: [
|
||||
EndpointType.Method,
|
||||
EndpointType.Function,
|
||||
EndpointType.Constructor,
|
||||
EndpointType.ClassMethod,
|
||||
EndpointType.StaticMethod,
|
||||
],
|
||||
supportedEndpointTypes: [EndpointType.Method, EndpointType.Function],
|
||||
// extensible predicate sourceModel(
|
||||
// string type, string path, string kind
|
||||
// );
|
||||
generateMethodDefinition: (method) => [
|
||||
pythonType(method.packageName, method.typeName, method.endpointType),
|
||||
pythonPath(method.methodName, method.output),
|
||||
method.packageName,
|
||||
pythonPath(
|
||||
method.typeName,
|
||||
method.methodName,
|
||||
method.endpointType,
|
||||
method.output,
|
||||
),
|
||||
method.kind,
|
||||
],
|
||||
readModeledMethod: (row) => {
|
||||
const packageName = row[0] as string;
|
||||
const {
|
||||
packageName,
|
||||
typeName,
|
||||
methodName,
|
||||
endpointType,
|
||||
path: output,
|
||||
} = parsePythonTypeAndPath(row[0] as string, row[1] as string);
|
||||
} = parsePythonAccessPath(row[1] as string);
|
||||
return {
|
||||
type: "source",
|
||||
output,
|
||||
@@ -63,31 +59,30 @@ export const python: ModelsAsDataLanguage = {
|
||||
sink: {
|
||||
extensiblePredicate: sharedExtensiblePredicates.sink,
|
||||
supportedKinds: sharedKinds.sink,
|
||||
supportedEndpointTypes: [
|
||||
EndpointType.Method,
|
||||
EndpointType.Function,
|
||||
EndpointType.Constructor,
|
||||
EndpointType.ClassMethod,
|
||||
EndpointType.StaticMethod,
|
||||
],
|
||||
supportedEndpointTypes: [EndpointType.Method, EndpointType.Function],
|
||||
// extensible predicate sinkModel(
|
||||
// string type, string path, string kind
|
||||
// );
|
||||
generateMethodDefinition: (method) => {
|
||||
return [
|
||||
pythonType(method.packageName, method.typeName, method.endpointType),
|
||||
pythonPath(method.methodName, method.input),
|
||||
method.packageName,
|
||||
pythonPath(
|
||||
method.typeName,
|
||||
method.methodName,
|
||||
method.endpointType,
|
||||
method.input,
|
||||
),
|
||||
method.kind,
|
||||
];
|
||||
},
|
||||
readModeledMethod: (row) => {
|
||||
const packageName = row[0] as string;
|
||||
const {
|
||||
packageName,
|
||||
typeName,
|
||||
methodName,
|
||||
endpointType,
|
||||
path: input,
|
||||
} = parsePythonTypeAndPath(row[0] as string, row[1] as string);
|
||||
} = parsePythonAccessPath(row[1] as string);
|
||||
return {
|
||||
type: "sink",
|
||||
input,
|
||||
@@ -105,26 +100,25 @@ export const python: ModelsAsDataLanguage = {
|
||||
summary: {
|
||||
extensiblePredicate: sharedExtensiblePredicates.summary,
|
||||
supportedKinds: sharedKinds.summary,
|
||||
supportedEndpointTypes: [
|
||||
EndpointType.Method,
|
||||
EndpointType.Function,
|
||||
EndpointType.Constructor,
|
||||
EndpointType.ClassMethod,
|
||||
EndpointType.StaticMethod,
|
||||
],
|
||||
supportedEndpointTypes: [EndpointType.Method, EndpointType.Function],
|
||||
// extensible predicate summaryModel(
|
||||
// string type, string path, string input, string output, string kind
|
||||
// );
|
||||
generateMethodDefinition: (method) => [
|
||||
pythonType(method.packageName, method.typeName, method.endpointType),
|
||||
pythonMethodPath(method.methodName),
|
||||
method.packageName,
|
||||
pythonMethodPath(
|
||||
method.typeName,
|
||||
method.methodName,
|
||||
method.endpointType,
|
||||
),
|
||||
method.input,
|
||||
method.output,
|
||||
method.kind,
|
||||
],
|
||||
readModeledMethod: (row) => {
|
||||
const { packageName, typeName, methodName, endpointType, path } =
|
||||
parsePythonTypeAndPath(row[0] as string, row[1] as string);
|
||||
const packageName = row[0] as string;
|
||||
const { typeName, methodName, endpointType, path } =
|
||||
parsePythonAccessPath(row[1] as string);
|
||||
if (path !== "") {
|
||||
throw new Error("Summary path must be a method");
|
||||
}
|
||||
@@ -150,13 +144,18 @@ export const python: ModelsAsDataLanguage = {
|
||||
// string type, string path, string kind
|
||||
// );
|
||||
generateMethodDefinition: (method) => [
|
||||
pythonType(method.packageName, method.typeName, method.endpointType),
|
||||
pythonMethodPath(method.methodName),
|
||||
method.packageName,
|
||||
pythonMethodPath(
|
||||
method.typeName,
|
||||
method.methodName,
|
||||
method.endpointType,
|
||||
),
|
||||
method.kind,
|
||||
],
|
||||
readModeledMethod: (row) => {
|
||||
const { packageName, typeName, methodName, endpointType, path } =
|
||||
parsePythonTypeAndPath(row[0] as string, row[1] as string);
|
||||
const packageName = row[0] as string;
|
||||
const { typeName, methodName, endpointType, path } =
|
||||
parsePythonAccessPath(row[1] as string);
|
||||
if (path !== "") {
|
||||
throw new Error("Neutral path must be a method");
|
||||
}
|
||||
@@ -173,46 +172,25 @@ export const python: ModelsAsDataLanguage = {
|
||||
};
|
||||
},
|
||||
},
|
||||
type: {
|
||||
extensiblePredicate: "typeModel",
|
||||
// extensible predicate typeModel(string type1, string type2, string path);
|
||||
generateMethodDefinition: (method) => [
|
||||
method.relatedTypeName,
|
||||
pythonType(method.packageName, method.typeName, method.endpointType),
|
||||
pythonPath(method.methodName, method.path),
|
||||
],
|
||||
readModeledMethod: (row) => {
|
||||
const { packageName, typeName, methodName, endpointType, path } =
|
||||
parsePythonTypeAndPath(row[1] as string, row[2] as string);
|
||||
|
||||
return {
|
||||
type: "type",
|
||||
relatedTypeName: row[0] as string,
|
||||
path,
|
||||
signature: pythonMethodSignature(typeName, methodName),
|
||||
endpointType,
|
||||
packageName,
|
||||
typeName,
|
||||
methodName,
|
||||
methodParameters: "",
|
||||
};
|
||||
},
|
||||
},
|
||||
},
|
||||
getArgumentOptions: (method) => {
|
||||
// Argument and Parameter are equivalent in Python, but we'll use Argument in the model editor
|
||||
const argumentsList = getArgumentsList(method.methodParameters).map(
|
||||
(argument, index): MethodArgument => {
|
||||
if (hasPythonSelfArgument(method.endpointType) && index === 0) {
|
||||
if (
|
||||
method.endpointType === EndpointType.Method &&
|
||||
argument === "self" &&
|
||||
index === 0
|
||||
) {
|
||||
return {
|
||||
path: "Argument[self]",
|
||||
label: `Argument[self]: ${argument}`,
|
||||
label: "Argument[self]: self",
|
||||
};
|
||||
}
|
||||
|
||||
// If this endpoint has a self argument, self does not count as an argument index so we
|
||||
// If this is a method, self does not count as an argument index, so we
|
||||
// should start at 0 for the second argument
|
||||
if (hasPythonSelfArgument(method.endpointType)) {
|
||||
if (method.endpointType === EndpointType.Method) {
|
||||
index -= 1;
|
||||
}
|
||||
|
||||
|
||||
@@ -70,6 +70,8 @@ export class MethodModelingViewProvider extends AbstractWebviewViewProvider<
|
||||
method: Method,
|
||||
modeledMethods: readonly ModeledMethod[],
|
||||
isModified: boolean,
|
||||
isInProgress: boolean,
|
||||
processedByAutoModel: boolean,
|
||||
): Promise<void> {
|
||||
this.method = method;
|
||||
this.databaseItem = databaseItem;
|
||||
@@ -80,6 +82,8 @@ export class MethodModelingViewProvider extends AbstractWebviewViewProvider<
|
||||
method,
|
||||
modeledMethods,
|
||||
isModified,
|
||||
isInProgress,
|
||||
processedByAutoModel,
|
||||
});
|
||||
}
|
||||
|
||||
@@ -100,6 +104,8 @@ export class MethodModelingViewProvider extends AbstractWebviewViewProvider<
|
||||
selectedMethod.method,
|
||||
selectedMethod.modeledMethods,
|
||||
selectedMethod.isModified,
|
||||
selectedMethod.isInProgress,
|
||||
selectedMethod.processedByAutoModel,
|
||||
);
|
||||
}
|
||||
}
|
||||
@@ -197,6 +203,8 @@ export class MethodModelingViewProvider extends AbstractWebviewViewProvider<
|
||||
e.method,
|
||||
e.modeledMethods,
|
||||
e.isModified,
|
||||
e.isInProgress,
|
||||
e.processedByAutoModel,
|
||||
);
|
||||
}
|
||||
}),
|
||||
@@ -224,6 +232,36 @@ export class MethodModelingViewProvider extends AbstractWebviewViewProvider<
|
||||
}
|
||||
}),
|
||||
);
|
||||
|
||||
this.push(
|
||||
this.modelingEvents.onInProgressMethodsChanged(async (e) => {
|
||||
if (this.method && this.databaseItem) {
|
||||
const dbUri = this.databaseItem.databaseUri.toString();
|
||||
if (e.dbUri === dbUri) {
|
||||
const inProgress = e.methods.has(this.method.signature);
|
||||
await this.postMessage({
|
||||
t: "setInProgress",
|
||||
inProgress,
|
||||
});
|
||||
}
|
||||
}
|
||||
}),
|
||||
);
|
||||
|
||||
this.push(
|
||||
this.modelingEvents.onProcessedByAutoModelMethodsChanged(async (e) => {
|
||||
if (this.method && this.databaseItem) {
|
||||
const dbUri = this.databaseItem.databaseUri.toString();
|
||||
if (e.dbUri === dbUri) {
|
||||
const processedByAutoModel = e.methods.has(this.method.signature);
|
||||
await this.postMessage({
|
||||
t: "setProcessedByAutoModel",
|
||||
processedByAutoModel,
|
||||
});
|
||||
}
|
||||
}
|
||||
}),
|
||||
);
|
||||
}
|
||||
|
||||
private registerToModelConfigEvents(): void {
|
||||
|
||||
@@ -29,8 +29,6 @@ export enum EndpointType {
|
||||
Method = "method",
|
||||
Constructor = "constructor",
|
||||
Function = "function",
|
||||
StaticMethod = "staticMethod",
|
||||
ClassMethod = "classMethod",
|
||||
}
|
||||
|
||||
export interface MethodDefinition {
|
||||
|
||||
@@ -204,6 +204,7 @@ export class ModelEditorModule extends DisposableObject {
|
||||
this.app.logger,
|
||||
queryDir,
|
||||
language,
|
||||
this.modelConfig,
|
||||
initialMode,
|
||||
);
|
||||
if (!success) {
|
||||
|
||||
@@ -8,6 +8,7 @@ import {
|
||||
syntheticQueryPackName,
|
||||
} from "./model-editor-queries";
|
||||
import type { CodeQLCliServer } from "../codeql-cli/cli";
|
||||
import type { ModelConfig } from "../config";
|
||||
import type { Mode } from "./shared/mode";
|
||||
import type { NotificationLogger } from "../common/logging";
|
||||
|
||||
@@ -29,6 +30,7 @@ import type { NotificationLogger } from "../common/logging";
|
||||
* @param logger The logger to use.
|
||||
* @param queryDir The directory to set up.
|
||||
* @param language The language to use for the queries.
|
||||
* @param modelConfig The model config to use.
|
||||
* @param initialMode The initial mode to use to check the existence of the queries.
|
||||
* @returns true if the setup was successful, false otherwise.
|
||||
*/
|
||||
@@ -37,6 +39,7 @@ export async function setUpPack(
|
||||
logger: NotificationLogger,
|
||||
queryDir: string,
|
||||
language: QueryLanguage,
|
||||
modelConfig: ModelConfig,
|
||||
initialMode: Mode,
|
||||
): Promise<boolean> {
|
||||
// Download the required query packs
|
||||
@@ -88,5 +91,10 @@ export async function setUpPack(
|
||||
await cliServer.packInstall(queryDir);
|
||||
}
|
||||
|
||||
// Download any other required packs
|
||||
if (language === "java" && modelConfig.llmGeneration) {
|
||||
await cliServer.packDownload([`codeql/${language}-automodel-queries`]);
|
||||
}
|
||||
|
||||
return true;
|
||||
}
|
||||
|
||||
@@ -49,6 +49,7 @@ import {
|
||||
import { pickExtensionPack } from "./extension-pack-picker";
|
||||
import type { QueryLanguage } from "../common/query-language";
|
||||
import { getLanguageDisplayName } from "../common/query-language";
|
||||
import { AutoModeler } from "./auto-modeler";
|
||||
import { telemetryListener } from "../common/vscode/telemetry";
|
||||
import type { ModelingStore } from "./modeling-store";
|
||||
import type { ModelingEvents } from "./modeling-events";
|
||||
@@ -76,6 +77,7 @@ export class ModelEditorView extends AbstractWebview<
|
||||
ToModelEditorMessage,
|
||||
FromModelEditorMessage
|
||||
> {
|
||||
private readonly autoModeler: AutoModeler;
|
||||
private readonly modelEvaluator: ModelEvaluator;
|
||||
private readonly languageDefinition: ModelsAsDataLanguage;
|
||||
// Cancellation token source that can be used for passing into long-running operations. Should only
|
||||
@@ -112,6 +114,19 @@ export class ModelEditorView extends AbstractWebview<
|
||||
this.registerToModelingEvents();
|
||||
this.registerToModelConfigEvents();
|
||||
|
||||
this.autoModeler = new AutoModeler(
|
||||
app,
|
||||
cliServer,
|
||||
queryRunner,
|
||||
this.modelConfig,
|
||||
modelingStore,
|
||||
queryStorageDir,
|
||||
databaseItem,
|
||||
language,
|
||||
async (modeledMethods) => {
|
||||
this.addModeledMethods(modeledMethods);
|
||||
},
|
||||
);
|
||||
this.languageDefinition = getModelsAsDataLanguage(language);
|
||||
|
||||
this.modelEvaluator = new ModelEvaluator(
|
||||
@@ -302,6 +317,21 @@ export class ModelEditorView extends AbstractWebview<
|
||||
"model-editor-generate-modeled-methods",
|
||||
);
|
||||
|
||||
break;
|
||||
case "generateMethodsFromLlm":
|
||||
await this.generateModeledMethodsFromLlm(
|
||||
msg.packageName,
|
||||
msg.methodSignatures,
|
||||
);
|
||||
void telemetryListener?.sendUIInteraction(
|
||||
"model-editor-generate-methods-from-llm",
|
||||
);
|
||||
break;
|
||||
case "stopGeneratingMethodsFromLlm":
|
||||
await this.autoModeler.stopModeling(msg.packageName);
|
||||
void telemetryListener?.sendUIInteraction(
|
||||
"model-editor-stop-generating-methods-from-llm",
|
||||
);
|
||||
break;
|
||||
case "modelDependency":
|
||||
await this.modelDependency();
|
||||
@@ -408,6 +438,9 @@ export class ModelEditorView extends AbstractWebview<
|
||||
this.modelConfig.flowGeneration) &&
|
||||
!!modelsAsDataLanguage.modelGeneration;
|
||||
|
||||
const showLlmButton =
|
||||
this.databaseItem.language === "java" && this.modelConfig.llmGeneration;
|
||||
|
||||
const showEvaluationUi = this.modelConfig.modelEvaluation;
|
||||
|
||||
const sourceArchiveAvailable =
|
||||
@@ -423,6 +456,7 @@ export class ModelEditorView extends AbstractWebview<
|
||||
extensionPack: this.extensionPack,
|
||||
language: this.language,
|
||||
showGenerateButton,
|
||||
showLlmButton,
|
||||
showEvaluationUi,
|
||||
mode: this.modelingStore.getMode(this.databaseItem),
|
||||
showModeSwitchButton,
|
||||
@@ -771,6 +805,33 @@ export class ModelEditorView extends AbstractWebview<
|
||||
);
|
||||
}
|
||||
|
||||
private async generateModeledMethodsFromLlm(
|
||||
packageName: string,
|
||||
methodSignatures: string[],
|
||||
): Promise<void> {
|
||||
const methods = this.modelingStore.getMethods(
|
||||
this.databaseItem,
|
||||
methodSignatures,
|
||||
);
|
||||
const modeledMethods = this.modelingStore.getModeledMethods(
|
||||
this.databaseItem,
|
||||
methodSignatures,
|
||||
);
|
||||
const processedByAutoModelMethods =
|
||||
this.modelingStore.getProcessedByAutoModelMethods(
|
||||
this.databaseItem,
|
||||
methodSignatures,
|
||||
);
|
||||
const mode = this.modelingStore.getMode(this.databaseItem);
|
||||
await this.autoModeler.startModeling(
|
||||
packageName,
|
||||
methods,
|
||||
modeledMethods,
|
||||
processedByAutoModelMethods,
|
||||
mode,
|
||||
);
|
||||
}
|
||||
|
||||
private async modelDependency(): Promise<void> {
|
||||
return withProgress(
|
||||
async (progress, token) => {
|
||||
@@ -922,6 +983,30 @@ export class ModelEditorView extends AbstractWebview<
|
||||
}),
|
||||
);
|
||||
|
||||
this.push(
|
||||
this.modelingEvents.onInProgressMethodsChanged(async (event) => {
|
||||
if (event.dbUri === this.databaseItem.databaseUri.toString()) {
|
||||
await this.postMessage({
|
||||
t: "setInProgressMethods",
|
||||
methods: Array.from(event.methods),
|
||||
});
|
||||
}
|
||||
}),
|
||||
);
|
||||
|
||||
this.push(
|
||||
this.modelingEvents.onProcessedByAutoModelMethodsChanged(
|
||||
async (event) => {
|
||||
if (event.dbUri === this.databaseItem.databaseUri.toString()) {
|
||||
await this.postMessage({
|
||||
t: "setProcessedByAutoModelMethods",
|
||||
methods: Array.from(event.methods),
|
||||
});
|
||||
}
|
||||
},
|
||||
),
|
||||
);
|
||||
|
||||
this.push(
|
||||
this.modelingEvents.onRevealInModelEditor(async (event) => {
|
||||
if (event.dbUri === this.databaseItem.databaseUri.toString()) {
|
||||
|
||||
@@ -1,5 +1,6 @@
|
||||
import { assertNever } from "../common/helpers-pure";
|
||||
import type { MethodSignature } from "./method";
|
||||
import type { ModelingStatus } from "./shared/modeling-status";
|
||||
|
||||
export type ModeledMethodType =
|
||||
| "none"
|
||||
@@ -14,6 +15,10 @@ export type Provenance =
|
||||
| "df-generated"
|
||||
// Generated by the dataflow model and manually edited
|
||||
| "df-manual"
|
||||
// Generated by the auto-model
|
||||
| "ai-generated"
|
||||
// Generated by the auto-model and manually edited
|
||||
| "ai-manual"
|
||||
// Entered by the user in the editor manually
|
||||
| "manual";
|
||||
|
||||
@@ -107,6 +112,30 @@ export function modeledMethodSupportsProvenance(
|
||||
);
|
||||
}
|
||||
|
||||
export function isModelPending(
|
||||
modeledMethod: ModeledMethod | undefined,
|
||||
modelingStatus: ModelingStatus,
|
||||
processedByAutoModel: boolean,
|
||||
): boolean {
|
||||
if (
|
||||
(!modeledMethod || modeledMethod.type === "none") &&
|
||||
processedByAutoModel
|
||||
) {
|
||||
return true;
|
||||
}
|
||||
|
||||
if (!modeledMethod) {
|
||||
return false;
|
||||
}
|
||||
|
||||
return (
|
||||
modelingStatus === "unsaved" &&
|
||||
modeledMethod.type !== "none" &&
|
||||
modeledMethodSupportsProvenance(modeledMethod) &&
|
||||
modeledMethod.provenance === "ai-generated"
|
||||
);
|
||||
}
|
||||
|
||||
/**
|
||||
* Calculates the new provenance for a modeled method based on the current provenance.
|
||||
* @param modeledMethod The modeled method if there is one.
|
||||
@@ -129,6 +158,13 @@ export function calculateNewProvenance(
|
||||
case "df-manual":
|
||||
// If the method has had manual edits, we want the provenance to stay the same.
|
||||
return "df-manual";
|
||||
case "ai-generated":
|
||||
// If the method has been generated and there has been a change, we assume
|
||||
// that the user has manually edited it.
|
||||
return "ai-manual";
|
||||
case "ai-manual":
|
||||
// If the method has had manual edits, we want the provenance to stay the same.
|
||||
return "ai-manual";
|
||||
default:
|
||||
// The method has been modeled manually.
|
||||
return "manual";
|
||||
|
||||
@@ -37,6 +37,18 @@ interface SelectedMethodChangedEvent {
|
||||
readonly usage: Usage;
|
||||
readonly modeledMethods: readonly ModeledMethod[];
|
||||
readonly isModified: boolean;
|
||||
readonly isInProgress: boolean;
|
||||
readonly processedByAutoModel: boolean;
|
||||
}
|
||||
|
||||
interface InProgressMethodsChangedEvent {
|
||||
readonly dbUri: string;
|
||||
readonly methods: ReadonlySet<string>;
|
||||
}
|
||||
|
||||
interface ProcessedByAutoModelMethodsChangedEvent {
|
||||
readonly dbUri: string;
|
||||
readonly methods: ReadonlySet<string>;
|
||||
}
|
||||
|
||||
interface ModelEvaluationRunChangedEvent {
|
||||
@@ -71,6 +83,8 @@ export class ModelingEvents extends DisposableObject {
|
||||
public readonly onModeChanged: AppEvent<ModeChangedEvent>;
|
||||
public readonly onModeledAndModifiedMethodsChanged: AppEvent<ModeledAndModifiedMethodsChangedEvent>;
|
||||
public readonly onSelectedMethodChanged: AppEvent<SelectedMethodChangedEvent>;
|
||||
public readonly onInProgressMethodsChanged: AppEvent<InProgressMethodsChangedEvent>;
|
||||
public readonly onProcessedByAutoModelMethodsChanged: AppEvent<ProcessedByAutoModelMethodsChangedEvent>;
|
||||
public readonly onModelEvaluationRunChanged: AppEvent<ModelEvaluationRunChangedEvent>;
|
||||
public readonly onRevealInModelEditor: AppEvent<RevealInModelEditorEvent>;
|
||||
public readonly onFocusModelEditor: AppEvent<FocusModelEditorEvent>;
|
||||
@@ -85,6 +99,8 @@ export class ModelingEvents extends DisposableObject {
|
||||
private readonly onModeChangedEventEmitter: AppEventEmitter<ModeChangedEvent>;
|
||||
private readonly onModeledAndModifiedMethodsChangedEventEmitter: AppEventEmitter<ModeledAndModifiedMethodsChangedEvent>;
|
||||
private readonly onSelectedMethodChangedEventEmitter: AppEventEmitter<SelectedMethodChangedEvent>;
|
||||
private readonly onInProgressMethodsChangedEventEmitter: AppEventEmitter<InProgressMethodsChangedEvent>;
|
||||
private readonly onProcessedByAutoModelMethodsChangedEventEmitter: AppEventEmitter<ProcessedByAutoModelMethodsChangedEvent>;
|
||||
private readonly onModelEvaluationRunChangedEventEmitter: AppEventEmitter<ModelEvaluationRunChangedEvent>;
|
||||
private readonly onRevealInModelEditorEventEmitter: AppEventEmitter<RevealInModelEditorEvent>;
|
||||
private readonly onFocusModelEditorEventEmitter: AppEventEmitter<FocusModelEditorEvent>;
|
||||
@@ -135,6 +151,18 @@ export class ModelingEvents extends DisposableObject {
|
||||
this.onSelectedMethodChanged =
|
||||
this.onSelectedMethodChangedEventEmitter.event;
|
||||
|
||||
this.onInProgressMethodsChangedEventEmitter = this.push(
|
||||
app.createEventEmitter<InProgressMethodsChangedEvent>(),
|
||||
);
|
||||
this.onInProgressMethodsChanged =
|
||||
this.onInProgressMethodsChangedEventEmitter.event;
|
||||
|
||||
this.onProcessedByAutoModelMethodsChangedEventEmitter = this.push(
|
||||
app.createEventEmitter<ProcessedByAutoModelMethodsChangedEvent>(),
|
||||
);
|
||||
this.onProcessedByAutoModelMethodsChanged =
|
||||
this.onProcessedByAutoModelMethodsChangedEventEmitter.event;
|
||||
|
||||
this.onModelEvaluationRunChangedEventEmitter = this.push(
|
||||
app.createEventEmitter<ModelEvaluationRunChangedEvent>(),
|
||||
);
|
||||
@@ -226,6 +254,8 @@ export class ModelingEvents extends DisposableObject {
|
||||
usage: Usage,
|
||||
modeledMethods: ModeledMethod[],
|
||||
isModified: boolean,
|
||||
isInProgress: boolean,
|
||||
processedByAutoModel: boolean,
|
||||
) {
|
||||
this.onSelectedMethodChangedEventEmitter.fire({
|
||||
databaseItem,
|
||||
@@ -233,6 +263,28 @@ export class ModelingEvents extends DisposableObject {
|
||||
usage,
|
||||
modeledMethods,
|
||||
isModified,
|
||||
isInProgress,
|
||||
processedByAutoModel,
|
||||
});
|
||||
}
|
||||
|
||||
public fireInProgressMethodsChangedEvent(
|
||||
dbUri: string,
|
||||
methods: ReadonlySet<string>,
|
||||
) {
|
||||
this.onInProgressMethodsChangedEventEmitter.fire({
|
||||
dbUri,
|
||||
methods,
|
||||
});
|
||||
}
|
||||
|
||||
public fireProcessedByAutoModelMethodsChangedEvent(
|
||||
dbUri: string,
|
||||
methods: ReadonlySet<string>,
|
||||
) {
|
||||
this.onProcessedByAutoModelMethodsChangedEventEmitter.fire({
|
||||
dbUri,
|
||||
methods,
|
||||
});
|
||||
}
|
||||
|
||||
|
||||
@@ -15,6 +15,8 @@ interface InternalDbModelingState {
|
||||
mode: Mode;
|
||||
modeledMethods: Record<string, ModeledMethod[]>;
|
||||
modifiedMethodSignatures: Set<string>;
|
||||
inProgressMethods: Set<string>;
|
||||
processedByAutoModelMethods: Set<string>;
|
||||
selectedMethod: Method | undefined;
|
||||
selectedUsage: Usage | undefined;
|
||||
modelEvaluationRun: ModelEvaluationRun | undefined;
|
||||
@@ -28,6 +30,8 @@ export interface DbModelingState {
|
||||
readonly mode: Mode;
|
||||
readonly modeledMethods: Readonly<Record<string, readonly ModeledMethod[]>>;
|
||||
readonly modifiedMethodSignatures: ReadonlySet<string>;
|
||||
readonly inProgressMethods: ReadonlySet<string>;
|
||||
readonly processedByAutoModelMethods: ReadonlySet<string>;
|
||||
readonly selectedMethod: Method | undefined;
|
||||
readonly selectedUsage: Usage | undefined;
|
||||
readonly modelEvaluationRun: ModelEvaluationRun | undefined;
|
||||
@@ -40,6 +44,8 @@ export interface SelectedMethodDetails {
|
||||
readonly usage: Usage | undefined;
|
||||
readonly modeledMethods: readonly ModeledMethod[];
|
||||
readonly isModified: boolean;
|
||||
readonly isInProgress: boolean;
|
||||
readonly processedByAutoModel: boolean;
|
||||
}
|
||||
|
||||
export class ModelingStore extends DisposableObject {
|
||||
@@ -62,8 +68,10 @@ export class ModelingStore extends DisposableObject {
|
||||
mode,
|
||||
modeledMethods: {},
|
||||
modifiedMethodSignatures: new Set(),
|
||||
processedByAutoModelMethods: new Set(),
|
||||
selectedMethod: undefined,
|
||||
selectedUsage: undefined,
|
||||
inProgressMethods: new Set(),
|
||||
modelEvaluationRun: undefined,
|
||||
isModelAlertsViewOpen: false,
|
||||
});
|
||||
@@ -152,6 +160,7 @@ export class ModelingStore extends DisposableObject {
|
||||
methods,
|
||||
state.modeledMethods,
|
||||
state.modifiedMethodSignatures,
|
||||
state.processedByAutoModelMethods,
|
||||
);
|
||||
});
|
||||
}
|
||||
@@ -162,6 +171,7 @@ export class ModelingStore extends DisposableObject {
|
||||
state.methods,
|
||||
state.modeledMethods,
|
||||
state.modifiedMethodSignatures,
|
||||
state.processedByAutoModelMethods,
|
||||
);
|
||||
});
|
||||
}
|
||||
@@ -298,15 +308,75 @@ export class ModelingStore extends DisposableObject {
|
||||
|
||||
const modeledMethods = dbState.modeledMethods[method.signature] ?? [];
|
||||
const isModified = dbState.modifiedMethodSignatures.has(method.signature);
|
||||
const isInProgress = dbState.inProgressMethods.has(method.signature);
|
||||
const processedByAutoModel = dbState.processedByAutoModelMethods.has(
|
||||
method.signature,
|
||||
);
|
||||
this.modelingEvents.fireSelectedMethodChangedEvent(
|
||||
dbItem,
|
||||
method,
|
||||
usage,
|
||||
modeledMethods,
|
||||
isModified,
|
||||
isInProgress,
|
||||
processedByAutoModel,
|
||||
);
|
||||
}
|
||||
|
||||
public addInProgressMethods(
|
||||
dbItem: DatabaseItem,
|
||||
inProgressMethods: string[],
|
||||
) {
|
||||
this.changeInProgressMethods(dbItem, (state) => {
|
||||
state.inProgressMethods = new Set([
|
||||
...state.inProgressMethods,
|
||||
...inProgressMethods,
|
||||
]);
|
||||
});
|
||||
}
|
||||
|
||||
public removeInProgressMethods(
|
||||
dbItem: DatabaseItem,
|
||||
methodSignatures: string[],
|
||||
) {
|
||||
this.changeInProgressMethods(dbItem, (state) => {
|
||||
state.inProgressMethods = new Set(
|
||||
Array.from(state.inProgressMethods).filter(
|
||||
(s) => !methodSignatures.includes(s),
|
||||
),
|
||||
);
|
||||
});
|
||||
}
|
||||
|
||||
public getProcessedByAutoModelMethods(
|
||||
dbItem: DatabaseItem,
|
||||
methodSignatures?: string[],
|
||||
): Set<string> {
|
||||
const processedByAutoModelMethods =
|
||||
this.getState(dbItem).processedByAutoModelMethods;
|
||||
if (!methodSignatures) {
|
||||
return processedByAutoModelMethods;
|
||||
}
|
||||
return new Set(
|
||||
Array.from(processedByAutoModelMethods).filter((x) =>
|
||||
methodSignatures.includes(x),
|
||||
),
|
||||
);
|
||||
}
|
||||
|
||||
public addProcessedByAutoModelMethods(
|
||||
dbItem: DatabaseItem,
|
||||
processedByAutoModelMethods: string[],
|
||||
) {
|
||||
this.changeProcessedByAutoModelMethods(dbItem, (state) => {
|
||||
state.processedByAutoModelMethods = new Set([
|
||||
...state.processedByAutoModelMethods,
|
||||
...processedByAutoModelMethods,
|
||||
]);
|
||||
});
|
||||
this.updateMethodSorting(dbItem);
|
||||
}
|
||||
|
||||
public updateModelEvaluationRun(
|
||||
dbItem: DatabaseItem,
|
||||
evaluationRun: ModelEvaluationRun | undefined,
|
||||
@@ -335,6 +405,10 @@ export class ModelingStore extends DisposableObject {
|
||||
isModified: dbState.modifiedMethodSignatures.has(
|
||||
selectedMethod.signature,
|
||||
),
|
||||
isInProgress: dbState.inProgressMethods.has(selectedMethod.signature),
|
||||
processedByAutoModel: dbState.processedByAutoModelMethods.has(
|
||||
selectedMethod.signature,
|
||||
),
|
||||
};
|
||||
}
|
||||
|
||||
@@ -386,6 +460,34 @@ export class ModelingStore extends DisposableObject {
|
||||
);
|
||||
}
|
||||
|
||||
private changeInProgressMethods(
|
||||
dbItem: DatabaseItem,
|
||||
updateState: (state: InternalDbModelingState) => void,
|
||||
) {
|
||||
const state = this.getState(dbItem);
|
||||
|
||||
updateState(state);
|
||||
|
||||
this.modelingEvents.fireInProgressMethodsChangedEvent(
|
||||
dbItem.databaseUri.toString(),
|
||||
state.inProgressMethods,
|
||||
);
|
||||
}
|
||||
|
||||
private changeProcessedByAutoModelMethods(
|
||||
dbItem: DatabaseItem,
|
||||
updateState: (state: InternalDbModelingState) => void,
|
||||
) {
|
||||
const state = this.getState(dbItem);
|
||||
|
||||
updateState(state);
|
||||
|
||||
this.modelingEvents.fireProcessedByAutoModelMethodsChangedEvent(
|
||||
dbItem.databaseUri.toString(),
|
||||
state.processedByAutoModelMethods,
|
||||
);
|
||||
}
|
||||
|
||||
private changeModelEvaluationRun(
|
||||
dbItem: DatabaseItem,
|
||||
updateState: (state: InternalDbModelingState) => void,
|
||||
|
||||
@@ -17,7 +17,6 @@ export type Query = {
|
||||
* - libraryVersion: the version of the library that contains the external API. This is a string and can be empty if the version cannot be determined.
|
||||
* - type: the modeled kind of the method, either "sink", "source", "summary", or "neutral"
|
||||
* - classification: the classification of the use of the method, either "source", "test", "generated", or "unknown"
|
||||
* - kind: the kind of the endpoint, language-specific, e.g. "method" or "function"
|
||||
*/
|
||||
applicationModeQuery: string;
|
||||
/**
|
||||
@@ -33,7 +32,6 @@ export type Query = {
|
||||
* - supported: whether this method is modeled. This should be a string representation of a boolean to satify the result pattern for a problem query.
|
||||
* - libraryName: the name of the file or library that contains the method. This is a string and usually the basename of a file.
|
||||
* - type: the modeled kind of the method, either "sink", "source", "summary", or "neutral"
|
||||
* - kind: the kind of the endpoint, language-specific, e.g. "method" or "function"
|
||||
*/
|
||||
frameworkModeQuery: string;
|
||||
dependencies?: {
|
||||
@@ -52,7 +50,6 @@ export type ApplicationModeTuple = [
|
||||
string,
|
||||
ModeledMethodType,
|
||||
CallClassification,
|
||||
string | BqrsEntityValue | undefined,
|
||||
];
|
||||
|
||||
export type FrameworkModeTuple = [
|
||||
@@ -64,5 +61,4 @@ export type FrameworkModeTuple = [
|
||||
boolean,
|
||||
string,
|
||||
ModeledMethodType,
|
||||
string | BqrsEntityValue | undefined,
|
||||
];
|
||||
|
||||
@@ -0,0 +1,48 @@
|
||||
import type { Method, MethodSignature } from "../method";
|
||||
import type { ModeledMethod } from "../modeled-method";
|
||||
import type { Mode } from "./mode";
|
||||
import { groupMethods, sortGroupNames } from "./sorting";
|
||||
|
||||
/**
|
||||
* Return the candidates that the model should be run on. This includes limiting the number of
|
||||
* candidates to the candidate limit and filtering out anything that is already modeled and respecting
|
||||
* the order in the UI.
|
||||
* @param mode Whether it is application or framework mode.
|
||||
* @param methods all methods.
|
||||
* @param modeledMethodsBySignature the currently modeled methods.
|
||||
* @returns list of modeled methods that are candidates for modeling.
|
||||
*/
|
||||
|
||||
export function getCandidates(
|
||||
mode: Mode,
|
||||
methods: readonly Method[],
|
||||
modeledMethodsBySignature: Record<string, readonly ModeledMethod[]>,
|
||||
processedByAutoModelMethods: Set<string>,
|
||||
): MethodSignature[] {
|
||||
const candidateMethods = methods.filter((method) => {
|
||||
// Filter out any methods already processed by auto-model
|
||||
if (processedByAutoModelMethods.has(method.signature)) {
|
||||
return false;
|
||||
}
|
||||
|
||||
const modeledMethods: ModeledMethod[] = [
|
||||
...(modeledMethodsBySignature[method.signature] ?? []),
|
||||
];
|
||||
|
||||
// Anything that is modeled is not a candidate
|
||||
if (modeledMethods.some((m) => m.type !== "none")) {
|
||||
return false;
|
||||
}
|
||||
|
||||
// A method that is supported is modeled outside of the model file, so it is not a candidate.
|
||||
if (method.supported) {
|
||||
return false;
|
||||
}
|
||||
|
||||
return true;
|
||||
});
|
||||
|
||||
// Sort the same way as the UI so we send the first ones listed in the UI first
|
||||
const grouped = groupMethods(candidateMethods, mode);
|
||||
return sortGroupNames(grouped).flatMap((name) => grouped[name]);
|
||||
}
|
||||
@@ -47,6 +47,7 @@ export function sortMethods(
|
||||
methods: readonly Method[],
|
||||
modeledMethodsMap: Record<string, readonly ModeledMethod[]>,
|
||||
modifiedSignatures: ReadonlySet<string>,
|
||||
processedByAutoModelMethods: ReadonlySet<string>,
|
||||
): Method[] {
|
||||
const sortedMethods = [...methods];
|
||||
sortedMethods.sort((a, b) => {
|
||||
@@ -55,11 +56,13 @@ export function sortMethods(
|
||||
a,
|
||||
modeledMethodsMap[a.signature] ?? [],
|
||||
modifiedSignatures.has(a.signature),
|
||||
processedByAutoModelMethods.has(a.signature),
|
||||
);
|
||||
const methodBPrimarySortOrdinal = getMethodPrimarySortOrdinal(
|
||||
b,
|
||||
modeledMethodsMap[b.signature] ?? [],
|
||||
modifiedSignatures.has(b.signature),
|
||||
processedByAutoModelMethods.has(b.signature),
|
||||
);
|
||||
if (methodAPrimarySortOrdinal !== methodBPrimarySortOrdinal) {
|
||||
return methodAPrimarySortOrdinal - methodBPrimarySortOrdinal;
|
||||
@@ -79,25 +82,32 @@ export function sortMethods(
|
||||
|
||||
/**
|
||||
* Assigns numbers to the following classes of methods:
|
||||
* - Unsaved manual models + unmodeled methods => 0
|
||||
* - Saved models from this model pack (AutoModel and manual) => 1
|
||||
* - Methods not modelable in this model pack => 2
|
||||
* - Unsaved positive AutoModel predictions => 0
|
||||
* - Negative AutoModel predictions => 1
|
||||
* - Unsaved manual models + unmodeled methods => 2
|
||||
* - Saved models from this model pack (AutoModel and manual) => 3
|
||||
* - Methods not modelable in this model pack => 4
|
||||
*/
|
||||
function getMethodPrimarySortOrdinal(
|
||||
method: Method,
|
||||
modeledMethods: readonly ModeledMethod[],
|
||||
isUnsaved: boolean,
|
||||
isProcessedByAutoModel: boolean,
|
||||
): number {
|
||||
const canBeModeled = canMethodBeModeled(method, modeledMethods, isUnsaved);
|
||||
const isModeled = modeledMethods.length > 0;
|
||||
if (canBeModeled) {
|
||||
if ((isModeled && isUnsaved) || !isModeled) {
|
||||
if (isModeled && isUnsaved && isProcessedByAutoModel) {
|
||||
return 0;
|
||||
} else {
|
||||
} else if (!isModeled && isProcessedByAutoModel) {
|
||||
return 1;
|
||||
} else if ((isModeled && isUnsaved) || !isModeled) {
|
||||
return 2;
|
||||
} else {
|
||||
return 3;
|
||||
}
|
||||
} else {
|
||||
return 2;
|
||||
return 4;
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
@@ -7,6 +7,7 @@ export interface ModelEditorViewState {
|
||||
extensionPack: ExtensionPack;
|
||||
language: QueryLanguage;
|
||||
showGenerateButton: boolean;
|
||||
showLlmButton: boolean;
|
||||
showEvaluationUi: boolean;
|
||||
mode: Mode;
|
||||
showModeSwitchButton: boolean;
|
||||
|
||||
@@ -9,16 +9,20 @@ export const SUPPORTED_LANGUAGES: QueryLanguage[] = [
|
||||
QueryLanguage.Java,
|
||||
QueryLanguage.CSharp,
|
||||
QueryLanguage.Ruby,
|
||||
QueryLanguage.Python,
|
||||
];
|
||||
|
||||
export function isSupportedLanguage(
|
||||
language: QueryLanguage,
|
||||
_modelConfig: ModelConfig,
|
||||
modelConfig: ModelConfig,
|
||||
) {
|
||||
if (SUPPORTED_LANGUAGES.includes(language)) {
|
||||
return true;
|
||||
}
|
||||
|
||||
if (language === QueryLanguage.Python) {
|
||||
// Python is only enabled when the config setting is set
|
||||
return modelConfig.enablePython;
|
||||
}
|
||||
|
||||
return false;
|
||||
}
|
||||
|
||||
@@ -845,7 +845,7 @@ export class QueryHistoryManager extends DisposableObject {
|
||||
evalLogData,
|
||||
);
|
||||
this.evalLogViewer.updateRoots(await evalLogTreeBuilder.getRoots());
|
||||
} catch {
|
||||
} catch (e) {
|
||||
throw new Error(
|
||||
`Could not read evaluator log summary JSON file to generate viewer data at ${item.jsonEvalLogSummaryLocation}.`,
|
||||
);
|
||||
|
||||
@@ -48,11 +48,6 @@ function mapVariantAnalysisDtoToDto(
|
||||
): VariantAnalysisDto {
|
||||
return {
|
||||
id: variantAnalysis.id,
|
||||
controllerRepo: {
|
||||
id: variantAnalysis.controllerRepo.id,
|
||||
fullName: variantAnalysis.controllerRepo.fullName,
|
||||
private: variantAnalysis.controllerRepo.private,
|
||||
},
|
||||
query: {
|
||||
name: variantAnalysis.query.name,
|
||||
filePath: variantAnalysis.query.filePath,
|
||||
|
||||
@@ -48,12 +48,12 @@ function mapVariantAnalysisToDomainModel(
|
||||
): VariantAnalysis {
|
||||
return {
|
||||
id: variantAnalysis.id,
|
||||
controllerRepo: {
|
||||
id: variantAnalysis.controllerRepo.id,
|
||||
fullName: variantAnalysis.controllerRepo.fullName,
|
||||
private: variantAnalysis.controllerRepo.private,
|
||||
},
|
||||
language: mapQueryLanguageToDomainModel(variantAnalysis.query.language),
|
||||
controllerRepo: {
|
||||
id: 0,
|
||||
fullName: "",
|
||||
private: false,
|
||||
},
|
||||
query: {
|
||||
name: variantAnalysis.query.name,
|
||||
filePath: variantAnalysis.query.filePath,
|
||||
|
||||
@@ -15,11 +15,6 @@ export interface QueryHistoryVariantAnalysisDto {
|
||||
|
||||
export interface VariantAnalysisDto {
|
||||
id: number;
|
||||
controllerRepo: {
|
||||
id: number;
|
||||
fullName: string;
|
||||
private: boolean;
|
||||
};
|
||||
query: {
|
||||
name: string;
|
||||
filePath: string;
|
||||
|
||||
@@ -61,7 +61,7 @@ export class TestRunner extends DisposableObject {
|
||||
})) {
|
||||
await eventHandler(event);
|
||||
}
|
||||
} catch {
|
||||
} catch (e) {
|
||||
// CodeQL testing can throw exception even in normal scenarios. For example, if the test run
|
||||
// produces no output (which is normal), the testing command would throw an exception on
|
||||
// unexpected EOF during json parsing. So nothing needs to be done here - all the relevant
|
||||
|
||||
@@ -544,16 +544,9 @@ export async function generateEvalLogSummaries(
|
||||
await cliServer.generateJsonLogSummary(log, jsonSummary);
|
||||
|
||||
if (humanReadableSummary !== undefined) {
|
||||
progress(progressUpdate(3, 3, "Generating summary symbols file"));
|
||||
summarySymbols = outputDir.evalLogSummarySymbolsPath;
|
||||
if (
|
||||
!(await cliServer.cliConstraints.supportsGenerateSummarySymbolMap())
|
||||
) {
|
||||
// We're using an old CLI that cannot generate the summary symbols file while generating the
|
||||
// human-readable log summary. As a fallback, create it by parsing the human-readable
|
||||
// summary.
|
||||
progress(progressUpdate(3, 3, "Generating summary symbols file"));
|
||||
await generateSummarySymbolsFile(humanReadableSummary, summarySymbols);
|
||||
}
|
||||
await generateSummarySymbolsFile(humanReadableSummary, summarySymbols);
|
||||
}
|
||||
}
|
||||
|
||||
@@ -607,7 +600,7 @@ export async function logEndSummary(
|
||||
const endSummaryContent = await readFile(endSummary, "utf-8");
|
||||
void logger.log(" --- Evaluator Log Summary --- ");
|
||||
void logger.log(endSummaryContent);
|
||||
} catch {
|
||||
} catch (e) {
|
||||
void showAndLogWarningMessage(
|
||||
extLogger,
|
||||
`Could not read structured evaluator log end of summary file at ${endSummary}.`,
|
||||
|
||||
@@ -7,7 +7,7 @@ export default {
|
||||
component: ResponsiveContainerComponent,
|
||||
} as Meta<typeof ResponsiveContainerComponent>;
|
||||
|
||||
const Template: StoryFn<typeof ResponsiveContainerComponent> = () => (
|
||||
const Template: StoryFn<typeof ResponsiveContainerComponent> = (args) => (
|
||||
<ResponsiveContainerComponent>
|
||||
<span>Hello</span>
|
||||
</ResponsiveContainerComponent>
|
||||
|
||||
@@ -53,3 +53,20 @@ FullyModeledMethod.args = {
|
||||
method,
|
||||
modeledMethod,
|
||||
};
|
||||
|
||||
export const ModelingInProgress = Template.bind({});
|
||||
ModelingInProgress.args = {
|
||||
method,
|
||||
modeledMethod,
|
||||
isModelingInProgress: true,
|
||||
};
|
||||
|
||||
const generatedModeledMethod = createSinkModeledMethod({
|
||||
provenance: "ai-generated",
|
||||
});
|
||||
export const ModelingNotAccepted = Template.bind({});
|
||||
ModelingNotAccepted.args = {
|
||||
method,
|
||||
modeledMethod: generatedModeledMethod,
|
||||
modelPending: true,
|
||||
};
|
||||
|
||||
@@ -0,0 +1,14 @@
|
||||
import type { Meta, StoryFn } from "@storybook/react";
|
||||
|
||||
import { InProgressDropdown as InProgressDropdownComponent } from "../../view/model-editor/InProgressDropdown";
|
||||
|
||||
export default {
|
||||
title: "CodeQL Model Editor/In Progress Dropdown",
|
||||
component: InProgressDropdownComponent,
|
||||
} as Meta<typeof InProgressDropdownComponent>;
|
||||
|
||||
const Template: StoryFn<typeof InProgressDropdownComponent> = (args) => (
|
||||
<InProgressDropdownComponent />
|
||||
);
|
||||
|
||||
export const InProgressDropdown = Template.bind({});
|
||||
@@ -220,8 +220,10 @@ LibraryRow.args = {
|
||||
],
|
||||
},
|
||||
modifiedSignatures: new Set(["org.sql2o.Sql2o#Sql2o(String)"]),
|
||||
inProgressMethods: new Set(),
|
||||
viewState: createMockModelEditorViewState({
|
||||
showGenerateButton: true,
|
||||
showLlmButton: true,
|
||||
}),
|
||||
hideModeledMethods: false,
|
||||
};
|
||||
|
||||
@@ -96,6 +96,7 @@ const modeledMethod: ModeledMethod = {
|
||||
|
||||
const viewState = createMockModelEditorViewState({
|
||||
showGenerateButton: true,
|
||||
showLlmButton: true,
|
||||
});
|
||||
|
||||
export const Unmodeled = Template.bind({});
|
||||
@@ -145,6 +146,15 @@ AlreadyModeled.args = {
|
||||
viewState,
|
||||
};
|
||||
|
||||
export const ModelingInProgress = Template.bind({});
|
||||
ModelingInProgress.args = {
|
||||
method,
|
||||
modeledMethods: [modeledMethod],
|
||||
modelingInProgress: true,
|
||||
methodCanBeModeled: true,
|
||||
viewState,
|
||||
};
|
||||
|
||||
export const MultipleModelings = Template.bind({});
|
||||
MultipleModelings.args = {
|
||||
method,
|
||||
|
||||
@@ -27,6 +27,7 @@ ModelEditor.args = {
|
||||
dataExtensions: [],
|
||||
},
|
||||
showGenerateButton: true,
|
||||
showLlmButton: true,
|
||||
}),
|
||||
initialMethods: [
|
||||
{
|
||||
|
||||
@@ -2,7 +2,7 @@
|
||||
"compilerOptions": {
|
||||
"module": "esnext",
|
||||
"moduleResolution": "node",
|
||||
"target": "es2021",
|
||||
"target": "es6",
|
||||
"outDir": "out",
|
||||
"lib": ["ES2021", "dom"],
|
||||
"jsx": "react-jsx",
|
||||
|
||||
@@ -160,7 +160,7 @@ async function exportVariantAnalysisAnalysisResults(
|
||||
expectedAnalysesResultsCount: number,
|
||||
exportFormat: "gist" | "local",
|
||||
commandManager: AppCommandManager,
|
||||
credentials: Credentials,
|
||||
_credentials: Credentials,
|
||||
progress: ProgressCallback,
|
||||
token: CancellationToken,
|
||||
) {
|
||||
@@ -191,7 +191,6 @@ async function exportVariantAnalysisAnalysisResults(
|
||||
markdownFiles,
|
||||
exportFormat,
|
||||
commandManager,
|
||||
credentials,
|
||||
progress,
|
||||
token,
|
||||
);
|
||||
@@ -236,7 +235,6 @@ async function exportResults(
|
||||
markdownFiles: MarkdownFile[],
|
||||
exportFormat: "gist" | "local",
|
||||
commandManager: AppCommandManager,
|
||||
credentials: Credentials,
|
||||
progress?: ProgressCallback,
|
||||
token?: CancellationToken,
|
||||
) {
|
||||
@@ -249,7 +247,6 @@ async function exportResults(
|
||||
description,
|
||||
markdownFiles,
|
||||
commandManager,
|
||||
credentials,
|
||||
progress,
|
||||
token,
|
||||
);
|
||||
@@ -268,7 +265,6 @@ async function exportToGist(
|
||||
description: string,
|
||||
markdownFiles: MarkdownFile[],
|
||||
commandManager: AppCommandManager,
|
||||
credentials: Credentials,
|
||||
progress?: ProgressCallback,
|
||||
token?: CancellationToken,
|
||||
) {
|
||||
@@ -291,7 +287,7 @@ async function exportToGist(
|
||||
{} as { [key: string]: { content: string } },
|
||||
);
|
||||
|
||||
const gistUrl = await createGist(credentials, description, gistFiles);
|
||||
const gistUrl = await createGist(description, gistFiles);
|
||||
if (gistUrl) {
|
||||
// This needs to use .then to ensure we aren't keeping the progress notification open. We shouldn't await the
|
||||
// "Open gist" button click.
|
||||
|
||||
@@ -1,5 +1,4 @@
|
||||
import type { OctokitResponse } from "@octokit/types/dist-types";
|
||||
import type { Credentials } from "../../common/authentication";
|
||||
import { getGitHubInstanceUrl } from "../../config";
|
||||
import type { VariantAnalysisSubmission } from "../shared/variant-analysis";
|
||||
import type {
|
||||
VariantAnalysis,
|
||||
@@ -7,84 +6,142 @@ import type {
|
||||
VariantAnalysisSubmissionRequest,
|
||||
} from "./variant-analysis";
|
||||
import type { Repository } from "./repository";
|
||||
import { extLogger } from "../../common/logging/vscode";
|
||||
|
||||
export async function submitVariantAnalysis(
|
||||
credentials: Credentials,
|
||||
submissionDetails: VariantAnalysisSubmission,
|
||||
): Promise<VariantAnalysis> {
|
||||
const octokit = await credentials.getOctokit();
|
||||
|
||||
const { actionRepoRef, language, pack, databases, controllerRepoId } =
|
||||
submissionDetails;
|
||||
|
||||
const data: VariantAnalysisSubmissionRequest = {
|
||||
action_repo_ref: actionRepoRef,
|
||||
language,
|
||||
query_pack: pack,
|
||||
repositories: databases.repositories,
|
||||
repository_lists: databases.repositoryLists,
|
||||
repository_owners: databases.repositoryOwners,
|
||||
};
|
||||
|
||||
const response: OctokitResponse<VariantAnalysis> = await octokit.request(
|
||||
"POST /repositories/:controllerRepoId/code-scanning/codeql/variant-analyses",
|
||||
{
|
||||
controllerRepoId,
|
||||
data,
|
||||
},
|
||||
);
|
||||
|
||||
return response.data;
|
||||
function getOctokitBaseUrl(): string {
|
||||
let apiUrl = getGitHubInstanceUrl().toString();
|
||||
if (apiUrl.endsWith("/")) {
|
||||
apiUrl = apiUrl.slice(0, -1);
|
||||
}
|
||||
if (apiUrl.startsWith("https://")) {
|
||||
apiUrl = apiUrl.replace("https://", "http://");
|
||||
}
|
||||
return apiUrl;
|
||||
}
|
||||
|
||||
export async function submitVariantAnalysis(
|
||||
submissionDetails: VariantAnalysisSubmission,
|
||||
): Promise<VariantAnalysis> {
|
||||
try {
|
||||
console.log("Getting base URL...");
|
||||
const baseUrl = getOctokitBaseUrl();
|
||||
void extLogger.log(`Base URL: ${baseUrl}`);
|
||||
|
||||
const { actionRepoRef, language, pack, databases, controllerRepoId } =
|
||||
submissionDetails;
|
||||
|
||||
const data: VariantAnalysisSubmissionRequest = {
|
||||
action_repo_ref: actionRepoRef,
|
||||
language,
|
||||
query_pack: pack,
|
||||
repositories: databases.repositories,
|
||||
repository_lists: databases.repositoryLists,
|
||||
repository_owners: databases.repositoryOwners,
|
||||
};
|
||||
|
||||
void extLogger.log(
|
||||
`Sending fetch request with data: ${JSON.stringify(data)}`,
|
||||
);
|
||||
|
||||
void extLogger.log(
|
||||
`Fetch request URL: ${baseUrl}/repositories/${controllerRepoId}/code-scanning/codeql/variant-analyses`,
|
||||
);
|
||||
|
||||
const response = await fetch(
|
||||
`${baseUrl}/repositories/${controllerRepoId}/code-scanning/codeql/variant-analyses`,
|
||||
{
|
||||
method: "POST",
|
||||
headers: {
|
||||
"Content-Type": "application/json",
|
||||
},
|
||||
body: JSON.stringify(data),
|
||||
},
|
||||
);
|
||||
|
||||
void extLogger.log(`Response status: ${response.status}`);
|
||||
|
||||
if (!response.ok) {
|
||||
throw new Error(
|
||||
`Error submitting variant analysis: ${response.statusText}`,
|
||||
);
|
||||
}
|
||||
|
||||
const responseData = await response.json();
|
||||
void extLogger.log(`Response data: ${responseData}`);
|
||||
|
||||
return responseData;
|
||||
} catch (error) {
|
||||
void extLogger.log(`Error: ${error}`);
|
||||
throw error;
|
||||
}
|
||||
}
|
||||
export async function getVariantAnalysis(
|
||||
credentials: Credentials,
|
||||
controllerRepoId: number,
|
||||
variantAnalysisId: number,
|
||||
): Promise<VariantAnalysis> {
|
||||
const octokit = await credentials.getOctokit();
|
||||
const baseUrl = getOctokitBaseUrl();
|
||||
|
||||
const response: OctokitResponse<VariantAnalysis> = await octokit.request(
|
||||
"GET /repositories/:controllerRepoId/code-scanning/codeql/variant-analyses/:variantAnalysisId",
|
||||
const response = await fetch(
|
||||
`${baseUrl}/repositories/${controllerRepoId}/code-scanning/codeql/variant-analyses/${variantAnalysisId}`,
|
||||
{
|
||||
controllerRepoId,
|
||||
variantAnalysisId,
|
||||
method: "GET",
|
||||
headers: {
|
||||
"Content-Type": "application/json",
|
||||
},
|
||||
},
|
||||
);
|
||||
|
||||
return response.data;
|
||||
if (!response.ok) {
|
||||
throw new Error(`Error getting variant analysis: ${response.statusText}`);
|
||||
}
|
||||
|
||||
return response.json();
|
||||
}
|
||||
|
||||
export async function getVariantAnalysisRepo(
|
||||
credentials: Credentials,
|
||||
controllerRepoId: number,
|
||||
variantAnalysisId: number,
|
||||
repoId: number,
|
||||
): Promise<VariantAnalysisRepoTask> {
|
||||
const octokit = await credentials.getOctokit();
|
||||
const baseUrl = getOctokitBaseUrl();
|
||||
|
||||
const response: OctokitResponse<VariantAnalysisRepoTask> =
|
||||
await octokit.request(
|
||||
"GET /repositories/:controllerRepoId/code-scanning/codeql/variant-analyses/:variantAnalysisId/repositories/:repoId",
|
||||
{
|
||||
controllerRepoId,
|
||||
variantAnalysisId,
|
||||
repoId,
|
||||
const response = await fetch(
|
||||
`${baseUrl}/repositories/${controllerRepoId}/code-scanning/codeql/variant-analyses/${variantAnalysisId}/repositories/${repoId}`,
|
||||
{
|
||||
method: "GET",
|
||||
headers: {
|
||||
"Content-Type": "application/json",
|
||||
},
|
||||
);
|
||||
},
|
||||
);
|
||||
|
||||
return response.data;
|
||||
if (!response.ok) {
|
||||
throw new Error(
|
||||
`Error getting variant analysis repo: ${response.statusText}`,
|
||||
);
|
||||
}
|
||||
|
||||
return response.json();
|
||||
}
|
||||
|
||||
export async function getRepositoryFromNwo(
|
||||
credentials: Credentials,
|
||||
owner: string,
|
||||
repo: string,
|
||||
): Promise<Repository> {
|
||||
const octokit = await credentials.getOctokit();
|
||||
const baseUrl = getOctokitBaseUrl();
|
||||
|
||||
const response = await octokit.rest.repos.get({ owner, repo });
|
||||
return response.data as Repository;
|
||||
const response = await fetch(`${baseUrl}/repos/${owner}/${repo}`, {
|
||||
method: "GET",
|
||||
headers: {
|
||||
"Content-Type": "application/json",
|
||||
},
|
||||
});
|
||||
|
||||
if (!response.ok) {
|
||||
throw new Error(`Error getting repository: ${response.statusText}`);
|
||||
}
|
||||
|
||||
return response.json();
|
||||
}
|
||||
|
||||
/**
|
||||
@@ -92,22 +149,29 @@ export async function getRepositoryFromNwo(
|
||||
* Returns the URL of the created gist.
|
||||
*/
|
||||
export async function createGist(
|
||||
credentials: Credentials,
|
||||
description: string,
|
||||
files: { [key: string]: { content: string } },
|
||||
): Promise<string | undefined> {
|
||||
const octokit = await credentials.getOctokit();
|
||||
const response = await octokit.request("POST /gists", {
|
||||
description,
|
||||
files,
|
||||
public: false,
|
||||
const baseUrl = getOctokitBaseUrl();
|
||||
|
||||
const response = await fetch(`${baseUrl}/gists`, {
|
||||
method: "POST",
|
||||
headers: {
|
||||
"Content-Type": "application/json",
|
||||
},
|
||||
body: JSON.stringify({
|
||||
description,
|
||||
files,
|
||||
public: false,
|
||||
}),
|
||||
});
|
||||
if (response.status >= 300) {
|
||||
|
||||
if (!response.ok) {
|
||||
throw new Error(
|
||||
`Error exporting variant analysis results: ${response.status} ${
|
||||
response?.data || ""
|
||||
}`,
|
||||
`Error creating gist: ${response.status} ${response.statusText}`,
|
||||
);
|
||||
}
|
||||
return response.data.html_url;
|
||||
|
||||
const data = await response.json();
|
||||
return data.html_url;
|
||||
}
|
||||
|
||||
@@ -26,7 +26,7 @@ export async function readRepoStates(
|
||||
const repoStates = mapRepoStatesToDomainModel(repoStatesData);
|
||||
|
||||
return repoStates;
|
||||
} catch {
|
||||
} catch (e) {
|
||||
// Ignore this error, we simply might not have downloaded anything yet
|
||||
return undefined;
|
||||
}
|
||||
|
||||
@@ -1,5 +1,5 @@
|
||||
import type { CancellationToken } from "vscode";
|
||||
import { Uri, window } from "vscode";
|
||||
import { Uri } from "vscode";
|
||||
import { join, sep, basename, relative } from "path";
|
||||
import { dump, load } from "js-yaml";
|
||||
import { copy, writeFile, readFile, mkdirp } from "fs-extra";
|
||||
@@ -7,26 +7,17 @@ import type { DirectoryResult } from "tmp-promise";
|
||||
import { dir, tmpName } from "tmp-promise";
|
||||
import { tmpDir } from "../tmp-dir";
|
||||
import { getOnDiskWorkspaceFolders } from "../common/vscode/workspace-folders";
|
||||
import type { Credentials } from "../common/authentication";
|
||||
import type { CodeQLCliServer } from "../codeql-cli/cli";
|
||||
import { extLogger } from "../common/logging/vscode";
|
||||
import {
|
||||
getActionBranch,
|
||||
getRemoteControllerRepo,
|
||||
setRemoteControllerRepo,
|
||||
} from "../config";
|
||||
import { getActionBranch } from "../config";
|
||||
import type { ProgressCallback } from "../common/vscode/progress";
|
||||
import { UserCancellationException } from "../common/vscode/progress";
|
||||
import type { RequestError } from "@octokit/types/dist-types";
|
||||
import type { QueryMetadata } from "../common/interface-types";
|
||||
import { getErrorMessage, REPO_REGEX } from "../common/helpers-pure";
|
||||
import { getRepositoryFromNwo } from "./gh-api/gh-api-client";
|
||||
import type { RepositorySelection } from "./repository-selection";
|
||||
import {
|
||||
getRepositorySelection,
|
||||
isValidSelection,
|
||||
} from "./repository-selection";
|
||||
import type { Repository } from "./shared/repository";
|
||||
import type { DbManager } from "../databases/db-manager";
|
||||
import {
|
||||
getQlPackFilePath,
|
||||
@@ -116,6 +107,15 @@ async function generateQueryPack(
|
||||
|
||||
let precompilationOpts: string[];
|
||||
if (cliSupportsMrvaPackCreate) {
|
||||
if (
|
||||
qlPackDetails.queryFiles.length > 1 &&
|
||||
!(await cliServer.cliConstraints.supportsPackCreateWithMultipleQueries())
|
||||
) {
|
||||
throw new Error(
|
||||
`Installed CLI version does not allow creating a MRVA pack with multiple queries`,
|
||||
);
|
||||
}
|
||||
|
||||
const queryOpts = qlPackDetails.queryFiles.flatMap((q) => [
|
||||
"--query",
|
||||
join(targetPackPath, relative(qlPackDetails.qlPackRootPath, q)),
|
||||
@@ -276,13 +276,11 @@ interface PreparedRemoteQuery {
|
||||
base64Pack: string;
|
||||
modelPacks: ModelPackDetails[];
|
||||
repoSelection: RepositorySelection;
|
||||
controllerRepo: Repository;
|
||||
queryStartTime: number;
|
||||
}
|
||||
|
||||
export async function prepareRemoteQueryRun(
|
||||
cliServer: CodeQLCliServer,
|
||||
credentials: Credentials,
|
||||
qlPackDetails: QlPackDetails,
|
||||
progress: ProgressCallback,
|
||||
token: CancellationToken,
|
||||
@@ -313,8 +311,6 @@ export async function prepareRemoteQueryRun(
|
||||
message: "Determining controller repo",
|
||||
});
|
||||
|
||||
const controllerRepo = await getControllerRepo(credentials);
|
||||
|
||||
progress({
|
||||
maxStep: 4,
|
||||
step: 3,
|
||||
@@ -358,7 +354,6 @@ export async function prepareRemoteQueryRun(
|
||||
base64Pack: generatedPack.base64Pack,
|
||||
modelPacks: generatedPack.modelPacks,
|
||||
repoSelection,
|
||||
controllerRepo,
|
||||
queryStartTime,
|
||||
};
|
||||
}
|
||||
@@ -485,84 +480,6 @@ export function getQueryName(
|
||||
return queryMetadata?.name ?? basename(queryFilePath);
|
||||
}
|
||||
|
||||
export async function getControllerRepo(
|
||||
credentials: Credentials,
|
||||
): Promise<Repository> {
|
||||
// Get the controller repo from the config, if it exists.
|
||||
// If it doesn't exist, prompt the user to enter it, check
|
||||
// whether the repo exists, and save the nwo to the config.
|
||||
|
||||
let shouldSetControllerRepo = false;
|
||||
let controllerRepoNwo: string | undefined;
|
||||
controllerRepoNwo = getRemoteControllerRepo();
|
||||
if (!controllerRepoNwo || !REPO_REGEX.test(controllerRepoNwo)) {
|
||||
void extLogger.log(
|
||||
controllerRepoNwo
|
||||
? "Invalid controller repository name."
|
||||
: "No controller repository defined.",
|
||||
);
|
||||
controllerRepoNwo = await window.showInputBox({
|
||||
title:
|
||||
"Controller repository in which to run GitHub Actions workflows for variant analyses",
|
||||
placeHolder: "<owner>/<repo>",
|
||||
prompt:
|
||||
"Enter the name of a GitHub repository in the format <owner>/<repo>. You can change this in the extension settings.",
|
||||
ignoreFocusOut: true,
|
||||
});
|
||||
if (!controllerRepoNwo) {
|
||||
throw new UserCancellationException("No controller repository entered.");
|
||||
} else if (!REPO_REGEX.test(controllerRepoNwo)) {
|
||||
// Check if user entered invalid input
|
||||
throw new UserCancellationException(
|
||||
"Invalid repository format. Must be a valid GitHub repository in the format <owner>/<repo>.",
|
||||
);
|
||||
}
|
||||
|
||||
shouldSetControllerRepo = true;
|
||||
}
|
||||
|
||||
void extLogger.log(`Using controller repository: ${controllerRepoNwo}`);
|
||||
const controllerRepo = await getControllerRepoFromApi(
|
||||
credentials,
|
||||
controllerRepoNwo,
|
||||
);
|
||||
|
||||
if (shouldSetControllerRepo) {
|
||||
void extLogger.log(
|
||||
`Setting the controller repository as: ${controllerRepoNwo}`,
|
||||
);
|
||||
await setRemoteControllerRepo(controllerRepoNwo);
|
||||
}
|
||||
|
||||
return controllerRepo;
|
||||
}
|
||||
|
||||
async function getControllerRepoFromApi(
|
||||
credentials: Credentials,
|
||||
nwo: string,
|
||||
): Promise<Repository> {
|
||||
const [owner, repo] = nwo.split("/");
|
||||
try {
|
||||
const controllerRepo = await getRepositoryFromNwo(credentials, owner, repo);
|
||||
void extLogger.log(`Controller repository ID: ${controllerRepo.id}`);
|
||||
return {
|
||||
id: controllerRepo.id,
|
||||
fullName: controllerRepo.full_name,
|
||||
private: controllerRepo.private,
|
||||
};
|
||||
} catch (e) {
|
||||
if ((e as RequestError).status === 404) {
|
||||
throw new Error(`Controller repository "${owner}/${repo}" not found`);
|
||||
} else {
|
||||
throw new Error(
|
||||
`Error getting controller repository "${owner}/${repo}": ${getErrorMessage(
|
||||
e,
|
||||
)}`,
|
||||
);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
function removeWorkspaceRefs(qlpack: QlPackFile) {
|
||||
if (!qlpack.dependencies) {
|
||||
return;
|
||||
|
||||
@@ -78,6 +78,7 @@ import {
|
||||
REPO_STATES_FILENAME,
|
||||
writeRepoStates,
|
||||
} from "./repo-states-store";
|
||||
import { FetchError } from "node-fetch";
|
||||
import {
|
||||
showAndLogExceptionWithTelemetry,
|
||||
showAndLogInformationMessage,
|
||||
@@ -372,16 +373,19 @@ export class VariantAnalysisManager
|
||||
);
|
||||
}
|
||||
|
||||
// log to extLogger
|
||||
void this.app.logger.log(
|
||||
`Running variant analysis with query: ${queryName}, language: ${variantAnalysisLanguage}`,
|
||||
);
|
||||
|
||||
const {
|
||||
actionBranch,
|
||||
base64Pack,
|
||||
modelPacks,
|
||||
repoSelection,
|
||||
controllerRepo,
|
||||
queryStartTime,
|
||||
} = await prepareRemoteQueryRun(
|
||||
this.cliServer,
|
||||
this.app.credentials,
|
||||
qlPackDetails,
|
||||
progress,
|
||||
token,
|
||||
@@ -398,12 +402,15 @@ export class VariantAnalysisManager
|
||||
count: qlPackDetails.queryFiles.length,
|
||||
};
|
||||
|
||||
// log that submitting
|
||||
void this.app.logger.log("Submitting variant analysis");
|
||||
|
||||
const variantAnalysisSubmission: VariantAnalysisSubmission = {
|
||||
startTime: queryStartTime,
|
||||
actionRepoRef: actionBranch,
|
||||
controllerRepoId: controllerRepo.id,
|
||||
language: variantAnalysisLanguage,
|
||||
pack: base64Pack,
|
||||
controllerRepoId: 0,
|
||||
query: {
|
||||
name: queryName,
|
||||
filePath: firstQueryFile,
|
||||
@@ -421,7 +428,6 @@ export class VariantAnalysisManager
|
||||
let variantAnalysisResponse: ApiVariantAnalysis;
|
||||
try {
|
||||
variantAnalysisResponse = await submitVariantAnalysis(
|
||||
this.app.credentials,
|
||||
variantAnalysisSubmission,
|
||||
);
|
||||
} catch (e: unknown) {
|
||||
@@ -430,9 +436,17 @@ export class VariantAnalysisManager
|
||||
e instanceof RequestError &&
|
||||
handleRequestError(e, this.config.githubUrl, this.app.logger)
|
||||
) {
|
||||
// log
|
||||
void this.app.logger.log(
|
||||
`Error submitting variant analysis: ${getErrorMessage(e)}`,
|
||||
);
|
||||
return undefined;
|
||||
}
|
||||
|
||||
// throwing
|
||||
void this.app.logger.log(
|
||||
`Error submitting variant analysis: ${getErrorMessage(e)}`,
|
||||
);
|
||||
throw e;
|
||||
}
|
||||
|
||||
@@ -561,7 +575,7 @@ export class VariantAnalysisManager
|
||||
});
|
||||
const doc = await workspace.openTextDocument(uri);
|
||||
await Window.showTextDocument(doc, { preview: false });
|
||||
} catch {
|
||||
} catch (error) {
|
||||
void showAndLogWarningMessage(
|
||||
this.app.logger,
|
||||
"Could not open variant analysis query text. Failed to open text document.",
|
||||
@@ -585,7 +599,7 @@ export class VariantAnalysisManager
|
||||
variantAnalysis.query.filePath,
|
||||
);
|
||||
await Window.showTextDocument(textDocument, ViewColumn.One);
|
||||
} catch {
|
||||
} catch (error) {
|
||||
void showAndLogWarningMessage(
|
||||
this.app.logger,
|
||||
`Could not open file: ${variantAnalysis.query.filePath}`,
|
||||
@@ -805,8 +819,7 @@ export class VariantAnalysisManager
|
||||
let repoTask: VariantAnalysisRepositoryTask;
|
||||
try {
|
||||
const repoTaskResponse = await getVariantAnalysisRepo(
|
||||
this.app.credentials,
|
||||
variantAnalysis.controllerRepo.id,
|
||||
0,
|
||||
variantAnalysis.id,
|
||||
scannedRepo.repository.id,
|
||||
);
|
||||
@@ -858,9 +871,7 @@ export class VariantAnalysisManager
|
||||
} catch (e) {
|
||||
if (
|
||||
retry++ < maxRetryCount &&
|
||||
e &&
|
||||
typeof e === "object" &&
|
||||
"code" in e &&
|
||||
e instanceof FetchError &&
|
||||
(e.code === "ETIMEDOUT" || e.code === "ECONNRESET")
|
||||
) {
|
||||
void this.app.logger.log(
|
||||
|
||||
@@ -62,7 +62,6 @@ export class VariantAnalysisMonitor extends DisposableObject {
|
||||
try {
|
||||
await this._monitorVariantAnalysis(
|
||||
variantAnalysis.id,
|
||||
variantAnalysis.controllerRepo.id,
|
||||
variantAnalysis.executionStartTime,
|
||||
variantAnalysis.query.name,
|
||||
variantAnalysis.language,
|
||||
@@ -74,7 +73,6 @@ export class VariantAnalysisMonitor extends DisposableObject {
|
||||
|
||||
private async _monitorVariantAnalysis(
|
||||
variantAnalysisId: number,
|
||||
controllerRepoId: number,
|
||||
executionStartTime: number,
|
||||
queryName: string,
|
||||
language: QueryLanguage,
|
||||
@@ -97,11 +95,7 @@ export class VariantAnalysisMonitor extends DisposableObject {
|
||||
|
||||
let variantAnalysisSummary: ApiVariantAnalysis;
|
||||
try {
|
||||
variantAnalysisSummary = await getVariantAnalysis(
|
||||
this.app.credentials,
|
||||
controllerRepoId,
|
||||
variantAnalysisId,
|
||||
);
|
||||
variantAnalysisSummary = await getVariantAnalysis(0, variantAnalysisId);
|
||||
} catch (e) {
|
||||
const errorMessage = getErrorMessage(e);
|
||||
|
||||
|
||||
Some files were not shown because too many files have changed in this diff Show More
Reference in New Issue
Block a user