mirror of
https://github.com/mountain-loop/yaak.git
synced 2026-02-27 04:10:03 -05:00
Compare commits
37 Commits
cli-improv
...
codex/cli-
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
b9b90188a0 | ||
|
|
e64404d7a5 | ||
|
|
072b486857 | ||
|
|
23e9cbb376 | ||
|
|
9c09e32a56 | ||
|
|
be26cc4db4 | ||
|
|
a2f12aef35 | ||
|
|
e34301ccab | ||
|
|
8f0062f917 | ||
|
|
68d68035a1 | ||
|
|
ffc80d234c | ||
|
|
644c683714 | ||
|
|
d2c1bd79ac | ||
|
|
020589f2e6 | ||
|
|
b2a70d8938 | ||
|
|
64c626ed30 | ||
|
|
35d9ed901a | ||
|
|
f04b34be1a | ||
|
|
1e7e1232da | ||
|
|
c31d477a90 | ||
|
|
443e1b8262 | ||
|
|
c6b7cb2e32 | ||
|
|
4aef826a80 | ||
|
|
50c7992b42 | ||
|
|
5e9aebda6f | ||
|
|
a1e84c7785 | ||
|
|
fea4411afa | ||
|
|
8315e4afad | ||
|
|
a19ee9b502 | ||
|
|
0130bdee6f | ||
|
|
71ae9f41ed | ||
|
|
d06b6ce636 | ||
|
|
f5727b28c4 | ||
|
|
c62db7be06 | ||
|
|
4e56daa555 | ||
|
|
746bedf885 | ||
|
|
949c4a445a |
59
.github/workflows/release-api-npm.yml
vendored
Normal file
59
.github/workflows/release-api-npm.yml
vendored
Normal file
@@ -0,0 +1,59 @@
|
|||||||
|
name: Release API to NPM
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
tags: [yaak-api-*]
|
||||||
|
workflow_dispatch:
|
||||||
|
inputs:
|
||||||
|
version:
|
||||||
|
description: API version to publish (for example 0.9.0 or v0.9.0)
|
||||||
|
required: true
|
||||||
|
type: string
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
publish-npm:
|
||||||
|
name: Publish @yaakapp/api
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
id-token: write
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Setup Node
|
||||||
|
uses: actions/setup-node@v4
|
||||||
|
with:
|
||||||
|
node-version: lts/*
|
||||||
|
registry-url: https://registry.npmjs.org
|
||||||
|
|
||||||
|
- name: Install dependencies
|
||||||
|
run: npm ci
|
||||||
|
|
||||||
|
- name: Set @yaakapp/api version
|
||||||
|
shell: bash
|
||||||
|
env:
|
||||||
|
WORKFLOW_VERSION: ${{ inputs.version }}
|
||||||
|
run: |
|
||||||
|
set -euo pipefail
|
||||||
|
if [ "${{ github.event_name }}" = "workflow_dispatch" ]; then
|
||||||
|
VERSION="$WORKFLOW_VERSION"
|
||||||
|
else
|
||||||
|
VERSION="${GITHUB_REF_NAME#yaak-api-}"
|
||||||
|
fi
|
||||||
|
VERSION="${VERSION#v}"
|
||||||
|
echo "Preparing @yaakapp/api version: $VERSION"
|
||||||
|
cd packages/plugin-runtime-types
|
||||||
|
npm version "$VERSION" --no-git-tag-version --allow-same-version
|
||||||
|
|
||||||
|
- name: Build @yaakapp/api
|
||||||
|
working-directory: packages/plugin-runtime-types
|
||||||
|
run: npm run build
|
||||||
|
|
||||||
|
- name: Publish @yaakapp/api
|
||||||
|
working-directory: packages/plugin-runtime-types
|
||||||
|
run: npm publish --provenance --access public
|
||||||
@@ -1,4 +1,4 @@
|
|||||||
name: Generate Artifacts
|
name: Release App Artifacts
|
||||||
on:
|
on:
|
||||||
push:
|
push:
|
||||||
tags: [v*]
|
tags: [v*]
|
||||||
218
.github/workflows/release-cli-npm.yml
vendored
Normal file
218
.github/workflows/release-cli-npm.yml
vendored
Normal file
@@ -0,0 +1,218 @@
|
|||||||
|
name: Release CLI to NPM
|
||||||
|
|
||||||
|
on:
|
||||||
|
push:
|
||||||
|
tags: [yaak-cli-*]
|
||||||
|
workflow_dispatch:
|
||||||
|
inputs:
|
||||||
|
version:
|
||||||
|
description: CLI version to publish (for example 0.4.0 or v0.4.0)
|
||||||
|
required: true
|
||||||
|
type: string
|
||||||
|
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
|
||||||
|
jobs:
|
||||||
|
prepare-vendored-assets:
|
||||||
|
name: Prepare vendored plugin assets
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Setup Node
|
||||||
|
uses: actions/setup-node@v4
|
||||||
|
with:
|
||||||
|
node-version: lts/*
|
||||||
|
|
||||||
|
- name: Install Rust stable
|
||||||
|
uses: dtolnay/rust-toolchain@stable
|
||||||
|
|
||||||
|
- name: Install dependencies
|
||||||
|
run: npm ci
|
||||||
|
|
||||||
|
- name: Build plugin assets
|
||||||
|
env:
|
||||||
|
SKIP_WASM_BUILD: "1"
|
||||||
|
run: |
|
||||||
|
npm run build
|
||||||
|
npm run vendor:vendor-plugins
|
||||||
|
|
||||||
|
- name: Upload vendored assets
|
||||||
|
uses: actions/upload-artifact@v4
|
||||||
|
with:
|
||||||
|
name: vendored-assets
|
||||||
|
path: |
|
||||||
|
crates-tauri/yaak-app/vendored/plugin-runtime/index.cjs
|
||||||
|
crates-tauri/yaak-app/vendored/plugins
|
||||||
|
if-no-files-found: error
|
||||||
|
|
||||||
|
build-binaries:
|
||||||
|
name: Build ${{ matrix.pkg }}
|
||||||
|
needs: prepare-vendored-assets
|
||||||
|
runs-on: ${{ matrix.runner }}
|
||||||
|
strategy:
|
||||||
|
fail-fast: false
|
||||||
|
matrix:
|
||||||
|
include:
|
||||||
|
- pkg: cli-darwin-arm64
|
||||||
|
runner: macos-latest
|
||||||
|
target: aarch64-apple-darwin
|
||||||
|
binary: yaak
|
||||||
|
- pkg: cli-darwin-x64
|
||||||
|
runner: macos-latest
|
||||||
|
target: x86_64-apple-darwin
|
||||||
|
binary: yaak
|
||||||
|
- pkg: cli-linux-arm64
|
||||||
|
runner: ubuntu-22.04-arm
|
||||||
|
target: aarch64-unknown-linux-gnu
|
||||||
|
binary: yaak
|
||||||
|
- pkg: cli-linux-x64
|
||||||
|
runner: ubuntu-22.04
|
||||||
|
target: x86_64-unknown-linux-gnu
|
||||||
|
binary: yaak
|
||||||
|
- pkg: cli-win32-arm64
|
||||||
|
runner: windows-latest
|
||||||
|
target: aarch64-pc-windows-msvc
|
||||||
|
binary: yaak.exe
|
||||||
|
- pkg: cli-win32-x64
|
||||||
|
runner: windows-latest
|
||||||
|
target: x86_64-pc-windows-msvc
|
||||||
|
binary: yaak.exe
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Install Rust stable
|
||||||
|
uses: dtolnay/rust-toolchain@stable
|
||||||
|
with:
|
||||||
|
targets: ${{ matrix.target }}
|
||||||
|
|
||||||
|
- name: Restore Rust cache
|
||||||
|
uses: Swatinem/rust-cache@v2
|
||||||
|
with:
|
||||||
|
shared-key: release-cli-npm
|
||||||
|
cache-on-failure: true
|
||||||
|
|
||||||
|
- name: Install Linux build dependencies
|
||||||
|
if: startsWith(matrix.runner, 'ubuntu')
|
||||||
|
run: |
|
||||||
|
sudo apt-get update
|
||||||
|
sudo apt-get install -y pkg-config libdbus-1-dev
|
||||||
|
|
||||||
|
- name: Download vendored assets
|
||||||
|
uses: actions/download-artifact@v4
|
||||||
|
with:
|
||||||
|
name: vendored-assets
|
||||||
|
path: crates-tauri/yaak-app/vendored
|
||||||
|
|
||||||
|
- name: Set CLI build version
|
||||||
|
shell: bash
|
||||||
|
env:
|
||||||
|
WORKFLOW_VERSION: ${{ inputs.version }}
|
||||||
|
run: |
|
||||||
|
set -euo pipefail
|
||||||
|
if [ "${{ github.event_name }}" = "workflow_dispatch" ]; then
|
||||||
|
VERSION="$WORKFLOW_VERSION"
|
||||||
|
else
|
||||||
|
VERSION="${GITHUB_REF_NAME#yaak-cli-}"
|
||||||
|
fi
|
||||||
|
VERSION="${VERSION#v}"
|
||||||
|
echo "Building yaak version: $VERSION"
|
||||||
|
echo "YAAK_CLI_VERSION=$VERSION" >> "$GITHUB_ENV"
|
||||||
|
|
||||||
|
- name: Build yaak
|
||||||
|
run: cargo build --locked --release -p yaak-cli --bin yaak --target ${{ matrix.target }}
|
||||||
|
|
||||||
|
- name: Stage binary artifact
|
||||||
|
shell: bash
|
||||||
|
run: |
|
||||||
|
set -euo pipefail
|
||||||
|
mkdir -p "npm/dist/${{ matrix.pkg }}"
|
||||||
|
cp "target/${{ matrix.target }}/release/${{ matrix.binary }}" "npm/dist/${{ matrix.pkg }}/${{ matrix.binary }}"
|
||||||
|
|
||||||
|
- name: Upload binary artifact
|
||||||
|
uses: actions/upload-artifact@v4
|
||||||
|
with:
|
||||||
|
name: ${{ matrix.pkg }}
|
||||||
|
path: npm/dist/${{ matrix.pkg }}/${{ matrix.binary }}
|
||||||
|
if-no-files-found: error
|
||||||
|
|
||||||
|
publish-npm:
|
||||||
|
name: Publish @yaakapp/cli packages
|
||||||
|
needs: build-binaries
|
||||||
|
runs-on: ubuntu-latest
|
||||||
|
permissions:
|
||||||
|
contents: read
|
||||||
|
id-token: write
|
||||||
|
|
||||||
|
steps:
|
||||||
|
- name: Checkout
|
||||||
|
uses: actions/checkout@v4
|
||||||
|
|
||||||
|
- name: Setup Node
|
||||||
|
uses: actions/setup-node@v4
|
||||||
|
with:
|
||||||
|
node-version: lts/*
|
||||||
|
registry-url: https://registry.npmjs.org
|
||||||
|
|
||||||
|
- name: Download binary artifacts
|
||||||
|
uses: actions/download-artifact@v4
|
||||||
|
with:
|
||||||
|
pattern: cli-*
|
||||||
|
path: npm/dist
|
||||||
|
merge-multiple: false
|
||||||
|
|
||||||
|
- name: Prepare npm packages
|
||||||
|
shell: bash
|
||||||
|
env:
|
||||||
|
WORKFLOW_VERSION: ${{ inputs.version }}
|
||||||
|
run: |
|
||||||
|
set -euo pipefail
|
||||||
|
if [ "${{ github.event_name }}" = "workflow_dispatch" ]; then
|
||||||
|
VERSION="$WORKFLOW_VERSION"
|
||||||
|
else
|
||||||
|
VERSION="${GITHUB_REF_NAME#yaak-cli-}"
|
||||||
|
fi
|
||||||
|
VERSION="${VERSION#v}"
|
||||||
|
if [[ "$VERSION" == *-* ]]; then
|
||||||
|
PRERELEASE="${VERSION#*-}"
|
||||||
|
NPM_TAG="${PRERELEASE%%.*}"
|
||||||
|
else
|
||||||
|
NPM_TAG="latest"
|
||||||
|
fi
|
||||||
|
echo "Preparing CLI npm packages for version: $VERSION"
|
||||||
|
echo "Publishing with npm dist-tag: $NPM_TAG"
|
||||||
|
echo "NPM_TAG=$NPM_TAG" >> "$GITHUB_ENV"
|
||||||
|
YAAK_CLI_VERSION="$VERSION" node npm/prepare-publish.js
|
||||||
|
|
||||||
|
- name: Publish @yaakapp/cli-darwin-arm64
|
||||||
|
run: npm publish --provenance --access public --tag "$NPM_TAG"
|
||||||
|
working-directory: npm/cli-darwin-arm64
|
||||||
|
|
||||||
|
- name: Publish @yaakapp/cli-darwin-x64
|
||||||
|
run: npm publish --provenance --access public --tag "$NPM_TAG"
|
||||||
|
working-directory: npm/cli-darwin-x64
|
||||||
|
|
||||||
|
- name: Publish @yaakapp/cli-linux-arm64
|
||||||
|
run: npm publish --provenance --access public --tag "$NPM_TAG"
|
||||||
|
working-directory: npm/cli-linux-arm64
|
||||||
|
|
||||||
|
- name: Publish @yaakapp/cli-linux-x64
|
||||||
|
run: npm publish --provenance --access public --tag "$NPM_TAG"
|
||||||
|
working-directory: npm/cli-linux-x64
|
||||||
|
|
||||||
|
- name: Publish @yaakapp/cli-win32-arm64
|
||||||
|
run: npm publish --provenance --access public --tag "$NPM_TAG"
|
||||||
|
working-directory: npm/cli-win32-arm64
|
||||||
|
|
||||||
|
- name: Publish @yaakapp/cli-win32-x64
|
||||||
|
run: npm publish --provenance --access public --tag "$NPM_TAG"
|
||||||
|
working-directory: npm/cli-win32-x64
|
||||||
|
|
||||||
|
- name: Publish @yaakapp/cli
|
||||||
|
run: npm publish --provenance --access public --tag "$NPM_TAG"
|
||||||
|
working-directory: npm/cli
|
||||||
1988
Cargo.lock
generated
1988
Cargo.lock
generated
File diff suppressed because it is too large
Load Diff
@@ -22,7 +22,7 @@
|
|||||||
<!-- sponsors-premium --><a href="https://github.com/MVST-Solutions"><img src="https://github.com/MVST-Solutions.png" width="80px" alt="User avatar: MVST-Solutions" /></a> <a href="https://github.com/dharsanb"><img src="https://github.com/dharsanb.png" width="80px" alt="User avatar: dharsanb" /></a> <a href="https://github.com/railwayapp"><img src="https://github.com/railwayapp.png" width="80px" alt="User avatar: railwayapp" /></a> <a href="https://github.com/caseyamcl"><img src="https://github.com/caseyamcl.png" width="80px" alt="User avatar: caseyamcl" /></a> <a href="https://github.com/bytebase"><img src="https://github.com/bytebase.png" width="80px" alt="User avatar: bytebase" /></a> <a href="https://github.com/"><img src="https://raw.githubusercontent.com/JamesIves/github-sponsors-readme-action/dev/.github/assets/placeholder.png" width="80px" alt="User avatar: " /></a> <!-- sponsors-premium -->
|
<!-- sponsors-premium --><a href="https://github.com/MVST-Solutions"><img src="https://github.com/MVST-Solutions.png" width="80px" alt="User avatar: MVST-Solutions" /></a> <a href="https://github.com/dharsanb"><img src="https://github.com/dharsanb.png" width="80px" alt="User avatar: dharsanb" /></a> <a href="https://github.com/railwayapp"><img src="https://github.com/railwayapp.png" width="80px" alt="User avatar: railwayapp" /></a> <a href="https://github.com/caseyamcl"><img src="https://github.com/caseyamcl.png" width="80px" alt="User avatar: caseyamcl" /></a> <a href="https://github.com/bytebase"><img src="https://github.com/bytebase.png" width="80px" alt="User avatar: bytebase" /></a> <a href="https://github.com/"><img src="https://raw.githubusercontent.com/JamesIves/github-sponsors-readme-action/dev/.github/assets/placeholder.png" width="80px" alt="User avatar: " /></a> <!-- sponsors-premium -->
|
||||||
</p>
|
</p>
|
||||||
<p align="center">
|
<p align="center">
|
||||||
<!-- sponsors-base --><a href="https://github.com/seanwash"><img src="https://github.com/seanwash.png" width="50px" alt="User avatar: seanwash" /></a> <a href="https://github.com/jerath"><img src="https://github.com/jerath.png" width="50px" alt="User avatar: jerath" /></a> <a href="https://github.com/itsa-sh"><img src="https://github.com/itsa-sh.png" width="50px" alt="User avatar: itsa-sh" /></a> <a href="https://github.com/dmmulroy"><img src="https://github.com/dmmulroy.png" width="50px" alt="User avatar: dmmulroy" /></a> <a href="https://github.com/timcole"><img src="https://github.com/timcole.png" width="50px" alt="User avatar: timcole" /></a> <a href="https://github.com/VLZH"><img src="https://github.com/VLZH.png" width="50px" alt="User avatar: VLZH" /></a> <a href="https://github.com/terasaka2k"><img src="https://github.com/terasaka2k.png" width="50px" alt="User avatar: terasaka2k" /></a> <a href="https://github.com/andriyor"><img src="https://github.com/andriyor.png" width="50px" alt="User avatar: andriyor" /></a> <a href="https://github.com/majudhu"><img src="https://github.com/majudhu.png" width="50px" alt="User avatar: majudhu" /></a> <a href="https://github.com/axelrindle"><img src="https://github.com/axelrindle.png" width="50px" alt="User avatar: axelrindle" /></a> <a href="https://github.com/jirizverina"><img src="https://github.com/jirizverina.png" width="50px" alt="User avatar: jirizverina" /></a> <a href="https://github.com/chip-well"><img src="https://github.com/chip-well.png" width="50px" alt="User avatar: chip-well" /></a> <a href="https://github.com/GRAYAH"><img src="https://github.com/GRAYAH.png" width="50px" alt="User avatar: GRAYAH" /></a> <a href="https://github.com/flashblaze"><img src="https://github.com/flashblaze.png" width="50px" alt="User avatar: flashblaze" /></a> <!-- sponsors-base -->
|
<!-- sponsors-base --><a href="https://github.com/seanwash"><img src="https://github.com/seanwash.png" width="50px" alt="User avatar: seanwash" /></a> <a href="https://github.com/jerath"><img src="https://github.com/jerath.png" width="50px" alt="User avatar: jerath" /></a> <a href="https://github.com/itsa-sh"><img src="https://github.com/itsa-sh.png" width="50px" alt="User avatar: itsa-sh" /></a> <a href="https://github.com/dmmulroy"><img src="https://github.com/dmmulroy.png" width="50px" alt="User avatar: dmmulroy" /></a> <a href="https://github.com/timcole"><img src="https://github.com/timcole.png" width="50px" alt="User avatar: timcole" /></a> <a href="https://github.com/VLZH"><img src="https://github.com/VLZH.png" width="50px" alt="User avatar: VLZH" /></a> <a href="https://github.com/terasaka2k"><img src="https://github.com/terasaka2k.png" width="50px" alt="User avatar: terasaka2k" /></a> <a href="https://github.com/andriyor"><img src="https://github.com/andriyor.png" width="50px" alt="User avatar: andriyor" /></a> <a href="https://github.com/majudhu"><img src="https://github.com/majudhu.png" width="50px" alt="User avatar: majudhu" /></a> <a href="https://github.com/axelrindle"><img src="https://github.com/axelrindle.png" width="50px" alt="User avatar: axelrindle" /></a> <a href="https://github.com/jirizverina"><img src="https://github.com/jirizverina.png" width="50px" alt="User avatar: jirizverina" /></a> <a href="https://github.com/chip-well"><img src="https://github.com/chip-well.png" width="50px" alt="User avatar: chip-well" /></a> <a href="https://github.com/GRAYAH"><img src="https://github.com/GRAYAH.png" width="50px" alt="User avatar: GRAYAH" /></a> <a href="https://github.com/flashblaze"><img src="https://github.com/flashblaze.png" width="50px" alt="User avatar: flashblaze" /></a> <a href="https://github.com/Frostist"><img src="https://github.com/Frostist.png" width="50px" alt="User avatar: Frostist" /></a> <!-- sponsors-base -->
|
||||||
</p>
|
</p>
|
||||||
|
|
||||||

|

|
||||||
|
|||||||
@@ -5,19 +5,32 @@ edition = "2024"
|
|||||||
publish = false
|
publish = false
|
||||||
|
|
||||||
[[bin]]
|
[[bin]]
|
||||||
name = "yaakcli"
|
name = "yaak"
|
||||||
path = "src/main.rs"
|
path = "src/main.rs"
|
||||||
|
|
||||||
[dependencies]
|
[dependencies]
|
||||||
|
base64 = "0.22"
|
||||||
clap = { version = "4", features = ["derive"] }
|
clap = { version = "4", features = ["derive"] }
|
||||||
|
console = "0.15"
|
||||||
dirs = "6"
|
dirs = "6"
|
||||||
env_logger = "0.11"
|
env_logger = "0.11"
|
||||||
futures = "0.3"
|
futures = "0.3"
|
||||||
|
hex = { workspace = true }
|
||||||
|
include_dir = "0.7"
|
||||||
|
keyring = { workspace = true, features = ["apple-native", "windows-native", "sync-secret-service"] }
|
||||||
log = { workspace = true }
|
log = { workspace = true }
|
||||||
|
rand = "0.8"
|
||||||
|
reqwest = { workspace = true }
|
||||||
|
rolldown = "0.1.0"
|
||||||
|
oxc_resolver = "=11.10.0"
|
||||||
schemars = { workspace = true }
|
schemars = { workspace = true }
|
||||||
serde = { workspace = true }
|
serde = { workspace = true }
|
||||||
serde_json = { workspace = true }
|
serde_json = { workspace = true }
|
||||||
tokio = { workspace = true, features = ["rt-multi-thread", "macros"] }
|
sha2 = { workspace = true }
|
||||||
|
tokio = { workspace = true, features = ["rt-multi-thread", "macros", "io-util", "net", "signal", "time"] }
|
||||||
|
walkdir = "2"
|
||||||
|
webbrowser = "1"
|
||||||
|
zip = "4"
|
||||||
yaak = { workspace = true }
|
yaak = { workspace = true }
|
||||||
yaak-crypto = { workspace = true }
|
yaak-crypto = { workspace = true }
|
||||||
yaak-http = { workspace = true }
|
yaak-http = { workspace = true }
|
||||||
|
|||||||
@@ -1,87 +1,66 @@
|
|||||||
# yaak-cli
|
# Yaak CLI
|
||||||
|
|
||||||
Command-line interface for Yaak.
|
The `yaak` CLI for publishing plugins and creating/updating/sending requests.
|
||||||
|
|
||||||
## Command Overview
|
## Installation
|
||||||
|
|
||||||
Current top-level commands:
|
```sh
|
||||||
|
npm install @yaakapp/cli
|
||||||
|
```
|
||||||
|
|
||||||
|
## Agentic Workflows
|
||||||
|
|
||||||
|
The `yaak` CLI is primarily meant to be used by AI agents, and has the following features:
|
||||||
|
|
||||||
|
- `schema` subcommands to get the JSON Schema for any model (eg. `yaak request schema http`)
|
||||||
|
- `--json '{...}'` input format to create and update data
|
||||||
|
- `--verbose` mode for extracting debug info while sending requests
|
||||||
|
- The ability to send entire workspaces and folders (Supports `--parallel` and `--fail-fast`)
|
||||||
|
|
||||||
|
### Example Prompts
|
||||||
|
|
||||||
|
Use the `yaak` CLI with agents like Claude or Codex to do useful things for you.
|
||||||
|
|
||||||
|
Here are some example prompts:
|
||||||
|
|
||||||
```text
|
```text
|
||||||
yaakcli send <request_id>
|
Scan my API routes and create a workspace (using yaak cli) with
|
||||||
yaakcli workspace list
|
all the requests needed for me to do manual testing?
|
||||||
yaakcli workspace show <workspace_id>
|
|
||||||
yaakcli workspace create --name <name>
|
|
||||||
yaakcli workspace create --json '{"name":"My Workspace"}'
|
|
||||||
yaakcli workspace create '{"name":"My Workspace"}'
|
|
||||||
yaakcli workspace update --json '{"id":"wk_abc","description":"Updated"}'
|
|
||||||
yaakcli workspace delete <workspace_id> [--yes]
|
|
||||||
yaakcli request list <workspace_id>
|
|
||||||
yaakcli request show <request_id>
|
|
||||||
yaakcli request send <request_id>
|
|
||||||
yaakcli request create <workspace_id> --name <name> --url <url> [--method GET]
|
|
||||||
yaakcli request create --json '{"workspaceId":"wk_abc","name":"Users","url":"https://api.example.com/users"}'
|
|
||||||
yaakcli request create '{"workspaceId":"wk_abc","name":"Users","url":"https://api.example.com/users"}'
|
|
||||||
yaakcli request update --json '{"id":"rq_abc","name":"Users v2"}'
|
|
||||||
yaakcli request delete <request_id> [--yes]
|
|
||||||
yaakcli folder list <workspace_id>
|
|
||||||
yaakcli folder show <folder_id>
|
|
||||||
yaakcli folder create <workspace_id> --name <name>
|
|
||||||
yaakcli folder create --json '{"workspaceId":"wk_abc","name":"Auth"}'
|
|
||||||
yaakcli folder create '{"workspaceId":"wk_abc","name":"Auth"}'
|
|
||||||
yaakcli folder update --json '{"id":"fl_abc","name":"Auth v2"}'
|
|
||||||
yaakcli folder delete <folder_id> [--yes]
|
|
||||||
yaakcli environment list <workspace_id>
|
|
||||||
yaakcli environment show <environment_id>
|
|
||||||
yaakcli environment create <workspace_id> --name <name>
|
|
||||||
yaakcli environment create --json '{"workspaceId":"wk_abc","name":"Production"}'
|
|
||||||
yaakcli environment create '{"workspaceId":"wk_abc","name":"Production"}'
|
|
||||||
yaakcli environment update --json '{"id":"ev_abc","color":"#00ff00"}'
|
|
||||||
yaakcli environment delete <environment_id> [--yes]
|
|
||||||
```
|
```
|
||||||
|
|
||||||
Global options:
|
```text
|
||||||
|
Send all the GraphQL requests in my workspace
|
||||||
- `--data-dir <path>`: use a custom data directory
|
|
||||||
- `-e, --environment <id>`: environment to use during request rendering/sending
|
|
||||||
- `-v, --verbose`: verbose logging and send output
|
|
||||||
|
|
||||||
Notes:
|
|
||||||
|
|
||||||
- `send` is currently a shortcut for sending an HTTP request ID.
|
|
||||||
- `delete` commands prompt for confirmation unless `--yes` is provided.
|
|
||||||
- In non-interactive mode, `delete` commands require `--yes`.
|
|
||||||
- `create` and `update` commands support `--json` and positional JSON shorthand.
|
|
||||||
- `update` uses JSON Merge Patch semantics (RFC 7386) for partial updates.
|
|
||||||
|
|
||||||
## Examples
|
|
||||||
|
|
||||||
```bash
|
|
||||||
yaakcli workspace list
|
|
||||||
yaakcli workspace create --name "My Workspace"
|
|
||||||
yaakcli workspace show wk_abc
|
|
||||||
yaakcli workspace update --json '{"id":"wk_abc","description":"Team workspace"}'
|
|
||||||
yaakcli request list wk_abc
|
|
||||||
yaakcli request show rq_abc
|
|
||||||
yaakcli request create wk_abc --name "Users" --url "https://api.example.com/users"
|
|
||||||
yaakcli request update --json '{"id":"rq_abc","name":"Users v2"}'
|
|
||||||
yaakcli request send rq_abc -e ev_abc
|
|
||||||
yaakcli request delete rq_abc --yes
|
|
||||||
yaakcli folder create wk_abc --name "Auth"
|
|
||||||
yaakcli folder update --json '{"id":"fl_abc","name":"Auth v2"}'
|
|
||||||
yaakcli environment create wk_abc --name "Production"
|
|
||||||
yaakcli environment update --json '{"id":"ev_abc","color":"#00ff00"}'
|
|
||||||
```
|
```
|
||||||
|
|
||||||
## Roadmap
|
## Description
|
||||||
|
|
||||||
Planned command expansion (request schema and polymorphic send) is tracked in `PLAN.md`.
|
Here's the current print of `yaak --help`
|
||||||
|
|
||||||
When command behavior changes, update this README and verify with:
|
```text
|
||||||
|
Yaak CLI - API client from the command line
|
||||||
|
|
||||||
```bash
|
Usage: yaak [OPTIONS] <COMMAND>
|
||||||
cargo run -q -p yaak-cli -- --help
|
|
||||||
cargo run -q -p yaak-cli -- request --help
|
Commands:
|
||||||
cargo run -q -p yaak-cli -- workspace --help
|
auth Authentication commands
|
||||||
cargo run -q -p yaak-cli -- folder --help
|
plugin Plugin development and publishing commands
|
||||||
cargo run -q -p yaak-cli -- environment --help
|
send Send a request, folder, or workspace by ID
|
||||||
|
workspace Workspace commands
|
||||||
|
request Request commands
|
||||||
|
folder Folder commands
|
||||||
|
environment Environment commands
|
||||||
|
|
||||||
|
Options:
|
||||||
|
--data-dir <DATA_DIR> Use a custom data directory
|
||||||
|
-e, --environment <ENVIRONMENT> Environment ID to use for variable substitution
|
||||||
|
-v, --verbose Enable verbose send output (events and streamed response body)
|
||||||
|
--log [<LEVEL>] Enable CLI logging; optionally set level (error|warn|info|debug|trace) [possible values: error, warn, info, debug, trace]
|
||||||
|
-h, --help Print help
|
||||||
|
-V, --version Print version
|
||||||
|
|
||||||
|
Agent Hints:
|
||||||
|
- Template variable syntax is ${[ my_var ]}, not {{ ... }}
|
||||||
|
- Template function syntax is ${[ namespace.my_func(a='aaa',b='bbb') ]}
|
||||||
|
- View JSONSchema for models before creating or updating (eg. `yaak request schema http`)
|
||||||
|
- Deletion requires confirmation (--yes for non-interactive environments)
|
||||||
```
|
```
|
||||||
|
|||||||
@@ -2,8 +2,16 @@ use clap::{Args, Parser, Subcommand, ValueEnum};
|
|||||||
use std::path::PathBuf;
|
use std::path::PathBuf;
|
||||||
|
|
||||||
#[derive(Parser)]
|
#[derive(Parser)]
|
||||||
#[command(name = "yaakcli")]
|
#[command(name = "yaak")]
|
||||||
#[command(about = "Yaak CLI - API client from the command line")]
|
#[command(about = "Yaak CLI - API client from the command line")]
|
||||||
|
#[command(version = crate::version::cli_version())]
|
||||||
|
#[command(disable_help_subcommand = true)]
|
||||||
|
#[command(after_help = r#"Agent Hints:
|
||||||
|
- Template variable syntax is ${[ my_var ]}, not {{ ... }}
|
||||||
|
- Template function syntax is ${[ namespace.my_func(a='aaa',b='bbb') ]}
|
||||||
|
- View JSONSchema for models before creating or updating (eg. `yaak request schema http`)
|
||||||
|
- Deletion requires confirmation (--yes for non-interactive environments)
|
||||||
|
"#)]
|
||||||
pub struct Cli {
|
pub struct Cli {
|
||||||
/// Use a custom data directory
|
/// Use a custom data directory
|
||||||
#[arg(long, global = true)]
|
#[arg(long, global = true)]
|
||||||
@@ -13,16 +21,32 @@ pub struct Cli {
|
|||||||
#[arg(long, short, global = true)]
|
#[arg(long, short, global = true)]
|
||||||
pub environment: Option<String>,
|
pub environment: Option<String>,
|
||||||
|
|
||||||
/// Enable verbose logging
|
/// Enable verbose send output (events and streamed response body)
|
||||||
#[arg(long, short, global = true)]
|
#[arg(long, short, global = true)]
|
||||||
pub verbose: bool,
|
pub verbose: bool,
|
||||||
|
|
||||||
|
/// Enable CLI logging; optionally set level (error|warn|info|debug|trace)
|
||||||
|
#[arg(long, global = true, value_name = "LEVEL", num_args = 0..=1, ignore_case = true)]
|
||||||
|
pub log: Option<Option<LogLevel>>,
|
||||||
|
|
||||||
#[command(subcommand)]
|
#[command(subcommand)]
|
||||||
pub command: Commands,
|
pub command: Commands,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Subcommand)]
|
#[derive(Subcommand)]
|
||||||
pub enum Commands {
|
pub enum Commands {
|
||||||
|
/// Authentication commands
|
||||||
|
Auth(AuthArgs),
|
||||||
|
|
||||||
|
/// Plugin development and publishing commands
|
||||||
|
Plugin(PluginArgs),
|
||||||
|
|
||||||
|
#[command(hide = true)]
|
||||||
|
Build(PluginPathArg),
|
||||||
|
|
||||||
|
#[command(hide = true)]
|
||||||
|
Dev(PluginPathArg),
|
||||||
|
|
||||||
/// Send a request, folder, or workspace by ID
|
/// Send a request, folder, or workspace by ID
|
||||||
Send(SendArgs),
|
Send(SendArgs),
|
||||||
|
|
||||||
@@ -44,12 +68,8 @@ pub struct SendArgs {
|
|||||||
/// Request, folder, or workspace ID
|
/// Request, folder, or workspace ID
|
||||||
pub id: String,
|
pub id: String,
|
||||||
|
|
||||||
/// Execute requests sequentially (default)
|
|
||||||
#[arg(long, conflicts_with = "parallel")]
|
|
||||||
pub sequential: bool,
|
|
||||||
|
|
||||||
/// Execute requests in parallel
|
/// Execute requests in parallel
|
||||||
#[arg(long, conflicts_with = "sequential")]
|
#[arg(long)]
|
||||||
pub parallel: bool,
|
pub parallel: bool,
|
||||||
|
|
||||||
/// Stop on first request failure when sending folders/workspaces
|
/// Stop on first request failure when sending folders/workspaces
|
||||||
@@ -58,6 +78,7 @@ pub struct SendArgs {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Args)]
|
#[derive(Args)]
|
||||||
|
#[command(disable_help_subcommand = true)]
|
||||||
pub struct WorkspaceArgs {
|
pub struct WorkspaceArgs {
|
||||||
#[command(subcommand)]
|
#[command(subcommand)]
|
||||||
pub command: WorkspaceCommands,
|
pub command: WorkspaceCommands,
|
||||||
@@ -68,6 +89,13 @@ pub enum WorkspaceCommands {
|
|||||||
/// List all workspaces
|
/// List all workspaces
|
||||||
List,
|
List,
|
||||||
|
|
||||||
|
/// Output JSON schema for workspace create/update payloads
|
||||||
|
Schema {
|
||||||
|
/// Pretty-print schema JSON output
|
||||||
|
#[arg(long)]
|
||||||
|
pretty: bool,
|
||||||
|
},
|
||||||
|
|
||||||
/// Show a workspace as JSON
|
/// Show a workspace as JSON
|
||||||
Show {
|
Show {
|
||||||
/// Workspace ID
|
/// Workspace ID
|
||||||
@@ -112,6 +140,7 @@ pub enum WorkspaceCommands {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Args)]
|
#[derive(Args)]
|
||||||
|
#[command(disable_help_subcommand = true)]
|
||||||
pub struct RequestArgs {
|
pub struct RequestArgs {
|
||||||
#[command(subcommand)]
|
#[command(subcommand)]
|
||||||
pub command: RequestCommands,
|
pub command: RequestCommands,
|
||||||
@@ -141,6 +170,10 @@ pub enum RequestCommands {
|
|||||||
Schema {
|
Schema {
|
||||||
#[arg(value_enum)]
|
#[arg(value_enum)]
|
||||||
request_type: RequestSchemaType,
|
request_type: RequestSchemaType,
|
||||||
|
|
||||||
|
/// Pretty-print schema JSON output
|
||||||
|
#[arg(long)]
|
||||||
|
pretty: bool,
|
||||||
},
|
},
|
||||||
|
|
||||||
/// Create a new HTTP request
|
/// Create a new HTTP request
|
||||||
@@ -194,7 +227,29 @@ pub enum RequestSchemaType {
|
|||||||
Websocket,
|
Websocket,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Clone, Copy, Debug, ValueEnum)]
|
||||||
|
pub enum LogLevel {
|
||||||
|
Error,
|
||||||
|
Warn,
|
||||||
|
Info,
|
||||||
|
Debug,
|
||||||
|
Trace,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl LogLevel {
|
||||||
|
pub fn as_filter(self) -> log::LevelFilter {
|
||||||
|
match self {
|
||||||
|
LogLevel::Error => log::LevelFilter::Error,
|
||||||
|
LogLevel::Warn => log::LevelFilter::Warn,
|
||||||
|
LogLevel::Info => log::LevelFilter::Info,
|
||||||
|
LogLevel::Debug => log::LevelFilter::Debug,
|
||||||
|
LogLevel::Trace => log::LevelFilter::Trace,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
#[derive(Args)]
|
#[derive(Args)]
|
||||||
|
#[command(disable_help_subcommand = true)]
|
||||||
pub struct FolderArgs {
|
pub struct FolderArgs {
|
||||||
#[command(subcommand)]
|
#[command(subcommand)]
|
||||||
pub command: FolderCommands,
|
pub command: FolderCommands,
|
||||||
@@ -251,6 +306,7 @@ pub enum FolderCommands {
|
|||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Args)]
|
#[derive(Args)]
|
||||||
|
#[command(disable_help_subcommand = true)]
|
||||||
pub struct EnvironmentArgs {
|
pub struct EnvironmentArgs {
|
||||||
#[command(subcommand)]
|
#[command(subcommand)]
|
||||||
pub command: EnvironmentCommands,
|
pub command: EnvironmentCommands,
|
||||||
@@ -264,6 +320,13 @@ pub enum EnvironmentCommands {
|
|||||||
workspace_id: String,
|
workspace_id: String,
|
||||||
},
|
},
|
||||||
|
|
||||||
|
/// Output JSON schema for environment create/update payloads
|
||||||
|
Schema {
|
||||||
|
/// Pretty-print schema JSON output
|
||||||
|
#[arg(long)]
|
||||||
|
pretty: bool,
|
||||||
|
},
|
||||||
|
|
||||||
/// Show an environment as JSON
|
/// Show an environment as JSON
|
||||||
Show {
|
Show {
|
||||||
/// Environment ID
|
/// Environment ID
|
||||||
@@ -271,15 +334,22 @@ pub enum EnvironmentCommands {
|
|||||||
},
|
},
|
||||||
|
|
||||||
/// Create an environment
|
/// Create an environment
|
||||||
|
#[command(after_help = r#"Modes (choose one):
|
||||||
|
1) yaak environment create <workspace_id> --name <name>
|
||||||
|
2) yaak environment create --json '{"workspaceId":"wk_abc","name":"Production"}'
|
||||||
|
3) yaak environment create '{"workspaceId":"wk_abc","name":"Production"}'
|
||||||
|
4) yaak environment create <workspace_id> --json '{"name":"Production"}'
|
||||||
|
"#)]
|
||||||
Create {
|
Create {
|
||||||
/// Workspace ID (or positional JSON payload shorthand)
|
/// Workspace ID for flag-based mode, or positional JSON payload shorthand
|
||||||
|
#[arg(value_name = "WORKSPACE_ID_OR_JSON")]
|
||||||
workspace_id: Option<String>,
|
workspace_id: Option<String>,
|
||||||
|
|
||||||
/// Environment name
|
/// Environment name
|
||||||
#[arg(short, long)]
|
#[arg(short, long)]
|
||||||
name: Option<String>,
|
name: Option<String>,
|
||||||
|
|
||||||
/// JSON payload
|
/// JSON payload (use instead of WORKSPACE_ID/--name)
|
||||||
#[arg(long)]
|
#[arg(long)]
|
||||||
json: Option<String>,
|
json: Option<String>,
|
||||||
},
|
},
|
||||||
@@ -305,3 +375,61 @@ pub enum EnvironmentCommands {
|
|||||||
yes: bool,
|
yes: bool,
|
||||||
},
|
},
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[derive(Args)]
|
||||||
|
#[command(disable_help_subcommand = true)]
|
||||||
|
pub struct AuthArgs {
|
||||||
|
#[command(subcommand)]
|
||||||
|
pub command: AuthCommands,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Subcommand)]
|
||||||
|
pub enum AuthCommands {
|
||||||
|
/// Login to Yaak via web browser
|
||||||
|
Login,
|
||||||
|
|
||||||
|
/// Sign out of the Yaak CLI
|
||||||
|
Logout,
|
||||||
|
|
||||||
|
/// Print the current logged-in user's info
|
||||||
|
Whoami,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Args)]
|
||||||
|
#[command(disable_help_subcommand = true)]
|
||||||
|
pub struct PluginArgs {
|
||||||
|
#[command(subcommand)]
|
||||||
|
pub command: PluginCommands,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Subcommand)]
|
||||||
|
pub enum PluginCommands {
|
||||||
|
/// Transpile code into a runnable plugin bundle
|
||||||
|
Build(PluginPathArg),
|
||||||
|
|
||||||
|
/// Build plugin bundle continuously when the filesystem changes
|
||||||
|
Dev(PluginPathArg),
|
||||||
|
|
||||||
|
/// Generate a "Hello World" Yaak plugin
|
||||||
|
Generate(GenerateArgs),
|
||||||
|
|
||||||
|
/// Publish a Yaak plugin version to the plugin registry
|
||||||
|
Publish(PluginPathArg),
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Args, Clone)]
|
||||||
|
pub struct PluginPathArg {
|
||||||
|
/// Path to plugin directory (defaults to current working directory)
|
||||||
|
pub path: Option<PathBuf>,
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Args, Clone)]
|
||||||
|
pub struct GenerateArgs {
|
||||||
|
/// Plugin name (defaults to a generated name in interactive mode)
|
||||||
|
#[arg(long)]
|
||||||
|
pub name: Option<String>,
|
||||||
|
|
||||||
|
/// Output directory for the generated plugin (defaults to ./<name> in interactive mode)
|
||||||
|
#[arg(long)]
|
||||||
|
pub dir: Option<PathBuf>,
|
||||||
|
}
|
||||||
|
|||||||
528
crates-cli/yaak-cli/src/commands/auth.rs
Normal file
528
crates-cli/yaak-cli/src/commands/auth.rs
Normal file
@@ -0,0 +1,528 @@
|
|||||||
|
use crate::cli::{AuthArgs, AuthCommands};
|
||||||
|
use crate::ui;
|
||||||
|
use crate::utils::http;
|
||||||
|
use base64::Engine as _;
|
||||||
|
use keyring::Entry;
|
||||||
|
use rand::RngCore;
|
||||||
|
use rand::rngs::OsRng;
|
||||||
|
use reqwest::Url;
|
||||||
|
use serde_json::Value;
|
||||||
|
use sha2::{Digest, Sha256};
|
||||||
|
use std::io::{self, IsTerminal, Write};
|
||||||
|
use std::time::Duration;
|
||||||
|
use tokio::io::{AsyncReadExt, AsyncWriteExt};
|
||||||
|
use tokio::net::{TcpListener, TcpStream};
|
||||||
|
|
||||||
|
const OAUTH_CLIENT_ID: &str = "a1fe44800c2d7e803cad1b4bf07a291c";
|
||||||
|
const KEYRING_USER: &str = "yaak";
|
||||||
|
const AUTH_TIMEOUT: Duration = Duration::from_secs(300);
|
||||||
|
const MAX_REQUEST_BYTES: usize = 16 * 1024;
|
||||||
|
|
||||||
|
type CommandResult<T = ()> = std::result::Result<T, String>;
|
||||||
|
|
||||||
|
#[derive(Clone, Copy, Debug, Eq, PartialEq)]
|
||||||
|
enum Environment {
|
||||||
|
Production,
|
||||||
|
Staging,
|
||||||
|
Development,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Environment {
|
||||||
|
fn app_base_url(self) -> &'static str {
|
||||||
|
match self {
|
||||||
|
Environment::Production => "https://yaak.app",
|
||||||
|
Environment::Staging => "https://todo.yaak.app",
|
||||||
|
Environment::Development => "http://localhost:9444",
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn api_base_url(self) -> &'static str {
|
||||||
|
match self {
|
||||||
|
Environment::Production => "https://api.yaak.app",
|
||||||
|
Environment::Staging => "https://todo.yaak.app",
|
||||||
|
Environment::Development => "http://localhost:9444",
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn keyring_service(self) -> &'static str {
|
||||||
|
match self {
|
||||||
|
Environment::Production => "app.yaak.cli.Token",
|
||||||
|
Environment::Staging => "app.yaak.cli.staging.Token",
|
||||||
|
Environment::Development => "app.yaak.cli.dev.Token",
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
struct OAuthFlow {
|
||||||
|
app_base_url: String,
|
||||||
|
auth_url: Url,
|
||||||
|
token_url: String,
|
||||||
|
redirect_url: String,
|
||||||
|
state: String,
|
||||||
|
code_verifier: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn run(args: AuthArgs) -> i32 {
|
||||||
|
let result = match args.command {
|
||||||
|
AuthCommands::Login => login().await,
|
||||||
|
AuthCommands::Logout => logout(),
|
||||||
|
AuthCommands::Whoami => whoami().await,
|
||||||
|
};
|
||||||
|
|
||||||
|
match result {
|
||||||
|
Ok(()) => 0,
|
||||||
|
Err(error) => {
|
||||||
|
ui::error(&error);
|
||||||
|
1
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn login() -> CommandResult {
|
||||||
|
let environment = current_environment();
|
||||||
|
|
||||||
|
let listener = TcpListener::bind("127.0.0.1:0")
|
||||||
|
.await
|
||||||
|
.map_err(|e| format!("Failed to start OAuth callback server: {e}"))?;
|
||||||
|
let port = listener
|
||||||
|
.local_addr()
|
||||||
|
.map_err(|e| format!("Failed to determine callback server port: {e}"))?
|
||||||
|
.port();
|
||||||
|
|
||||||
|
let oauth = build_oauth_flow(environment, port)?;
|
||||||
|
|
||||||
|
ui::info(&format!("Initiating login to {}", oauth.auth_url));
|
||||||
|
if !confirm_open_browser()? {
|
||||||
|
ui::info("Login canceled");
|
||||||
|
return Ok(());
|
||||||
|
}
|
||||||
|
|
||||||
|
if let Err(err) = webbrowser::open(oauth.auth_url.as_ref()) {
|
||||||
|
ui::warning(&format!("Failed to open browser: {err}"));
|
||||||
|
ui::info(&format!("Open this URL manually:\n{}", oauth.auth_url));
|
||||||
|
}
|
||||||
|
ui::info("Waiting for authentication...");
|
||||||
|
|
||||||
|
let code = tokio::select! {
|
||||||
|
result = receive_oauth_code(listener, &oauth.state, &oauth.app_base_url) => result?,
|
||||||
|
_ = tokio::signal::ctrl_c() => {
|
||||||
|
return Err("Interrupted by user".to_string());
|
||||||
|
}
|
||||||
|
_ = tokio::time::sleep(AUTH_TIMEOUT) => {
|
||||||
|
return Err("Timeout waiting for authentication".to_string());
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
let token = exchange_access_token(&oauth, &code).await?;
|
||||||
|
store_auth_token(environment, &token)?;
|
||||||
|
ui::success("Authentication successful!");
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn logout() -> CommandResult {
|
||||||
|
delete_auth_token(current_environment())?;
|
||||||
|
ui::success("Signed out of Yaak");
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn whoami() -> CommandResult {
|
||||||
|
let environment = current_environment();
|
||||||
|
let token = match get_auth_token(environment)? {
|
||||||
|
Some(token) => token,
|
||||||
|
None => {
|
||||||
|
ui::warning("Not logged in");
|
||||||
|
ui::info("Please run `yaak auth login`");
|
||||||
|
return Ok(());
|
||||||
|
}
|
||||||
|
};
|
||||||
|
|
||||||
|
let url = format!("{}/api/v1/whoami", environment.api_base_url());
|
||||||
|
let response = http::build_client(Some(&token))?
|
||||||
|
.get(url)
|
||||||
|
.send()
|
||||||
|
.await
|
||||||
|
.map_err(|e| format!("Failed to call whoami endpoint: {e}"))?;
|
||||||
|
|
||||||
|
let status = response.status();
|
||||||
|
let body =
|
||||||
|
response.text().await.map_err(|e| format!("Failed to read whoami response body: {e}"))?;
|
||||||
|
|
||||||
|
if !status.is_success() {
|
||||||
|
if status.as_u16() == 401 {
|
||||||
|
let _ = delete_auth_token(environment);
|
||||||
|
return Err(
|
||||||
|
"Unauthorized to access CLI. Run `yaak auth login` to refresh credentials."
|
||||||
|
.to_string(),
|
||||||
|
);
|
||||||
|
}
|
||||||
|
return Err(http::parse_api_error(status.as_u16(), &body));
|
||||||
|
}
|
||||||
|
|
||||||
|
println!("{body}");
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn current_environment() -> Environment {
|
||||||
|
let value = std::env::var("ENVIRONMENT").ok();
|
||||||
|
parse_environment(value.as_deref())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn parse_environment(value: Option<&str>) -> Environment {
|
||||||
|
match value {
|
||||||
|
Some("staging") => Environment::Staging,
|
||||||
|
Some("development") => Environment::Development,
|
||||||
|
_ => Environment::Production,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn build_oauth_flow(environment: Environment, callback_port: u16) -> CommandResult<OAuthFlow> {
|
||||||
|
let code_verifier = random_hex(32);
|
||||||
|
let state = random_hex(24);
|
||||||
|
let redirect_url = format!("http://127.0.0.1:{callback_port}/oauth/callback");
|
||||||
|
|
||||||
|
let code_challenge = base64::engine::general_purpose::URL_SAFE_NO_PAD
|
||||||
|
.encode(Sha256::digest(code_verifier.as_bytes()));
|
||||||
|
|
||||||
|
let mut auth_url = Url::parse(&format!("{}/login/oauth/authorize", environment.app_base_url()))
|
||||||
|
.map_err(|e| format!("Failed to build OAuth authorize URL: {e}"))?;
|
||||||
|
auth_url
|
||||||
|
.query_pairs_mut()
|
||||||
|
.append_pair("response_type", "code")
|
||||||
|
.append_pair("client_id", OAUTH_CLIENT_ID)
|
||||||
|
.append_pair("redirect_uri", &redirect_url)
|
||||||
|
.append_pair("state", &state)
|
||||||
|
.append_pair("code_challenge_method", "S256")
|
||||||
|
.append_pair("code_challenge", &code_challenge);
|
||||||
|
|
||||||
|
Ok(OAuthFlow {
|
||||||
|
app_base_url: environment.app_base_url().to_string(),
|
||||||
|
auth_url,
|
||||||
|
token_url: format!("{}/login/oauth/access_token", environment.app_base_url()),
|
||||||
|
redirect_url,
|
||||||
|
state,
|
||||||
|
code_verifier,
|
||||||
|
})
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn receive_oauth_code(
|
||||||
|
listener: TcpListener,
|
||||||
|
expected_state: &str,
|
||||||
|
app_base_url: &str,
|
||||||
|
) -> CommandResult<String> {
|
||||||
|
loop {
|
||||||
|
let (mut stream, _) = listener
|
||||||
|
.accept()
|
||||||
|
.await
|
||||||
|
.map_err(|e| format!("OAuth callback server accept error: {e}"))?;
|
||||||
|
|
||||||
|
match parse_callback_request(&mut stream).await {
|
||||||
|
Ok((state, code)) => {
|
||||||
|
if state != expected_state {
|
||||||
|
let _ = write_bad_request(&mut stream, "Invalid OAuth state").await;
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
let success_redirect = format!("{app_base_url}/login/oauth/success");
|
||||||
|
write_redirect(&mut stream, &success_redirect)
|
||||||
|
.await
|
||||||
|
.map_err(|e| format!("Failed responding to OAuth callback: {e}"))?;
|
||||||
|
return Ok(code);
|
||||||
|
}
|
||||||
|
Err(error) => {
|
||||||
|
let _ = write_bad_request(&mut stream, &error).await;
|
||||||
|
if error.starts_with("OAuth provider returned error:") {
|
||||||
|
return Err(error);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn parse_callback_request(stream: &mut TcpStream) -> CommandResult<(String, String)> {
|
||||||
|
let target = read_http_target(stream).await?;
|
||||||
|
if !target.starts_with("/oauth/callback") {
|
||||||
|
return Err("Expected /oauth/callback path".to_string());
|
||||||
|
}
|
||||||
|
|
||||||
|
let url = Url::parse(&format!("http://127.0.0.1{target}"))
|
||||||
|
.map_err(|e| format!("Failed to parse callback URL: {e}"))?;
|
||||||
|
let mut state: Option<String> = None;
|
||||||
|
let mut code: Option<String> = None;
|
||||||
|
let mut oauth_error: Option<String> = None;
|
||||||
|
let mut oauth_error_description: Option<String> = None;
|
||||||
|
|
||||||
|
for (k, v) in url.query_pairs() {
|
||||||
|
if k == "state" {
|
||||||
|
state = Some(v.into_owned());
|
||||||
|
} else if k == "code" {
|
||||||
|
code = Some(v.into_owned());
|
||||||
|
} else if k == "error" {
|
||||||
|
oauth_error = Some(v.into_owned());
|
||||||
|
} else if k == "error_description" {
|
||||||
|
oauth_error_description = Some(v.into_owned());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if let Some(error) = oauth_error {
|
||||||
|
let mut message = format!("OAuth provider returned error: {error}");
|
||||||
|
if let Some(description) = oauth_error_description.filter(|d| !d.is_empty()) {
|
||||||
|
message.push_str(&format!(" ({description})"));
|
||||||
|
}
|
||||||
|
return Err(message);
|
||||||
|
}
|
||||||
|
|
||||||
|
let state = state.ok_or_else(|| "Missing 'state' query parameter".to_string())?;
|
||||||
|
let code = code.ok_or_else(|| "Missing 'code' query parameter".to_string())?;
|
||||||
|
|
||||||
|
if code.is_empty() {
|
||||||
|
return Err("Missing 'code' query parameter".to_string());
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok((state, code))
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn read_http_target(stream: &mut TcpStream) -> CommandResult<String> {
|
||||||
|
let mut buf = vec![0_u8; MAX_REQUEST_BYTES];
|
||||||
|
let mut total_read = 0_usize;
|
||||||
|
|
||||||
|
loop {
|
||||||
|
let n = stream
|
||||||
|
.read(&mut buf[total_read..])
|
||||||
|
.await
|
||||||
|
.map_err(|e| format!("Failed reading callback request: {e}"))?;
|
||||||
|
if n == 0 {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
total_read += n;
|
||||||
|
|
||||||
|
if buf[..total_read].windows(4).any(|w| w == b"\r\n\r\n") {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
if total_read == MAX_REQUEST_BYTES {
|
||||||
|
return Err("OAuth callback request too large".to_string());
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
let req = String::from_utf8_lossy(&buf[..total_read]);
|
||||||
|
let request_line =
|
||||||
|
req.lines().next().ok_or_else(|| "Invalid callback request line".to_string())?;
|
||||||
|
let mut parts = request_line.split_whitespace();
|
||||||
|
let method = parts.next().unwrap_or_default();
|
||||||
|
let target = parts.next().unwrap_or_default();
|
||||||
|
|
||||||
|
if method != "GET" {
|
||||||
|
return Err(format!("Expected GET callback request, got '{method}'"));
|
||||||
|
}
|
||||||
|
if target.is_empty() {
|
||||||
|
return Err("Missing callback request target".to_string());
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(target.to_string())
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn write_bad_request(stream: &mut TcpStream, message: &str) -> std::io::Result<()> {
|
||||||
|
let body = format!("Failed to authenticate: {message}");
|
||||||
|
let response = format!(
|
||||||
|
"HTTP/1.1 400 Bad Request\r\nContent-Type: text/plain; charset=utf-8\r\nContent-Length: {}\r\nConnection: close\r\n\r\n{}",
|
||||||
|
body.len(),
|
||||||
|
body
|
||||||
|
);
|
||||||
|
stream.write_all(response.as_bytes()).await?;
|
||||||
|
stream.shutdown().await
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn write_redirect(stream: &mut TcpStream, location: &str) -> std::io::Result<()> {
|
||||||
|
let response = format!(
|
||||||
|
"HTTP/1.1 302 Found\r\nLocation: {location}\r\nContent-Length: 0\r\nConnection: close\r\n\r\n"
|
||||||
|
);
|
||||||
|
stream.write_all(response.as_bytes()).await?;
|
||||||
|
stream.shutdown().await
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn exchange_access_token(oauth: &OAuthFlow, code: &str) -> CommandResult<String> {
|
||||||
|
let response = http::build_client(None)?
|
||||||
|
.post(&oauth.token_url)
|
||||||
|
.form(&[
|
||||||
|
("grant_type", "authorization_code"),
|
||||||
|
("client_id", OAUTH_CLIENT_ID),
|
||||||
|
("code", code),
|
||||||
|
("redirect_uri", oauth.redirect_url.as_str()),
|
||||||
|
("code_verifier", oauth.code_verifier.as_str()),
|
||||||
|
])
|
||||||
|
.send()
|
||||||
|
.await
|
||||||
|
.map_err(|e| format!("Failed to exchange OAuth code for access token: {e}"))?;
|
||||||
|
|
||||||
|
let status = response.status();
|
||||||
|
let body =
|
||||||
|
response.text().await.map_err(|e| format!("Failed to read token response body: {e}"))?;
|
||||||
|
|
||||||
|
if !status.is_success() {
|
||||||
|
return Err(format!(
|
||||||
|
"Failed to fetch access token: status={} body={}",
|
||||||
|
status.as_u16(),
|
||||||
|
body
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
|
let parsed: Value =
|
||||||
|
serde_json::from_str(&body).map_err(|e| format!("Invalid token response JSON: {e}"))?;
|
||||||
|
let token = parsed
|
||||||
|
.get("access_token")
|
||||||
|
.and_then(Value::as_str)
|
||||||
|
.filter(|s| !s.is_empty())
|
||||||
|
.ok_or_else(|| format!("Token response missing access_token: {body}"))?;
|
||||||
|
|
||||||
|
Ok(token.to_string())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn keyring_entry(environment: Environment) -> CommandResult<Entry> {
|
||||||
|
Entry::new(environment.keyring_service(), KEYRING_USER)
|
||||||
|
.map_err(|e| format!("Failed to initialize auth keyring entry: {e}"))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_auth_token(environment: Environment) -> CommandResult<Option<String>> {
|
||||||
|
let entry = keyring_entry(environment)?;
|
||||||
|
match entry.get_password() {
|
||||||
|
Ok(token) => Ok(Some(token)),
|
||||||
|
Err(keyring::Error::NoEntry) => Ok(None),
|
||||||
|
Err(err) => Err(format!("Failed to read auth token: {err}")),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn store_auth_token(environment: Environment, token: &str) -> CommandResult {
|
||||||
|
let entry = keyring_entry(environment)?;
|
||||||
|
entry.set_password(token).map_err(|e| format!("Failed to store auth token: {e}"))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn delete_auth_token(environment: Environment) -> CommandResult {
|
||||||
|
let entry = keyring_entry(environment)?;
|
||||||
|
match entry.delete_credential() {
|
||||||
|
Ok(()) | Err(keyring::Error::NoEntry) => Ok(()),
|
||||||
|
Err(err) => Err(format!("Failed to delete auth token: {err}")),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn random_hex(bytes: usize) -> String {
|
||||||
|
let mut data = vec![0_u8; bytes];
|
||||||
|
OsRng.fill_bytes(&mut data);
|
||||||
|
hex::encode(data)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn confirm_open_browser() -> CommandResult<bool> {
|
||||||
|
if !io::stdin().is_terminal() {
|
||||||
|
return Ok(true);
|
||||||
|
}
|
||||||
|
|
||||||
|
loop {
|
||||||
|
print!("Open default browser? [Y/n]: ");
|
||||||
|
io::stdout().flush().map_err(|e| format!("Failed to flush stdout: {e}"))?;
|
||||||
|
|
||||||
|
let mut input = String::new();
|
||||||
|
io::stdin().read_line(&mut input).map_err(|e| format!("Failed to read input: {e}"))?;
|
||||||
|
|
||||||
|
match input.trim().to_ascii_lowercase().as_str() {
|
||||||
|
"" | "y" | "yes" => return Ok(true),
|
||||||
|
"n" | "no" => return Ok(false),
|
||||||
|
_ => ui::warning("Please answer y or n"),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use super::*;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn environment_mapping() {
|
||||||
|
assert_eq!(parse_environment(Some("staging")), Environment::Staging);
|
||||||
|
assert_eq!(parse_environment(Some("development")), Environment::Development);
|
||||||
|
assert_eq!(parse_environment(Some("production")), Environment::Production);
|
||||||
|
assert_eq!(parse_environment(None), Environment::Production);
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn parses_callback_request() {
|
||||||
|
let listener = TcpListener::bind("127.0.0.1:0").await.expect("bind");
|
||||||
|
let addr = listener.local_addr().expect("local addr");
|
||||||
|
|
||||||
|
let server = tokio::spawn(async move {
|
||||||
|
let (mut stream, _) = listener.accept().await.expect("accept");
|
||||||
|
parse_callback_request(&mut stream).await
|
||||||
|
});
|
||||||
|
|
||||||
|
let mut client = TcpStream::connect(addr).await.expect("connect");
|
||||||
|
client
|
||||||
|
.write_all(
|
||||||
|
b"GET /oauth/callback?code=abc123&state=xyz HTTP/1.1\r\nHost: localhost\r\n\r\n",
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
.expect("write");
|
||||||
|
|
||||||
|
let parsed = server.await.expect("join").expect("parse");
|
||||||
|
assert_eq!(parsed.0, "xyz");
|
||||||
|
assert_eq!(parsed.1, "abc123");
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn parse_callback_request_oauth_error() {
|
||||||
|
let listener = TcpListener::bind("127.0.0.1:0").await.expect("bind");
|
||||||
|
let addr = listener.local_addr().expect("local addr");
|
||||||
|
|
||||||
|
let server = tokio::spawn(async move {
|
||||||
|
let (mut stream, _) = listener.accept().await.expect("accept");
|
||||||
|
parse_callback_request(&mut stream).await
|
||||||
|
});
|
||||||
|
|
||||||
|
let mut client = TcpStream::connect(addr).await.expect("connect");
|
||||||
|
client
|
||||||
|
.write_all(
|
||||||
|
b"GET /oauth/callback?error=access_denied&error_description=User%20denied&state=xyz HTTP/1.1\r\nHost: localhost\r\n\r\n",
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
.expect("write");
|
||||||
|
|
||||||
|
let err = server.await.expect("join").expect_err("should fail");
|
||||||
|
assert!(err.contains("OAuth provider returned error: access_denied"));
|
||||||
|
assert!(err.contains("User denied"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[tokio::test]
|
||||||
|
async fn receive_oauth_code_fails_fast_on_provider_error() {
|
||||||
|
let listener = TcpListener::bind("127.0.0.1:0").await.expect("bind");
|
||||||
|
let addr = listener.local_addr().expect("local addr");
|
||||||
|
|
||||||
|
let server = tokio::spawn(async move {
|
||||||
|
receive_oauth_code(listener, "expected-state", "http://localhost:9444").await
|
||||||
|
});
|
||||||
|
|
||||||
|
let mut client = TcpStream::connect(addr).await.expect("connect");
|
||||||
|
client
|
||||||
|
.write_all(
|
||||||
|
b"GET /oauth/callback?error=access_denied&state=expected-state HTTP/1.1\r\nHost: localhost\r\n\r\n",
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
.expect("write");
|
||||||
|
|
||||||
|
let result = tokio::time::timeout(std::time::Duration::from_secs(2), server)
|
||||||
|
.await
|
||||||
|
.expect("should not timeout")
|
||||||
|
.expect("join");
|
||||||
|
let err = result.expect_err("should return oauth error");
|
||||||
|
assert!(err.contains("OAuth provider returned error: access_denied"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn builds_oauth_flow_with_pkce() {
|
||||||
|
let flow = build_oauth_flow(Environment::Development, 8080).expect("flow");
|
||||||
|
assert!(flow.auth_url.as_str().contains("code_challenge_method=S256"));
|
||||||
|
assert!(
|
||||||
|
flow.auth_url
|
||||||
|
.as_str()
|
||||||
|
.contains("redirect_uri=http%3A%2F%2F127.0.0.1%3A8080%2Foauth%2Fcallback")
|
||||||
|
);
|
||||||
|
assert_eq!(flow.redirect_url, "http://127.0.0.1:8080/oauth/callback");
|
||||||
|
assert_eq!(flow.token_url, "http://localhost:9444/login/oauth/access_token");
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -2,9 +2,11 @@ use crate::cli::{EnvironmentArgs, EnvironmentCommands};
|
|||||||
use crate::context::CliContext;
|
use crate::context::CliContext;
|
||||||
use crate::utils::confirm::confirm_delete;
|
use crate::utils::confirm::confirm_delete;
|
||||||
use crate::utils::json::{
|
use crate::utils::json::{
|
||||||
apply_merge_patch, is_json_shorthand, parse_optional_json, parse_required_json, require_id,
|
apply_merge_patch, is_json_shorthand, merge_workspace_id_arg, parse_optional_json,
|
||||||
validate_create_id,
|
parse_required_json, require_id, validate_create_id,
|
||||||
};
|
};
|
||||||
|
use crate::utils::schema::append_agent_hints;
|
||||||
|
use schemars::schema_for;
|
||||||
use yaak_models::models::Environment;
|
use yaak_models::models::Environment;
|
||||||
use yaak_models::util::UpdateSource;
|
use yaak_models::util::UpdateSource;
|
||||||
|
|
||||||
@@ -13,6 +15,7 @@ type CommandResult<T = ()> = std::result::Result<T, String>;
|
|||||||
pub fn run(ctx: &CliContext, args: EnvironmentArgs) -> i32 {
|
pub fn run(ctx: &CliContext, args: EnvironmentArgs) -> i32 {
|
||||||
let result = match args.command {
|
let result = match args.command {
|
||||||
EnvironmentCommands::List { workspace_id } => list(ctx, &workspace_id),
|
EnvironmentCommands::List { workspace_id } => list(ctx, &workspace_id),
|
||||||
|
EnvironmentCommands::Schema { pretty } => schema(pretty),
|
||||||
EnvironmentCommands::Show { environment_id } => show(ctx, &environment_id),
|
EnvironmentCommands::Show { environment_id } => show(ctx, &environment_id),
|
||||||
EnvironmentCommands::Create { workspace_id, name, json } => {
|
EnvironmentCommands::Create { workspace_id, name, json } => {
|
||||||
create(ctx, workspace_id, name, json)
|
create(ctx, workspace_id, name, json)
|
||||||
@@ -30,6 +33,18 @@ pub fn run(ctx: &CliContext, args: EnvironmentArgs) -> i32 {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn schema(pretty: bool) -> CommandResult {
|
||||||
|
let mut schema = serde_json::to_value(schema_for!(Environment))
|
||||||
|
.map_err(|e| format!("Failed to serialize environment schema: {e}"))?;
|
||||||
|
append_agent_hints(&mut schema);
|
||||||
|
|
||||||
|
let output =
|
||||||
|
if pretty { serde_json::to_string_pretty(&schema) } else { serde_json::to_string(&schema) }
|
||||||
|
.map_err(|e| format!("Failed to format environment schema JSON: {e}"))?;
|
||||||
|
println!("{output}");
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
fn list(ctx: &CliContext, workspace_id: &str) -> CommandResult {
|
fn list(ctx: &CliContext, workspace_id: &str) -> CommandResult {
|
||||||
let environments = ctx
|
let environments = ctx
|
||||||
.db()
|
.db()
|
||||||
@@ -63,17 +78,11 @@ fn create(
|
|||||||
name: Option<String>,
|
name: Option<String>,
|
||||||
json: Option<String>,
|
json: Option<String>,
|
||||||
) -> CommandResult {
|
) -> CommandResult {
|
||||||
if json.is_some() && workspace_id.as_deref().is_some_and(|v| !is_json_shorthand(v)) {
|
let json_shorthand =
|
||||||
return Err(
|
workspace_id.as_deref().filter(|v| is_json_shorthand(v)).map(str::to_owned);
|
||||||
"environment create cannot combine workspace_id with --json payload".to_string()
|
let workspace_id_arg = workspace_id.filter(|v| !is_json_shorthand(v));
|
||||||
);
|
|
||||||
}
|
|
||||||
|
|
||||||
let payload = parse_optional_json(
|
let payload = parse_optional_json(json, json_shorthand, "environment create")?;
|
||||||
json,
|
|
||||||
workspace_id.clone().filter(|v| is_json_shorthand(v)),
|
|
||||||
"environment create",
|
|
||||||
)?;
|
|
||||||
|
|
||||||
if let Some(payload) = payload {
|
if let Some(payload) = payload {
|
||||||
if name.is_some() {
|
if name.is_some() {
|
||||||
@@ -83,10 +92,11 @@ fn create(
|
|||||||
validate_create_id(&payload, "environment")?;
|
validate_create_id(&payload, "environment")?;
|
||||||
let mut environment: Environment = serde_json::from_value(payload)
|
let mut environment: Environment = serde_json::from_value(payload)
|
||||||
.map_err(|e| format!("Failed to parse environment create JSON: {e}"))?;
|
.map_err(|e| format!("Failed to parse environment create JSON: {e}"))?;
|
||||||
|
merge_workspace_id_arg(
|
||||||
if environment.workspace_id.is_empty() {
|
workspace_id_arg.as_deref(),
|
||||||
return Err("environment create JSON requires non-empty \"workspaceId\"".to_string());
|
&mut environment.workspace_id,
|
||||||
}
|
"environment create",
|
||||||
|
)?;
|
||||||
|
|
||||||
if environment.parent_model.is_empty() {
|
if environment.parent_model.is_empty() {
|
||||||
environment.parent_model = "environment".to_string();
|
environment.parent_model = "environment".to_string();
|
||||||
@@ -101,7 +111,7 @@ fn create(
|
|||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
|
|
||||||
let workspace_id = workspace_id.ok_or_else(|| {
|
let workspace_id = workspace_id_arg.ok_or_else(|| {
|
||||||
"environment create requires workspace_id unless JSON payload is provided".to_string()
|
"environment create requires workspace_id unless JSON payload is provided".to_string()
|
||||||
})?;
|
})?;
|
||||||
let name = name.ok_or_else(|| {
|
let name = name.ok_or_else(|| {
|
||||||
|
|||||||
@@ -2,8 +2,8 @@ use crate::cli::{FolderArgs, FolderCommands};
|
|||||||
use crate::context::CliContext;
|
use crate::context::CliContext;
|
||||||
use crate::utils::confirm::confirm_delete;
|
use crate::utils::confirm::confirm_delete;
|
||||||
use crate::utils::json::{
|
use crate::utils::json::{
|
||||||
apply_merge_patch, is_json_shorthand, parse_optional_json, parse_required_json, require_id,
|
apply_merge_patch, is_json_shorthand, merge_workspace_id_arg, parse_optional_json,
|
||||||
validate_create_id,
|
parse_required_json, require_id, validate_create_id,
|
||||||
};
|
};
|
||||||
use yaak_models::models::Folder;
|
use yaak_models::models::Folder;
|
||||||
use yaak_models::util::UpdateSource;
|
use yaak_models::util::UpdateSource;
|
||||||
@@ -58,15 +58,11 @@ fn create(
|
|||||||
name: Option<String>,
|
name: Option<String>,
|
||||||
json: Option<String>,
|
json: Option<String>,
|
||||||
) -> CommandResult {
|
) -> CommandResult {
|
||||||
if json.is_some() && workspace_id.as_deref().is_some_and(|v| !is_json_shorthand(v)) {
|
let json_shorthand =
|
||||||
return Err("folder create cannot combine workspace_id with --json payload".to_string());
|
workspace_id.as_deref().filter(|v| is_json_shorthand(v)).map(str::to_owned);
|
||||||
}
|
let workspace_id_arg = workspace_id.filter(|v| !is_json_shorthand(v));
|
||||||
|
|
||||||
let payload = parse_optional_json(
|
let payload = parse_optional_json(json, json_shorthand, "folder create")?;
|
||||||
json,
|
|
||||||
workspace_id.clone().filter(|v| is_json_shorthand(v)),
|
|
||||||
"folder create",
|
|
||||||
)?;
|
|
||||||
|
|
||||||
if let Some(payload) = payload {
|
if let Some(payload) = payload {
|
||||||
if name.is_some() {
|
if name.is_some() {
|
||||||
@@ -74,12 +70,13 @@ fn create(
|
|||||||
}
|
}
|
||||||
|
|
||||||
validate_create_id(&payload, "folder")?;
|
validate_create_id(&payload, "folder")?;
|
||||||
let folder: Folder = serde_json::from_value(payload)
|
let mut folder: Folder = serde_json::from_value(payload)
|
||||||
.map_err(|e| format!("Failed to parse folder create JSON: {e}"))?;
|
.map_err(|e| format!("Failed to parse folder create JSON: {e}"))?;
|
||||||
|
merge_workspace_id_arg(
|
||||||
if folder.workspace_id.is_empty() {
|
workspace_id_arg.as_deref(),
|
||||||
return Err("folder create JSON requires non-empty \"workspaceId\"".to_string());
|
&mut folder.workspace_id,
|
||||||
}
|
"folder create",
|
||||||
|
)?;
|
||||||
|
|
||||||
let created = ctx
|
let created = ctx
|
||||||
.db()
|
.db()
|
||||||
@@ -90,7 +87,7 @@ fn create(
|
|||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
|
|
||||||
let workspace_id = workspace_id.ok_or_else(|| {
|
let workspace_id = workspace_id_arg.ok_or_else(|| {
|
||||||
"folder create requires workspace_id unless JSON payload is provided".to_string()
|
"folder create requires workspace_id unless JSON payload is provided".to_string()
|
||||||
})?;
|
})?;
|
||||||
let name = name.ok_or_else(|| {
|
let name = name.ok_or_else(|| {
|
||||||
|
|||||||
@@ -1,5 +1,7 @@
|
|||||||
|
pub mod auth;
|
||||||
pub mod environment;
|
pub mod environment;
|
||||||
pub mod folder;
|
pub mod folder;
|
||||||
|
pub mod plugin;
|
||||||
pub mod request;
|
pub mod request;
|
||||||
pub mod send;
|
pub mod send;
|
||||||
pub mod workspace;
|
pub mod workspace;
|
||||||
|
|||||||
525
crates-cli/yaak-cli/src/commands/plugin.rs
Normal file
525
crates-cli/yaak-cli/src/commands/plugin.rs
Normal file
@@ -0,0 +1,525 @@
|
|||||||
|
use crate::cli::{GenerateArgs, PluginArgs, PluginCommands, PluginPathArg};
|
||||||
|
use crate::ui;
|
||||||
|
use crate::utils::http;
|
||||||
|
use keyring::Entry;
|
||||||
|
use rand::Rng;
|
||||||
|
use rolldown::{
|
||||||
|
Bundler, BundlerOptions, ExperimentalOptions, InputItem, LogLevel, OutputFormat, Platform,
|
||||||
|
WatchOption, Watcher,
|
||||||
|
};
|
||||||
|
use serde::Deserialize;
|
||||||
|
use std::collections::HashSet;
|
||||||
|
use std::fs;
|
||||||
|
use std::io::{self, IsTerminal, Read, Write};
|
||||||
|
use std::path::{Path, PathBuf};
|
||||||
|
use std::sync::Arc;
|
||||||
|
use tokio::sync::Mutex;
|
||||||
|
use walkdir::WalkDir;
|
||||||
|
use zip::CompressionMethod;
|
||||||
|
use zip::write::SimpleFileOptions;
|
||||||
|
|
||||||
|
type CommandResult<T = ()> = std::result::Result<T, String>;
|
||||||
|
|
||||||
|
const KEYRING_USER: &str = "yaak";
|
||||||
|
|
||||||
|
#[derive(Clone, Copy, Debug, Eq, PartialEq)]
|
||||||
|
enum Environment {
|
||||||
|
Production,
|
||||||
|
Staging,
|
||||||
|
Development,
|
||||||
|
}
|
||||||
|
|
||||||
|
impl Environment {
|
||||||
|
fn api_base_url(self) -> &'static str {
|
||||||
|
match self {
|
||||||
|
Environment::Production => "https://api.yaak.app",
|
||||||
|
Environment::Staging => "https://todo.yaak.app",
|
||||||
|
Environment::Development => "http://localhost:9444",
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn keyring_service(self) -> &'static str {
|
||||||
|
match self {
|
||||||
|
Environment::Production => "app.yaak.cli.Token",
|
||||||
|
Environment::Staging => "app.yaak.cli.staging.Token",
|
||||||
|
Environment::Development => "app.yaak.cli.dev.Token",
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn run_build(args: PluginPathArg) -> i32 {
|
||||||
|
match build(args).await {
|
||||||
|
Ok(()) => 0,
|
||||||
|
Err(error) => {
|
||||||
|
ui::error(&error);
|
||||||
|
1
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn run(args: PluginArgs) -> i32 {
|
||||||
|
match args.command {
|
||||||
|
PluginCommands::Build(args) => run_build(args).await,
|
||||||
|
PluginCommands::Dev(args) => run_dev(args).await,
|
||||||
|
PluginCommands::Generate(args) => run_generate(args).await,
|
||||||
|
PluginCommands::Publish(args) => run_publish(args).await,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn run_dev(args: PluginPathArg) -> i32 {
|
||||||
|
match dev(args).await {
|
||||||
|
Ok(()) => 0,
|
||||||
|
Err(error) => {
|
||||||
|
ui::error(&error);
|
||||||
|
1
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn run_generate(args: GenerateArgs) -> i32 {
|
||||||
|
match generate(args) {
|
||||||
|
Ok(()) => 0,
|
||||||
|
Err(error) => {
|
||||||
|
ui::error(&error);
|
||||||
|
1
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub async fn run_publish(args: PluginPathArg) -> i32 {
|
||||||
|
match publish(args).await {
|
||||||
|
Ok(()) => 0,
|
||||||
|
Err(error) => {
|
||||||
|
ui::error(&error);
|
||||||
|
1
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn build(args: PluginPathArg) -> CommandResult {
|
||||||
|
let plugin_dir = resolve_plugin_dir(args.path)?;
|
||||||
|
ensure_plugin_build_inputs(&plugin_dir)?;
|
||||||
|
|
||||||
|
ui::info(&format!("Building plugin {}...", plugin_dir.display()));
|
||||||
|
let warnings = build_plugin_bundle(&plugin_dir).await?;
|
||||||
|
for warning in warnings {
|
||||||
|
ui::warning(&warning);
|
||||||
|
}
|
||||||
|
ui::success(&format!("Built plugin bundle at {}", plugin_dir.join("build/index.js").display()));
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn dev(args: PluginPathArg) -> CommandResult {
|
||||||
|
let plugin_dir = resolve_plugin_dir(args.path)?;
|
||||||
|
ensure_plugin_build_inputs(&plugin_dir)?;
|
||||||
|
|
||||||
|
ui::info(&format!("Watching plugin {}...", plugin_dir.display()));
|
||||||
|
ui::info("Press Ctrl-C to stop");
|
||||||
|
|
||||||
|
let bundler = Bundler::new(bundler_options(&plugin_dir, true))
|
||||||
|
.map_err(|err| format!("Failed to initialize Rolldown watcher: {err}"))?;
|
||||||
|
let watcher = Watcher::new(vec![Arc::new(Mutex::new(bundler))], None)
|
||||||
|
.map_err(|err| format!("Failed to start Rolldown watcher: {err}"))?;
|
||||||
|
|
||||||
|
watcher.start().await;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn generate(args: GenerateArgs) -> CommandResult {
|
||||||
|
let default_name = random_name();
|
||||||
|
let name = match args.name {
|
||||||
|
Some(name) => name,
|
||||||
|
None => prompt_with_default("Plugin name", &default_name)?,
|
||||||
|
};
|
||||||
|
|
||||||
|
let default_dir = format!("./{name}");
|
||||||
|
let output_dir = match args.dir {
|
||||||
|
Some(dir) => dir,
|
||||||
|
None => PathBuf::from(prompt_with_default("Plugin dir", &default_dir)?),
|
||||||
|
};
|
||||||
|
|
||||||
|
if output_dir.exists() {
|
||||||
|
return Err(format!("Plugin directory already exists: {}", output_dir.display()));
|
||||||
|
}
|
||||||
|
|
||||||
|
ui::info(&format!("Generating plugin in {}", output_dir.display()));
|
||||||
|
fs::create_dir_all(output_dir.join("src"))
|
||||||
|
.map_err(|e| format!("Failed creating plugin directory {}: {e}", output_dir.display()))?;
|
||||||
|
|
||||||
|
write_file(&output_dir.join(".gitignore"), TEMPLATE_GITIGNORE)?;
|
||||||
|
write_file(
|
||||||
|
&output_dir.join("package.json"),
|
||||||
|
&TEMPLATE_PACKAGE_JSON.replace("yaak-plugin-name", &name),
|
||||||
|
)?;
|
||||||
|
write_file(&output_dir.join("tsconfig.json"), TEMPLATE_TSCONFIG)?;
|
||||||
|
write_file(&output_dir.join("README.md"), &TEMPLATE_README.replace("yaak-plugin-name", &name))?;
|
||||||
|
write_file(
|
||||||
|
&output_dir.join("src/index.ts"),
|
||||||
|
&TEMPLATE_INDEX_TS.replace("yaak-plugin-name", &name),
|
||||||
|
)?;
|
||||||
|
write_file(&output_dir.join("src/index.test.ts"), TEMPLATE_INDEX_TEST_TS)?;
|
||||||
|
|
||||||
|
ui::success("Plugin scaffold generated");
|
||||||
|
ui::info("Next steps:");
|
||||||
|
println!(" 1. cd {}", output_dir.display());
|
||||||
|
println!(" 2. npm install");
|
||||||
|
println!(" 3. yaak plugin build");
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn publish(args: PluginPathArg) -> CommandResult {
|
||||||
|
let plugin_dir = resolve_plugin_dir(args.path)?;
|
||||||
|
ensure_plugin_build_inputs(&plugin_dir)?;
|
||||||
|
|
||||||
|
let environment = current_environment();
|
||||||
|
let token = get_auth_token(environment)?
|
||||||
|
.ok_or_else(|| "Not logged in. Run `yaak auth login`.".to_string())?;
|
||||||
|
|
||||||
|
ui::info(&format!("Building plugin {}...", plugin_dir.display()));
|
||||||
|
let warnings = build_plugin_bundle(&plugin_dir).await?;
|
||||||
|
for warning in warnings {
|
||||||
|
ui::warning(&warning);
|
||||||
|
}
|
||||||
|
|
||||||
|
ui::info("Archiving plugin");
|
||||||
|
let archive = create_publish_archive(&plugin_dir)?;
|
||||||
|
|
||||||
|
ui::info("Uploading plugin");
|
||||||
|
let url = format!("{}/api/v1/plugins/publish", environment.api_base_url());
|
||||||
|
let response = http::build_client(Some(&token))?
|
||||||
|
.post(url)
|
||||||
|
.header(reqwest::header::CONTENT_TYPE, "application/zip")
|
||||||
|
.body(archive)
|
||||||
|
.send()
|
||||||
|
.await
|
||||||
|
.map_err(|e| format!("Failed to upload plugin: {e}"))?;
|
||||||
|
|
||||||
|
let status = response.status();
|
||||||
|
let body =
|
||||||
|
response.text().await.map_err(|e| format!("Failed reading publish response body: {e}"))?;
|
||||||
|
|
||||||
|
if !status.is_success() {
|
||||||
|
return Err(http::parse_api_error(status.as_u16(), &body));
|
||||||
|
}
|
||||||
|
|
||||||
|
let published: PublishResponse = serde_json::from_str(&body)
|
||||||
|
.map_err(|e| format!("Failed parsing publish response JSON: {e}\nResponse: {body}"))?;
|
||||||
|
ui::success(&format!("Plugin published {}", published.version));
|
||||||
|
println!(" -> {}", published.url);
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
#[derive(Deserialize)]
|
||||||
|
struct PublishResponse {
|
||||||
|
version: String,
|
||||||
|
url: String,
|
||||||
|
}
|
||||||
|
|
||||||
|
async fn build_plugin_bundle(plugin_dir: &Path) -> CommandResult<Vec<String>> {
|
||||||
|
prepare_build_output_dir(plugin_dir)?;
|
||||||
|
let mut bundler = Bundler::new(bundler_options(plugin_dir, false))
|
||||||
|
.map_err(|err| format!("Failed to initialize Rolldown: {err}"))?;
|
||||||
|
let output = bundler.write().await.map_err(|err| format!("Plugin build failed:\n{err}"))?;
|
||||||
|
|
||||||
|
Ok(output.warnings.into_iter().map(|w| w.to_string()).collect())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn prepare_build_output_dir(plugin_dir: &Path) -> CommandResult {
|
||||||
|
let build_dir = plugin_dir.join("build");
|
||||||
|
if build_dir.exists() {
|
||||||
|
fs::remove_dir_all(&build_dir)
|
||||||
|
.map_err(|e| format!("Failed to clean build directory {}: {e}", build_dir.display()))?;
|
||||||
|
}
|
||||||
|
fs::create_dir_all(&build_dir)
|
||||||
|
.map_err(|e| format!("Failed to create build directory {}: {e}", build_dir.display()))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn bundler_options(plugin_dir: &Path, watch: bool) -> BundlerOptions {
|
||||||
|
BundlerOptions {
|
||||||
|
input: Some(vec![InputItem { import: "./src/index.ts".to_string(), ..Default::default() }]),
|
||||||
|
cwd: Some(plugin_dir.to_path_buf()),
|
||||||
|
file: Some("build/index.js".to_string()),
|
||||||
|
format: Some(OutputFormat::Cjs),
|
||||||
|
platform: Some(Platform::Node),
|
||||||
|
log_level: Some(LogLevel::Info),
|
||||||
|
experimental: watch
|
||||||
|
.then_some(ExperimentalOptions { incremental_build: Some(true), ..Default::default() }),
|
||||||
|
watch: watch.then_some(WatchOption::default()),
|
||||||
|
..Default::default()
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn resolve_plugin_dir(path: Option<PathBuf>) -> CommandResult<PathBuf> {
|
||||||
|
let cwd =
|
||||||
|
std::env::current_dir().map_err(|e| format!("Failed to read current directory: {e}"))?;
|
||||||
|
let candidate = match path {
|
||||||
|
Some(path) if path.is_absolute() => path,
|
||||||
|
Some(path) => cwd.join(path),
|
||||||
|
None => cwd,
|
||||||
|
};
|
||||||
|
|
||||||
|
if !candidate.exists() {
|
||||||
|
return Err(format!("Plugin directory does not exist: {}", candidate.display()));
|
||||||
|
}
|
||||||
|
if !candidate.is_dir() {
|
||||||
|
return Err(format!("Plugin path is not a directory: {}", candidate.display()));
|
||||||
|
}
|
||||||
|
|
||||||
|
candidate
|
||||||
|
.canonicalize()
|
||||||
|
.map_err(|e| format!("Failed to resolve plugin directory {}: {e}", candidate.display()))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn ensure_plugin_build_inputs(plugin_dir: &Path) -> CommandResult {
|
||||||
|
let package_json = plugin_dir.join("package.json");
|
||||||
|
if !package_json.is_file() {
|
||||||
|
return Err(format!(
|
||||||
|
"{} does not exist. Ensure that you are in a plugin directory.",
|
||||||
|
package_json.display()
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
|
let entry = plugin_dir.join("src/index.ts");
|
||||||
|
if !entry.is_file() {
|
||||||
|
return Err(format!("Required entrypoint missing: {}", entry.display()));
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn create_publish_archive(plugin_dir: &Path) -> CommandResult<Vec<u8>> {
|
||||||
|
let required_files = [
|
||||||
|
"README.md",
|
||||||
|
"package.json",
|
||||||
|
"build/index.js",
|
||||||
|
"src/index.ts",
|
||||||
|
];
|
||||||
|
let optional_files = ["package-lock.json"];
|
||||||
|
|
||||||
|
let mut selected = HashSet::new();
|
||||||
|
for required in required_files {
|
||||||
|
let required_path = plugin_dir.join(required);
|
||||||
|
if !required_path.is_file() {
|
||||||
|
return Err(format!("Missing required file: {required}"));
|
||||||
|
}
|
||||||
|
selected.insert(required.to_string());
|
||||||
|
}
|
||||||
|
for optional in optional_files {
|
||||||
|
selected.insert(optional.to_string());
|
||||||
|
}
|
||||||
|
|
||||||
|
let cursor = std::io::Cursor::new(Vec::new());
|
||||||
|
let mut zip = zip::ZipWriter::new(cursor);
|
||||||
|
let options = SimpleFileOptions::default().compression_method(CompressionMethod::Deflated);
|
||||||
|
|
||||||
|
for entry in WalkDir::new(plugin_dir) {
|
||||||
|
let entry = entry.map_err(|e| format!("Failed walking plugin directory: {e}"))?;
|
||||||
|
if !entry.file_type().is_file() {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
let path = entry.path();
|
||||||
|
let rel = path
|
||||||
|
.strip_prefix(plugin_dir)
|
||||||
|
.map_err(|e| format!("Failed deriving relative path for {}: {e}", path.display()))?;
|
||||||
|
let rel = rel.to_string_lossy().replace('\\', "/");
|
||||||
|
|
||||||
|
let keep = rel.starts_with("src/") || rel.starts_with("build/") || selected.contains(&rel);
|
||||||
|
if !keep {
|
||||||
|
continue;
|
||||||
|
}
|
||||||
|
|
||||||
|
zip.start_file(rel, options).map_err(|e| format!("Failed adding file to archive: {e}"))?;
|
||||||
|
let mut file = fs::File::open(path)
|
||||||
|
.map_err(|e| format!("Failed opening file {}: {e}", path.display()))?;
|
||||||
|
let mut contents = Vec::new();
|
||||||
|
file.read_to_end(&mut contents)
|
||||||
|
.map_err(|e| format!("Failed reading file {}: {e}", path.display()))?;
|
||||||
|
zip.write_all(&contents).map_err(|e| format!("Failed writing archive contents: {e}"))?;
|
||||||
|
}
|
||||||
|
|
||||||
|
let cursor = zip.finish().map_err(|e| format!("Failed finalizing plugin archive: {e}"))?;
|
||||||
|
Ok(cursor.into_inner())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn write_file(path: &Path, contents: &str) -> CommandResult {
|
||||||
|
if let Some(parent) = path.parent() {
|
||||||
|
fs::create_dir_all(parent)
|
||||||
|
.map_err(|e| format!("Failed creating directory {}: {e}", parent.display()))?;
|
||||||
|
}
|
||||||
|
fs::write(path, contents).map_err(|e| format!("Failed writing file {}: {e}", path.display()))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn prompt_with_default(label: &str, default: &str) -> CommandResult<String> {
|
||||||
|
if !io::stdin().is_terminal() {
|
||||||
|
return Ok(default.to_string());
|
||||||
|
}
|
||||||
|
|
||||||
|
print!("{label} [{default}]: ");
|
||||||
|
io::stdout().flush().map_err(|e| format!("Failed to flush stdout: {e}"))?;
|
||||||
|
|
||||||
|
let mut input = String::new();
|
||||||
|
io::stdin().read_line(&mut input).map_err(|e| format!("Failed to read input: {e}"))?;
|
||||||
|
let trimmed = input.trim();
|
||||||
|
|
||||||
|
if trimmed.is_empty() { Ok(default.to_string()) } else { Ok(trimmed.to_string()) }
|
||||||
|
}
|
||||||
|
|
||||||
|
fn current_environment() -> Environment {
|
||||||
|
match std::env::var("ENVIRONMENT").as_deref() {
|
||||||
|
Ok("staging") => Environment::Staging,
|
||||||
|
Ok("development") => Environment::Development,
|
||||||
|
_ => Environment::Production,
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn keyring_entry(environment: Environment) -> CommandResult<Entry> {
|
||||||
|
Entry::new(environment.keyring_service(), KEYRING_USER)
|
||||||
|
.map_err(|e| format!("Failed to initialize auth keyring entry: {e}"))
|
||||||
|
}
|
||||||
|
|
||||||
|
fn get_auth_token(environment: Environment) -> CommandResult<Option<String>> {
|
||||||
|
let entry = keyring_entry(environment)?;
|
||||||
|
match entry.get_password() {
|
||||||
|
Ok(token) => Ok(Some(token)),
|
||||||
|
Err(keyring::Error::NoEntry) => Ok(None),
|
||||||
|
Err(err) => Err(format!("Failed to read auth token: {err}")),
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn random_name() -> String {
|
||||||
|
const ADJECTIVES: &[&str] = &[
|
||||||
|
"young", "youthful", "yellow", "yielding", "yappy", "yawning", "yummy", "yucky", "yearly",
|
||||||
|
"yester", "yeasty", "yelling",
|
||||||
|
];
|
||||||
|
const NOUNS: &[&str] = &[
|
||||||
|
"yak", "yarn", "year", "yell", "yoke", "yoga", "yam", "yacht", "yodel",
|
||||||
|
];
|
||||||
|
|
||||||
|
let mut rng = rand::thread_rng();
|
||||||
|
let adjective = ADJECTIVES[rng.gen_range(0..ADJECTIVES.len())];
|
||||||
|
let noun = NOUNS[rng.gen_range(0..NOUNS.len())];
|
||||||
|
format!("{adjective}-{noun}")
|
||||||
|
}
|
||||||
|
|
||||||
|
const TEMPLATE_GITIGNORE: &str = "node_modules\n";
|
||||||
|
|
||||||
|
const TEMPLATE_PACKAGE_JSON: &str = r#"{
|
||||||
|
"name": "yaak-plugin-name",
|
||||||
|
"private": true,
|
||||||
|
"version": "0.0.1",
|
||||||
|
"scripts": {
|
||||||
|
"build": "yaak plugin build",
|
||||||
|
"dev": "yaak plugin dev"
|
||||||
|
},
|
||||||
|
"devDependencies": {
|
||||||
|
"@types/node": "^24.10.1",
|
||||||
|
"typescript": "^5.9.3",
|
||||||
|
"vitest": "^4.0.14"
|
||||||
|
},
|
||||||
|
"dependencies": {
|
||||||
|
"@yaakapp/api": "^0.7.0"
|
||||||
|
}
|
||||||
|
}
|
||||||
|
"#;
|
||||||
|
|
||||||
|
const TEMPLATE_TSCONFIG: &str = r#"{
|
||||||
|
"compilerOptions": {
|
||||||
|
"target": "es2021",
|
||||||
|
"lib": ["DOM", "DOM.Iterable", "ESNext"],
|
||||||
|
"useDefineForClassFields": true,
|
||||||
|
"allowJs": false,
|
||||||
|
"skipLibCheck": true,
|
||||||
|
"esModuleInterop": false,
|
||||||
|
"allowSyntheticDefaultImports": true,
|
||||||
|
"strict": true,
|
||||||
|
"noUncheckedIndexedAccess": true,
|
||||||
|
"forceConsistentCasingInFileNames": true,
|
||||||
|
"module": "ESNext",
|
||||||
|
"moduleResolution": "Node",
|
||||||
|
"resolveJsonModule": true,
|
||||||
|
"isolatedModules": true,
|
||||||
|
"noEmit": true,
|
||||||
|
"jsx": "react-jsx"
|
||||||
|
},
|
||||||
|
"include": ["src"]
|
||||||
|
}
|
||||||
|
"#;
|
||||||
|
|
||||||
|
const TEMPLATE_README: &str = r#"# yaak-plugin-name
|
||||||
|
|
||||||
|
Describe what your plugin does.
|
||||||
|
"#;
|
||||||
|
|
||||||
|
const TEMPLATE_INDEX_TS: &str = r#"import type { PluginDefinition } from "@yaakapp/api";
|
||||||
|
|
||||||
|
export const plugin: PluginDefinition = {
|
||||||
|
httpRequestActions: [
|
||||||
|
{
|
||||||
|
label: "Hello, From Plugin",
|
||||||
|
icon: "info",
|
||||||
|
async onSelect(ctx, args) {
|
||||||
|
await ctx.toast.show({
|
||||||
|
color: "success",
|
||||||
|
message: `You clicked the request ${args.httpRequest.id}`,
|
||||||
|
});
|
||||||
|
},
|
||||||
|
},
|
||||||
|
],
|
||||||
|
};
|
||||||
|
"#;
|
||||||
|
|
||||||
|
const TEMPLATE_INDEX_TEST_TS: &str = r#"import { describe, expect, test } from "vitest";
|
||||||
|
import { plugin } from "./index";
|
||||||
|
|
||||||
|
describe("Example Plugin", () => {
|
||||||
|
test("Exports plugin object", () => {
|
||||||
|
expect(plugin).toBeTypeOf("object");
|
||||||
|
});
|
||||||
|
});
|
||||||
|
"#;
|
||||||
|
|
||||||
|
#[cfg(test)]
|
||||||
|
mod tests {
|
||||||
|
use super::create_publish_archive;
|
||||||
|
use std::collections::HashSet;
|
||||||
|
use std::fs;
|
||||||
|
use std::io::Cursor;
|
||||||
|
use tempfile::TempDir;
|
||||||
|
use zip::ZipArchive;
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn publish_archive_includes_required_and_optional_files() {
|
||||||
|
let dir = TempDir::new().expect("temp dir");
|
||||||
|
let root = dir.path();
|
||||||
|
|
||||||
|
fs::create_dir_all(root.join("src")).expect("create src");
|
||||||
|
fs::create_dir_all(root.join("build")).expect("create build");
|
||||||
|
fs::create_dir_all(root.join("ignored")).expect("create ignored");
|
||||||
|
|
||||||
|
fs::write(root.join("README.md"), "# Demo\n").expect("write README");
|
||||||
|
fs::write(root.join("package.json"), "{}").expect("write package.json");
|
||||||
|
fs::write(root.join("package-lock.json"), "{}").expect("write package-lock.json");
|
||||||
|
fs::write(root.join("src/index.ts"), "export const plugin = {};\n")
|
||||||
|
.expect("write src/index.ts");
|
||||||
|
fs::write(root.join("build/index.js"), "exports.plugin = {};\n")
|
||||||
|
.expect("write build/index.js");
|
||||||
|
fs::write(root.join("ignored/secret.txt"), "do-not-ship").expect("write ignored file");
|
||||||
|
|
||||||
|
let archive = create_publish_archive(root).expect("create archive");
|
||||||
|
let mut zip = ZipArchive::new(Cursor::new(archive)).expect("open zip");
|
||||||
|
|
||||||
|
let mut names = HashSet::new();
|
||||||
|
for i in 0..zip.len() {
|
||||||
|
let file = zip.by_index(i).expect("zip entry");
|
||||||
|
names.insert(file.name().to_string());
|
||||||
|
}
|
||||||
|
|
||||||
|
assert!(names.contains("README.md"));
|
||||||
|
assert!(names.contains("package.json"));
|
||||||
|
assert!(names.contains("package-lock.json"));
|
||||||
|
assert!(names.contains("src/index.ts"));
|
||||||
|
assert!(names.contains("build/index.js"));
|
||||||
|
assert!(!names.contains("ignored/secret.txt"));
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -2,14 +2,17 @@ use crate::cli::{RequestArgs, RequestCommands, RequestSchemaType};
|
|||||||
use crate::context::CliContext;
|
use crate::context::CliContext;
|
||||||
use crate::utils::confirm::confirm_delete;
|
use crate::utils::confirm::confirm_delete;
|
||||||
use crate::utils::json::{
|
use crate::utils::json::{
|
||||||
apply_merge_patch, is_json_shorthand, parse_optional_json, parse_required_json, require_id,
|
apply_merge_patch, is_json_shorthand, merge_workspace_id_arg, parse_optional_json,
|
||||||
validate_create_id,
|
parse_required_json, require_id, validate_create_id,
|
||||||
};
|
};
|
||||||
|
use crate::utils::schema::append_agent_hints;
|
||||||
use schemars::schema_for;
|
use schemars::schema_for;
|
||||||
use serde_json::{Map, Value, json};
|
use serde_json::{Map, Value, json};
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
|
use std::io::Write;
|
||||||
use tokio::sync::mpsc;
|
use tokio::sync::mpsc;
|
||||||
use yaak::send::{SendHttpRequestByIdWithPluginsParams, send_http_request_by_id_with_plugins};
|
use yaak::send::{SendHttpRequestByIdWithPluginsParams, send_http_request_by_id_with_plugins};
|
||||||
|
use yaak_http::sender::HttpResponseEvent as SenderHttpResponseEvent;
|
||||||
use yaak_models::models::{GrpcRequest, HttpRequest, WebsocketRequest};
|
use yaak_models::models::{GrpcRequest, HttpRequest, WebsocketRequest};
|
||||||
use yaak_models::queries::any_request::AnyRequest;
|
use yaak_models::queries::any_request::AnyRequest;
|
||||||
use yaak_models::util::UpdateSource;
|
use yaak_models::util::UpdateSource;
|
||||||
@@ -35,8 +38,8 @@ pub async fn run(
|
|||||||
}
|
}
|
||||||
};
|
};
|
||||||
}
|
}
|
||||||
RequestCommands::Schema { request_type } => {
|
RequestCommands::Schema { request_type, pretty } => {
|
||||||
return match schema(ctx, request_type).await {
|
return match schema(ctx, request_type, pretty).await {
|
||||||
Ok(()) => 0,
|
Ok(()) => 0,
|
||||||
Err(error) => {
|
Err(error) => {
|
||||||
eprintln!("Error: {error}");
|
eprintln!("Error: {error}");
|
||||||
@@ -75,7 +78,7 @@ fn list(ctx: &CliContext, workspace_id: &str) -> CommandResult {
|
|||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
async fn schema(ctx: &CliContext, request_type: RequestSchemaType) -> CommandResult {
|
async fn schema(ctx: &CliContext, request_type: RequestSchemaType, pretty: bool) -> CommandResult {
|
||||||
let mut schema = match request_type {
|
let mut schema = match request_type {
|
||||||
RequestSchemaType::Http => serde_json::to_value(schema_for!(HttpRequest))
|
RequestSchemaType::Http => serde_json::to_value(schema_for!(HttpRequest))
|
||||||
.map_err(|e| format!("Failed to serialize HTTP request schema: {e}"))?,
|
.map_err(|e| format!("Failed to serialize HTTP request schema: {e}"))?,
|
||||||
@@ -85,16 +88,51 @@ async fn schema(ctx: &CliContext, request_type: RequestSchemaType) -> CommandRes
|
|||||||
.map_err(|e| format!("Failed to serialize WebSocket request schema: {e}"))?,
|
.map_err(|e| format!("Failed to serialize WebSocket request schema: {e}"))?,
|
||||||
};
|
};
|
||||||
|
|
||||||
|
enrich_schema_guidance(&mut schema, request_type);
|
||||||
|
append_agent_hints(&mut schema);
|
||||||
|
|
||||||
if let Err(error) = merge_auth_schema_from_plugins(ctx, &mut schema).await {
|
if let Err(error) = merge_auth_schema_from_plugins(ctx, &mut schema).await {
|
||||||
eprintln!("Warning: Failed to enrich authentication schema from plugins: {error}");
|
eprintln!("Warning: Failed to enrich authentication schema from plugins: {error}");
|
||||||
}
|
}
|
||||||
|
|
||||||
let output = serde_json::to_string_pretty(&schema)
|
let output =
|
||||||
.map_err(|e| format!("Failed to format schema JSON: {e}"))?;
|
if pretty { serde_json::to_string_pretty(&schema) } else { serde_json::to_string(&schema) }
|
||||||
|
.map_err(|e| format!("Failed to format schema JSON: {e}"))?;
|
||||||
println!("{output}");
|
println!("{output}");
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn enrich_schema_guidance(schema: &mut Value, request_type: RequestSchemaType) {
|
||||||
|
if !matches!(request_type, RequestSchemaType::Http) {
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
let Some(properties) = schema.get_mut("properties").and_then(Value::as_object_mut) else {
|
||||||
|
return;
|
||||||
|
};
|
||||||
|
|
||||||
|
if let Some(url_schema) = properties.get_mut("url").and_then(Value::as_object_mut) {
|
||||||
|
append_description(
|
||||||
|
url_schema,
|
||||||
|
"For path segments like `/foo/:id/comments/:commentId`, put concrete values in `urlParameters` using names without `:` (for example `id`, `commentId`).",
|
||||||
|
);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
fn append_description(schema: &mut Map<String, Value>, extra: &str) {
|
||||||
|
match schema.get_mut("description") {
|
||||||
|
Some(Value::String(existing)) if !existing.trim().is_empty() => {
|
||||||
|
if !existing.ends_with(' ') {
|
||||||
|
existing.push(' ');
|
||||||
|
}
|
||||||
|
existing.push_str(extra);
|
||||||
|
}
|
||||||
|
_ => {
|
||||||
|
schema.insert("description".to_string(), Value::String(extra.to_string()));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
async fn merge_auth_schema_from_plugins(
|
async fn merge_auth_schema_from_plugins(
|
||||||
ctx: &CliContext,
|
ctx: &CliContext,
|
||||||
schema: &mut Value,
|
schema: &mut Value,
|
||||||
@@ -298,15 +336,11 @@ fn create(
|
|||||||
url: Option<String>,
|
url: Option<String>,
|
||||||
json: Option<String>,
|
json: Option<String>,
|
||||||
) -> CommandResult {
|
) -> CommandResult {
|
||||||
if json.is_some() && workspace_id.as_deref().is_some_and(|v| !is_json_shorthand(v)) {
|
let json_shorthand =
|
||||||
return Err("request create cannot combine workspace_id with --json payload".to_string());
|
workspace_id.as_deref().filter(|v| is_json_shorthand(v)).map(str::to_owned);
|
||||||
}
|
let workspace_id_arg = workspace_id.filter(|v| !is_json_shorthand(v));
|
||||||
|
|
||||||
let payload = parse_optional_json(
|
let payload = parse_optional_json(json, json_shorthand, "request create")?;
|
||||||
json,
|
|
||||||
workspace_id.clone().filter(|v| is_json_shorthand(v)),
|
|
||||||
"request create",
|
|
||||||
)?;
|
|
||||||
|
|
||||||
if let Some(payload) = payload {
|
if let Some(payload) = payload {
|
||||||
if name.is_some() || method.is_some() || url.is_some() {
|
if name.is_some() || method.is_some() || url.is_some() {
|
||||||
@@ -314,12 +348,13 @@ fn create(
|
|||||||
}
|
}
|
||||||
|
|
||||||
validate_create_id(&payload, "request")?;
|
validate_create_id(&payload, "request")?;
|
||||||
let request: HttpRequest = serde_json::from_value(payload)
|
let mut request: HttpRequest = serde_json::from_value(payload)
|
||||||
.map_err(|e| format!("Failed to parse request create JSON: {e}"))?;
|
.map_err(|e| format!("Failed to parse request create JSON: {e}"))?;
|
||||||
|
merge_workspace_id_arg(
|
||||||
if request.workspace_id.is_empty() {
|
workspace_id_arg.as_deref(),
|
||||||
return Err("request create JSON requires non-empty \"workspaceId\"".to_string());
|
&mut request.workspace_id,
|
||||||
}
|
"request create",
|
||||||
|
)?;
|
||||||
|
|
||||||
let created = ctx
|
let created = ctx
|
||||||
.db()
|
.db()
|
||||||
@@ -330,7 +365,7 @@ fn create(
|
|||||||
return Ok(());
|
return Ok(());
|
||||||
}
|
}
|
||||||
|
|
||||||
let workspace_id = workspace_id.ok_or_else(|| {
|
let workspace_id = workspace_id_arg.ok_or_else(|| {
|
||||||
"request create requires workspace_id unless JSON payload is provided".to_string()
|
"request create requires workspace_id unless JSON payload is provided".to_string()
|
||||||
})?;
|
})?;
|
||||||
let name = name.unwrap_or_default();
|
let name = name.unwrap_or_default();
|
||||||
@@ -434,14 +469,24 @@ async fn send_http_request_by_id(
|
|||||||
) -> Result<(), String> {
|
) -> Result<(), String> {
|
||||||
let plugin_context = PluginContext::new(None, Some(workspace_id.to_string()));
|
let plugin_context = PluginContext::new(None, Some(workspace_id.to_string()));
|
||||||
|
|
||||||
let (event_tx, mut event_rx) = mpsc::channel(100);
|
let (event_tx, mut event_rx) = mpsc::channel::<SenderHttpResponseEvent>(100);
|
||||||
|
let (body_chunk_tx, mut body_chunk_rx) = mpsc::unbounded_channel::<Vec<u8>>();
|
||||||
let event_handle = tokio::spawn(async move {
|
let event_handle = tokio::spawn(async move {
|
||||||
while let Some(event) = event_rx.recv().await {
|
while let Some(event) = event_rx.recv().await {
|
||||||
if verbose {
|
if verbose && !matches!(event, SenderHttpResponseEvent::ChunkReceived { .. }) {
|
||||||
println!("{}", event);
|
println!("{}", event);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
});
|
});
|
||||||
|
let body_handle = tokio::task::spawn_blocking(move || {
|
||||||
|
let mut stdout = std::io::stdout();
|
||||||
|
while let Some(chunk) = body_chunk_rx.blocking_recv() {
|
||||||
|
if stdout.write_all(&chunk).is_err() {
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
let _ = stdout.flush();
|
||||||
|
}
|
||||||
|
});
|
||||||
let response_dir = ctx.data_dir().join("responses");
|
let response_dir = ctx.data_dir().join("responses");
|
||||||
|
|
||||||
let result = send_http_request_by_id_with_plugins(SendHttpRequestByIdWithPluginsParams {
|
let result = send_http_request_by_id_with_plugins(SendHttpRequestByIdWithPluginsParams {
|
||||||
@@ -453,6 +498,7 @@ async fn send_http_request_by_id(
|
|||||||
cookie_jar_id: None,
|
cookie_jar_id: None,
|
||||||
response_dir: &response_dir,
|
response_dir: &response_dir,
|
||||||
emit_events_to: Some(event_tx),
|
emit_events_to: Some(event_tx),
|
||||||
|
emit_response_body_chunks_to: Some(body_chunk_tx),
|
||||||
plugin_manager: ctx.plugin_manager(),
|
plugin_manager: ctx.plugin_manager(),
|
||||||
encryption_manager: ctx.encryption_manager.clone(),
|
encryption_manager: ctx.encryption_manager.clone(),
|
||||||
plugin_context: &plugin_context,
|
plugin_context: &plugin_context,
|
||||||
@@ -462,24 +508,7 @@ async fn send_http_request_by_id(
|
|||||||
.await;
|
.await;
|
||||||
|
|
||||||
let _ = event_handle.await;
|
let _ = event_handle.await;
|
||||||
let result = result.map_err(|e| e.to_string())?;
|
let _ = body_handle.await;
|
||||||
|
result.map_err(|e| e.to_string())?;
|
||||||
if verbose {
|
|
||||||
println!();
|
|
||||||
}
|
|
||||||
println!(
|
|
||||||
"HTTP {} {}",
|
|
||||||
result.response.status,
|
|
||||||
result.response.status_reason.as_deref().unwrap_or("")
|
|
||||||
);
|
|
||||||
if verbose {
|
|
||||||
for header in &result.response.headers {
|
|
||||||
println!("{}: {}", header.name, header.value);
|
|
||||||
}
|
|
||||||
println!();
|
|
||||||
}
|
|
||||||
let body = String::from_utf8(result.response_body)
|
|
||||||
.map_err(|e| format!("Failed to read response body: {e}"))?;
|
|
||||||
println!("{}", body);
|
|
||||||
Ok(())
|
Ok(())
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -4,6 +4,8 @@ use crate::utils::confirm::confirm_delete;
|
|||||||
use crate::utils::json::{
|
use crate::utils::json::{
|
||||||
apply_merge_patch, parse_optional_json, parse_required_json, require_id, validate_create_id,
|
apply_merge_patch, parse_optional_json, parse_required_json, require_id, validate_create_id,
|
||||||
};
|
};
|
||||||
|
use crate::utils::schema::append_agent_hints;
|
||||||
|
use schemars::schema_for;
|
||||||
use yaak_models::models::Workspace;
|
use yaak_models::models::Workspace;
|
||||||
use yaak_models::util::UpdateSource;
|
use yaak_models::util::UpdateSource;
|
||||||
|
|
||||||
@@ -12,6 +14,7 @@ type CommandResult<T = ()> = std::result::Result<T, String>;
|
|||||||
pub fn run(ctx: &CliContext, args: WorkspaceArgs) -> i32 {
|
pub fn run(ctx: &CliContext, args: WorkspaceArgs) -> i32 {
|
||||||
let result = match args.command {
|
let result = match args.command {
|
||||||
WorkspaceCommands::List => list(ctx),
|
WorkspaceCommands::List => list(ctx),
|
||||||
|
WorkspaceCommands::Schema { pretty } => schema(pretty),
|
||||||
WorkspaceCommands::Show { workspace_id } => show(ctx, &workspace_id),
|
WorkspaceCommands::Show { workspace_id } => show(ctx, &workspace_id),
|
||||||
WorkspaceCommands::Create { name, json, json_input } => create(ctx, name, json, json_input),
|
WorkspaceCommands::Create { name, json, json_input } => create(ctx, name, json, json_input),
|
||||||
WorkspaceCommands::Update { json, json_input } => update(ctx, json, json_input),
|
WorkspaceCommands::Update { json, json_input } => update(ctx, json, json_input),
|
||||||
@@ -27,6 +30,23 @@ pub fn run(ctx: &CliContext, args: WorkspaceArgs) -> i32 {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn schema(pretty: bool) -> CommandResult {
|
||||||
|
let mut schema =
|
||||||
|
serde_json::to_value(schema_for!(Workspace)).map_err(|e| format!(
|
||||||
|
"Failed to serialize workspace schema: {e}"
|
||||||
|
))?;
|
||||||
|
append_agent_hints(&mut schema);
|
||||||
|
|
||||||
|
let output = if pretty {
|
||||||
|
serde_json::to_string_pretty(&schema)
|
||||||
|
} else {
|
||||||
|
serde_json::to_string(&schema)
|
||||||
|
}
|
||||||
|
.map_err(|e| format!("Failed to format workspace schema JSON: {e}"))?;
|
||||||
|
println!("{output}");
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
fn list(ctx: &CliContext) -> CommandResult {
|
fn list(ctx: &CliContext) -> CommandResult {
|
||||||
let workspaces =
|
let workspaces =
|
||||||
ctx.db().list_workspaces().map_err(|e| format!("Failed to list workspaces: {e}"))?;
|
ctx.db().list_workspaces().map_err(|e| format!("Failed to list workspaces: {e}"))?;
|
||||||
|
|||||||
@@ -1,4 +1,6 @@
|
|||||||
use crate::plugin_events::CliPluginEventBridge;
|
use crate::plugin_events::CliPluginEventBridge;
|
||||||
|
use include_dir::{Dir, include_dir};
|
||||||
|
use std::fs;
|
||||||
use std::path::{Path, PathBuf};
|
use std::path::{Path, PathBuf};
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use tokio::sync::Mutex;
|
use tokio::sync::Mutex;
|
||||||
@@ -9,6 +11,13 @@ use yaak_models::query_manager::QueryManager;
|
|||||||
use yaak_plugins::events::PluginContext;
|
use yaak_plugins::events::PluginContext;
|
||||||
use yaak_plugins::manager::PluginManager;
|
use yaak_plugins::manager::PluginManager;
|
||||||
|
|
||||||
|
const EMBEDDED_PLUGIN_RUNTIME: &str = include_str!(concat!(
|
||||||
|
env!("CARGO_MANIFEST_DIR"),
|
||||||
|
"/../../crates-tauri/yaak-app/vendored/plugin-runtime/index.cjs"
|
||||||
|
));
|
||||||
|
static EMBEDDED_VENDORED_PLUGINS: Dir<'_> =
|
||||||
|
include_dir!("$CARGO_MANIFEST_DIR/../../crates-tauri/yaak-app/vendored/plugins");
|
||||||
|
|
||||||
pub struct CliContext {
|
pub struct CliContext {
|
||||||
data_dir: PathBuf,
|
data_dir: PathBuf,
|
||||||
query_manager: QueryManager,
|
query_manager: QueryManager,
|
||||||
@@ -29,41 +38,40 @@ impl CliContext {
|
|||||||
let encryption_manager = Arc::new(EncryptionManager::new(query_manager.clone(), app_id));
|
let encryption_manager = Arc::new(EncryptionManager::new(query_manager.clone(), app_id));
|
||||||
|
|
||||||
let plugin_manager = if with_plugins {
|
let plugin_manager = if with_plugins {
|
||||||
let vendored_plugin_dir = data_dir.join("vendored-plugins");
|
let embedded_vendored_plugin_dir = data_dir.join("vendored-plugins");
|
||||||
|
let bundled_plugin_dir =
|
||||||
|
resolve_bundled_plugin_dir_for_cli(&embedded_vendored_plugin_dir);
|
||||||
let installed_plugin_dir = data_dir.join("installed-plugins");
|
let installed_plugin_dir = data_dir.join("installed-plugins");
|
||||||
let node_bin_path = PathBuf::from("node");
|
let node_bin_path = PathBuf::from("node");
|
||||||
|
|
||||||
let plugin_runtime_main =
|
if bundled_plugin_dir == embedded_vendored_plugin_dir {
|
||||||
std::env::var("YAAK_PLUGIN_RUNTIME").map(PathBuf::from).unwrap_or_else(|_| {
|
prepare_embedded_vendored_plugins(&embedded_vendored_plugin_dir)
|
||||||
PathBuf::from(env!("CARGO_MANIFEST_DIR"))
|
.expect("Failed to prepare bundled plugins");
|
||||||
.join("../../crates-tauri/yaak-app/vendored/plugin-runtime/index.cjs")
|
|
||||||
});
|
|
||||||
|
|
||||||
let plugin_manager = Arc::new(
|
|
||||||
PluginManager::new(
|
|
||||||
vendored_plugin_dir,
|
|
||||||
installed_plugin_dir,
|
|
||||||
node_bin_path,
|
|
||||||
plugin_runtime_main,
|
|
||||||
false,
|
|
||||||
)
|
|
||||||
.await,
|
|
||||||
);
|
|
||||||
|
|
||||||
let plugins = query_manager.connect().list_plugins().unwrap_or_default();
|
|
||||||
if !plugins.is_empty() {
|
|
||||||
let errors = plugin_manager
|
|
||||||
.initialize_all_plugins(plugins, &PluginContext::new_empty())
|
|
||||||
.await;
|
|
||||||
for (plugin_dir, error_msg) in errors {
|
|
||||||
eprintln!(
|
|
||||||
"Warning: Failed to initialize plugin '{}': {}",
|
|
||||||
plugin_dir, error_msg
|
|
||||||
);
|
|
||||||
}
|
|
||||||
}
|
}
|
||||||
|
|
||||||
Some(plugin_manager)
|
let plugin_runtime_main =
|
||||||
|
std::env::var("YAAK_PLUGIN_RUNTIME").map(PathBuf::from).unwrap_or_else(|_| {
|
||||||
|
prepare_embedded_plugin_runtime(&data_dir)
|
||||||
|
.expect("Failed to prepare embedded plugin runtime")
|
||||||
|
});
|
||||||
|
|
||||||
|
match PluginManager::new(
|
||||||
|
bundled_plugin_dir,
|
||||||
|
embedded_vendored_plugin_dir,
|
||||||
|
installed_plugin_dir,
|
||||||
|
node_bin_path,
|
||||||
|
plugin_runtime_main,
|
||||||
|
&query_manager,
|
||||||
|
&PluginContext::new_empty(),
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
{
|
||||||
|
Ok(plugin_manager) => Some(Arc::new(plugin_manager)),
|
||||||
|
Err(err) => {
|
||||||
|
eprintln!("Warning: Failed to initialize plugins: {err}");
|
||||||
|
None
|
||||||
|
}
|
||||||
|
}
|
||||||
} else {
|
} else {
|
||||||
None
|
None
|
||||||
};
|
};
|
||||||
@@ -113,3 +121,34 @@ impl CliContext {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn prepare_embedded_plugin_runtime(data_dir: &Path) -> std::io::Result<PathBuf> {
|
||||||
|
let runtime_dir = data_dir.join("vendored").join("plugin-runtime");
|
||||||
|
fs::create_dir_all(&runtime_dir)?;
|
||||||
|
let runtime_main = runtime_dir.join("index.cjs");
|
||||||
|
fs::write(&runtime_main, EMBEDDED_PLUGIN_RUNTIME)?;
|
||||||
|
Ok(runtime_main)
|
||||||
|
}
|
||||||
|
|
||||||
|
fn prepare_embedded_vendored_plugins(vendored_plugin_dir: &Path) -> std::io::Result<()> {
|
||||||
|
fs::create_dir_all(vendored_plugin_dir)?;
|
||||||
|
EMBEDDED_VENDORED_PLUGINS.extract(vendored_plugin_dir)?;
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn resolve_bundled_plugin_dir_for_cli(embedded_vendored_plugin_dir: &Path) -> PathBuf {
|
||||||
|
if !cfg!(debug_assertions) {
|
||||||
|
return embedded_vendored_plugin_dir.to_path_buf();
|
||||||
|
}
|
||||||
|
|
||||||
|
let plugins_dir = match std::env::current_dir() {
|
||||||
|
Ok(cwd) => cwd.join("plugins"),
|
||||||
|
Err(_) => return embedded_vendored_plugin_dir.to_path_buf(),
|
||||||
|
};
|
||||||
|
|
||||||
|
if !plugins_dir.is_dir() {
|
||||||
|
return embedded_vendored_plugin_dir.to_path_buf();
|
||||||
|
}
|
||||||
|
|
||||||
|
plugins_dir.canonicalize().unwrap_or(plugins_dir)
|
||||||
|
}
|
||||||
|
|||||||
@@ -2,7 +2,9 @@ mod cli;
|
|||||||
mod commands;
|
mod commands;
|
||||||
mod context;
|
mod context;
|
||||||
mod plugin_events;
|
mod plugin_events;
|
||||||
|
mod ui;
|
||||||
mod utils;
|
mod utils;
|
||||||
|
mod version;
|
||||||
|
|
||||||
use clap::Parser;
|
use clap::Parser;
|
||||||
use cli::{Cli, Commands, RequestCommands};
|
use cli::{Cli, Commands, RequestCommands};
|
||||||
@@ -10,10 +12,18 @@ use context::CliContext;
|
|||||||
|
|
||||||
#[tokio::main]
|
#[tokio::main]
|
||||||
async fn main() {
|
async fn main() {
|
||||||
let Cli { data_dir, environment, verbose, command } = Cli::parse();
|
let Cli { data_dir, environment, verbose, log, command } = Cli::parse();
|
||||||
|
|
||||||
if verbose {
|
if let Some(log_level) = log {
|
||||||
env_logger::Builder::from_env(env_logger::Env::default().default_filter_or("info")).init();
|
match log_level {
|
||||||
|
Some(level) => {
|
||||||
|
env_logger::Builder::new().filter_level(level.as_filter()).init();
|
||||||
|
}
|
||||||
|
None => {
|
||||||
|
env_logger::Builder::from_env(env_logger::Env::default().default_filter_or("info"))
|
||||||
|
.init();
|
||||||
|
}
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
let app_id = if cfg!(debug_assertions) { "app.yaak.desktop.dev" } else { "app.yaak.desktop" };
|
let app_id = if cfg!(debug_assertions) { "app.yaak.desktop.dev" } else { "app.yaak.desktop" };
|
||||||
@@ -22,6 +32,15 @@ async fn main() {
|
|||||||
dirs::data_dir().expect("Could not determine data directory").join(app_id)
|
dirs::data_dir().expect("Could not determine data directory").join(app_id)
|
||||||
});
|
});
|
||||||
|
|
||||||
|
let needs_context = matches!(
|
||||||
|
&command,
|
||||||
|
Commands::Send(_)
|
||||||
|
| Commands::Workspace(_)
|
||||||
|
| Commands::Request(_)
|
||||||
|
| Commands::Folder(_)
|
||||||
|
| Commands::Environment(_)
|
||||||
|
);
|
||||||
|
|
||||||
let needs_plugins = matches!(
|
let needs_plugins = matches!(
|
||||||
&command,
|
&command,
|
||||||
Commands::Send(_)
|
Commands::Send(_)
|
||||||
@@ -30,21 +49,51 @@ async fn main() {
|
|||||||
})
|
})
|
||||||
);
|
);
|
||||||
|
|
||||||
let context = CliContext::initialize(data_dir, app_id, needs_plugins).await;
|
let context = if needs_context {
|
||||||
|
Some(CliContext::initialize(data_dir, app_id, needs_plugins).await)
|
||||||
let exit_code = match command {
|
} else {
|
||||||
Commands::Send(args) => {
|
None
|
||||||
commands::send::run(&context, args, environment.as_deref(), verbose).await
|
|
||||||
}
|
|
||||||
Commands::Workspace(args) => commands::workspace::run(&context, args),
|
|
||||||
Commands::Request(args) => {
|
|
||||||
commands::request::run(&context, args, environment.as_deref(), verbose).await
|
|
||||||
}
|
|
||||||
Commands::Folder(args) => commands::folder::run(&context, args),
|
|
||||||
Commands::Environment(args) => commands::environment::run(&context, args),
|
|
||||||
};
|
};
|
||||||
|
|
||||||
context.shutdown().await;
|
let exit_code = match command {
|
||||||
|
Commands::Auth(args) => commands::auth::run(args).await,
|
||||||
|
Commands::Plugin(args) => commands::plugin::run(args).await,
|
||||||
|
Commands::Build(args) => commands::plugin::run_build(args).await,
|
||||||
|
Commands::Dev(args) => commands::plugin::run_dev(args).await,
|
||||||
|
Commands::Send(args) => {
|
||||||
|
commands::send::run(
|
||||||
|
context.as_ref().expect("context initialized for send"),
|
||||||
|
args,
|
||||||
|
environment.as_deref(),
|
||||||
|
verbose,
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
}
|
||||||
|
Commands::Workspace(args) => commands::workspace::run(
|
||||||
|
context.as_ref().expect("context initialized for workspace"),
|
||||||
|
args,
|
||||||
|
),
|
||||||
|
Commands::Request(args) => {
|
||||||
|
commands::request::run(
|
||||||
|
context.as_ref().expect("context initialized for request"),
|
||||||
|
args,
|
||||||
|
environment.as_deref(),
|
||||||
|
verbose,
|
||||||
|
)
|
||||||
|
.await
|
||||||
|
}
|
||||||
|
Commands::Folder(args) => {
|
||||||
|
commands::folder::run(context.as_ref().expect("context initialized for folder"), args)
|
||||||
|
}
|
||||||
|
Commands::Environment(args) => commands::environment::run(
|
||||||
|
context.as_ref().expect("context initialized for environment"),
|
||||||
|
args,
|
||||||
|
),
|
||||||
|
};
|
||||||
|
|
||||||
|
if let Some(context) = &context {
|
||||||
|
context.shutdown().await;
|
||||||
|
}
|
||||||
|
|
||||||
if exit_code != 0 {
|
if exit_code != 0 {
|
||||||
std::process::exit(exit_code);
|
std::process::exit(exit_code);
|
||||||
|
|||||||
34
crates-cli/yaak-cli/src/ui.rs
Normal file
34
crates-cli/yaak-cli/src/ui.rs
Normal file
@@ -0,0 +1,34 @@
|
|||||||
|
use console::style;
|
||||||
|
use std::io::{self, IsTerminal};
|
||||||
|
|
||||||
|
pub fn info(message: &str) {
|
||||||
|
if io::stdout().is_terminal() {
|
||||||
|
println!("{:<8} {}", style("INFO").cyan().bold(), style(message).cyan());
|
||||||
|
} else {
|
||||||
|
println!("INFO {message}");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn warning(message: &str) {
|
||||||
|
if io::stdout().is_terminal() {
|
||||||
|
println!("{:<8} {}", style("WARNING").yellow().bold(), style(message).yellow());
|
||||||
|
} else {
|
||||||
|
println!("WARNING {message}");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn success(message: &str) {
|
||||||
|
if io::stdout().is_terminal() {
|
||||||
|
println!("{:<8} {}", style("SUCCESS").green().bold(), style(message).green());
|
||||||
|
} else {
|
||||||
|
println!("SUCCESS {message}");
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn error(message: &str) {
|
||||||
|
if io::stderr().is_terminal() {
|
||||||
|
eprintln!("{:<8} {}", style("ERROR").red().bold(), style(message).red());
|
||||||
|
} else {
|
||||||
|
eprintln!("Error: {message}");
|
||||||
|
}
|
||||||
|
}
|
||||||
47
crates-cli/yaak-cli/src/utils/http.rs
Normal file
47
crates-cli/yaak-cli/src/utils/http.rs
Normal file
@@ -0,0 +1,47 @@
|
|||||||
|
use reqwest::Client;
|
||||||
|
use reqwest::header::{HeaderMap, HeaderName, HeaderValue, USER_AGENT};
|
||||||
|
use serde_json::Value;
|
||||||
|
|
||||||
|
pub fn build_client(session_token: Option<&str>) -> Result<Client, String> {
|
||||||
|
let mut headers = HeaderMap::new();
|
||||||
|
let user_agent = HeaderValue::from_str(&user_agent())
|
||||||
|
.map_err(|e| format!("Failed to build user-agent header: {e}"))?;
|
||||||
|
headers.insert(USER_AGENT, user_agent);
|
||||||
|
|
||||||
|
if let Some(token) = session_token {
|
||||||
|
let token_value = HeaderValue::from_str(token)
|
||||||
|
.map_err(|e| format!("Failed to build session header: {e}"))?;
|
||||||
|
headers.insert(HeaderName::from_static("x-yaak-session"), token_value);
|
||||||
|
}
|
||||||
|
|
||||||
|
Client::builder()
|
||||||
|
.default_headers(headers)
|
||||||
|
.build()
|
||||||
|
.map_err(|e| format!("Failed to initialize HTTP client: {e}"))
|
||||||
|
}
|
||||||
|
|
||||||
|
pub fn parse_api_error(status: u16, body: &str) -> String {
|
||||||
|
if let Ok(value) = serde_json::from_str::<Value>(body) {
|
||||||
|
if let Some(message) = value.get("message").and_then(Value::as_str) {
|
||||||
|
return message.to_string();
|
||||||
|
}
|
||||||
|
if let Some(error) = value.get("error").and_then(Value::as_str) {
|
||||||
|
return error.to_string();
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
format!("API error {status}: {body}")
|
||||||
|
}
|
||||||
|
|
||||||
|
fn user_agent() -> String {
|
||||||
|
format!("YaakCli/{} ({})", crate::version::cli_version(), ua_platform())
|
||||||
|
}
|
||||||
|
|
||||||
|
fn ua_platform() -> &'static str {
|
||||||
|
match std::env::consts::OS {
|
||||||
|
"windows" => "Win",
|
||||||
|
"darwin" => "Mac",
|
||||||
|
"linux" => "Linux",
|
||||||
|
_ => "Unknown",
|
||||||
|
}
|
||||||
|
}
|
||||||
@@ -63,6 +63,30 @@ pub fn validate_create_id(payload: &Value, context: &str) -> JsonResult<()> {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
pub fn merge_workspace_id_arg(
|
||||||
|
workspace_id_from_arg: Option<&str>,
|
||||||
|
payload_workspace_id: &mut String,
|
||||||
|
context: &str,
|
||||||
|
) -> JsonResult<()> {
|
||||||
|
if let Some(workspace_id_arg) = workspace_id_from_arg {
|
||||||
|
if payload_workspace_id.is_empty() {
|
||||||
|
*payload_workspace_id = workspace_id_arg.to_string();
|
||||||
|
} else if payload_workspace_id != workspace_id_arg {
|
||||||
|
return Err(format!(
|
||||||
|
"{context} got conflicting workspace_id values between positional arg and JSON payload"
|
||||||
|
));
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if payload_workspace_id.is_empty() {
|
||||||
|
return Err(format!(
|
||||||
|
"{context} requires non-empty \"workspaceId\" in JSON payload or positional workspace_id"
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(())
|
||||||
|
}
|
||||||
|
|
||||||
pub fn apply_merge_patch<T>(existing: &T, patch: &Value, id: &str, context: &str) -> JsonResult<T>
|
pub fn apply_merge_patch<T>(existing: &T, patch: &Value, id: &str, context: &str) -> JsonResult<T>
|
||||||
where
|
where
|
||||||
T: Serialize + DeserializeOwned,
|
T: Serialize + DeserializeOwned,
|
||||||
|
|||||||
@@ -1,2 +1,4 @@
|
|||||||
pub mod confirm;
|
pub mod confirm;
|
||||||
|
pub mod http;
|
||||||
pub mod json;
|
pub mod json;
|
||||||
|
pub mod schema;
|
||||||
|
|||||||
15
crates-cli/yaak-cli/src/utils/schema.rs
Normal file
15
crates-cli/yaak-cli/src/utils/schema.rs
Normal file
@@ -0,0 +1,15 @@
|
|||||||
|
use serde_json::{Value, json};
|
||||||
|
|
||||||
|
pub fn append_agent_hints(schema: &mut Value) {
|
||||||
|
let Some(schema_obj) = schema.as_object_mut() else {
|
||||||
|
return;
|
||||||
|
};
|
||||||
|
|
||||||
|
schema_obj.insert(
|
||||||
|
"x-yaak-agent-hints".to_string(),
|
||||||
|
json!({
|
||||||
|
"templateVariableSyntax": "${[ my_var ]}",
|
||||||
|
"templateFunctionSyntax": "${[ namespace.my_func(a='aaa',b='bbb') ]}",
|
||||||
|
}),
|
||||||
|
);
|
||||||
|
}
|
||||||
3
crates-cli/yaak-cli/src/version.rs
Normal file
3
crates-cli/yaak-cli/src/version.rs
Normal file
@@ -0,0 +1,3 @@
|
|||||||
|
pub fn cli_version() -> &'static str {
|
||||||
|
option_env!("YAAK_CLI_VERSION").unwrap_or(env!("CARGO_PKG_VERSION"))
|
||||||
|
}
|
||||||
@@ -10,7 +10,7 @@ use yaak_models::query_manager::QueryManager;
|
|||||||
use yaak_models::util::UpdateSource;
|
use yaak_models::util::UpdateSource;
|
||||||
|
|
||||||
pub fn cli_cmd(data_dir: &Path) -> Command {
|
pub fn cli_cmd(data_dir: &Path) -> Command {
|
||||||
let mut cmd = cargo_bin_cmd!("yaakcli");
|
let mut cmd = cargo_bin_cmd!("yaak");
|
||||||
cmd.arg("--data-dir").arg(data_dir);
|
cmd.arg("--data-dir").arg(data_dir);
|
||||||
cmd
|
cmd
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -78,3 +78,69 @@ fn json_create_and_update_merge_patch_round_trip() {
|
|||||||
.stdout(contains("\"name\": \"Json Environment\""))
|
.stdout(contains("\"name\": \"Json Environment\""))
|
||||||
.stdout(contains("\"color\": \"#00ff00\""));
|
.stdout(contains("\"color\": \"#00ff00\""));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn create_merges_positional_workspace_id_into_json_payload() {
|
||||||
|
let temp_dir = TempDir::new().expect("Failed to create temp dir");
|
||||||
|
let data_dir = temp_dir.path();
|
||||||
|
seed_workspace(data_dir, "wk_test");
|
||||||
|
|
||||||
|
let create_assert = cli_cmd(data_dir)
|
||||||
|
.args([
|
||||||
|
"environment",
|
||||||
|
"create",
|
||||||
|
"wk_test",
|
||||||
|
"--json",
|
||||||
|
r#"{"name":"Merged Environment"}"#,
|
||||||
|
])
|
||||||
|
.assert()
|
||||||
|
.success();
|
||||||
|
let environment_id = parse_created_id(&create_assert.get_output().stdout, "environment create");
|
||||||
|
|
||||||
|
cli_cmd(data_dir)
|
||||||
|
.args(["environment", "show", &environment_id])
|
||||||
|
.assert()
|
||||||
|
.success()
|
||||||
|
.stdout(contains("\"workspaceId\": \"wk_test\""))
|
||||||
|
.stdout(contains("\"name\": \"Merged Environment\""));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn create_rejects_conflicting_workspace_ids_between_arg_and_json() {
|
||||||
|
let temp_dir = TempDir::new().expect("Failed to create temp dir");
|
||||||
|
let data_dir = temp_dir.path();
|
||||||
|
seed_workspace(data_dir, "wk_test");
|
||||||
|
seed_workspace(data_dir, "wk_other");
|
||||||
|
|
||||||
|
cli_cmd(data_dir)
|
||||||
|
.args([
|
||||||
|
"environment",
|
||||||
|
"create",
|
||||||
|
"wk_test",
|
||||||
|
"--json",
|
||||||
|
r#"{"workspaceId":"wk_other","name":"Mismatch"}"#,
|
||||||
|
])
|
||||||
|
.assert()
|
||||||
|
.failure()
|
||||||
|
.stderr(contains(
|
||||||
|
"environment create got conflicting workspace_id values between positional arg and JSON payload",
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn environment_schema_outputs_json_schema() {
|
||||||
|
let temp_dir = TempDir::new().expect("Failed to create temp dir");
|
||||||
|
let data_dir = temp_dir.path();
|
||||||
|
|
||||||
|
cli_cmd(data_dir)
|
||||||
|
.args(["environment", "schema"])
|
||||||
|
.assert()
|
||||||
|
.success()
|
||||||
|
.stdout(contains("\"type\":\"object\""))
|
||||||
|
.stdout(contains("\"x-yaak-agent-hints\""))
|
||||||
|
.stdout(contains("\"templateVariableSyntax\":\"${[ my_var ]}\""))
|
||||||
|
.stdout(contains(
|
||||||
|
"\"templateFunctionSyntax\":\"${[ namespace.my_func(a='aaa',b='bbb') ]}\"",
|
||||||
|
))
|
||||||
|
.stdout(contains("\"workspaceId\""));
|
||||||
|
}
|
||||||
|
|||||||
@@ -72,3 +72,51 @@ fn json_create_and_update_merge_patch_round_trip() {
|
|||||||
.stdout(contains("\"name\": \"Json Folder\""))
|
.stdout(contains("\"name\": \"Json Folder\""))
|
||||||
.stdout(contains("\"description\": \"Folder Description\""));
|
.stdout(contains("\"description\": \"Folder Description\""));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn create_merges_positional_workspace_id_into_json_payload() {
|
||||||
|
let temp_dir = TempDir::new().expect("Failed to create temp dir");
|
||||||
|
let data_dir = temp_dir.path();
|
||||||
|
seed_workspace(data_dir, "wk_test");
|
||||||
|
|
||||||
|
let create_assert = cli_cmd(data_dir)
|
||||||
|
.args([
|
||||||
|
"folder",
|
||||||
|
"create",
|
||||||
|
"wk_test",
|
||||||
|
"--json",
|
||||||
|
r#"{"name":"Merged Folder"}"#,
|
||||||
|
])
|
||||||
|
.assert()
|
||||||
|
.success();
|
||||||
|
let folder_id = parse_created_id(&create_assert.get_output().stdout, "folder create");
|
||||||
|
|
||||||
|
cli_cmd(data_dir)
|
||||||
|
.args(["folder", "show", &folder_id])
|
||||||
|
.assert()
|
||||||
|
.success()
|
||||||
|
.stdout(contains("\"workspaceId\": \"wk_test\""))
|
||||||
|
.stdout(contains("\"name\": \"Merged Folder\""));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn create_rejects_conflicting_workspace_ids_between_arg_and_json() {
|
||||||
|
let temp_dir = TempDir::new().expect("Failed to create temp dir");
|
||||||
|
let data_dir = temp_dir.path();
|
||||||
|
seed_workspace(data_dir, "wk_test");
|
||||||
|
seed_workspace(data_dir, "wk_other");
|
||||||
|
|
||||||
|
cli_cmd(data_dir)
|
||||||
|
.args([
|
||||||
|
"folder",
|
||||||
|
"create",
|
||||||
|
"wk_test",
|
||||||
|
"--json",
|
||||||
|
r#"{"workspaceId":"wk_other","name":"Mismatch"}"#,
|
||||||
|
])
|
||||||
|
.assert()
|
||||||
|
.failure()
|
||||||
|
.stderr(contains(
|
||||||
|
"folder create got conflicting workspace_id values between positional arg and JSON payload",
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|||||||
@@ -130,6 +130,54 @@ fn create_allows_workspace_only_with_empty_defaults() {
|
|||||||
assert_eq!(request.url, "");
|
assert_eq!(request.url, "");
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn create_merges_positional_workspace_id_into_json_payload() {
|
||||||
|
let temp_dir = TempDir::new().expect("Failed to create temp dir");
|
||||||
|
let data_dir = temp_dir.path();
|
||||||
|
seed_workspace(data_dir, "wk_test");
|
||||||
|
|
||||||
|
let create_assert = cli_cmd(data_dir)
|
||||||
|
.args([
|
||||||
|
"request",
|
||||||
|
"create",
|
||||||
|
"wk_test",
|
||||||
|
"--json",
|
||||||
|
r#"{"name":"Merged Request","url":"https://example.com"}"#,
|
||||||
|
])
|
||||||
|
.assert()
|
||||||
|
.success();
|
||||||
|
let request_id = parse_created_id(&create_assert.get_output().stdout, "request create");
|
||||||
|
|
||||||
|
cli_cmd(data_dir)
|
||||||
|
.args(["request", "show", &request_id])
|
||||||
|
.assert()
|
||||||
|
.success()
|
||||||
|
.stdout(contains("\"workspaceId\": \"wk_test\""))
|
||||||
|
.stdout(contains("\"name\": \"Merged Request\""));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn create_rejects_conflicting_workspace_ids_between_arg_and_json() {
|
||||||
|
let temp_dir = TempDir::new().expect("Failed to create temp dir");
|
||||||
|
let data_dir = temp_dir.path();
|
||||||
|
seed_workspace(data_dir, "wk_test");
|
||||||
|
seed_workspace(data_dir, "wk_other");
|
||||||
|
|
||||||
|
cli_cmd(data_dir)
|
||||||
|
.args([
|
||||||
|
"request",
|
||||||
|
"create",
|
||||||
|
"wk_test",
|
||||||
|
"--json",
|
||||||
|
r#"{"workspaceId":"wk_other","name":"Mismatch"}"#,
|
||||||
|
])
|
||||||
|
.assert()
|
||||||
|
.failure()
|
||||||
|
.stderr(contains(
|
||||||
|
"request create got conflicting workspace_id values between positional arg and JSON payload",
|
||||||
|
));
|
||||||
|
}
|
||||||
|
|
||||||
#[test]
|
#[test]
|
||||||
fn request_send_persists_response_body_and_events() {
|
fn request_send_persists_response_body_and_events() {
|
||||||
let temp_dir = TempDir::new().expect("Failed to create temp dir");
|
let temp_dir = TempDir::new().expect("Failed to create temp dir");
|
||||||
@@ -156,7 +204,6 @@ fn request_send_persists_response_body_and_events() {
|
|||||||
.args(["request", "send", &request_id])
|
.args(["request", "send", &request_id])
|
||||||
.assert()
|
.assert()
|
||||||
.success()
|
.success()
|
||||||
.stdout(contains("HTTP 200 OK"))
|
|
||||||
.stdout(contains("hello from integration test"));
|
.stdout(contains("hello from integration test"));
|
||||||
|
|
||||||
let qm = query_manager(data_dir);
|
let qm = query_manager(data_dir);
|
||||||
@@ -189,6 +236,26 @@ fn request_schema_http_outputs_json_schema() {
|
|||||||
.args(["request", "schema", "http"])
|
.args(["request", "schema", "http"])
|
||||||
.assert()
|
.assert()
|
||||||
.success()
|
.success()
|
||||||
|
.stdout(contains("\"type\":\"object\""))
|
||||||
|
.stdout(contains("\"x-yaak-agent-hints\""))
|
||||||
|
.stdout(contains("\"templateVariableSyntax\":\"${[ my_var ]}\""))
|
||||||
|
.stdout(contains(
|
||||||
|
"\"templateFunctionSyntax\":\"${[ namespace.my_func(a='aaa',b='bbb') ]}\"",
|
||||||
|
))
|
||||||
|
.stdout(contains("\"authentication\":"))
|
||||||
|
.stdout(contains("/foo/:id/comments/:commentId"))
|
||||||
|
.stdout(contains("put concrete values in `urlParameters`"));
|
||||||
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn request_schema_http_pretty_prints_with_flag() {
|
||||||
|
let temp_dir = TempDir::new().expect("Failed to create temp dir");
|
||||||
|
let data_dir = temp_dir.path();
|
||||||
|
|
||||||
|
cli_cmd(data_dir)
|
||||||
|
.args(["request", "schema", "http", "--pretty"])
|
||||||
|
.assert()
|
||||||
|
.success()
|
||||||
.stdout(contains("\"type\": \"object\""))
|
.stdout(contains("\"type\": \"object\""))
|
||||||
.stdout(contains("\"authentication\""));
|
.stdout(contains("\"authentication\""));
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -31,7 +31,6 @@ fn top_level_send_workspace_sends_http_requests_and_prints_summary() {
|
|||||||
.args(["send", "wk_test"])
|
.args(["send", "wk_test"])
|
||||||
.assert()
|
.assert()
|
||||||
.success()
|
.success()
|
||||||
.stdout(contains("HTTP 200 OK"))
|
|
||||||
.stdout(contains("workspace bulk send"))
|
.stdout(contains("workspace bulk send"))
|
||||||
.stdout(contains("Send summary: 1 succeeded, 0 failed"));
|
.stdout(contains("Send summary: 1 succeeded, 0 failed"));
|
||||||
}
|
}
|
||||||
@@ -62,7 +61,6 @@ fn top_level_send_folder_sends_http_requests_and_prints_summary() {
|
|||||||
.args(["send", "fl_test"])
|
.args(["send", "fl_test"])
|
||||||
.assert()
|
.assert()
|
||||||
.success()
|
.success()
|
||||||
.stdout(contains("HTTP 200 OK"))
|
|
||||||
.stdout(contains("folder bulk send"))
|
.stdout(contains("folder bulk send"))
|
||||||
.stdout(contains("Send summary: 1 succeeded, 0 failed"));
|
.stdout(contains("Send summary: 1 succeeded, 0 failed"));
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -57,3 +57,19 @@ fn json_create_and_update_merge_patch_round_trip() {
|
|||||||
.stdout(contains("\"name\": \"Json Workspace\""))
|
.stdout(contains("\"name\": \"Json Workspace\""))
|
||||||
.stdout(contains("\"description\": \"Updated via JSON\""));
|
.stdout(contains("\"description\": \"Updated via JSON\""));
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#[test]
|
||||||
|
fn workspace_schema_outputs_json_schema() {
|
||||||
|
let temp_dir = TempDir::new().expect("Failed to create temp dir");
|
||||||
|
let data_dir = temp_dir.path();
|
||||||
|
|
||||||
|
cli_cmd(data_dir)
|
||||||
|
.args(["workspace", "schema"])
|
||||||
|
.assert()
|
||||||
|
.success()
|
||||||
|
.stdout(contains("\"type\":\"object\""))
|
||||||
|
.stdout(contains("\"x-yaak-agent-hints\""))
|
||||||
|
.stdout(contains("\"templateVariableSyntax\":\"${[ my_var ]}\""))
|
||||||
|
.stdout(contains("\"templateFunctionSyntax\":\"${[ namespace.my_func(a='aaa',b='bbb') ]}\""))
|
||||||
|
.stdout(contains("\"name\""));
|
||||||
|
}
|
||||||
|
|||||||
@@ -154,6 +154,7 @@ async fn send_http_request_inner<R: Runtime>(
|
|||||||
cookie_jar_id,
|
cookie_jar_id,
|
||||||
response_dir: &response_dir,
|
response_dir: &response_dir,
|
||||||
emit_events_to: None,
|
emit_events_to: None,
|
||||||
|
emit_response_body_chunks_to: None,
|
||||||
existing_response: Some(response_ctx.response().clone()),
|
existing_response: Some(response_ctx.response().clone()),
|
||||||
plugin_manager,
|
plugin_manager,
|
||||||
encryption_manager,
|
encryption_manager,
|
||||||
|
|||||||
@@ -362,7 +362,7 @@ async fn handle_host_plugin_request<R: Runtime>(
|
|||||||
workspace_id: http_request.workspace_id.clone(),
|
workspace_id: http_request.workspace_id.clone(),
|
||||||
..Default::default()
|
..Default::default()
|
||||||
},
|
},
|
||||||
&UpdateSource::Plugin,
|
&UpdateSource::from_window_label(window.label()),
|
||||||
&blobs,
|
&blobs,
|
||||||
)?
|
)?
|
||||||
};
|
};
|
||||||
|
|||||||
@@ -10,6 +10,7 @@ use crate::error::Result;
|
|||||||
use crate::models_ext::QueryManagerExt;
|
use crate::models_ext::QueryManagerExt;
|
||||||
use log::{error, info, warn};
|
use log::{error, info, warn};
|
||||||
use serde::Serialize;
|
use serde::Serialize;
|
||||||
|
use std::path::PathBuf;
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use std::sync::atomic::{AtomicBool, Ordering};
|
use std::sync::atomic::{AtomicBool, Ordering};
|
||||||
use std::time::{Duration, Instant};
|
use std::time::{Duration, Instant};
|
||||||
@@ -23,12 +24,11 @@ use tokio::sync::Mutex;
|
|||||||
use ts_rs::TS;
|
use ts_rs::TS;
|
||||||
use yaak_api::yaak_api_client;
|
use yaak_api::yaak_api_client;
|
||||||
use yaak_models::models::Plugin;
|
use yaak_models::models::Plugin;
|
||||||
use yaak_models::util::UpdateSource;
|
|
||||||
use yaak_plugins::api::{
|
use yaak_plugins::api::{
|
||||||
PluginNameVersion, PluginSearchResponse, PluginUpdatesResponse, check_plugin_updates,
|
PluginNameVersion, PluginSearchResponse, PluginUpdatesResponse, check_plugin_updates,
|
||||||
search_plugins,
|
search_plugins,
|
||||||
};
|
};
|
||||||
use yaak_plugins::events::{Color, Icon, PluginContext, ShowToastRequest};
|
use yaak_plugins::events::PluginContext;
|
||||||
use yaak_plugins::install::{delete_and_uninstall, download_and_install};
|
use yaak_plugins::install::{delete_and_uninstall, download_and_install};
|
||||||
use yaak_plugins::manager::PluginManager;
|
use yaak_plugins::manager::PluginManager;
|
||||||
use yaak_plugins::plugin_meta::get_plugin_meta;
|
use yaak_plugins::plugin_meta::get_plugin_meta;
|
||||||
@@ -244,6 +244,11 @@ pub fn init<R: Runtime>() -> TauriPlugin<R> {
|
|||||||
.path()
|
.path()
|
||||||
.resolve("vendored/plugins", BaseDirectory::Resource)
|
.resolve("vendored/plugins", BaseDirectory::Resource)
|
||||||
.expect("failed to resolve plugin directory resource");
|
.expect("failed to resolve plugin directory resource");
|
||||||
|
let bundled_plugin_dir = if is_dev() {
|
||||||
|
resolve_workspace_plugins_dir().unwrap_or_else(|| vendored_plugin_dir.clone())
|
||||||
|
} else {
|
||||||
|
vendored_plugin_dir.clone()
|
||||||
|
};
|
||||||
|
|
||||||
let installed_plugin_dir = app_handle
|
let installed_plugin_dir = app_handle
|
||||||
.path()
|
.path()
|
||||||
@@ -267,63 +272,23 @@ pub fn init<R: Runtime>() -> TauriPlugin<R> {
|
|||||||
.expect("failed to resolve plugin runtime")
|
.expect("failed to resolve plugin runtime")
|
||||||
.join("index.cjs");
|
.join("index.cjs");
|
||||||
|
|
||||||
let dev_mode = is_dev();
|
let query_manager =
|
||||||
|
app_handle.state::<yaak_models::query_manager::QueryManager>().inner().clone();
|
||||||
|
|
||||||
// Create plugin manager asynchronously
|
// Create plugin manager asynchronously
|
||||||
let app_handle_clone = app_handle.clone();
|
let app_handle_clone = app_handle.clone();
|
||||||
tauri::async_runtime::block_on(async move {
|
tauri::async_runtime::block_on(async move {
|
||||||
let manager = PluginManager::new(
|
let manager = PluginManager::new(
|
||||||
|
bundled_plugin_dir,
|
||||||
vendored_plugin_dir,
|
vendored_plugin_dir,
|
||||||
installed_plugin_dir,
|
installed_plugin_dir,
|
||||||
node_bin_path,
|
node_bin_path,
|
||||||
plugin_runtime_main,
|
plugin_runtime_main,
|
||||||
dev_mode,
|
&query_manager,
|
||||||
|
&PluginContext::new_empty(),
|
||||||
)
|
)
|
||||||
.await;
|
.await
|
||||||
|
.expect("Failed to initialize plugins");
|
||||||
// Initialize all plugins after manager is created
|
|
||||||
let bundled_dirs = manager
|
|
||||||
.list_bundled_plugin_dirs()
|
|
||||||
.await
|
|
||||||
.expect("Failed to list bundled plugins");
|
|
||||||
|
|
||||||
// Ensure all bundled plugins make it into the database
|
|
||||||
let db = app_handle_clone.db();
|
|
||||||
for dir in &bundled_dirs {
|
|
||||||
if db.get_plugin_by_directory(dir).is_none() {
|
|
||||||
db.upsert_plugin(
|
|
||||||
&Plugin {
|
|
||||||
directory: dir.clone(),
|
|
||||||
enabled: true,
|
|
||||||
url: None,
|
|
||||||
..Default::default()
|
|
||||||
},
|
|
||||||
&UpdateSource::Background,
|
|
||||||
)
|
|
||||||
.expect("Failed to upsert bundled plugin");
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
// Get all plugins from database and initialize
|
|
||||||
let plugins = db.list_plugins().expect("Failed to list plugins from database");
|
|
||||||
drop(db); // Explicitly drop the connection before await
|
|
||||||
|
|
||||||
let errors =
|
|
||||||
manager.initialize_all_plugins(plugins, &PluginContext::new_empty()).await;
|
|
||||||
|
|
||||||
// Show toast for any failed plugins
|
|
||||||
for (plugin_dir, error_msg) in errors {
|
|
||||||
let plugin_name = plugin_dir.split('/').last().unwrap_or(&plugin_dir);
|
|
||||||
let toast = ShowToastRequest {
|
|
||||||
message: format!("Failed to start plugin '{}': {}", plugin_name, error_msg),
|
|
||||||
color: Some(Color::Danger),
|
|
||||||
icon: Some(Icon::AlertTriangle),
|
|
||||||
timeout: Some(10000),
|
|
||||||
};
|
|
||||||
if let Err(emit_err) = app_handle_clone.emit("show_toast", toast) {
|
|
||||||
error!("Failed to emit toast for plugin error: {emit_err:?}");
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
app_handle_clone.manage(manager);
|
app_handle_clone.manage(manager);
|
||||||
});
|
});
|
||||||
@@ -362,3 +327,11 @@ pub fn init<R: Runtime>() -> TauriPlugin<R> {
|
|||||||
})
|
})
|
||||||
.build()
|
.build()
|
||||||
}
|
}
|
||||||
|
|
||||||
|
fn resolve_workspace_plugins_dir() -> Option<PathBuf> {
|
||||||
|
PathBuf::from(env!("CARGO_MANIFEST_DIR"))
|
||||||
|
.join("../..")
|
||||||
|
.join("plugins")
|
||||||
|
.canonicalize()
|
||||||
|
.ok()
|
||||||
|
}
|
||||||
|
|||||||
@@ -55,6 +55,7 @@ mod tests {
|
|||||||
|
|
||||||
let mut out = Vec::new();
|
let mut out = Vec::new();
|
||||||
super::collect_any_types(json, &mut out);
|
super::collect_any_types(json, &mut out);
|
||||||
|
out.sort();
|
||||||
assert_eq!(out, vec!["foo.bar", "mount_source.MountSourceRBDVolume"]);
|
assert_eq!(out, vec!["foo.bar", "mount_source.MountSourceRBDVolume"]);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -36,7 +36,6 @@ impl HttpConnectionManager {
|
|||||||
connections.retain(|_, (_, last_used)| last_used.elapsed() <= self.ttl);
|
connections.retain(|_, (_, last_used)| last_used.elapsed() <= self.ttl);
|
||||||
|
|
||||||
if let Some((cached, last_used)) = connections.get_mut(&id) {
|
if let Some((cached, last_used)) = connections.get_mut(&id) {
|
||||||
info!("Re-using HTTP client {id}");
|
|
||||||
*last_used = Instant::now();
|
*last_used = Instant::now();
|
||||||
return Ok(CachedClient {
|
return Ok(CachedClient {
|
||||||
client: cached.client.clone(),
|
client: cached.client.clone(),
|
||||||
|
|||||||
@@ -74,7 +74,7 @@ pub struct ClientCertificate {
|
|||||||
pub enabled: bool,
|
pub enabled: bool,
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone, PartialEq, Serialize, Deserialize, Default, TS)]
|
#[derive(Debug, Clone, PartialEq, Serialize, Deserialize, Default, JsonSchema, TS)]
|
||||||
#[serde(rename_all = "camelCase")]
|
#[serde(rename_all = "camelCase")]
|
||||||
#[ts(export, export_to = "gen_models.ts")]
|
#[ts(export, export_to = "gen_models.ts")]
|
||||||
pub struct DnsOverride {
|
pub struct DnsOverride {
|
||||||
@@ -293,7 +293,7 @@ impl UpsertModelInfo for Settings {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone, PartialEq, Serialize, Deserialize, Default, TS)]
|
#[derive(Debug, Clone, PartialEq, Serialize, Deserialize, Default, JsonSchema, TS)]
|
||||||
#[serde(default, rename_all = "camelCase")]
|
#[serde(default, rename_all = "camelCase")]
|
||||||
#[ts(export, export_to = "gen_models.ts")]
|
#[ts(export, export_to = "gen_models.ts")]
|
||||||
#[enum_def(table_name = "workspaces")]
|
#[enum_def(table_name = "workspaces")]
|
||||||
@@ -590,7 +590,7 @@ impl UpsertModelInfo for CookieJar {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone, PartialEq, Serialize, Deserialize, Default, TS)]
|
#[derive(Debug, Clone, PartialEq, Serialize, Deserialize, Default, JsonSchema, TS)]
|
||||||
#[serde(default, rename_all = "camelCase")]
|
#[serde(default, rename_all = "camelCase")]
|
||||||
#[ts(export, export_to = "gen_models.ts")]
|
#[ts(export, export_to = "gen_models.ts")]
|
||||||
#[enum_def(table_name = "environments")]
|
#[enum_def(table_name = "environments")]
|
||||||
@@ -611,6 +611,8 @@ pub struct Environment {
|
|||||||
pub base: bool,
|
pub base: bool,
|
||||||
pub parent_model: String,
|
pub parent_model: String,
|
||||||
pub parent_id: Option<String>,
|
pub parent_id: Option<String>,
|
||||||
|
/// Variables defined in this environment scope.
|
||||||
|
/// Child environments override parent variables by name.
|
||||||
pub variables: Vec<EnvironmentVariable>,
|
pub variables: Vec<EnvironmentVariable>,
|
||||||
pub color: Option<String>,
|
pub color: Option<String>,
|
||||||
pub sort_priority: f64,
|
pub sort_priority: f64,
|
||||||
@@ -698,7 +700,7 @@ impl UpsertModelInfo for Environment {
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
#[derive(Debug, Clone, PartialEq, Serialize, Deserialize, Default, TS)]
|
#[derive(Debug, Clone, PartialEq, Serialize, Deserialize, Default, JsonSchema, TS)]
|
||||||
#[serde(default, rename_all = "camelCase")]
|
#[serde(default, rename_all = "camelCase")]
|
||||||
#[ts(export, export_to = "gen_models.ts")]
|
#[ts(export, export_to = "gen_models.ts")]
|
||||||
pub struct EnvironmentVariable {
|
pub struct EnvironmentVariable {
|
||||||
@@ -845,6 +847,8 @@ pub struct HttpUrlParameter {
|
|||||||
#[serde(default = "default_true")]
|
#[serde(default = "default_true")]
|
||||||
#[ts(optional, as = "Option<bool>")]
|
#[ts(optional, as = "Option<bool>")]
|
||||||
pub enabled: bool,
|
pub enabled: bool,
|
||||||
|
/// Colon-prefixed parameters are treated as path parameters if they match, like `/users/:id`
|
||||||
|
/// Other entries are appended as query parameters
|
||||||
pub name: String,
|
pub name: String,
|
||||||
pub value: String,
|
pub value: String,
|
||||||
#[ts(optional, as = "Option<String>")]
|
#[ts(optional, as = "Option<String>")]
|
||||||
@@ -877,6 +881,7 @@ pub struct HttpRequest {
|
|||||||
pub name: String,
|
pub name: String,
|
||||||
pub sort_priority: f64,
|
pub sort_priority: f64,
|
||||||
pub url: String,
|
pub url: String,
|
||||||
|
/// URL parameters used for both path placeholders (`:id`) and query string entries.
|
||||||
pub url_parameters: Vec<HttpUrlParameter>,
|
pub url_parameters: Vec<HttpUrlParameter>,
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1118,6 +1123,7 @@ pub struct WebsocketRequest {
|
|||||||
pub name: String,
|
pub name: String,
|
||||||
pub sort_priority: f64,
|
pub sort_priority: f64,
|
||||||
pub url: String,
|
pub url: String,
|
||||||
|
/// URL parameters used for both path placeholders (`:id`) and query string entries.
|
||||||
pub url_parameters: Vec<HttpUrlParameter>,
|
pub url_parameters: Vec<HttpUrlParameter>,
|
||||||
}
|
}
|
||||||
|
|
||||||
@@ -1728,6 +1734,7 @@ pub struct GrpcRequest {
|
|||||||
pub name: String,
|
pub name: String,
|
||||||
pub service: Option<String>,
|
pub service: Option<String>,
|
||||||
pub sort_priority: f64,
|
pub sort_priority: f64,
|
||||||
|
/// Server URL (http for plaintext or https for secure)
|
||||||
pub url: String,
|
pub url: String,
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -24,7 +24,6 @@ use crate::plugin_handle::PluginHandle;
|
|||||||
use crate::server_ws::PluginRuntimeServerWebsocket;
|
use crate::server_ws::PluginRuntimeServerWebsocket;
|
||||||
use log::{error, info, warn};
|
use log::{error, info, warn};
|
||||||
use std::collections::HashMap;
|
use std::collections::HashMap;
|
||||||
use std::env;
|
|
||||||
use std::path::{Path, PathBuf};
|
use std::path::{Path, PathBuf};
|
||||||
use std::sync::Arc;
|
use std::sync::Arc;
|
||||||
use std::time::Duration;
|
use std::time::Duration;
|
||||||
@@ -34,7 +33,8 @@ use tokio::sync::mpsc::error::TrySendError;
|
|||||||
use tokio::sync::{Mutex, mpsc, oneshot};
|
use tokio::sync::{Mutex, mpsc, oneshot};
|
||||||
use tokio::time::{Instant, timeout};
|
use tokio::time::{Instant, timeout};
|
||||||
use yaak_models::models::Plugin;
|
use yaak_models::models::Plugin;
|
||||||
use yaak_models::util::generate_id;
|
use yaak_models::query_manager::QueryManager;
|
||||||
|
use yaak_models::util::{UpdateSource, generate_id};
|
||||||
use yaak_templates::error::Error::RenderError;
|
use yaak_templates::error::Error::RenderError;
|
||||||
use yaak_templates::error::Result as TemplateResult;
|
use yaak_templates::error::Result as TemplateResult;
|
||||||
|
|
||||||
@@ -45,9 +45,9 @@ pub struct PluginManager {
|
|||||||
kill_tx: tokio::sync::watch::Sender<bool>,
|
kill_tx: tokio::sync::watch::Sender<bool>,
|
||||||
killed_rx: Arc<Mutex<Option<oneshot::Receiver<()>>>>,
|
killed_rx: Arc<Mutex<Option<oneshot::Receiver<()>>>>,
|
||||||
ws_service: Arc<PluginRuntimeServerWebsocket>,
|
ws_service: Arc<PluginRuntimeServerWebsocket>,
|
||||||
|
bundled_plugin_dir: PathBuf,
|
||||||
vendored_plugin_dir: PathBuf,
|
vendored_plugin_dir: PathBuf,
|
||||||
pub(crate) installed_plugin_dir: PathBuf,
|
pub(crate) installed_plugin_dir: PathBuf,
|
||||||
dev_mode: bool,
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Callback for plugin initialization events (e.g., toast notifications)
|
/// Callback for plugin initialization events (e.g., toast notifications)
|
||||||
@@ -57,18 +57,22 @@ impl PluginManager {
|
|||||||
/// Create a new PluginManager with the given paths.
|
/// Create a new PluginManager with the given paths.
|
||||||
///
|
///
|
||||||
/// # Arguments
|
/// # Arguments
|
||||||
|
/// * `bundled_plugin_dir` - Directory to scan for bundled plugins
|
||||||
/// * `vendored_plugin_dir` - Path to vendored plugins directory
|
/// * `vendored_plugin_dir` - Path to vendored plugins directory
|
||||||
/// * `installed_plugin_dir` - Path to installed plugins directory
|
/// * `installed_plugin_dir` - Path to installed plugins directory
|
||||||
/// * `node_bin_path` - Path to the yaaknode binary
|
/// * `node_bin_path` - Path to the yaaknode binary
|
||||||
/// * `plugin_runtime_main` - Path to the plugin runtime index.cjs
|
/// * `plugin_runtime_main` - Path to the plugin runtime index.cjs
|
||||||
/// * `dev_mode` - Whether the app is in dev mode (affects plugin loading)
|
/// * `query_manager` - Query manager for bundled plugin registration and loading
|
||||||
|
/// * `plugin_context` - Context to use while initializing plugins
|
||||||
pub async fn new(
|
pub async fn new(
|
||||||
|
bundled_plugin_dir: PathBuf,
|
||||||
vendored_plugin_dir: PathBuf,
|
vendored_plugin_dir: PathBuf,
|
||||||
installed_plugin_dir: PathBuf,
|
installed_plugin_dir: PathBuf,
|
||||||
node_bin_path: PathBuf,
|
node_bin_path: PathBuf,
|
||||||
plugin_runtime_main: PathBuf,
|
plugin_runtime_main: PathBuf,
|
||||||
dev_mode: bool,
|
query_manager: &QueryManager,
|
||||||
) -> PluginManager {
|
plugin_context: &PluginContext,
|
||||||
|
) -> Result<PluginManager> {
|
||||||
let (events_tx, mut events_rx) = mpsc::channel(2048);
|
let (events_tx, mut events_rx) = mpsc::channel(2048);
|
||||||
let (kill_server_tx, kill_server_rx) = tokio::sync::watch::channel(false);
|
let (kill_server_tx, kill_server_rx) = tokio::sync::watch::channel(false);
|
||||||
let (killed_tx, killed_rx) = oneshot::channel();
|
let (killed_tx, killed_rx) = oneshot::channel();
|
||||||
@@ -84,9 +88,9 @@ impl PluginManager {
|
|||||||
ws_service: Arc::new(ws_service.clone()),
|
ws_service: Arc::new(ws_service.clone()),
|
||||||
kill_tx: kill_server_tx,
|
kill_tx: kill_server_tx,
|
||||||
killed_rx: Arc::new(Mutex::new(Some(killed_rx))),
|
killed_rx: Arc::new(Mutex::new(Some(killed_rx))),
|
||||||
|
bundled_plugin_dir,
|
||||||
vendored_plugin_dir,
|
vendored_plugin_dir,
|
||||||
installed_plugin_dir,
|
installed_plugin_dir,
|
||||||
dev_mode,
|
|
||||||
};
|
};
|
||||||
|
|
||||||
// Forward events to subscribers
|
// Forward events to subscribers
|
||||||
@@ -151,33 +155,47 @@ impl PluginManager {
|
|||||||
&kill_server_rx,
|
&kill_server_rx,
|
||||||
killed_tx,
|
killed_tx,
|
||||||
)
|
)
|
||||||
.await
|
.await?;
|
||||||
.unwrap();
|
|
||||||
info!("Waiting for plugins to initialize");
|
info!("Waiting for plugins to initialize");
|
||||||
init_plugins_task.await.unwrap();
|
init_plugins_task.await.map_err(|e| PluginErr(e.to_string()))?;
|
||||||
|
|
||||||
plugin_manager
|
let bundled_dirs = plugin_manager.list_bundled_plugin_dirs().await?;
|
||||||
}
|
let db = query_manager.connect();
|
||||||
|
for dir in bundled_dirs {
|
||||||
/// Get the vendored plugin directory path (resolves dev mode path if applicable)
|
if db.get_plugin_by_directory(&dir).is_none() {
|
||||||
pub fn get_plugins_dir(&self) -> PathBuf {
|
db.upsert_plugin(
|
||||||
if self.dev_mode {
|
&Plugin {
|
||||||
// Use plugins directly for easy development
|
directory: dir,
|
||||||
// Tauri runs from crates-tauri/yaak-app/, so go up two levels to reach project root
|
enabled: true,
|
||||||
env::current_dir()
|
url: None,
|
||||||
.map(|cwd| cwd.join("../../plugins").canonicalize().unwrap())
|
..Default::default()
|
||||||
.unwrap_or_else(|_| self.vendored_plugin_dir.clone())
|
},
|
||||||
} else {
|
&UpdateSource::Background,
|
||||||
self.vendored_plugin_dir.clone()
|
)?;
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
let plugins = db.list_plugins()?;
|
||||||
|
drop(db);
|
||||||
|
|
||||||
|
let init_errors = plugin_manager.initialize_all_plugins(plugins, plugin_context).await;
|
||||||
|
if !init_errors.is_empty() {
|
||||||
|
let joined = init_errors
|
||||||
|
.into_iter()
|
||||||
|
.map(|(dir, err)| format!("{dir}: {err}"))
|
||||||
|
.collect::<Vec<_>>()
|
||||||
|
.join("; ");
|
||||||
|
return Err(PluginErr(format!("Failed to initialize plugin(s): {joined}")));
|
||||||
|
}
|
||||||
|
|
||||||
|
Ok(plugin_manager)
|
||||||
}
|
}
|
||||||
|
|
||||||
/// Read plugin directories from disk and return their paths.
|
/// Read plugin directories from disk and return their paths.
|
||||||
/// This is useful for discovering bundled plugins.
|
/// This is useful for discovering bundled plugins.
|
||||||
pub async fn list_bundled_plugin_dirs(&self) -> Result<Vec<String>> {
|
pub async fn list_bundled_plugin_dirs(&self) -> Result<Vec<String>> {
|
||||||
let plugins_dir = self.get_plugins_dir();
|
info!("Loading bundled plugins from {:?}", self.bundled_plugin_dir);
|
||||||
info!("Loading bundled plugins from {plugins_dir:?}");
|
read_plugins_dir(&self.bundled_plugin_dir).await
|
||||||
read_plugins_dir(&plugins_dir).await
|
|
||||||
}
|
}
|
||||||
|
|
||||||
pub async fn uninstall(&self, plugin_context: &PluginContext, dir: &str) -> Result<()> {
|
pub async fn uninstall(&self, plugin_context: &PluginContext, dir: &str) -> Result<()> {
|
||||||
|
|||||||
@@ -273,6 +273,5 @@ pub fn find_client_certificate(
|
|||||||
});
|
});
|
||||||
}
|
}
|
||||||
|
|
||||||
debug!("No matching client certificate found for {}", url_string);
|
|
||||||
None
|
None
|
||||||
}
|
}
|
||||||
|
|||||||
@@ -239,6 +239,7 @@ pub struct SendHttpRequestByIdParams<'a, T: TemplateCallback> {
|
|||||||
pub cookie_jar_id: Option<String>,
|
pub cookie_jar_id: Option<String>,
|
||||||
pub response_dir: &'a Path,
|
pub response_dir: &'a Path,
|
||||||
pub emit_events_to: Option<mpsc::Sender<SenderHttpResponseEvent>>,
|
pub emit_events_to: Option<mpsc::Sender<SenderHttpResponseEvent>>,
|
||||||
|
pub emit_response_body_chunks_to: Option<mpsc::UnboundedSender<Vec<u8>>>,
|
||||||
pub cancelled_rx: Option<watch::Receiver<bool>>,
|
pub cancelled_rx: Option<watch::Receiver<bool>>,
|
||||||
pub prepare_sendable_request: Option<&'a dyn PrepareSendableRequest>,
|
pub prepare_sendable_request: Option<&'a dyn PrepareSendableRequest>,
|
||||||
pub executor: Option<&'a dyn SendRequestExecutor>,
|
pub executor: Option<&'a dyn SendRequestExecutor>,
|
||||||
@@ -255,6 +256,7 @@ pub struct SendHttpRequestParams<'a, T: TemplateCallback> {
|
|||||||
pub cookie_jar_id: Option<String>,
|
pub cookie_jar_id: Option<String>,
|
||||||
pub response_dir: &'a Path,
|
pub response_dir: &'a Path,
|
||||||
pub emit_events_to: Option<mpsc::Sender<SenderHttpResponseEvent>>,
|
pub emit_events_to: Option<mpsc::Sender<SenderHttpResponseEvent>>,
|
||||||
|
pub emit_response_body_chunks_to: Option<mpsc::UnboundedSender<Vec<u8>>>,
|
||||||
pub cancelled_rx: Option<watch::Receiver<bool>>,
|
pub cancelled_rx: Option<watch::Receiver<bool>>,
|
||||||
pub auth_context_id: Option<String>,
|
pub auth_context_id: Option<String>,
|
||||||
pub existing_response: Option<HttpResponse>,
|
pub existing_response: Option<HttpResponse>,
|
||||||
@@ -271,6 +273,7 @@ pub struct SendHttpRequestWithPluginsParams<'a> {
|
|||||||
pub cookie_jar_id: Option<String>,
|
pub cookie_jar_id: Option<String>,
|
||||||
pub response_dir: &'a Path,
|
pub response_dir: &'a Path,
|
||||||
pub emit_events_to: Option<mpsc::Sender<SenderHttpResponseEvent>>,
|
pub emit_events_to: Option<mpsc::Sender<SenderHttpResponseEvent>>,
|
||||||
|
pub emit_response_body_chunks_to: Option<mpsc::UnboundedSender<Vec<u8>>>,
|
||||||
pub existing_response: Option<HttpResponse>,
|
pub existing_response: Option<HttpResponse>,
|
||||||
pub plugin_manager: Arc<PluginManager>,
|
pub plugin_manager: Arc<PluginManager>,
|
||||||
pub encryption_manager: Arc<EncryptionManager>,
|
pub encryption_manager: Arc<EncryptionManager>,
|
||||||
@@ -288,6 +291,7 @@ pub struct SendHttpRequestByIdWithPluginsParams<'a> {
|
|||||||
pub cookie_jar_id: Option<String>,
|
pub cookie_jar_id: Option<String>,
|
||||||
pub response_dir: &'a Path,
|
pub response_dir: &'a Path,
|
||||||
pub emit_events_to: Option<mpsc::Sender<SenderHttpResponseEvent>>,
|
pub emit_events_to: Option<mpsc::Sender<SenderHttpResponseEvent>>,
|
||||||
|
pub emit_response_body_chunks_to: Option<mpsc::UnboundedSender<Vec<u8>>>,
|
||||||
pub plugin_manager: Arc<PluginManager>,
|
pub plugin_manager: Arc<PluginManager>,
|
||||||
pub encryption_manager: Arc<EncryptionManager>,
|
pub encryption_manager: Arc<EncryptionManager>,
|
||||||
pub plugin_context: &'a PluginContext,
|
pub plugin_context: &'a PluginContext,
|
||||||
@@ -353,6 +357,7 @@ pub async fn send_http_request_by_id_with_plugins(
|
|||||||
cookie_jar_id: params.cookie_jar_id,
|
cookie_jar_id: params.cookie_jar_id,
|
||||||
response_dir: params.response_dir,
|
response_dir: params.response_dir,
|
||||||
emit_events_to: params.emit_events_to,
|
emit_events_to: params.emit_events_to,
|
||||||
|
emit_response_body_chunks_to: params.emit_response_body_chunks_to,
|
||||||
existing_response: None,
|
existing_response: None,
|
||||||
plugin_manager: params.plugin_manager,
|
plugin_manager: params.plugin_manager,
|
||||||
encryption_manager: params.encryption_manager,
|
encryption_manager: params.encryption_manager,
|
||||||
@@ -397,6 +402,7 @@ pub async fn send_http_request_with_plugins(
|
|||||||
cookie_jar_id: params.cookie_jar_id,
|
cookie_jar_id: params.cookie_jar_id,
|
||||||
response_dir: params.response_dir,
|
response_dir: params.response_dir,
|
||||||
emit_events_to: params.emit_events_to,
|
emit_events_to: params.emit_events_to,
|
||||||
|
emit_response_body_chunks_to: params.emit_response_body_chunks_to,
|
||||||
cancelled_rx: params.cancelled_rx,
|
cancelled_rx: params.cancelled_rx,
|
||||||
auth_context_id: None,
|
auth_context_id: None,
|
||||||
existing_response: params.existing_response,
|
existing_response: params.existing_response,
|
||||||
@@ -427,6 +433,7 @@ pub async fn send_http_request_by_id<T: TemplateCallback>(
|
|||||||
cookie_jar_id: params.cookie_jar_id,
|
cookie_jar_id: params.cookie_jar_id,
|
||||||
response_dir: params.response_dir,
|
response_dir: params.response_dir,
|
||||||
emit_events_to: params.emit_events_to,
|
emit_events_to: params.emit_events_to,
|
||||||
|
emit_response_body_chunks_to: params.emit_response_body_chunks_to,
|
||||||
cancelled_rx: params.cancelled_rx,
|
cancelled_rx: params.cancelled_rx,
|
||||||
existing_response: None,
|
existing_response: None,
|
||||||
prepare_sendable_request: params.prepare_sendable_request,
|
prepare_sendable_request: params.prepare_sendable_request,
|
||||||
@@ -687,13 +694,17 @@ pub async fn send_http_request<T: TemplateCallback>(
|
|||||||
Ok(n) => {
|
Ok(n) => {
|
||||||
written_bytes += n;
|
written_bytes += n;
|
||||||
let start_idx = response_body.len() - n;
|
let start_idx = response_body.len() - n;
|
||||||
file.write_all(&response_body[start_idx..]).await.map_err(|source| {
|
let chunk = &response_body[start_idx..];
|
||||||
|
file.write_all(chunk).await.map_err(|source| {
|
||||||
SendHttpRequestError::WriteResponseBody { path: body_path.clone(), source }
|
SendHttpRequestError::WriteResponseBody { path: body_path.clone(), source }
|
||||||
})?;
|
})?;
|
||||||
file.flush().await.map_err(|source| SendHttpRequestError::WriteResponseBody {
|
file.flush().await.map_err(|source| SendHttpRequestError::WriteResponseBody {
|
||||||
path: body_path.clone(),
|
path: body_path.clone(),
|
||||||
source,
|
source,
|
||||||
})?;
|
})?;
|
||||||
|
if let Some(tx) = params.emit_response_body_chunks_to.as_ref() {
|
||||||
|
let _ = tx.send(chunk.to_vec());
|
||||||
|
}
|
||||||
|
|
||||||
let now = Instant::now();
|
let now = Instant::now();
|
||||||
let should_update = now.duration_since(last_progress_update).as_millis()
|
let should_update = now.duration_since(last_progress_update).as_millis()
|
||||||
|
|||||||
0
npm/cli-darwin-arm64/bin/.gitkeep
Normal file
0
npm/cli-darwin-arm64/bin/.gitkeep
Normal file
10
npm/cli-darwin-arm64/package.json
Normal file
10
npm/cli-darwin-arm64/package.json
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
{
|
||||||
|
"name": "@yaakapp/cli-darwin-arm64",
|
||||||
|
"version": "0.0.1",
|
||||||
|
"repository": {
|
||||||
|
"type": "git",
|
||||||
|
"url": "git+https://github.com/mountain-loop/yaak.git"
|
||||||
|
},
|
||||||
|
"os": ["darwin"],
|
||||||
|
"cpu": ["arm64"]
|
||||||
|
}
|
||||||
0
npm/cli-darwin-x64/bin/.gitkeep
Normal file
0
npm/cli-darwin-x64/bin/.gitkeep
Normal file
10
npm/cli-darwin-x64/package.json
Normal file
10
npm/cli-darwin-x64/package.json
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
{
|
||||||
|
"name": "@yaakapp/cli-darwin-x64",
|
||||||
|
"version": "0.0.1",
|
||||||
|
"repository": {
|
||||||
|
"type": "git",
|
||||||
|
"url": "git+https://github.com/mountain-loop/yaak.git"
|
||||||
|
},
|
||||||
|
"os": ["darwin"],
|
||||||
|
"cpu": ["x64"]
|
||||||
|
}
|
||||||
0
npm/cli-linux-arm64/bin/.gitkeep
Normal file
0
npm/cli-linux-arm64/bin/.gitkeep
Normal file
10
npm/cli-linux-arm64/package.json
Normal file
10
npm/cli-linux-arm64/package.json
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
{
|
||||||
|
"name": "@yaakapp/cli-linux-arm64",
|
||||||
|
"version": "0.0.1",
|
||||||
|
"repository": {
|
||||||
|
"type": "git",
|
||||||
|
"url": "git+https://github.com/mountain-loop/yaak.git"
|
||||||
|
},
|
||||||
|
"os": ["linux"],
|
||||||
|
"cpu": ["arm64"]
|
||||||
|
}
|
||||||
0
npm/cli-linux-x64/bin/.gitkeep
Normal file
0
npm/cli-linux-x64/bin/.gitkeep
Normal file
10
npm/cli-linux-x64/package.json
Normal file
10
npm/cli-linux-x64/package.json
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
{
|
||||||
|
"name": "@yaakapp/cli-linux-x64",
|
||||||
|
"version": "0.0.1",
|
||||||
|
"repository": {
|
||||||
|
"type": "git",
|
||||||
|
"url": "git+https://github.com/mountain-loop/yaak.git"
|
||||||
|
},
|
||||||
|
"os": ["linux"],
|
||||||
|
"cpu": ["x64"]
|
||||||
|
}
|
||||||
0
npm/cli-win32-arm64/bin/.gitkeep
Normal file
0
npm/cli-win32-arm64/bin/.gitkeep
Normal file
10
npm/cli-win32-arm64/package.json
Normal file
10
npm/cli-win32-arm64/package.json
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
{
|
||||||
|
"name": "@yaakapp/cli-win32-arm64",
|
||||||
|
"version": "0.0.1",
|
||||||
|
"repository": {
|
||||||
|
"type": "git",
|
||||||
|
"url": "git+https://github.com/mountain-loop/yaak.git"
|
||||||
|
},
|
||||||
|
"os": ["win32"],
|
||||||
|
"cpu": ["arm64"]
|
||||||
|
}
|
||||||
0
npm/cli-win32-x64/bin/.gitkeep
Normal file
0
npm/cli-win32-x64/bin/.gitkeep
Normal file
10
npm/cli-win32-x64/package.json
Normal file
10
npm/cli-win32-x64/package.json
Normal file
@@ -0,0 +1,10 @@
|
|||||||
|
{
|
||||||
|
"name": "@yaakapp/cli-win32-x64",
|
||||||
|
"version": "0.0.1",
|
||||||
|
"repository": {
|
||||||
|
"type": "git",
|
||||||
|
"url": "git+https://github.com/mountain-loop/yaak.git"
|
||||||
|
},
|
||||||
|
"os": ["win32"],
|
||||||
|
"cpu": ["x64"]
|
||||||
|
}
|
||||||
2
npm/cli/.gitignore
vendored
Normal file
2
npm/cli/.gitignore
vendored
Normal file
@@ -0,0 +1,2 @@
|
|||||||
|
yaak
|
||||||
|
yaak.exe
|
||||||
30
npm/cli/bin/cli.js
Executable file
30
npm/cli/bin/cli.js
Executable file
@@ -0,0 +1,30 @@
|
|||||||
|
#!/usr/bin/env node
|
||||||
|
|
||||||
|
const path = require("path");
|
||||||
|
const childProcess = require("child_process");
|
||||||
|
const { BINARY_NAME, PLATFORM_SPECIFIC_PACKAGE_NAME } = require("../common");
|
||||||
|
|
||||||
|
function getBinaryPath() {
|
||||||
|
try {
|
||||||
|
if (!PLATFORM_SPECIFIC_PACKAGE_NAME) {
|
||||||
|
throw new Error("unsupported platform");
|
||||||
|
}
|
||||||
|
return require.resolve(`${PLATFORM_SPECIFIC_PACKAGE_NAME}/bin/${BINARY_NAME}`);
|
||||||
|
} catch (_) {
|
||||||
|
return path.join(__dirname, "..", BINARY_NAME);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
const result = childProcess.spawnSync(getBinaryPath(), process.argv.slice(2), {
|
||||||
|
stdio: "inherit"
|
||||||
|
});
|
||||||
|
|
||||||
|
if (result.error) {
|
||||||
|
throw result.error;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (result.signal) {
|
||||||
|
process.kill(process.pid, result.signal);
|
||||||
|
}
|
||||||
|
|
||||||
|
process.exit(result.status ?? 1);
|
||||||
20
npm/cli/common.js
Normal file
20
npm/cli/common.js
Normal file
@@ -0,0 +1,20 @@
|
|||||||
|
const BINARY_DISTRIBUTION_PACKAGES = {
|
||||||
|
darwin_arm64: "@yaakapp/cli-darwin-arm64",
|
||||||
|
darwin_x64: "@yaakapp/cli-darwin-x64",
|
||||||
|
linux_arm64: "@yaakapp/cli-linux-arm64",
|
||||||
|
linux_x64: "@yaakapp/cli-linux-x64",
|
||||||
|
win32_x64: "@yaakapp/cli-win32-x64",
|
||||||
|
win32_arm64: "@yaakapp/cli-win32-arm64"
|
||||||
|
};
|
||||||
|
|
||||||
|
const BINARY_DISTRIBUTION_VERSION = require("./package.json").version;
|
||||||
|
const BINARY_NAME = process.platform === "win32" ? "yaak.exe" : "yaak";
|
||||||
|
const PLATFORM_SPECIFIC_PACKAGE_NAME =
|
||||||
|
BINARY_DISTRIBUTION_PACKAGES[`${process.platform}_${process.arch}`];
|
||||||
|
|
||||||
|
module.exports = {
|
||||||
|
BINARY_DISTRIBUTION_PACKAGES,
|
||||||
|
BINARY_DISTRIBUTION_VERSION,
|
||||||
|
BINARY_NAME,
|
||||||
|
PLATFORM_SPECIFIC_PACKAGE_NAME
|
||||||
|
};
|
||||||
20
npm/cli/index.js
Normal file
20
npm/cli/index.js
Normal file
@@ -0,0 +1,20 @@
|
|||||||
|
const path = require("path");
|
||||||
|
const childProcess = require("child_process");
|
||||||
|
const { PLATFORM_SPECIFIC_PACKAGE_NAME, BINARY_NAME } = require("./common");
|
||||||
|
|
||||||
|
function getBinaryPath() {
|
||||||
|
try {
|
||||||
|
if (!PLATFORM_SPECIFIC_PACKAGE_NAME) {
|
||||||
|
throw new Error("unsupported platform");
|
||||||
|
}
|
||||||
|
return require.resolve(`${PLATFORM_SPECIFIC_PACKAGE_NAME}/bin/${BINARY_NAME}`);
|
||||||
|
} catch (_) {
|
||||||
|
return path.join(__dirname, BINARY_NAME);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
module.exports.runBinary = function runBinary(...args) {
|
||||||
|
childProcess.execFileSync(getBinaryPath(), args, {
|
||||||
|
stdio: "inherit"
|
||||||
|
});
|
||||||
|
};
|
||||||
97
npm/cli/install.js
Normal file
97
npm/cli/install.js
Normal file
@@ -0,0 +1,97 @@
|
|||||||
|
const fs = require("node:fs");
|
||||||
|
const path = require("node:path");
|
||||||
|
const zlib = require("node:zlib");
|
||||||
|
const https = require("node:https");
|
||||||
|
const {
|
||||||
|
BINARY_DISTRIBUTION_VERSION,
|
||||||
|
BINARY_NAME,
|
||||||
|
PLATFORM_SPECIFIC_PACKAGE_NAME
|
||||||
|
} = require("./common");
|
||||||
|
|
||||||
|
const fallbackBinaryPath = path.join(__dirname, BINARY_NAME);
|
||||||
|
|
||||||
|
function makeRequest(url) {
|
||||||
|
return new Promise((resolve, reject) => {
|
||||||
|
https
|
||||||
|
.get(url, (response) => {
|
||||||
|
if (response.statusCode >= 200 && response.statusCode < 300) {
|
||||||
|
const chunks = [];
|
||||||
|
response.on("data", (chunk) => chunks.push(chunk));
|
||||||
|
response.on("end", () => resolve(Buffer.concat(chunks)));
|
||||||
|
} else if (
|
||||||
|
response.statusCode >= 300 &&
|
||||||
|
response.statusCode < 400 &&
|
||||||
|
response.headers.location
|
||||||
|
) {
|
||||||
|
makeRequest(response.headers.location).then(resolve, reject);
|
||||||
|
} else {
|
||||||
|
reject(
|
||||||
|
new Error(
|
||||||
|
`npm responded with status code ${response.statusCode} when downloading package ${url}`
|
||||||
|
)
|
||||||
|
);
|
||||||
|
}
|
||||||
|
})
|
||||||
|
.on("error", (error) => reject(error));
|
||||||
|
});
|
||||||
|
}
|
||||||
|
|
||||||
|
function extractFileFromTarball(tarballBuffer, filepath) {
|
||||||
|
let offset = 0;
|
||||||
|
while (offset < tarballBuffer.length) {
|
||||||
|
const header = tarballBuffer.subarray(offset, offset + 512);
|
||||||
|
offset += 512;
|
||||||
|
|
||||||
|
const fileName = header.toString("utf-8", 0, 100).replace(/\0.*/g, "");
|
||||||
|
const fileSize = parseInt(header.toString("utf-8", 124, 136).replace(/\0.*/g, ""), 8);
|
||||||
|
|
||||||
|
if (fileName === filepath) {
|
||||||
|
return tarballBuffer.subarray(offset, offset + fileSize);
|
||||||
|
}
|
||||||
|
|
||||||
|
offset = (offset + fileSize + 511) & ~511;
|
||||||
|
}
|
||||||
|
|
||||||
|
return null;
|
||||||
|
}
|
||||||
|
|
||||||
|
async function downloadBinaryFromNpm() {
|
||||||
|
if (!PLATFORM_SPECIFIC_PACKAGE_NAME) {
|
||||||
|
throw new Error(`Unsupported platform: ${process.platform}/${process.arch}`);
|
||||||
|
}
|
||||||
|
|
||||||
|
const packageNameWithoutScope = PLATFORM_SPECIFIC_PACKAGE_NAME.split("/")[1];
|
||||||
|
const tarballUrl = `https://registry.npmjs.org/${PLATFORM_SPECIFIC_PACKAGE_NAME}/-/${packageNameWithoutScope}-${BINARY_DISTRIBUTION_VERSION}.tgz`;
|
||||||
|
const tarballDownloadBuffer = await makeRequest(tarballUrl);
|
||||||
|
const tarballBuffer = zlib.unzipSync(tarballDownloadBuffer);
|
||||||
|
|
||||||
|
const binary = extractFileFromTarball(tarballBuffer, `package/bin/${BINARY_NAME}`);
|
||||||
|
if (!binary) {
|
||||||
|
throw new Error(`Could not find package/bin/${BINARY_NAME} in tarball`);
|
||||||
|
}
|
||||||
|
|
||||||
|
fs.writeFileSync(fallbackBinaryPath, binary);
|
||||||
|
fs.chmodSync(fallbackBinaryPath, "755");
|
||||||
|
}
|
||||||
|
|
||||||
|
function isPlatformSpecificPackageInstalled() {
|
||||||
|
try {
|
||||||
|
if (!PLATFORM_SPECIFIC_PACKAGE_NAME) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
require.resolve(`${PLATFORM_SPECIFIC_PACKAGE_NAME}/bin/${BINARY_NAME}`);
|
||||||
|
return true;
|
||||||
|
} catch (_) {
|
||||||
|
return false;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!isPlatformSpecificPackageInstalled()) {
|
||||||
|
console.log("Platform package missing. Downloading Yaak CLI binary from npm...");
|
||||||
|
downloadBinaryFromNpm().catch((err) => {
|
||||||
|
console.error("Failed to install Yaak CLI binary:", err);
|
||||||
|
process.exitCode = 1;
|
||||||
|
});
|
||||||
|
} else {
|
||||||
|
console.log("Platform package present. Using bundled Yaak CLI binary.");
|
||||||
|
}
|
||||||
25
npm/cli/package.json
Normal file
25
npm/cli/package.json
Normal file
@@ -0,0 +1,25 @@
|
|||||||
|
{
|
||||||
|
"name": "@yaakapp/cli",
|
||||||
|
"version": "0.0.1",
|
||||||
|
"main": "./index.js",
|
||||||
|
"repository": {
|
||||||
|
"type": "git",
|
||||||
|
"url": "git+https://github.com/mountain-loop/yaak.git"
|
||||||
|
},
|
||||||
|
"scripts": {
|
||||||
|
"postinstall": "node ./install.js",
|
||||||
|
"prepublishOnly": "node ./prepublish.js"
|
||||||
|
},
|
||||||
|
"bin": {
|
||||||
|
"yaak": "bin/cli.js",
|
||||||
|
"yaakcli": "bin/cli.js"
|
||||||
|
},
|
||||||
|
"optionalDependencies": {
|
||||||
|
"@yaakapp/cli-darwin-x64": "0.0.1",
|
||||||
|
"@yaakapp/cli-darwin-arm64": "0.0.1",
|
||||||
|
"@yaakapp/cli-linux-arm64": "0.0.1",
|
||||||
|
"@yaakapp/cli-linux-x64": "0.0.1",
|
||||||
|
"@yaakapp/cli-win32-x64": "0.0.1",
|
||||||
|
"@yaakapp/cli-win32-arm64": "0.0.1"
|
||||||
|
}
|
||||||
|
}
|
||||||
5
npm/cli/prepublish.js
Normal file
5
npm/cli/prepublish.js
Normal file
@@ -0,0 +1,5 @@
|
|||||||
|
const fs = require("node:fs");
|
||||||
|
const path = require("node:path");
|
||||||
|
|
||||||
|
const cliReadme = path.join(__dirname, "..", "..", "crates-cli", "yaak-cli", "README.md");
|
||||||
|
fs.copyFileSync(cliReadme, path.join(__dirname, "README.md"));
|
||||||
77
npm/prepare-publish.js
Normal file
77
npm/prepare-publish.js
Normal file
@@ -0,0 +1,77 @@
|
|||||||
|
const { chmodSync, copyFileSync, existsSync, readFileSync, writeFileSync } = require("node:fs");
|
||||||
|
const { join } = require("node:path");
|
||||||
|
|
||||||
|
const version = process.env.YAAK_CLI_VERSION?.replace(/^v/, "");
|
||||||
|
if (!version) {
|
||||||
|
console.error("YAAK_CLI_VERSION is not set");
|
||||||
|
process.exit(1);
|
||||||
|
}
|
||||||
|
|
||||||
|
const packages = [
|
||||||
|
"cli",
|
||||||
|
"cli-darwin-arm64",
|
||||||
|
"cli-darwin-x64",
|
||||||
|
"cli-linux-arm64",
|
||||||
|
"cli-linux-x64",
|
||||||
|
"cli-win32-arm64",
|
||||||
|
"cli-win32-x64"
|
||||||
|
];
|
||||||
|
|
||||||
|
const binaries = [
|
||||||
|
{
|
||||||
|
src: join(__dirname, "dist", "cli-darwin-arm64", "yaak"),
|
||||||
|
dest: join(__dirname, "cli-darwin-arm64", "bin", "yaak")
|
||||||
|
},
|
||||||
|
{
|
||||||
|
src: join(__dirname, "dist", "cli-darwin-x64", "yaak"),
|
||||||
|
dest: join(__dirname, "cli-darwin-x64", "bin", "yaak")
|
||||||
|
},
|
||||||
|
{
|
||||||
|
src: join(__dirname, "dist", "cli-linux-arm64", "yaak"),
|
||||||
|
dest: join(__dirname, "cli-linux-arm64", "bin", "yaak")
|
||||||
|
},
|
||||||
|
{
|
||||||
|
src: join(__dirname, "dist", "cli-linux-x64", "yaak"),
|
||||||
|
dest: join(__dirname, "cli-linux-x64", "bin", "yaak")
|
||||||
|
},
|
||||||
|
{
|
||||||
|
src: join(__dirname, "dist", "cli-win32-arm64", "yaak.exe"),
|
||||||
|
dest: join(__dirname, "cli-win32-arm64", "bin", "yaak.exe")
|
||||||
|
},
|
||||||
|
{
|
||||||
|
src: join(__dirname, "dist", "cli-win32-x64", "yaak.exe"),
|
||||||
|
dest: join(__dirname, "cli-win32-x64", "bin", "yaak.exe")
|
||||||
|
}
|
||||||
|
];
|
||||||
|
|
||||||
|
for (const { src, dest } of binaries) {
|
||||||
|
if (!existsSync(src)) {
|
||||||
|
console.error(`Missing binary artifact: ${src}`);
|
||||||
|
process.exit(1);
|
||||||
|
}
|
||||||
|
copyFileSync(src, dest);
|
||||||
|
if (!dest.endsWith(".exe")) {
|
||||||
|
chmodSync(dest, 0o755);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
for (const pkg of packages) {
|
||||||
|
const filepath = join(__dirname, pkg, "package.json");
|
||||||
|
const json = JSON.parse(readFileSync(filepath, "utf-8"));
|
||||||
|
json.version = version;
|
||||||
|
|
||||||
|
if (json.name === "@yaakapp/cli") {
|
||||||
|
json.optionalDependencies = {
|
||||||
|
"@yaakapp/cli-darwin-x64": version,
|
||||||
|
"@yaakapp/cli-darwin-arm64": version,
|
||||||
|
"@yaakapp/cli-linux-arm64": version,
|
||||||
|
"@yaakapp/cli-linux-x64": version,
|
||||||
|
"@yaakapp/cli-win32-x64": version,
|
||||||
|
"@yaakapp/cli-win32-arm64": version
|
||||||
|
};
|
||||||
|
}
|
||||||
|
|
||||||
|
writeFileSync(filepath, `${JSON.stringify(json, null, 2)}\n`);
|
||||||
|
}
|
||||||
|
|
||||||
|
console.log(`Prepared @yaakapp/cli npm packages for ${version}`);
|
||||||
66
package-lock.json
generated
66
package-lock.json
generated
@@ -37,7 +37,6 @@
|
|||||||
"plugins/template-function-cookie",
|
"plugins/template-function-cookie",
|
||||||
"plugins/template-function-ctx",
|
"plugins/template-function-ctx",
|
||||||
"plugins/template-function-encode",
|
"plugins/template-function-encode",
|
||||||
"plugins/template-function-faker",
|
|
||||||
"plugins/template-function-fs",
|
"plugins/template-function-fs",
|
||||||
"plugins/template-function-hash",
|
"plugins/template-function-hash",
|
||||||
"plugins/template-function-json",
|
"plugins/template-function-json",
|
||||||
@@ -74,7 +73,7 @@
|
|||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@biomejs/biome": "^2.3.13",
|
"@biomejs/biome": "^2.3.13",
|
||||||
"@tauri-apps/cli": "^2.9.6",
|
"@tauri-apps/cli": "^2.9.6",
|
||||||
"@yaakapp/cli": "^0.3.4",
|
"@yaakapp/cli": "^0.4.0",
|
||||||
"dotenv-cli": "^11.0.0",
|
"dotenv-cli": "^11.0.0",
|
||||||
"husky": "^9.1.7",
|
"husky": "^9.1.7",
|
||||||
"nodejs-file-downloader": "^4.13.0",
|
"nodejs-file-downloader": "^4.13.0",
|
||||||
@@ -4327,27 +4326,28 @@
|
|||||||
"link": true
|
"link": true
|
||||||
},
|
},
|
||||||
"node_modules/@yaakapp/cli": {
|
"node_modules/@yaakapp/cli": {
|
||||||
"version": "0.3.4",
|
"version": "0.4.0",
|
||||||
"resolved": "https://registry.npmjs.org/@yaakapp/cli/-/cli-0.3.4.tgz",
|
"resolved": "https://registry.npmjs.org/@yaakapp/cli/-/cli-0.4.0.tgz",
|
||||||
"integrity": "sha512-bSSL3noEfyoPC0M+bj34jbBZbB+gwYLCHL9cf6BYHgkRQKlHFpvN6z8M2jQZljb+CTQdHK0NzosmwHLpjMmAVA==",
|
"integrity": "sha512-8xnu2oFWlgV+xeIAHMuEgsqX6Sxq4UYrSH2WbafwDLbSep6fxpO74tiBH7xp4wakt/7Bcy9a2Q5R9nkAc1ZUdA==",
|
||||||
"dev": true,
|
"dev": true,
|
||||||
"hasInstallScript": true,
|
"hasInstallScript": true,
|
||||||
"bin": {
|
"bin": {
|
||||||
|
"yaak": "bin/cli.js",
|
||||||
"yaakcli": "bin/cli.js"
|
"yaakcli": "bin/cli.js"
|
||||||
},
|
},
|
||||||
"optionalDependencies": {
|
"optionalDependencies": {
|
||||||
"@yaakapp/cli-darwin-arm64": "0.3.4",
|
"@yaakapp/cli-darwin-arm64": "0.4.0",
|
||||||
"@yaakapp/cli-darwin-x64": "0.3.4",
|
"@yaakapp/cli-darwin-x64": "0.4.0",
|
||||||
"@yaakapp/cli-linux-arm64": "0.3.4",
|
"@yaakapp/cli-linux-arm64": "0.4.0",
|
||||||
"@yaakapp/cli-linux-x64": "0.3.4",
|
"@yaakapp/cli-linux-x64": "0.4.0",
|
||||||
"@yaakapp/cli-win32-arm64": "0.3.4",
|
"@yaakapp/cli-win32-arm64": "0.4.0",
|
||||||
"@yaakapp/cli-win32-x64": "0.3.4"
|
"@yaakapp/cli-win32-x64": "0.4.0"
|
||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/@yaakapp/cli-darwin-arm64": {
|
"node_modules/@yaakapp/cli-darwin-arm64": {
|
||||||
"version": "0.3.4",
|
"version": "0.4.0",
|
||||||
"resolved": "https://registry.npmjs.org/@yaakapp/cli-darwin-arm64/-/cli-darwin-arm64-0.3.4.tgz",
|
"resolved": "https://registry.npmjs.org/@yaakapp/cli-darwin-arm64/-/cli-darwin-arm64-0.4.0.tgz",
|
||||||
"integrity": "sha512-iTohEO7XSVZwSvTgEQE9my3wGyWtTl1q8yfol7hHwVFTX7G8Geh8X2j2vVokHhj7J9OZL9jtYQWIsM1ekOHSEQ==",
|
"integrity": "sha512-bl8+VQNPMabXNGQCa7u6w0JGe3CmzYZPsGE8Q+5wGSxa3trGf1bmq/fMW5JXrMi1P7Laepnyad0TGGP/2C8uwQ==",
|
||||||
"cpu": [
|
"cpu": [
|
||||||
"arm64"
|
"arm64"
|
||||||
],
|
],
|
||||||
@@ -4358,9 +4358,9 @@
|
|||||||
]
|
]
|
||||||
},
|
},
|
||||||
"node_modules/@yaakapp/cli-darwin-x64": {
|
"node_modules/@yaakapp/cli-darwin-x64": {
|
||||||
"version": "0.3.4",
|
"version": "0.4.0",
|
||||||
"resolved": "https://registry.npmjs.org/@yaakapp/cli-darwin-x64/-/cli-darwin-x64-0.3.4.tgz",
|
"resolved": "https://registry.npmjs.org/@yaakapp/cli-darwin-x64/-/cli-darwin-x64-0.4.0.tgz",
|
||||||
"integrity": "sha512-gz7IcjFGKA0cCAum1Aq8kmVg7erYYSrZ9pliDw0NZyObjrBysJcsDXLodEU437u0pihtdCfoLsq3rsYYs8uwCA==",
|
"integrity": "sha512-R+ETXNBWvmA3W88ZoTk/JtG/PZaUb85y3SwBgMbwcgdhBVwNS/g+DbCspcTFI5zs8Txsf5VuiFU+dW9M9olZ6A==",
|
||||||
"cpu": [
|
"cpu": [
|
||||||
"x64"
|
"x64"
|
||||||
],
|
],
|
||||||
@@ -4371,9 +4371,9 @@
|
|||||||
]
|
]
|
||||||
},
|
},
|
||||||
"node_modules/@yaakapp/cli-linux-arm64": {
|
"node_modules/@yaakapp/cli-linux-arm64": {
|
||||||
"version": "0.3.4",
|
"version": "0.4.0",
|
||||||
"resolved": "https://registry.npmjs.org/@yaakapp/cli-linux-arm64/-/cli-linux-arm64-0.3.4.tgz",
|
"resolved": "https://registry.npmjs.org/@yaakapp/cli-linux-arm64/-/cli-linux-arm64-0.4.0.tgz",
|
||||||
"integrity": "sha512-Yiwz8PBkXngmr0lTMW1pgy+F/kUISkzvqofdoBseXTrS/GDxoW3ILnG3If30LuIyWWPgqpuU+qKMtbVDzuncPQ==",
|
"integrity": "sha512-Pf7VyQf4r85FsI0qYnnst7URQF8/RxSZZj79cXLai0FnN3fDiypX4CmHx765bJxgfQZlBvqVmvPAaMW/TeiJEQ==",
|
||||||
"cpu": [
|
"cpu": [
|
||||||
"arm64"
|
"arm64"
|
||||||
],
|
],
|
||||||
@@ -4384,9 +4384,9 @@
|
|||||||
]
|
]
|
||||||
},
|
},
|
||||||
"node_modules/@yaakapp/cli-linux-x64": {
|
"node_modules/@yaakapp/cli-linux-x64": {
|
||||||
"version": "0.3.4",
|
"version": "0.4.0",
|
||||||
"resolved": "https://registry.npmjs.org/@yaakapp/cli-linux-x64/-/cli-linux-x64-0.3.4.tgz",
|
"resolved": "https://registry.npmjs.org/@yaakapp/cli-linux-x64/-/cli-linux-x64-0.4.0.tgz",
|
||||||
"integrity": "sha512-j7/r18UYNlFChDVU5N5ye3mmL+OR9Uu3LY72JxW+s/SyV69Bo8Griii75Wt19z/jj2ES8pxD+4IJq56VF3wJ7w==",
|
"integrity": "sha512-bYWWfHAIW81A+ydJChjH1Qo3+aihz9gFLh7/9MOa6CJgnC6H3V5cnapmh50Hddt9l5ic02aA1FB8ORQOXxb01A==",
|
||||||
"cpu": [
|
"cpu": [
|
||||||
"x64"
|
"x64"
|
||||||
],
|
],
|
||||||
@@ -4397,9 +4397,9 @@
|
|||||||
]
|
]
|
||||||
},
|
},
|
||||||
"node_modules/@yaakapp/cli-win32-arm64": {
|
"node_modules/@yaakapp/cli-win32-arm64": {
|
||||||
"version": "0.3.4",
|
"version": "0.4.0",
|
||||||
"resolved": "https://registry.npmjs.org/@yaakapp/cli-win32-arm64/-/cli-win32-arm64-0.3.4.tgz",
|
"resolved": "https://registry.npmjs.org/@yaakapp/cli-win32-arm64/-/cli-win32-arm64-0.4.0.tgz",
|
||||||
"integrity": "sha512-OUSKOKrSnzrTAGW0c+2ZCwA4yhgw/bA+gyeTvpf7cELVuB0qooGkEcJ3lM7fPMKmUbFU0r+K/Ggq1QMUr7cJLQ==",
|
"integrity": "sha512-8X12xkyidyYZ5vtarZGFSYR6HJbUMFUsNxYPNQccnYJIY+soNkjJHOWDjaRvBzCbR8MLT9N04Y5PE/Jv20gXpA==",
|
||||||
"cpu": [
|
"cpu": [
|
||||||
"arm64"
|
"arm64"
|
||||||
],
|
],
|
||||||
@@ -4410,9 +4410,9 @@
|
|||||||
]
|
]
|
||||||
},
|
},
|
||||||
"node_modules/@yaakapp/cli-win32-x64": {
|
"node_modules/@yaakapp/cli-win32-x64": {
|
||||||
"version": "0.3.4",
|
"version": "0.4.0",
|
||||||
"resolved": "https://registry.npmjs.org/@yaakapp/cli-win32-x64/-/cli-win32-x64-0.3.4.tgz",
|
"resolved": "https://registry.npmjs.org/@yaakapp/cli-win32-x64/-/cli-win32-x64-0.4.0.tgz",
|
||||||
"integrity": "sha512-sVYnW1rROLbzFUCyeZ++ibN+8gJS7FdPnBRHIE0KORfeI4e7Gw/aMUji2qpSZ1gt3DrAU95DDNjBkDvGBAgqag==",
|
"integrity": "sha512-wansfrCCycFcFclowQQxfsNLIAyATyqnnbITED5gUfUrBf8NFHrG0sWVCWlXUhHU7YvpmqL7CsdtlMkIGiZCPQ==",
|
||||||
"cpu": [
|
"cpu": [
|
||||||
"x64"
|
"x64"
|
||||||
],
|
],
|
||||||
@@ -7985,9 +7985,9 @@
|
|||||||
}
|
}
|
||||||
},
|
},
|
||||||
"node_modules/hono": {
|
"node_modules/hono": {
|
||||||
"version": "4.11.7",
|
"version": "4.11.10",
|
||||||
"resolved": "https://registry.npmjs.org/hono/-/hono-4.11.7.tgz",
|
"resolved": "https://registry.npmjs.org/hono/-/hono-4.11.10.tgz",
|
||||||
"integrity": "sha512-l7qMiNee7t82bH3SeyUCt9UF15EVmaBvsppY2zQtrbIhl/yzBTny+YUxsVjSjQ6gaqaeVtZmGocom8TzBlA4Yw==",
|
"integrity": "sha512-kyWP5PAiMooEvGrA9jcD3IXF7ATu8+o7B3KCbPXid5se52NPqnOpM/r9qeW2heMnOekF4kqR1fXJqCYeCLKrZg==",
|
||||||
"license": "MIT",
|
"license": "MIT",
|
||||||
"engines": {
|
"engines": {
|
||||||
"node": ">=16.9.0"
|
"node": ">=16.9.0"
|
||||||
@@ -16020,7 +16020,7 @@
|
|||||||
"@hono/mcp": "^0.2.3",
|
"@hono/mcp": "^0.2.3",
|
||||||
"@hono/node-server": "^1.19.7",
|
"@hono/node-server": "^1.19.7",
|
||||||
"@modelcontextprotocol/sdk": "^1.26.0",
|
"@modelcontextprotocol/sdk": "^1.26.0",
|
||||||
"hono": "^4.11.7",
|
"hono": "^4.11.10",
|
||||||
"zod": "^3.25.76"
|
"zod": "^3.25.76"
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
|
|||||||
@@ -70,7 +70,6 @@
|
|||||||
"app-dev": "node scripts/run-dev.mjs",
|
"app-dev": "node scripts/run-dev.mjs",
|
||||||
"migration": "node scripts/create-migration.cjs",
|
"migration": "node scripts/create-migration.cjs",
|
||||||
"build": "npm run --workspaces --if-present build",
|
"build": "npm run --workspaces --if-present build",
|
||||||
"build-plugins": "npm run --workspaces --if-present build",
|
|
||||||
"test": "npm run --workspaces --if-present test",
|
"test": "npm run --workspaces --if-present test",
|
||||||
"icons": "run-p icons:*",
|
"icons": "run-p icons:*",
|
||||||
"icons:dev": "tauri icon crates-tauri/yaak-app/icons/icon-dev.png --output crates-tauri/yaak-app/icons/dev",
|
"icons:dev": "tauri icon crates-tauri/yaak-app/icons/icon-dev.png --output crates-tauri/yaak-app/icons/dev",
|
||||||
@@ -98,7 +97,7 @@
|
|||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
"@biomejs/biome": "^2.3.13",
|
"@biomejs/biome": "^2.3.13",
|
||||||
"@tauri-apps/cli": "^2.9.6",
|
"@tauri-apps/cli": "^2.9.6",
|
||||||
"@yaakapp/cli": "^0.3.4",
|
"@yaakapp/cli": "^0.4.0",
|
||||||
"dotenv-cli": "^11.0.0",
|
"dotenv-cli": "^11.0.0",
|
||||||
"husky": "^9.1.7",
|
"husky": "^9.1.7",
|
||||||
"nodejs-file-downloader": "^4.13.0",
|
"nodejs-file-downloader": "^4.13.0",
|
||||||
|
|||||||
@@ -27,7 +27,6 @@
|
|||||||
"build:copy-types": "run-p build:copy-types:*",
|
"build:copy-types": "run-p build:copy-types:*",
|
||||||
"build:copy-types:root": "cpy --flat ../../crates/yaak-plugins/bindings/*.ts ./src/bindings",
|
"build:copy-types:root": "cpy --flat ../../crates/yaak-plugins/bindings/*.ts ./src/bindings",
|
||||||
"build:copy-types:next": "cpy --flat ../../crates/yaak-plugins/bindings/serde_json/*.ts ./src/bindings/serde_json",
|
"build:copy-types:next": "cpy --flat ../../crates/yaak-plugins/bindings/serde_json/*.ts ./src/bindings/serde_json",
|
||||||
"publish": "npm publish",
|
|
||||||
"prepublishOnly": "npm run build"
|
"prepublishOnly": "npm run build"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
|
|||||||
@@ -29,6 +29,7 @@ const modules = [
|
|||||||
|
|
||||||
function normalizeResult(result: unknown): string {
|
function normalizeResult(result: unknown): string {
|
||||||
if (typeof result === 'string') return result;
|
if (typeof result === 'string') return result;
|
||||||
|
if (result instanceof Date) return result.toISOString();
|
||||||
return JSON.stringify(result);
|
return JSON.stringify(result);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|||||||
@@ -9,4 +9,18 @@ describe('template-function-faker', () => {
|
|||||||
// accidental additions, removals, or renames across faker upgrades.
|
// accidental additions, removals, or renames across faker upgrades.
|
||||||
expect(names).toMatchSnapshot();
|
expect(names).toMatchSnapshot();
|
||||||
});
|
});
|
||||||
|
|
||||||
|
it('renders date results as unquoted ISO strings', async () => {
|
||||||
|
const { plugin } = await import('../src/index');
|
||||||
|
const fn = plugin.templateFunctions?.find((fn) => fn.name === 'faker.date.future');
|
||||||
|
|
||||||
|
expect(fn?.onRender).toBeTypeOf('function');
|
||||||
|
|
||||||
|
const result = await fn!.onRender!(
|
||||||
|
{} as Parameters<NonNullable<typeof fn.onRender>>[0],
|
||||||
|
{ values: {} } as Parameters<NonNullable<typeof fn.onRender>>[1],
|
||||||
|
);
|
||||||
|
|
||||||
|
expect(result).toMatch(/^\d{4}-\d{2}-\d{2}T\d{2}:\d{2}:\d{2}\.\d{3}Z$/);
|
||||||
|
});
|
||||||
});
|
});
|
||||||
|
|||||||
@@ -18,7 +18,7 @@
|
|||||||
"@hono/mcp": "^0.2.3",
|
"@hono/mcp": "^0.2.3",
|
||||||
"@hono/node-server": "^1.19.7",
|
"@hono/node-server": "^1.19.7",
|
||||||
"@modelcontextprotocol/sdk": "^1.26.0",
|
"@modelcontextprotocol/sdk": "^1.26.0",
|
||||||
"hono": "^4.11.7",
|
"hono": "^4.11.10",
|
||||||
"zod": "^3.25.76"
|
"zod": "^3.25.76"
|
||||||
},
|
},
|
||||||
"devDependencies": {
|
"devDependencies": {
|
||||||
|
|||||||
@@ -1,7 +1,7 @@
|
|||||||
{
|
{
|
||||||
"name": "@yaak/action-send-folder",
|
"name": "@yaak/action-send-folder",
|
||||||
"displayName": "Send All",
|
"displayName": "Send All",
|
||||||
"description": "Send all HTTP requests in a folder sequentially",
|
"description": "Send all HTTP requests in a folder sequentially in tree order",
|
||||||
"repository": {
|
"repository": {
|
||||||
"type": "git",
|
"type": "git",
|
||||||
"url": "https://github.com/mountain-loop/yaak.git",
|
"url": "https://github.com/mountain-loop/yaak.git",
|
||||||
|
|||||||
@@ -14,22 +14,44 @@ export const plugin: PluginDefinition = {
|
|||||||
ctx.httpRequest.list(),
|
ctx.httpRequest.list(),
|
||||||
]);
|
]);
|
||||||
|
|
||||||
// Build a set of all folder IDs that are descendants of the target folder
|
// Build the send order to match tree ordering:
|
||||||
const folderIds = new Set<string>([targetFolder.id]);
|
// sort siblings by sortPriority then updatedAt, and traverse folders depth-first.
|
||||||
const addDescendants = (parentId: string) => {
|
const compareByOrder = (
|
||||||
for (const folder of allFolders) {
|
a: Pick<typeof allFolders[number], 'sortPriority' | 'updatedAt'>,
|
||||||
if (folder.folderId === parentId && !folderIds.has(folder.id)) {
|
b: Pick<typeof allFolders[number], 'sortPriority' | 'updatedAt'>,
|
||||||
folderIds.add(folder.id);
|
) => {
|
||||||
addDescendants(folder.id);
|
if (a.sortPriority === b.sortPriority) {
|
||||||
|
return a.updatedAt > b.updatedAt ? 1 : -1;
|
||||||
|
}
|
||||||
|
return a.sortPriority - b.sortPriority;
|
||||||
|
};
|
||||||
|
|
||||||
|
const childrenByFolderId = new Map<string, Array<typeof allFolders[number] | typeof allRequests[number]>>();
|
||||||
|
for (const folder of allFolders) {
|
||||||
|
if (folder.folderId == null) continue;
|
||||||
|
const children = childrenByFolderId.get(folder.folderId) ?? [];
|
||||||
|
children.push(folder);
|
||||||
|
childrenByFolderId.set(folder.folderId, children);
|
||||||
|
}
|
||||||
|
for (const request of allRequests) {
|
||||||
|
if (request.folderId == null) continue;
|
||||||
|
const children = childrenByFolderId.get(request.folderId) ?? [];
|
||||||
|
children.push(request);
|
||||||
|
childrenByFolderId.set(request.folderId, children);
|
||||||
|
}
|
||||||
|
|
||||||
|
const requestsToSend: typeof allRequests = [];
|
||||||
|
const collectRequests = (folderId: string) => {
|
||||||
|
const children = (childrenByFolderId.get(folderId) ?? []).slice().sort(compareByOrder);
|
||||||
|
for (const child of children) {
|
||||||
|
if (child.model === 'folder') {
|
||||||
|
collectRequests(child.id);
|
||||||
|
} else if (child.model === 'http_request') {
|
||||||
|
requestsToSend.push(child);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
};
|
};
|
||||||
addDescendants(targetFolder.id);
|
collectRequests(targetFolder.id);
|
||||||
|
|
||||||
// Filter HTTP requests to those in the target folder or its descendants
|
|
||||||
const requestsToSend = allRequests.filter(
|
|
||||||
(req) => req.folderId != null && folderIds.has(req.folderId),
|
|
||||||
);
|
|
||||||
|
|
||||||
if (requestsToSend.length === 0) {
|
if (requestsToSend.length === 0) {
|
||||||
await ctx.toast.show({
|
await ctx.toast.show({
|
||||||
@@ -40,7 +62,7 @@ export const plugin: PluginDefinition = {
|
|||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
// Send each request sequentially
|
// Send requests sequentially in the calculated folder order.
|
||||||
let successCount = 0;
|
let successCount = 0;
|
||||||
let errorCount = 0;
|
let errorCount = 0;
|
||||||
|
|
||||||
|
|||||||
@@ -11,7 +11,8 @@
|
|||||||
"version": "0.1.0",
|
"version": "0.1.0",
|
||||||
"scripts": {
|
"scripts": {
|
||||||
"build": "yaakcli build",
|
"build": "yaakcli build",
|
||||||
"dev": "yaakcli dev"
|
"dev": "yaakcli dev",
|
||||||
|
"test": "vitest --run tests"
|
||||||
},
|
},
|
||||||
"dependencies": {
|
"dependencies": {
|
||||||
"httpntlm": "^1.8.13"
|
"httpntlm": "^1.8.13"
|
||||||
|
|||||||
@@ -2,6 +2,16 @@ import type { PluginDefinition } from '@yaakapp/api';
|
|||||||
|
|
||||||
import { ntlm } from 'httpntlm';
|
import { ntlm } from 'httpntlm';
|
||||||
|
|
||||||
|
function extractNtlmChallenge(headers: Array<{ name: string; value: string }>): string | null {
|
||||||
|
const authValues = headers
|
||||||
|
.filter((h) => h.name.toLowerCase() === 'www-authenticate')
|
||||||
|
.flatMap((h) => h.value.split(','))
|
||||||
|
.map((v) => v.trim())
|
||||||
|
.filter(Boolean);
|
||||||
|
|
||||||
|
return authValues.find((v) => /^NTLM\s+\S+/i.test(v)) ?? null;
|
||||||
|
}
|
||||||
|
|
||||||
export const plugin: PluginDefinition = {
|
export const plugin: PluginDefinition = {
|
||||||
authentication: {
|
authentication: {
|
||||||
name: 'windows',
|
name: 'windows',
|
||||||
@@ -68,15 +78,12 @@ export const plugin: PluginDefinition = {
|
|||||||
},
|
},
|
||||||
});
|
});
|
||||||
|
|
||||||
const wwwAuthenticateHeader = negotiateResponse.headers.find(
|
const ntlmChallenge = extractNtlmChallenge(negotiateResponse.headers);
|
||||||
(h) => h.name.toLowerCase() === 'www-authenticate',
|
if (ntlmChallenge == null) {
|
||||||
);
|
throw new Error('Unable to find NTLM challenge in WWW-Authenticate response headers');
|
||||||
|
|
||||||
if (!wwwAuthenticateHeader?.value) {
|
|
||||||
throw new Error('Unable to find www-authenticate response header for NTLM');
|
|
||||||
}
|
}
|
||||||
|
|
||||||
const type2 = ntlm.parseType2Message(wwwAuthenticateHeader.value, (err: Error | null) => {
|
const type2 = ntlm.parseType2Message(ntlmChallenge, (err: Error | null) => {
|
||||||
if (err != null) throw err;
|
if (err != null) throw err;
|
||||||
});
|
});
|
||||||
const type3 = ntlm.createType3Message(type2, options);
|
const type3 = ntlm.createType3Message(type2, options);
|
||||||
|
|||||||
84
plugins/auth-ntlm/tests/index.test.ts
Normal file
84
plugins/auth-ntlm/tests/index.test.ts
Normal file
@@ -0,0 +1,84 @@
|
|||||||
|
import type { Context } from '@yaakapp/api';
|
||||||
|
import { beforeEach, describe, expect, test, vi } from 'vitest';
|
||||||
|
|
||||||
|
const ntlmMock = vi.hoisted(() => ({
|
||||||
|
createType1Message: vi.fn(),
|
||||||
|
parseType2Message: vi.fn(),
|
||||||
|
createType3Message: vi.fn(),
|
||||||
|
}));
|
||||||
|
|
||||||
|
vi.mock('httpntlm', () => ({ ntlm: ntlmMock }));
|
||||||
|
|
||||||
|
import { plugin } from '../src';
|
||||||
|
|
||||||
|
describe('auth-ntlm', () => {
|
||||||
|
beforeEach(() => {
|
||||||
|
ntlmMock.createType1Message.mockReset();
|
||||||
|
ntlmMock.parseType2Message.mockReset();
|
||||||
|
ntlmMock.createType3Message.mockReset();
|
||||||
|
ntlmMock.createType1Message.mockReturnValue('NTLM TYPE1');
|
||||||
|
ntlmMock.parseType2Message.mockReturnValue({} as any);
|
||||||
|
ntlmMock.createType3Message.mockReturnValue('NTLM TYPE3');
|
||||||
|
});
|
||||||
|
|
||||||
|
test('uses NTLM challenge when Negotiate and NTLM headers are separate', async () => {
|
||||||
|
const send = vi.fn().mockResolvedValue({
|
||||||
|
headers: [
|
||||||
|
{ name: 'WWW-Authenticate', value: 'Negotiate' },
|
||||||
|
{ name: 'WWW-Authenticate', value: 'NTLM TlRMTVNTUAACAAAAAA==' },
|
||||||
|
],
|
||||||
|
});
|
||||||
|
const ctx = { httpRequest: { send } } as unknown as Context;
|
||||||
|
|
||||||
|
const result = await plugin.authentication?.onApply(ctx, {
|
||||||
|
values: {},
|
||||||
|
headers: [],
|
||||||
|
url: 'https://example.local/resource',
|
||||||
|
method: 'GET',
|
||||||
|
contextId: 'ctx',
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(ntlmMock.parseType2Message).toHaveBeenCalledWith(
|
||||||
|
'NTLM TlRMTVNTUAACAAAAAA==',
|
||||||
|
expect.any(Function),
|
||||||
|
);
|
||||||
|
expect(result).toEqual({ setHeaders: [{ name: 'Authorization', value: 'NTLM TYPE3' }] });
|
||||||
|
});
|
||||||
|
|
||||||
|
test('uses NTLM challenge when auth schemes are comma-separated in one header', async () => {
|
||||||
|
const send = vi.fn().mockResolvedValue({
|
||||||
|
headers: [{ name: 'www-authenticate', value: 'Negotiate, NTLM TlRMTVNTUAACAAAAAA==' }],
|
||||||
|
});
|
||||||
|
const ctx = { httpRequest: { send } } as unknown as Context;
|
||||||
|
|
||||||
|
await plugin.authentication?.onApply(ctx, {
|
||||||
|
values: {},
|
||||||
|
headers: [],
|
||||||
|
url: 'https://example.local/resource',
|
||||||
|
method: 'GET',
|
||||||
|
contextId: 'ctx',
|
||||||
|
});
|
||||||
|
|
||||||
|
expect(ntlmMock.parseType2Message).toHaveBeenCalledWith(
|
||||||
|
'NTLM TlRMTVNTUAACAAAAAA==',
|
||||||
|
expect.any(Function),
|
||||||
|
);
|
||||||
|
});
|
||||||
|
|
||||||
|
test('throws a clear error when NTLM challenge is missing', async () => {
|
||||||
|
const send = vi.fn().mockResolvedValue({
|
||||||
|
headers: [{ name: 'WWW-Authenticate', value: 'Negotiate' }],
|
||||||
|
});
|
||||||
|
const ctx = { httpRequest: { send } } as unknown as Context;
|
||||||
|
|
||||||
|
await expect(
|
||||||
|
plugin.authentication?.onApply(ctx, {
|
||||||
|
values: {},
|
||||||
|
headers: [],
|
||||||
|
url: 'https://example.local/resource',
|
||||||
|
method: 'GET',
|
||||||
|
contextId: 'ctx',
|
||||||
|
}),
|
||||||
|
).rejects.toThrow('Unable to find NTLM challenge in WWW-Authenticate response headers');
|
||||||
|
});
|
||||||
|
});
|
||||||
@@ -72,6 +72,10 @@ export const plugin: PluginDefinition = {
|
|||||||
name: 'header',
|
name: 'header',
|
||||||
label: 'Header Name',
|
label: 'Header Name',
|
||||||
async dynamic(ctx, args) {
|
async dynamic(ctx, args) {
|
||||||
|
// Dynamic form config also runs during send-time rendering.
|
||||||
|
// Keep this preview-only to avoid side-effect request sends.
|
||||||
|
if (args.purpose !== 'preview') return null;
|
||||||
|
|
||||||
const response = await getResponse(ctx, {
|
const response = await getResponse(ctx, {
|
||||||
requestId: String(args.values.request || ''),
|
requestId: String(args.values.request || ''),
|
||||||
purpose: args.purpose,
|
purpose: args.purpose,
|
||||||
@@ -146,6 +150,10 @@ export const plugin: PluginDefinition = {
|
|||||||
label: 'JSONPath or XPath',
|
label: 'JSONPath or XPath',
|
||||||
placeholder: '$.books[0].id or /books[0]/id',
|
placeholder: '$.books[0].id or /books[0]/id',
|
||||||
dynamic: async (ctx, args) => {
|
dynamic: async (ctx, args) => {
|
||||||
|
// Dynamic form config also runs during send-time rendering.
|
||||||
|
// Keep this preview-only to avoid side-effect request sends.
|
||||||
|
if (args.purpose !== 'preview') return null;
|
||||||
|
|
||||||
const resp = await getResponse(ctx, {
|
const resp = await getResponse(ctx, {
|
||||||
requestId: String(args.values.request || ''),
|
requestId: String(args.values.request || ''),
|
||||||
purpose: 'preview',
|
purpose: 'preview',
|
||||||
|
|||||||
@@ -1,4 +1,4 @@
|
|||||||
const { readdirSync, cpSync, existsSync } = require('node:fs');
|
const { readdirSync, cpSync, existsSync, mkdirSync } = require('node:fs');
|
||||||
const path = require('node:path');
|
const path = require('node:path');
|
||||||
|
|
||||||
const pluginsDir = path.join(__dirname, '..', 'plugins');
|
const pluginsDir = path.join(__dirname, '..', 'plugins');
|
||||||
@@ -24,6 +24,7 @@ for (const name of readdirSync(pluginsDir)) {
|
|||||||
continue;
|
continue;
|
||||||
}
|
}
|
||||||
const destDir = path.join(__dirname, '../crates-tauri/yaak-app/vendored/plugins/', name);
|
const destDir = path.join(__dirname, '../crates-tauri/yaak-app/vendored/plugins/', name);
|
||||||
|
mkdirSync(destDir, { recursive: true });
|
||||||
console.log(`Copying ${name} to ${destDir}`);
|
console.log(`Copying ${name} to ${destDir}`);
|
||||||
cpSync(path.join(dir, 'package.json'), path.join(destDir, 'package.json'));
|
cpSync(path.join(dir, 'package.json'), path.join(destDir, 'package.json'));
|
||||||
cpSync(path.join(dir, 'build'), path.join(destDir, 'build'), { recursive: true });
|
cpSync(path.join(dir, 'build'), path.join(destDir, 'build'), { recursive: true });
|
||||||
|
|||||||
Reference in New Issue
Block a user