Compare commits
32 Commits
sidecar-v1
...
v1.4.18
| Author | SHA1 | Date | |
|---|---|---|---|
|
|
d9d90563cc | ||
|
|
5a674ed199 | ||
|
|
9d78fce3f0 | ||
|
|
a8de39de84 | ||
|
|
bc82584dff | ||
|
|
4d0b4ee1c5 | ||
|
|
c73e9de0ac | ||
|
|
288c6ad6a3 | ||
|
|
af8046f9b1 | ||
|
|
6003885519 | ||
|
|
8829846b53 | ||
|
|
cf449d9338 | ||
|
|
5a6910834c | ||
|
|
a6c7eb5d5e | ||
|
|
135d5d534b | ||
|
|
76f34fe17d | ||
|
|
68ad31b6a7 | ||
|
|
fcbe405e23 | ||
|
|
4adfd2adc6 | ||
|
|
f3843d59f1 | ||
|
|
ad68251e04 | ||
|
|
9468d01a88 | ||
|
|
a3151ad55e | ||
|
|
5bff40e9b4 | ||
|
|
0ccb02ba27 | ||
|
|
aa4033b412 | ||
|
|
b4b9435317 | ||
|
|
ee1d4f8643 | ||
|
|
4a186d1de6 | ||
|
|
fff37992b1 | ||
|
|
8afe3230d3 | ||
|
|
04e7fb1a99 |
103
.gitea/workflows/build-app-linux.yml
Normal file
103
.gitea/workflows/build-app-linux.yml
Normal file
@@ -0,0 +1,103 @@
|
||||
name: Build App (Linux)
|
||||
|
||||
on:
|
||||
workflow_dispatch:
|
||||
inputs:
|
||||
tag:
|
||||
description: 'Release tag to build (e.g. v1.4.5)'
|
||||
required: true
|
||||
|
||||
jobs:
|
||||
build-linux:
|
||||
name: Build App (Linux)
|
||||
runs-on: ubuntu-latest
|
||||
env:
|
||||
NODE_VERSION: "20"
|
||||
steps:
|
||||
- name: Determine tag
|
||||
id: tag
|
||||
run: |
|
||||
TAG="${{ inputs.tag }}"
|
||||
if [ -z "$TAG" ]; then
|
||||
TAG="${{ github.event.inputs.tag }}"
|
||||
fi
|
||||
if [ -z "$TAG" ]; then
|
||||
TAG=$(git ls-remote --tags --sort=-v:refname origin 'refs/tags/v*' | head -1 | sed 's|.*refs/tags/||')
|
||||
fi
|
||||
echo "Building for tag: ${TAG}"
|
||||
echo "tag=${TAG}" >> $GITHUB_OUTPUT
|
||||
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
ref: ${{ steps.tag.outputs.tag }}
|
||||
|
||||
- name: Set up Node.js
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: ${{ env.NODE_VERSION }}
|
||||
|
||||
- name: Install Rust stable
|
||||
run: |
|
||||
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh -s -- -y --default-toolchain stable
|
||||
echo "$HOME/.cargo/bin" >> $GITHUB_PATH
|
||||
|
||||
- name: Install system dependencies
|
||||
run: |
|
||||
sudo apt-get update
|
||||
sudo apt-get install -y libgtk-3-dev libwebkit2gtk-4.1-dev libappindicator3-dev librsvg2-dev patchelf xdg-utils rpm
|
||||
|
||||
- name: Install npm dependencies
|
||||
run: npm ci
|
||||
|
||||
- name: Build Tauri app
|
||||
run: npm run tauri build
|
||||
|
||||
- name: Upload to release
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
sudo apt-get install -y jq
|
||||
REPO_API="${GITHUB_SERVER_URL}/api/v1/repos/${GITHUB_REPOSITORY}"
|
||||
TAG="${{ steps.tag.outputs.tag }}"
|
||||
echo "Release tag: ${TAG}"
|
||||
|
||||
echo "Waiting for release ${TAG} to be available..."
|
||||
RELEASE_ID=""
|
||||
for i in $(seq 1 30); do
|
||||
RELEASE_JSON=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/tags/${TAG}")
|
||||
RELEASE_ID=$(echo "$RELEASE_JSON" | jq -r '.id // empty')
|
||||
|
||||
if [ -n "${RELEASE_ID}" ] && [ "${RELEASE_ID}" != "null" ]; then
|
||||
echo "Found release: ${TAG} (ID: ${RELEASE_ID})"
|
||||
break
|
||||
fi
|
||||
|
||||
echo "Attempt ${i}/30: Release not ready yet, retrying in 10s..."
|
||||
sleep 10
|
||||
done
|
||||
|
||||
if [ -z "${RELEASE_ID}" ] || [ "${RELEASE_ID}" = "null" ]; then
|
||||
echo "ERROR: Failed to find release for tag ${TAG} after 30 attempts."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
find src-tauri/target/release/bundle -type f \( -name "*.deb" -o -name "*.rpm" -o -name "*.AppImage" \) | while IFS= read -r file; do
|
||||
filename=$(basename "$file")
|
||||
encoded_name=$(echo "$filename" | sed 's/ /%20/g')
|
||||
echo "Uploading ${filename} ($(du -h "$file" | cut -f1))..."
|
||||
|
||||
ASSET_ID=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets" | jq -r ".[] | select(.name == \"${filename}\") | .id // empty")
|
||||
if [ -n "${ASSET_ID}" ]; then
|
||||
curl -s -X DELETE -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets/${ASSET_ID}"
|
||||
fi
|
||||
|
||||
HTTP_CODE=$(curl -s -o /dev/null -w "%{http_code}" -X POST \
|
||||
-H "Authorization: token ${BUILD_TOKEN}" \
|
||||
-H "Content-Type: application/octet-stream" \
|
||||
-T "$file" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets?name=${encoded_name}")
|
||||
echo "Upload response: HTTP ${HTTP_CODE}"
|
||||
done
|
||||
101
.gitea/workflows/build-app-macos.yml
Normal file
101
.gitea/workflows/build-app-macos.yml
Normal file
@@ -0,0 +1,101 @@
|
||||
name: Build App (macOS)
|
||||
|
||||
on:
|
||||
workflow_dispatch:
|
||||
inputs:
|
||||
tag:
|
||||
description: 'Release tag to build (e.g. v1.4.5)'
|
||||
required: true
|
||||
|
||||
jobs:
|
||||
build-macos:
|
||||
name: Build App (macOS)
|
||||
runs-on: macos-latest
|
||||
env:
|
||||
NODE_VERSION: "20"
|
||||
steps:
|
||||
- name: Determine tag
|
||||
id: tag
|
||||
run: |
|
||||
TAG="${{ inputs.tag }}"
|
||||
if [ -z "$TAG" ]; then
|
||||
TAG="${{ github.event.inputs.tag }}"
|
||||
fi
|
||||
if [ -z "$TAG" ]; then
|
||||
TAG=$(git ls-remote --tags --sort=-v:refname origin 'refs/tags/v*' | head -1 | sed 's|.*refs/tags/||')
|
||||
fi
|
||||
echo "Building for tag: ${TAG}"
|
||||
echo "tag=${TAG}" >> $GITHUB_OUTPUT
|
||||
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
ref: ${{ steps.tag.outputs.tag }}
|
||||
|
||||
- name: Set up Node.js
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: ${{ env.NODE_VERSION }}
|
||||
|
||||
- name: Install Rust stable
|
||||
run: |
|
||||
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh -s -- -y --default-toolchain stable
|
||||
echo "$HOME/.cargo/bin" >> $GITHUB_PATH
|
||||
|
||||
- name: Install system dependencies
|
||||
run: brew install --quiet create-dmg || true
|
||||
|
||||
- name: Install npm dependencies
|
||||
run: npm ci
|
||||
|
||||
- name: Build Tauri app
|
||||
run: npm run tauri build
|
||||
|
||||
- name: Upload to release
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
which jq || brew install jq
|
||||
REPO_API="${GITHUB_SERVER_URL}/api/v1/repos/${GITHUB_REPOSITORY}"
|
||||
TAG="${{ steps.tag.outputs.tag }}"
|
||||
echo "Release tag: ${TAG}"
|
||||
|
||||
echo "Waiting for release ${TAG} to be available..."
|
||||
RELEASE_ID=""
|
||||
for i in $(seq 1 30); do
|
||||
RELEASE_JSON=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/tags/${TAG}")
|
||||
RELEASE_ID=$(echo "$RELEASE_JSON" | jq -r '.id // empty')
|
||||
|
||||
if [ -n "${RELEASE_ID}" ] && [ "${RELEASE_ID}" != "null" ]; then
|
||||
echo "Found release: ${TAG} (ID: ${RELEASE_ID})"
|
||||
break
|
||||
fi
|
||||
|
||||
echo "Attempt ${i}/30: Release not ready yet, retrying in 10s..."
|
||||
sleep 10
|
||||
done
|
||||
|
||||
if [ -z "${RELEASE_ID}" ] || [ "${RELEASE_ID}" = "null" ]; then
|
||||
echo "ERROR: Failed to find release for tag ${TAG} after 30 attempts."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
find src-tauri/target/release/bundle -type f -name "*.dmg" | while IFS= read -r file; do
|
||||
filename=$(basename "$file")
|
||||
encoded_name=$(echo "$filename" | sed 's/ /%20/g')
|
||||
echo "Uploading ${filename} ($(du -h "$file" | cut -f1))..."
|
||||
|
||||
ASSET_ID=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets" | jq -r ".[] | select(.name == \"${filename}\") | .id // empty")
|
||||
if [ -n "${ASSET_ID}" ]; then
|
||||
curl -s -X DELETE -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets/${ASSET_ID}"
|
||||
fi
|
||||
|
||||
HTTP_CODE=$(curl -s -o /dev/null -w "%{http_code}" -X POST \
|
||||
-H "Authorization: token ${BUILD_TOKEN}" \
|
||||
-H "Content-Type: application/octet-stream" \
|
||||
-T "$file" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets?name=${encoded_name}")
|
||||
echo "Upload response: HTTP ${HTTP_CODE}"
|
||||
done
|
||||
117
.gitea/workflows/build-app-windows.yml
Normal file
117
.gitea/workflows/build-app-windows.yml
Normal file
@@ -0,0 +1,117 @@
|
||||
name: Build App (Windows)
|
||||
|
||||
on:
|
||||
workflow_dispatch:
|
||||
inputs:
|
||||
tag:
|
||||
description: 'Release tag to build (e.g. v1.4.5)'
|
||||
required: true
|
||||
|
||||
env:
|
||||
NODE_VERSION: "20"
|
||||
|
||||
jobs:
|
||||
build-windows:
|
||||
name: Build App (Windows)
|
||||
runs-on: windows-latest
|
||||
env:
|
||||
RELEASE_TAG: ${{ inputs.tag }}
|
||||
steps:
|
||||
- name: Show tag
|
||||
shell: powershell
|
||||
run: |
|
||||
Write-Host "Building for tag: $env:RELEASE_TAG"
|
||||
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
ref: ${{ inputs.tag }}
|
||||
|
||||
- name: Set up Node.js
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: ${{ env.NODE_VERSION }}
|
||||
|
||||
- name: Install Rust stable
|
||||
shell: powershell
|
||||
run: |
|
||||
if (Get-Command rustup -ErrorAction SilentlyContinue) {
|
||||
rustup default stable
|
||||
} else {
|
||||
Invoke-WebRequest -Uri https://win.rustup.rs/x86_64 -OutFile rustup-init.exe
|
||||
.\rustup-init.exe -y --default-toolchain stable
|
||||
echo "$env:USERPROFILE\.cargo\bin" | Out-File -FilePath $env:GITHUB_PATH -Encoding utf8 -Append
|
||||
}
|
||||
|
||||
- name: Install npm dependencies
|
||||
shell: powershell
|
||||
run: npm ci
|
||||
|
||||
- name: Build Tauri app
|
||||
shell: powershell
|
||||
run: npm run tauri build
|
||||
|
||||
- name: Upload to release
|
||||
shell: powershell
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
$REPO_API = "${{ github.server_url }}/api/v1/repos/${{ github.repository }}"
|
||||
$Headers = @{ "Authorization" = "token $env:BUILD_TOKEN" }
|
||||
$TAG = $env:RELEASE_TAG
|
||||
Write-Host "Release tag: $TAG"
|
||||
|
||||
if (-not $TAG) {
|
||||
Write-Host "ERROR: RELEASE_TAG is empty"
|
||||
exit 1
|
||||
}
|
||||
|
||||
Write-Host "Waiting for release $TAG to be available..."
|
||||
$RELEASE_ID = $null
|
||||
|
||||
for ($i = 1; $i -le 30; $i++) {
|
||||
try {
|
||||
$release = Invoke-RestMethod -Uri "$REPO_API/releases/tags/$TAG" -Headers $Headers -ErrorAction Stop
|
||||
$RELEASE_ID = $release.id
|
||||
|
||||
if ($RELEASE_ID) {
|
||||
Write-Host "Found release: $TAG (ID: $RELEASE_ID)"
|
||||
break
|
||||
}
|
||||
} catch {}
|
||||
|
||||
Write-Host "Attempt ${i}/30: Release not ready yet, retrying in 10s..."
|
||||
Start-Sleep -Seconds 10
|
||||
}
|
||||
|
||||
if (-not $RELEASE_ID) {
|
||||
Write-Host "ERROR: Failed to find release for tag $TAG after 30 attempts."
|
||||
exit 1
|
||||
}
|
||||
|
||||
Get-ChildItem -Path src-tauri\target\release\bundle -Recurse -Include *.msi,*-setup.exe | ForEach-Object {
|
||||
$filename = $_.Name
|
||||
$encodedName = [System.Uri]::EscapeDataString($filename)
|
||||
$size = [math]::Round($_.Length / 1MB, 1)
|
||||
Write-Host "Uploading $filename ($size MB)..."
|
||||
|
||||
try {
|
||||
$assets = Invoke-RestMethod -Uri "$REPO_API/releases/$RELEASE_ID/assets" -Headers $Headers
|
||||
$existing = $assets | Where-Object { $_.name -eq $filename }
|
||||
if ($existing) {
|
||||
Invoke-RestMethod -Uri "$REPO_API/releases/$RELEASE_ID/assets/$($existing.id)" -Method Delete -Headers $Headers
|
||||
}
|
||||
} catch {}
|
||||
|
||||
$uploadUrl = "$REPO_API/releases/$RELEASE_ID/assets?name=$encodedName"
|
||||
$result = curl.exe --fail --silent --show-error `
|
||||
-X POST `
|
||||
-H "Authorization: token $env:BUILD_TOKEN" `
|
||||
-H "Content-Type: application/octet-stream" `
|
||||
-T "$($_.FullName)" `
|
||||
"$uploadUrl" 2>&1
|
||||
if ($LASTEXITCODE -eq 0) {
|
||||
Write-Host "Upload successful: $filename"
|
||||
} else {
|
||||
Write-Host "WARNING: Upload failed for ${filename}: $result"
|
||||
}
|
||||
}
|
||||
118
.gitea/workflows/build-sidecar-linux.yml
Normal file
118
.gitea/workflows/build-sidecar-linux.yml
Normal file
@@ -0,0 +1,118 @@
|
||||
name: Build Sidecar (Linux)
|
||||
|
||||
on:
|
||||
workflow_dispatch:
|
||||
inputs:
|
||||
tag:
|
||||
description: 'Sidecar release tag to build (e.g. sidecar-v1.0.3)'
|
||||
required: true
|
||||
|
||||
jobs:
|
||||
build-sidecar-linux:
|
||||
name: Build Sidecar (Linux)
|
||||
runs-on: ubuntu-latest
|
||||
env:
|
||||
PYTHON_VERSION: "3.11"
|
||||
steps:
|
||||
- name: Determine tag
|
||||
id: tag
|
||||
run: |
|
||||
TAG="${{ inputs.tag }}"
|
||||
if [ -z "$TAG" ]; then
|
||||
TAG="${{ github.event.inputs.tag }}"
|
||||
fi
|
||||
if [ -z "$TAG" ]; then
|
||||
TAG=$(git ls-remote --tags --sort=-v:refname origin 'refs/tags/sidecar-v*' | head -1 | sed 's|.*refs/tags/||')
|
||||
fi
|
||||
echo "Building for tag: ${TAG}"
|
||||
echo "tag=${TAG}" >> $GITHUB_OUTPUT
|
||||
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
ref: ${{ steps.tag.outputs.tag }}
|
||||
|
||||
- name: Install uv
|
||||
run: |
|
||||
if command -v uv &> /dev/null; then
|
||||
echo "uv already installed: $(uv --version)"
|
||||
else
|
||||
curl -LsSf https://astral.sh/uv/install.sh | sh
|
||||
echo "$HOME/.local/bin" >> $GITHUB_PATH
|
||||
fi
|
||||
|
||||
- name: Set up Python
|
||||
run: uv python install ${{ env.PYTHON_VERSION }}
|
||||
|
||||
- name: Install system dependencies
|
||||
run: |
|
||||
sudo apt-get update
|
||||
sudo apt-get install -y portaudio19-dev
|
||||
|
||||
- name: Build sidecar (CUDA)
|
||||
run: |
|
||||
uv sync --frozen || uv sync
|
||||
uv run pyinstaller local-transcription-headless.spec
|
||||
|
||||
- name: Package sidecar (CUDA)
|
||||
run: |
|
||||
cd dist/local-transcription-backend && zip -r ../../sidecar-linux-x86_64-cuda.zip .
|
||||
|
||||
- name: Build sidecar (CPU)
|
||||
run: |
|
||||
rm -rf dist/local-transcription-backend build/
|
||||
uv pip install torch torchaudio --index-url https://download.pytorch.org/whl/cpu --force-reinstall
|
||||
# Run pyinstaller directly from venv to prevent uv run from
|
||||
# re-resolving torch back to the CUDA version via pyproject.toml sources
|
||||
.venv/bin/pyinstaller local-transcription-headless.spec
|
||||
|
||||
- name: Package sidecar (CPU)
|
||||
run: |
|
||||
cd dist/local-transcription-backend && zip -r ../../sidecar-linux-x86_64-cpu.zip .
|
||||
|
||||
- name: Upload to sidecar release
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
sudo apt-get install -y jq
|
||||
REPO_API="${GITHUB_SERVER_URL}/api/v1/repos/${GITHUB_REPOSITORY}"
|
||||
TAG="${{ steps.tag.outputs.tag }}"
|
||||
|
||||
echo "Waiting for sidecar release ${TAG} to be available..."
|
||||
for i in $(seq 1 30); do
|
||||
RELEASE_JSON=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/tags/${TAG}")
|
||||
RELEASE_ID=$(echo "$RELEASE_JSON" | jq -r '.id // empty')
|
||||
|
||||
if [ -n "${RELEASE_ID}" ] && [ "${RELEASE_ID}" != "null" ]; then
|
||||
echo "Found sidecar release: ${TAG} (ID: ${RELEASE_ID})"
|
||||
break
|
||||
fi
|
||||
|
||||
echo "Attempt ${i}/30: Release not ready yet, retrying in 10s..."
|
||||
sleep 10
|
||||
done
|
||||
|
||||
if [ -z "${RELEASE_ID}" ] || [ "${RELEASE_ID}" = "null" ]; then
|
||||
echo "ERROR: Failed to find sidecar release for tag ${TAG} after 30 attempts."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
for file in sidecar-*.zip; do
|
||||
filename=$(basename "$file")
|
||||
encoded_name=$(echo "$filename" | sed 's/ /%20/g')
|
||||
echo "Uploading ${filename} ($(du -h "$file" | cut -f1))..."
|
||||
|
||||
ASSET_ID=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets" | jq -r ".[] | select(.name == \"${filename}\") | .id // empty")
|
||||
if [ -n "${ASSET_ID}" ]; then
|
||||
curl -s -X DELETE -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets/${ASSET_ID}"
|
||||
fi
|
||||
|
||||
HTTP_CODE=$(curl -s -o /dev/null -w "%{http_code}" -X POST \
|
||||
-H "Authorization: token ${BUILD_TOKEN}" \
|
||||
-H "Content-Type: application/octet-stream" \
|
||||
-T "$file" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets?name=${encoded_name}")
|
||||
echo "Upload response: HTTP ${HTTP_CODE}"
|
||||
done
|
||||
109
.gitea/workflows/build-sidecar-macos.yml
Normal file
109
.gitea/workflows/build-sidecar-macos.yml
Normal file
@@ -0,0 +1,109 @@
|
||||
name: Build Sidecar (macOS)
|
||||
|
||||
on:
|
||||
workflow_dispatch:
|
||||
inputs:
|
||||
tag:
|
||||
description: 'Sidecar release tag to build (e.g. sidecar-v1.0.3)'
|
||||
required: true
|
||||
|
||||
jobs:
|
||||
build-sidecar-macos:
|
||||
name: Build Sidecar (macOS)
|
||||
runs-on: macos-latest
|
||||
env:
|
||||
PYTHON_VERSION: "3.11"
|
||||
steps:
|
||||
- name: Determine tag
|
||||
id: tag
|
||||
run: |
|
||||
TAG="${{ inputs.tag }}"
|
||||
if [ -z "$TAG" ]; then
|
||||
TAG="${{ github.event.inputs.tag }}"
|
||||
fi
|
||||
if [ -z "$TAG" ]; then
|
||||
TAG=$(git ls-remote --tags --sort=-v:refname origin 'refs/tags/sidecar-v*' | head -1 | sed 's|.*refs/tags/||')
|
||||
fi
|
||||
echo "Building for tag: ${TAG}"
|
||||
echo "tag=${TAG}" >> $GITHUB_OUTPUT
|
||||
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
ref: ${{ steps.tag.outputs.tag }}
|
||||
|
||||
- name: Install uv
|
||||
run: |
|
||||
if command -v uv &> /dev/null; then
|
||||
echo "uv already installed: $(uv --version)"
|
||||
else
|
||||
curl -LsSf https://astral.sh/uv/install.sh | sh
|
||||
echo "$HOME/.local/bin" >> $GITHUB_PATH
|
||||
fi
|
||||
|
||||
- name: Set up Python
|
||||
run: uv python install ${{ env.PYTHON_VERSION }}
|
||||
|
||||
- name: Install system dependencies
|
||||
run: brew install portaudio
|
||||
|
||||
- name: Build sidecar (CPU)
|
||||
env:
|
||||
UV_NO_SOURCES: "1"
|
||||
run: |
|
||||
# UV_NO_SOURCES bypasses pyproject.toml's [tool.uv.sources] which forces
|
||||
# torch from the CUDA index (no macOS ARM wheels there).
|
||||
# Default PyPI torch includes MPS (Apple Silicon GPU) support.
|
||||
uv sync
|
||||
.venv/bin/pyinstaller local-transcription-headless.spec
|
||||
|
||||
- name: Package sidecar (CPU)
|
||||
run: |
|
||||
cd dist/local-transcription-backend && zip -r ../../sidecar-macos-aarch64-cpu.zip .
|
||||
|
||||
- name: Upload to sidecar release
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
which jq || brew install jq
|
||||
REPO_API="${GITHUB_SERVER_URL}/api/v1/repos/${GITHUB_REPOSITORY}"
|
||||
TAG="${{ steps.tag.outputs.tag }}"
|
||||
|
||||
echo "Waiting for sidecar release ${TAG} to be available..."
|
||||
for i in $(seq 1 30); do
|
||||
RELEASE_JSON=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/tags/${TAG}")
|
||||
RELEASE_ID=$(echo "$RELEASE_JSON" | jq -r '.id // empty')
|
||||
|
||||
if [ -n "${RELEASE_ID}" ] && [ "${RELEASE_ID}" != "null" ]; then
|
||||
echo "Found sidecar release: ${TAG} (ID: ${RELEASE_ID})"
|
||||
break
|
||||
fi
|
||||
|
||||
echo "Attempt ${i}/30: Release not ready yet, retrying in 10s..."
|
||||
sleep 10
|
||||
done
|
||||
|
||||
if [ -z "${RELEASE_ID}" ] || [ "${RELEASE_ID}" = "null" ]; then
|
||||
echo "ERROR: Failed to find sidecar release for tag ${TAG} after 30 attempts."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
for file in sidecar-*.zip; do
|
||||
filename=$(basename "$file")
|
||||
encoded_name=$(echo "$filename" | sed 's/ /%20/g')
|
||||
echo "Uploading ${filename} ($(du -h "$file" | cut -f1))..."
|
||||
|
||||
ASSET_ID=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets" | jq -r ".[] | select(.name == \"${filename}\") | .id // empty")
|
||||
if [ -n "${ASSET_ID}" ]; then
|
||||
curl -s -X DELETE -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets/${ASSET_ID}"
|
||||
fi
|
||||
|
||||
HTTP_CODE=$(curl -s -o /dev/null -w "%{http_code}" -X POST \
|
||||
-H "Authorization: token ${BUILD_TOKEN}" \
|
||||
-H "Content-Type: application/octet-stream" \
|
||||
-T "$file" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets?name=${encoded_name}")
|
||||
echo "Upload response: HTTP ${HTTP_CODE}"
|
||||
done
|
||||
145
.gitea/workflows/build-sidecar-windows.yml
Normal file
145
.gitea/workflows/build-sidecar-windows.yml
Normal file
@@ -0,0 +1,145 @@
|
||||
name: Build Sidecar (Windows)
|
||||
|
||||
on:
|
||||
workflow_dispatch:
|
||||
inputs:
|
||||
tag:
|
||||
description: 'Sidecar release tag to build (e.g. sidecar-v1.0.3)'
|
||||
required: true
|
||||
|
||||
jobs:
|
||||
build-sidecar-windows:
|
||||
name: Build Sidecar (Windows)
|
||||
runs-on: windows-latest
|
||||
env:
|
||||
PYTHON_VERSION: "3.11"
|
||||
RELEASE_TAG: ${{ inputs.tag }}
|
||||
steps:
|
||||
- name: Show tag
|
||||
shell: powershell
|
||||
run: |
|
||||
Write-Host "Building for tag: $env:RELEASE_TAG"
|
||||
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
ref: ${{ inputs.tag }}
|
||||
|
||||
- name: Install uv
|
||||
shell: powershell
|
||||
run: |
|
||||
if (Get-Command uv -ErrorAction SilentlyContinue) {
|
||||
Write-Host "uv already installed: $(uv --version)"
|
||||
} else {
|
||||
irm https://astral.sh/uv/install.ps1 | iex
|
||||
$uvPaths = @(
|
||||
"$env:USERPROFILE\.local\bin",
|
||||
"$env:USERPROFILE\.cargo\bin",
|
||||
"$env:LOCALAPPDATA\uv\bin"
|
||||
)
|
||||
foreach ($p in $uvPaths) {
|
||||
if (Test-Path $p) {
|
||||
echo $p | Out-File -FilePath $env:GITHUB_PATH -Encoding utf8 -Append
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
- name: Set up Python
|
||||
shell: powershell
|
||||
run: uv python install ${{ env.PYTHON_VERSION }}
|
||||
|
||||
- name: Install 7-Zip
|
||||
shell: powershell
|
||||
run: |
|
||||
if (-not (Get-Command 7z -ErrorAction SilentlyContinue)) {
|
||||
choco install 7zip -y
|
||||
}
|
||||
|
||||
- name: Build sidecar (CUDA)
|
||||
shell: powershell
|
||||
run: |
|
||||
uv sync --frozen
|
||||
if ($LASTEXITCODE -ne 0) { uv sync }
|
||||
uv run pyinstaller local-transcription-headless.spec
|
||||
|
||||
- name: Package sidecar (CUDA)
|
||||
shell: powershell
|
||||
run: |
|
||||
7z a -tzip -mx=5 sidecar-windows-x86_64-cuda.zip .\dist\local-transcription-backend\*
|
||||
|
||||
- name: Build sidecar (CPU)
|
||||
shell: powershell
|
||||
run: |
|
||||
Remove-Item -Recurse -Force dist\local-transcription-backend, build -ErrorAction SilentlyContinue
|
||||
uv pip install torch torchaudio --index-url https://download.pytorch.org/whl/cpu --force-reinstall
|
||||
.venv\Scripts\pyinstaller.exe local-transcription-headless.spec
|
||||
|
||||
- name: Package sidecar (CPU)
|
||||
shell: powershell
|
||||
run: |
|
||||
7z a -tzip -mx=5 sidecar-windows-x86_64-cpu.zip .\dist\local-transcription-backend\*
|
||||
|
||||
- name: Upload to sidecar release
|
||||
shell: powershell
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
$REPO_API = "${{ github.server_url }}/api/v1/repos/${{ github.repository }}"
|
||||
$Headers = @{ "Authorization" = "token $env:BUILD_TOKEN" }
|
||||
$TAG = $env:RELEASE_TAG
|
||||
Write-Host "Release tag: $TAG"
|
||||
|
||||
if (-not $TAG) {
|
||||
Write-Host "ERROR: RELEASE_TAG is empty"
|
||||
exit 1
|
||||
}
|
||||
|
||||
Write-Host "Waiting for sidecar release $TAG to be available..."
|
||||
$RELEASE_ID = $null
|
||||
|
||||
for ($i = 1; $i -le 30; $i++) {
|
||||
try {
|
||||
$release = Invoke-RestMethod -Uri "$REPO_API/releases/tags/$TAG" -Headers $Headers -ErrorAction Stop
|
||||
$RELEASE_ID = $release.id
|
||||
|
||||
if ($RELEASE_ID) {
|
||||
Write-Host "Found sidecar release: $TAG (ID: $RELEASE_ID)"
|
||||
break
|
||||
}
|
||||
} catch {}
|
||||
|
||||
Write-Host "Attempt ${i}/30: Release not ready yet, retrying in 10s..."
|
||||
Start-Sleep -Seconds 10
|
||||
}
|
||||
|
||||
if (-not $RELEASE_ID) {
|
||||
Write-Host "ERROR: Failed to find sidecar release for tag $TAG after 30 attempts."
|
||||
exit 1
|
||||
}
|
||||
|
||||
Get-ChildItem -Path . -Filter "sidecar-*.zip" | ForEach-Object {
|
||||
$filename = $_.Name
|
||||
$encodedName = [System.Uri]::EscapeDataString($filename)
|
||||
$size = [math]::Round($_.Length / 1MB, 1)
|
||||
Write-Host "Uploading $filename ($size MB)..."
|
||||
|
||||
try {
|
||||
$assets = Invoke-RestMethod -Uri "$REPO_API/releases/$RELEASE_ID/assets" -Headers $Headers
|
||||
$existing = $assets | Where-Object { $_.name -eq $filename }
|
||||
if ($existing) {
|
||||
Invoke-RestMethod -Uri "$REPO_API/releases/$RELEASE_ID/assets/$($existing.id)" -Method Delete -Headers $Headers
|
||||
}
|
||||
} catch {}
|
||||
|
||||
$uploadUrl = "$REPO_API/releases/$RELEASE_ID/assets?name=$encodedName"
|
||||
$result = curl.exe --fail --silent --show-error `
|
||||
-X POST `
|
||||
-H "Authorization: token $env:BUILD_TOKEN" `
|
||||
-H "Content-Type: application/octet-stream" `
|
||||
-T "$($_.FullName)" `
|
||||
"$uploadUrl" 2>&1
|
||||
if ($LASTEXITCODE -eq 0) {
|
||||
Write-Host "Upload successful: $filename"
|
||||
} else {
|
||||
Write-Host "WARNING: Upload failed for ${filename}: $result"
|
||||
}
|
||||
}
|
||||
@@ -1,425 +0,0 @@
|
||||
name: Build Sidecars
|
||||
|
||||
on:
|
||||
push:
|
||||
branches: [main]
|
||||
paths:
|
||||
- 'client/**'
|
||||
- 'server/**'
|
||||
- 'backend/**'
|
||||
- 'pyproject.toml'
|
||||
- 'local-transcription-headless.spec'
|
||||
workflow_dispatch:
|
||||
|
||||
jobs:
|
||||
bump-sidecar-version:
|
||||
name: Bump sidecar version and tag
|
||||
if: "!contains(github.event.head_commit.message, '[skip ci]')"
|
||||
runs-on: ubuntu-latest
|
||||
outputs:
|
||||
version: ${{ steps.bump.outputs.version }}
|
||||
tag: ${{ steps.bump.outputs.tag }}
|
||||
has_changes: ${{ steps.check_changes.outputs.has_changes }}
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
fetch-depth: 2
|
||||
|
||||
- name: Check for backend changes
|
||||
id: check_changes
|
||||
run: |
|
||||
# If triggered by workflow_dispatch, always build
|
||||
if [ "${{ github.event_name }}" = "workflow_dispatch" ]; then
|
||||
echo "has_changes=true" >> $GITHUB_OUTPUT
|
||||
exit 0
|
||||
fi
|
||||
# Check if relevant files changed in this commit
|
||||
CHANGED=$(git diff --name-only HEAD~1 HEAD -- client/ server/ backend/ pyproject.toml local-transcription-headless.spec 2>/dev/null || echo "")
|
||||
if [ -n "$CHANGED" ]; then
|
||||
echo "has_changes=true" >> $GITHUB_OUTPUT
|
||||
echo "Backend changes detected: $CHANGED"
|
||||
else
|
||||
echo "has_changes=false" >> $GITHUB_OUTPUT
|
||||
echo "No backend changes detected, skipping sidecar build"
|
||||
fi
|
||||
|
||||
- name: Configure git
|
||||
if: steps.check_changes.outputs.has_changes == 'true'
|
||||
run: |
|
||||
git config user.name "Gitea Actions"
|
||||
git config user.email "actions@gitea.local"
|
||||
|
||||
- name: Bump sidecar patch version
|
||||
if: steps.check_changes.outputs.has_changes == 'true'
|
||||
id: bump
|
||||
run: |
|
||||
# Read current version from pyproject.toml
|
||||
CURRENT=$(grep '^version = ' pyproject.toml | head -1 | sed 's/version = "\(.*\)"/\1/')
|
||||
echo "Current sidecar version: ${CURRENT}"
|
||||
|
||||
# Increment patch number
|
||||
MAJOR=$(echo "${CURRENT}" | cut -d. -f1)
|
||||
MINOR=$(echo "${CURRENT}" | cut -d. -f2)
|
||||
PATCH=$(echo "${CURRENT}" | cut -d. -f3)
|
||||
NEW_PATCH=$((PATCH + 1))
|
||||
NEW_VERSION="${MAJOR}.${MINOR}.${NEW_PATCH}"
|
||||
echo "New sidecar version: ${NEW_VERSION}"
|
||||
|
||||
# Update pyproject.toml
|
||||
sed -i "s/^version = \"${CURRENT}\"/version = \"${NEW_VERSION}\"/" pyproject.toml
|
||||
|
||||
# Update version.py
|
||||
sed -i "s/__version__ = \"${CURRENT}\"/__version__ = \"${NEW_VERSION}\"/" version.py
|
||||
sed -i "s/__version_info__ = .*/__version_info__ = (${MAJOR}, ${MINOR}, ${NEW_PATCH})/" version.py
|
||||
|
||||
echo "version=${NEW_VERSION}" >> $GITHUB_OUTPUT
|
||||
echo "tag=sidecar-v${NEW_VERSION}" >> $GITHUB_OUTPUT
|
||||
|
||||
- name: Commit and tag
|
||||
if: steps.check_changes.outputs.has_changes == 'true'
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
NEW_VERSION="${{ steps.bump.outputs.version }}"
|
||||
TAG="${{ steps.bump.outputs.tag }}"
|
||||
git add pyproject.toml version.py
|
||||
git commit -m "chore: bump sidecar version to ${NEW_VERSION} [skip ci]"
|
||||
git tag "${TAG}"
|
||||
|
||||
REMOTE_URL=$(git remote get-url origin | sed "s|://|://gitea-actions:${BUILD_TOKEN}@|")
|
||||
git pull --rebase "${REMOTE_URL}" main || true
|
||||
git push "${REMOTE_URL}" HEAD:main
|
||||
git push "${REMOTE_URL}" "${TAG}"
|
||||
|
||||
- name: Create Gitea release
|
||||
if: steps.check_changes.outputs.has_changes == 'true'
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
REPO_API="${GITHUB_SERVER_URL}/api/v1/repos/${GITHUB_REPOSITORY}"
|
||||
TAG="${{ steps.bump.outputs.tag }}"
|
||||
VERSION="${{ steps.bump.outputs.version }}"
|
||||
RELEASE_NAME="Sidecar v${VERSION}"
|
||||
|
||||
curl -s -X POST \
|
||||
-H "Authorization: token ${BUILD_TOKEN}" \
|
||||
-H "Content-Type: application/json" \
|
||||
-d "{\"tag_name\": \"${TAG}\", \"name\": \"${RELEASE_NAME}\", \"body\": \"Automated sidecar build.\", \"draft\": false, \"prerelease\": false}" \
|
||||
"${REPO_API}/releases"
|
||||
echo "Created release: ${RELEASE_NAME}"
|
||||
|
||||
# ── Linux sidecar (CUDA + CPU) ──
|
||||
|
||||
build-sidecar-linux:
|
||||
name: Build Sidecar (Linux)
|
||||
needs: bump-sidecar-version
|
||||
if: needs.bump-sidecar-version.outputs.has_changes == 'true'
|
||||
runs-on: ubuntu-latest
|
||||
env:
|
||||
PYTHON_VERSION: "3.11"
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
ref: ${{ needs.bump-sidecar-version.outputs.tag }}
|
||||
|
||||
- name: Install uv
|
||||
run: |
|
||||
if command -v uv &> /dev/null; then
|
||||
echo "uv already installed: $(uv --version)"
|
||||
else
|
||||
curl -LsSf https://astral.sh/uv/install.sh | sh
|
||||
echo "$HOME/.local/bin" >> $GITHUB_PATH
|
||||
fi
|
||||
|
||||
- name: Set up Python
|
||||
run: uv python install ${{ env.PYTHON_VERSION }}
|
||||
|
||||
- name: Install system dependencies
|
||||
run: |
|
||||
sudo apt-get update
|
||||
sudo apt-get install -y portaudio19-dev
|
||||
|
||||
- name: Build sidecar (CUDA)
|
||||
run: |
|
||||
uv sync --frozen || uv sync
|
||||
uv run pyinstaller local-transcription-headless.spec
|
||||
|
||||
- name: Package sidecar (CUDA)
|
||||
run: |
|
||||
cd dist/local-transcription-backend && zip -r ../../sidecar-linux-x86_64-cuda.zip .
|
||||
|
||||
- name: Build sidecar (CPU)
|
||||
run: |
|
||||
rm -rf dist/local-transcription-backend build/
|
||||
uv pip install torch torchaudio --index-url https://download.pytorch.org/whl/cpu --force-reinstall
|
||||
uv run pyinstaller local-transcription-headless.spec
|
||||
|
||||
- name: Package sidecar (CPU)
|
||||
run: |
|
||||
cd dist/local-transcription-backend && zip -r ../../sidecar-linux-x86_64-cpu.zip .
|
||||
|
||||
- name: Upload to sidecar release
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
sudo apt-get install -y jq
|
||||
REPO_API="${GITHUB_SERVER_URL}/api/v1/repos/${GITHUB_REPOSITORY}"
|
||||
TAG="${{ needs.bump-sidecar-version.outputs.tag }}"
|
||||
|
||||
echo "Waiting for sidecar release ${TAG} to be available..."
|
||||
for i in $(seq 1 30); do
|
||||
RELEASE_JSON=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/tags/${TAG}")
|
||||
RELEASE_ID=$(echo "$RELEASE_JSON" | jq -r '.id // empty')
|
||||
|
||||
if [ -n "${RELEASE_ID}" ] && [ "${RELEASE_ID}" != "null" ]; then
|
||||
echo "Found sidecar release: ${TAG} (ID: ${RELEASE_ID})"
|
||||
break
|
||||
fi
|
||||
|
||||
echo "Attempt ${i}/30: Release not ready yet, retrying in 10s..."
|
||||
sleep 10
|
||||
done
|
||||
|
||||
if [ -z "${RELEASE_ID}" ] || [ "${RELEASE_ID}" = "null" ]; then
|
||||
echo "ERROR: Failed to find sidecar release for tag ${TAG} after 30 attempts."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
for file in sidecar-*.zip; do
|
||||
filename=$(basename "$file")
|
||||
encoded_name=$(echo "$filename" | sed 's/ /%20/g')
|
||||
echo "Uploading ${filename} ($(du -h "$file" | cut -f1))..."
|
||||
|
||||
ASSET_ID=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets" | jq -r ".[] | select(.name == \"${filename}\") | .id // empty")
|
||||
if [ -n "${ASSET_ID}" ]; then
|
||||
curl -s -X DELETE -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets/${ASSET_ID}"
|
||||
fi
|
||||
|
||||
HTTP_CODE=$(curl -s -o /dev/null -w "%{http_code}" -X POST \
|
||||
-H "Authorization: token ${BUILD_TOKEN}" \
|
||||
-H "Content-Type: application/octet-stream" \
|
||||
-T "$file" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets?name=${encoded_name}")
|
||||
echo "Upload response: HTTP ${HTTP_CODE}"
|
||||
done
|
||||
|
||||
# ── Windows sidecar (CUDA + CPU) ──
|
||||
|
||||
build-sidecar-windows:
|
||||
name: Build Sidecar (Windows)
|
||||
needs: bump-sidecar-version
|
||||
if: needs.bump-sidecar-version.outputs.has_changes == 'true'
|
||||
runs-on: windows-latest
|
||||
env:
|
||||
PYTHON_VERSION: "3.11"
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
ref: ${{ needs.bump-sidecar-version.outputs.tag }}
|
||||
|
||||
- name: Install uv
|
||||
shell: powershell
|
||||
run: |
|
||||
if (Get-Command uv -ErrorAction SilentlyContinue) {
|
||||
Write-Host "uv already installed: $(uv --version)"
|
||||
} else {
|
||||
irm https://astral.sh/uv/install.ps1 | iex
|
||||
# Add both possible uv install locations to PATH
|
||||
$uvPaths = @(
|
||||
"$env:USERPROFILE\.local\bin",
|
||||
"$env:USERPROFILE\.cargo\bin",
|
||||
"$env:LOCALAPPDATA\uv\bin"
|
||||
)
|
||||
foreach ($p in $uvPaths) {
|
||||
if (Test-Path $p) {
|
||||
echo $p | Out-File -FilePath $env:GITHUB_PATH -Encoding utf8 -Append
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
- name: Set up Python
|
||||
shell: powershell
|
||||
run: uv python install ${{ env.PYTHON_VERSION }}
|
||||
|
||||
- name: Install 7-Zip
|
||||
shell: powershell
|
||||
run: |
|
||||
if (-not (Get-Command 7z -ErrorAction SilentlyContinue)) {
|
||||
choco install 7zip -y
|
||||
}
|
||||
|
||||
- name: Build sidecar (CUDA)
|
||||
shell: powershell
|
||||
run: |
|
||||
uv sync --frozen
|
||||
if ($LASTEXITCODE -ne 0) { uv sync }
|
||||
uv run pyinstaller local-transcription-headless.spec
|
||||
|
||||
- name: Package sidecar (CUDA)
|
||||
shell: powershell
|
||||
run: |
|
||||
7z a -tzip -mx=5 sidecar-windows-x86_64-cuda.zip .\dist\local-transcription-backend\*
|
||||
|
||||
- name: Build sidecar (CPU)
|
||||
shell: powershell
|
||||
run: |
|
||||
Remove-Item -Recurse -Force dist\local-transcription-backend, build -ErrorAction SilentlyContinue
|
||||
uv pip install torch torchaudio --index-url https://download.pytorch.org/whl/cpu --force-reinstall
|
||||
uv run pyinstaller local-transcription-headless.spec
|
||||
|
||||
- name: Package sidecar (CPU)
|
||||
shell: powershell
|
||||
run: |
|
||||
7z a -tzip -mx=5 sidecar-windows-x86_64-cpu.zip .\dist\local-transcription-backend\*
|
||||
|
||||
- name: Upload to sidecar release
|
||||
shell: powershell
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
$REPO_API = "${{ github.server_url }}/api/v1/repos/${{ github.repository }}"
|
||||
$Headers = @{ "Authorization" = "token $env:BUILD_TOKEN" }
|
||||
$TAG = "${{ needs.bump-sidecar-version.outputs.tag }}"
|
||||
|
||||
Write-Host "Waiting for sidecar release ${TAG} to be available..."
|
||||
$RELEASE_ID = $null
|
||||
|
||||
for ($i = 1; $i -le 30; $i++) {
|
||||
try {
|
||||
$release = Invoke-RestMethod -Uri "${REPO_API}/releases/tags/${TAG}" -Headers $Headers -ErrorAction Stop
|
||||
$RELEASE_ID = $release.id
|
||||
|
||||
if ($RELEASE_ID) {
|
||||
Write-Host "Found sidecar release: ${TAG} (ID: ${RELEASE_ID})"
|
||||
break
|
||||
}
|
||||
} catch {}
|
||||
|
||||
Write-Host "Attempt ${i}/30: Release not ready yet, retrying in 10s..."
|
||||
Start-Sleep -Seconds 10
|
||||
}
|
||||
|
||||
if (-not $RELEASE_ID) {
|
||||
Write-Host "ERROR: Failed to find sidecar release for tag ${TAG} after 30 attempts."
|
||||
exit 1
|
||||
}
|
||||
|
||||
Get-ChildItem -Path . -Filter "sidecar-*.zip" | ForEach-Object {
|
||||
$filename = $_.Name
|
||||
$encodedName = [System.Uri]::EscapeDataString($filename)
|
||||
$size = [math]::Round($_.Length / 1MB, 1)
|
||||
Write-Host "Uploading ${filename} (${size} MB)..."
|
||||
|
||||
try {
|
||||
$assets = Invoke-RestMethod -Uri "${REPO_API}/releases/${RELEASE_ID}/assets" -Headers $Headers
|
||||
$existing = $assets | Where-Object { $_.name -eq $filename }
|
||||
if ($existing) {
|
||||
Invoke-RestMethod -Uri "${REPO_API}/releases/${RELEASE_ID}/assets/$($existing.id)" -Method Delete -Headers $Headers
|
||||
}
|
||||
} catch {}
|
||||
|
||||
$uploadUrl = "${REPO_API}/releases/${RELEASE_ID}/assets?name=${encodedName}"
|
||||
$result = curl.exe --fail --silent --show-error `
|
||||
-X POST `
|
||||
-H "Authorization: token $env:BUILD_TOKEN" `
|
||||
-H "Content-Type: application/octet-stream" `
|
||||
-T "$($_.FullName)" `
|
||||
"$uploadUrl" 2>&1
|
||||
if ($LASTEXITCODE -eq 0) {
|
||||
Write-Host "Upload successful: ${filename}"
|
||||
} else {
|
||||
Write-Host "WARNING: Upload failed for ${filename}: ${result}"
|
||||
}
|
||||
}
|
||||
|
||||
# ── macOS sidecar (CPU only — no CUDA on macOS) ──
|
||||
|
||||
build-sidecar-macos:
|
||||
name: Build Sidecar (macOS)
|
||||
needs: bump-sidecar-version
|
||||
if: needs.bump-sidecar-version.outputs.has_changes == 'true'
|
||||
runs-on: macos-latest
|
||||
env:
|
||||
PYTHON_VERSION: "3.11"
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
ref: ${{ needs.bump-sidecar-version.outputs.tag }}
|
||||
|
||||
- name: Install uv
|
||||
run: |
|
||||
if command -v uv &> /dev/null; then
|
||||
echo "uv already installed: $(uv --version)"
|
||||
else
|
||||
curl -LsSf https://astral.sh/uv/install.sh | sh
|
||||
echo "$HOME/.local/bin" >> $GITHUB_PATH
|
||||
fi
|
||||
|
||||
- name: Set up Python
|
||||
run: uv python install ${{ env.PYTHON_VERSION }}
|
||||
|
||||
- name: Install system dependencies
|
||||
run: brew install portaudio
|
||||
|
||||
- name: Build sidecar (CPU)
|
||||
run: |
|
||||
# --no-sources bypasses pyproject.toml's [tool.uv.sources] which forces
|
||||
# torch from the CUDA index (no macOS ARM wheels there)
|
||||
# Default PyPI torch includes MPS (Apple Silicon GPU) support
|
||||
uv sync --no-sources
|
||||
uv run pyinstaller local-transcription-headless.spec
|
||||
|
||||
- name: Package sidecar (CPU)
|
||||
run: |
|
||||
cd dist/local-transcription-backend && zip -r ../../sidecar-macos-aarch64-cpu.zip .
|
||||
|
||||
- name: Upload to sidecar release
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
which jq || brew install jq
|
||||
REPO_API="${GITHUB_SERVER_URL}/api/v1/repos/${GITHUB_REPOSITORY}"
|
||||
TAG="${{ needs.bump-sidecar-version.outputs.tag }}"
|
||||
|
||||
echo "Waiting for sidecar release ${TAG} to be available..."
|
||||
for i in $(seq 1 30); do
|
||||
RELEASE_JSON=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/tags/${TAG}")
|
||||
RELEASE_ID=$(echo "$RELEASE_JSON" | jq -r '.id // empty')
|
||||
|
||||
if [ -n "${RELEASE_ID}" ] && [ "${RELEASE_ID}" != "null" ]; then
|
||||
echo "Found sidecar release: ${TAG} (ID: ${RELEASE_ID})"
|
||||
break
|
||||
fi
|
||||
|
||||
echo "Attempt ${i}/30: Release not ready yet, retrying in 10s..."
|
||||
sleep 10
|
||||
done
|
||||
|
||||
if [ -z "${RELEASE_ID}" ] || [ "${RELEASE_ID}" = "null" ]; then
|
||||
echo "ERROR: Failed to find sidecar release for tag ${TAG} after 30 attempts."
|
||||
exit 1
|
||||
fi
|
||||
|
||||
for file in sidecar-*.zip; do
|
||||
filename=$(basename "$file")
|
||||
encoded_name=$(echo "$filename" | sed 's/ /%20/g')
|
||||
echo "Uploading ${filename} ($(du -h "$file" | cut -f1))..."
|
||||
|
||||
ASSET_ID=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets" | jq -r ".[] | select(.name == \"${filename}\") | .id // empty")
|
||||
if [ -n "${ASSET_ID}" ]; then
|
||||
curl -s -X DELETE -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets/${ASSET_ID}"
|
||||
fi
|
||||
|
||||
HTTP_CODE=$(curl -s -o /dev/null -w "%{http_code}" -X POST \
|
||||
-H "Authorization: token ${BUILD_TOKEN}" \
|
||||
-H "Content-Type: application/octet-stream" \
|
||||
-T "$file" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets?name=${encoded_name}")
|
||||
echo "Upload response: HTTP ${HTTP_CODE}"
|
||||
done
|
||||
@@ -25,11 +25,9 @@ jobs:
|
||||
- name: Bump patch version
|
||||
id: bump
|
||||
run: |
|
||||
# Read current version from package.json
|
||||
CURRENT=$(grep '"version"' package.json | head -1 | sed 's/.*"version": *"\([^"]*\)".*/\1/')
|
||||
echo "Current version: ${CURRENT}"
|
||||
|
||||
# Increment patch number
|
||||
MAJOR=$(echo "${CURRENT}" | cut -d. -f1)
|
||||
MINOR=$(echo "${CURRENT}" | cut -d. -f2)
|
||||
PATCH=$(echo "${CURRENT}" | cut -d. -f3)
|
||||
@@ -37,16 +35,9 @@ jobs:
|
||||
NEW_VERSION="${MAJOR}.${MINOR}.${NEW_PATCH}"
|
||||
echo "New version: ${NEW_VERSION}"
|
||||
|
||||
# Update package.json
|
||||
sed -i "s/\"version\": \"${CURRENT}\"/\"version\": \"${NEW_VERSION}\"/" package.json
|
||||
|
||||
# Update src-tauri/tauri.conf.json
|
||||
sed -i "s/\"version\": \"${CURRENT}\"/\"version\": \"${NEW_VERSION}\"/" src-tauri/tauri.conf.json
|
||||
|
||||
# Update src-tauri/Cargo.toml
|
||||
sed -i "s/^version = \"${CURRENT}\"/version = \"${NEW_VERSION}\"/" src-tauri/Cargo.toml
|
||||
|
||||
# Update version.py
|
||||
sed -i "s/__version__ = \"${CURRENT}\"/__version__ = \"${NEW_VERSION}\"/" version.py
|
||||
sed -i "s/__version_info__ = .*/__version_info__ = (${MAJOR}, ${MINOR}, ${NEW_PATCH})/" version.py
|
||||
|
||||
@@ -82,219 +73,19 @@ jobs:
|
||||
"${REPO_API}/releases"
|
||||
echo "Created release: ${RELEASE_NAME}"
|
||||
|
||||
# ── Platform builds (run after version bump) ──
|
||||
|
||||
build-linux:
|
||||
name: Build App (Linux)
|
||||
needs: bump-version
|
||||
runs-on: ubuntu-latest
|
||||
env:
|
||||
NODE_VERSION: "20"
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
ref: ${{ needs.bump-version.outputs.tag }}
|
||||
|
||||
- name: Set up Node.js
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: ${{ env.NODE_VERSION }}
|
||||
|
||||
- name: Install Rust stable
|
||||
run: |
|
||||
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh -s -- -y --default-toolchain stable
|
||||
echo "$HOME/.cargo/bin" >> $GITHUB_PATH
|
||||
|
||||
- name: Install system dependencies
|
||||
run: |
|
||||
sudo apt-get update
|
||||
sudo apt-get install -y libgtk-3-dev libwebkit2gtk-4.1-dev libappindicator3-dev librsvg2-dev patchelf xdg-utils rpm
|
||||
|
||||
- name: Install npm dependencies
|
||||
run: npm ci
|
||||
|
||||
- name: Build Tauri app
|
||||
run: npm run tauri build
|
||||
|
||||
- name: Upload to release
|
||||
- name: Trigger per-OS app builds
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
sudo apt-get install -y jq
|
||||
REPO_API="${GITHUB_SERVER_URL}/api/v1/repos/${GITHUB_REPOSITORY}"
|
||||
TAG="${{ needs.bump-version.outputs.tag }}"
|
||||
echo "Release tag: ${TAG}"
|
||||
|
||||
RELEASE_ID=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/tags/${TAG}" | jq -r '.id // empty')
|
||||
|
||||
if [ -z "${RELEASE_ID}" ] || [ "${RELEASE_ID}" = "null" ]; then
|
||||
echo "ERROR: Failed to find release for tag ${TAG}."
|
||||
exit 1
|
||||
fi
|
||||
echo "Release ID: ${RELEASE_ID}"
|
||||
|
||||
find src-tauri/target/release/bundle -type f \( -name "*.deb" -o -name "*.rpm" -o -name "*.AppImage" \) | while IFS= read -r file; do
|
||||
filename=$(basename "$file")
|
||||
encoded_name=$(echo "$filename" | sed 's/ /%20/g')
|
||||
echo "Uploading ${filename} ($(du -h "$file" | cut -f1))..."
|
||||
|
||||
ASSET_ID=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets" | jq -r ".[] | select(.name == \"${filename}\") | .id // empty")
|
||||
if [ -n "${ASSET_ID}" ]; then
|
||||
curl -s -X DELETE -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets/${ASSET_ID}"
|
||||
fi
|
||||
TAG="${{ steps.bump.outputs.tag }}"
|
||||
|
||||
for workflow in build-app-linux.yml build-app-windows.yml build-app-macos.yml; do
|
||||
echo "Dispatching ${workflow} for ${TAG}..."
|
||||
HTTP_CODE=$(curl -s -o /dev/null -w "%{http_code}" -X POST \
|
||||
-H "Authorization: token ${BUILD_TOKEN}" \
|
||||
-H "Content-Type: application/octet-stream" \
|
||||
-T "$file" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets?name=${encoded_name}")
|
||||
echo "Upload response: HTTP ${HTTP_CODE}"
|
||||
done
|
||||
|
||||
build-windows:
|
||||
name: Build App (Windows)
|
||||
needs: bump-version
|
||||
runs-on: windows-latest
|
||||
env:
|
||||
NODE_VERSION: "20"
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
ref: ${{ needs.bump-version.outputs.tag }}
|
||||
|
||||
- name: Set up Node.js
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: ${{ env.NODE_VERSION }}
|
||||
|
||||
- name: Install Rust stable
|
||||
shell: powershell
|
||||
run: |
|
||||
if (Get-Command rustup -ErrorAction SilentlyContinue) {
|
||||
rustup default stable
|
||||
} else {
|
||||
Invoke-WebRequest -Uri https://win.rustup.rs/x86_64 -OutFile rustup-init.exe
|
||||
.\rustup-init.exe -y --default-toolchain stable
|
||||
echo "$env:USERPROFILE\.cargo\bin" | Out-File -FilePath $env:GITHUB_PATH -Encoding utf8 -Append
|
||||
}
|
||||
|
||||
- name: Install npm dependencies
|
||||
shell: powershell
|
||||
run: npm ci
|
||||
|
||||
- name: Build Tauri app
|
||||
shell: powershell
|
||||
run: npm run tauri build
|
||||
|
||||
- name: Upload to release
|
||||
shell: powershell
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
$REPO_API = "${{ github.server_url }}/api/v1/repos/${{ github.repository }}"
|
||||
$Headers = @{ "Authorization" = "token $env:BUILD_TOKEN" }
|
||||
$TAG = "${{ needs.bump-version.outputs.tag }}"
|
||||
Write-Host "Release tag: ${TAG}"
|
||||
|
||||
$release = Invoke-RestMethod -Uri "${REPO_API}/releases/tags/${TAG}" -Headers $Headers -ErrorAction Stop
|
||||
$RELEASE_ID = $release.id
|
||||
Write-Host "Release ID: ${RELEASE_ID}"
|
||||
|
||||
Get-ChildItem -Path src-tauri\target\release\bundle -Recurse -Include *.msi,*-setup.exe | ForEach-Object {
|
||||
$filename = $_.Name
|
||||
$encodedName = [System.Uri]::EscapeDataString($filename)
|
||||
$size = [math]::Round($_.Length / 1MB, 1)
|
||||
Write-Host "Uploading ${filename} (${size} MB)..."
|
||||
|
||||
try {
|
||||
$assets = Invoke-RestMethod -Uri "${REPO_API}/releases/${RELEASE_ID}/assets" -Headers $Headers
|
||||
$existing = $assets | Where-Object { $_.name -eq $filename }
|
||||
if ($existing) {
|
||||
Invoke-RestMethod -Uri "${REPO_API}/releases/${RELEASE_ID}/assets/$($existing.id)" -Method Delete -Headers $Headers
|
||||
}
|
||||
} catch {}
|
||||
|
||||
$uploadUrl = "${REPO_API}/releases/${RELEASE_ID}/assets?name=${encodedName}"
|
||||
$result = curl.exe --fail --silent --show-error `
|
||||
-X POST `
|
||||
-H "Authorization: token $env:BUILD_TOKEN" `
|
||||
-H "Content-Type: application/octet-stream" `
|
||||
-T "$($_.FullName)" `
|
||||
"$uploadUrl" 2>&1
|
||||
if ($LASTEXITCODE -eq 0) {
|
||||
Write-Host "Upload successful: ${filename}"
|
||||
} else {
|
||||
Write-Host "WARNING: Upload failed for ${filename}: ${result}"
|
||||
}
|
||||
}
|
||||
|
||||
build-macos:
|
||||
name: Build App (macOS)
|
||||
needs: bump-version
|
||||
runs-on: macos-latest
|
||||
env:
|
||||
NODE_VERSION: "20"
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
ref: ${{ needs.bump-version.outputs.tag }}
|
||||
|
||||
- name: Set up Node.js
|
||||
uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: ${{ env.NODE_VERSION }}
|
||||
|
||||
- name: Install Rust stable
|
||||
run: |
|
||||
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh -s -- -y --default-toolchain stable
|
||||
echo "$HOME/.cargo/bin" >> $GITHUB_PATH
|
||||
|
||||
- name: Install system dependencies
|
||||
run: brew install --quiet create-dmg || true
|
||||
|
||||
- name: Install npm dependencies
|
||||
run: npm ci
|
||||
|
||||
- name: Build Tauri app
|
||||
run: npm run tauri build
|
||||
|
||||
- name: Upload to release
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
which jq || brew install jq
|
||||
REPO_API="${GITHUB_SERVER_URL}/api/v1/repos/${GITHUB_REPOSITORY}"
|
||||
TAG="${{ needs.bump-version.outputs.tag }}"
|
||||
echo "Release tag: ${TAG}"
|
||||
|
||||
RELEASE_ID=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/tags/${TAG}" | jq -r '.id // empty')
|
||||
|
||||
if [ -z "${RELEASE_ID}" ] || [ "${RELEASE_ID}" = "null" ]; then
|
||||
echo "ERROR: Failed to find release for tag ${TAG}."
|
||||
exit 1
|
||||
fi
|
||||
echo "Release ID: ${RELEASE_ID}"
|
||||
|
||||
find src-tauri/target/release/bundle -type f -name "*.dmg" | while IFS= read -r file; do
|
||||
filename=$(basename "$file")
|
||||
encoded_name=$(echo "$filename" | sed 's/ /%20/g')
|
||||
echo "Uploading ${filename} ($(du -h "$file" | cut -f1))..."
|
||||
|
||||
ASSET_ID=$(curl -s -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets" | jq -r ".[] | select(.name == \"${filename}\") | .id // empty")
|
||||
if [ -n "${ASSET_ID}" ]; then
|
||||
curl -s -X DELETE -H "Authorization: token ${BUILD_TOKEN}" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets/${ASSET_ID}"
|
||||
fi
|
||||
|
||||
HTTP_CODE=$(curl -s -o /dev/null -w "%{http_code}" -X POST \
|
||||
-H "Authorization: token ${BUILD_TOKEN}" \
|
||||
-H "Content-Type: application/octet-stream" \
|
||||
-T "$file" \
|
||||
"${REPO_API}/releases/${RELEASE_ID}/assets?name=${encoded_name}")
|
||||
echo "Upload response: HTTP ${HTTP_CODE}"
|
||||
-H "Content-Type: application/json" \
|
||||
-d "{\"ref\": \"main\", \"inputs\": {\"tag\": \"${TAG}\"}}" \
|
||||
"${REPO_API}/actions/workflows/${workflow}/dispatches")
|
||||
echo " -> HTTP ${HTTP_CODE}"
|
||||
done
|
||||
|
||||
120
.gitea/workflows/sidecar-release.yml
Normal file
120
.gitea/workflows/sidecar-release.yml
Normal file
@@ -0,0 +1,120 @@
|
||||
name: Sidecar Release
|
||||
|
||||
on:
|
||||
push:
|
||||
branches: [main]
|
||||
paths:
|
||||
- 'client/**'
|
||||
- 'server/**'
|
||||
- 'backend/**'
|
||||
- 'pyproject.toml'
|
||||
- 'local-transcription-headless.spec'
|
||||
workflow_dispatch:
|
||||
|
||||
jobs:
|
||||
bump-sidecar-version:
|
||||
name: Bump sidecar version and tag
|
||||
if: "!contains(github.event.head_commit.message, '[skip ci]')"
|
||||
runs-on: ubuntu-latest
|
||||
outputs:
|
||||
version: ${{ steps.bump.outputs.version }}
|
||||
tag: ${{ steps.bump.outputs.tag }}
|
||||
has_changes: ${{ steps.check_changes.outputs.has_changes }}
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
with:
|
||||
fetch-depth: 2
|
||||
|
||||
- name: Check for backend changes
|
||||
id: check_changes
|
||||
run: |
|
||||
if [ "${{ github.event_name }}" = "workflow_dispatch" ]; then
|
||||
echo "has_changes=true" >> $GITHUB_OUTPUT
|
||||
exit 0
|
||||
fi
|
||||
CHANGED=$(git diff --name-only HEAD~1 HEAD -- client/ server/ backend/ pyproject.toml local-transcription-headless.spec 2>/dev/null || echo "")
|
||||
if [ -n "$CHANGED" ]; then
|
||||
echo "has_changes=true" >> $GITHUB_OUTPUT
|
||||
echo "Backend changes detected: $CHANGED"
|
||||
else
|
||||
echo "has_changes=false" >> $GITHUB_OUTPUT
|
||||
echo "No backend changes detected, skipping sidecar build"
|
||||
fi
|
||||
|
||||
- name: Configure git
|
||||
if: steps.check_changes.outputs.has_changes == 'true'
|
||||
run: |
|
||||
git config user.name "Gitea Actions"
|
||||
git config user.email "actions@gitea.local"
|
||||
|
||||
- name: Bump sidecar patch version
|
||||
if: steps.check_changes.outputs.has_changes == 'true'
|
||||
id: bump
|
||||
run: |
|
||||
CURRENT=$(grep '^version = ' pyproject.toml | head -1 | sed 's/version = "\(.*\)"/\1/')
|
||||
echo "Current sidecar version: ${CURRENT}"
|
||||
|
||||
MAJOR=$(echo "${CURRENT}" | cut -d. -f1)
|
||||
MINOR=$(echo "${CURRENT}" | cut -d. -f2)
|
||||
PATCH=$(echo "${CURRENT}" | cut -d. -f3)
|
||||
NEW_PATCH=$((PATCH + 1))
|
||||
NEW_VERSION="${MAJOR}.${MINOR}.${NEW_PATCH}"
|
||||
echo "New sidecar version: ${NEW_VERSION}"
|
||||
|
||||
sed -i "s/^version = \"${CURRENT}\"/version = \"${NEW_VERSION}\"/" pyproject.toml
|
||||
sed -i "s/__version__ = \"${CURRENT}\"/__version__ = \"${NEW_VERSION}\"/" version.py
|
||||
sed -i "s/__version_info__ = .*/__version_info__ = (${MAJOR}, ${MINOR}, ${NEW_PATCH})/" version.py
|
||||
|
||||
echo "version=${NEW_VERSION}" >> $GITHUB_OUTPUT
|
||||
echo "tag=sidecar-v${NEW_VERSION}" >> $GITHUB_OUTPUT
|
||||
|
||||
- name: Commit and tag
|
||||
if: steps.check_changes.outputs.has_changes == 'true'
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
NEW_VERSION="${{ steps.bump.outputs.version }}"
|
||||
TAG="${{ steps.bump.outputs.tag }}"
|
||||
git add pyproject.toml version.py
|
||||
git commit -m "chore: bump sidecar version to ${NEW_VERSION} [skip ci]"
|
||||
git tag "${TAG}"
|
||||
|
||||
REMOTE_URL=$(git remote get-url origin | sed "s|://|://gitea-actions:${BUILD_TOKEN}@|")
|
||||
git pull --rebase "${REMOTE_URL}" main || true
|
||||
git push "${REMOTE_URL}" HEAD:main
|
||||
git push "${REMOTE_URL}" "${TAG}"
|
||||
|
||||
- name: Create Gitea release
|
||||
if: steps.check_changes.outputs.has_changes == 'true'
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
REPO_API="${GITHUB_SERVER_URL}/api/v1/repos/${GITHUB_REPOSITORY}"
|
||||
TAG="${{ steps.bump.outputs.tag }}"
|
||||
VERSION="${{ steps.bump.outputs.version }}"
|
||||
RELEASE_NAME="Sidecar v${VERSION}"
|
||||
|
||||
curl -s -X POST \
|
||||
-H "Authorization: token ${BUILD_TOKEN}" \
|
||||
-H "Content-Type: application/json" \
|
||||
-d "{\"tag_name\": \"${TAG}\", \"name\": \"${RELEASE_NAME}\", \"body\": \"Automated sidecar build.\", \"draft\": false, \"prerelease\": false}" \
|
||||
"${REPO_API}/releases"
|
||||
echo "Created release: ${RELEASE_NAME}"
|
||||
|
||||
- name: Trigger per-OS sidecar builds
|
||||
if: steps.check_changes.outputs.has_changes == 'true'
|
||||
env:
|
||||
BUILD_TOKEN: ${{ secrets.BUILD_TOKEN }}
|
||||
run: |
|
||||
REPO_API="${GITHUB_SERVER_URL}/api/v1/repos/${GITHUB_REPOSITORY}"
|
||||
TAG="${{ steps.bump.outputs.tag }}"
|
||||
|
||||
for workflow in build-sidecar-linux.yml build-sidecar-windows.yml build-sidecar-macos.yml; do
|
||||
echo "Dispatching ${workflow} for ${TAG}..."
|
||||
HTTP_CODE=$(curl -s -o /dev/null -w "%{http_code}" -X POST \
|
||||
-H "Authorization: token ${BUILD_TOKEN}" \
|
||||
-H "Content-Type: application/json" \
|
||||
-d "{\"ref\": \"main\", \"inputs\": {\"tag\": \"${TAG}\"}}" \
|
||||
"${REPO_API}/actions/workflows/${workflow}/dispatches")
|
||||
echo " -> HTTP ${HTTP_CODE}"
|
||||
done
|
||||
57
.gitea/workflows/test.yml
Normal file
57
.gitea/workflows/test.yml
Normal file
@@ -0,0 +1,57 @@
|
||||
name: Tests
|
||||
|
||||
on:
|
||||
push:
|
||||
branches: [main]
|
||||
pull_request:
|
||||
branches: [main]
|
||||
|
||||
jobs:
|
||||
python-tests:
|
||||
name: Python Backend Tests
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
|
||||
- name: Install test dependencies
|
||||
run: |
|
||||
pip install --break-system-packages pytest httpx pytest-asyncio anyio fastapi pydantic pyyaml uvicorn requests
|
||||
|
||||
- name: Run pytest
|
||||
run: python3 -m pytest backend/tests/ client/tests/ -v --tb=short
|
||||
|
||||
frontend-tests:
|
||||
name: Frontend Tests
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
|
||||
- uses: actions/setup-node@v4
|
||||
with:
|
||||
node-version: 20
|
||||
|
||||
- name: Install dependencies
|
||||
run: npm ci
|
||||
|
||||
- name: Run Vitest
|
||||
run: npx vitest run
|
||||
|
||||
rust-tests:
|
||||
name: Rust Sidecar Tests
|
||||
runs-on: ubuntu-latest
|
||||
steps:
|
||||
- uses: actions/checkout@v4
|
||||
|
||||
- name: Install Rust
|
||||
run: |
|
||||
curl --proto '=https' --tlsv1.2 -sSf https://sh.rustup.rs | sh -s -- -y --default-toolchain stable
|
||||
echo "$HOME/.cargo/bin" >> $GITHUB_PATH
|
||||
|
||||
- name: Install Tauri system dependencies
|
||||
run: |
|
||||
sudo apt-get update
|
||||
sudo apt-get install -y libgtk-3-dev libwebkit2gtk-4.1-dev libappindicator3-dev librsvg2-dev patchelf
|
||||
|
||||
- name: Run cargo test
|
||||
working-directory: src-tauri
|
||||
run: cargo test
|
||||
27
CLAUDE.md
27
CLAUDE.md
@@ -64,8 +64,14 @@ local-transcription/
|
||||
│ ├── web_display.py # FastAPI OBS display server (WebSocket + HTML)
|
||||
│ └── nodejs/ # Optional multi-user sync server
|
||||
├── .gitea/workflows/ # CI/CD
|
||||
│ ├── release.yml # Tauri app builds (Linux/Windows/macOS)
|
||||
│ └── build-sidecar.yml # Python sidecar builds (CUDA + CPU)
|
||||
│ ├── release.yml # Coordinator: version bump, tag, release creation
|
||||
│ ├── build-app-linux.yml # Linux Tauri app build (triggered by v* tag)
|
||||
│ ├── build-app-windows.yml # Windows Tauri app build (triggered by v* tag)
|
||||
│ ├── build-app-macos.yml # macOS Tauri app build (triggered by v* tag)
|
||||
│ ├── sidecar-release.yml # Sidecar coordinator: version bump, tag, release
|
||||
│ ├── build-sidecar-linux.yml # Linux sidecar build (triggered by sidecar-v* tag)
|
||||
│ ├── build-sidecar-windows.yml # Windows sidecar build (triggered by sidecar-v* tag)
|
||||
│ └── build-sidecar-macos.yml # macOS sidecar build (triggered by sidecar-v* tag)
|
||||
├── config/default_config.yaml # Default settings template
|
||||
├── main.py # Legacy PySide6 GUI entry point
|
||||
├── main_cli.py # CLI version for testing
|
||||
@@ -205,12 +211,21 @@ Uses Svelte 5 runes throughout (`$state`, `$derived`, `$effect`, `$props`). No S
|
||||
|
||||
## CI/CD
|
||||
|
||||
Two Gitea Actions workflows in `.gitea/workflows/`:
|
||||
Eight Gitea Actions workflows in `.gitea/workflows/`, split into coordinators and per-OS builders:
|
||||
|
||||
- **`release.yml`**: Triggers on push to `main`. Auto-bumps version, builds Tauri app on Linux/Windows/macOS, uploads `.deb`, `.rpm`, `.msi`, `.dmg` to Gitea release.
|
||||
- **`build-sidecar.yml`**: Triggers on changes to `client/`, `server/`, `backend/`, `pyproject.toml`. Builds headless Python sidecar via PyInstaller. CUDA + CPU for Linux/Windows, CPU-only for macOS.
|
||||
**App release (Tauri):**
|
||||
- **`release.yml`**: Coordinator. Triggers on push to `main`. Auto-bumps version in package.json/tauri.conf.json/Cargo.toml/version.py, commits, tags `v{VERSION}`, creates Gitea release.
|
||||
- **`build-app-linux.yml`**: Triggers on `v*` tag push or `workflow_dispatch`. Builds Tauri app, uploads `.deb`/`.rpm`/`.AppImage`.
|
||||
- **`build-app-windows.yml`**: Triggers on `v*` tag push or `workflow_dispatch`. Builds Tauri app, uploads `.msi`/`*-setup.exe`.
|
||||
- **`build-app-macos.yml`**: Triggers on `v*` tag push or `workflow_dispatch`. Builds Tauri app, uploads `.dmg`.
|
||||
|
||||
Both require a `BUILD_TOKEN` secret (Gitea API token with release write access).
|
||||
**Sidecar release (Python backend):**
|
||||
- **`sidecar-release.yml`**: Coordinator. Triggers on push to `main` with changes in `client/`, `server/`, `backend/`, `pyproject.toml`, or `local-transcription-headless.spec`. Bumps version in pyproject.toml/version.py, tags `sidecar-v{VERSION}`, creates Gitea release.
|
||||
- **`build-sidecar-linux.yml`**: Triggers on `sidecar-v*` tag push or `workflow_dispatch`. Builds CUDA + CPU sidecars via PyInstaller.
|
||||
- **`build-sidecar-windows.yml`**: Triggers on `sidecar-v*` tag push or `workflow_dispatch`. Builds CUDA + CPU sidecars via PyInstaller.
|
||||
- **`build-sidecar-macos.yml`**: Triggers on `sidecar-v*` tag push or `workflow_dispatch`. Builds CPU-only sidecar via PyInstaller.
|
||||
|
||||
All per-OS build workflows can be re-run independently via `workflow_dispatch` with an optional `tag` input. All require a `BUILD_TOKEN` secret (Gitea API token with release write access).
|
||||
|
||||
## Common Patterns
|
||||
|
||||
|
||||
@@ -99,11 +99,19 @@ class APIServer:
|
||||
|
||||
self.controller.on_credits_low = on_credits_low
|
||||
|
||||
def set_event_loop(self, loop: asyncio.AbstractEventLoop):
|
||||
"""Set the event loop used for broadcasting (call from uvicorn startup)."""
|
||||
self._event_loop = loop
|
||||
|
||||
def _broadcast_control(self, data: dict):
|
||||
"""Send a message to all connected /ws/control clients."""
|
||||
if not self.control_connections:
|
||||
return
|
||||
|
||||
loop = getattr(self, '_event_loop', None)
|
||||
if loop is None:
|
||||
return
|
||||
|
||||
message = json.dumps(data)
|
||||
disconnected = []
|
||||
|
||||
@@ -111,7 +119,7 @@ class APIServer:
|
||||
try:
|
||||
asyncio.run_coroutine_threadsafe(
|
||||
ws.send_text(message),
|
||||
asyncio.get_event_loop(),
|
||||
loop,
|
||||
)
|
||||
except Exception:
|
||||
disconnected.append(ws)
|
||||
@@ -124,6 +132,10 @@ class APIServer:
|
||||
app = self.app
|
||||
ctrl = self.controller
|
||||
|
||||
@app.on_event("startup")
|
||||
async def on_startup():
|
||||
self.set_event_loop(asyncio.get_event_loop())
|
||||
|
||||
# ── Status ─────────────────────────────────────────────
|
||||
|
||||
@app.get("/api/status")
|
||||
|
||||
@@ -88,11 +88,16 @@ def main():
|
||||
# Create API server wrapping the controller
|
||||
api_server = APIServer(controller)
|
||||
|
||||
# Determine actual port (web server may have shifted if port was in use)
|
||||
actual_port = controller.actual_web_port or args.port
|
||||
# OBS display runs on the configured port, API server on port+1
|
||||
obs_port = controller.actual_web_port or args.port
|
||||
api_port = obs_port + 1
|
||||
|
||||
# Print ready event so Tauri can discover the port
|
||||
print(json.dumps({"event": "ready", "port": actual_port}), flush=True)
|
||||
# Print ready event so Tauri can discover the API port
|
||||
print(json.dumps({
|
||||
"event": "ready",
|
||||
"port": api_port,
|
||||
"obs_port": obs_port,
|
||||
}), flush=True)
|
||||
|
||||
# Run the API server (blocks)
|
||||
import uvicorn
|
||||
@@ -104,7 +109,7 @@ def main():
|
||||
uvicorn.run(
|
||||
api_server.app,
|
||||
host=args.host,
|
||||
port=actual_port + 1, # API on port+1, OBS display on the main port
|
||||
port=api_port,
|
||||
log_level="error",
|
||||
access_log=False,
|
||||
)
|
||||
|
||||
0
backend/tests/__init__.py
Normal file
0
backend/tests/__init__.py
Normal file
159
backend/tests/conftest.py
Normal file
159
backend/tests/conftest.py
Normal file
@@ -0,0 +1,159 @@
|
||||
"""Shared fixtures for backend tests.
|
||||
|
||||
Heavy third-party modules (torch, sounddevice, numpy, RealtimeSTT, etc.) are
|
||||
stubbed at the *sys.modules* level before any backend code is imported. This
|
||||
lets the test suite run on a plain Python install without GPU drivers, audio
|
||||
hardware, or heavyweight ML libraries.
|
||||
"""
|
||||
|
||||
import sys
|
||||
import types
|
||||
from pathlib import Path
|
||||
from unittest.mock import MagicMock
|
||||
|
||||
import pytest
|
||||
|
||||
# ── Project root on sys.path ────────────────────────────────────────
|
||||
project_root = Path(__file__).resolve().parent.parent.parent
|
||||
if str(project_root) not in sys.path:
|
||||
sys.path.insert(0, str(project_root))
|
||||
|
||||
|
||||
# ── Stub heavy modules before anything imports them ─────────────────
|
||||
|
||||
def _stub(name: str) -> types.ModuleType:
|
||||
"""Create a stub module and register it in sys.modules if not already present."""
|
||||
if name in sys.modules:
|
||||
return sys.modules[name]
|
||||
mod = types.ModuleType(name)
|
||||
sys.modules[name] = mod
|
||||
return mod
|
||||
|
||||
|
||||
# numpy -- must behave like a real module for `import numpy as np`
|
||||
_np = _stub("numpy")
|
||||
_np.float32 = float
|
||||
_np.float64 = float
|
||||
_np.int16 = int
|
||||
_np.ndarray = MagicMock
|
||||
_np.array = MagicMock(return_value=MagicMock())
|
||||
_np.zeros = MagicMock(return_value=MagicMock())
|
||||
_np.frombuffer = MagicMock(return_value=MagicMock())
|
||||
|
||||
# torch + sub-modules
|
||||
_torch = _stub("torch")
|
||||
_torch.cuda = MagicMock()
|
||||
_torch.cuda.is_available = MagicMock(return_value=False)
|
||||
_torch.backends = MagicMock()
|
||||
_torch.backends.mps = MagicMock()
|
||||
_torch.backends.mps.is_available = MagicMock(return_value=False)
|
||||
_stub("torch.cuda")
|
||||
_stub("torch.backends")
|
||||
_stub("torch.backends.mps")
|
||||
_stub("torchaudio")
|
||||
|
||||
# sounddevice
|
||||
_sd = _stub("sounddevice")
|
||||
_sd.query_devices = MagicMock(return_value=[])
|
||||
|
||||
# RealtimeSTT (imported by transcription_engine_realtime)
|
||||
_rtstt = _stub("RealtimeSTT")
|
||||
_rtstt.AudioToTextRecorder = MagicMock
|
||||
|
||||
# faster_whisper (sometimes imported transitively)
|
||||
_stub("faster_whisper")
|
||||
|
||||
# noisereduce
|
||||
_stub("noisereduce")
|
||||
|
||||
# scipy
|
||||
_scipy = _stub("scipy")
|
||||
_stub("scipy.signal")
|
||||
_stub("scipy.io")
|
||||
_stub("scipy.io.wavfile")
|
||||
|
||||
# webrtcvad
|
||||
_stub("webrtcvad")
|
||||
|
||||
# openwakeword
|
||||
_stub("openwakeword")
|
||||
|
||||
# pvporcupine
|
||||
_stub("pvporcupine")
|
||||
|
||||
# PySide6 (should not be needed, but just in case)
|
||||
_stub("PySide6")
|
||||
_stub("PySide6.QtWidgets")
|
||||
_stub("PySide6.QtCore")
|
||||
_stub("PySide6.QtGui")
|
||||
|
||||
# websockets
|
||||
_ws = _stub("websockets")
|
||||
_ws.connect = MagicMock
|
||||
|
||||
# deepgram (cloud transcription)
|
||||
_stub("deepgram")
|
||||
|
||||
|
||||
# ── Fixtures ────────────────────────────────────────────────────────
|
||||
|
||||
@pytest.fixture
|
||||
def mock_config(tmp_path):
|
||||
"""Return a Config object backed by a temporary file.
|
||||
|
||||
This avoids touching the real user config at ~/.local-transcription/.
|
||||
"""
|
||||
config_file = tmp_path / "test_config.yaml"
|
||||
from client.config import Config
|
||||
|
||||
config = Config(config_path=str(config_file))
|
||||
return config
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def controller(mock_config):
|
||||
"""Return an AppController wired to *mock_config* without starting heavy
|
||||
subsystems (engine, web server, device manager).
|
||||
|
||||
The transcription engine, web server thread, and DeviceManager are all
|
||||
replaced with lightweight mocks so the test suite can run without a GPU,
|
||||
audio hardware, or a free network port.
|
||||
"""
|
||||
from unittest.mock import patch
|
||||
|
||||
with patch("backend.app_controller.DeviceManager") as MockDM, \
|
||||
patch("backend.app_controller.RealtimeTranscriptionEngine"), \
|
||||
patch("backend.app_controller.DeepgramTranscriptionEngine"), \
|
||||
patch("backend.app_controller.TranscriptionWebServer"), \
|
||||
patch("backend.app_controller.ServerSyncClient"):
|
||||
|
||||
# DeviceManager stub
|
||||
dm_instance = MagicMock()
|
||||
dm_instance.get_device_info.return_value = [("cpu", "CPU")]
|
||||
dm_instance.get_device_for_whisper.return_value = "cpu"
|
||||
MockDM.return_value = dm_instance
|
||||
|
||||
from backend.app_controller import AppController
|
||||
|
||||
ctrl = AppController(config=mock_config)
|
||||
yield ctrl
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def api_client(controller):
|
||||
"""Return an httpx.AsyncClient speaking ASGI to the APIServer's FastAPI app.
|
||||
|
||||
Usage in tests::
|
||||
|
||||
async def test_something(api_client):
|
||||
resp = await api_client.get("/api/status")
|
||||
assert resp.status_code == 200
|
||||
"""
|
||||
from backend.api_server import APIServer
|
||||
import httpx
|
||||
|
||||
api = APIServer(controller)
|
||||
|
||||
transport = httpx.ASGITransport(app=api.app)
|
||||
client = httpx.AsyncClient(transport=transport, base_url="http://testserver")
|
||||
return client
|
||||
150
backend/tests/test_api_server.py
Normal file
150
backend/tests/test_api_server.py
Normal file
@@ -0,0 +1,150 @@
|
||||
"""Tests for backend.api_server.APIServer REST endpoints."""
|
||||
|
||||
import sys
|
||||
from pathlib import Path
|
||||
from unittest.mock import MagicMock, patch
|
||||
|
||||
import pytest
|
||||
import pytest_asyncio
|
||||
|
||||
# Ensure project root is on path
|
||||
project_root = Path(__file__).resolve().parent.parent.parent
|
||||
sys.path.insert(0, str(project_root))
|
||||
|
||||
|
||||
# ── GET /api/status ─────────────────────────────────────────────────
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_get_status(api_client):
|
||||
resp = await api_client.get("/api/status")
|
||||
assert resp.status_code == 200
|
||||
data = resp.json()
|
||||
assert "state" in data
|
||||
assert "is_transcribing" in data
|
||||
assert "version" in data
|
||||
assert "web_server" in data
|
||||
|
||||
|
||||
# ── GET /api/config ─────────────────────────────────────────────────
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_get_config(api_client):
|
||||
resp = await api_client.get("/api/config")
|
||||
assert resp.status_code == 200
|
||||
data = resp.json()
|
||||
# The config should be a dict (the raw config mapping)
|
||||
assert isinstance(data, dict)
|
||||
|
||||
|
||||
# ── PUT /api/config ─────────────────────────────────────────────────
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_put_config(api_client, controller):
|
||||
"""Updating config via PUT should persist and return success."""
|
||||
# Patch reload_engine to avoid heavy lifting
|
||||
controller.reload_engine = MagicMock(return_value=(True, "ok"))
|
||||
controller.current_model_size = "base.en"
|
||||
controller.current_device_config = "auto"
|
||||
controller.config.set("transcription.model", "base.en")
|
||||
controller.config.set("transcription.device", "auto")
|
||||
|
||||
resp = await api_client.put(
|
||||
"/api/config",
|
||||
json={"settings": {"display.font_size": 24}},
|
||||
)
|
||||
assert resp.status_code == 200
|
||||
body = resp.json()
|
||||
assert body["status"] == "ok"
|
||||
|
||||
# Verify the value was actually saved
|
||||
assert controller.config.get("display.font_size") == 24
|
||||
|
||||
|
||||
# ── POST /api/start (engine not ready) ─────────────────────────────
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_start_when_not_ready(api_client, controller):
|
||||
"""Starting transcription without an engine should return 400."""
|
||||
controller.transcription_engine = None
|
||||
resp = await api_client.post("/api/start")
|
||||
assert resp.status_code == 400
|
||||
|
||||
|
||||
# ── POST /api/clear ─────────────────────────────────────────────────
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_clear(api_client, controller):
|
||||
from client.transcription_engine_realtime import TranscriptionResult
|
||||
from datetime import datetime
|
||||
|
||||
controller.transcriptions = [
|
||||
TranscriptionResult(text="One", is_final=True, timestamp=datetime.now(), user_name="U"),
|
||||
]
|
||||
|
||||
resp = await api_client.post("/api/clear")
|
||||
assert resp.status_code == 200
|
||||
body = resp.json()
|
||||
assert body["cleared"] == 1
|
||||
assert len(controller.transcriptions) == 0
|
||||
|
||||
|
||||
# ── GET /api/audio-devices ──────────────────────────────────────────
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_get_audio_devices(api_client, controller):
|
||||
"""Audio devices endpoint should return a list, even when mocked."""
|
||||
# Mock sounddevice so the test works without audio hardware
|
||||
with patch("backend.app_controller.AppController.get_audio_devices",
|
||||
return_value=[{"index": 0, "name": "Mock Mic"}]):
|
||||
resp = await api_client.get("/api/audio-devices")
|
||||
assert resp.status_code == 200
|
||||
data = resp.json()
|
||||
assert "devices" in data
|
||||
assert len(data["devices"]) >= 1
|
||||
|
||||
|
||||
# ── GET /api/compute-devices ────────────────────────────────────────
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_get_compute_devices(api_client, controller):
|
||||
resp = await api_client.get("/api/compute-devices")
|
||||
assert resp.status_code == 200
|
||||
data = resp.json()
|
||||
assert "devices" in data
|
||||
# At minimum we get the "Auto-detect" entry
|
||||
assert any(d["id"] == "auto" for d in data["devices"])
|
||||
|
||||
|
||||
# ── GET /api/check-update ──────────────────────────────────────────
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_check_update(api_client, controller):
|
||||
"""check-update should return a dict with an 'available' key."""
|
||||
with patch.object(controller, "check_for_updates",
|
||||
return_value={"available": False, "current_version": "1.0.0"}):
|
||||
resp = await api_client.get("/api/check-update")
|
||||
assert resp.status_code == 200
|
||||
data = resp.json()
|
||||
assert "available" in data
|
||||
|
||||
|
||||
# ── GET /api/version ────────────────────────────────────────────────
|
||||
|
||||
|
||||
@pytest.mark.asyncio
|
||||
async def test_version(api_client):
|
||||
resp = await api_client.get("/api/version")
|
||||
assert resp.status_code == 200
|
||||
data = resp.json()
|
||||
assert "version" in data
|
||||
# Should be a non-empty string
|
||||
assert isinstance(data["version"], str)
|
||||
assert len(data["version"]) > 0
|
||||
181
backend/tests/test_app_controller.py
Normal file
181
backend/tests/test_app_controller.py
Normal file
@@ -0,0 +1,181 @@
|
||||
"""Tests for backend.app_controller.AppController."""
|
||||
|
||||
import sys
|
||||
from datetime import datetime
|
||||
from pathlib import Path
|
||||
from unittest.mock import MagicMock, patch
|
||||
|
||||
import pytest
|
||||
|
||||
# Ensure project root is on path
|
||||
project_root = Path(__file__).resolve().parent.parent.parent
|
||||
sys.path.insert(0, str(project_root))
|
||||
|
||||
from backend.app_controller import AppState
|
||||
|
||||
|
||||
# ── basic state ─────────────────────────────────────────────────────
|
||||
|
||||
|
||||
def test_initial_state(controller):
|
||||
"""A freshly constructed controller should be INITIALIZING and not transcribing."""
|
||||
assert controller.state == AppState.INITIALIZING
|
||||
assert controller.is_transcribing is False
|
||||
|
||||
|
||||
# ── start / stop ────────────────────────────────────────────────────
|
||||
|
||||
|
||||
def test_start_transcription_without_engine(controller):
|
||||
"""Starting transcription before the engine is ready should fail gracefully."""
|
||||
controller.transcription_engine = None
|
||||
success, message = controller.start_transcription()
|
||||
assert success is False
|
||||
assert "not ready" in message.lower()
|
||||
|
||||
|
||||
def test_start_stop_cycle(controller):
|
||||
"""Full start -> stop cycle with a mocked engine that reports ready."""
|
||||
engine = MagicMock()
|
||||
engine.is_ready.return_value = True
|
||||
engine.start_recording.return_value = True
|
||||
controller.transcription_engine = engine
|
||||
|
||||
# Start
|
||||
ok, msg = controller.start_transcription()
|
||||
assert ok is True
|
||||
assert controller.is_transcribing is True
|
||||
assert controller.state == AppState.TRANSCRIBING
|
||||
|
||||
# Stop
|
||||
ok, msg = controller.stop_transcription()
|
||||
assert ok is True
|
||||
assert controller.is_transcribing is False
|
||||
engine.stop_recording.assert_called_once()
|
||||
|
||||
|
||||
def test_double_start_rejected(controller):
|
||||
"""Calling start_transcription twice should reject the second call."""
|
||||
engine = MagicMock()
|
||||
engine.is_ready.return_value = True
|
||||
engine.start_recording.return_value = True
|
||||
controller.transcription_engine = engine
|
||||
|
||||
controller.start_transcription()
|
||||
success, message = controller.start_transcription()
|
||||
assert success is False
|
||||
assert "already" in message.lower()
|
||||
|
||||
|
||||
# ── transcription storage ───────────────────────────────────────────
|
||||
|
||||
|
||||
def test_clear_transcriptions(controller):
|
||||
"""clear_transcriptions should empty the list and return the count."""
|
||||
from client.transcription_engine_realtime import TranscriptionResult
|
||||
|
||||
controller.transcriptions = [
|
||||
TranscriptionResult(text="Hello", is_final=True, timestamp=datetime.now(), user_name="Alice"),
|
||||
TranscriptionResult(text="World", is_final=True, timestamp=datetime.now(), user_name="Bob"),
|
||||
]
|
||||
count = controller.clear_transcriptions()
|
||||
assert count == 2
|
||||
assert len(controller.transcriptions) == 0
|
||||
|
||||
|
||||
def test_get_transcriptions_text_with_timestamps(controller):
|
||||
"""get_transcriptions_text should include [HH:MM:SS] prefixes when requested."""
|
||||
from client.transcription_engine_realtime import TranscriptionResult
|
||||
|
||||
ts = datetime(2025, 1, 15, 10, 30, 45)
|
||||
controller.transcriptions = [
|
||||
TranscriptionResult(text="Test line", is_final=True, timestamp=ts, user_name="User"),
|
||||
]
|
||||
|
||||
text = controller.get_transcriptions_text(include_timestamps=True)
|
||||
assert "[10:30:45]" in text
|
||||
assert "User:" in text
|
||||
assert "Test line" in text
|
||||
|
||||
|
||||
# ── settings / engine reload ────────────────────────────────────────
|
||||
|
||||
|
||||
def test_apply_settings_triggers_reload_on_model_change(controller):
|
||||
"""Changing the transcription model should trigger an engine reload."""
|
||||
controller.current_model_size = "base.en"
|
||||
controller.current_device_config = "auto"
|
||||
|
||||
# Patch reload_engine so it doesn't actually try to spin up threads
|
||||
controller.reload_engine = MagicMock(return_value=(True, "reloaded"))
|
||||
|
||||
reloaded, msg = controller.apply_settings({
|
||||
"transcription.model": "small.en",
|
||||
})
|
||||
|
||||
assert reloaded is True
|
||||
controller.reload_engine.assert_called_once()
|
||||
|
||||
|
||||
def test_apply_settings_no_reload_when_same(controller):
|
||||
"""If model and device haven't changed, no reload should happen."""
|
||||
controller.current_model_size = "base.en"
|
||||
controller.current_device_config = "auto"
|
||||
|
||||
# Ensure config returns the same values
|
||||
controller.config.set("transcription.model", "base.en")
|
||||
controller.config.set("transcription.device", "auto")
|
||||
|
||||
controller.reload_engine = MagicMock(return_value=(True, "reloaded"))
|
||||
|
||||
reloaded, msg = controller.apply_settings({
|
||||
"display.font_size": 20,
|
||||
})
|
||||
|
||||
assert reloaded is False
|
||||
controller.reload_engine.assert_not_called()
|
||||
|
||||
|
||||
# ── transcription callbacks ─────────────────────────────────────────
|
||||
|
||||
|
||||
def test_on_final_transcription_callback_fires(controller):
|
||||
"""_on_final_transcription should append and invoke on_transcription callback."""
|
||||
from client.transcription_engine_realtime import TranscriptionResult
|
||||
|
||||
received = []
|
||||
controller.on_transcription = lambda data: received.append(data)
|
||||
|
||||
controller.is_transcribing = True
|
||||
controller._set_state(AppState.TRANSCRIBING)
|
||||
|
||||
result = TranscriptionResult(
|
||||
text="Hello world",
|
||||
is_final=True,
|
||||
timestamp=datetime.now(),
|
||||
user_name="Tester",
|
||||
)
|
||||
controller._on_final_transcription(result)
|
||||
|
||||
assert len(controller.transcriptions) == 1
|
||||
assert len(received) == 1
|
||||
assert received[0]["text"] == "Hello world"
|
||||
assert received[0]["user_name"] == "Tester"
|
||||
assert received[0]["is_preview"] is False
|
||||
|
||||
|
||||
def test_on_final_transcription_ignored_when_not_transcribing(controller):
|
||||
"""If the controller is not in transcribing state the callback should be a no-op."""
|
||||
from client.transcription_engine_realtime import TranscriptionResult
|
||||
|
||||
controller.is_transcribing = False
|
||||
|
||||
result = TranscriptionResult(
|
||||
text="Should be ignored",
|
||||
is_final=True,
|
||||
timestamp=datetime.now(),
|
||||
user_name="Ghost",
|
||||
)
|
||||
controller._on_final_transcription(result)
|
||||
|
||||
assert len(controller.transcriptions) == 0
|
||||
56
backend/tests/test_main_headless.py
Normal file
56
backend/tests/test_main_headless.py
Normal file
@@ -0,0 +1,56 @@
|
||||
"""Tests for backend.main_headless ready-event JSON format."""
|
||||
|
||||
import sys
|
||||
from pathlib import Path
|
||||
|
||||
import pytest
|
||||
|
||||
# Ensure project root is on path
|
||||
project_root = Path(__file__).resolve().parent.parent.parent
|
||||
sys.path.insert(0, str(project_root))
|
||||
|
||||
|
||||
def test_ready_event_reports_api_port_not_obs_port():
|
||||
"""The ready JSON printed by main_headless must set ``port`` to
|
||||
``obs_port + 1`` (the API port), not the OBS display port.
|
||||
|
||||
From main_headless.py::
|
||||
|
||||
obs_port = controller.actual_web_port or args.port
|
||||
api_port = obs_port + 1
|
||||
|
||||
print(json.dumps({
|
||||
"event": "ready",
|
||||
"port": api_port,
|
||||
"obs_port": obs_port,
|
||||
}), flush=True)
|
||||
|
||||
We verify this contract by reading the source and checking the
|
||||
structure directly (running main() would start a real server).
|
||||
"""
|
||||
import ast
|
||||
import textwrap
|
||||
|
||||
source_path = project_root / "backend" / "main_headless.py"
|
||||
source = source_path.read_text()
|
||||
|
||||
# Verify the key relationships exist in the source:
|
||||
# 1. api_port = obs_port + 1
|
||||
assert "api_port = obs_port + 1" in source, (
|
||||
"Expected `api_port = obs_port + 1` in main_headless.py"
|
||||
)
|
||||
|
||||
# 2. The ready event JSON uses api_port for "port", not obs_port
|
||||
assert '"port": api_port' in source or "'port': api_port" in source, (
|
||||
"The ready event should report api_port as 'port'"
|
||||
)
|
||||
|
||||
# 3. obs_port is also included separately
|
||||
assert '"obs_port": obs_port' in source or "'obs_port': obs_port" in source, (
|
||||
"The ready event should also include 'obs_port'"
|
||||
)
|
||||
|
||||
# 4. Verify the event name
|
||||
assert '"event": "ready"' in source or "'event': 'ready'" in source, (
|
||||
"The ready event should have event='ready'"
|
||||
)
|
||||
0
client/tests/__init__.py
Normal file
0
client/tests/__init__.py
Normal file
78
client/tests/test_config.py
Normal file
78
client/tests/test_config.py
Normal file
@@ -0,0 +1,78 @@
|
||||
"""Tests for client.config.Config."""
|
||||
|
||||
import sys
|
||||
from pathlib import Path
|
||||
|
||||
import pytest
|
||||
|
||||
# Ensure project root is on path
|
||||
project_root = Path(__file__).resolve().parent.parent.parent
|
||||
sys.path.insert(0, str(project_root))
|
||||
|
||||
from client.config import Config
|
||||
|
||||
|
||||
@pytest.fixture
|
||||
def cfg(tmp_path):
|
||||
"""A Config backed by a temp file so we never touch the real user config."""
|
||||
return Config(config_path=str(tmp_path / "test_config.yaml"))
|
||||
|
||||
|
||||
# ── dot-notation get ────────────────────────────────────────────────
|
||||
|
||||
|
||||
def test_dot_notation_get(cfg):
|
||||
"""Config.get should traverse nested dicts using dot-separated keys."""
|
||||
cfg.config = {"audio": {"sample_rate": 16000}}
|
||||
assert cfg.get("audio.sample_rate") == 16000
|
||||
|
||||
|
||||
# ── dot-notation set ────────────────────────────────────────────────
|
||||
|
||||
|
||||
def test_dot_notation_set(cfg):
|
||||
"""Config.set should create/update nested values via dot-separated keys."""
|
||||
cfg.set("audio.sample_rate", 44100)
|
||||
assert cfg.config["audio"]["sample_rate"] == 44100
|
||||
|
||||
# Also readable via .get
|
||||
assert cfg.get("audio.sample_rate") == 44100
|
||||
|
||||
|
||||
# ── missing key returns default ─────────────────────────────────────
|
||||
|
||||
|
||||
def test_missing_key_returns_default(cfg):
|
||||
"""Accessing a nonexistent key should return the supplied default."""
|
||||
assert cfg.get("nonexistent.path", "fallback") == "fallback"
|
||||
assert cfg.get("also.missing") is None # default default is None
|
||||
|
||||
|
||||
# ── nested set creates intermediate dicts ───────────────────────────
|
||||
|
||||
|
||||
def test_nested_set_creates_path(cfg):
|
||||
"""Setting a deeply nested key should create all intermediate dicts."""
|
||||
cfg.config = {}
|
||||
cfg.set("a.b.c.d", 42)
|
||||
|
||||
assert cfg.config["a"]["b"]["c"]["d"] == 42
|
||||
assert cfg.get("a.b.c.d") == 42
|
||||
|
||||
|
||||
# ── save and reload round-trip ──────────────────────────────────────
|
||||
|
||||
|
||||
def test_save_and_reload(tmp_path):
|
||||
"""Values persisted via save() should survive a fresh Config load."""
|
||||
config_file = str(tmp_path / "roundtrip.yaml")
|
||||
|
||||
# Create and populate
|
||||
cfg1 = Config(config_path=config_file)
|
||||
cfg1.set("user.name", "TestUser")
|
||||
cfg1.set("transcription.model", "tiny.en")
|
||||
|
||||
# Load a fresh instance from the same file
|
||||
cfg2 = Config(config_path=config_file)
|
||||
assert cfg2.get("user.name") == "TestUser"
|
||||
assert cfg2.get("transcription.model") == "tiny.en"
|
||||
1085
package-lock.json
generated
1085
package-lock.json
generated
File diff suppressed because it is too large
Load Diff
11
package.json
11
package.json
@@ -1,7 +1,7 @@
|
||||
{
|
||||
"name": "local-transcription",
|
||||
"private": true,
|
||||
"version": "1.4.3",
|
||||
"version": "1.4.18",
|
||||
"type": "module",
|
||||
"scripts": {
|
||||
"dev": "vite dev",
|
||||
@@ -12,16 +12,19 @@
|
||||
"devDependencies": {
|
||||
"@sveltejs/vite-plugin-svelte": "^5.0.0",
|
||||
"@tauri-apps/cli": "^2.0.0",
|
||||
"@testing-library/svelte": "^5.3.1",
|
||||
"@tsconfig/svelte": "^5.0.0",
|
||||
"jsdom": "^29.0.2",
|
||||
"svelte": "^5.0.0",
|
||||
"svelte-check": "^4.0.0",
|
||||
"typescript": "~5.6.0",
|
||||
"vite": "^6.0.0"
|
||||
"vite": "^6.0.0",
|
||||
"vitest": "^4.1.3"
|
||||
},
|
||||
"dependencies": {
|
||||
"@tauri-apps/api": "^2.0.0",
|
||||
"@tauri-apps/plugin-dialog": "^2.0.0",
|
||||
"@tauri-apps/plugin-shell": "^2.0.0",
|
||||
"@tauri-apps/plugin-process": "^2.0.0"
|
||||
"@tauri-apps/plugin-process": "^2.0.0",
|
||||
"@tauri-apps/plugin-shell": "^2.0.0"
|
||||
}
|
||||
}
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
[project]
|
||||
name = "local-transcription"
|
||||
version = "1.0.2"
|
||||
version = "1.0.3"
|
||||
description = "A standalone desktop application for real-time speech-to-text transcription using Whisper models"
|
||||
readme = "README.md"
|
||||
requires-python = ">=3.9"
|
||||
|
||||
498
src-tauri/Cargo.lock
generated
498
src-tauri/Cargo.lock
generated
@@ -47,6 +47,15 @@ version = "1.0.102"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "7f202df86484c868dbad7eaa557ef785d5c66295e41b460ef922eca0723b842c"
|
||||
|
||||
[[package]]
|
||||
name = "arbitrary"
|
||||
version = "1.4.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "c3d036a3c4ab069c7b410a2ce876bd74808d2d0888a82667669f8e783a898bf1"
|
||||
dependencies = [
|
||||
"derive_arbitrary",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "atk"
|
||||
version = "0.18.2"
|
||||
@@ -307,8 +316,10 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "c673075a2e0e5f4a1dde27ce9dee1ea4558c7ffe648f576438a20ca1d2acc4b0"
|
||||
dependencies = [
|
||||
"iana-time-zone",
|
||||
"js-sys",
|
||||
"num-traits",
|
||||
"serde",
|
||||
"wasm-bindgen",
|
||||
"windows-link 0.2.1",
|
||||
]
|
||||
|
||||
@@ -338,6 +349,16 @@ dependencies = [
|
||||
"version_check",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "core-foundation"
|
||||
version = "0.9.4"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "91e195e091a93c46f7102ec7818a2aa394e1e1771c3ab4825963fa03e45afb8f"
|
||||
dependencies = [
|
||||
"core-foundation-sys",
|
||||
"libc",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "core-foundation"
|
||||
version = "0.10.1"
|
||||
@@ -361,9 +382,9 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "064badf302c3194842cf2c5d61f56cc88e54a759313879cdf03abdd27d0c3b97"
|
||||
dependencies = [
|
||||
"bitflags 2.11.0",
|
||||
"core-foundation",
|
||||
"core-foundation 0.10.1",
|
||||
"core-graphics-types",
|
||||
"foreign-types",
|
||||
"foreign-types 0.5.0",
|
||||
"libc",
|
||||
]
|
||||
|
||||
@@ -374,7 +395,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "3d44a101f213f6c4cdc1853d4b78aef6db6bdfa3468798cc1d9912f4735013eb"
|
||||
dependencies = [
|
||||
"bitflags 2.11.0",
|
||||
"core-foundation",
|
||||
"core-foundation 0.10.1",
|
||||
"libc",
|
||||
]
|
||||
|
||||
@@ -515,6 +536,17 @@ dependencies = [
|
||||
"serde_core",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "derive_arbitrary"
|
||||
version = "1.4.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "1e567bd82dcff979e4b03460c307b3cdc9e96fde3d73bed1496d2bc75d9dd62a"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.117",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "derive_more"
|
||||
version = "0.99.20"
|
||||
@@ -792,6 +824,15 @@ version = "0.2.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "77ce24cb58228fbb8aa041425bb1050850ac19177686ea6e0f41a70416f56fdb"
|
||||
|
||||
[[package]]
|
||||
name = "foreign-types"
|
||||
version = "0.3.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "f6f339eb8adc052cd2ca78910fda869aefa38d22d5cb648e6485e4d3fc06f3b1"
|
||||
dependencies = [
|
||||
"foreign-types-shared 0.1.1",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "foreign-types"
|
||||
version = "0.5.0"
|
||||
@@ -799,7 +840,7 @@ source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "d737d9aa519fb7b749cbc3b962edcf310a8dd1f4b67c91c4f83975dbdd17d965"
|
||||
dependencies = [
|
||||
"foreign-types-macros",
|
||||
"foreign-types-shared",
|
||||
"foreign-types-shared 0.3.1",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -813,6 +854,12 @@ dependencies = [
|
||||
"syn 2.0.117",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "foreign-types-shared"
|
||||
version = "0.1.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "00b0228411908ca8685dba7fc2cdd70ec9990a6e753e89b6ac91a84c40fbaf4b"
|
||||
|
||||
[[package]]
|
||||
name = "foreign-types-shared"
|
||||
version = "0.3.1"
|
||||
@@ -1222,6 +1269,25 @@ dependencies = [
|
||||
"syn 2.0.117",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "h2"
|
||||
version = "0.4.13"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "2f44da3a8150a6703ed5d34e164b875fd14c2cdab9af1252a9a1020bde2bdc54"
|
||||
dependencies = [
|
||||
"atomic-waker",
|
||||
"bytes",
|
||||
"fnv",
|
||||
"futures-core",
|
||||
"futures-sink",
|
||||
"http",
|
||||
"indexmap 2.13.1",
|
||||
"slab",
|
||||
"tokio",
|
||||
"tokio-util",
|
||||
"tracing",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "hashbrown"
|
||||
version = "0.12.3"
|
||||
@@ -1332,6 +1398,7 @@ dependencies = [
|
||||
"bytes",
|
||||
"futures-channel",
|
||||
"futures-core",
|
||||
"h2",
|
||||
"http",
|
||||
"http-body",
|
||||
"httparse",
|
||||
@@ -1342,6 +1409,38 @@ dependencies = [
|
||||
"want",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "hyper-rustls"
|
||||
version = "0.27.7"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "e3c93eb611681b207e1fe55d5a71ecf91572ec8a6705cdb6857f7d8d5242cf58"
|
||||
dependencies = [
|
||||
"http",
|
||||
"hyper",
|
||||
"hyper-util",
|
||||
"rustls",
|
||||
"rustls-pki-types",
|
||||
"tokio",
|
||||
"tokio-rustls",
|
||||
"tower-service",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "hyper-tls"
|
||||
version = "0.6.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "70206fc6890eaca9fde8a0bf71caa2ddfc9fe045ac9e5c70df101a7dbde866e0"
|
||||
dependencies = [
|
||||
"bytes",
|
||||
"http-body-util",
|
||||
"hyper",
|
||||
"hyper-util",
|
||||
"native-tls",
|
||||
"tokio",
|
||||
"tokio-native-tls",
|
||||
"tower-service",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "hyper-util"
|
||||
version = "0.1.20"
|
||||
@@ -1360,9 +1459,11 @@ dependencies = [
|
||||
"percent-encoding",
|
||||
"pin-project-lite",
|
||||
"socket2",
|
||||
"system-configuration",
|
||||
"tokio",
|
||||
"tower-service",
|
||||
"tracing",
|
||||
"windows-registry",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -1766,6 +1867,12 @@ dependencies = [
|
||||
"libc",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "linux-raw-sys"
|
||||
version = "0.12.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "32a66949e030da00e8c7d4434b251670a91556f4144941d37452769c25d58a53"
|
||||
|
||||
[[package]]
|
||||
name = "litemap"
|
||||
version = "0.8.2"
|
||||
@@ -1774,8 +1881,12 @@ checksum = "92daf443525c4cce67b150400bc2316076100ce0b3686209eb8cf3c31612e6f0"
|
||||
|
||||
[[package]]
|
||||
name = "local-transcription"
|
||||
version = "1.4.0"
|
||||
version = "1.4.16"
|
||||
dependencies = [
|
||||
"bytes",
|
||||
"chrono",
|
||||
"futures-util",
|
||||
"reqwest 0.12.28",
|
||||
"serde",
|
||||
"serde_json",
|
||||
"tauri",
|
||||
@@ -1783,6 +1894,9 @@ dependencies = [
|
||||
"tauri-plugin-dialog",
|
||||
"tauri-plugin-process",
|
||||
"tauri-plugin-shell",
|
||||
"tempfile",
|
||||
"tokio",
|
||||
"zip",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
@@ -1911,6 +2025,23 @@ dependencies = [
|
||||
"windows-sys 0.60.2",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "native-tls"
|
||||
version = "0.2.18"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "465500e14ea162429d264d44189adc38b199b62b1c21eea9f69e4b73cb03bbf2"
|
||||
dependencies = [
|
||||
"libc",
|
||||
"log",
|
||||
"openssl",
|
||||
"openssl-probe",
|
||||
"openssl-sys",
|
||||
"schannel",
|
||||
"security-framework",
|
||||
"security-framework-sys",
|
||||
"tempfile",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "ndk"
|
||||
version = "0.9.0"
|
||||
@@ -2132,6 +2263,50 @@ dependencies = [
|
||||
"pathdiff",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "openssl"
|
||||
version = "0.10.76"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "951c002c75e16ea2c65b8c7e4d3d51d5530d8dfa7d060b4776828c88cfb18ecf"
|
||||
dependencies = [
|
||||
"bitflags 2.11.0",
|
||||
"cfg-if",
|
||||
"foreign-types 0.3.2",
|
||||
"libc",
|
||||
"once_cell",
|
||||
"openssl-macros",
|
||||
"openssl-sys",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "openssl-macros"
|
||||
version = "0.1.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "a948666b637a0f465e8564c73e89d4dde00d72d4d473cc972f390fc3dcee7d9c"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.117",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "openssl-probe"
|
||||
version = "0.2.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "7c87def4c32ab89d880effc9e097653c8da5d6ef28e6b539d313baaacfbafcbe"
|
||||
|
||||
[[package]]
|
||||
name = "openssl-sys"
|
||||
version = "0.9.112"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "57d55af3b3e226502be1526dfdba67ab0e9c96fc293004e79576b2b9edb0dbdb"
|
||||
dependencies = [
|
||||
"cc",
|
||||
"libc",
|
||||
"pkg-config",
|
||||
"vcpkg",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "option-ext"
|
||||
version = "0.2.0"
|
||||
@@ -2727,6 +2902,49 @@ version = "0.8.10"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "dc897dd8d9e8bd1ed8cdad82b5966c3e0ecae09fb1907d58efaa013543185d0a"
|
||||
|
||||
[[package]]
|
||||
name = "reqwest"
|
||||
version = "0.12.28"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "eddd3ca559203180a307f12d114c268abf583f59b03cb906fd0b3ff8646c1147"
|
||||
dependencies = [
|
||||
"base64 0.22.1",
|
||||
"bytes",
|
||||
"encoding_rs",
|
||||
"futures-core",
|
||||
"futures-util",
|
||||
"h2",
|
||||
"http",
|
||||
"http-body",
|
||||
"http-body-util",
|
||||
"hyper",
|
||||
"hyper-rustls",
|
||||
"hyper-tls",
|
||||
"hyper-util",
|
||||
"js-sys",
|
||||
"log",
|
||||
"mime",
|
||||
"native-tls",
|
||||
"percent-encoding",
|
||||
"pin-project-lite",
|
||||
"rustls-pki-types",
|
||||
"serde",
|
||||
"serde_json",
|
||||
"serde_urlencoded",
|
||||
"sync_wrapper",
|
||||
"tokio",
|
||||
"tokio-native-tls",
|
||||
"tokio-util",
|
||||
"tower",
|
||||
"tower-http",
|
||||
"tower-service",
|
||||
"url",
|
||||
"wasm-bindgen",
|
||||
"wasm-bindgen-futures",
|
||||
"wasm-streams 0.4.2",
|
||||
"web-sys",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "reqwest"
|
||||
version = "0.13.2"
|
||||
@@ -2757,7 +2975,7 @@ dependencies = [
|
||||
"url",
|
||||
"wasm-bindgen",
|
||||
"wasm-bindgen-futures",
|
||||
"wasm-streams",
|
||||
"wasm-streams 0.5.0",
|
||||
"web-sys",
|
||||
]
|
||||
|
||||
@@ -2785,6 +3003,20 @@ dependencies = [
|
||||
"windows-sys 0.60.2",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "ring"
|
||||
version = "0.17.14"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "a4689e6c2294d81e88dc6261c768b63bc4fcdb852be6d1352498b114f61383b7"
|
||||
dependencies = [
|
||||
"cc",
|
||||
"cfg-if",
|
||||
"getrandom 0.2.17",
|
||||
"libc",
|
||||
"untrusted",
|
||||
"windows-sys 0.52.0",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "rustc-hash"
|
||||
version = "2.1.2"
|
||||
@@ -2800,12 +3032,64 @@ dependencies = [
|
||||
"semver",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "rustix"
|
||||
version = "1.1.4"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "b6fe4565b9518b83ef4f91bb47ce29620ca828bd32cb7e408f0062e9930ba190"
|
||||
dependencies = [
|
||||
"bitflags 2.11.0",
|
||||
"errno",
|
||||
"libc",
|
||||
"linux-raw-sys",
|
||||
"windows-sys 0.61.2",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "rustls"
|
||||
version = "0.23.37"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "758025cb5fccfd3bc2fd74708fd4682be41d99e5dff73c377c0646c6012c73a4"
|
||||
dependencies = [
|
||||
"once_cell",
|
||||
"rustls-pki-types",
|
||||
"rustls-webpki",
|
||||
"subtle",
|
||||
"zeroize",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "rustls-pki-types"
|
||||
version = "1.14.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "be040f8b0a225e40375822a563fa9524378b9d63112f53e19ffff34df5d33fdd"
|
||||
dependencies = [
|
||||
"zeroize",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "rustls-webpki"
|
||||
version = "0.103.10"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "df33b2b81ac578cabaf06b89b0631153a3f416b0a886e8a7a1707fb51abbd1ef"
|
||||
dependencies = [
|
||||
"ring",
|
||||
"rustls-pki-types",
|
||||
"untrusted",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "rustversion"
|
||||
version = "1.0.22"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "b39cdef0fa800fc44525c84ccb54a029961a8215f9619753635a9c0d2538d46d"
|
||||
|
||||
[[package]]
|
||||
name = "ryu"
|
||||
version = "1.0.23"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "9774ba4a74de5f7b1c1451ed6cd5285a32eddb5cccb8cc655a4e50009e06477f"
|
||||
|
||||
[[package]]
|
||||
name = "same-file"
|
||||
version = "1.0.6"
|
||||
@@ -2815,6 +3099,15 @@ dependencies = [
|
||||
"winapi-util",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "schannel"
|
||||
version = "0.1.29"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "91c1b7e4904c873ef0710c1f407dde2e6287de2bebc1bbbf7d430bb7cbffd939"
|
||||
dependencies = [
|
||||
"windows-sys 0.61.2",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "schemars"
|
||||
version = "0.8.22"
|
||||
@@ -2872,6 +3165,29 @@ version = "1.2.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "94143f37725109f92c262ed2cf5e59bce7498c01bcc1502d7b9afe439a4e9f49"
|
||||
|
||||
[[package]]
|
||||
name = "security-framework"
|
||||
version = "3.7.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "b7f4bc775c73d9a02cde8bf7b2ec4c9d12743edf609006c7facc23998404cd1d"
|
||||
dependencies = [
|
||||
"bitflags 2.11.0",
|
||||
"core-foundation 0.10.1",
|
||||
"core-foundation-sys",
|
||||
"libc",
|
||||
"security-framework-sys",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "security-framework-sys"
|
||||
version = "2.17.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "6ce2691df843ecc5d231c0b14ece2acc3efb62c0a398c7e1d875f3983ce020e3"
|
||||
dependencies = [
|
||||
"core-foundation-sys",
|
||||
"libc",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "selectors"
|
||||
version = "0.24.0"
|
||||
@@ -3014,6 +3330,18 @@ dependencies = [
|
||||
"serde_core",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "serde_urlencoded"
|
||||
version = "0.7.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "d3491c14715ca2294c4d6a88f15e84739788c1d030eed8c110436aafdaa2f3fd"
|
||||
dependencies = [
|
||||
"form_urlencoded",
|
||||
"itoa",
|
||||
"ryu",
|
||||
"serde",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "serde_with"
|
||||
version = "3.18.0"
|
||||
@@ -3294,6 +3622,12 @@ version = "0.11.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "7da8b5736845d9f2fcb837ea5d9e2628564b3b043a70948a3f0b778838c5fb4f"
|
||||
|
||||
[[package]]
|
||||
name = "subtle"
|
||||
version = "2.6.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "13c2bddecc57b384dee18652358fb23172facb8a2c51ccc10d74c157bdea3292"
|
||||
|
||||
[[package]]
|
||||
name = "swift-rs"
|
||||
version = "1.0.7"
|
||||
@@ -3347,6 +3681,27 @@ dependencies = [
|
||||
"syn 2.0.117",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "system-configuration"
|
||||
version = "0.7.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "a13f3d0daba03132c0aa9767f98351b3488edc2c100cda2d2ec2b04f3d8d3c8b"
|
||||
dependencies = [
|
||||
"bitflags 2.11.0",
|
||||
"core-foundation 0.9.4",
|
||||
"system-configuration-sys",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "system-configuration-sys"
|
||||
version = "0.6.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "8e1d1b10ced5ca923a1fcb8d03e96b8d3268065d724548c0211415ff6ac6bac4"
|
||||
dependencies = [
|
||||
"core-foundation-sys",
|
||||
"libc",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "system-deps"
|
||||
version = "6.2.2"
|
||||
@@ -3368,7 +3723,7 @@ checksum = "9103edf55f2da3c82aea4c7fab7c4241032bfeea0e71fa557d98e00e7ce7cc20"
|
||||
dependencies = [
|
||||
"bitflags 2.11.0",
|
||||
"block2",
|
||||
"core-foundation",
|
||||
"core-foundation 0.10.1",
|
||||
"core-graphics",
|
||||
"crossbeam-channel",
|
||||
"dispatch2",
|
||||
@@ -3445,7 +3800,7 @@ dependencies = [
|
||||
"percent-encoding",
|
||||
"plist",
|
||||
"raw-window-handle",
|
||||
"reqwest",
|
||||
"reqwest 0.13.2",
|
||||
"serde",
|
||||
"serde_json",
|
||||
"serde_repr",
|
||||
@@ -3719,6 +4074,19 @@ dependencies = [
|
||||
"toml 0.9.12+spec-1.1.0",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tempfile"
|
||||
version = "3.27.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "32497e9a4c7b38532efcdebeef879707aa9f794296a4f0244f6f69e9bc8574bd"
|
||||
dependencies = [
|
||||
"fastrand",
|
||||
"getrandom 0.4.2",
|
||||
"once_cell",
|
||||
"rustix",
|
||||
"windows-sys 0.61.2",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tendril"
|
||||
version = "0.4.3"
|
||||
@@ -3830,11 +4198,45 @@ dependencies = [
|
||||
"bytes",
|
||||
"libc",
|
||||
"mio",
|
||||
"parking_lot",
|
||||
"pin-project-lite",
|
||||
"signal-hook-registry",
|
||||
"socket2",
|
||||
"tokio-macros",
|
||||
"windows-sys 0.61.2",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tokio-macros"
|
||||
version = "2.7.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "385a6cb71ab9ab790c5fe8d67f1645e6c450a7ce006a33de03daa956cf70a496"
|
||||
dependencies = [
|
||||
"proc-macro2",
|
||||
"quote",
|
||||
"syn 2.0.117",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tokio-native-tls"
|
||||
version = "0.3.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "bbae76ab933c85776efabc971569dd6119c580d8f5d448769dec1764bf796ef2"
|
||||
dependencies = [
|
||||
"native-tls",
|
||||
"tokio",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tokio-rustls"
|
||||
version = "0.26.4"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "1729aa945f29d91ba541258c8df89027d5792d85a8841fb65e8bf0f4ede4ef61"
|
||||
dependencies = [
|
||||
"rustls",
|
||||
"tokio",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "tokio-util"
|
||||
version = "0.7.18"
|
||||
@@ -4116,6 +4518,12 @@ version = "0.2.6"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "ebc1c04c71510c7f702b52b7c350734c9ff1295c464a03335b00bb84fc54f853"
|
||||
|
||||
[[package]]
|
||||
name = "untrusted"
|
||||
version = "0.9.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "8ecb6da28b8a351d773b68d5825ac39017e680750f980f3a1a85cd8dd28a47c1"
|
||||
|
||||
[[package]]
|
||||
name = "url"
|
||||
version = "2.5.8"
|
||||
@@ -4165,6 +4573,12 @@ dependencies = [
|
||||
"wasm-bindgen",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "vcpkg"
|
||||
version = "0.2.15"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "accd4ea62f7bb7a82fe23066fb0957d48ef677f6eeb8215f372f52e48bb32426"
|
||||
|
||||
[[package]]
|
||||
name = "version-compare"
|
||||
version = "0.2.1"
|
||||
@@ -4323,6 +4737,19 @@ dependencies = [
|
||||
"wasmparser",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "wasm-streams"
|
||||
version = "0.4.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "15053d8d85c7eccdbefef60f06769760a563c7f0a9d6902a13d35c7800b0ad65"
|
||||
dependencies = [
|
||||
"futures-util",
|
||||
"js-sys",
|
||||
"wasm-bindgen",
|
||||
"wasm-bindgen-futures",
|
||||
"web-sys",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "wasm-streams"
|
||||
version = "0.5.0"
|
||||
@@ -4599,6 +5026,17 @@ dependencies = [
|
||||
"windows-link 0.1.3",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "windows-registry"
|
||||
version = "0.6.1"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "02752bf7fbdcce7f2a27a742f798510f3e5ad88dbe84871e5168e2120c3d5720"
|
||||
dependencies = [
|
||||
"windows-link 0.2.1",
|
||||
"windows-result 0.4.1",
|
||||
"windows-strings 0.5.1",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "windows-result"
|
||||
version = "0.3.4"
|
||||
@@ -4644,6 +5082,15 @@ dependencies = [
|
||||
"windows-targets 0.42.2",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "windows-sys"
|
||||
version = "0.52.0"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "282be5f36a8ce781fad8c8ae18fa3f9beff57ec1b52cb3de0789201425d9a33d"
|
||||
dependencies = [
|
||||
"windows-targets 0.52.6",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "windows-sys"
|
||||
version = "0.59.0"
|
||||
@@ -5132,6 +5579,12 @@ dependencies = [
|
||||
"synstructure",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "zeroize"
|
||||
version = "1.8.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "b97154e67e32c85465826e8bcc1c59429aaaf107c1e4a9e53c8d8ccd5eff88d0"
|
||||
|
||||
[[package]]
|
||||
name = "zerotrie"
|
||||
version = "0.2.4"
|
||||
@@ -5165,8 +5618,37 @@ dependencies = [
|
||||
"syn 2.0.117",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "zip"
|
||||
version = "2.4.2"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "fabe6324e908f85a1c52063ce7aa26b68dcb7eb6dbc83a2d148403c9bc3eba50"
|
||||
dependencies = [
|
||||
"arbitrary",
|
||||
"crc32fast",
|
||||
"crossbeam-utils",
|
||||
"displaydoc",
|
||||
"flate2",
|
||||
"indexmap 2.13.1",
|
||||
"memchr",
|
||||
"thiserror 2.0.18",
|
||||
"zopfli",
|
||||
]
|
||||
|
||||
[[package]]
|
||||
name = "zmij"
|
||||
version = "1.0.21"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "b8848ee67ecc8aedbaf3e4122217aff892639231befc6a1b58d29fff4c2cabaa"
|
||||
|
||||
[[package]]
|
||||
name = "zopfli"
|
||||
version = "0.8.3"
|
||||
source = "registry+https://github.com/rust-lang/crates.io-index"
|
||||
checksum = "f05cd8797d63865425ff89b5c4a48804f35ba0ce8d125800027ad6017d2b5249"
|
||||
dependencies = [
|
||||
"bumpalo",
|
||||
"crc32fast",
|
||||
"log",
|
||||
"simd-adler32",
|
||||
]
|
||||
|
||||
@@ -1,6 +1,6 @@
|
||||
[package]
|
||||
name = "local-transcription"
|
||||
version = "1.4.3"
|
||||
version = "1.4.18"
|
||||
description = "Real-time speech-to-text transcription for streamers"
|
||||
authors = ["Local Transcription Contributors"]
|
||||
edition = "2021"
|
||||
@@ -19,3 +19,12 @@ tauri-plugin-dialog = "2"
|
||||
tauri-plugin-process = "2"
|
||||
serde = { version = "1", features = ["derive"] }
|
||||
serde_json = "1"
|
||||
reqwest = { version = "0.12", features = ["json", "stream"] }
|
||||
futures-util = "0.3"
|
||||
zip = { version = "2", default-features = false, features = ["deflate"] }
|
||||
bytes = "1"
|
||||
tokio = { version = "1", features = ["full"] }
|
||||
chrono = "0.4"
|
||||
|
||||
[dev-dependencies]
|
||||
tempfile = "3"
|
||||
|
||||
14
src-tauri/capabilities/default.json
Normal file
14
src-tauri/capabilities/default.json
Normal file
@@ -0,0 +1,14 @@
|
||||
{
|
||||
"identifier": "default",
|
||||
"description": "Default permissions for the main window",
|
||||
"windows": ["main"],
|
||||
"permissions": [
|
||||
"core:default",
|
||||
"core:event:default",
|
||||
"core:event:allow-listen",
|
||||
"core:event:allow-emit",
|
||||
"shell:default",
|
||||
"dialog:default",
|
||||
"process:default"
|
||||
]
|
||||
}
|
||||
@@ -1,9 +1,75 @@
|
||||
mod sidecar;
|
||||
|
||||
use std::sync::Mutex;
|
||||
use tauri::Manager;
|
||||
|
||||
/// App log directory, set during setup.
|
||||
static LOG_DIR: std::sync::OnceLock<std::path::PathBuf> = std::sync::OnceLock::new();
|
||||
|
||||
/// Write a log message to the app's log file (for debugging).
|
||||
#[tauri::command]
|
||||
fn write_log(message: String) {
|
||||
if let Some(log_dir) = LOG_DIR.get() {
|
||||
let log_path = log_dir.join("frontend.log");
|
||||
use std::io::Write;
|
||||
if let Ok(mut f) = std::fs::OpenOptions::new()
|
||||
.create(true)
|
||||
.append(true)
|
||||
.open(&log_path)
|
||||
{
|
||||
let _ = writeln!(f, "[{}] {}", chrono::Local::now().format("%H:%M:%S%.3f"), message);
|
||||
}
|
||||
}
|
||||
eprintln!("[frontend] {}", message);
|
||||
}
|
||||
|
||||
#[cfg_attr(mobile, tauri::mobile_entry_point)]
|
||||
pub fn run() {
|
||||
tauri::Builder::default()
|
||||
.plugin(tauri_plugin_shell::init())
|
||||
.plugin(tauri_plugin_dialog::init())
|
||||
.plugin(tauri_plugin_process::init())
|
||||
.manage(sidecar::ManagedSidecar(Mutex::new(
|
||||
sidecar::SidecarManager::new(),
|
||||
)))
|
||||
.setup(|app| {
|
||||
let resource_dir = app
|
||||
.path()
|
||||
.resource_dir()
|
||||
.expect("failed to resolve resource dir");
|
||||
let data_dir = app
|
||||
.path()
|
||||
.app_data_dir()
|
||||
.expect("failed to resolve app data dir");
|
||||
|
||||
std::fs::create_dir_all(&data_dir).expect("failed to create app data dir");
|
||||
|
||||
// Set up logging
|
||||
LOG_DIR.set(data_dir.clone()).ok();
|
||||
let log_path = data_dir.join("app.log");
|
||||
if let Ok(mut f) = std::fs::OpenOptions::new()
|
||||
.create(true)
|
||||
.append(true)
|
||||
.open(&log_path)
|
||||
{
|
||||
use std::io::Write;
|
||||
let _ = writeln!(f, "\n=== App started at {} ===", chrono::Local::now());
|
||||
let _ = writeln!(f, "Resource dir: {}", resource_dir.display());
|
||||
let _ = writeln!(f, "Data dir: {}", data_dir.display());
|
||||
}
|
||||
|
||||
sidecar::init_dirs(resource_dir, data_dir);
|
||||
Ok(())
|
||||
})
|
||||
.invoke_handler(tauri::generate_handler![
|
||||
sidecar::check_sidecar,
|
||||
sidecar::download_sidecar,
|
||||
sidecar::check_sidecar_update,
|
||||
sidecar::get_sidecar_port,
|
||||
sidecar::start_sidecar,
|
||||
sidecar::stop_sidecar,
|
||||
write_log,
|
||||
])
|
||||
.run(tauri::generate_context!())
|
||||
.expect("error while running tauri application");
|
||||
}
|
||||
|
||||
941
src-tauri/src/sidecar/mod.rs
Normal file
941
src-tauri/src/sidecar/mod.rs
Normal file
@@ -0,0 +1,941 @@
|
||||
use std::io::BufRead;
|
||||
use std::path::PathBuf;
|
||||
use std::sync::Mutex;
|
||||
|
||||
use serde::{Deserialize, Serialize};
|
||||
use tauri::{AppHandle, Emitter};
|
||||
|
||||
const REPO_API: &str =
|
||||
"https://repo.anhonesthost.net/api/v1/repos/streamer-tools/local-transcription";
|
||||
|
||||
const BINARY_NAME: &str = if cfg!(windows) {
|
||||
"local-transcription-backend.exe"
|
||||
} else {
|
||||
"local-transcription-backend"
|
||||
};
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// Directory state (initialised once during Tauri setup)
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
static DIRS: std::sync::OnceLock<SidecarDirs> = std::sync::OnceLock::new();
|
||||
|
||||
struct SidecarDirs {
|
||||
#[allow(dead_code)]
|
||||
resource_dir: PathBuf,
|
||||
data_dir: PathBuf,
|
||||
}
|
||||
|
||||
/// Called from Tauri `setup` to persist the resource / data directories.
|
||||
pub fn init_dirs(resource_dir: PathBuf, data_dir: PathBuf) {
|
||||
let _ = DIRS.set(SidecarDirs {
|
||||
resource_dir,
|
||||
data_dir,
|
||||
});
|
||||
}
|
||||
|
||||
fn data_dir() -> &'static PathBuf {
|
||||
&DIRS.get().expect("sidecar::init_dirs not called").data_dir
|
||||
}
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// Version helpers
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
fn version_file() -> PathBuf {
|
||||
data_dir().join("sidecar-version.txt")
|
||||
}
|
||||
|
||||
fn read_installed_version() -> Option<String> {
|
||||
std::fs::read_to_string(version_file())
|
||||
.ok()
|
||||
.map(|s| s.trim().to_string())
|
||||
.filter(|s| !s.is_empty())
|
||||
}
|
||||
|
||||
fn sidecar_dir_for_version(version: &str) -> PathBuf {
|
||||
data_dir().join(format!("sidecar-{version}"))
|
||||
}
|
||||
|
||||
fn binary_path_for_version(version: &str) -> PathBuf {
|
||||
sidecar_dir_for_version(version).join(BINARY_NAME)
|
||||
}
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// Gitea API types
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
#[derive(Debug, Deserialize)]
|
||||
struct GiteaRelease {
|
||||
tag_name: String,
|
||||
assets: Vec<GiteaAsset>,
|
||||
}
|
||||
|
||||
#[derive(Debug, Deserialize)]
|
||||
struct GiteaAsset {
|
||||
name: String,
|
||||
browser_download_url: String,
|
||||
size: u64,
|
||||
}
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// Platform / arch detection
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
fn platform_token() -> &'static str {
|
||||
if cfg!(target_os = "windows") {
|
||||
"windows"
|
||||
} else if cfg!(target_os = "macos") {
|
||||
"macos"
|
||||
} else {
|
||||
"linux"
|
||||
}
|
||||
}
|
||||
|
||||
fn arch_token() -> &'static str {
|
||||
if cfg!(target_arch = "aarch64") {
|
||||
"aarch64"
|
||||
} else {
|
||||
"x86_64"
|
||||
}
|
||||
}
|
||||
|
||||
/// Build the expected asset prefix, e.g. `sidecar-linux-x86_64-cuda`.
|
||||
fn asset_prefix(variant: &str) -> String {
|
||||
format!("sidecar-{}-{}-{}", platform_token(), arch_token(), variant)
|
||||
}
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// Tauri commands
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
/// Returns `true` when a sidecar binary is installed and the file exists.
|
||||
#[tauri::command]
|
||||
pub fn check_sidecar() -> bool {
|
||||
if let Some(version) = read_installed_version() {
|
||||
binary_path_for_version(&version).exists()
|
||||
} else {
|
||||
false
|
||||
}
|
||||
}
|
||||
|
||||
/// Download progress payload emitted via `sidecar-download-progress`.
|
||||
#[derive(Clone, Serialize)]
|
||||
struct DownloadProgress {
|
||||
downloaded: u64,
|
||||
total: u64,
|
||||
phase: String, // "downloading" | "extracting" | "done" | "error"
|
||||
message: String,
|
||||
}
|
||||
|
||||
/// Download & install the latest sidecar release.
|
||||
///
|
||||
/// `variant` is typically `"cuda"` or `"cpu"`.
|
||||
#[tauri::command]
|
||||
pub async fn download_sidecar(app: AppHandle, variant: String) -> Result<String, String> {
|
||||
use futures_util::StreamExt;
|
||||
|
||||
let emit = |progress: DownloadProgress| {
|
||||
let _ = app.emit("sidecar-download-progress", progress);
|
||||
};
|
||||
|
||||
// 1. Fetch releases from Gitea (filter to sidecar-v* tags) ---------------
|
||||
emit(DownloadProgress {
|
||||
downloaded: 0,
|
||||
total: 0,
|
||||
phase: "downloading".into(),
|
||||
message: "Fetching release info...".into(),
|
||||
});
|
||||
|
||||
let releases_url = format!("{REPO_API}/releases?limit=20");
|
||||
let client = reqwest::Client::new();
|
||||
let releases: Vec<GiteaRelease> = client
|
||||
.get(&releases_url)
|
||||
.send()
|
||||
.await
|
||||
.map_err(|e| format!("Failed to fetch releases: {e}"))?
|
||||
.json()
|
||||
.await
|
||||
.map_err(|e| format!("Failed to parse releases: {e}"))?;
|
||||
|
||||
// Find the latest release whose tag starts with `sidecar-v`
|
||||
let release = releases
|
||||
.into_iter()
|
||||
.find(|r| r.tag_name.starts_with("sidecar-v"))
|
||||
.ok_or_else(|| "No sidecar release found".to_string())?;
|
||||
|
||||
let version = release.tag_name.clone(); // e.g. "sidecar-v1.0.2"
|
||||
|
||||
// 2. Find matching asset ----------------------------------------------------
|
||||
let prefix = asset_prefix(&variant);
|
||||
let asset = release
|
||||
.assets
|
||||
.iter()
|
||||
.find(|a| a.name.starts_with(&prefix) && a.name.ends_with(".zip"))
|
||||
.ok_or_else(|| {
|
||||
format!(
|
||||
"No asset matching '{}' in release {}. Available: {}",
|
||||
prefix,
|
||||
version,
|
||||
release
|
||||
.assets
|
||||
.iter()
|
||||
.map(|a| a.name.as_str())
|
||||
.collect::<Vec<_>>()
|
||||
.join(", ")
|
||||
)
|
||||
})?;
|
||||
|
||||
let total_size = asset.size;
|
||||
let download_url = asset.browser_download_url.clone();
|
||||
|
||||
// 3. Stream download ---------------------------------------------------------
|
||||
emit(DownloadProgress {
|
||||
downloaded: 0,
|
||||
total: total_size,
|
||||
phase: "downloading".into(),
|
||||
message: format!("Downloading {}...", asset.name),
|
||||
});
|
||||
|
||||
let response = client
|
||||
.get(&download_url)
|
||||
.send()
|
||||
.await
|
||||
.map_err(|e| format!("Download request failed: {e}"))?;
|
||||
|
||||
if !response.status().is_success() {
|
||||
return Err(format!("Download failed with status {}", response.status()));
|
||||
}
|
||||
|
||||
let tmp_zip = data_dir().join("_sidecar_download.zip");
|
||||
let mut file = tokio::fs::File::create(&tmp_zip)
|
||||
.await
|
||||
.map_err(|e| format!("Cannot create temp file: {e}"))?;
|
||||
|
||||
let mut stream = response.bytes_stream();
|
||||
let mut downloaded: u64 = 0;
|
||||
|
||||
use tokio::io::AsyncWriteExt;
|
||||
while let Some(chunk) = stream.next().await {
|
||||
let chunk = chunk.map_err(|e| format!("Download stream error: {e}"))?;
|
||||
file.write_all(&chunk)
|
||||
.await
|
||||
.map_err(|e| format!("Write error: {e}"))?;
|
||||
downloaded += chunk.len() as u64;
|
||||
|
||||
emit(DownloadProgress {
|
||||
downloaded,
|
||||
total: total_size,
|
||||
phase: "downloading".into(),
|
||||
message: format!(
|
||||
"Downloading... {:.1} / {:.1} MB",
|
||||
downloaded as f64 / 1_048_576.0,
|
||||
total_size as f64 / 1_048_576.0
|
||||
),
|
||||
});
|
||||
}
|
||||
|
||||
file.flush()
|
||||
.await
|
||||
.map_err(|e| format!("Flush error: {e}"))?;
|
||||
drop(file);
|
||||
|
||||
// 4. Extract zip -------------------------------------------------------------
|
||||
emit(DownloadProgress {
|
||||
downloaded,
|
||||
total: total_size,
|
||||
phase: "extracting".into(),
|
||||
message: "Extracting sidecar...".into(),
|
||||
});
|
||||
|
||||
let dest_dir = sidecar_dir_for_version(&version);
|
||||
if dest_dir.exists() {
|
||||
std::fs::remove_dir_all(&dest_dir)
|
||||
.map_err(|e| format!("Cannot clean old dir: {e}"))?;
|
||||
}
|
||||
std::fs::create_dir_all(&dest_dir)
|
||||
.map_err(|e| format!("Cannot create sidecar dir: {e}"))?;
|
||||
|
||||
// Extraction is blocking I/O -- offload to a spawn_blocking thread.
|
||||
let zip_path = tmp_zip.clone();
|
||||
let dest = dest_dir.clone();
|
||||
tokio::task::spawn_blocking(move || extract_zip(&zip_path, &dest))
|
||||
.await
|
||||
.map_err(|e| format!("Join error: {e}"))?
|
||||
.map_err(|e| format!("Extraction error: {e}"))?;
|
||||
|
||||
// Remove the temp zip
|
||||
let _ = std::fs::remove_file(&tmp_zip);
|
||||
|
||||
// 5. Set executable permissions on Unix -------------------------------------
|
||||
#[cfg(unix)]
|
||||
{
|
||||
use std::os::unix::fs::PermissionsExt;
|
||||
let bin = dest_dir.join(BINARY_NAME);
|
||||
if bin.exists() {
|
||||
let mut perms = std::fs::metadata(&bin)
|
||||
.map_err(|e| format!("metadata error: {e}"))?
|
||||
.permissions();
|
||||
perms.set_mode(0o755);
|
||||
std::fs::set_permissions(&bin, perms)
|
||||
.map_err(|e| format!("chmod error: {e}"))?;
|
||||
}
|
||||
}
|
||||
|
||||
// 6. Write version file & clean up old versions ----------------------------
|
||||
std::fs::write(version_file(), &version)
|
||||
.map_err(|e| format!("Failed to write version file: {e}"))?;
|
||||
|
||||
cleanup_old_versions(&version);
|
||||
|
||||
emit(DownloadProgress {
|
||||
downloaded,
|
||||
total: total_size,
|
||||
phase: "done".into(),
|
||||
message: "Sidecar installed successfully".into(),
|
||||
});
|
||||
|
||||
Ok(version)
|
||||
}
|
||||
|
||||
/// Check if there is a newer sidecar release than the installed one.
|
||||
/// Returns `Some(tag_name)` when an update is available, or `None`.
|
||||
#[tauri::command]
|
||||
pub async fn check_sidecar_update() -> Result<Option<String>, String> {
|
||||
let installed = match read_installed_version() {
|
||||
Some(v) => v,
|
||||
None => return Ok(None),
|
||||
};
|
||||
|
||||
let releases_url = format!("{REPO_API}/releases?limit=20");
|
||||
let releases: Vec<GiteaRelease> = reqwest::Client::new()
|
||||
.get(&releases_url)
|
||||
.send()
|
||||
.await
|
||||
.map_err(|e| format!("Failed to fetch releases: {e}"))?
|
||||
.json()
|
||||
.await
|
||||
.map_err(|e| format!("Failed to parse releases: {e}"))?;
|
||||
|
||||
let latest = releases
|
||||
.iter()
|
||||
.find(|r| r.tag_name.starts_with("sidecar-v"));
|
||||
|
||||
match latest {
|
||||
Some(rel) if rel.tag_name != installed => Ok(Some(rel.tag_name.clone())),
|
||||
_ => Ok(None),
|
||||
}
|
||||
}
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// Zip extraction helper
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
fn extract_zip(zip_path: &std::path::Path, dest: &std::path::Path) -> Result<(), String> {
|
||||
let file =
|
||||
std::fs::File::open(zip_path).map_err(|e| format!("Cannot open zip: {e}"))?;
|
||||
let mut archive =
|
||||
zip::ZipArchive::new(file).map_err(|e| format!("Invalid zip: {e}"))?;
|
||||
|
||||
for i in 0..archive.len() {
|
||||
let mut entry = archive
|
||||
.by_index(i)
|
||||
.map_err(|e| format!("Zip entry error: {e}"))?;
|
||||
let entry_path = match entry.enclosed_name() {
|
||||
Some(p) => p.to_owned(),
|
||||
None => continue,
|
||||
};
|
||||
|
||||
let out_path = dest.join(&entry_path);
|
||||
|
||||
if entry.is_dir() {
|
||||
std::fs::create_dir_all(&out_path)
|
||||
.map_err(|e| format!("mkdir error: {e}"))?;
|
||||
} else {
|
||||
if let Some(parent) = out_path.parent() {
|
||||
std::fs::create_dir_all(parent)
|
||||
.map_err(|e| format!("mkdir error: {e}"))?;
|
||||
}
|
||||
let mut outfile = std::fs::File::create(&out_path)
|
||||
.map_err(|e| format!("create file error: {e}"))?;
|
||||
std::io::copy(&mut entry, &mut outfile)
|
||||
.map_err(|e| format!("copy error: {e}"))?;
|
||||
}
|
||||
}
|
||||
Ok(())
|
||||
}
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// Cleanup old versions
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
fn cleanup_old_versions(current_version: &str) {
|
||||
let data = data_dir();
|
||||
let current_dir_name = format!("sidecar-{current_version}");
|
||||
if let Ok(entries) = std::fs::read_dir(data) {
|
||||
for entry in entries.flatten() {
|
||||
let name = entry.file_name().to_string_lossy().to_string();
|
||||
if name.starts_with("sidecar-v") // e.g. sidecar-v1.0.1
|
||||
&& name != current_dir_name
|
||||
&& entry.path().is_dir()
|
||||
{
|
||||
let _ = std::fs::remove_dir_all(entry.path());
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// SidecarManager — launch / stop / query the backend process
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
#[derive(Debug, Serialize, Deserialize)]
|
||||
struct ReadyEvent {
|
||||
event: String,
|
||||
port: u16,
|
||||
}
|
||||
|
||||
pub struct SidecarManager {
|
||||
child: Option<std::process::Child>,
|
||||
port: Option<u16>,
|
||||
}
|
||||
|
||||
impl SidecarManager {
|
||||
pub fn new() -> Self {
|
||||
Self {
|
||||
child: None,
|
||||
port: None,
|
||||
}
|
||||
}
|
||||
|
||||
/// Returns `true` when the child process is still alive.
|
||||
pub fn is_running(&mut self) -> bool {
|
||||
match &mut self.child {
|
||||
Some(child) => match child.try_wait() {
|
||||
Ok(Some(_)) => {
|
||||
// Process has exited
|
||||
self.child = None;
|
||||
self.port = None;
|
||||
false
|
||||
}
|
||||
Ok(None) => true,
|
||||
Err(_) => false,
|
||||
},
|
||||
None => false,
|
||||
}
|
||||
}
|
||||
|
||||
/// Start the sidecar if it is not already running. Returns the port.
|
||||
pub fn ensure_running(&mut self) -> Result<u16, String> {
|
||||
if self.is_running() {
|
||||
return self
|
||||
.port
|
||||
.ok_or_else(|| "Sidecar running but port unknown".into());
|
||||
}
|
||||
|
||||
let is_dev = cfg!(debug_assertions)
|
||||
|| std::env::var("LOCAL_TRANSCRIPTION_DEV")
|
||||
.map(|v| v == "1")
|
||||
.unwrap_or(false);
|
||||
|
||||
let mut cmd = if is_dev {
|
||||
self.build_dev_command()?
|
||||
} else {
|
||||
self.build_prod_command()?
|
||||
};
|
||||
|
||||
// Hide the console window on Windows in release mode.
|
||||
#[cfg(windows)]
|
||||
{
|
||||
use std::os::windows::process::CommandExt;
|
||||
const CREATE_NO_WINDOW: u32 = 0x08000000;
|
||||
cmd.creation_flags(CREATE_NO_WINDOW);
|
||||
}
|
||||
|
||||
cmd.stdout(std::process::Stdio::piped());
|
||||
cmd.stderr(std::process::Stdio::piped());
|
||||
|
||||
let mut child = cmd.spawn().map_err(|e| format!("Failed to spawn sidecar: {e}"))?;
|
||||
|
||||
// Wait for the `{"event":"ready","port":...}` line on stdout.
|
||||
let stdout = child
|
||||
.stdout
|
||||
.take()
|
||||
.ok_or("Failed to capture sidecar stdout")?;
|
||||
|
||||
let port = Self::wait_for_ready(stdout)?;
|
||||
|
||||
self.child = Some(child);
|
||||
self.port = Some(port);
|
||||
Ok(port)
|
||||
}
|
||||
|
||||
/// Stop the sidecar process if running.
|
||||
pub fn stop(&mut self) {
|
||||
if let Some(mut child) = self.child.take() {
|
||||
let _ = child.kill();
|
||||
let _ = child.wait();
|
||||
}
|
||||
self.port = None;
|
||||
}
|
||||
|
||||
/// Return the port the sidecar is listening on, if known.
|
||||
pub fn port(&self) -> Option<u16> {
|
||||
self.port
|
||||
}
|
||||
|
||||
// -- private helpers -------------------------------------------------------
|
||||
|
||||
fn build_dev_command(&self) -> Result<std::process::Command, String> {
|
||||
let mut cmd = std::process::Command::new("python");
|
||||
cmd.args(["-m", "backend.main_headless"]);
|
||||
|
||||
// Try to find the project root (parent of src-tauri)
|
||||
if let Some(dirs) = DIRS.get() {
|
||||
let project_root = dirs
|
||||
.resource_dir
|
||||
.parent() // src-tauri
|
||||
.and_then(|p| p.parent()); // project root
|
||||
if let Some(root) = project_root {
|
||||
cmd.current_dir(root);
|
||||
}
|
||||
}
|
||||
|
||||
Ok(cmd)
|
||||
}
|
||||
|
||||
fn build_prod_command(&self) -> Result<std::process::Command, String> {
|
||||
let version = read_installed_version()
|
||||
.ok_or("No sidecar version installed")?;
|
||||
let bin = binary_path_for_version(&version);
|
||||
if !bin.exists() {
|
||||
return Err(format!("Sidecar binary not found at {}", bin.display()));
|
||||
}
|
||||
let mut cmd = std::process::Command::new(&bin);
|
||||
cmd.current_dir(
|
||||
bin.parent()
|
||||
.ok_or("Cannot determine sidecar parent dir")?,
|
||||
);
|
||||
Ok(cmd)
|
||||
}
|
||||
|
||||
fn wait_for_ready(stdout: std::process::ChildStdout) -> Result<u16, String> {
|
||||
let reader = std::io::BufReader::new(stdout);
|
||||
let timeout = std::time::Duration::from_secs(120);
|
||||
let start = std::time::Instant::now();
|
||||
|
||||
for line in reader.lines() {
|
||||
if start.elapsed() > timeout {
|
||||
return Err("Timed out waiting for sidecar ready event".into());
|
||||
}
|
||||
let line = line.map_err(|e| format!("IO error reading stdout: {e}"))?;
|
||||
if let Ok(evt) = serde_json::from_str::<ReadyEvent>(&line) {
|
||||
if evt.event == "ready" {
|
||||
return Ok(evt.port);
|
||||
}
|
||||
}
|
||||
// Ignore other lines (e.g. log output)
|
||||
}
|
||||
Err("Sidecar process exited before sending ready event".into())
|
||||
}
|
||||
}
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// Tauri-managed SidecarManager state & commands
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
/// Wrapper so we can store `SidecarManager` in Tauri's managed state.
|
||||
pub struct ManagedSidecar(pub Mutex<SidecarManager>);
|
||||
|
||||
#[tauri::command]
|
||||
pub fn get_sidecar_port(state: tauri::State<'_, ManagedSidecar>) -> Result<Option<u16>, String> {
|
||||
let mut mgr = state
|
||||
.0
|
||||
.lock()
|
||||
.map_err(|e| format!("Lock error: {e}"))?;
|
||||
// Refresh running status before returning port
|
||||
if !mgr.is_running() {
|
||||
return Ok(None);
|
||||
}
|
||||
Ok(mgr.port())
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn start_sidecar(state: tauri::State<'_, ManagedSidecar>) -> Result<u16, String> {
|
||||
let mut mgr = state
|
||||
.0
|
||||
.lock()
|
||||
.map_err(|e| format!("Lock error: {e}"))?;
|
||||
mgr.ensure_running()
|
||||
}
|
||||
|
||||
#[tauri::command]
|
||||
pub fn stop_sidecar(state: tauri::State<'_, ManagedSidecar>) -> Result<(), String> {
|
||||
let mut mgr = state
|
||||
.0
|
||||
.lock()
|
||||
.map_err(|e| format!("Lock error: {e}"))?;
|
||||
mgr.stop();
|
||||
Ok(())
|
||||
}
|
||||
|
||||
// ---------------------------------------------------------------------------
|
||||
// Tests
|
||||
// ---------------------------------------------------------------------------
|
||||
|
||||
#[cfg(test)]
|
||||
mod tests {
|
||||
use super::*;
|
||||
use std::io::Write;
|
||||
|
||||
// -----------------------------------------------------------------------
|
||||
// 1. Platform / arch detection
|
||||
// -----------------------------------------------------------------------
|
||||
|
||||
#[test]
|
||||
fn platform_token_returns_valid_value() {
|
||||
let token = platform_token();
|
||||
assert!(
|
||||
["windows", "macos", "linux"].contains(&token),
|
||||
"unexpected platform token: {token}"
|
||||
);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn arch_token_returns_valid_value() {
|
||||
let token = arch_token();
|
||||
assert!(
|
||||
["x86_64", "aarch64"].contains(&token),
|
||||
"unexpected arch token: {token}"
|
||||
);
|
||||
}
|
||||
|
||||
// -----------------------------------------------------------------------
|
||||
// 2. Asset name construction
|
||||
// -----------------------------------------------------------------------
|
||||
|
||||
#[test]
|
||||
fn asset_prefix_cpu() {
|
||||
let prefix = asset_prefix("cpu");
|
||||
let expected = format!("sidecar-{}-{}-cpu", platform_token(), arch_token());
|
||||
assert_eq!(prefix, expected);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn asset_prefix_cuda() {
|
||||
let prefix = asset_prefix("cuda");
|
||||
let expected = format!("sidecar-{}-{}-cuda", platform_token(), arch_token());
|
||||
assert_eq!(prefix, expected);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn asset_prefix_format_matches_zip_convention() {
|
||||
// The download function looks for assets matching
|
||||
// `{prefix}*.zip`, so verify the prefix starts with "sidecar-"
|
||||
// and contains exactly three hyphens (sidecar-OS-ARCH-VARIANT).
|
||||
let prefix = asset_prefix("cpu");
|
||||
assert!(prefix.starts_with("sidecar-"));
|
||||
assert_eq!(prefix.matches('-').count(), 3, "expected 3 hyphens in '{prefix}'");
|
||||
}
|
||||
|
||||
// -----------------------------------------------------------------------
|
||||
// 3. Version parsing — tag_name format "sidecar-vX.Y.Z"
|
||||
// -----------------------------------------------------------------------
|
||||
|
||||
#[test]
|
||||
fn sidecar_tag_starts_with_expected_prefix() {
|
||||
// The code filters releases by `tag_name.starts_with("sidecar-v")`.
|
||||
// Verify the convention: a version string like "sidecar-v1.0.2" passes
|
||||
// the filter, while "v1.0.2" does not.
|
||||
let tag = "sidecar-v1.0.2";
|
||||
assert!(tag.starts_with("sidecar-v"));
|
||||
|
||||
let bad_tag = "v1.0.2";
|
||||
assert!(!bad_tag.starts_with("sidecar-v"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn strip_sidecar_v_prefix() {
|
||||
// The codebase stores the full tag as the version (e.g. "sidecar-v1.0.2").
|
||||
// Verify we can strip the prefix to get just "1.0.2" when needed.
|
||||
let tag = "sidecar-v1.0.2";
|
||||
let semver = tag.strip_prefix("sidecar-v").unwrap();
|
||||
assert_eq!(semver, "1.0.2");
|
||||
}
|
||||
|
||||
// -----------------------------------------------------------------------
|
||||
// 4. ReadyEvent deserialization
|
||||
// -----------------------------------------------------------------------
|
||||
|
||||
#[test]
|
||||
fn ready_event_deserializes_basic() {
|
||||
let json = r#"{"event": "ready", "port": 8081}"#;
|
||||
let evt: ReadyEvent = serde_json::from_str(json).unwrap();
|
||||
assert_eq!(evt.event, "ready");
|
||||
assert_eq!(evt.port, 8081);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn ready_event_deserializes_with_extra_fields() {
|
||||
// The backend may emit additional fields like `obs_port`.
|
||||
// serde should ignore unknown fields by default (deny_unknown_fields
|
||||
// is NOT set on ReadyEvent).
|
||||
let json = r#"{"event": "ready", "port": 8081, "obs_port": 8080}"#;
|
||||
let evt: ReadyEvent = serde_json::from_str(json).unwrap();
|
||||
assert_eq!(evt.event, "ready");
|
||||
assert_eq!(evt.port, 8081);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn ready_event_rejects_missing_port() {
|
||||
let json = r#"{"event": "ready"}"#;
|
||||
assert!(serde_json::from_str::<ReadyEvent>(json).is_err());
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn ready_event_rejects_invalid_port_type() {
|
||||
let json = r#"{"event": "ready", "port": "not_a_number"}"#;
|
||||
assert!(serde_json::from_str::<ReadyEvent>(json).is_err());
|
||||
}
|
||||
|
||||
// -----------------------------------------------------------------------
|
||||
// Helper: initialise DIRS with a temp directory so path-related functions
|
||||
// work. Because OnceLock can only be set once per process, all tests that
|
||||
// need DIRS must coordinate. We use std::sync::Once + a global temp path.
|
||||
// -----------------------------------------------------------------------
|
||||
|
||||
static TEST_DATA_DIR: std::sync::OnceLock<PathBuf> = std::sync::OnceLock::new();
|
||||
|
||||
/// Ensure `DIRS` is initialised (idempotent within a test run).
|
||||
/// Returns the data_dir path.
|
||||
fn ensure_dirs_initialised() -> PathBuf {
|
||||
TEST_DATA_DIR
|
||||
.get_or_init(|| {
|
||||
let tmp = tempfile::tempdir().expect("create tempdir");
|
||||
let data = tmp.path().to_path_buf();
|
||||
// We intentionally leak `tmp` so the directory lives for the
|
||||
// entire test-run.
|
||||
std::mem::forget(tmp);
|
||||
let resource = data.join("resource"); // dummy
|
||||
std::fs::create_dir_all(&resource).ok();
|
||||
init_dirs(resource, data.clone());
|
||||
data
|
||||
})
|
||||
.clone()
|
||||
}
|
||||
|
||||
// -----------------------------------------------------------------------
|
||||
// 5. Path construction (requires init_dirs)
|
||||
// -----------------------------------------------------------------------
|
||||
|
||||
#[test]
|
||||
fn version_file_path_is_in_data_dir() {
|
||||
let data = ensure_dirs_initialised();
|
||||
let vf = version_file();
|
||||
assert_eq!(vf, data.join("sidecar-version.txt"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn sidecar_dir_for_version_contains_version() {
|
||||
let data = ensure_dirs_initialised();
|
||||
let dir = sidecar_dir_for_version("sidecar-v1.2.3");
|
||||
assert_eq!(dir, data.join("sidecar-sidecar-v1.2.3"));
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn binary_path_for_version_has_correct_filename() {
|
||||
let _data = ensure_dirs_initialised();
|
||||
let bin = binary_path_for_version("sidecar-v1.2.3");
|
||||
assert_eq!(bin.file_name().unwrap(), BINARY_NAME);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn read_installed_version_none_when_missing() {
|
||||
let _data = ensure_dirs_initialised();
|
||||
// The version file should not exist yet (clean temp dir).
|
||||
// If another test wrote it, this still validates the function
|
||||
// doesn't panic.
|
||||
let _ = read_installed_version(); // should not panic
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn write_then_read_installed_version() {
|
||||
let _data = ensure_dirs_initialised();
|
||||
let vf = version_file();
|
||||
std::fs::write(&vf, "sidecar-v2.0.0\n").unwrap();
|
||||
let v = read_installed_version().expect("should read version");
|
||||
assert_eq!(v, "sidecar-v2.0.0");
|
||||
}
|
||||
|
||||
// -----------------------------------------------------------------------
|
||||
// 6. Cleanup old versions
|
||||
// -----------------------------------------------------------------------
|
||||
|
||||
/// Cleanup tests are combined into one function because
|
||||
/// `cleanup_old_versions` operates on the shared `data_dir()` and
|
||||
/// tests run in parallel, so separate tests would race.
|
||||
#[test]
|
||||
fn cleanup_old_versions_behaviour() {
|
||||
let data = ensure_dirs_initialised();
|
||||
|
||||
// -- Part A: removes stale version dirs, keeps current ----------------
|
||||
let dirs_to_create = ["sidecar-v1.0.0", "sidecar-v1.0.1", "sidecar-v1.0.2"];
|
||||
for d in &dirs_to_create {
|
||||
std::fs::create_dir_all(data.join(d)).unwrap();
|
||||
}
|
||||
|
||||
// cleanup_old_versions builds `current_dir_name = "sidecar-{version}"`.
|
||||
// Passing "v1.0.2" produces "sidecar-v1.0.2" which matches our dir name.
|
||||
cleanup_old_versions("v1.0.2");
|
||||
|
||||
assert!(
|
||||
!data.join("sidecar-v1.0.0").exists(),
|
||||
"sidecar-v1.0.0 should be removed"
|
||||
);
|
||||
assert!(
|
||||
!data.join("sidecar-v1.0.1").exists(),
|
||||
"sidecar-v1.0.1 should be removed"
|
||||
);
|
||||
assert!(
|
||||
data.join("sidecar-v1.0.2").exists(),
|
||||
"sidecar-v1.0.2 should be kept"
|
||||
);
|
||||
|
||||
// -- Part B: ignores non-sidecar directories --------------------------
|
||||
let other = data.join("some-other-dir");
|
||||
std::fs::create_dir_all(&other).unwrap();
|
||||
|
||||
cleanup_old_versions("v1.0.2"); // run again — should leave other alone
|
||||
|
||||
assert!(other.exists(), "non-sidecar dir should not be removed");
|
||||
|
||||
// Clean up so we don't affect other tests that share data_dir.
|
||||
let _ = std::fs::remove_dir_all(data.join("sidecar-v1.0.2"));
|
||||
let _ = std::fs::remove_dir_all(&other);
|
||||
}
|
||||
|
||||
// -----------------------------------------------------------------------
|
||||
// 7. Zip extraction
|
||||
// -----------------------------------------------------------------------
|
||||
|
||||
#[test]
|
||||
fn extract_zip_creates_files() {
|
||||
let tmp = tempfile::tempdir().unwrap();
|
||||
let zip_path = tmp.path().join("test.zip");
|
||||
let dest_dir = tmp.path().join("output");
|
||||
std::fs::create_dir_all(&dest_dir).unwrap();
|
||||
|
||||
// Build a simple zip in memory.
|
||||
{
|
||||
let file = std::fs::File::create(&zip_path).unwrap();
|
||||
let mut writer = zip::ZipWriter::new(file);
|
||||
let options = zip::write::SimpleFileOptions::default()
|
||||
.compression_method(zip::CompressionMethod::Deflated);
|
||||
|
||||
writer.start_file("hello.txt", options).unwrap();
|
||||
writer.write_all(b"Hello, world!").unwrap();
|
||||
|
||||
writer.start_file("subdir/nested.txt", options).unwrap();
|
||||
writer.write_all(b"Nested content").unwrap();
|
||||
|
||||
writer.finish().unwrap();
|
||||
}
|
||||
|
||||
extract_zip(&zip_path, &dest_dir).expect("extraction should succeed");
|
||||
|
||||
let hello = dest_dir.join("hello.txt");
|
||||
assert!(hello.exists(), "hello.txt should exist");
|
||||
assert_eq!(std::fs::read_to_string(&hello).unwrap(), "Hello, world!");
|
||||
|
||||
let nested = dest_dir.join("subdir/nested.txt");
|
||||
assert!(nested.exists(), "subdir/nested.txt should exist");
|
||||
assert_eq!(std::fs::read_to_string(&nested).unwrap(), "Nested content");
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn extract_zip_error_on_invalid_file() {
|
||||
let tmp = tempfile::tempdir().unwrap();
|
||||
let bad_zip = tmp.path().join("bad.zip");
|
||||
std::fs::write(&bad_zip, b"not a zip file").unwrap();
|
||||
let dest = tmp.path().join("dest");
|
||||
std::fs::create_dir_all(&dest).unwrap();
|
||||
|
||||
let result = extract_zip(&bad_zip, &dest);
|
||||
assert!(result.is_err(), "should fail on invalid zip");
|
||||
}
|
||||
|
||||
// -----------------------------------------------------------------------
|
||||
// SidecarManager unit tests (no process spawning)
|
||||
// -----------------------------------------------------------------------
|
||||
|
||||
#[test]
|
||||
fn sidecar_manager_new_is_not_running() {
|
||||
let mut mgr = SidecarManager::new();
|
||||
assert!(!mgr.is_running());
|
||||
assert!(mgr.port().is_none());
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn sidecar_manager_stop_when_not_running() {
|
||||
let mut mgr = SidecarManager::new();
|
||||
mgr.stop(); // should not panic
|
||||
assert!(!mgr.is_running());
|
||||
}
|
||||
|
||||
// -----------------------------------------------------------------------
|
||||
// GiteaRelease / GiteaAsset deserialization
|
||||
// -----------------------------------------------------------------------
|
||||
|
||||
#[test]
|
||||
fn gitea_release_deserializes() {
|
||||
let json = r#"{
|
||||
"tag_name": "sidecar-v1.0.0",
|
||||
"assets": [
|
||||
{
|
||||
"name": "sidecar-linux-x86_64-cuda.zip",
|
||||
"browser_download_url": "https://example.com/file.zip",
|
||||
"size": 12345
|
||||
}
|
||||
]
|
||||
}"#;
|
||||
let release: GiteaRelease = serde_json::from_str(json).unwrap();
|
||||
assert_eq!(release.tag_name, "sidecar-v1.0.0");
|
||||
assert_eq!(release.assets.len(), 1);
|
||||
assert_eq!(release.assets[0].name, "sidecar-linux-x86_64-cuda.zip");
|
||||
assert_eq!(release.assets[0].size, 12345);
|
||||
}
|
||||
|
||||
#[test]
|
||||
fn gitea_release_with_extra_fields() {
|
||||
// Gitea responses include many more fields; serde should ignore them.
|
||||
let json = r#"{
|
||||
"id": 42,
|
||||
"tag_name": "sidecar-v2.0.0",
|
||||
"name": "Release 2.0.0",
|
||||
"body": "changelog here",
|
||||
"draft": false,
|
||||
"prerelease": false,
|
||||
"assets": []
|
||||
}"#;
|
||||
let release: GiteaRelease = serde_json::from_str(json).unwrap();
|
||||
assert_eq!(release.tag_name, "sidecar-v2.0.0");
|
||||
assert!(release.assets.is_empty());
|
||||
}
|
||||
|
||||
// -----------------------------------------------------------------------
|
||||
// DownloadProgress serialization round-trip
|
||||
// -----------------------------------------------------------------------
|
||||
|
||||
#[test]
|
||||
fn download_progress_serializes() {
|
||||
let progress = DownloadProgress {
|
||||
downloaded: 1024,
|
||||
total: 4096,
|
||||
phase: "downloading".into(),
|
||||
message: "50%".into(),
|
||||
};
|
||||
let json = serde_json::to_string(&progress).unwrap();
|
||||
assert!(json.contains("\"downloaded\":1024"));
|
||||
assert!(json.contains("\"phase\":\"downloading\""));
|
||||
}
|
||||
}
|
||||
@@ -1,6 +1,6 @@
|
||||
{
|
||||
"productName": "Local Transcription",
|
||||
"version": "1.4.3",
|
||||
"version": "1.4.18",
|
||||
"identifier": "net.anhonesthost.local-transcription",
|
||||
"build": {
|
||||
"frontendDist": "../dist",
|
||||
|
||||
222
src/App.svelte
222
src/App.svelte
@@ -5,13 +5,20 @@
|
||||
import Controls from "$lib/components/Controls.svelte";
|
||||
import TranscriptionDisplay from "$lib/components/TranscriptionDisplay.svelte";
|
||||
import Settings from "$lib/components/Settings.svelte";
|
||||
import SidecarSetup from "$lib/components/SidecarSetup.svelte";
|
||||
import { backendStore } from "$lib/stores/backend";
|
||||
import { configStore } from "$lib/stores/config";
|
||||
|
||||
type SidecarState = "checking" | "needs_setup" | "starting" | "connected";
|
||||
|
||||
let showSettings = $state(false);
|
||||
let sidecarState = $state<SidecarState>("checking");
|
||||
let debugLog = $state("");
|
||||
|
||||
let obsDisplayUrl = $derived(backendStore.obsUrl);
|
||||
let syncDisplayUrl = $derived(backendStore.syncUrl);
|
||||
let isConnected = $derived(backendStore.connectionState === "connected");
|
||||
let connectionState = $derived(backendStore.connectionState);
|
||||
|
||||
function openSettings() {
|
||||
showSettings = true;
|
||||
@@ -21,9 +28,72 @@
|
||||
showSettings = false;
|
||||
}
|
||||
|
||||
let tauriInvoke: ((cmd: string, args?: Record<string, unknown>) => Promise<unknown>) | null = null;
|
||||
|
||||
function log(msg: string) {
|
||||
console.log(`[App] ${msg}`);
|
||||
debugLog = msg;
|
||||
// Also write to file via Tauri if available
|
||||
tauriInvoke?.("write_log", { message: msg });
|
||||
}
|
||||
|
||||
async function checkAndLaunchSidecar() {
|
||||
try {
|
||||
log("Importing Tauri API...");
|
||||
const { invoke } = await import("@tauri-apps/api/core");
|
||||
tauriInvoke = invoke;
|
||||
|
||||
log("Checking if sidecar is installed...");
|
||||
sidecarState = "checking";
|
||||
const installed = await invoke<boolean>("check_sidecar");
|
||||
log(`Sidecar installed: ${installed}`);
|
||||
|
||||
if (!installed) {
|
||||
sidecarState = "needs_setup";
|
||||
return;
|
||||
}
|
||||
|
||||
await launchSidecar();
|
||||
} catch (err) {
|
||||
// Not running in Tauri (browser dev mode) - skip sidecar check
|
||||
// and connect directly to localhost:8081
|
||||
log(`Tauri not available (${err}), using dev mode`);
|
||||
sidecarState = "starting";
|
||||
backendStore.setPort(8081);
|
||||
backendStore.connect();
|
||||
configStore.fetchConfig();
|
||||
}
|
||||
}
|
||||
|
||||
async function launchSidecar() {
|
||||
try {
|
||||
const { invoke } = await import("@tauri-apps/api/core");
|
||||
|
||||
log("Starting sidecar...");
|
||||
sidecarState = "starting";
|
||||
await invoke("start_sidecar");
|
||||
|
||||
log("Getting sidecar port...");
|
||||
const port = await invoke<number>("get_sidecar_port");
|
||||
log(`Sidecar ready on port ${port}`);
|
||||
backendStore.setPort(port);
|
||||
backendStore.connect();
|
||||
configStore.fetchConfig();
|
||||
} catch (err) {
|
||||
// If sidecar launch fails, still try connecting to default port
|
||||
log(`Sidecar launch failed: ${err}, trying default port`);
|
||||
sidecarState = "starting";
|
||||
backendStore.connect();
|
||||
configStore.fetchConfig();
|
||||
}
|
||||
}
|
||||
|
||||
async function onSidecarReady() {
|
||||
await launchSidecar();
|
||||
}
|
||||
|
||||
onMount(() => {
|
||||
backendStore.connect();
|
||||
configStore.loadConfig();
|
||||
checkAndLaunchSidecar();
|
||||
|
||||
return () => {
|
||||
backendStore.disconnect();
|
||||
@@ -31,33 +101,141 @@
|
||||
});
|
||||
</script>
|
||||
|
||||
<div class="app-shell">
|
||||
<Header onSettingsClick={openSettings} />
|
||||
<StatusBar />
|
||||
|
||||
<div class="display-links">
|
||||
<span class="link-label">OBS:</span>
|
||||
<a href={obsDisplayUrl} target="_blank" rel="noopener">{obsDisplayUrl}</a>
|
||||
{#if syncDisplayUrl}
|
||||
<span class="link-separator">|</span>
|
||||
<span class="link-label">Sync:</span>
|
||||
<a href={syncDisplayUrl} target="_blank" rel="noopener"
|
||||
>{syncDisplayUrl}</a
|
||||
>
|
||||
{/if}
|
||||
{#if sidecarState === "checking"}
|
||||
<div class="connecting-overlay" style="background:#1e1e1e;color:#e0e0e0;display:flex;align-items:center;justify-content:center;height:100%;width:100%;">
|
||||
<div class="connecting-content" style="text-align:center;">
|
||||
<div class="connecting-icon">
|
||||
<div class="spinner"></div>
|
||||
</div>
|
||||
<h2 style="font-size:20px;margin:16px 0 8px;">Local Transcription</h2>
|
||||
<p style="color:#a0a0a0;font-size:14px;">Checking setup...</p>
|
||||
{#if debugLog}
|
||||
<p style="color:#707070;font-size:11px;margin-top:12px;">{debugLog}</p>
|
||||
{/if}
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<TranscriptionDisplay />
|
||||
<Controls />
|
||||
{:else if sidecarState === "needs_setup"}
|
||||
<SidecarSetup onComplete={onSidecarReady} />
|
||||
|
||||
<div class="version-label">v{backendStore.version}</div>
|
||||
</div>
|
||||
{:else if !isConnected}
|
||||
<div class="connecting-overlay" style="background:#1e1e1e;color:#e0e0e0;display:flex;align-items:center;justify-content:center;height:100%;width:100%;">
|
||||
<div class="connecting-content" style="text-align:center;">
|
||||
<div class="connecting-icon">
|
||||
{#if connectionState === "error"}
|
||||
<svg width="48" height="48" viewBox="0 0 24 24" fill="none" stroke="#e74c3c" stroke-width="2">
|
||||
<circle cx="12" cy="12" r="10"/>
|
||||
<line x1="15" y1="9" x2="9" y2="15"/>
|
||||
<line x1="9" y1="9" x2="15" y2="15"/>
|
||||
</svg>
|
||||
{:else}
|
||||
<div class="spinner"></div>
|
||||
{/if}
|
||||
</div>
|
||||
<h2 style="font-size:20px;margin:16px 0 8px;">Local Transcription</h2>
|
||||
{#if connectionState === "error"}
|
||||
<p style="color:#a0a0a0;">Cannot connect to backend</p>
|
||||
<p class="hint">Make sure the Python backend is running:<br>
|
||||
<code>uv run python -m backend.main_headless</code></p>
|
||||
{:else}
|
||||
<p style="color:#a0a0a0;">Connecting to backend...</p>
|
||||
{/if}
|
||||
{#if debugLog}
|
||||
<p style="color:#707070;font-size:11px;margin-top:12px;">{debugLog}</p>
|
||||
{/if}
|
||||
</div>
|
||||
</div>
|
||||
|
||||
{#if showSettings}
|
||||
<Settings onClose={closeSettings} />
|
||||
{:else}
|
||||
<div class="app-shell">
|
||||
<Header onSettingsClick={openSettings} />
|
||||
<StatusBar />
|
||||
|
||||
<div class="display-links">
|
||||
<span class="link-label">OBS:</span>
|
||||
<a href={obsDisplayUrl} target="_blank" rel="noopener">{obsDisplayUrl}</a>
|
||||
{#if syncDisplayUrl}
|
||||
<span class="link-separator">|</span>
|
||||
<span class="link-label">Sync:</span>
|
||||
<a href={syncDisplayUrl} target="_blank" rel="noopener"
|
||||
>{syncDisplayUrl}</a
|
||||
>
|
||||
{/if}
|
||||
</div>
|
||||
|
||||
<TranscriptionDisplay />
|
||||
<Controls />
|
||||
|
||||
<div class="version-label">v{backendStore.version}</div>
|
||||
</div>
|
||||
|
||||
{#if showSettings}
|
||||
<Settings onClose={closeSettings} />
|
||||
{/if}
|
||||
{/if}
|
||||
|
||||
<style>
|
||||
.connecting-overlay {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
justify-content: center;
|
||||
height: 100%;
|
||||
width: 100%;
|
||||
background-color: var(--bg-primary);
|
||||
}
|
||||
|
||||
.connecting-content {
|
||||
text-align: center;
|
||||
color: var(--text-primary);
|
||||
}
|
||||
|
||||
.connecting-content h2 {
|
||||
margin: 16px 0 8px;
|
||||
font-size: 20px;
|
||||
font-weight: 600;
|
||||
}
|
||||
|
||||
.connecting-content p {
|
||||
margin: 4px 0;
|
||||
color: var(--text-secondary);
|
||||
font-size: 14px;
|
||||
}
|
||||
|
||||
.connecting-content .hint {
|
||||
margin-top: 16px;
|
||||
font-size: 12px;
|
||||
color: var(--text-muted);
|
||||
}
|
||||
|
||||
.connecting-content code {
|
||||
display: inline-block;
|
||||
margin-top: 4px;
|
||||
padding: 4px 8px;
|
||||
background: var(--bg-tertiary);
|
||||
border-radius: 4px;
|
||||
font-size: 12px;
|
||||
color: var(--text-primary);
|
||||
}
|
||||
|
||||
.connecting-icon {
|
||||
display: flex;
|
||||
justify-content: center;
|
||||
margin-bottom: 8px;
|
||||
}
|
||||
|
||||
.spinner {
|
||||
width: 40px;
|
||||
height: 40px;
|
||||
border: 3px solid var(--border-color);
|
||||
border-top-color: var(--accent-color, #4CAF50);
|
||||
border-radius: 50%;
|
||||
animation: spin 0.8s linear infinite;
|
||||
}
|
||||
|
||||
@keyframes spin {
|
||||
to { transform: rotate(360deg); }
|
||||
}
|
||||
|
||||
.app-shell {
|
||||
display: flex;
|
||||
flex-direction: column;
|
||||
|
||||
@@ -37,10 +37,14 @@
|
||||
let syncPassphrase = $state("");
|
||||
let remoteMode = $state("local");
|
||||
let remoteServerUrl = $state("");
|
||||
let byokApiKey = $state("");
|
||||
let managedEmail = $state("");
|
||||
let managedPassword = $state("");
|
||||
let autoCheckUpdates = $state(true);
|
||||
|
||||
let saving = $state(false);
|
||||
let saveMessage = $state("");
|
||||
|
||||
// Fetched device lists
|
||||
let audioDevices = $state<{ id: string; name: string }[]>([]);
|
||||
let computeDevices = $state<{ id: string; name: string }[]>([]);
|
||||
@@ -107,6 +111,7 @@
|
||||
syncPassphrase = cfg.server_sync.passphrase;
|
||||
remoteMode = cfg.remote.mode;
|
||||
remoteServerUrl = cfg.remote.server_url;
|
||||
byokApiKey = cfg.remote.byok_api_key ?? "";
|
||||
autoCheckUpdates = cfg.updates.auto_check;
|
||||
});
|
||||
|
||||
@@ -183,17 +188,23 @@
|
||||
remote: {
|
||||
mode: remoteMode,
|
||||
server_url: remoteServerUrl,
|
||||
byok_api_key: byokApiKey,
|
||||
},
|
||||
updates: {
|
||||
auto_check: autoCheckUpdates,
|
||||
},
|
||||
};
|
||||
|
||||
saving = true;
|
||||
saveMessage = "";
|
||||
try {
|
||||
await configStore.saveConfig(updates);
|
||||
onClose();
|
||||
await configStore.updateConfig(updates);
|
||||
saveMessage = "Settings saved!";
|
||||
setTimeout(() => onClose(), 600);
|
||||
} catch (err) {
|
||||
console.error("Failed to save settings:", err);
|
||||
saveMessage = `Error: ${err}`;
|
||||
saving = false;
|
||||
}
|
||||
}
|
||||
|
||||
@@ -203,7 +214,7 @@
|
||||
|
||||
async function handleCheckUpdates() {
|
||||
try {
|
||||
await backendStore.apiPost("/api/check-updates");
|
||||
await backendStore.apiGet("/api/check-update");
|
||||
} catch (err) {
|
||||
console.error("Failed to check for updates:", err);
|
||||
}
|
||||
@@ -211,7 +222,7 @@
|
||||
|
||||
async function handleManagedLogin() {
|
||||
try {
|
||||
await backendStore.apiPost("/api/remote/login", {
|
||||
await backendStore.apiPost("/api/login", {
|
||||
email: managedEmail,
|
||||
password: managedPassword,
|
||||
});
|
||||
@@ -222,7 +233,7 @@
|
||||
|
||||
async function handleManagedRegister() {
|
||||
try {
|
||||
await backendStore.apiPost("/api/remote/register", {
|
||||
await backendStore.apiPost("/api/register", {
|
||||
email: managedEmail,
|
||||
password: managedPassword,
|
||||
});
|
||||
@@ -572,7 +583,7 @@
|
||||
BYOK (Bring Your Own Key)
|
||||
</label>
|
||||
</div>
|
||||
{#if remoteMode !== "local"}
|
||||
{#if remoteMode === "managed"}
|
||||
<div class="field">
|
||||
<label for="remote-url">Server URL</label>
|
||||
<input
|
||||
@@ -583,6 +594,20 @@
|
||||
/>
|
||||
</div>
|
||||
{/if}
|
||||
{#if remoteMode === "byok"}
|
||||
<div class="field">
|
||||
<label for="byok-key">Deepgram API Key</label>
|
||||
<input
|
||||
id="byok-key"
|
||||
type="password"
|
||||
bind:value={byokApiKey}
|
||||
placeholder="Enter your Deepgram API key"
|
||||
/>
|
||||
<p style="font-size: 11px; color: var(--text-muted); margin-top: 4px;">
|
||||
Get a key at <a href="https://console.deepgram.com" target="_blank" rel="noopener" style="color: var(--accent-blue);">console.deepgram.com</a>
|
||||
</p>
|
||||
</div>
|
||||
{/if}
|
||||
{#if remoteMode === "managed"}
|
||||
<div class="managed-auth">
|
||||
<div class="field">
|
||||
@@ -626,8 +651,13 @@
|
||||
</div>
|
||||
|
||||
<div class="settings-footer">
|
||||
<button onclick={handleCancel}>Cancel</button>
|
||||
<button class="primary" onclick={handleSave}>Save</button>
|
||||
{#if saveMessage}
|
||||
<span class="save-message" class:error={saveMessage.startsWith("Error")}>{saveMessage}</span>
|
||||
{/if}
|
||||
<button onclick={handleCancel} disabled={saving}>Cancel</button>
|
||||
<button class="primary" onclick={handleSave} disabled={saving}>
|
||||
{saving ? "Saving..." : "Save"}
|
||||
</button>
|
||||
</div>
|
||||
</div>
|
||||
</div>
|
||||
@@ -771,10 +801,21 @@
|
||||
|
||||
.settings-footer {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
justify-content: flex-end;
|
||||
gap: 8px;
|
||||
padding: 16px 20px;
|
||||
border-top: 1px solid var(--border-color);
|
||||
flex-shrink: 0;
|
||||
}
|
||||
|
||||
.save-message {
|
||||
margin-right: auto;
|
||||
font-size: 13px;
|
||||
color: #4CAF50;
|
||||
}
|
||||
|
||||
.save-message.error {
|
||||
color: #f44336;
|
||||
}
|
||||
</style>
|
||||
|
||||
384
src/lib/components/SidecarSetup.svelte
Normal file
384
src/lib/components/SidecarSetup.svelte
Normal file
@@ -0,0 +1,384 @@
|
||||
<script lang="ts">
|
||||
import { invoke } from "@tauri-apps/api/core";
|
||||
import { listen } from "@tauri-apps/api/event";
|
||||
import { onMount } from "svelte";
|
||||
|
||||
interface Props {
|
||||
onComplete: () => void;
|
||||
}
|
||||
|
||||
let { onComplete }: Props = $props();
|
||||
|
||||
type SetupState = "choose" | "downloading" | "error" | "success";
|
||||
|
||||
let setupState = $state<SetupState>("choose");
|
||||
let variant = $state<"cpu" | "cuda">("cpu");
|
||||
let progress = $state(0);
|
||||
let progressMessage = $state("");
|
||||
let errorMessage = $state("");
|
||||
|
||||
let unlisten: (() => void) | null = null;
|
||||
|
||||
onMount(() => {
|
||||
return () => {
|
||||
if (unlisten) {
|
||||
unlisten();
|
||||
unlisten = null;
|
||||
}
|
||||
};
|
||||
});
|
||||
|
||||
async function startDownload() {
|
||||
setupState = "downloading";
|
||||
progress = 0;
|
||||
progressMessage = "Starting download...";
|
||||
errorMessage = "";
|
||||
|
||||
try {
|
||||
// Listen for progress events from the Tauri backend
|
||||
unlisten = await listen<{ progress: number; message: string }>(
|
||||
"sidecar-download-progress",
|
||||
(event) => {
|
||||
progress = event.payload.progress;
|
||||
progressMessage = event.payload.message;
|
||||
}
|
||||
);
|
||||
|
||||
await invoke("download_sidecar", { variant });
|
||||
|
||||
// Download complete
|
||||
setupState = "success";
|
||||
if (unlisten) {
|
||||
unlisten();
|
||||
unlisten = null;
|
||||
}
|
||||
|
||||
// Brief pause to show success, then proceed
|
||||
setTimeout(() => {
|
||||
onComplete();
|
||||
}, 1500);
|
||||
} catch (err) {
|
||||
setupState = "error";
|
||||
errorMessage = err instanceof Error ? err.message : String(err);
|
||||
if (unlisten) {
|
||||
unlisten();
|
||||
unlisten = null;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
function retry() {
|
||||
setupState = "choose";
|
||||
progress = 0;
|
||||
progressMessage = "";
|
||||
errorMessage = "";
|
||||
}
|
||||
</script>
|
||||
|
||||
<div class="setup-overlay">
|
||||
<div class="setup-card">
|
||||
<div class="setup-header">
|
||||
<h1 class="app-title">Local Transcription</h1>
|
||||
<h2 class="setup-heading">First-Time Setup</h2>
|
||||
</div>
|
||||
|
||||
{#if setupState === "choose"}
|
||||
<p class="setup-description">
|
||||
The app needs to download its transcription engine before you can start.
|
||||
Choose the version that best fits your hardware.
|
||||
</p>
|
||||
|
||||
<div class="variant-options">
|
||||
<label class="variant-option" class:selected={variant === "cpu"}>
|
||||
<input
|
||||
type="radio"
|
||||
name="variant"
|
||||
value="cpu"
|
||||
bind:group={variant}
|
||||
/>
|
||||
<div class="variant-info">
|
||||
<span class="variant-name">Standard (CPU)</span>
|
||||
<span class="variant-desc">Works on all computers (~500 MB download)</span>
|
||||
</div>
|
||||
</label>
|
||||
|
||||
<label class="variant-option" class:selected={variant === "cuda"}>
|
||||
<input
|
||||
type="radio"
|
||||
name="variant"
|
||||
value="cuda"
|
||||
bind:group={variant}
|
||||
/>
|
||||
<div class="variant-info">
|
||||
<span class="variant-name">GPU Accelerated (CUDA)</span>
|
||||
<span class="variant-desc">Faster transcription with NVIDIA GPU (~2 GB download)</span>
|
||||
</div>
|
||||
</label>
|
||||
</div>
|
||||
|
||||
<button class="download-btn" onclick={startDownload}>
|
||||
Download & Install
|
||||
</button>
|
||||
|
||||
{:else if setupState === "downloading"}
|
||||
<div class="progress-section">
|
||||
<p class="progress-message">{progressMessage}</p>
|
||||
<div class="progress-bar-track">
|
||||
<div
|
||||
class="progress-bar-fill"
|
||||
style="width: {progress}%"
|
||||
></div>
|
||||
</div>
|
||||
<p class="progress-percent">{Math.round(progress)}%</p>
|
||||
</div>
|
||||
|
||||
{:else if setupState === "error"}
|
||||
<div class="error-section">
|
||||
<div class="error-icon">
|
||||
<svg width="48" height="48" viewBox="0 0 24 24" fill="none" stroke="#f44336" stroke-width="2">
|
||||
<circle cx="12" cy="12" r="10"/>
|
||||
<line x1="15" y1="9" x2="9" y2="15"/>
|
||||
<line x1="9" y1="9" x2="15" y2="15"/>
|
||||
</svg>
|
||||
</div>
|
||||
<p class="error-title">Download Failed</p>
|
||||
<p class="error-message">{errorMessage}</p>
|
||||
<button class="retry-btn" onclick={retry}>
|
||||
Try Again
|
||||
</button>
|
||||
</div>
|
||||
|
||||
{:else if setupState === "success"}
|
||||
<div class="success-section">
|
||||
<div class="success-icon">
|
||||
<svg width="48" height="48" viewBox="0 0 24 24" fill="none" stroke="#4CAF50" stroke-width="2">
|
||||
<circle cx="12" cy="12" r="10"/>
|
||||
<polyline points="16 9 10.5 15 8 12.5"/>
|
||||
</svg>
|
||||
</div>
|
||||
<p class="success-title">Setup Complete</p>
|
||||
<p class="success-message">The transcription engine is ready to go.</p>
|
||||
</div>
|
||||
{/if}
|
||||
</div>
|
||||
</div>
|
||||
|
||||
<style>
|
||||
.setup-overlay {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
justify-content: center;
|
||||
height: 100%;
|
||||
width: 100%;
|
||||
background-color: #1e1e1e;
|
||||
}
|
||||
|
||||
.setup-card {
|
||||
background-color: #2a2a2a;
|
||||
border-radius: 12px;
|
||||
padding: 40px;
|
||||
max-width: 480px;
|
||||
width: 100%;
|
||||
margin: 20px;
|
||||
box-shadow: 0 8px 32px rgba(0, 0, 0, 0.4);
|
||||
}
|
||||
|
||||
.setup-header {
|
||||
text-align: center;
|
||||
margin-bottom: 24px;
|
||||
}
|
||||
|
||||
.app-title {
|
||||
font-size: 24px;
|
||||
font-weight: 700;
|
||||
color: #e0e0e0;
|
||||
margin-bottom: 4px;
|
||||
}
|
||||
|
||||
.setup-heading {
|
||||
font-size: 16px;
|
||||
font-weight: 500;
|
||||
color: #a0a0a0;
|
||||
}
|
||||
|
||||
.setup-description {
|
||||
font-size: 14px;
|
||||
color: #a0a0a0;
|
||||
line-height: 1.6;
|
||||
text-align: center;
|
||||
margin-bottom: 24px;
|
||||
}
|
||||
|
||||
.variant-options {
|
||||
display: flex;
|
||||
flex-direction: column;
|
||||
gap: 12px;
|
||||
margin-bottom: 24px;
|
||||
}
|
||||
|
||||
.variant-option {
|
||||
display: flex;
|
||||
align-items: center;
|
||||
gap: 12px;
|
||||
padding: 14px 16px;
|
||||
border: 2px solid #444;
|
||||
border-radius: 8px;
|
||||
cursor: pointer;
|
||||
transition: border-color 0.15s ease, background-color 0.15s ease;
|
||||
}
|
||||
|
||||
.variant-option:hover {
|
||||
background-color: #333;
|
||||
border-color: #555;
|
||||
}
|
||||
|
||||
.variant-option.selected {
|
||||
border-color: #4CAF50;
|
||||
background-color: rgba(76, 175, 80, 0.08);
|
||||
}
|
||||
|
||||
.variant-option input[type="radio"] {
|
||||
width: 18px;
|
||||
height: 18px;
|
||||
flex-shrink: 0;
|
||||
}
|
||||
|
||||
.variant-info {
|
||||
display: flex;
|
||||
flex-direction: column;
|
||||
gap: 2px;
|
||||
}
|
||||
|
||||
.variant-name {
|
||||
font-size: 14px;
|
||||
font-weight: 600;
|
||||
color: #e0e0e0;
|
||||
}
|
||||
|
||||
.variant-desc {
|
||||
font-size: 12px;
|
||||
color: #888;
|
||||
}
|
||||
|
||||
.download-btn {
|
||||
display: block;
|
||||
width: 100%;
|
||||
padding: 12px 24px;
|
||||
font-size: 15px;
|
||||
font-weight: 600;
|
||||
color: white;
|
||||
background-color: #4CAF50;
|
||||
border: none;
|
||||
border-radius: 8px;
|
||||
cursor: pointer;
|
||||
transition: background-color 0.15s ease;
|
||||
}
|
||||
|
||||
.download-btn:hover {
|
||||
background-color: #45a049;
|
||||
}
|
||||
|
||||
.download-btn:active {
|
||||
transform: scale(0.98);
|
||||
}
|
||||
|
||||
/* Progress state */
|
||||
.progress-section {
|
||||
text-align: center;
|
||||
padding: 20px 0;
|
||||
}
|
||||
|
||||
.progress-message {
|
||||
font-size: 14px;
|
||||
color: #a0a0a0;
|
||||
margin-bottom: 16px;
|
||||
}
|
||||
|
||||
.progress-bar-track {
|
||||
width: 100%;
|
||||
height: 8px;
|
||||
background-color: #3a3a3a;
|
||||
border-radius: 4px;
|
||||
overflow: hidden;
|
||||
margin-bottom: 8px;
|
||||
}
|
||||
|
||||
.progress-bar-fill {
|
||||
height: 100%;
|
||||
background-color: #4CAF50;
|
||||
border-radius: 4px;
|
||||
transition: width 0.3s ease;
|
||||
}
|
||||
|
||||
.progress-percent {
|
||||
font-size: 13px;
|
||||
color: #707070;
|
||||
}
|
||||
|
||||
/* Error state */
|
||||
.error-section {
|
||||
text-align: center;
|
||||
padding: 10px 0;
|
||||
}
|
||||
|
||||
.error-icon {
|
||||
display: flex;
|
||||
justify-content: center;
|
||||
margin-bottom: 12px;
|
||||
}
|
||||
|
||||
.error-title {
|
||||
font-size: 18px;
|
||||
font-weight: 600;
|
||||
color: #f44336;
|
||||
margin-bottom: 8px;
|
||||
}
|
||||
|
||||
.error-message {
|
||||
font-size: 13px;
|
||||
color: #a0a0a0;
|
||||
margin-bottom: 20px;
|
||||
word-break: break-word;
|
||||
}
|
||||
|
||||
.retry-btn {
|
||||
display: inline-block;
|
||||
padding: 10px 28px;
|
||||
font-size: 14px;
|
||||
font-weight: 600;
|
||||
color: white;
|
||||
background-color: #4CAF50;
|
||||
border: none;
|
||||
border-radius: 8px;
|
||||
cursor: pointer;
|
||||
transition: background-color 0.15s ease;
|
||||
}
|
||||
|
||||
.retry-btn:hover {
|
||||
background-color: #45a049;
|
||||
}
|
||||
|
||||
/* Success state */
|
||||
.success-section {
|
||||
text-align: center;
|
||||
padding: 20px 0;
|
||||
}
|
||||
|
||||
.success-icon {
|
||||
display: flex;
|
||||
justify-content: center;
|
||||
margin-bottom: 12px;
|
||||
}
|
||||
|
||||
.success-title {
|
||||
font-size: 18px;
|
||||
font-weight: 600;
|
||||
color: #4CAF50;
|
||||
margin-bottom: 4px;
|
||||
}
|
||||
|
||||
.success-message {
|
||||
font-size: 14px;
|
||||
color: #a0a0a0;
|
||||
}
|
||||
</style>
|
||||
@@ -54,6 +54,35 @@ async function apiFetch(path: string, options?: RequestInit): Promise<Response>
|
||||
return fetch(url, { ...options, headers });
|
||||
}
|
||||
|
||||
// ── Status polling ──────────────────────────────────────────────────
|
||||
|
||||
let statusPollTimer: ReturnType<typeof setTimeout> | null = null;
|
||||
|
||||
async function pollStatus() {
|
||||
try {
|
||||
const resp = await fetch(apiUrl("/api/status"));
|
||||
if (resp.ok) {
|
||||
const data = await resp.json();
|
||||
if (data.state) {
|
||||
state.appState = data.state as AppState;
|
||||
}
|
||||
if (data.engine_device) {
|
||||
state.deviceInfo = data.engine_device;
|
||||
}
|
||||
if (data.version) {
|
||||
state.version = data.version;
|
||||
}
|
||||
}
|
||||
} catch {
|
||||
// API not ready yet, will retry
|
||||
}
|
||||
|
||||
// Keep polling every 2s while still initializing
|
||||
if (state.appState === "initializing" && state.connectionState === "connected") {
|
||||
statusPollTimer = setTimeout(pollStatus, 2000);
|
||||
}
|
||||
}
|
||||
|
||||
// ── WebSocket management ─────────────────────────────────────────────
|
||||
|
||||
function connectWebSocket() {
|
||||
@@ -80,6 +109,9 @@ function _openSocket() {
|
||||
clearTimeout(reconnectTimer);
|
||||
reconnectTimer = null;
|
||||
}
|
||||
// Poll status to catch engine ready state that may have been
|
||||
// missed (engine can finish before WebSocket connects)
|
||||
pollStatus();
|
||||
};
|
||||
|
||||
ws.onmessage = (event) => {
|
||||
@@ -132,6 +164,10 @@ function _scheduleReconnect() {
|
||||
}
|
||||
|
||||
function disconnect() {
|
||||
if (statusPollTimer) {
|
||||
clearTimeout(statusPollTimer);
|
||||
statusPollTimer = null;
|
||||
}
|
||||
if (reconnectTimer) {
|
||||
clearTimeout(reconnectTimer);
|
||||
reconnectTimer = null;
|
||||
@@ -255,6 +291,14 @@ export const backendStore = {
|
||||
get wsUrl() {
|
||||
return `ws://localhost:${state.port}/ws/control`;
|
||||
},
|
||||
get obsUrl() {
|
||||
// OBS display runs on the web server port (one below the API port)
|
||||
const obsPort = state.port > 0 ? state.port - 1 : 8080;
|
||||
return `http://localhost:${obsPort}`;
|
||||
},
|
||||
get syncUrl() {
|
||||
return "";
|
||||
},
|
||||
setPort,
|
||||
connect: connectWebSocket,
|
||||
disconnect,
|
||||
77
src/lib/stores/backend.test.ts
Normal file
77
src/lib/stores/backend.test.ts
Normal file
@@ -0,0 +1,77 @@
|
||||
import { describe, it, expect, vi, beforeEach } from "vitest";
|
||||
import { backendStore } from "./backend.svelte.ts";
|
||||
|
||||
// Mock WebSocket globally so the store module can reference it
|
||||
class MockWebSocket {
|
||||
onopen: ((ev: Event) => void) | null = null;
|
||||
onclose: ((ev: CloseEvent) => void) | null = null;
|
||||
onmessage: ((ev: MessageEvent) => void) | null = null;
|
||||
onerror: ((ev: Event) => void) | null = null;
|
||||
close = vi.fn();
|
||||
}
|
||||
vi.stubGlobal("WebSocket", MockWebSocket);
|
||||
|
||||
// Mock fetch to prevent real network calls
|
||||
vi.stubGlobal(
|
||||
"fetch",
|
||||
vi.fn(() =>
|
||||
Promise.resolve({
|
||||
ok: true,
|
||||
json: () => Promise.resolve({}),
|
||||
})
|
||||
)
|
||||
);
|
||||
|
||||
describe("backend store", () => {
|
||||
beforeEach(() => {
|
||||
backendStore.disconnect();
|
||||
backendStore.setPort(8081);
|
||||
});
|
||||
|
||||
it("test_exports_expected_properties", () => {
|
||||
expect(backendStore).toHaveProperty("port");
|
||||
expect(backendStore).toHaveProperty("connectionState");
|
||||
expect(backendStore).toHaveProperty("connected");
|
||||
expect(backendStore).toHaveProperty("appState");
|
||||
expect(backendStore).toHaveProperty("stateMessage");
|
||||
expect(backendStore).toHaveProperty("deviceInfo");
|
||||
expect(backendStore).toHaveProperty("version");
|
||||
expect(backendStore).toHaveProperty("lastError");
|
||||
expect(backendStore).toHaveProperty("apiBaseUrl");
|
||||
expect(backendStore).toHaveProperty("wsUrl");
|
||||
expect(backendStore).toHaveProperty("obsUrl");
|
||||
expect(backendStore).toHaveProperty("syncUrl");
|
||||
});
|
||||
|
||||
it("test_exports_expected_methods", () => {
|
||||
expect(typeof backendStore.setPort).toBe("function");
|
||||
expect(typeof backendStore.connect).toBe("function");
|
||||
expect(typeof backendStore.disconnect).toBe("function");
|
||||
expect(typeof backendStore.apiUrl).toBe("function");
|
||||
expect(typeof backendStore.apiFetch).toBe("function");
|
||||
expect(typeof backendStore.apiGet).toBe("function");
|
||||
expect(typeof backendStore.apiPost).toBe("function");
|
||||
expect(typeof backendStore.apiPut).toBe("function");
|
||||
});
|
||||
|
||||
it("test_obsUrl_derives_from_port", () => {
|
||||
backendStore.setPort(8081);
|
||||
expect(backendStore.obsUrl).toBe("http://localhost:8080");
|
||||
});
|
||||
|
||||
it("test_apiBaseUrl_uses_port", () => {
|
||||
backendStore.setPort(8081);
|
||||
expect(backendStore.apiBaseUrl).toBe("http://localhost:8081");
|
||||
});
|
||||
|
||||
it("test_wsUrl_uses_port", () => {
|
||||
backendStore.setPort(8081);
|
||||
expect(backendStore.wsUrl).toBe("ws://localhost:8081/ws/control");
|
||||
});
|
||||
|
||||
it("test_initial_state", () => {
|
||||
// After disconnect() in beforeEach, state should be disconnected
|
||||
expect(backendStore.connectionState).toBe("disconnected");
|
||||
expect(backendStore.appState).toBe("initializing");
|
||||
});
|
||||
});
|
||||
48
src/lib/stores/config.test.ts
Normal file
48
src/lib/stores/config.test.ts
Normal file
@@ -0,0 +1,48 @@
|
||||
import { describe, it, expect, vi } from "vitest";
|
||||
|
||||
// Mock fetch so the backend store module doesn't make real requests
|
||||
vi.stubGlobal(
|
||||
"fetch",
|
||||
vi.fn(() =>
|
||||
Promise.resolve({
|
||||
ok: true,
|
||||
json: () => Promise.resolve({}),
|
||||
})
|
||||
)
|
||||
);
|
||||
|
||||
// Mock WebSocket for the backend store dependency
|
||||
class MockWebSocket {
|
||||
onopen: ((ev: Event) => void) | null = null;
|
||||
onclose: ((ev: CloseEvent) => void) | null = null;
|
||||
onmessage: ((ev: MessageEvent) => void) | null = null;
|
||||
onerror: ((ev: Event) => void) | null = null;
|
||||
close = vi.fn();
|
||||
}
|
||||
vi.stubGlobal("WebSocket", MockWebSocket);
|
||||
|
||||
import { configStore } from "./config.svelte.ts";
|
||||
|
||||
describe("config store", () => {
|
||||
it("test_has_fetchConfig_method", () => {
|
||||
expect(typeof configStore.fetchConfig).toBe("function");
|
||||
});
|
||||
|
||||
it("test_has_updateConfig_method", () => {
|
||||
expect(typeof configStore.updateConfig).toBe("function");
|
||||
});
|
||||
|
||||
it("test_config_defaults_have_expected_keys", () => {
|
||||
const cfg = configStore.config;
|
||||
expect(cfg).toHaveProperty("user");
|
||||
expect(cfg).toHaveProperty("audio");
|
||||
expect(cfg).toHaveProperty("transcription");
|
||||
expect(cfg).toHaveProperty("display");
|
||||
expect(cfg).toHaveProperty("remote");
|
||||
expect(cfg).toHaveProperty("updates");
|
||||
});
|
||||
|
||||
it("test_remote_config_has_byok_api_key", () => {
|
||||
expect(configStore.config.remote.byok_api_key).toBeDefined();
|
||||
});
|
||||
});
|
||||
34
src/lib/stores/file-extension.test.ts
Normal file
34
src/lib/stores/file-extension.test.ts
Normal file
@@ -0,0 +1,34 @@
|
||||
import { describe, it, expect } from "vitest";
|
||||
import * as fs from "node:fs";
|
||||
import * as path from "node:path";
|
||||
|
||||
describe("store file extensions", () => {
|
||||
it("test_store_files_use_svelte_ts_extension", () => {
|
||||
const storesDir = path.resolve(__dirname);
|
||||
const files = fs.readdirSync(storesDir);
|
||||
|
||||
// Find .ts files that are NOT .svelte.ts and NOT test files
|
||||
const plainTsFiles = files.filter(
|
||||
(f) =>
|
||||
f.endsWith(".ts") &&
|
||||
!f.endsWith(".svelte.ts") &&
|
||||
!f.endsWith(".test.ts")
|
||||
);
|
||||
|
||||
for (const file of plainTsFiles) {
|
||||
const content = fs.readFileSync(path.join(storesDir, file), "utf-8");
|
||||
expect(content).not.toMatch(
|
||||
/\$state\s*[<(]/,
|
||||
`${file} uses $state() but does not have .svelte.ts extension`
|
||||
);
|
||||
expect(content).not.toMatch(
|
||||
/\$derived\s*[<(]/,
|
||||
`${file} uses $derived() but does not have .svelte.ts extension`
|
||||
);
|
||||
expect(content).not.toMatch(
|
||||
/\$effect\s*[<(]/,
|
||||
`${file} uses $effect() but does not have .svelte.ts extension`
|
||||
);
|
||||
}
|
||||
});
|
||||
});
|
||||
71
src/lib/stores/transcriptions.test.ts
Normal file
71
src/lib/stores/transcriptions.test.ts
Normal file
@@ -0,0 +1,71 @@
|
||||
import { describe, it, expect, vi, beforeEach } from "vitest";
|
||||
|
||||
// Mock WebSocket for the backend store dependency (loaded transitively)
|
||||
class MockWebSocket {
|
||||
onopen: ((ev: Event) => void) | null = null;
|
||||
onclose: ((ev: CloseEvent) => void) | null = null;
|
||||
onmessage: ((ev: MessageEvent) => void) | null = null;
|
||||
onerror: ((ev: Event) => void) | null = null;
|
||||
close = vi.fn();
|
||||
}
|
||||
vi.stubGlobal("WebSocket", MockWebSocket);
|
||||
|
||||
vi.stubGlobal(
|
||||
"fetch",
|
||||
vi.fn(() =>
|
||||
Promise.resolve({
|
||||
ok: true,
|
||||
json: () => Promise.resolve({}),
|
||||
})
|
||||
)
|
||||
);
|
||||
|
||||
import { transcriptionStore } from "./transcriptions.svelte.ts";
|
||||
|
||||
describe("transcriptions store", () => {
|
||||
beforeEach(() => {
|
||||
transcriptionStore.clearAll();
|
||||
});
|
||||
|
||||
it("test_addTranscription", () => {
|
||||
transcriptionStore.addTranscription({
|
||||
text: "Hello world",
|
||||
user_name: "TestUser",
|
||||
timestamp: "12:00:00",
|
||||
});
|
||||
|
||||
expect(transcriptionStore.items.length).toBe(1);
|
||||
expect(transcriptionStore.items[0].text).toBe("Hello world");
|
||||
expect(transcriptionStore.items[0].userName).toBe("TestUser");
|
||||
expect(transcriptionStore.items[0].timestamp).toBe("12:00:00");
|
||||
expect(transcriptionStore.items[0].isPreview).toBe(false);
|
||||
});
|
||||
|
||||
it("test_clearAll", () => {
|
||||
transcriptionStore.addTranscription({ text: "One" });
|
||||
transcriptionStore.addTranscription({ text: "Two" });
|
||||
expect(transcriptionStore.items.length).toBe(2);
|
||||
|
||||
transcriptionStore.clearAll();
|
||||
expect(transcriptionStore.items.length).toBe(0);
|
||||
});
|
||||
|
||||
it("test_getPlainText", () => {
|
||||
transcriptionStore.addTranscription({
|
||||
text: "Hello",
|
||||
user_name: "Alice",
|
||||
timestamp: "10:00",
|
||||
});
|
||||
transcriptionStore.addTranscription({
|
||||
text: "World",
|
||||
user_name: "Bob",
|
||||
timestamp: "10:01",
|
||||
});
|
||||
|
||||
const text = transcriptionStore.getPlainText();
|
||||
expect(text).toContain("[10:00] Alice: Hello");
|
||||
expect(text).toContain("[10:01] Bob: World");
|
||||
// Lines separated by newline
|
||||
expect(text.split("\n").length).toBe(2);
|
||||
});
|
||||
});
|
||||
@@ -1,7 +1,7 @@
|
||||
"""Version information for Local Transcription."""
|
||||
|
||||
__version__ = "1.4.3"
|
||||
__version_info__ = (1, 0, 2)
|
||||
__version__ = "1.4.18"
|
||||
__version_info__ = (1, 4, 18)
|
||||
|
||||
# Version history:
|
||||
# 1.4.0 - Auto-update feature:
|
||||
|
||||
@@ -10,6 +10,7 @@ export default defineConfig({
|
||||
alias: {
|
||||
$lib: path.resolve("./src/lib"),
|
||||
},
|
||||
extensions: [".svelte.ts", ".ts", ".svelte", ".js", ".mjs", ".mts"],
|
||||
},
|
||||
server: {
|
||||
port: 1420,
|
||||
@@ -18,4 +19,8 @@ export default defineConfig({
|
||||
ignored: ["**/src-tauri/**", "**/client/**", "**/server/**", "**/backend/**", "**/gui/**"],
|
||||
},
|
||||
},
|
||||
test: {
|
||||
environment: "jsdom",
|
||||
include: ["src/**/*.test.ts"],
|
||||
},
|
||||
});
|
||||
|
||||
Reference in New Issue
Block a user