Increase scope of embedding cli tests

This commit is contained in:
Sam Malayek 2025-11-02 13:55:01 -08:00
parent c1c3d99ef6
commit 2de1e6871f
3 changed files with 23 additions and 13 deletions

View File

@ -4,19 +4,25 @@ name: Embedding CLI
on: on:
workflow_dispatch: workflow_dispatch:
push: push:
branches: branches: [master, feature/**]
- feature/*
- master
paths: paths:
- '.github/workflows/embeddings.yml' - '.github/workflows/embedding.yml'
- 'examples/embedding/**' - 'examples/**'
- 'examples/tests/**' - 'src/**'
- 'ggml/**'
- 'include/**'
- '**/CMakeLists.txt'
- 'tests/e2e/embedding/**'
pull_request: pull_request:
types: [opened, synchronize, reopened] types: [opened, synchronize, reopened]
paths: paths:
- '.github/workflows/embeddings.yml' - '.github/workflows/embedding.yml'
- 'examples/embedding/**' - 'examples/**'
- 'examples/tests/**' - 'src/**'
- 'ggml/**'
- 'include/**'
- '**/CMakeLists.txt'
- 'tests/e2e/embedding/**'
jobs: jobs:
embedding-cli-tests: embedding-cli-tests:
@ -56,4 +62,4 @@ jobs:
- name: Run embedding tests - name: Run embedding tests
run: | run: |
pytest -v examples/tests pytest -v tests/e2e/embedding

View File

@ -1,14 +1,17 @@
import os, json, subprocess, hashlib import json
import hashlib
import os
import pytest
import subprocess
from pathlib import Path from pathlib import Path
import numpy as np import numpy as np
import pytest
# --------------------------------------------------------------------------- # ---------------------------------------------------------------------------
# Configuration constants # Configuration constants
# --------------------------------------------------------------------------- # ---------------------------------------------------------------------------
EPS = 1e-3 EPS = 1e-3
REPO_ROOT = Path(__file__).resolve().parents[2] REPO_ROOT = Path(__file__).resolve().parents[3]
EXE = REPO_ROOT / ("build/bin/llama-embedding.exe" if os.name == "nt" else "build/bin/llama-embedding") EXE = REPO_ROOT / ("build/bin/llama-embedding.exe" if os.name == "nt" else "build/bin/llama-embedding")
DEFAULT_ENV = {**os.environ, "LLAMA_CACHE": os.environ.get("LLAMA_CACHE", "tmp")} DEFAULT_ENV = {**os.environ, "LLAMA_CACHE": os.environ.get("LLAMA_CACHE", "tmp")}
SEED = "42" SEED = "42"
@ -96,6 +99,7 @@ def embedding_hash(vec: np.ndarray) -> str:
# Register custom mark so pytest doesn't warn about it # Register custom mark so pytest doesn't warn about it
pytestmark = pytest.mark.filterwarnings("ignore::pytest.PytestUnknownMarkWarning") pytestmark = pytest.mark.filterwarnings("ignore::pytest.PytestUnknownMarkWarning")
@pytest.mark.slow @pytest.mark.slow
@pytest.mark.parametrize("fmt", ["raw", "json"]) @pytest.mark.parametrize("fmt", ["raw", "json"])
@pytest.mark.parametrize("text", ["hello world", "hi 🌎", "line1\nline2\nline3"]) @pytest.mark.parametrize("text", ["hello world", "hi 🌎", "line1\nline2\nline3"])