Skip to content

Commit b4db84f

Browse files
committed
(spike) using rhoai lls
ls -la /Users/lance/Documents/RedHat/lightspeed-stack/.git/.COMMIT_EDITMSG*
1 parent 714ce2b commit b4db84f

File tree

10 files changed

+45
-51
lines changed

10 files changed

+45
-51
lines changed

Dockerfile.rh-llama-stack

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,9 @@
1+
# Custom Red Hat llama-stack image with missing dependencies
2+
FROM quay.io/opendatahub/llama-stack:rhoai-v2.25-latest
3+
4+
# Install missing dependencies
5+
USER root
6+
RUN pip install faiss-cpu==1.11.0
7+
8+
# Switch back to the original user
9+
USER 1001

README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -245,7 +245,7 @@ version = "0.1.0"
245245
description = "Llama Stack runner"
246246
authors = []
247247
dependencies = [
248-
"llama-stack==0.2.21",
248+
"llama-stack==0.2.23",
249249
"fastapi>=0.115.12",
250250
"opentelemetry-sdk>=1.34.0",
251251
"opentelemetry-exporter-otlp>=1.34.0",

docker-compose.yaml

Lines changed: 12 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -1,23 +1,27 @@
11
services:
2+
# Red Hat llama-stack distribution with FAISS
23
llama-stack:
34
build:
45
context: .
5-
dockerfile: test.containerfile
6+
dockerfile: Dockerfile.rh-llama-stack
7+
platform: linux/amd64
68
container_name: llama-stack
79
ports:
8-
- "8321:8321" # Expose llama-stack on 8321 (adjust if needed)
10+
- "8321:8321"
911
volumes:
10-
- ./run.yaml:/app-root/run.yaml:Z
12+
- ./run.yaml:/opt/app-root/run.yaml:Z
1113
environment:
1214
- OPENAI_API_KEY=${OPENAI_API_KEY}
15+
- BRAVE_SEARCH_API_KEY=${BRAVE_SEARCH_API_KEY:-}
16+
- TAVILY_SEARCH_API_KEY=${TAVILY_SEARCH_API_KEY:-}
1317
networks:
1418
- lightspeednet
1519
healthcheck:
1620
test: ["CMD", "curl", "-f", "http://localhost:8321/v1/health"]
17-
interval: 10s # how often to run the check
18-
timeout: 5s # how long to wait before considering it failed
19-
retries: 3 # how many times to retry before marking as unhealthy
20-
start_period: 15s # time to wait before starting checks
21+
interval: 10s
22+
timeout: 5s
23+
retries: 3
24+
start_period: 30s
2125

2226
lightspeed-stack:
2327
build:
@@ -44,4 +48,4 @@ services:
4448

4549
networks:
4650
lightspeednet:
47-
driver: bridge
51+
driver: bridge

docs/deployment_guide.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -676,7 +676,7 @@ a4982f43195537b9eb1cec510fe6655f245d6d4b7236a4759808115d5d719972
676676
description = "Default template for PDM package"
677677
authors = []
678678
dependencies = [
679-
"llama-stack==0.2.18",
679+
"llama-stack==0.2.23",
680680
"fastapi>=0.115.12",
681681
"opentelemetry-sdk>=1.34.0",
682682
"opentelemetry-exporter-otlp>=1.34.0",

docs/getting_started.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ It is possible to run Lightspeed Core Stack service with Llama Stack "embedded"
2424
1. Add and install all required dependencies
2525
```bash
2626
uv add \
27-
"llama-stack==0.2.21" \
27+
"llama-stack==0.2.23" \
2828
"fastapi>=0.115.12" \
2929
"opentelemetry-sdk>=1.34.0" \
3030
"opentelemetry-exporter-otlp>=1.34.0" \

examples/pyproject.llamastack.toml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ version = "0.1.0"
44
description = "Default template for PDM package"
55
authors = []
66
dependencies = [
7-
"llama-stack==0.2.21",
7+
"llama-stack==0.2.23",
88
"fastapi>=0.115.12",
99
"opentelemetry-sdk>=1.34.0",
1010
"opentelemetry-exporter-otlp>=1.34.0",

pyproject.toml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -28,8 +28,8 @@ dependencies = [
2828
# Used by authentication/k8s integration
2929
"kubernetes>=30.1.0",
3030
# Used to call Llama Stack APIs
31-
"llama-stack==0.2.21",
32-
"llama-stack-client==0.2.21",
31+
"llama-stack==0.2.23",
32+
"llama-stack-client==0.2.23",
3333
# Used by Logger
3434
"rich>=14.0.0",
3535
# Used by JWK token auth handler

run.yaml

Lines changed: 9 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -12,6 +12,7 @@ apis:
1212
- telemetry
1313
- tool_runtime
1414
- vector_io
15+
- files
1516
benchmarks: []
1617
container_image: null
1718
datasets: []
@@ -105,6 +106,14 @@ providers:
105106
- provider_id: rag-runtime
106107
provider_type: inline::rag-runtime
107108
config: {}
109+
files:
110+
- provider_id: meta-reference-files
111+
provider_type: inline::localfs
112+
config:
113+
storage_dir: .llama/distributions/ollama/files
114+
metadata_store:
115+
type: sqlite
116+
db_path: .llama/distributions/ollama/files_metadata.db
108117
vector_io:
109118
- config:
110119
kvstore:

test.containerfile

Lines changed: 0 additions & 28 deletions
This file was deleted.

uv.lock

Lines changed: 9 additions & 9 deletions
Some generated files are not rendered by default. Learn more about customizing how changed files appear on GitHub.

0 commit comments

Comments
 (0)