|
| 1 | +# olsconfig.yaml sample for local ollama server |
| 2 | +# |
| 3 | +# 1. install local ollama server from https://ollama.com/ |
| 4 | +# 2. install llama3.1:latest model with: |
| 5 | +# ollama pull llama3.1:latest |
| 6 | +# 3. Copy this file to the project root of cloned lightspeed-service repo |
| 7 | +# 4. Install dependencies with: |
| 8 | +# make install-deps |
| 9 | +# 5. Start lightspeed-service with: |
| 10 | +# OPENAI_API_KEY=IGNORED make run |
| 11 | +# 6. Open https://localhost:8080/ui in your web browser |
| 12 | +# |
| 13 | +llm_providers: |
| 14 | + - name: ollama |
| 15 | + type: openai |
| 16 | + url: "http://localhost:11434/v1/" |
| 17 | + models: |
| 18 | + - name: "mistral" |
| 19 | + - name: 'llama3.2:latest' |
| 20 | + - name: my_rhoai |
| 21 | + type: openai |
| 22 | + url: "https://granite3-8b-wisdom-model-staging.apps.stage2-west.v2dz.p1.openshiftapps.com/v1" |
| 23 | + credentials_path: ols_api_key.txt |
| 24 | + models: |
| 25 | + - name: granite3-8b |
| 26 | +ols_config: |
| 27 | + # max_workers: 1 |
| 28 | + reference_content: |
| 29 | + # product_docs_index_path: "./vector_db/vector_db/aap_product_docs/2.5" |
| 30 | + # product_docs_index_id: aap-product-docs-2_5 |
| 31 | + # embeddings_model_path: "./vector_db/embeddings_model" |
| 32 | + conversation_cache: |
| 33 | + type: memory |
| 34 | + memory: |
| 35 | + max_entries: 1000 |
| 36 | + logging_config: |
| 37 | + app_log_level: info |
| 38 | + lib_log_level: warning |
| 39 | + uvicorn_log_level: info |
| 40 | + default_provider: ollama |
| 41 | + default_model: 'llama3.2:latest' |
| 42 | + query_validation_method: llm |
| 43 | + user_data_collection: |
| 44 | + feedback_disabled: false |
| 45 | + feedback_storage: "/tmp/data/feedback" |
| 46 | + transcripts_disabled: false |
| 47 | + transcripts_storage: "/tmp/data/transcripts" |
| 48 | +dev_config: |
| 49 | + # config options specific to dev environment - launching OLS in local |
| 50 | + enable_dev_ui: true |
| 51 | + disable_auth: true |
| 52 | + disable_tls: true |
| 53 | + pyroscope_url: "https://pyroscope.pyroscope.svc.cluster.local:4040" |
| 54 | + # llm_params: |
| 55 | + # temperature_override: 0 |
| 56 | + # k8s_auth_token: optional_token_when_no_available_kube_config |
0 commit comments