Skip to content

[feat] [doc improvement]: Where does context chat backend get its compute? #233

@tareko

Description

@tareko

Describe the feature you'd like to request
I finally got context chat working, and it has tremendous potential. I am running our server on a digitalocean droplet, and do our computing on a separate ollama server.

I looked at the configuration file /nc_app_context_chat_backend_data/config.yaml, but I don't have the skills to understand where the computing is happening. If I had to guess, I would guess it's all happening locally. For sure any request takes a very long time (up to 60 minutes) and doesn't seem to register on the GPU server. I can't tell, and the documentation doesn't help too much.

Describe the solution you'd like

  1. Please take a moment to explain to me where models are being computed
  2. If it is as I suspect and everything happens local, then a small section in the documentation that explains this would be useful - I'm happy to write this and submit a PR
  3. Is it possible to use other computing resources? If so, how?
  4. If yes to above, then this should also be documented - again happy to write this and submit a PR

Describe alternatives you've considered

Metadata

Metadata

Assignees

No one assigned

    Labels

    documentationImprovements or additions to documentationenhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions