Chatbox AI app backend support #1047
                  
                    
                      magikRUKKOLA
                    
                  
                
                  started this conversation in
                Show and tell
              
            Replies: 1 comment 6 replies
-
| 
         Is there a way to fold the <think> produced by R1 series? My local deployment is too thinkable that its <think> is longer than its answer...  | 
  
Beta Was this translation helpful? Give feedback.
                  
                    6 replies
                  
                
            
  
    Sign up for free
    to join this conversation on GitHub.
    Already have an account?
    Sign in to comment
  
        
    
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
In case someone would want to use ktransformers (multiple ones, running at different machines) as a backend for apps like Chatbox AI that support Ollama API they could use the following.
[EDIT] (link to the latest version of nginx config updated)
chatboxai/chatbox#2221 (comment)
the picture below shows the usage of the llms from three different backends (ollama and 2 ktransformers).
Beta Was this translation helpful? Give feedback.
All reactions