Replies: 1 comment 4 replies
-
Here is llama.cpp: an llm serving framework and http server written in c++ and licensed by MIT. |
Beta Was this translation helpful? Give feedback.
4 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I know Llama 3 is open source, but when I go to websites like replicate, it shows a price per 1 million tokens if I want to use that API. So I don't think I fully understand the concept. I am trying to develop an application that uses Llama 3, but I want to grasp the foundational knowledge first. Could someone explain this to me? And what kind of costs would be involved if I want to use llama for my application? (after calculation, per user is expected to use 15-20k tokens per day)
Beta Was this translation helpful? Give feedback.
All reactions