You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardExpand all lines: README.md
+23Lines changed: 23 additions & 0 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -112,6 +112,29 @@ kubectl get svc -n rayserve-vllm
112
112
113
113
This solution supports building agentic AI applications that can leverage either CPU-based (llama.cpp) or GPU-based (vLLM) model inference backends. The agent architecture enables models to call external functions and services.
114
114
115
+
### Understanding Agentic AI and Function Calling
116
+
117
+
Agentic AI refers to AI systems that can act autonomously to achieve specific goals by making decisions and taking actions. In this solution, we implement agentic capabilities through function calling, which allows language models to:
118
+
119
+
1.**Recognize when to use tools**: The model identifies when external data or capabilities are needed to fulfill a user request
120
+
2.**Structure function calls**: The model generates properly formatted function calls with appropriate parameters
121
+
3.**Process function results**: The model incorporates returned data into its responses
122
+
123
+
Function calling enables models to bridge the gap between natural language understanding and external systems, allowing them to:
124
+
125
+
- Access real-time information (like weather data in our example)
126
+
- Perform calculations or data transformations
127
+
- Interact with external APIs and services
128
+
- Execute specific actions based on user requests
129
+
130
+
Our implementation provides a framework where the model:
131
+
- Parses user intent from natural language
132
+
- Determines which function to call and with what parameters
133
+
- Makes the API call through a dedicated service
134
+
- Processes the returned information to generate a coherent response
135
+
136
+
This approach significantly extends the capabilities of language models beyond their pre-trained knowledge, making them more useful for real-world applications.
0 commit comments