Skip to content

Commit 5c48f7e

Browse files
authored
Merge pull request #279657 from MicrosoftDocs/repo_sync_working_branch
Confirm merge from repo_sync_working_branch to main to sync with https://github.com/MicrosoftDocs/azure-docs (branch main)
2 parents 10f5197 + d9763e4 commit 5c48f7e

File tree

4 files changed

+10
-9
lines changed

4 files changed

+10
-9
lines changed

articles/aks/gpu-cluster.md

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -177,9 +177,10 @@ To use Azure Linux, you specify the OS SKU by setting `os-sku` to `AzureLinux` d
177177
name: nvidia-device-plugin-ds
178178
spec:
179179
tolerations:
180-
- key: nvidia.com/gpu
181-
operator: Exists
182-
effect: NoSchedule
180+
- key: "sku"
181+
operator: "Equal"
182+
value: "gpu"
183+
effect: "NoSchedule"
183184
# Mark this pod as a critical add-on; when enabled, the critical add-on
184185
# scheduler reserves resources for critical add-on pods so that they can
185186
# be rescheduled after a failure.

articles/aks/private-clusters.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -58,7 +58,7 @@ Create a private cluster with default basic networking using the [`az aks create
5858
```azurecli-interactive
5959
az aks create \
6060
--name <private-cluster-name> \
61-
--resource-group-name <private-cluster-resource-group> \
61+
--resource-group <private-cluster-resource-group> \
6262
--load-balancer-sku standard \
6363
--enable-private-cluster \
6464
--generate-ssh-keys

articles/azure-cache-for-redis/cache-troubleshoot-timeouts.md

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -145,11 +145,11 @@ There are several changes you can make to mitigate high server load:
145145

146146
- Investigate what is causing high server load such as [long-running commands](#long-running-commands), noted in this article, because of high memory pressure.
147147
- [Scale](cache-how-to-scale.md) out to more shards to distribute load across multiple Redis processes or scale up to a larger cache size with more CPU cores. For more information, see [Azure Cache for Redis planning FAQs](./cache-planning-faq.yml).
148-
- If your production workload on a _C1_ cache is negatively affected by extra latency from virus scanning, you can reduce the effect by to pay for a higher tier offering with multiple CPU cores, such as _C2_.
148+
- If your production workload on a _C1_ cache is negatively affected by extra latency from some internal defender scan runs, you can reduce the effect by scaling to a higher tier offering with multiple CPU cores, such as _C2_.
149149

150150
#### Spikes in server load
151151

152-
On _C0_ and _C1_ caches, you might see short spikes in server load not caused by an increase in requests a couple times a day while virus scanning is running on the VMs. You see higher latency for requests while virus scanning is happening on these tiers. Caches on the _C0_ and _C1_ tiers only have a single core to multitask, dividing the work of serving virus scanning and Redis requests.
152+
On _C0_ and _C1_ caches, you might see short spikes in server load not caused by an increase in requests a couple times a day while internal defender scanning is running on the VMs. You see higher latency for requests while internal defender scans happen on these tiers. Caches on the _C0_ and _C1_ tiers only have a single core to multitask, dividing the work of serving internal defender scanning and Redis requests.
153153

154154
### High memory usage
155155

articles/cosmos-db/ai-agents.md

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -396,7 +396,7 @@ from langchain_core.runnables.history import RunnableWithMessageHistory
396396
from langchain.agents import AgentExecutor, create_openai_tools_agent
397397
from service import TravelAgentTools as agent_tools
398398

399-
load_dotenv(override=True)
399+
load_dotenv(override=False)
400400

401401

402402
chat : ChatOpenAI | None=None
@@ -438,7 +438,7 @@ def LLM_init():
438438
LLM_init()
439439
```
440440

441-
The **init.py** file commences by initiating the loading of environment variables from a **.env** file utilizing the ```load_dotenv(override=True)``` method. Then, a global variable named ```agent_with_chat_history``` is instantiated for the agent, intended for use by our **TravelAgent.py**. The ```LLM_init()``` method is invoked during module initialization to configure our AI agent for conversation via the API web layer. The OpenAI Chat object is instantiated using the GPT-3.5 model, incorporating specific parameters such as model name and temperature. The chat object, tools list, and prompt template are combined to generate an ```AgentExecutor```, which operates as our AI Travel Agent. Lastly, the agent with history, ```agent_with_chat_history```, is established using ```RunnableWithMessageHistory``` with chat history (MongoDBChatMessageHistory), enabling it to maintain a complete conversation history via Azure Cosmos DB.
441+
The **init.py** file commences by initiating the loading of environment variables from a **.env** file utilizing the ```load_dotenv(override=False)``` method. Then, a global variable named ```agent_with_chat_history``` is instantiated for the agent, intended for use by our **TravelAgent.py**. The ```LLM_init()``` method is invoked during module initialization to configure our AI agent for conversation via the API web layer. The OpenAI Chat object is instantiated using the GPT-3.5 model, incorporating specific parameters such as model name and temperature. The chat object, tools list, and prompt template are combined to generate an ```AgentExecutor```, which operates as our AI Travel Agent. Lastly, the agent with history, ```agent_with_chat_history```, is established using ```RunnableWithMessageHistory``` with chat history (MongoDBChatMessageHistory), enabling it to maintain a complete conversation history via Azure Cosmos DB.
442442

443443
#### Prompt
444444

@@ -507,7 +507,7 @@ from model.prompt import PromptResponse
507507
import time
508508
from dotenv import load_dotenv
509509

510-
load_dotenv(override=True)
510+
load_dotenv(override=False)
511511

512512

513513
def agent_chat(input:str, session_id:str)->str:

0 commit comments

Comments
 (0)