File tree Expand file tree Collapse file tree 1 file changed +2
-2
lines changed
articles/ai-foundry/foundry-local/includes/use-langchain Expand file tree Collapse file tree 1 file changed +2
-2
lines changed Original file line number Diff line number Diff line change @@ -36,15 +36,15 @@ import { ChatPromptTemplate } from "@langchain/core/prompts";
36
36
// to your end-user's device.
37
37
// TIP: You can find a list of available models by running the
38
38
// following command in your terminal: `foundry model list`.
39
- const modelAlias = " phi-3-mini-4k" ;
39
+ const alias = " phi-3-mini-4k" ;
40
40
41
41
// Create a FoundryLocalManager instance. This will start the Foundry
42
42
// Local service if it is not already running.
43
43
const foundryLocalManager = new FoundryLocalManager ()
44
44
45
45
// Initialize the manager with a model. This will download the model
46
46
// if it is not already present on the user's device.
47
- const modelInfo = await foundryLocalManager .init (modelAlias )
47
+ const modelInfo = await foundryLocalManager .init (alias )
48
48
console .log (" Model Info:" , modelInfo)
49
49
50
50
// Configure ChatOpenAI to use your locally-running model
You can’t perform that action at this time.
0 commit comments