1+ {
2+ "recipe_id" : " cpu_inference" ,
3+ "recipe_mode" : " service" ,
4+ "deployment_name" : " Cpu Inference A2" ,
5+ "recipe_image_uri" : " ghcr.io/amperecomputingai/ollama-ampere:1.0.0-ol9" ,
6+ "recipe_node_shape" : " VM.Standard.A2.Flex" ,
7+ "recipe_flex_shape_ocpu_count" : 8 ,
8+ "recipe_flex_shape_memory_size_in_gbs" : 64 ,
9+ "input_object_storage" : [
10+ {
11+ "par" : " https://objectstorage.us-ashburn-1.oraclecloud.com/p/PhxpLO7eu4bDXCB_wshp9jJXsrwxFTAJTEDY6it6tmDdEqIJf-CcZMugp3jfHnb5/n/axusxi89ppdg/b/blueprints-models/o/" ,
12+ "mount_location" : " /models" ,
13+ "volume_size_in_gbs" : 20
14+ }
15+ ],
16+ "recipe_container_env" : [
17+ {
18+ "key" : " OLLAMA_MODELS" ,
19+ "value" : " /models"
20+ },
21+ {
22+ "key" : " MODEL_NAME" ,
23+ "value" : " llama3.1:8b-instruct-q8_0"
24+ },
25+ {
26+ "key" : " PROMPT" ,
27+ "value" : " What is the capital of Spain?"
28+ }
29+ ],
30+ "recipe_replica_count" : 1 ,
31+ "recipe_container_port" : " 11434" ,
32+ "recipe_node_pool_size" : 1 ,
33+ "recipe_node_boot_volume_size_in_gbs" : 200 ,
34+ "recipe_ephemeral_storage_size" : 100
35+ }
0 commit comments