You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository was archived by the owner on Oct 10, 2025. It is now read-only.
Copy file name to clipboardExpand all lines: README.md
+228-2Lines changed: 228 additions & 2 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -52,7 +52,233 @@ the AWS Well-Architected Framework. This solution uses the following AWS CDK Sol
52
52
53
53
You can launch this solution with one click from [AWS Solutions Implementations](https://aws.amazon.com/solutions/implementations/maintaining-personalized-experiences-with-ml).
54
54
55
-
To customize the solution, or to contribute to the solution, follow the steps below:
55
+
To customize the solution, or to contribute to the solution, see [Creating a custom build](#creating-a-custom-build)
56
+
57
+
## Configuration
58
+
59
+
This solution uses **parameter files**. The parameter file contains all the necessary information to create and maintain
60
+
your resources in Amazon Personalize.
61
+
62
+
The file can contain the following sections
63
+
-`datasetGroup`
64
+
-`datasets`
65
+
-`solutions` (can contain `campaigns` and `batchInferenceJobs`)
66
+
-`eventTracker`
67
+
-`filters`
68
+
69
+
<details>
70
+
<summary>See a sample of the parameter file</summary>
"filterExpression": "INCLUDE ItemID WHERE Interactions.EVENT_TYPE in (\"click\", \"stream\")"
218
+
}
219
+
},
220
+
{
221
+
"serviceConfig": {
222
+
"name": "interacted",
223
+
"filterExpression": "INCLUDE ItemID WHERE Interactions.EVENT_TYPE in (\"*\")"
224
+
}
225
+
}
226
+
]
227
+
}
228
+
```
229
+
230
+
</details>
231
+
232
+
This solution allows you to manage multiple dataset groups through the use of multiple parameter files. All .json files
233
+
discovered under the `train/` prefix will trigger the workflow however, the following structure is recommended:
234
+
235
+
```
236
+
train/
237
+
│
238
+
├── <dataset_group_1>/ (option 1 - single csv files for data import)
239
+
│ ├── config.json
240
+
│ ├── interactions.csv
241
+
│ ├── items.csv (optional)
242
+
│ └── users.csv (optional)
243
+
│
244
+
└── <dataset_group_2>/ (option 2 - multiple csv files for data import)
245
+
├── config.json
246
+
├── interactions/
247
+
│ ├── <interactions_part_1>.csv
248
+
│ ├── <interactions_part_2>.csv
249
+
│ └── <interactions_part_n>.csv
250
+
├── users/ (optional)
251
+
│ ├── <users_part_1>.csv
252
+
│ ├── <users_part_2>.csv
253
+
│ └── <users_part_n>.csv
254
+
└── items/ (optional)
255
+
├── <items_part_1>.csv
256
+
├── <items_part_2>.csv
257
+
└── <items_part_n>.csv
258
+
```
259
+
260
+
If batch inference jobs are required, [batch inference job configuration files](https://docs.aws.amazon.com/personalize/latest/dg/recommendations-batch.html#batch-data-upload)
261
+
must also be uploaded to the following lcoation:
262
+
263
+
```
264
+
batch/
265
+
│
266
+
└── <dataset_group_name>/
267
+
└── <solution_name>/
268
+
└── job_config.json
269
+
```
270
+
271
+
Batch inference output will be produced at the following location:
272
+
273
+
```
274
+
batch/
275
+
│
276
+
└── <dataset_group_name>/
277
+
└── <solution_name>/
278
+
└── <solution_name_YYYY_MM_DD_HH_MM_SS>/
279
+
├── _CHECK
280
+
└── job_config.json.out
281
+
```
56
282
57
283
## Creating a custom build
58
284
To customize the solution, follow the steps below:
@@ -134,7 +360,7 @@ build-s3-cdk-dist \
134
360
S3 bucket where the name is `<DIST_BUCKET_PREFIX>-<REGION_NAME>`. The solution's CloudFormation template will expect the
135
361
source code to be located in the bucket matching that name.
136
362
-`$SOLUTION_NAME` - The name of This solution (example: personalize-solution-customization)
137
-
-`$VERSION` - The version number to use (example: v1.0.0)
363
+
-`$VERSION` - The version number to use (example: v1.0.1)
138
364
-`$REGION_NAME` - The region name to use (example: us-east-1)
139
365
140
366
This will result in all global assets being pushed to the `DIST_BUCKET_PREFIX`, and all regional assets being pushed to
0 commit comments