@@ -572,10 +572,6 @@ production: &base
572
572
geo_migrated_local_files_clean_up_worker :
573
573
cron : " 15 */6 * * *"
574
574
575
- # Export pseudonymized data in CSV format for analysis
576
- pseudonymizer_worker :
577
- cron : " 0 * * * *"
578
-
579
575
# Elasticsearch bulk updater for incremental updates.
580
576
# NOTE: This will only take effect if elasticsearch is enabled.
581
577
elastic_index_bulk_cron_worker :
@@ -1154,20 +1150,6 @@ production: &base
1154
1150
remote_directory : ' {{GCS_BACKUP_BUCKET}}'
1155
1151
# end-gcs
1156
1152
1157
- # # Pseudonymizer exporter
1158
- pseudonymizer :
1159
- # Tables manifest that specifies the fields to extract and pseudonymize.
1160
- manifest : config/pseudonymizer.yml
1161
- upload :
1162
- remote_directory : ' gitlab-elt'
1163
- # Fog storage connection settings, see http://fog.io/storage/ .
1164
- connection :
1165
- # provider: AWS
1166
- # region: eu-west-1
1167
- # aws_access_key_id: AKIAKIAKI
1168
- # aws_secret_access_key: 'secret123'
1169
- # # The remote 'directory' to store the CSV files. For S3, this would be the bucket name.
1170
-
1171
1153
# # GitLab Shell settings
1172
1154
gitlab_shell :
1173
1155
path : {{GITLAB_SHELL_INSTALL_DIR}}/
@@ -1416,17 +1398,6 @@ test:
1416
1398
secret_file : tmp/gitlab_workhorse_test_secret
1417
1399
backup :
1418
1400
path : tmp/tests/backups
1419
- pseudonymizer :
1420
- manifest : config/pseudonymizer.yml
1421
- upload :
1422
- # The remote 'directory' to store the CSV files. For S3, this would be the bucket name.
1423
- remote_directory : gitlab-elt.test
1424
- # Fog storage connection settings, see http://fog.io/storage/
1425
- connection :
1426
- provider : AWS # Only AWS supported at the moment
1427
- aws_access_key_id : AWS_ACCESS_KEY_ID
1428
- aws_secret_access_key : AWS_SECRET_ACCESS_KEY
1429
- region : us-east-1
1430
1401
gitlab_shell :
1431
1402
path : tmp/tests/gitlab-shell/
1432
1403
authorized_keys_file : tmp/tests/authorized_keys
0 commit comments