Skip to content

Commit 074154e

Browse files
committed
formatting
1 parent 851503e commit 074154e

File tree

1 file changed

+24
-35
lines changed

1 file changed

+24
-35
lines changed

scripts/migration/migrate_users.py

Lines changed: 24 additions & 35 deletions
Original file line numberDiff line numberDiff line change
@@ -1,31 +1,25 @@
1+
import asyncio
12
import os
3+
from secrets import token_urlsafe
24

3-
from bson import ObjectId
4-
from faker import Faker
5-
from dotenv import dotenv_values, load_dotenv
5+
import nest_asyncio
66
import requests
7-
import asyncio
8-
from pymongo import MongoClient
97
from app.config import settings
10-
from itsdangerous.url_safe import URLSafeSerializer
11-
from secrets import token_urlsafe
8+
from app.models.files import FileDB, FileDBViewList, FileOut, LocalFileIn, StorageType
129
from app.routers.files import add_file_entry
13-
from app.models.files import (
14-
FileOut,
15-
FileDB,
16-
FileDBViewList,
17-
LocalFileIn,
18-
StorageType,
19-
)
10+
from bson import ObjectId
11+
from dotenv import dotenv_values, load_dotenv
12+
from faker import Faker
13+
from itsdangerous.url_safe import URLSafeSerializer
2014
from minio import Minio
21-
import nest_asyncio
15+
from pymongo import MongoClient
2216

2317
nest_asyncio.apply()
24-
from app.models.users import UserDB, UserOut, UserAPIKeyDB
18+
from app import dependencies
19+
from app.models.users import UserAPIKeyDB, UserDB, UserOut
2520
from beanie import init_beanie
26-
from fastapi import FastAPI, APIRouter, Depends
2721
from elasticsearch import Elasticsearch
28-
from app import dependencies
22+
from fastapi import APIRouter, Depends, FastAPI
2923
from motor.motor_asyncio import AsyncIOMotorClient
3024
from pika.adapters.blocking_connection import BlockingChannel
3125

@@ -370,10 +364,10 @@ async def process_user(
370364
current_headers=user_headers_v2,
371365
)
372366
dataset_files_endpoint = (
373-
CLOWDER_V1
374-
+ "api/datasets/"
375-
+ dataset["id"]
376-
+ "/files?=superAdmin=true"
367+
CLOWDER_V1
368+
+ "api/datasets/"
369+
+ dataset["id"]
370+
+ "/files?=superAdmin=true"
377371
)
378372
r_files = requests.get(
379373
dataset_files_endpoint, headers=clowder_headers_v1, verify=False
@@ -387,12 +381,10 @@ async def process_user(
387381
file_folder = file["folders"]
388382
# TODO download the file from v1 using api routes
389383
v1_download_url = (
390-
CLOWDER_V1 + "api/files/" + file_id + "?superAdmin=true"
384+
CLOWDER_V1 + "api/files/" + file_id + "?superAdmin=true"
391385
)
392386
print("downloading file", filename)
393-
download = requests.get(
394-
v1_download_url, headers=clowder_headers_v1
395-
)
387+
download = requests.get(v1_download_url, headers=clowder_headers_v1)
396388
with open(filename, "wb") as f:
397389
f.write(download.content)
398390
file_data = {"file": open(filename, "rb")}
@@ -404,11 +396,11 @@ async def process_user(
404396
if matching_folder:
405397
upload_files = {"files": open(filename, "rb")}
406398
dataset_file_upload_endpoint = (
407-
CLOWDER_V2
408-
+ "api/v2/datasets/"
409-
+ dataset_v2_id
410-
+ "/filesMultiple?folder_id="
411-
+ matching_folder["id"]
399+
CLOWDER_V2
400+
+ "api/v2/datasets/"
401+
+ dataset_v2_id
402+
+ "/filesMultiple?folder_id="
403+
+ matching_folder["id"]
412404
)
413405
response = requests.post(
414406
dataset_file_upload_endpoint,
@@ -418,10 +410,7 @@ async def process_user(
418410

419411
else:
420412
dataset_file_upload_endpoint = (
421-
CLOWDER_V2
422-
+ "api/v2/datasets/"
423-
+ dataset_v2_id
424-
+ "/files"
413+
CLOWDER_V2 + "api/v2/datasets/" + dataset_v2_id + "/files"
425414
)
426415
response = requests.post(
427416
dataset_file_upload_endpoint,

0 commit comments

Comments
 (0)