1+ import asyncio
12import os
3+ from secrets import token_urlsafe
24
3- from bson import ObjectId
4- from faker import Faker
5- from dotenv import dotenv_values , load_dotenv
5+ import nest_asyncio
66import requests
7- import asyncio
8- from pymongo import MongoClient
97from app .config import settings
10- from itsdangerous .url_safe import URLSafeSerializer
11- from secrets import token_urlsafe
8+ from app .models .files import FileDB , FileDBViewList , FileOut , LocalFileIn , StorageType
129from app .routers .files import add_file_entry
13- from app .models .files import (
14- FileOut ,
15- FileDB ,
16- FileDBViewList ,
17- LocalFileIn ,
18- StorageType ,
19- )
10+ from bson import ObjectId
11+ from dotenv import dotenv_values , load_dotenv
12+ from faker import Faker
13+ from itsdangerous .url_safe import URLSafeSerializer
2014from minio import Minio
21- import nest_asyncio
15+ from pymongo import MongoClient
2216
2317nest_asyncio .apply ()
24- from app .models .users import UserDB , UserOut , UserAPIKeyDB
18+ from app import dependencies
19+ from app .models .users import UserAPIKeyDB , UserDB , UserOut
2520from beanie import init_beanie
26- from fastapi import FastAPI , APIRouter , Depends
2721from elasticsearch import Elasticsearch
28- from app import dependencies
22+ from fastapi import APIRouter , Depends , FastAPI
2923from motor .motor_asyncio import AsyncIOMotorClient
3024from pika .adapters .blocking_connection import BlockingChannel
3125
@@ -370,10 +364,10 @@ async def process_user(
370364 current_headers = user_headers_v2 ,
371365 )
372366 dataset_files_endpoint = (
373- CLOWDER_V1
374- + "api/datasets/"
375- + dataset ["id" ]
376- + "/files?=superAdmin=true"
367+ CLOWDER_V1
368+ + "api/datasets/"
369+ + dataset ["id" ]
370+ + "/files?=superAdmin=true"
377371 )
378372 r_files = requests .get (
379373 dataset_files_endpoint , headers = clowder_headers_v1 , verify = False
@@ -387,12 +381,10 @@ async def process_user(
387381 file_folder = file ["folders" ]
388382 # TODO download the file from v1 using api routes
389383 v1_download_url = (
390- CLOWDER_V1 + "api/files/" + file_id + "?superAdmin=true"
384+ CLOWDER_V1 + "api/files/" + file_id + "?superAdmin=true"
391385 )
392386 print ("downloading file" , filename )
393- download = requests .get (
394- v1_download_url , headers = clowder_headers_v1
395- )
387+ download = requests .get (v1_download_url , headers = clowder_headers_v1 )
396388 with open (filename , "wb" ) as f :
397389 f .write (download .content )
398390 file_data = {"file" : open (filename , "rb" )}
@@ -404,11 +396,11 @@ async def process_user(
404396 if matching_folder :
405397 upload_files = {"files" : open (filename , "rb" )}
406398 dataset_file_upload_endpoint = (
407- CLOWDER_V2
408- + "api/v2/datasets/"
409- + dataset_v2_id
410- + "/filesMultiple?folder_id="
411- + matching_folder ["id" ]
399+ CLOWDER_V2
400+ + "api/v2/datasets/"
401+ + dataset_v2_id
402+ + "/filesMultiple?folder_id="
403+ + matching_folder ["id" ]
412404 )
413405 response = requests .post (
414406 dataset_file_upload_endpoint ,
@@ -418,10 +410,7 @@ async def process_user(
418410
419411 else :
420412 dataset_file_upload_endpoint = (
421- CLOWDER_V2
422- + "api/v2/datasets/"
423- + dataset_v2_id
424- + "/files"
413+ CLOWDER_V2 + "api/v2/datasets/" + dataset_v2_id + "/files"
425414 )
426415 response = requests .post (
427416 dataset_file_upload_endpoint ,
0 commit comments