20
20
get_search_result ,
21
21
)
22
22
23
- # from mcim_sync.models.database.file_cdn import File as FileCDN
24
23
from mcim_sync .utils .constans import ProjectDetail
25
24
from mcim_sync .utils .loger import log
26
25
from mcim_sync .utils .model_submitter import ModelSubmitter
33
32
config = Config .load ()
34
33
35
34
API = config .curseforge_api
36
- MAX_LENGTH = config .max_file_size
37
- MIN_DOWNLOAD_COUNT = 0
38
35
HEADERS = {
39
36
"x-api-key" : config .curseforge_api_key ,
40
37
"User-Agent" : "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/116.0.0.0 Safari/537.36 Edg/116.0.1938.54" ,
41
38
}
42
39
43
40
44
41
def append_model_from_files_res (
45
- res , latestFiles : List [dict ], need_to_cache : bool = True
42
+ res , latestFiles : List [dict ]
46
43
):
47
44
with ModelSubmitter () as submitter :
48
45
for file in res ["data" ]:
@@ -54,36 +51,11 @@ def append_model_from_files_res(
54
51
latestFiles = latestFiles , # type: ignore
55
52
)
56
53
)
57
- # file_sha1 = find_hash_in_curseforge_hashes(file["hashes"], 1)
58
-
59
- # if config.file_cdn:
60
- # if (
61
- # file_model.fileLength is not None
62
- # and file_model.downloadCount is not None
63
- # and file_model.downloadUrl is not None
64
- # and file_sha1
65
- # ):
66
- # if (
67
- # need_to_cache
68
- # and file_model.gameId == 432
69
- # and file_model.fileLength <= MAX_LENGTH
70
- # and file_model.downloadCount >= MIN_DOWNLOAD_COUNT
71
- # ):
72
- # submitter.add(
73
- # FileCDN(
74
- # sha1=file_sha1,
75
- # url=file_model.downloadUrl,
76
- # path=file_sha1,
77
- # size=file_model.fileLength,
78
- # mtime=int(time.time()),
79
- # ) # type: ignore
80
- # )
81
- # file_model.file_cdn_cached = True
82
54
submitter .add (file_model )
83
55
84
56
85
57
def sync_mod_all_files (
86
- modId : int , latestFiles : List [dict ], need_to_cache : bool = True
58
+ modId : int , latestFiles : List [dict ]
87
59
) -> int :
88
60
params = {"index" : 0 , "pageSize" : 50 }
89
61
file_id_list = []
@@ -93,7 +65,7 @@ def sync_mod_all_files(
93
65
while True :
94
66
res = get_mod_files (modId , params ["index" ], params ["pageSize" ])
95
67
append_model_from_files_res (
96
- res , latestFiles = latestFiles , need_to_cache = need_to_cache
68
+ res , latestFiles = latestFiles
97
69
)
98
70
file_id_list .extend ([file ["id" ] for file in res ["data" ]])
99
71
@@ -118,7 +90,7 @@ def sync_mod_all_files(
118
90
119
91
120
92
def sync_mod_all_files_at_once (
121
- modId : int , latestFiles : List [dict ], need_to_cache : bool = True
93
+ modId : int , latestFiles : List [dict ]
122
94
) -> Optional [int ]:
123
95
max_retries = 3
124
96
page_size = 10000
@@ -147,7 +119,7 @@ def sync_mod_all_files_at_once(
147
119
original_files_count = mongodb_engine .count (File , File .modId == modId )
148
120
149
121
append_model_from_files_res (
150
- res , latestFiles = latestFiles , need_to_cache = need_to_cache
122
+ res , latestFiles = latestFiles
151
123
)
152
124
153
125
removed_count = mongodb_engine .remove (
@@ -170,13 +142,11 @@ def sync_mod(modId: int) -> Optional[ProjectDetail]:
170
142
# version_count = sync_mod_all_files(
171
143
# modId,
172
144
# latestFiles=res["latestFiles"],
173
- # need_to_cache=True if res["classId"] == 6 else False,
174
145
# )
175
146
176
147
version_count = sync_mod_all_files_at_once (
177
148
modId ,
178
149
latestFiles = res ["latestFiles" ],
179
- need_to_cache = True if res ["classId" ] == 6 else False ,
180
150
)
181
151
182
152
if version_count is None :
0 commit comments