- fix tests

This commit is contained in:
tiefseetauchner 2024-08-16 13:25:50 +02:00
parent d71863e35e
commit 5dbff72ca7
2 changed files with 8 additions and 7 deletions

View File

@ -190,11 +190,12 @@ describe('checkForDuplicates', () => {
}); });
}); });
it('throws error when check duplicates retry is failed', async () => { it('returns results when check duplicates retry is failed', async () => {
vi.mocked(checkBulkUpload).mockRejectedValue(new Error('Network error')); vi.mocked(checkBulkUpload).mockRejectedValue(new Error('Network error'));
await expect(checkForDuplicates([testFilePath], { concurrency: 1 })).rejects.toThrowError( await expect(checkForDuplicates([testFilePath], { concurrency: 1 })).resolves.toEqual({
'An error occurred while checking for duplicates: Network error', duplicates: [],
); newFiles: [],
});
}); });
}); });

View File

@ -101,11 +101,12 @@ export const checkForDuplicates = async (files: string[], { concurrency, skipHas
const newFiles: string[] = []; const newFiles: string[] = [];
const duplicates: Asset[] = []; const duplicates: Asset[] = [];
const checkBulkUploadQueue = new Queue<AssetBulkUploadCheckResults, any>( const checkBulkUploadQueue = new Queue<AssetBulkUploadCheckResults, void>(
async (assets: AssetBulkUploadCheckResults) => { async (assets: AssetBulkUploadCheckResults) => {
const response = await checkBulkUpload({ assetBulkUploadCheckDto: { assets } }); const response = await checkBulkUpload({ assetBulkUploadCheckDto: { assets } });
const results = response.results as AssetBulkUploadCheckResults; const results = response.results as AssetBulkUploadCheckResults;
for (const { id: filepath, assetId, action } of results) { for (const { id: filepath, assetId, action } of results) {
if (action === Action.Accept) { if (action === Action.Accept) {
newFiles.push(filepath); newFiles.push(filepath);
@ -116,7 +117,6 @@ export const checkForDuplicates = async (files: string[], { concurrency, skipHas
} }
checkProgressBar.increment(assets.length); checkProgressBar.increment(assets.length);
return { message: 'success' };
}, },
{ concurrency, retry: 3 }, { concurrency, retry: 3 },
); );
@ -130,7 +130,7 @@ export const checkForDuplicates = async (files: string[], { concurrency, skipHas
results.push(dto); results.push(dto);
checkBulkUploadRequests.push(dto); checkBulkUploadRequests.push(dto);
if (checkBulkUploadRequests.length >= concurrency) { if (checkBulkUploadRequests.length > concurrency) {
void checkBulkUploadQueue.push([...checkBulkUploadRequests]); void checkBulkUploadQueue.push([...checkBulkUploadRequests]);
checkBulkUploadRequests = []; checkBulkUploadRequests = [];
} }