Added a new method for triggering jobs rather than scheduling ad-hoc.

This commit is contained in:
Joseph Milazzo 2021-03-14 09:39:48 -05:00
parent a7516d313d
commit d23a569446
4 changed files with 51 additions and 16 deletions

View File

@ -1,18 +1,50 @@
using Xunit; using System.Collections.Generic;
using System.IO;
using API.Data;
using API.Entities;
using API.Interfaces;
using API.Interfaces.Services;
using API.Services;
using Microsoft.Extensions.Logging;
using NSubstitute;
using Xunit;
namespace API.Tests.Services namespace API.Tests.Services
{ {
public class CacheServiceTests public class CacheServiceTests
{ {
// private readonly CacheService _cacheService; private readonly CacheService _cacheService;
// private readonly ILogger<CacheService> _logger = Substitute.For<ILogger<CacheService>>(); private readonly ILogger<CacheService> _logger = Substitute.For<ILogger<CacheService>>();
// private readonly IUnitOfWork _unitOfWork = Substitute.For<IUnitOfWork>(); private readonly IUnitOfWork _unitOfWork = Substitute.For<IUnitOfWork>();
// private readonly IArchiveService _archiveService = Substitute.For<IArchiveService>(); private readonly IArchiveService _archiveService = Substitute.For<IArchiveService>();
// private readonly IDirectoryService _directoryService = Substitute.For<DirectoryService>(); private readonly IDirectoryService _directoryService = Substitute.For<DirectoryService>();
// public CacheServiceTests() public CacheServiceTests()
{
_cacheService = new CacheService(_logger, _unitOfWork, _archiveService, _directoryService);
}
// [Fact]
// public async void Ensure_ShouldExtractArchive(int chapterId)
// { // {
// //_cacheService = new CacheService(_logger, _unitOfWork, _archiveService, _directoryService); //
// // CacheDirectory needs to be customized.
// _unitOfWork.VolumeRepository.GetChapterAsync(chapterId).Returns(new Chapter
// {
// Id = 1,
// Files = new List<MangaFile>()
// {
// new MangaFile()
// {
// FilePath = ""
// }
// }
// });
//
// await _cacheService.Ensure(1);
//
// var testDirectory = Path.Join(Directory.GetCurrentDirectory(), "../../../Services/Test Data/CacheService/Archives");
//
// } // }
//string GetCachedPagePath(Volume volume, int page) //string GetCachedPagePath(Volume volume, int page)

View File

@ -1,4 +1,5 @@
using System.IO; using System.IO;
using System.Linq;
using System.Threading.Tasks; using System.Threading.Tasks;
using API.Entities.Enums; using API.Entities.Enums;
using API.Helpers.Converters; using API.Helpers.Converters;
@ -53,33 +54,34 @@ namespace API.Services
if (setting != null) if (setting != null)
{ {
_logger.LogDebug("Scheduling Scan Library Task for {Cron}", setting); _logger.LogDebug("Scheduling Scan Library Task for {Cron}", setting);
RecurringJob.AddOrUpdate(() => _scannerService.ScanLibraries(), () => CronConverter.ConvertToCronNotation(setting)); RecurringJob.AddOrUpdate("scan-libraries", () => _scannerService.ScanLibraries(), () => CronConverter.ConvertToCronNotation(setting));
} }
else else
{ {
RecurringJob.AddOrUpdate(() => _scannerService.ScanLibraries(), Cron.Daily); RecurringJob.AddOrUpdate("scan-libraries", () => _scannerService.ScanLibraries(), Cron.Daily);
} }
setting = Task.Run(() => _unitOfWork.SettingsRepository.GetSettingAsync(ServerSettingKey.TaskBackup)).Result.Value; setting = Task.Run(() => _unitOfWork.SettingsRepository.GetSettingAsync(ServerSettingKey.TaskBackup)).Result.Value;
if (setting != null) if (setting != null)
{ {
_logger.LogDebug("Scheduling Backup Task for {Cron}", setting); _logger.LogDebug("Scheduling Backup Task for {Cron}", setting);
RecurringJob.AddOrUpdate(() => _backupService.BackupDatabase(), () => CronConverter.ConvertToCronNotation(setting)); RecurringJob.AddOrUpdate("backup", () => _backupService.BackupDatabase(), () => CronConverter.ConvertToCronNotation(setting));
} }
else else
{ {
RecurringJob.AddOrUpdate(() => _backupService.BackupDatabase(), Cron.Weekly); RecurringJob.AddOrUpdate("backup", () => _backupService.BackupDatabase(), Cron.Weekly);
} }
RecurringJob.AddOrUpdate(() => _cleanupService.Cleanup(), Cron.Daily); RecurringJob.AddOrUpdate("cleanup", () => _cleanupService.Cleanup(), Cron.Daily);
} }
public void ScanLibrary(int libraryId, bool forceUpdate = false) public void ScanLibrary(int libraryId, bool forceUpdate = false)
{ {
_logger.LogInformation("Enqueuing library scan for: {LibraryId}", libraryId); _logger.LogInformation("Enqueuing library scan for: {LibraryId}", libraryId);
BackgroundJob.Enqueue(() => _scannerService.ScanLibrary(libraryId, forceUpdate)); BackgroundJob.Enqueue(() => _scannerService.ScanLibrary(libraryId, forceUpdate));
BackgroundJob.Enqueue(() => _cleanupService.Cleanup()); // When we do a scan, force cache to re-unpack in case page numbers change //BackgroundJob.Enqueue(() => _cleanupService.Cleanup()); // When we do a scan, force cache to re-unpack in case page numbers change
RecurringJob.Trigger("cleanup"); // TODO: Alternate way to trigger jobs. Test this out and see if we should switch.
} }
public void CleanupChapters(int[] chapterIds) public void CleanupChapters(int[] chapterIds)

View File

@ -63,7 +63,8 @@ namespace API.Services.Tasks
_scannedSeries = null; _scannedSeries = null;
} }
//[DisableConcurrentExecution(timeoutInSeconds: 360)] [DisableConcurrentExecution(5)]
[AutomaticRetry(Attempts = 0, LogEvents = false, OnAttemptsExceeded = AttemptsExceededAction.Delete)]
public void ScanLibrary(int libraryId, bool forceUpdate) public void ScanLibrary(int libraryId, bool forceUpdate)
{ {
_forceUpdate = forceUpdate; _forceUpdate = forceUpdate;