mirror of
https://github.com/Kareadita/Kavita.git
synced 2025-05-24 00:52:23 -04:00
Sort Order Fix on Volume Detail Page (#3216)
This commit is contained in:
parent
ee253820f6
commit
0a13cc8454
@ -112,8 +112,18 @@ public class ScannerServiceTests : AbstractDbTest
|
||||
|
||||
Assert.NotNull(postLib);
|
||||
Assert.Equal(4, postLib.Series.Count);
|
||||
}
|
||||
|
||||
Assert.True(true);
|
||||
[Fact]
|
||||
public async Task ScanLibrary_ShouldCombineNestedFolder()
|
||||
{
|
||||
var testcase = "Series and Series-Series Combined - Manga.json";
|
||||
var postLib = await GenerateScannerData(testcase);
|
||||
|
||||
Assert.NotNull(postLib);
|
||||
Assert.Single(postLib.Series);
|
||||
Assert.Single(postLib.Series);
|
||||
Assert.Equal(2, postLib.Series.First().Volumes.Count);
|
||||
}
|
||||
|
||||
private async Task<Library> GenerateScannerData(string testcase)
|
||||
|
@ -0,0 +1,4 @@
|
||||
[
|
||||
"Dress Up Darling/Dress Up Darling Ch 01.cbz",
|
||||
"Dress Up Darling/Dress Up Darling/Dress Up Darling Vol 01.cbz"
|
||||
]
|
@ -127,9 +127,18 @@ public class VolumeRepository : IVolumeRepository
|
||||
|
||||
if (includeChapters)
|
||||
{
|
||||
query = query.Include(v => v.Chapters).AsSplitQuery();
|
||||
query = query
|
||||
.Includes(VolumeIncludes.Chapters)
|
||||
.AsSplitQuery();
|
||||
}
|
||||
return await query.ToListAsync();
|
||||
var volumes = await query.ToListAsync();
|
||||
|
||||
foreach (var volume in volumes)
|
||||
{
|
||||
volume.Chapters = volume.Chapters.OrderBy(c => c.SortOrder).ToList();
|
||||
}
|
||||
|
||||
return volumes;
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
@ -142,12 +151,11 @@ public class VolumeRepository : IVolumeRepository
|
||||
{
|
||||
var volume = await _context.Volume
|
||||
.Where(vol => vol.Id == volumeId)
|
||||
.Include(vol => vol.Chapters)
|
||||
.ThenInclude(c => c.Files)
|
||||
.Includes(VolumeIncludes.Chapters | VolumeIncludes.Files)
|
||||
.AsSplitQuery()
|
||||
.OrderBy(v => v.MinNumber)
|
||||
.ProjectTo<VolumeDto>(_mapper.ConfigurationProvider)
|
||||
.SingleOrDefaultAsync(vol => vol.Id == volumeId);
|
||||
.FirstOrDefaultAsync(vol => vol.Id == volumeId);
|
||||
|
||||
if (volume == null) return null;
|
||||
|
||||
@ -166,8 +174,7 @@ public class VolumeRepository : IVolumeRepository
|
||||
{
|
||||
return await _context.Volume
|
||||
.Where(vol => vol.SeriesId == seriesId)
|
||||
.Include(vol => vol.Chapters)
|
||||
.ThenInclude(c => c.Files)
|
||||
.Includes(VolumeIncludes.Chapters | VolumeIncludes.Files)
|
||||
.AsSplitQuery()
|
||||
.OrderBy(vol => vol.MinNumber)
|
||||
.ToListAsync();
|
||||
@ -205,24 +212,19 @@ public class VolumeRepository : IVolumeRepository
|
||||
|
||||
await AddVolumeModifiers(userId, volumes);
|
||||
|
||||
foreach (var volume in volumes)
|
||||
{
|
||||
volume.Chapters = volume.Chapters.OrderBy(c => c.SortOrder).ToList();
|
||||
}
|
||||
|
||||
return volumes;
|
||||
}
|
||||
|
||||
public async Task<Volume?> GetVolumeByIdAsync(int volumeId)
|
||||
{
|
||||
return await _context.Volume.SingleOrDefaultAsync(x => x.Id == volumeId);
|
||||
return await _context.Volume.FirstOrDefaultAsync(x => x.Id == volumeId);
|
||||
}
|
||||
|
||||
public async Task<IList<Volume>> GetAllWithCoversInDifferentEncoding(EncodeFormat encodeFormat)
|
||||
{
|
||||
var extension = encodeFormat.GetExtension();
|
||||
return await _context.Volume
|
||||
.Include(v => v.Chapters)
|
||||
.Includes(VolumeIncludes.Chapters)
|
||||
.Where(c => !string.IsNullOrEmpty(c.CoverImage) && !c.CoverImage.EndsWith(extension))
|
||||
.AsSplitQuery()
|
||||
.ToListAsync();
|
||||
|
@ -80,25 +80,25 @@ public static class IncludesExtensions
|
||||
if (includes.HasFlag(VolumeIncludes.Files))
|
||||
{
|
||||
queryable = queryable
|
||||
.Include(vol => vol.Chapters.OrderBy(c => c.SortOrder))
|
||||
.Include(vol => vol.Chapters)
|
||||
.ThenInclude(c => c.Files);
|
||||
} else if (includes.HasFlag(VolumeIncludes.Chapters))
|
||||
{
|
||||
queryable = queryable
|
||||
.Include(vol => vol.Chapters.OrderBy(c => c.SortOrder));
|
||||
.Include(vol => vol.Chapters);
|
||||
}
|
||||
|
||||
if (includes.HasFlag(VolumeIncludes.People))
|
||||
{
|
||||
queryable = queryable
|
||||
.Include(vol => vol.Chapters.OrderBy(c => c.SortOrder))
|
||||
.Include(vol => vol.Chapters)
|
||||
.ThenInclude(c => c.People);
|
||||
}
|
||||
|
||||
if (includes.HasFlag(VolumeIncludes.Tags))
|
||||
{
|
||||
queryable = queryable
|
||||
.Include(vol => vol.Chapters.OrderBy(c => c.SortOrder))
|
||||
.Include(vol => vol.Chapters)
|
||||
.ThenInclude(c => c.Tags);
|
||||
}
|
||||
|
||||
|
@ -51,7 +51,10 @@ public class AutoMapperProfiles : Profile
|
||||
.ForMember(dest => dest.Series, opt => opt.MapFrom(src => src.Series));
|
||||
CreateMap<LibraryDto, Library>();
|
||||
CreateMap<Volume, VolumeDto>()
|
||||
.ForMember(dest => dest.Number, opt => opt.MapFrom(src => (int) src.MinNumber));
|
||||
.ForMember(dest => dest.Number,
|
||||
opt => opt.MapFrom(src => (int) src.MinNumber))
|
||||
.ForMember(dest => dest.Chapters,
|
||||
opt => opt.MapFrom(src => src.Chapters.OrderBy(c => c.SortOrder)));
|
||||
CreateMap<MangaFile, MangaFileDto>();
|
||||
CreateMap<Series, SeriesDto>();
|
||||
CreateMap<CollectionTag, CollectionTagDto>();
|
||||
|
@ -7,6 +7,7 @@ using API.Data;
|
||||
using API.DTOs.Reader;
|
||||
using API.Entities;
|
||||
using API.Entities.Enums;
|
||||
using API.Extensions;
|
||||
using Hangfire;
|
||||
using Microsoft.Extensions.Logging;
|
||||
|
||||
@ -90,6 +91,13 @@ public class BookmarkService : IBookmarkService
|
||||
var bookmark = await _unitOfWork.UserRepository.GetBookmarkAsync(bookmarkId);
|
||||
if (bookmark == null) return;
|
||||
|
||||
// Validate the bookmark isn't already in target format
|
||||
if (bookmark.FileName.EndsWith(encodeFormat.GetExtension()))
|
||||
{
|
||||
// Nothing to ddo
|
||||
return;
|
||||
}
|
||||
|
||||
bookmark.FileName = await _mediaConversionService.SaveAsEncodingFormat(bookmarkDirectory, bookmark.FileName,
|
||||
BookmarkStem(bookmark.AppUserId, bookmark.SeriesId, bookmark.ChapterId), encodeFormat);
|
||||
_unitOfWork.UserRepository.Update(bookmark);
|
||||
@ -137,7 +145,7 @@ public class BookmarkService : IBookmarkService
|
||||
_unitOfWork.UserRepository.Add(bookmark);
|
||||
await _unitOfWork.CommitAsync();
|
||||
|
||||
if (settings.EncodeMediaAs == EncodeFormat.WEBP)
|
||||
if (settings.EncodeMediaAs != EncodeFormat.PNG)
|
||||
{
|
||||
// Enqueue a task to convert the bookmark to webP
|
||||
BackgroundJob.Enqueue(() => ConvertBookmarkToEncoding(bookmark.Id));
|
||||
|
@ -453,10 +453,6 @@ public class SeriesService : ISeriesService
|
||||
continue;
|
||||
}
|
||||
|
||||
volume.Chapters = volume.Chapters
|
||||
.OrderBy(d => d.MinNumber, ChapterSortComparerDefaultLast.Default)
|
||||
.ToList();
|
||||
|
||||
if (RenameVolumeName(volume, libraryType, volumeLabel) || (bookTreatment && !volume.IsSpecial()))
|
||||
{
|
||||
processedVolumes.Add(volume);
|
||||
|
@ -220,6 +220,7 @@ public class TaskScheduler : ITaskScheduler
|
||||
|
||||
public void AnalyzeFilesForLibrary(int libraryId, bool forceUpdate = false)
|
||||
{
|
||||
_logger.LogInformation("Enqueuing library file analysis for: {LibraryId}", libraryId);
|
||||
BackgroundJob.Enqueue(() => _wordCountAnalyzerService.ScanLibrary(libraryId, forceUpdate));
|
||||
}
|
||||
|
||||
|
@ -435,7 +435,15 @@ public class ParseScannedFiles
|
||||
{
|
||||
if (scannedSeries[series].Count <= 0) continue;
|
||||
|
||||
UpdateSortOrder(scannedSeries, series);
|
||||
try
|
||||
{
|
||||
UpdateSortOrder(scannedSeries, series);
|
||||
}
|
||||
catch (Exception ex)
|
||||
{
|
||||
_logger.LogError(ex, "There was an issue setting IssueOrder");
|
||||
}
|
||||
|
||||
|
||||
processedScannedSeries.Add(new ScannedSeriesResult()
|
||||
{
|
||||
@ -500,88 +508,90 @@ public class ParseScannedFiles
|
||||
}
|
||||
|
||||
|
||||
private void UpdateSortOrder(ConcurrentDictionary<ParsedSeries, List<ParserInfo>> scannedSeries, ParsedSeries series)
|
||||
public static void UpdateSortOrder(ConcurrentDictionary<ParsedSeries, List<ParserInfo>> scannedSeries, ParsedSeries series)
|
||||
{
|
||||
try
|
||||
// Set the Sort order per Volume
|
||||
var volumes = scannedSeries[series].GroupBy(info => info.Volumes);
|
||||
foreach (var volume in volumes)
|
||||
{
|
||||
// Set the Sort order per Volume
|
||||
var volumes = scannedSeries[series].GroupBy(info => info.Volumes);
|
||||
foreach (var volume in volumes)
|
||||
var infos = scannedSeries[series].Where(info => info.Volumes == volume.Key).ToList();
|
||||
IList<ParserInfo> chapters;
|
||||
var specialTreatment = infos.TrueForAll(info => info.IsSpecial);
|
||||
var hasAnySpMarker = infos.Exists(info => info.SpecialIndex > 0);
|
||||
var counter = 0f;
|
||||
|
||||
// Handle specials with SpecialIndex
|
||||
if (specialTreatment && hasAnySpMarker)
|
||||
{
|
||||
var infos = scannedSeries[series].Where(info => info.Volumes == volume.Key).ToList();
|
||||
IList<ParserInfo> chapters;
|
||||
var specialTreatment = infos.TrueForAll(info => info.IsSpecial);
|
||||
var hasAnySpMarker = infos.Exists(info => info.SpecialIndex > 0);
|
||||
var counter = 0f;
|
||||
|
||||
if (specialTreatment && hasAnySpMarker)
|
||||
{
|
||||
chapters = infos
|
||||
.OrderBy(info => info.SpecialIndex)
|
||||
.ToList();
|
||||
|
||||
foreach (var chapter in chapters)
|
||||
{
|
||||
chapter.IssueOrder = counter;
|
||||
counter++;
|
||||
}
|
||||
continue;
|
||||
}
|
||||
|
||||
|
||||
// If everything is a special but we don't have any SpecialIndex, then order naturally and use 0, 1, 2
|
||||
if (specialTreatment)
|
||||
{
|
||||
chapters = infos
|
||||
.OrderByNatural(info => Parser.Parser.RemoveExtensionIfSupported(info.Filename)!)
|
||||
.ToList();
|
||||
|
||||
foreach (var chapter in chapters)
|
||||
{
|
||||
chapter.IssueOrder = counter;
|
||||
counter++;
|
||||
}
|
||||
continue;
|
||||
}
|
||||
|
||||
chapters = infos
|
||||
.OrderByNatural(info => info.Chapters, StringComparer.InvariantCulture)
|
||||
.OrderBy(info => info.SpecialIndex)
|
||||
.ToList();
|
||||
|
||||
counter = 0f;
|
||||
var prevIssue = string.Empty;
|
||||
foreach (var chapter in chapters)
|
||||
{
|
||||
if (float.TryParse(chapter.Chapters, CultureInfo.InvariantCulture, out var parsedChapter))
|
||||
chapter.IssueOrder = counter;
|
||||
counter++;
|
||||
}
|
||||
continue;
|
||||
}
|
||||
|
||||
// Handle specials without SpecialIndex (natural order)
|
||||
if (specialTreatment)
|
||||
{
|
||||
chapters = infos
|
||||
.OrderByNatural(info => Parser.Parser.RemoveExtensionIfSupported(info.Filename)!)
|
||||
.ToList();
|
||||
|
||||
foreach (var chapter in chapters)
|
||||
{
|
||||
chapter.IssueOrder = counter;
|
||||
counter++;
|
||||
}
|
||||
continue;
|
||||
}
|
||||
|
||||
// Ensure chapters are sorted numerically when possible, otherwise push unparseable to the end
|
||||
chapters = infos
|
||||
.OrderBy(info => float.TryParse(info.Chapters, NumberStyles.Any, CultureInfo.InvariantCulture, out var val) ? val : float.MaxValue)
|
||||
.ToList();
|
||||
|
||||
counter = 0f;
|
||||
var prevIssue = string.Empty;
|
||||
foreach (var chapter in chapters)
|
||||
{
|
||||
if (float.TryParse(chapter.Chapters, NumberStyles.Any, CultureInfo.InvariantCulture, out var parsedChapter))
|
||||
{
|
||||
// Parsed successfully, use the numeric value
|
||||
counter = parsedChapter;
|
||||
chapter.IssueOrder = counter;
|
||||
|
||||
// Increment for next chapter (unless the next has a similar value, then add 0.1)
|
||||
if (!string.IsNullOrEmpty(prevIssue) && float.TryParse(prevIssue, CultureInfo.InvariantCulture, out var prevIssueFloat) && parsedChapter.Is(prevIssueFloat))
|
||||
{
|
||||
counter = parsedChapter;
|
||||
if (!string.IsNullOrEmpty(prevIssue) && float.TryParse(prevIssue, CultureInfo.InvariantCulture, out var prevIssueFloat) && parsedChapter.Is(prevIssueFloat))
|
||||
{
|
||||
// Bump by 0.1
|
||||
counter += 0.1f;
|
||||
}
|
||||
chapter.IssueOrder = counter;
|
||||
prevIssue = $"{parsedChapter}";
|
||||
counter += 0.1f; // bump if same value as the previous issue
|
||||
}
|
||||
else
|
||||
prevIssue = $"{parsedChapter}";
|
||||
}
|
||||
else
|
||||
{
|
||||
// Unparsed chapters: use the current counter and bump for the next
|
||||
if (!string.IsNullOrEmpty(prevIssue) && prevIssue == counter.ToString(CultureInfo.InvariantCulture))
|
||||
{
|
||||
// I need to bump by 0.1f as if the prevIssue matches counter
|
||||
if (!string.IsNullOrEmpty(prevIssue) && prevIssue == counter + "")
|
||||
{
|
||||
// Bump by 0.1
|
||||
counter += 0.1f;
|
||||
}
|
||||
chapter.IssueOrder = counter;
|
||||
counter++;
|
||||
prevIssue = chapter.Chapters;
|
||||
counter += 0.1f; // bump if same value as the previous issue
|
||||
}
|
||||
chapter.IssueOrder = counter;
|
||||
counter++;
|
||||
prevIssue = chapter.Chapters;
|
||||
}
|
||||
}
|
||||
}
|
||||
catch (Exception ex)
|
||||
{
|
||||
_logger.LogError(ex, "There was an issue setting IssueOrder");
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
private bool HasAllSeriesFolderNotChangedSinceLastScan(IList<SeriesModified> seriesFolders,
|
||||
string normalizedFolder)
|
||||
{
|
||||
return seriesFolders.All(f => HasSeriesFolderNotChangedSinceLastScan(f, normalizedFolder));
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
@ -603,12 +613,6 @@ public class ParseScannedFiles
|
||||
return false;
|
||||
}
|
||||
|
||||
private bool HasAllSeriesFolderNotChangedSinceLastScan(IList<SeriesModified> seriesFolders,
|
||||
string normalizedFolder)
|
||||
{
|
||||
return seriesFolders.All(f => HasSeriesFolderNotChangedSinceLastScan(f, normalizedFolder));
|
||||
}
|
||||
|
||||
private bool HasSeriesFolderNotChangedSinceLastScan(SeriesModified seriesModified, string normalizedFolder)
|
||||
{
|
||||
return seriesModified.LastScanned.Truncate(TimeSpan.TicksPerSecond) >=
|
||||
|
@ -812,7 +812,7 @@ export class MangaReaderComponent implements OnInit, AfterViewInit, OnDestroy {
|
||||
&& (this.readerService.imageUrlToChapterId(img.src) == chapterId || this.readerService.imageUrlToChapterId(img.src) === -1)
|
||||
);
|
||||
|
||||
console.log('Requesting page ', pageNum, ' found page: ', img, ' and app is requesting new image? ', forceNew);
|
||||
//console.log('Requesting page ', pageNum, ' found page: ', img, ' and app is requesting new image? ', forceNew);
|
||||
if (!img || forceNew) {
|
||||
img = new Image();
|
||||
img.src = this.getPageUrl(pageNum, chapterId);
|
||||
|
@ -118,7 +118,8 @@ export class GroupedTypeaheadComponent implements OnInit {
|
||||
|
||||
|
||||
@HostListener('window:click', ['$event'])
|
||||
handleDocumentClick(event: any) {
|
||||
handleDocumentClick(event: MouseEvent) {
|
||||
console.log('click: ', event)
|
||||
this.close();
|
||||
|
||||
}
|
||||
@ -197,7 +198,7 @@ export class GroupedTypeaheadComponent implements OnInit {
|
||||
}
|
||||
|
||||
toggleIncludeFiles(val: boolean) {
|
||||
const firstRun = val === false && val === this.includeChapterAndFiles;
|
||||
const firstRun = !val && val === this.includeChapterAndFiles;
|
||||
|
||||
this.includeChapterAndFiles = val;
|
||||
this.inputChanged.emit({value: this.searchTerm, includeFiles: this.includeChapterAndFiles});
|
||||
|
@ -666,7 +666,6 @@ export class VolumeDetailComponent implements OnInit {
|
||||
const chaptersWithProgress = this.volume.chapters.filter(c => c.pagesRead < c.pages);
|
||||
if (chaptersWithProgress.length > 0 && this.volume.chapters.length > 1) {
|
||||
this.currentlyReadingChapter = chaptersWithProgress[0];
|
||||
console.log('Updating currentlyReading chapter', this.currentlyReadingChapter)
|
||||
this.cdRef.markForCheck();
|
||||
} else {
|
||||
this.currentlyReadingChapter = undefined;
|
||||
|
Loading…
x
Reference in New Issue
Block a user