RED-4928: Changed licence generation to have the database filter the data by the dossier-ids and dates from the request

This commit is contained in:
Viktor Seifert 2022-08-12 18:56:13 +02:00
parent 0ad7592dd3
commit ca218e7bb3
4 changed files with 132 additions and 113 deletions

View File

@ -228,6 +228,10 @@ public class FileStatusPersistenceService {
return fileRepository.findByDossierId(dossierId);
}
public List<FileEntity> getStatusesForDossiersAndTimePeriod(Set<String> dossierIds, OffsetDateTime start, OffsetDateTime end) {
return fileRepository.findByDossierIdInAndAddedBetween(dossierIds, start, end);
}
public List<FileEntity> getActiveFiles(String dossierId) {

View File

@ -1,17 +1,19 @@
package com.iqser.red.service.persistence.management.v1.processor.service.persistence.repository;
import java.time.OffsetDateTime;
import java.util.List;
import java.util.Set;
import org.springframework.data.jpa.repository.JpaRepository;
import org.springframework.data.jpa.repository.Modifying;
import org.springframework.data.jpa.repository.Query;
import com.iqser.red.service.persistence.management.v1.processor.entity.dossier.FileEntity;
import com.iqser.red.service.persistence.management.v1.processor.service.persistence.projection.FilePageCountsProjection;
import com.iqser.red.service.persistence.management.v1.processor.service.persistence.projection.FileProcessingStatusProjection;
import com.iqser.red.service.persistence.management.v1.processor.service.persistence.projection.FileWorkflowStatusProjection;
import com.iqser.red.service.persistence.service.v1.api.model.dossiertemplate.dossier.file.ProcessingStatus;
import com.iqser.red.service.persistence.service.v1.api.model.dossiertemplate.dossier.file.WorkflowStatus;
import org.springframework.data.jpa.repository.JpaRepository;
import org.springframework.data.jpa.repository.Modifying;
import org.springframework.data.jpa.repository.Query;
import java.time.OffsetDateTime;
import java.util.List;
public interface FileRepository extends JpaRepository<FileEntity, String> {
@ -20,6 +22,8 @@ public interface FileRepository extends JpaRepository<FileEntity, String> {
List<FileEntity> findByDossierId(String dossierId);
List<FileEntity> findByDossierIdInAndAddedBetween(Set<String> dossierIds, OffsetDateTime start, OffsetDateTime end);
@Modifying
@Query("update FileEntity e set e.hasRedactions = :hasRedactions ," +

View File

@ -82,10 +82,7 @@ public class FileStatusService {
var fileEntities = fileStatusPersistenceService.getAllRelevantStatusesForReanalysisScheduler(fileManagementServiceSettings.getMaxErrorRetries());
var convertedList = convert(fileEntities, FileModel.class, new FileModelMapper());
return reanalysisRequiredStatusService.enhanceFileStatusWithAnalysisRequirements(convertedList)
.stream()
.filter(FileModel::isAnalysisRequired)
.collect(Collectors.toList());
return reanalysisRequiredStatusService.enhanceFileStatusWithAnalysisRequirements(convertedList).stream().filter(FileModel::isAnalysisRequired).collect(Collectors.toList());
}
@ -107,6 +104,14 @@ public class FileStatusService {
}
public List<FileModel> getStatusesForDossiersAndTimePeriod(Set<String> dossierIds, OffsetDateTime start, OffsetDateTime end) {
var fileEntities = fileStatusPersistenceService.getStatusesForDossiersAndTimePeriod(dossierIds, start, end);
var convertedList = convert(fileEntities, FileModel.class, new FileModelMapper());
return reanalysisRequiredStatusService.enhanceFileStatusWithAnalysisRequirements(convertedList);
}
public List<FileModel> getSoftDeletedForDossierList(List<String> dossierIds) {
var fileEntities = fileStatusPersistenceService.getSoftDeletedFiles(dossierIds);
@ -124,8 +129,7 @@ public class FileStatusService {
}
public void updateProcessingStatusPreprocessed(String dossierId, String fileId, boolean hasHighlights,
long fileSize) {
public void updateProcessingStatusPreprocessed(String dossierId, String fileId, boolean hasHighlights, long fileSize) {
fileStatusPersistenceService.updateProcessingStatusPreprocessed(fileId, hasHighlights, fileSize);
addToAnalysisQueue(dossierId, fileId, false, Set.of(), false);
@ -175,13 +179,22 @@ public class FileStatusService {
public void setStatusSuccessful(String fileId, AnalyzeResult analyzeResult) {
fileStatusPersistenceService.updateProcessingStatus(fileId, analyzeResult.getNumberOfPages(), analyzeResult.getDictionaryVersion(), analyzeResult.getRulesVersion(), analyzeResult.getLegalBasisVersion(), analyzeResult.getDuration(), analyzeResult.getDossierDictionaryVersion(), analyzeResult.getAnalysisVersion(), analyzeResult.getAnalysisNumber());
fileStatusPersistenceService.updateProcessingStatus(
fileId,
analyzeResult.getNumberOfPages(),
analyzeResult.getDictionaryVersion(),
analyzeResult.getRulesVersion(),
analyzeResult.getLegalBasisVersion(),
analyzeResult.getDuration(),
analyzeResult.getDossierDictionaryVersion(),
analyzeResult.getAnalysisVersion(),
analyzeResult.getAnalysisNumber());
}
@Transactional
public void setStatusReprocess(String dossierId, String fileId, boolean priority, Set<Integer> sectionsToReanalyse,
boolean triggeredManually, boolean manualRedactionReanalyse) {
public void setStatusReprocess(
String dossierId, String fileId, boolean priority, Set<Integer> sectionsToReanalyse, boolean triggeredManually, boolean manualRedactionReanalyse) {
log.info("Reprocessing file: {} from dossier {}", fileId, dossierId);
@ -215,8 +228,7 @@ public class FileStatusService {
@Transactional
public void setStatusFullReprocess(String dossierId, String fileId, boolean priority,
boolean requiresStructureAnalysis) {
public void setStatusFullReprocess(String dossierId, String fileId, boolean priority, boolean requiresStructureAnalysis) {
FileEntity fileStatus = fileStatusPersistenceService.getStatus(fileId);
@ -238,11 +250,7 @@ public class FileStatusService {
@SneakyThrows
public void addToPreprocessingQueue(String dossierId, String fileId, String filename) {
var processUntouchedDocumentRequest = ProcessUntouchedDocumentRequest.builder()
.dossierId(dossierId)
.fileId(fileId)
.fileName(filename)
.build();
var processUntouchedDocumentRequest = ProcessUntouchedDocumentRequest.builder().dossierId(dossierId).fileId(fileId).fileName(filename).build();
setStatusPreProcessingQueued(fileId);
@ -391,15 +399,18 @@ public class FileStatusService {
setStatusImageAnalyzing(fileId);
try {
rabbitTemplate.convertAndSend(MessagingConfiguration.IMAGE_SERVICE_QUEUE, objectMapper.writeValueAsString(ImageServiceRequest.builder()
.dossierId(dossierId)
.fileId(fileId)
.targetFileExtension(ImageServiceRequest.TARGET_FILE_EXTENSION)
.responseFileExtension(ImageServiceRequest.RESPONSE_FILE_EXTENSION)
.build()), message -> {
message.getMessageProperties().setPriority(1);
return message;
});
rabbitTemplate.convertAndSend(
MessagingConfiguration.IMAGE_SERVICE_QUEUE,
objectMapper.writeValueAsString(ImageServiceRequest.builder()
.dossierId(dossierId)
.fileId(fileId)
.targetFileExtension(ImageServiceRequest.TARGET_FILE_EXTENSION)
.responseFileExtension(ImageServiceRequest.RESPONSE_FILE_EXTENSION)
.build()),
message -> {
message.getMessageProperties().setPriority(1);
return message;
});
} catch (JsonProcessingException e) {
throw new RuntimeException(e);
}
@ -409,16 +420,19 @@ public class FileStatusService {
public void addToPdf2ImageQueue(String dossierId, String fileId, Set<Integer> pages) {
try {
rabbitTemplate.convertAndSend(MessagingConfiguration.PDF_2_IMAGE_QUEUE, objectMapper.writeValueAsString(Pdf2ImageServiceRequest.builder()
.dossierId(dossierId)
.fileId(fileId)
.targetFileExtension(Pdf2ImageServiceRequest.TARGET_FILE_EXTENSION)
.pages(pages)
.operation(Pdf2ImageServiceRequest.OPERATION_CONVERSION)
.build()), message -> {
message.getMessageProperties().setPriority(1);
return message;
});
rabbitTemplate.convertAndSend(
MessagingConfiguration.PDF_2_IMAGE_QUEUE,
objectMapper.writeValueAsString(Pdf2ImageServiceRequest.builder()
.dossierId(dossierId)
.fileId(fileId)
.targetFileExtension(Pdf2ImageServiceRequest.TARGET_FILE_EXTENSION)
.pages(pages)
.operation(Pdf2ImageServiceRequest.OPERATION_CONVERSION)
.build()),
message -> {
message.getMessageProperties().setPriority(1);
return message;
});
} catch (JsonProcessingException e) {
throw new RuntimeException(e);
}
@ -428,15 +442,18 @@ public class FileStatusService {
public void addToCvAnalysisRequestQueue(String dossierId, String fileId, Set<Integer> pages) {
try {
rabbitTemplate.convertAndSend(MessagingConfiguration.CV_ANALYSIS_QUEUE, objectMapper.writeValueAsString(CvAnalysisServiceRequest.builder()
.dossierId(dossierId)
.fileId(fileId)
.pages(pages)
.operation(CvAnalysisServiceRequest.OPERATION_TABLE_PARSING)
.build()), message -> {
message.getMessageProperties().setPriority(1);
return message;
});
rabbitTemplate.convertAndSend(
MessagingConfiguration.CV_ANALYSIS_QUEUE,
objectMapper.writeValueAsString(CvAnalysisServiceRequest.builder()
.dossierId(dossierId)
.fileId(fileId)
.pages(pages)
.operation(CvAnalysisServiceRequest.OPERATION_TABLE_PARSING)
.build()),
message -> {
message.getMessageProperties().setPriority(1);
return message;
});
} catch (JsonProcessingException e) {
throw new RuntimeException(e);
}
@ -450,8 +467,7 @@ public class FileStatusService {
@Transactional
public void overwriteFile(String dossierId, String fileId, String uploader, String filename,
boolean keepManualRedactions) {
public void overwriteFile(String dossierId, String fileId, String uploader, String filename, boolean keepManualRedactions) {
fileManagementStorageService.deleteObject(dossierId, fileId, FileType.ORIGIN);
fileManagementStorageService.deleteObject(dossierId, fileId, FileType.REDACTION_LOG);
@ -488,8 +504,7 @@ public class FileStatusService {
// wipe force redactions
var forceRedactions = forceRedactionPersistenceService.findForceRedactions(fileId, false);
forceRedactions.forEach(f -> forceRedactionPersistenceService.softDelete(fileId, f.getId()
.getAnnotationId(), now));
forceRedactions.forEach(f -> forceRedactionPersistenceService.softDelete(fileId, f.getId().getAnnotationId(), now));
// wipe add manual redactions
var addRedactions = addRedactionPersistenceService.findAddRedactions(fileId, false);
@ -497,23 +512,19 @@ public class FileStatusService {
// wipe removeRedactions
var removeRedactions = removeRedactionPersistenceService.findRemoveRedactions(fileId, false);
removeRedactions.forEach(f -> removeRedactionPersistenceService.softDelete(fileId, f.getId()
.getAnnotationId(), now));
removeRedactions.forEach(f -> removeRedactionPersistenceService.softDelete(fileId, f.getId().getAnnotationId(), now));
// wipe image recat
var imageRecategorizations = imageRecategorizationPersistenceService.findRecategorizations(fileId, false);
imageRecategorizations.forEach(f -> imageRecategorizationPersistenceService.softDelete(fileId, f.getId()
.getAnnotationId(), now));// wipe image recat
imageRecategorizations.forEach(f -> imageRecategorizationPersistenceService.softDelete(fileId, f.getId().getAnnotationId(), now));// wipe image recat
// wipe resize redactions
var resizeRedactions = resizeRedactionPersistenceService.findResizeRedactions(fileId, false);
resizeRedactions.forEach(f -> resizeRedactionPersistenceService.softDelete(fileId, f.getId()
.getAnnotationId(), now));
resizeRedactions.forEach(f -> resizeRedactionPersistenceService.softDelete(fileId, f.getId().getAnnotationId(), now));
// wipe legal basis changes
var legalBasisChanges = legalBasisChangePersistenceService.findLegalBasisChanges(fileId, false);
legalBasisChanges.forEach(f -> legalBasisChangePersistenceService.softDelete(fileId, f.getId()
.getAnnotationId(), now));
legalBasisChanges.forEach(f -> legalBasisChangePersistenceService.softDelete(fileId, f.getId().getAnnotationId(), now));
fileStatusPersistenceService.updateHasComments(fileId, false);
}
@ -547,8 +558,7 @@ public class FileStatusService {
@Transactional
protected void addToAnalysisQueue(String dossierId, String fileId, boolean priority,
Set<Integer> sectionsToReanalyse, boolean manualRedactionReanalyse) {
protected void addToAnalysisQueue(String dossierId, String fileId, boolean priority, Set<Integer> sectionsToReanalyse, boolean manualRedactionReanalyse) {
var dossier = dossierPersistenceService.getAndValidateDossier(dossierId);
var fileEntity = fileStatusPersistenceService.getStatus(fileId);
@ -619,23 +629,25 @@ public class FileStatusService {
setStatusNerAnalyzing(fileId);
try {
rabbitTemplate.convertAndSend(MessagingConfiguration.NER_SERVICE_QUEUE, objectMapper.writeValueAsString(NerServiceRequest.builder()
.dossierId(dossierId)
.fileId(fileId)
.targetFileExtension(NerServiceRequest.TARGET_FILE_EXTENSION)
.responseFileExtension(NerServiceRequest.RESPONSE_FILE_EXTENSION)
.build()), message -> {
message.getMessageProperties().setPriority(1);
return message;
});
rabbitTemplate.convertAndSend(
MessagingConfiguration.NER_SERVICE_QUEUE,
objectMapper.writeValueAsString(NerServiceRequest.builder()
.dossierId(dossierId)
.fileId(fileId)
.targetFileExtension(NerServiceRequest.TARGET_FILE_EXTENSION)
.responseFileExtension(NerServiceRequest.RESPONSE_FILE_EXTENSION)
.build()),
message -> {
message.getMessageProperties().setPriority(1);
return message;
});
} catch (JsonProcessingException e) {
throw new RuntimeException(e);
}
}
private MessageType calculateMessageType(String dossierId, String fileId, boolean reanalyse,
ProcessingStatus processingStatus, FileModel fileModel) {
private MessageType calculateMessageType(String dossierId, String fileId, boolean reanalyse, ProcessingStatus processingStatus, FileModel fileModel) {
if (!fileManagementStorageService.objectExists(dossierId, fileId, FileType.TEXT)) {
return MessageType.STRUCTURE_ANALYSE;
@ -644,8 +656,7 @@ public class FileStatusService {
return MessageType.ANALYSE;
}
if (reanalyse) {
if (fileModel.getLastFileAttributeChange() != null && fileModel.getLastProcessed()
.isBefore(fileModel.getLastFileAttributeChange())) {
if (fileModel.getLastFileAttributeChange() != null && fileModel.getLastProcessed().isBefore(fileModel.getLastFileAttributeChange())) {
return MessageType.ANALYSE;
}
return MessageType.REANALYSE;
@ -667,8 +678,7 @@ public class FileStatusService {
}
private List<com.iqser.red.service.redaction.v1.model.FileAttribute> convertAttributes(
List<FileAttributeEntity> fileAttributes, String dossierTemplateId) {
private List<com.iqser.red.service.redaction.v1.model.FileAttribute> convertAttributes(List<FileAttributeEntity> fileAttributes, String dossierTemplateId) {
List<com.iqser.red.service.redaction.v1.model.FileAttribute> fileAttributeList = new ArrayList<>();

View File

@ -4,11 +4,16 @@ import static java.util.stream.Collectors.toList;
import java.time.Duration;
import java.time.Instant;
import java.time.OffsetDateTime;
import java.time.ZoneId;
import java.util.Comparator;
import java.util.HashSet;
import java.util.List;
import java.util.Set;
import java.util.UUID;
import java.util.concurrent.TimeUnit;
import java.util.concurrent.atomic.AtomicInteger;
import java.util.stream.Collectors;
import org.apache.commons.collections4.CollectionUtils;
import org.apache.commons.lang3.StringUtils;
@ -32,6 +37,7 @@ import lombok.extern.slf4j.Slf4j;
@Slf4j
public class LicenseReportService {
private static final ZoneId UTC_ZONE_ID = ZoneId.of("Z");
private final FileStatusService fileStatusService;
private final DossierService dossierService;
@ -80,6 +86,7 @@ public class LicenseReportService {
}
private CachedReportData loadReport(LicenseReportRequest licenseReportRequest) {
log.debug("No licence-report found in cache, generating new report");
@ -88,49 +95,42 @@ public class LicenseReportService {
start = Instant.now();
}
List<String> dossierIds = licenseReportRequest.getDossierIds();
final Set<String> dossierIds;
if (CollectionUtils.isEmpty(licenseReportRequest.getDossierIds())) {
dossierIds = dossierService.getAllDossiers().stream().map(DossierEntity::getId).collect(toList());
dossierIds = dossierService.getAllDossiers().stream().map(DossierEntity::getId).collect(Collectors.toSet());
} else {
dossierIds = new HashSet<>(licenseReportRequest.getDossierIds());
}
AtomicInteger totalPagesAnalyzed = new AtomicInteger();
AtomicInteger totalPagesOcr = new AtomicInteger();
AtomicInteger totalNumberOfAnalyses = new AtomicInteger();
List<ReportData> data = dossierIds.stream()
.flatMap(dossierId -> fileStatusService.getDossierStatus(dossierId)
.stream()
.filter(fileStatus -> fileStatus.getAdded().toInstant().isAfter(licenseReportRequest.getStartDate())
&& fileStatus.getAdded().toInstant().isBefore(licenseReportRequest.getEndDate()))
.map(fileStatus -> {
ReportData reportData = new ReportData();
reportData.setDossier(dossierId);
reportData.setFileName(fileStatus.getFilename());
reportData.setAddedDate(fileStatus.getAdded().toInstant());
reportData.setLastUpdatedDate(fileStatus.getLastUpdated() == null ? null : fileStatus.getLastUpdated().toInstant());
reportData.setDeletedDate(fileStatus.getDeleted() == null ? null : fileStatus.getDeleted().toInstant());
reportData.setWorkflowStatus(fileStatus.getWorkflowStatus());
reportData.setNumberOfAnalyzedPages(fileStatus.getNumberOfPages());
reportData.setNumberOfOcrPages(fileStatus.getLastOCRTime() != null ? fileStatus.getNumberOfPages() : 0);
reportData.setAnalysisCount(fileStatus.getNumberOfAnalyses());
totalPagesAnalyzed.addAndGet(fileStatus.getNumberOfPages());
totalPagesOcr.addAndGet(fileStatus.getLastOCRTime() != null ? fileStatus.getNumberOfPages() : 0);
totalNumberOfAnalyses.addAndGet(fileStatus.getNumberOfAnalyses());
return reportData;
}))
.sorted(Comparator.comparing(ReportData::getFileName))
.collect(toList());
List<ReportData> data = fileStatusService.getStatusesForDossiersAndTimePeriod( //
dossierIds, //
OffsetDateTime.ofInstant(licenseReportRequest.getStartDate(), UTC_ZONE_ID), //
OffsetDateTime.ofInstant(licenseReportRequest.getEndDate(), UTC_ZONE_ID)) //
.stream().map(fileStatus -> {
ReportData reportData = new ReportData();
reportData.setDossier(fileStatus.getDossierId());
reportData.setFileName(fileStatus.getFilename());
reportData.setAddedDate(fileStatus.getAdded().toInstant());
reportData.setLastUpdatedDate(fileStatus.getLastUpdated() == null ? null : fileStatus.getLastUpdated().toInstant());
reportData.setDeletedDate(fileStatus.getDeleted() == null ? null : fileStatus.getDeleted().toInstant());
reportData.setWorkflowStatus(fileStatus.getWorkflowStatus());
reportData.setNumberOfAnalyzedPages(fileStatus.getNumberOfPages());
reportData.setNumberOfOcrPages(fileStatus.getLastOCRTime() != null ? fileStatus.getNumberOfPages() : 0);
reportData.setAnalysisCount(fileStatus.getNumberOfAnalyses());
totalPagesAnalyzed.addAndGet(fileStatus.getNumberOfPages());
totalPagesOcr.addAndGet(fileStatus.getLastOCRTime() != null ? fileStatus.getNumberOfPages() : 0);
totalNumberOfAnalyses.addAndGet(fileStatus.getNumberOfAnalyses());
return reportData;
}).sorted(Comparator.comparing(ReportData::getFileName)).collect(toList());
if (start != null) {
log.info("loadReport took {} to process", Duration.between(start, Instant.now()).toString());
}
return new CachedReportData(data,
dossierIds.size(),
totalNumberOfAnalyses.get(),
totalPagesAnalyzed.get(),
totalPagesOcr.get(),
licenseReportRequest.getStartDate(),
licenseReportRequest.getEndDate());
return new CachedReportData(data, dossierIds.size(), totalNumberOfAnalyses.get(), totalPagesAnalyzed.get(), totalPagesOcr.get(), licenseReportRequest.getStartDate(), licenseReportRequest.getEndDate());
}
@ -138,12 +138,13 @@ public class LicenseReportService {
@Value
class CachedReportData {
@NonNull
List<ReportData> data;
@NonNull List<ReportData> data;
int numberOfDossiers;
int totalNumberOfAnalyses;
int totalPagesAnalyzed;
int totalOcrPages;
Instant startDate;
Instant endDate;
}