RED-9466 - Adding annotation removes all AI based recommendations until forced re-analysis
- fix sonar issues
This commit is contained in:
parent
b9c6af69e2
commit
25c25ecf00
@ -223,7 +223,6 @@ public class EntityLogCreatorService {
|
||||
|
||||
String type = precursorEntity.getManualOverwrite().getType()
|
||||
.orElse(precursorEntity.getType());
|
||||
boolean isHint = isHint(precursorEntity.getEntityType());
|
||||
return EntityLogEntry.builder()
|
||||
.id(precursorEntity.getId())
|
||||
.reason(precursorEntity.buildReasonWithManualChangeDescriptions())
|
||||
|
||||
@ -73,9 +73,10 @@ public class ComponentDroolsExecutionService {
|
||||
entities.add(Entity.fromEntityLogEntry(entry, document, entry.getStartOffset(), entry.getEndOffset()));
|
||||
if (entry.getDuplicatedTextRanges() != null && !entry.getDuplicatedTextRanges().isEmpty()) {
|
||||
entry.getDuplicatedTextRanges()
|
||||
.forEach(duplicatedTextRange -> {
|
||||
entities.add(Entity.fromEntityLogEntry(entry, document, duplicatedTextRange.getStart(), duplicatedTextRange.getEnd()));
|
||||
});
|
||||
.forEach(duplicatedTextRange -> entities.add(Entity.fromEntityLogEntry(entry,
|
||||
document,
|
||||
duplicatedTextRange.getStart(),
|
||||
duplicatedTextRange.getEnd())));
|
||||
}
|
||||
return entities.stream();
|
||||
})
|
||||
|
||||
@ -178,10 +178,10 @@ public class RedactionAcceptanceTest extends AbstractRedactionIntegrationTest {
|
||||
|
||||
var publishedInformationEntry1 = findEntityByTypeAndValue(entityLog, "published_information", "Oxford University Press").findFirst()
|
||||
.orElseThrow();
|
||||
assertThat(publishedInformationEntry1.getSection().startsWith("Paragraph:"));
|
||||
assertThat(publishedInformationEntry1.getSection()).startsWith("Paragraph:");
|
||||
var asyaLyon1 = findEntityByTypeAndValueAndSectionNumber(entityLog, "CBI_author", "Asya Lyon", publishedInformationEntry1.getContainingNodeId()).findFirst()
|
||||
.orElseThrow();
|
||||
assertThat(asyaLyon1.getSection().startsWith("Paragraph:"));
|
||||
assertThat(asyaLyon1.getSection()).startsWith("Paragraph:");
|
||||
assertEquals(EntryState.SKIPPED, asyaLyon1.getState());
|
||||
|
||||
var idRemoval = buildIdRemoval(publishedInformationEntry1.getId());
|
||||
@ -232,10 +232,10 @@ public class RedactionAcceptanceTest extends AbstractRedactionIntegrationTest {
|
||||
|
||||
var publishedInformationEntry1 = findEntityByTypeAndValue(entityLog, "published_information", "Oxford University Press").findFirst()
|
||||
.orElseThrow();
|
||||
assertThat(publishedInformationEntry1.getSection().startsWith("Paragraph:"));
|
||||
assertThat(publishedInformationEntry1.getSection()).startsWith("Paragraph:");
|
||||
var asyaLyon1 = findEntityByTypeAndValueAndSectionNumber(entityLog, "CBI_author", "Asya Lyon", publishedInformationEntry1.getContainingNodeId()).findFirst()
|
||||
.orElseThrow();
|
||||
assertThat(publishedInformationEntry1.getSection().startsWith("Paragraph:"));
|
||||
assertThat(publishedInformationEntry1.getSection()).startsWith("Paragraph:");
|
||||
assertEquals(EntryState.SKIPPED, asyaLyon1.getState());
|
||||
|
||||
var idRemoval = buildIdRemoval(publishedInformationEntry1.getId());
|
||||
@ -388,7 +388,7 @@ public class RedactionAcceptanceTest extends AbstractRedactionIntegrationTest {
|
||||
|
||||
@Test
|
||||
@SneakyThrows
|
||||
public void testNerEntitiesAfterReanalysis() {
|
||||
void testNerEntitiesAfterReanalysis() {
|
||||
|
||||
String EFSA_SANITISATION_RULES = loadFromClassPath("drools/efsa_sanitisation.drl");
|
||||
when(rulesClient.getRules(TEST_DOSSIER_TEMPLATE_ID, RuleFileType.ENTITY)).thenReturn(JSONPrimitive.of(EFSA_SANITISATION_RULES));
|
||||
@ -414,7 +414,7 @@ public class RedactionAcceptanceTest extends AbstractRedactionIntegrationTest {
|
||||
String nerValue = "Osip S.";
|
||||
var nerEntity = findEntityByTypeAndValue(entityLog, DICTIONARY_AUTHOR, nerValue).findFirst()
|
||||
.orElseThrow();
|
||||
assertThat(nerEntity.getEngines().contains(Engine.NER));
|
||||
assertThat(nerEntity.getEngines()).contains(Engine.NER);
|
||||
|
||||
String dictionaryAddValue = "cooperation";
|
||||
ManualRedactionEntry manualRedactionEntry = ManualRedactionEntry.builder()
|
||||
@ -445,11 +445,11 @@ public class RedactionAcceptanceTest extends AbstractRedactionIntegrationTest {
|
||||
.filter(entityLogEntry -> entityLogEntry.getValue().equals(dictionaryAddValue))
|
||||
.findFirst()
|
||||
.get();
|
||||
assertEquals(entityLogEntryAdded.getState(), EntryState.APPLIED);
|
||||
assertEquals(EntryState.APPLIED, entityLogEntryAdded.getState());
|
||||
|
||||
nerEntity = findEntityByTypeAndValue(entityLog, DICTIONARY_AUTHOR, nerValue).findFirst()
|
||||
.orElseThrow();
|
||||
assertThat(nerEntity.getEngines().contains(Engine.NER));
|
||||
assertThat(nerEntity.getEngines()).contains(Engine.NER);
|
||||
dossierDictionary.get(DICTIONARY_AUTHOR).remove(dictionaryAddValue);
|
||||
|
||||
}
|
||||
|
||||
Loading…
x
Reference in New Issue
Block a user