Merge remote-tracking branch 'origin/dev' into dev

topo_dev
xueqingkun 4 months ago
commit 78e8e488cf

@ -311,20 +311,6 @@ SET name = CONCAT(IFNULL(start_entity_type, '无'), '-', IFNULL(rel_type, '无')
WHERE start_entity_type IS NOT NULL AND rel_type IS NOT NULL AND end_entity_type IS NOT NULL;
-- ------------------------------
INSERT INTO `nx_llm`.`com_dictionary` (`code`, `name`, `pid`, `type`, `status`, `note`, `priority`, `creator`, `update_time`, `create_time`, `updater`, `data_status`, `create_user_id`, `update_user_id`) VALUES (NULL, '提示词类型', 0, 'prompt_type', '1', NULL, 0, NULL, NULL, CURRENT_TIMESTAMP, NULL, '1', NULL, NULL);
SET @last_id = LAST_INSERT_ID();
INSERT INTO `nx_llm`.`com_dictionary` (`code`, `name`, `pid`, `type`, `status`, `note`, `priority`, `creator`, `update_time`, `create_time`, `updater`, `data_status`, `create_user_id`, `update_user_id`) VALUES ('1', '结构化推理', @last_id, 'prompt_type', '1', NULL, 0, NULL, NULL, CURRENT_TIMESTAMP, NULL, '1', NULL, NULL);
INSERT INTO `nx_llm`.`com_dictionary` (`code`, `name`, `pid`, `type`, `status`, `note`, `priority`, `creator`, `update_time`, `create_time`, `updater`, `data_status`, `create_user_id`, `update_user_id`) VALUES ('2', '图谱推理', @last_id, 'prompt_type', '1', NULL, 0, NULL, NULL, CURRENT_TIMESTAMP, NULL, '1', NULL, NULL);
INSERT INTO `nx_llm`.`com_dictionary` (`code`, `name`, `pid`, `type`, `status`, `note`, `priority`, `creator`, `update_time`, `create_time`, `updater`, `data_status`, `create_user_id`, `update_user_id`) VALUES ('3', '分类', @last_id, 'prompt_type', '1', NULL, 0, NULL, NULL, CURRENT_TIMESTAMP, NULL, '1', NULL, NULL);
INSERT INTO `nx_llm`.`com_dictionary` (`code`, `name`, `pid`, `type`, `status`, `note`, `priority`, `creator`, `update_time`, `create_time`, `updater`, `data_status`, `create_user_id`, `update_user_id`) VALUES (NULL, '提示词属性值类型', 0, 'prompt_attribute_valuetype', '1', NULL, 0, NULL, NULL, CURRENT_TIMESTAMP, NULL, '1', NULL, NULL);
SET @last_id = LAST_INSERT_ID();
INSERT INTO `nx_llm`.`com_dictionary` (`code`, `name`, `pid`, `type`, `status`, `note`, `priority`, `creator`, `update_time`, `create_time`, `updater`, `data_status`, `create_user_id`, `update_user_id`) VALUES ('1', '文本', @last_id, 'prompt_attribute_valuetype', '1', NULL, 0, NULL, NULL, CURRENT_TIMESTAMP, NULL, '1', NULL, NULL);
INSERT INTO `nx_llm`.`com_dictionary` (`code`, `name`, `pid`, `type`, `status`, `note`, `priority`, `creator`, `update_time`, `create_time`, `updater`, `data_status`, `create_user_id`, `update_user_id`) VALUES ('2', '数字', @last_id, 'prompt_attribute_valuetype', '1', NULL, 0, NULL, NULL, CURRENT_TIMESTAMP, NULL, '1', NULL, NULL);
INSERT INTO `nx_llm`.`com_dictionary` (`code`, `name`, `pid`, `type`, `status`, `note`, `priority`, `creator`, `update_time`, `create_time`, `updater`, `data_status`, `create_user_id`, `update_user_id`) VALUES ('3', '日期', @last_id, 'prompt_attribute_valuetype', '1', NULL, 0, NULL, NULL, CURRENT_TIMESTAMP, NULL, '1', NULL, NULL);
INSERT INTO `nx_llm`.`evidence_category` (`id`, `code`, `case_type`, `category_name`, `icon_url`, `parent_id`, `create_user_id`, `create_time`, `update_user_id`, `update_time`) VALUES (1, NULL, '1', '书证', NULL, NULL, NULL, CURRENT_TIMESTAMP, NULL, NULL);
INSERT INTO `nx_llm`.`evidence_category` (`id`, `code`, `case_type`, `category_name`, `icon_url`, `parent_id`, `create_user_id`, `create_time`, `update_user_id`, `update_time`) VALUES (2, NULL, '1', '电子证据', NULL, NULL, NULL, CURRENT_TIMESTAMP, NULL, NULL);
INSERT INTO `nx_llm`.`evidence_category` (`id`, `code`, `case_type`, `category_name`, `icon_url`, `parent_id`, `create_user_id`, `create_time`, `update_user_id`, `update_time`) VALUES (3, NULL, '1', '物证', NULL, NULL, NULL, CURRENT_TIMESTAMP, NULL, NULL);
@ -339,3 +325,4 @@ INSERT INTO nx_llm.evidence_category (id, code, case_type, category_name, icon_u
INSERT INTO nx_llm.evidence_category (id, code, case_type, category_name, icon_url, parent_id, create_user_id, create_time, update_user_id, update_time) VALUES (19, null, '1', '公文书证', null, '1', null, CURRENT_TIMESTAMP, NULL, NULL);
INSERT INTO `nx_llm`.`com_dictionary` (`name`, `code`, `pid`, `type`, `status`, `note`, `priority`, `creator`, `update_time`, `create_time`, `updater`, `data_status`, `create_user_id`, `update_user_id`) VALUES ('结构化查询', '5', 3, 'index_source', '1', NULL, 0, NULL, CURRENT_TIMESTAMP, NULL, NULL, '1', NULL, NULL);

@ -434,7 +434,7 @@ public class ModelCaseServiceImpl extends ServiceImpl<ModelCaseMapper, ModelCase
atomic.judgeWithIndexResult(relationalSymbol);
atomic.setSegmentation(nodeRecordSpliteMap);
atomic.setRecordIfSegmentationAbsent("无");
List<ModelAtomicResult> filterAtomicResultList = atomicResultList.stream().filter(v -> StrUtil.equals(atomic.getAtomicIndexId(), v.getAtomicId())).toList();
List<ModelAtomicResult> filterAtomicResultList = atomicResultList.stream().filter(v -> StrUtil.equals(atomic.getAtomicIndexId(), v.getAtomicId()) && StrUtil.equals(record.getIndexId(), v.getIndexId())).toList();
List<EvidentIndexResultDTO> evidentResult = getEvidentResult(filterAtomicResultList, caseEvidenceList, rootDirectory);
atomic.setEvidentResultList(evidentResult);

@ -22,7 +22,7 @@
<charset>UTF-8</charset>
</encoder>
<rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
<fileNamePattern>${LOG_PATH}/application.%d{yyyy-MM-dd}.log</fileNamePattern>
<fileNamePattern>${LOG_PATH}/application/application.%d{yyyy-MM-dd}.log</fileNamePattern>
<maxHistory>30</maxHistory>
<totalSizeCap>100MB</totalSizeCap>
</rollingPolicy>
@ -30,13 +30,13 @@
<!-- 异常错误日志单独输出 -->
<appender name="ERROR_FILE" class="ch.qos.logback.core.rolling.RollingFileAppender">
<file>${LOG_PATH}/error.log</file>
<file>${LOG_PATH}/error/error.log</file>
<encoder>
<pattern>%d{yyyy-MM-dd HH:mm:ss.SSS} [%thread] %-5level %logger{36} - %msg%n</pattern>
<charset>UTF-8</charset>
</encoder>
<rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
<fileNamePattern>${LOG_PATH}/error.%d{yyyy-MM-dd}.log</fileNamePattern>
<fileNamePattern>${LOG_PATH}/error/error.%d{yyyy-MM-dd}.log</fileNamePattern>
<maxHistory>30</maxHistory>
<totalSizeCap>100MB</totalSizeCap>
</rollingPolicy>
@ -45,15 +45,48 @@
</filter>
</appender>
<!-- 定义日志记录器 -->
<!-- <logger name="org.springframework" level="INFO" additivity="false">-->
<!-- <appender-ref ref="FILE"/>-->
<!-- </logger>-->
<!-- 提取三元组 -->
<appender name="EXTRACT_TRIPLE_META_LOG" class="ch.qos.logback.core.rolling.RollingFileAppender">
<file>${LOG_PATH}/extract_triple_meta/extract_triple_meta.log</file>
<encoder>
<pattern>%d{yyyy-MM-dd HH:mm:ss.SSS} [%thread] %-5level %logger{36} - %msg%n</pattern>
<charset>UTF-8</charset>
</encoder>
<rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
<fileNamePattern>${LOG_PATH}/extract_triple_meta/extract_triple_meta.%d{yyyy-MM-dd}.log</fileNamePattern>
<maxHistory>30</maxHistory>
<totalSizeCap>100MB</totalSizeCap>
</rollingPolicy>
</appender>
<!-- AOP -->
<appender name="AOP_LOG" class="ch.qos.logback.core.rolling.RollingFileAppender">
<file>${LOG_PATH}/AOP_LOG/AOP_LOG.log</file>
<encoder>
<pattern>%d{yyyy-MM-dd HH:mm:ss.SSS} [%thread] %-5level %logger{36} - %msg%n</pattern>
<charset>UTF-8</charset>
</encoder>
<rollingPolicy class="ch.qos.logback.core.rolling.TimeBasedRollingPolicy">
<fileNamePattern>${LOG_PATH}/AOP_LOG/AOP_LOG.%d{yyyy-MM-dd}.log</fileNamePattern>
<maxHistory>30</maxHistory>
<totalSizeCap>100MB</totalSizeCap>
</rollingPolicy>
</appender>
<!-- 提取三元组日志输出到extract_triple_meta文件不输出到默认日志 -->
<logger name="com.supervision.thread" level="INFO" additivity="false">
<appender-ref ref="EXTRACT_TRIPLE_META_LOG"/>
</logger>
<!-- 提取三元组日志输出到extract_triple_meta文件不输出到默认日志 -->
<logger name="com.supervision.police.service.impl.ExtractTripleInfoServiceImpl" level="INFO" additivity="false">
<appender-ref ref="EXTRACT_TRIPLE_META_LOG"/>
</logger>
<!-- <logger name="com.supervision" level="debug" additivity="false">-->
<!-- <appender-ref ref="FILE"/>-->
<!-- <appender-ref ref="CONSOLE"/>-->
<!-- </logger>-->
<!-- AOP日志输出到AOP文件不输出到默认日志 -->
<logger name="com.supervision.config.RequestLogConfig" level="INFO" additivity="false">
<appender-ref ref="AOP_LOG"/>
</logger>
<!-- 根日志记录器 -->
<root level="${LOGBACK_LOG_LEVEL}">

@ -65,8 +65,8 @@ public class CaseTest {
ruleCondition.setRelationalSymbol(atomicData.getRelationalSymbol());
modelAtomicIndices.stream().filter(modelAtomicIndex -> modelAtomicIndex.getId().equals(atomicData.getAtomicIndex())).findFirst().ifPresent(modelAtomicIndex -> {
if (modelAtomicIndex.getIndexSource().equals(IndexRuleConstants.OPERAND_TYPE_GRAPH)) {
ruleCondition.setOperator(">");
ruleCondition.setValue("0");
ruleCondition.setOperator(">=");
ruleCondition.setValue("1");
}
ruleCondition.setIndexSource(modelAtomicIndex.getIndexSource());
});

Loading…
Cancel
Save