try to: init lucene
This commit is contained in:
parent
bf5b1ef6c9
commit
90403f9b26
BIN
articlesIndex/index0/segments_2
Normal file
BIN
articlesIndex/index0/segments_2
Normal file
Binary file not shown.
0
articlesIndex/index0/write.lock
Normal file
0
articlesIndex/index0/write.lock
Normal file
41
pom.xml
41
pom.xml
@ -5,7 +5,7 @@
|
||||
<parent>
|
||||
<groupId>org.springframework.boot</groupId>
|
||||
<artifactId>spring-boot-starter-parent</artifactId>
|
||||
<version>2.2.6.RELEASE</version>
|
||||
<version>2.3.5.RELEASE</version>
|
||||
<relativePath/> <!-- lookup parent from repository -->
|
||||
</parent>
|
||||
<groupId>com.rymcu</groupId>
|
||||
@ -17,6 +17,7 @@
|
||||
|
||||
<properties>
|
||||
<java.version>1.8</java.version>
|
||||
<lucene.version>6.6.6</lucene.version>
|
||||
</properties>
|
||||
|
||||
<dependencies>
|
||||
@ -73,6 +74,11 @@
|
||||
<groupId>org.springframework.boot</groupId>
|
||||
<artifactId>spring-boot-starter-aop</artifactId>
|
||||
</dependency>
|
||||
<dependency>
|
||||
<groupId>org.springframework.boot</groupId>
|
||||
<artifactId>spring-boot-starter-log4j2</artifactId>
|
||||
</dependency>
|
||||
|
||||
<!-- 通用Mapper插件 -->
|
||||
<dependency>
|
||||
<groupId>tk.mybatis</groupId>
|
||||
@ -200,6 +206,39 @@
|
||||
</exclusion>
|
||||
</exclusions>
|
||||
</dependency>
|
||||
|
||||
<!-- lucene -->
|
||||
<dependency>
|
||||
<groupId>org.apache.lucene</groupId>
|
||||
<artifactId>lucene-queryparser</artifactId>
|
||||
<version>${lucene.version}</version>
|
||||
</dependency>
|
||||
|
||||
<dependency>
|
||||
<groupId>org.apache.lucene</groupId>
|
||||
<artifactId>lucene-analyzers-common</artifactId>
|
||||
<version>${lucene.version}</version>
|
||||
</dependency>
|
||||
|
||||
<dependency>
|
||||
<groupId>org.apache.lucene</groupId>
|
||||
<artifactId>lucene-highlighter</artifactId>
|
||||
<version>${lucene.version}</version>
|
||||
</dependency>
|
||||
|
||||
|
||||
<!-- https://mvnrepository.com/artifact/org.apache.lucene/lucene-suggest -->
|
||||
<dependency>
|
||||
<groupId>org.apache.lucene</groupId>
|
||||
<artifactId>lucene-suggest</artifactId>
|
||||
<version>${lucene.version}</version>
|
||||
</dependency>
|
||||
|
||||
<dependency>
|
||||
<groupId>cn.hutool</groupId>
|
||||
<artifactId>hutool-all</artifactId>
|
||||
<version>5.5.8</version>
|
||||
</dependency>
|
||||
</dependencies>
|
||||
|
||||
<build>
|
||||
|
82
src/main/java/com/rymcu/forest/config/WebLogAspect.java
Normal file
82
src/main/java/com/rymcu/forest/config/WebLogAspect.java
Normal file
@ -0,0 +1,82 @@
|
||||
package com.rymcu.forest.config;
|
||||
|
||||
import org.aspectj.lang.JoinPoint;
|
||||
import org.aspectj.lang.annotation.AfterReturning;
|
||||
import org.aspectj.lang.annotation.Aspect;
|
||||
import org.aspectj.lang.annotation.Before;
|
||||
import org.aspectj.lang.annotation.Pointcut;
|
||||
import org.slf4j.Logger;
|
||||
import org.slf4j.LoggerFactory;
|
||||
import org.springframework.stereotype.Component;
|
||||
import org.springframework.web.context.request.RequestContextHolder;
|
||||
import org.springframework.web.context.request.ServletRequestAttributes;
|
||||
|
||||
import javax.servlet.http.HttpServletRequest;
|
||||
import java.util.Arrays;
|
||||
import java.util.Enumeration;
|
||||
|
||||
/**
|
||||
* WebLogAspect 实现Web层的日志切面
|
||||
*
|
||||
* @author suwen
|
||||
* @date 2020/12/22 9:16 上午
|
||||
*/
|
||||
@Aspect
|
||||
@Component
|
||||
public class WebLogAspect {
|
||||
|
||||
private final Logger logger = LoggerFactory.getLogger(this.getClass());
|
||||
|
||||
ThreadLocal<Long> startTime = new ThreadLocal<Long>();
|
||||
|
||||
/**
|
||||
* 定义一个切入点. 解释下:
|
||||
*
|
||||
* <p>~ 第一个 * 代表任意修饰符及任意返回值. ~ 第二个 * 任意包名 ~ 第三个 * 代表任意方法.
|
||||
*
|
||||
* <p>~ 第四个 * 定义在web包或者子包 ~ 第五个 * 任意方法 ~ .. 匹配任意数量的参数. execution(*
|
||||
* xiao.ze.demo.service.impl.*.*(..))
|
||||
*/
|
||||
@Pointcut("execution(* com.rymcu.forest.*.api.*.*.*(..))")
|
||||
public void webLog() {
|
||||
}
|
||||
|
||||
@Before("webLog()")
|
||||
public void doBefore(JoinPoint joinPoint) {
|
||||
|
||||
startTime.set(System.currentTimeMillis());
|
||||
|
||||
// 接收到请求,记录请求内容
|
||||
logger.info("WebLogAspect.doBefore()");
|
||||
ServletRequestAttributes attributes =
|
||||
(ServletRequestAttributes) RequestContextHolder.getRequestAttributes();
|
||||
HttpServletRequest request = attributes.getRequest();
|
||||
|
||||
// 记录下请求内容
|
||||
logger.info("URL : " + request.getRequestURL().toString());
|
||||
logger.info("HTTP_METHOD : " + request.getMethod());
|
||||
logger.info("IP : " + request.getRemoteAddr());
|
||||
logger.info(
|
||||
"CLASS_METHOD : "
|
||||
+ joinPoint.getSignature().getDeclaringTypeName()
|
||||
+ "."
|
||||
+ joinPoint.getSignature().getName());
|
||||
logger.info("ARGS : " + Arrays.toString(joinPoint.getArgs()));
|
||||
|
||||
// 获取所有参数方法一:
|
||||
Enumeration<String> enu = request.getParameterNames();
|
||||
while (enu.hasMoreElements()) {
|
||||
String paraName = (String) enu.nextElement();
|
||||
logger.info(paraName + ": " + request.getParameter(paraName));
|
||||
}
|
||||
}
|
||||
|
||||
@AfterReturning("webLog()")
|
||||
public void doAfterReturning(JoinPoint joinPoint) {
|
||||
|
||||
// 处理完请求,返回内容
|
||||
logger.info("WebLogAspect.doAfterReturning()");
|
||||
logger.info("耗时(毫秒) : " + (System.currentTimeMillis() - startTime.get()));
|
||||
}
|
||||
}
|
||||
|
72
src/main/java/com/rymcu/forest/lucene/api/SearchController.java
Executable file
72
src/main/java/com/rymcu/forest/lucene/api/SearchController.java
Executable file
@ -0,0 +1,72 @@
|
||||
package com.rymcu.forest.lucene.api;
|
||||
|
||||
import com.rymcu.forest.dto.ArticleDTO;
|
||||
import com.rymcu.forest.dto.ArticleSearchDTO;
|
||||
import com.rymcu.forest.lucene.mapper.BaikeMapper;
|
||||
import com.rymcu.forest.lucene.model.Baike;
|
||||
import com.rymcu.forest.lucene.service.SearchService;
|
||||
import com.rymcu.forest.service.ArticleService;
|
||||
import org.springframework.beans.factory.annotation.Autowired;
|
||||
import org.springframework.web.bind.annotation.GetMapping;
|
||||
import org.springframework.web.bind.annotation.PathVariable;
|
||||
import org.springframework.web.bind.annotation.RequestMapping;
|
||||
import org.springframework.web.bind.annotation.RestController;
|
||||
import org.springframework.web.servlet.ModelAndView;
|
||||
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
|
||||
@RestController
|
||||
@RequestMapping("/api/v1/lucene")
|
||||
public class SearchController {
|
||||
@Autowired private BaikeMapper baikeMapper;
|
||||
@Autowired private SearchService searchService;
|
||||
@Autowired private ArticleService articleService;
|
||||
|
||||
@GetMapping("/index")
|
||||
public String createIndex(int limit, int offset) {
|
||||
// 拉取数据
|
||||
List<Baike> baikes = baikeMapper.getAllBaike(limit, offset);
|
||||
searchService.write(baikes);
|
||||
return "成功";
|
||||
}
|
||||
|
||||
@GetMapping("/indexArticle")
|
||||
public String createArticleIndex() {
|
||||
// 拉取数据
|
||||
List<ArticleDTO> list = articleService.findArticles(new ArticleSearchDTO());
|
||||
searchService.writeArticle(list);
|
||||
return "成功";
|
||||
}
|
||||
|
||||
/**
|
||||
* 搜索,实现高亮
|
||||
*
|
||||
* @param q
|
||||
* @return
|
||||
* @throws Exception
|
||||
*/
|
||||
@GetMapping("/search/{q}")
|
||||
public List<Map<String, String>> getSearchText(@PathVariable String q) throws Exception {
|
||||
return searchService.search(q);
|
||||
}
|
||||
|
||||
/**
|
||||
* 搜索,实现高亮
|
||||
*
|
||||
* @param q
|
||||
* @return
|
||||
* @throws Exception
|
||||
*/
|
||||
@GetMapping("/searchArticle/{q}")
|
||||
public List<?> searchArticle(@PathVariable String q) throws Exception {
|
||||
|
||||
return searchService.searchArticle(q);
|
||||
}
|
||||
|
||||
@GetMapping(value = "/search")
|
||||
public ModelAndView test(ModelAndView mv) {
|
||||
mv.setViewName("/search");
|
||||
return mv;
|
||||
}
|
||||
}
|
75
src/main/java/com/rymcu/forest/lucene/cfg/Configuration.java
Executable file
75
src/main/java/com/rymcu/forest/lucene/cfg/Configuration.java
Executable file
@ -0,0 +1,75 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.cfg;
|
||||
|
||||
import java.util.List;
|
||||
|
||||
/**
|
||||
*
|
||||
* 配置管理类接口
|
||||
*
|
||||
*/
|
||||
public interface Configuration {
|
||||
|
||||
/**
|
||||
* 返回useSmart标志位
|
||||
* useSmart =true ,分词器使用智能切分策略, =false则使用细粒度切分
|
||||
* @return useSmart
|
||||
*/
|
||||
boolean useSmart();
|
||||
|
||||
/**
|
||||
* 设置useSmart标志位
|
||||
* useSmart =true ,分词器使用智能切分策略, =false则使用细粒度切分
|
||||
* @param useSmart
|
||||
*/
|
||||
void setUseSmart(boolean useSmart);
|
||||
|
||||
/**
|
||||
* 获取主词典路径
|
||||
*
|
||||
* @return String 主词典路径
|
||||
*/
|
||||
String getMainDictionary();
|
||||
|
||||
/**
|
||||
* 获取量词词典路径
|
||||
* @return String 量词词典路径
|
||||
*/
|
||||
String getQuantifierDictionary();
|
||||
|
||||
/**
|
||||
* 获取扩展字典配置路径
|
||||
* @return List<String> 相对类加载器的路径
|
||||
*/
|
||||
List<String> getExtDictionary();
|
||||
|
||||
/**
|
||||
* 获取扩展停止词典配置路径
|
||||
* @return List<String> 相对类加载器的路径
|
||||
*/
|
||||
List<String> getExtStopWordDictionary();
|
||||
|
||||
}
|
159
src/main/java/com/rymcu/forest/lucene/cfg/DefaultConfig.java
Executable file
159
src/main/java/com/rymcu/forest/lucene/cfg/DefaultConfig.java
Executable file
@ -0,0 +1,159 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0 IK Analyzer release 5.0
|
||||
*
|
||||
* <p>Licensed to the Apache Software Foundation (ASF) under one or more contributor license
|
||||
* agreements. See the NOTICE file distributed with this work for additional information regarding
|
||||
* copyright ownership. The ASF licenses this file to You under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance with the License. You may obtain a
|
||||
* copy of the License at
|
||||
*
|
||||
* <p>http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* <p>Unless required by applicable law or agreed to in writing, software distributed under the
|
||||
* License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||
* express or implied. See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* <p>源代码由林良益(linliangyi2005@gmail.com)提供 版权声明 2012,乌龙茶工作室 provided by Linliangyi and copyright 2012
|
||||
* by Oolong studio
|
||||
*/
|
||||
package com.rymcu.forest.lucene.cfg;
|
||||
|
||||
import org.springframework.stereotype.Component;
|
||||
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
|
||||
/** Configuration 默认实现 2012-5-8 */
|
||||
@Component
|
||||
public class DefaultConfig implements Configuration {
|
||||
|
||||
/**
|
||||
* 分词器默认字典路径
|
||||
*/
|
||||
private static final String PATH_DIC_MAIN = "lucene/main2012.dic";
|
||||
private static final String PATH_DIC_QUANTIFIER = "lucene/quantifier.dic";
|
||||
|
||||
|
||||
/**
|
||||
* 分词器配置文件路径
|
||||
*/
|
||||
private static final String FILE_NAME = "IKAnalyzer.cfg.xml";
|
||||
// 配置属性——扩展字典
|
||||
private static final String EXT_DICT = "ext_dic";
|
||||
// 配置属性——扩展停止词典
|
||||
private static final String EXT_STOP = "ext_stopword";
|
||||
|
||||
private String extDic = "lucene/ext.dic";
|
||||
|
||||
private String extStopword = "lucene/stopword.dic";
|
||||
/*
|
||||
* 是否使用smart方式分词
|
||||
*/
|
||||
private boolean useSmart;
|
||||
|
||||
/**
|
||||
* 返回单例
|
||||
*
|
||||
* @return Configuration单例
|
||||
*/
|
||||
public static Configuration getInstance() {
|
||||
return new DefaultConfig();
|
||||
}
|
||||
|
||||
/**
|
||||
* 返回useSmart标志位 useSmart =true ,分词器使用智能切分策略, =false则使用细粒度切分
|
||||
*
|
||||
* @return useSmart
|
||||
*/
|
||||
@Override
|
||||
public boolean useSmart() {
|
||||
return useSmart;
|
||||
}
|
||||
|
||||
/**
|
||||
* 设置useSmart标志位 useSmart =true ,分词器使用智能切分策略, =false则使用细粒度切分
|
||||
*
|
||||
* @param useSmart
|
||||
*/
|
||||
@Override
|
||||
public void setUseSmart(boolean useSmart) {
|
||||
this.useSmart = useSmart;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取主词典路径
|
||||
*
|
||||
* @return String 主词典路径
|
||||
*/
|
||||
@Override
|
||||
public String getMainDictionary() {
|
||||
return PATH_DIC_MAIN;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取量词词典路径
|
||||
*
|
||||
* @return String 量词词典路径
|
||||
*/
|
||||
@Override
|
||||
public String getQuantifierDictionary() {
|
||||
return PATH_DIC_QUANTIFIER;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取扩展字典配置路径
|
||||
*
|
||||
* @return List<String> 相对类加载器的路径
|
||||
*/
|
||||
@Override
|
||||
public List<String> getExtDictionary() {
|
||||
List<String> extDictFiles = new ArrayList<String>(2);
|
||||
if (extDic != null) {
|
||||
// 使用;分割多个扩展字典配置
|
||||
String[] filePaths = extDic.split(";");
|
||||
for (String filePath : filePaths) {
|
||||
if (filePath != null && !"".equals(filePath.trim())) {
|
||||
extDictFiles.add(filePath.trim());
|
||||
}
|
||||
}
|
||||
}
|
||||
return extDictFiles;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取扩展停止词典配置路径
|
||||
*
|
||||
* @return List<String> 相对类加载器的路径
|
||||
*/
|
||||
@Override
|
||||
public List<String> getExtStopWordDictionary() {
|
||||
List<String> extStopWordDictFiles = new ArrayList<>(2);
|
||||
if (extStopword != null) {
|
||||
// 使用;分割多个扩展字典配置
|
||||
String[] filePaths = extStopword.split(";");
|
||||
for (String filePath : filePaths) {
|
||||
if (filePath != null && !"".equals(filePath.trim())) {
|
||||
extStopWordDictFiles.add(filePath.trim());
|
||||
}
|
||||
}
|
||||
}
|
||||
return extStopWordDictFiles;
|
||||
}
|
||||
|
||||
public String getExtDic() {
|
||||
return extDic;
|
||||
}
|
||||
|
||||
public void setExtDic(String extDic) {
|
||||
this.extDic = extDic;
|
||||
}
|
||||
|
||||
public String getExtStopword() {
|
||||
return extStopword;
|
||||
}
|
||||
|
||||
public void setExtStopword(String extStopword) {
|
||||
this.extStopword = extStopword;
|
||||
}
|
||||
}
|
379
src/main/java/com/rymcu/forest/lucene/core/AnalyzeContext.java
Executable file
379
src/main/java/com/rymcu/forest/lucene/core/AnalyzeContext.java
Executable file
@ -0,0 +1,379 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.core;
|
||||
|
||||
|
||||
import com.rymcu.forest.lucene.cfg.Configuration;
|
||||
import com.rymcu.forest.lucene.dic.Dictionary;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.Reader;
|
||||
import java.util.*;
|
||||
|
||||
/**
|
||||
*
|
||||
* 分词器上下文状态
|
||||
*
|
||||
*/
|
||||
class AnalyzeContext {
|
||||
|
||||
// 默认缓冲区大小
|
||||
private static final int BUFF_SIZE = 4096;
|
||||
// 缓冲区耗尽的临界值
|
||||
private static final int BUFF_EXHAUST_CRITICAL = 100;
|
||||
|
||||
// 字符窜读取缓冲
|
||||
private char[] segmentBuff;
|
||||
// 字符类型数组
|
||||
private int[] charTypes;
|
||||
|
||||
// 记录Reader内已分析的字串总长度
|
||||
// 在分多段分析词元时,该变量累计当前的segmentBuff相对于reader起始位置的位移
|
||||
private int buffOffset;
|
||||
// 当前缓冲区位置指针
|
||||
private int cursor;
|
||||
// 最近一次读入的,可处理的字串长度
|
||||
private int available;
|
||||
|
||||
// 子分词器锁
|
||||
// 该集合非空,说明有子分词器在占用segmentBuff
|
||||
private Set<String> buffLocker;
|
||||
|
||||
// 原始分词结果集合,未经歧义处理
|
||||
private QuickSortSet orgLexemes;
|
||||
// LexemePath位置索引表
|
||||
private Map<Integer, LexemePath> pathMap;
|
||||
// 最终分词结果集
|
||||
private LinkedList<Lexeme> results;
|
||||
|
||||
// 分词器配置项
|
||||
private Configuration cfg;
|
||||
|
||||
public AnalyzeContext(Configuration cfg) {
|
||||
this.cfg = cfg;
|
||||
this.segmentBuff = new char[BUFF_SIZE];
|
||||
this.charTypes = new int[BUFF_SIZE];
|
||||
this.buffLocker = new HashSet<String>();
|
||||
this.orgLexemes = new QuickSortSet();
|
||||
this.pathMap = new HashMap<Integer, LexemePath>();
|
||||
this.results = new LinkedList<Lexeme>();
|
||||
}
|
||||
|
||||
int getCursor() {
|
||||
return this.cursor;
|
||||
}
|
||||
|
||||
char[] getSegmentBuff() {
|
||||
return this.segmentBuff;
|
||||
}
|
||||
|
||||
char getCurrentChar() {
|
||||
return this.segmentBuff[this.cursor];
|
||||
}
|
||||
|
||||
int getCurrentCharType() {
|
||||
return this.charTypes[this.cursor];
|
||||
}
|
||||
|
||||
int getBufferOffset() {
|
||||
return this.buffOffset;
|
||||
}
|
||||
|
||||
/**
|
||||
* 根据context的上下文情况,填充segmentBuff
|
||||
* @param reader
|
||||
* @return 返回待分析的(有效的)字串长度
|
||||
* @throws IOException
|
||||
*/
|
||||
int fillBuffer(Reader reader) throws IOException {
|
||||
int readCount = 0;
|
||||
if (this.buffOffset == 0) {
|
||||
// 首次读取reader
|
||||
readCount = reader.read(segmentBuff);
|
||||
} else {
|
||||
int offset = this.available - this.cursor;
|
||||
if (offset > 0) {
|
||||
// 最近一次读取的>最近一次处理的,将未处理的字串拷贝到segmentBuff头部
|
||||
System.arraycopy(this.segmentBuff, this.cursor, this.segmentBuff, 0, offset);
|
||||
readCount = offset;
|
||||
}
|
||||
// 继续读取reader ,以onceReadIn - onceAnalyzed为起始位置,继续填充segmentBuff剩余的部分
|
||||
readCount += reader.read(this.segmentBuff, offset, BUFF_SIZE - offset);
|
||||
}
|
||||
// 记录最后一次从Reader中读入的可用字符长度
|
||||
this.available = readCount;
|
||||
// 重置当前指针
|
||||
this.cursor = 0;
|
||||
return readCount;
|
||||
}
|
||||
|
||||
/**
|
||||
* 初始化buff指针,处理第一个字符
|
||||
*/
|
||||
void initCursor() {
|
||||
this.cursor = 0;
|
||||
this.segmentBuff[this.cursor] = CharacterUtil.regularize(this.segmentBuff[this.cursor]);
|
||||
this.charTypes[this.cursor] = CharacterUtil.identifyCharType(this.segmentBuff[this.cursor]);
|
||||
}
|
||||
|
||||
/**
|
||||
* 指针+1
|
||||
* 成功返回 true; 指针已经到了buff尾部,不能前进,返回false
|
||||
* 并处理当前字符
|
||||
*/
|
||||
boolean moveCursor() {
|
||||
if (this.cursor < this.available - 1) {
|
||||
this.cursor++;
|
||||
this.segmentBuff[this.cursor] = CharacterUtil.regularize(this.segmentBuff[this.cursor]);
|
||||
this.charTypes[this.cursor] = CharacterUtil.identifyCharType(this.segmentBuff[this.cursor]);
|
||||
return true;
|
||||
} else {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 设置当前segmentBuff为锁定状态
|
||||
* 加入占用segmentBuff的子分词器名称,表示占用segmentBuff
|
||||
* @param segmenterName
|
||||
*/
|
||||
void lockBuffer(String segmenterName) {
|
||||
this.buffLocker.add(segmenterName);
|
||||
}
|
||||
|
||||
/**
|
||||
* 移除指定的子分词器名,释放对segmentBuff的占用
|
||||
* @param segmenterName
|
||||
*/
|
||||
void unlockBuffer(String segmenterName) {
|
||||
this.buffLocker.remove(segmenterName);
|
||||
}
|
||||
|
||||
/**
|
||||
* 只要buffLocker中存在segmenterName
|
||||
* 则buffer被锁定
|
||||
* @return boolean 缓冲去是否被锁定
|
||||
*/
|
||||
boolean isBufferLocked() {
|
||||
return this.buffLocker.size() > 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* 判断当前segmentBuff是否已经用完
|
||||
* 当前执针cursor移至segmentBuff末端this.available - 1
|
||||
* @return
|
||||
*/
|
||||
boolean isBufferConsumed() {
|
||||
return this.cursor == this.available - 1;
|
||||
}
|
||||
|
||||
/**
|
||||
* 判断segmentBuff是否需要读取新数据
|
||||
*
|
||||
* 满足一下条件时,
|
||||
* 1.available == BUFF_SIZE 表示buffer满载
|
||||
* 2.buffIndex < available - 1 && buffIndex > available - BUFF_EXHAUST_CRITICAL表示当前指针处于临界区内
|
||||
* 3.!context.isBufferLocked()表示没有segmenter在占用buffer
|
||||
* 要中断当前循环(buffer要进行移位,并再读取数据的操作)
|
||||
* @return
|
||||
*/
|
||||
boolean needRefillBuffer() {
|
||||
return this.available == BUFF_SIZE && this.cursor < this.available - 1
|
||||
&& this.cursor > this.available - BUFF_EXHAUST_CRITICAL && !this.isBufferLocked();
|
||||
}
|
||||
|
||||
/**
|
||||
* 累计当前的segmentBuff相对于reader起始位置的位移
|
||||
*/
|
||||
void markBufferOffset() {
|
||||
this.buffOffset += this.cursor;
|
||||
}
|
||||
|
||||
/**
|
||||
* 向分词结果集添加词元
|
||||
* @param lexeme
|
||||
*/
|
||||
void addLexeme(Lexeme lexeme) {
|
||||
this.orgLexemes.addLexeme(lexeme);
|
||||
}
|
||||
|
||||
/**
|
||||
* 添加分词结果路径
|
||||
* 路径起始位置 ---> 路径 映射表
|
||||
* @param path
|
||||
*/
|
||||
void addLexemePath(LexemePath path) {
|
||||
if (path != null) {
|
||||
this.pathMap.put(path.getPathBegin(), path);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 返回原始分词结果
|
||||
* @return
|
||||
*/
|
||||
QuickSortSet getOrgLexemes() {
|
||||
return this.orgLexemes;
|
||||
}
|
||||
|
||||
/**
|
||||
* 推送分词结果到结果集合
|
||||
* 1.从buff头部遍历到this.cursor已处理位置
|
||||
* 2.将map中存在的分词结果推入results
|
||||
* 3.将map中不存在的CJDK字符以单字方式推入results
|
||||
*/
|
||||
void outputToResult() {
|
||||
int index = 0;
|
||||
for (; index <= this.cursor;) {
|
||||
// 跳过非CJK字符
|
||||
if (CharacterUtil.CHAR_USELESS == this.charTypes[index]) {
|
||||
index++;
|
||||
continue;
|
||||
}
|
||||
// 从pathMap找出对应index位置的LexemePath
|
||||
LexemePath path = this.pathMap.get(index);
|
||||
if (path != null) {
|
||||
// 输出LexemePath中的lexeme到results集合
|
||||
Lexeme l = path.pollFirst();
|
||||
while (l != null) {
|
||||
this.results.add(l);
|
||||
// 将index移至lexeme后
|
||||
index = l.getBegin() + l.getLength();
|
||||
l = path.pollFirst();
|
||||
if (l != null) {
|
||||
// 输出path内部,词元间遗漏的单字
|
||||
for (; index < l.getBegin(); index++) {
|
||||
this.outputSingleCJK(index);
|
||||
}
|
||||
}
|
||||
}
|
||||
} else {// pathMap中找不到index对应的LexemePath
|
||||
// 单字输出
|
||||
this.outputSingleCJK(index);
|
||||
index++;
|
||||
}
|
||||
}
|
||||
// 清空当前的Map
|
||||
this.pathMap.clear();
|
||||
}
|
||||
|
||||
/**
|
||||
* 对CJK字符进行单字输出
|
||||
* @param index
|
||||
*/
|
||||
private void outputSingleCJK(int index) {
|
||||
if (CharacterUtil.CHAR_CHINESE == this.charTypes[index]) {
|
||||
Lexeme singleCharLexeme = new Lexeme(this.buffOffset, index, 1, Lexeme.TYPE_CNCHAR);
|
||||
this.results.add(singleCharLexeme);
|
||||
} else if (CharacterUtil.CHAR_OTHER_CJK == this.charTypes[index]) {
|
||||
Lexeme singleCharLexeme = new Lexeme(this.buffOffset, index, 1, Lexeme.TYPE_OTHER_CJK);
|
||||
this.results.add(singleCharLexeme);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 返回lexeme
|
||||
*
|
||||
* 同时处理合并
|
||||
* @return
|
||||
*/
|
||||
Lexeme getNextLexeme() {
|
||||
// 从结果集取出,并移除第一个Lexme
|
||||
Lexeme result = this.results.pollFirst();
|
||||
while (result != null) {
|
||||
// 数量词合并
|
||||
this.compound(result);
|
||||
if (Dictionary.getSingleton().isStopWord(this.segmentBuff, result.getBegin(),
|
||||
result.getLength())) {
|
||||
// 是停止词继续取列表的下一个
|
||||
result = this.results.pollFirst();
|
||||
} else {
|
||||
// 不是停止词, 生成lexeme的词元文本,输出
|
||||
result.setLexemeText(String.valueOf(segmentBuff, result.getBegin(), result.getLength()));
|
||||
break;
|
||||
}
|
||||
}
|
||||
return result;
|
||||
}
|
||||
|
||||
/**
|
||||
* 重置分词上下文状态
|
||||
*/
|
||||
void reset() {
|
||||
this.buffLocker.clear();
|
||||
this.orgLexemes = new QuickSortSet();
|
||||
this.available = 0;
|
||||
this.buffOffset = 0;
|
||||
this.charTypes = new int[BUFF_SIZE];
|
||||
this.cursor = 0;
|
||||
this.results.clear();
|
||||
this.segmentBuff = new char[BUFF_SIZE];
|
||||
this.pathMap.clear();
|
||||
}
|
||||
|
||||
/**
|
||||
* 组合词元
|
||||
*/
|
||||
private void compound(Lexeme result) {
|
||||
if (!this.cfg.useSmart()) {
|
||||
return;
|
||||
}
|
||||
// 数量词合并处理
|
||||
if (!this.results.isEmpty()) {
|
||||
|
||||
if (Lexeme.TYPE_ARABIC == result.getLexemeType()) {
|
||||
Lexeme nextLexeme = this.results.peekFirst();
|
||||
boolean appendOk = false;
|
||||
if (Lexeme.TYPE_CNUM == nextLexeme.getLexemeType()) {
|
||||
// 合并英文数词+中文数词
|
||||
appendOk = result.append(nextLexeme, Lexeme.TYPE_CNUM);
|
||||
} else if (Lexeme.TYPE_COUNT == nextLexeme.getLexemeType()) {
|
||||
// 合并英文数词+中文量词
|
||||
appendOk = result.append(nextLexeme, Lexeme.TYPE_CQUAN);
|
||||
}
|
||||
if (appendOk) {
|
||||
// 弹出
|
||||
this.results.pollFirst();
|
||||
}
|
||||
}
|
||||
|
||||
// 可能存在第二轮合并
|
||||
if (Lexeme.TYPE_CNUM == result.getLexemeType() && !this.results.isEmpty()) {
|
||||
Lexeme nextLexeme = this.results.peekFirst();
|
||||
boolean appendOk = false;
|
||||
if (Lexeme.TYPE_COUNT == nextLexeme.getLexemeType()) {
|
||||
// 合并中文数词+中文量词
|
||||
appendOk = result.append(nextLexeme, Lexeme.TYPE_CQUAN);
|
||||
}
|
||||
if (appendOk) {
|
||||
// 弹出
|
||||
this.results.pollFirst();
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
|
||||
}
|
132
src/main/java/com/rymcu/forest/lucene/core/CJKSegmenter.java
Executable file
132
src/main/java/com/rymcu/forest/lucene/core/CJKSegmenter.java
Executable file
@ -0,0 +1,132 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.core;
|
||||
|
||||
|
||||
|
||||
import com.rymcu.forest.lucene.dic.Dictionary;
|
||||
import com.rymcu.forest.lucene.dic.Hit;
|
||||
|
||||
import java.util.LinkedList;
|
||||
import java.util.List;
|
||||
|
||||
/**
|
||||
* 中文-日韩文子分词器
|
||||
*/
|
||||
class CJKSegmenter implements ISegmenter {
|
||||
|
||||
// 子分词器标签
|
||||
static final String SEGMENTER_NAME = "CJK_SEGMENTER";
|
||||
// 待处理的分词hit队列
|
||||
private List<Hit> tmpHits;
|
||||
|
||||
CJKSegmenter() {
|
||||
this.tmpHits = new LinkedList<Hit>();
|
||||
}
|
||||
|
||||
/*
|
||||
* (non-Javadoc)
|
||||
* @see org.wltea.analyzer.core.ISegmenter#analyze(org.wltea.analyzer.core.AnalyzeContext)
|
||||
*/
|
||||
@Override
|
||||
public void analyze(AnalyzeContext context) {
|
||||
if (CharacterUtil.CHAR_USELESS != context.getCurrentCharType()) {
|
||||
|
||||
// 优先处理tmpHits中的hit
|
||||
if (!this.tmpHits.isEmpty()) {
|
||||
// 处理词段队列
|
||||
Hit[] tmpArray = this.tmpHits.toArray(new Hit[this.tmpHits.size()]);
|
||||
for (Hit hit : tmpArray) {
|
||||
hit = Dictionary.getSingleton().matchWithHit(context.getSegmentBuff(),
|
||||
context.getCursor(), hit);
|
||||
if (hit.isMatch()) {
|
||||
// 输出当前的词
|
||||
Lexeme newLexeme = new Lexeme(context.getBufferOffset(), hit.getBegin(),
|
||||
context.getCursor() - hit.getBegin() + 1, Lexeme.TYPE_CNWORD);
|
||||
context.addLexeme(newLexeme);
|
||||
|
||||
if (!hit.isPrefix()) {// 不是词前缀,hit不需要继续匹配,移除
|
||||
this.tmpHits.remove(hit);
|
||||
}
|
||||
|
||||
} else if (hit.isUnmatch()) {
|
||||
// hit不是词,移除
|
||||
this.tmpHits.remove(hit);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// *********************************
|
||||
// 再对当前指针位置的字符进行单字匹配
|
||||
Hit singleCharHit = Dictionary.getSingleton().matchInMainDict(context.getSegmentBuff(),
|
||||
context.getCursor(), 1);
|
||||
if (singleCharHit.isMatch()) {// 首字成词
|
||||
// 输出当前的词
|
||||
Lexeme newLexeme = new Lexeme(context.getBufferOffset(), context.getCursor(), 1,
|
||||
Lexeme.TYPE_CNWORD);
|
||||
context.addLexeme(newLexeme);
|
||||
|
||||
// 同时也是词前缀
|
||||
if (singleCharHit.isPrefix()) {
|
||||
// 前缀匹配则放入hit列表
|
||||
this.tmpHits.add(singleCharHit);
|
||||
}
|
||||
} else if (singleCharHit.isPrefix()) {// 首字为词前缀
|
||||
// 前缀匹配则放入hit列表
|
||||
this.tmpHits.add(singleCharHit);
|
||||
}
|
||||
|
||||
} else {
|
||||
// 遇到CHAR_USELESS字符
|
||||
// 清空队列
|
||||
this.tmpHits.clear();
|
||||
}
|
||||
|
||||
// 判断缓冲区是否已经读完
|
||||
if (context.isBufferConsumed()) {
|
||||
// 清空队列
|
||||
this.tmpHits.clear();
|
||||
}
|
||||
|
||||
// 判断是否锁定缓冲区
|
||||
if (this.tmpHits.size() == 0) {
|
||||
context.unlockBuffer(SEGMENTER_NAME);
|
||||
|
||||
} else {
|
||||
context.lockBuffer(SEGMENTER_NAME);
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* (non-Javadoc)
|
||||
* @see org.wltea.analyzer.core.ISegmenter#reset()
|
||||
*/
|
||||
@Override
|
||||
public void reset() {
|
||||
// 清空队列
|
||||
this.tmpHits.clear();
|
||||
}
|
||||
|
||||
}
|
245
src/main/java/com/rymcu/forest/lucene/core/CN_QuantifierSegmenter.java
Executable file
245
src/main/java/com/rymcu/forest/lucene/core/CN_QuantifierSegmenter.java
Executable file
@ -0,0 +1,245 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.core;
|
||||
|
||||
|
||||
import com.rymcu.forest.lucene.dic.Dictionary;
|
||||
import com.rymcu.forest.lucene.dic.Hit;
|
||||
|
||||
import java.util.HashSet;
|
||||
import java.util.LinkedList;
|
||||
import java.util.List;
|
||||
import java.util.Set;
|
||||
|
||||
|
||||
/**
|
||||
*
|
||||
* 中文数量词子分词器
|
||||
*/
|
||||
class CN_QuantifierSegmenter implements ISegmenter {
|
||||
|
||||
// 子分词器标签
|
||||
static final String SEGMENTER_NAME = "QUAN_SEGMENTER";
|
||||
|
||||
// 中文数词
|
||||
private static String Chn_Num = "一二两三四五六七八九十零壹贰叁肆伍陆柒捌玖拾百千万亿拾佰仟萬億兆卅廿";// Cnum
|
||||
private static Set<Character> ChnNumberChars = new HashSet<Character>();
|
||||
static {
|
||||
char[] ca = Chn_Num.toCharArray();
|
||||
for (char nChar : ca) {
|
||||
ChnNumberChars.add(nChar);
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* 词元的开始位置, 同时作为子分词器状态标识 当start > -1 时,标识当前的分词器正在处理字符
|
||||
*/
|
||||
private int nStart;
|
||||
/*
|
||||
* 记录词元结束位置 end记录的是在词元中最后一个出现的合理的数词结束
|
||||
*/
|
||||
private int nEnd;
|
||||
|
||||
// 待处理的量词hit队列
|
||||
private List<Hit> countHits;
|
||||
|
||||
CN_QuantifierSegmenter() {
|
||||
nStart = -1;
|
||||
nEnd = -1;
|
||||
this.countHits = new LinkedList<Hit>();
|
||||
}
|
||||
|
||||
/**
|
||||
* 分词
|
||||
*/
|
||||
@Override
|
||||
public void analyze(AnalyzeContext context) {
|
||||
// 处理中文数词
|
||||
this.processCNumber(context);
|
||||
// 处理中文量词
|
||||
this.processCount(context);
|
||||
|
||||
// 判断是否锁定缓冲区
|
||||
if (this.nStart == -1 && this.nEnd == -1 && countHits.isEmpty()) {
|
||||
// 对缓冲区解锁
|
||||
context.unlockBuffer(SEGMENTER_NAME);
|
||||
} else {
|
||||
context.lockBuffer(SEGMENTER_NAME);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 重置子分词器状态
|
||||
*/
|
||||
@Override
|
||||
public void reset() {
|
||||
nStart = -1;
|
||||
nEnd = -1;
|
||||
countHits.clear();
|
||||
}
|
||||
|
||||
/**
|
||||
* 处理数词
|
||||
*/
|
||||
private void processCNumber(AnalyzeContext context) {
|
||||
if (nStart == -1 && nEnd == -1) {// 初始状态
|
||||
if (CharacterUtil.CHAR_CHINESE == context.getCurrentCharType()
|
||||
&& ChnNumberChars.contains(context.getCurrentChar())) {
|
||||
// 记录数词的起始、结束位置
|
||||
nStart = context.getCursor();
|
||||
nEnd = context.getCursor();
|
||||
}
|
||||
} else {// 正在处理状态
|
||||
if (CharacterUtil.CHAR_CHINESE == context.getCurrentCharType()
|
||||
&& ChnNumberChars.contains(context.getCurrentChar())) {
|
||||
// 记录数词的结束位置
|
||||
nEnd = context.getCursor();
|
||||
} else {
|
||||
// 输出数词
|
||||
this.outputNumLexeme(context);
|
||||
// 重置头尾指针
|
||||
nStart = -1;
|
||||
nEnd = -1;
|
||||
}
|
||||
}
|
||||
|
||||
// 缓冲区已经用完,还有尚未输出的数词
|
||||
if (context.isBufferConsumed()) {
|
||||
if (nStart != -1 && nEnd != -1) {
|
||||
// 输出数词
|
||||
outputNumLexeme(context);
|
||||
// 重置头尾指针
|
||||
nStart = -1;
|
||||
nEnd = -1;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 处理中文量词
|
||||
* @param context
|
||||
*/
|
||||
private void processCount(AnalyzeContext context) {
|
||||
// 判断是否需要启动量词扫描
|
||||
if (!this.needCountScan(context)) {
|
||||
return;
|
||||
}
|
||||
|
||||
if (CharacterUtil.CHAR_CHINESE == context.getCurrentCharType()) {
|
||||
|
||||
// 优先处理countHits中的hit
|
||||
if (!this.countHits.isEmpty()) {
|
||||
// 处理词段队列
|
||||
Hit[] tmpArray = this.countHits.toArray(new Hit[this.countHits.size()]);
|
||||
for (Hit hit : tmpArray) {
|
||||
hit = Dictionary.getSingleton().matchWithHit(context.getSegmentBuff(),
|
||||
context.getCursor(), hit);
|
||||
if (hit.isMatch()) {
|
||||
// 输出当前的词
|
||||
Lexeme newLexeme = new Lexeme(context.getBufferOffset(), hit.getBegin(),
|
||||
context.getCursor() - hit.getBegin() + 1, Lexeme.TYPE_COUNT);
|
||||
context.addLexeme(newLexeme);
|
||||
|
||||
if (!hit.isPrefix()) {// 不是词前缀,hit不需要继续匹配,移除
|
||||
this.countHits.remove(hit);
|
||||
}
|
||||
|
||||
} else if (hit.isUnmatch()) {
|
||||
// hit不是词,移除
|
||||
this.countHits.remove(hit);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
// *********************************
|
||||
// 对当前指针位置的字符进行单字匹配
|
||||
Hit singleCharHit = Dictionary.getSingleton().matchInQuantifierDict(context.getSegmentBuff(),
|
||||
context.getCursor(), 1);
|
||||
if (singleCharHit.isMatch()) {// 首字成量词词
|
||||
// 输出当前的词
|
||||
Lexeme newLexeme = new Lexeme(context.getBufferOffset(), context.getCursor(), 1,
|
||||
Lexeme.TYPE_COUNT);
|
||||
context.addLexeme(newLexeme);
|
||||
|
||||
// 同时也是词前缀
|
||||
if (singleCharHit.isPrefix()) {
|
||||
// 前缀匹配则放入hit列表
|
||||
this.countHits.add(singleCharHit);
|
||||
}
|
||||
} else if (singleCharHit.isPrefix()) {// 首字为量词前缀
|
||||
// 前缀匹配则放入hit列表
|
||||
this.countHits.add(singleCharHit);
|
||||
}
|
||||
|
||||
} else {
|
||||
// 输入的不是中文字符
|
||||
// 清空未成形的量词
|
||||
this.countHits.clear();
|
||||
}
|
||||
|
||||
// 缓冲区数据已经读完,还有尚未输出的量词
|
||||
if (context.isBufferConsumed()) {
|
||||
// 清空未成形的量词
|
||||
this.countHits.clear();
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 判断是否需要扫描量词
|
||||
* @return
|
||||
*/
|
||||
private boolean needCountScan(AnalyzeContext context) {
|
||||
if ((nStart != -1 && nEnd != -1) || !countHits.isEmpty()) {
|
||||
// 正在处理中文数词,或者正在处理量词
|
||||
return true;
|
||||
} else {
|
||||
// 找到一个相邻的数词
|
||||
if (!context.getOrgLexemes().isEmpty()) {
|
||||
Lexeme l = context.getOrgLexemes().peekLast();
|
||||
if (Lexeme.TYPE_CNUM == l.getLexemeType() || Lexeme.TYPE_ARABIC == l.getLexemeType()) {
|
||||
if (l.getBegin() + l.getLength() == context.getCursor()) {
|
||||
return true;
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
/**
|
||||
* 添加数词词元到结果集
|
||||
* @param context
|
||||
*/
|
||||
private void outputNumLexeme(AnalyzeContext context) {
|
||||
if (nStart > -1 && nEnd > -1) {
|
||||
// 输出数词
|
||||
Lexeme newLexeme = new Lexeme(context.getBufferOffset(), nStart, nEnd - nStart + 1,
|
||||
Lexeme.TYPE_CNUM);
|
||||
context.addLexeme(newLexeme);
|
||||
|
||||
}
|
||||
}
|
||||
|
||||
}
|
100
src/main/java/com/rymcu/forest/lucene/core/CharacterUtil.java
Executable file
100
src/main/java/com/rymcu/forest/lucene/core/CharacterUtil.java
Executable file
@ -0,0 +1,100 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
* 字符集识别工具类
|
||||
*/
|
||||
package com.rymcu.forest.lucene.core;
|
||||
|
||||
/**
|
||||
*
|
||||
* 字符集识别工具类
|
||||
*/
|
||||
class CharacterUtil {
|
||||
|
||||
public static final int CHAR_USELESS = 0;
|
||||
|
||||
public static final int CHAR_ARABIC = 0X00000001;
|
||||
|
||||
public static final int CHAR_ENGLISH = 0X00000002;
|
||||
|
||||
public static final int CHAR_CHINESE = 0X00000004;
|
||||
|
||||
public static final int CHAR_OTHER_CJK = 0X00000008;
|
||||
|
||||
/**
|
||||
* 识别字符类型
|
||||
* @param input
|
||||
* @return int CharacterUtil定义的字符类型常量
|
||||
*/
|
||||
static int identifyCharType(char input) {
|
||||
if (input >= '0' && input <= '9') {
|
||||
return CHAR_ARABIC;
|
||||
|
||||
} else if ((input >= 'a' && input <= 'z') || (input >= 'A' && input <= 'Z')) {
|
||||
return CHAR_ENGLISH;
|
||||
|
||||
} else {
|
||||
Character.UnicodeBlock ub = Character.UnicodeBlock.of(input);
|
||||
|
||||
if (ub == Character.UnicodeBlock.CJK_UNIFIED_IDEOGRAPHS
|
||||
|| ub == Character.UnicodeBlock.CJK_COMPATIBILITY_IDEOGRAPHS
|
||||
|| ub == Character.UnicodeBlock.CJK_UNIFIED_IDEOGRAPHS_EXTENSION_A) {
|
||||
// 目前已知的中文字符UTF-8集合
|
||||
return CHAR_CHINESE;
|
||||
|
||||
} else if (ub == Character.UnicodeBlock.HALFWIDTH_AND_FULLWIDTH_FORMS // 全角数字字符和日韩字符
|
||||
// 韩文字符集
|
||||
|| ub == Character.UnicodeBlock.HANGUL_SYLLABLES
|
||||
|| ub == Character.UnicodeBlock.HANGUL_JAMO
|
||||
|| ub == Character.UnicodeBlock.HANGUL_COMPATIBILITY_JAMO
|
||||
// 日文字符集
|
||||
|| ub == Character.UnicodeBlock.HIRAGANA // 平假名
|
||||
|| ub == Character.UnicodeBlock.KATAKANA // 片假名
|
||||
|| ub == Character.UnicodeBlock.KATAKANA_PHONETIC_EXTENSIONS) {
|
||||
return CHAR_OTHER_CJK;
|
||||
|
||||
}
|
||||
}
|
||||
// 其他的不做处理的字符
|
||||
return CHAR_USELESS;
|
||||
}
|
||||
|
||||
/**
|
||||
* 进行字符规格化(全角转半角,大写转小写处理)
|
||||
* @param input
|
||||
* @return char
|
||||
*/
|
||||
static char regularize(char input) {
|
||||
if (input == 12288) {
|
||||
input = (char) 32;
|
||||
|
||||
} else if (input > 65280 && input < 65375) {
|
||||
input = (char) (input - 65248);
|
||||
|
||||
} else if (input >= 'A' && input <= 'Z') {
|
||||
input += 32;
|
||||
}
|
||||
|
||||
return input;
|
||||
}
|
||||
}
|
152
src/main/java/com/rymcu/forest/lucene/core/IKArbitrator.java
Executable file
152
src/main/java/com/rymcu/forest/lucene/core/IKArbitrator.java
Executable file
@ -0,0 +1,152 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.core;
|
||||
|
||||
import java.util.Stack;
|
||||
import java.util.TreeSet;
|
||||
|
||||
/**
|
||||
* IK分词歧义裁决器
|
||||
*/
|
||||
class IKArbitrator {
|
||||
|
||||
IKArbitrator() {
|
||||
|
||||
}
|
||||
|
||||
/**
|
||||
* 分词歧义处理
|
||||
* @param orgLexemes
|
||||
* @param useSmart
|
||||
*/
|
||||
void process(AnalyzeContext context, boolean useSmart) {
|
||||
QuickSortSet orgLexemes = context.getOrgLexemes();
|
||||
Lexeme orgLexeme = orgLexemes.pollFirst();
|
||||
|
||||
LexemePath crossPath = new LexemePath();
|
||||
while (orgLexeme != null) {
|
||||
if (!crossPath.addCrossLexeme(orgLexeme)) {
|
||||
// 找到与crossPath不相交的下一个crossPath
|
||||
if (crossPath.size() == 1 || !useSmart) {
|
||||
// crossPath没有歧义 或者 不做歧义处理
|
||||
// 直接输出当前crossPath
|
||||
context.addLexemePath(crossPath);
|
||||
} else {
|
||||
// 对当前的crossPath进行歧义处理
|
||||
QuickSortSet.Cell headCell = crossPath.getHead();
|
||||
LexemePath judgeResult = this.judge(headCell, crossPath.getPathLength());
|
||||
// 输出歧义处理结果judgeResult
|
||||
context.addLexemePath(judgeResult);
|
||||
}
|
||||
|
||||
// 把orgLexeme加入新的crossPath中
|
||||
crossPath = new LexemePath();
|
||||
crossPath.addCrossLexeme(orgLexeme);
|
||||
}
|
||||
orgLexeme = orgLexemes.pollFirst();
|
||||
}
|
||||
|
||||
// 处理最后的path
|
||||
if (crossPath.size() == 1 || !useSmart) {
|
||||
// crossPath没有歧义 或者 不做歧义处理
|
||||
// 直接输出当前crossPath
|
||||
context.addLexemePath(crossPath);
|
||||
} else {
|
||||
// 对当前的crossPath进行歧义处理
|
||||
QuickSortSet.Cell headCell = crossPath.getHead();
|
||||
LexemePath judgeResult = this.judge(headCell, crossPath.getPathLength());
|
||||
// 输出歧义处理结果judgeResult
|
||||
context.addLexemePath(judgeResult);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 歧义识别
|
||||
* @param lexemeCell 歧义路径链表头
|
||||
* @param fullTextLength 歧义路径文本长度
|
||||
* @param option 候选结果路径
|
||||
* @return
|
||||
*/
|
||||
private LexemePath judge(QuickSortSet.Cell lexemeCell, int fullTextLength) {
|
||||
// 候选路径集合
|
||||
TreeSet<LexemePath> pathOptions = new TreeSet<LexemePath>();
|
||||
// 候选结果路径
|
||||
LexemePath option = new LexemePath();
|
||||
|
||||
// 对crossPath进行一次遍历,同时返回本次遍历中有冲突的Lexeme栈
|
||||
Stack<QuickSortSet.Cell> lexemeStack = this.forwardPath(lexemeCell, option);
|
||||
|
||||
// 当前词元链并非最理想的,加入候选路径集合
|
||||
pathOptions.add(option.copy());
|
||||
|
||||
// 存在歧义词,处理
|
||||
QuickSortSet.Cell c = null;
|
||||
while (!lexemeStack.isEmpty()) {
|
||||
c = lexemeStack.pop();
|
||||
// 回滚词元链
|
||||
this.backPath(c.getLexeme(), option);
|
||||
// 从歧义词位置开始,递归,生成可选方案
|
||||
this.forwardPath(c, option);
|
||||
pathOptions.add(option.copy());
|
||||
}
|
||||
|
||||
// 返回集合中的最优方案
|
||||
return pathOptions.first();
|
||||
|
||||
}
|
||||
|
||||
/**
|
||||
* 向前遍历,添加词元,构造一个无歧义词元组合
|
||||
* @param LexemePath path
|
||||
* @return
|
||||
*/
|
||||
private Stack<QuickSortSet.Cell> forwardPath(QuickSortSet.Cell lexemeCell, LexemePath option) {
|
||||
// 发生冲突的Lexeme栈
|
||||
Stack<QuickSortSet.Cell> conflictStack = new Stack<QuickSortSet.Cell>();
|
||||
QuickSortSet.Cell c = lexemeCell;
|
||||
// 迭代遍历Lexeme链表
|
||||
while (c != null && c.getLexeme() != null) {
|
||||
if (!option.addNotCrossLexeme(c.getLexeme())) {
|
||||
// 词元交叉,添加失败则加入lexemeStack栈
|
||||
conflictStack.push(c);
|
||||
}
|
||||
c = c.getNext();
|
||||
}
|
||||
return conflictStack;
|
||||
}
|
||||
|
||||
/**
|
||||
* 回滚词元链,直到它能够接受指定的词元
|
||||
* @param lexeme
|
||||
* @param l
|
||||
*/
|
||||
private void backPath(Lexeme l, LexemePath option) {
|
||||
while (option.checkCross(l)) {
|
||||
option.removeTail();
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
}
|
165
src/main/java/com/rymcu/forest/lucene/core/IKSegmenter.java
Executable file
165
src/main/java/com/rymcu/forest/lucene/core/IKSegmenter.java
Executable file
@ -0,0 +1,165 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*/
|
||||
package com.rymcu.forest.lucene.core;
|
||||
|
||||
import com.rymcu.forest.lucene.cfg.Configuration;
|
||||
import com.rymcu.forest.lucene.cfg.DefaultConfig;
|
||||
import com.rymcu.forest.lucene.dic.Dictionary;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.Reader;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
|
||||
/**
|
||||
* IK分词器主类
|
||||
*
|
||||
*/
|
||||
public final class IKSegmenter {
|
||||
|
||||
// 字符窜reader
|
||||
private Reader input;
|
||||
// 分词器配置项
|
||||
private Configuration cfg;
|
||||
// 分词器上下文
|
||||
private AnalyzeContext context;
|
||||
// 分词处理器列表
|
||||
private List<ISegmenter> segmenters;
|
||||
// 分词歧义裁决器
|
||||
private IKArbitrator arbitrator;
|
||||
|
||||
/**
|
||||
* IK分词器构造函数
|
||||
* @param input
|
||||
* @param useSmart 为true,使用智能分词策略
|
||||
*
|
||||
* 非智能分词:细粒度输出所有可能的切分结果
|
||||
* 智能分词: 合并数词和量词,对分词结果进行歧义判断
|
||||
*/
|
||||
public IKSegmenter(Reader input, boolean useSmart) {
|
||||
this.input = input;
|
||||
this.cfg = DefaultConfig.getInstance();
|
||||
this.cfg.setUseSmart(useSmart);
|
||||
this.init();
|
||||
}
|
||||
|
||||
/**
|
||||
* IK分词器构造函数
|
||||
* @param input
|
||||
* @param cfg 使用自定义的Configuration构造分词器
|
||||
*
|
||||
*/
|
||||
public IKSegmenter(Reader input, Configuration cfg) {
|
||||
this.input = input;
|
||||
this.cfg = cfg;
|
||||
this.init();
|
||||
}
|
||||
|
||||
/**
|
||||
* 初始化
|
||||
*/
|
||||
private void init() {
|
||||
// 初始化词典单例
|
||||
Dictionary.initial(this.cfg);
|
||||
// 初始化分词上下文
|
||||
this.context = new AnalyzeContext(this.cfg);
|
||||
// 加载子分词器
|
||||
this.segmenters = this.loadSegmenters();
|
||||
// 加载歧义裁决器
|
||||
this.arbitrator = new IKArbitrator();
|
||||
}
|
||||
|
||||
/**
|
||||
* 初始化词典,加载子分词器实现
|
||||
* @return List<ISegmenter>
|
||||
*/
|
||||
private List<ISegmenter> loadSegmenters() {
|
||||
List<ISegmenter> segmenters = new ArrayList<ISegmenter>(4);
|
||||
// 处理字母的子分词器
|
||||
segmenters.add(new LetterSegmenter());
|
||||
// 处理中文数量词的子分词器
|
||||
segmenters.add(new CN_QuantifierSegmenter());
|
||||
// 处理中文词的子分词器
|
||||
segmenters.add(new CJKSegmenter());
|
||||
return segmenters;
|
||||
}
|
||||
|
||||
/**
|
||||
* 分词,获取下一个词元
|
||||
* @return Lexeme 词元对象
|
||||
* @throws IOException
|
||||
*/
|
||||
public synchronized Lexeme next() throws IOException {
|
||||
Lexeme l = null;
|
||||
while ((l = context.getNextLexeme()) == null) {
|
||||
/*
|
||||
* 从reader中读取数据,填充buffer 如果reader是分次读入buffer的,那么buffer要 进行移位处理 移位处理上次读入的但未处理的数据
|
||||
*/
|
||||
int available = context.fillBuffer(this.input);
|
||||
if (available <= 0) {
|
||||
// reader已经读完
|
||||
context.reset();
|
||||
return null;
|
||||
|
||||
} else {
|
||||
// 初始化指针
|
||||
context.initCursor();
|
||||
do {
|
||||
// 遍历子分词器
|
||||
for (ISegmenter segmenter : segmenters) {
|
||||
segmenter.analyze(context);
|
||||
}
|
||||
// 字符缓冲区接近读完,需要读入新的字符
|
||||
if (context.needRefillBuffer()) {
|
||||
break;
|
||||
}
|
||||
// 向前移动指针
|
||||
} while (context.moveCursor());
|
||||
// 重置子分词器,为下轮循环进行初始化
|
||||
for (ISegmenter segmenter : segmenters) {
|
||||
segmenter.reset();
|
||||
}
|
||||
}
|
||||
// 对分词进行歧义处理
|
||||
this.arbitrator.process(context, this.cfg.useSmart());
|
||||
// 将分词结果输出到结果集,并处理未切分的单个CJK字符
|
||||
context.outputToResult();
|
||||
// 记录本次分词的缓冲区位移
|
||||
context.markBufferOffset();
|
||||
}
|
||||
return l;
|
||||
}
|
||||
|
||||
/**
|
||||
* 重置分词器到初始状态
|
||||
* @param input
|
||||
*/
|
||||
public synchronized void reset(Reader input) {
|
||||
this.input = input;
|
||||
context.reset();
|
||||
for (ISegmenter segmenter : segmenters) {
|
||||
segmenter.reset();
|
||||
}
|
||||
}
|
||||
}
|
44
src/main/java/com/rymcu/forest/lucene/core/ISegmenter.java
Executable file
44
src/main/java/com/rymcu/forest/lucene/core/ISegmenter.java
Executable file
@ -0,0 +1,44 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.core;
|
||||
|
||||
/**
|
||||
*
|
||||
* 子分词器接口
|
||||
*/
|
||||
interface ISegmenter {
|
||||
|
||||
/**
|
||||
* 从分析器读取下一个可能分解的词元对象
|
||||
* @param context 分词算法上下文
|
||||
*/
|
||||
void analyze(AnalyzeContext context);
|
||||
|
||||
/**
|
||||
* 重置子分析器状态
|
||||
*/
|
||||
void reset();
|
||||
|
||||
}
|
302
src/main/java/com/rymcu/forest/lucene/core/LetterSegmenter.java
Executable file
302
src/main/java/com/rymcu/forest/lucene/core/LetterSegmenter.java
Executable file
@ -0,0 +1,302 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.core;
|
||||
|
||||
import java.util.Arrays;
|
||||
|
||||
/**
|
||||
*
|
||||
* 英文字符及阿拉伯数字子分词器
|
||||
*/
|
||||
class LetterSegmenter implements ISegmenter {
|
||||
|
||||
// 子分词器标签
|
||||
static final String SEGMENTER_NAME = "LETTER_SEGMENTER";
|
||||
// 链接符号
|
||||
private static final char[] Letter_Connector = new char[] { '#', '&', '+', '-', '.', '@', '_' };
|
||||
|
||||
// 数字符号
|
||||
private static final char[] Num_Connector = new char[] { ',', '.' };
|
||||
|
||||
/*
|
||||
* 词元的开始位置, 同时作为子分词器状态标识 当start > -1 时,标识当前的分词器正在处理字符
|
||||
*/
|
||||
private int start;
|
||||
/*
|
||||
* 记录词元结束位置 end记录的是在词元中最后一个出现的Letter但非Sign_Connector的字符的位置
|
||||
*/
|
||||
private int end;
|
||||
|
||||
/*
|
||||
* 字母起始位置
|
||||
*/
|
||||
private int englishStart;
|
||||
|
||||
/*
|
||||
* 字母结束位置
|
||||
*/
|
||||
private int englishEnd;
|
||||
|
||||
/*
|
||||
* 阿拉伯数字起始位置
|
||||
*/
|
||||
private int arabicStart;
|
||||
|
||||
/*
|
||||
* 阿拉伯数字结束位置
|
||||
*/
|
||||
private int arabicEnd;
|
||||
|
||||
LetterSegmenter() {
|
||||
Arrays.sort(Letter_Connector);
|
||||
Arrays.sort(Num_Connector);
|
||||
this.start = -1;
|
||||
this.end = -1;
|
||||
this.englishStart = -1;
|
||||
this.englishEnd = -1;
|
||||
this.arabicStart = -1;
|
||||
this.arabicEnd = -1;
|
||||
}
|
||||
|
||||
/*
|
||||
* (non-Javadoc)
|
||||
* @see org.wltea.analyzer.core.ISegmenter#analyze(org.wltea.analyzer.core.AnalyzeContext)
|
||||
*/
|
||||
@Override
|
||||
public void analyze(AnalyzeContext context) {
|
||||
boolean bufferLockFlag = false;
|
||||
// 处理英文字母
|
||||
bufferLockFlag = this.processEnglishLetter(context) || bufferLockFlag;
|
||||
// 处理阿拉伯字母
|
||||
bufferLockFlag = this.processArabicLetter(context) || bufferLockFlag;
|
||||
// 处理混合字母(这个要放最后处理,可以通过QuickSortSet排除重复)
|
||||
bufferLockFlag = this.processMixLetter(context) || bufferLockFlag;
|
||||
|
||||
// 判断是否锁定缓冲区
|
||||
if (bufferLockFlag) {
|
||||
context.lockBuffer(SEGMENTER_NAME);
|
||||
} else {
|
||||
// 对缓冲区解锁
|
||||
context.unlockBuffer(SEGMENTER_NAME);
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* (non-Javadoc)
|
||||
* @see org.wltea.analyzer.core.ISegmenter#reset()
|
||||
*/
|
||||
@Override
|
||||
public void reset() {
|
||||
this.start = -1;
|
||||
this.end = -1;
|
||||
this.englishStart = -1;
|
||||
this.englishEnd = -1;
|
||||
this.arabicStart = -1;
|
||||
this.arabicEnd = -1;
|
||||
}
|
||||
|
||||
/**
|
||||
* 处理数字字母混合输出
|
||||
* 如:windos2000 | linliangyi2005@gmail.com
|
||||
* @param input
|
||||
* @param context
|
||||
* @return
|
||||
*/
|
||||
private boolean processMixLetter(AnalyzeContext context) {
|
||||
boolean needLock = false;
|
||||
|
||||
if (this.start == -1) {// 当前的分词器尚未开始处理字符
|
||||
if (CharacterUtil.CHAR_ARABIC == context.getCurrentCharType()
|
||||
|| CharacterUtil.CHAR_ENGLISH == context.getCurrentCharType()) {
|
||||
// 记录起始指针的位置,标明分词器进入处理状态
|
||||
this.start = context.getCursor();
|
||||
this.end = start;
|
||||
}
|
||||
|
||||
} else {// 当前的分词器正在处理字符
|
||||
if (CharacterUtil.CHAR_ARABIC == context.getCurrentCharType()
|
||||
|| CharacterUtil.CHAR_ENGLISH == context.getCurrentCharType()) {
|
||||
// 记录下可能的结束位置
|
||||
this.end = context.getCursor();
|
||||
|
||||
} else if (CharacterUtil.CHAR_USELESS == context.getCurrentCharType()
|
||||
&& this.isLetterConnector(context.getCurrentChar())) {
|
||||
// 记录下可能的结束位置
|
||||
this.end = context.getCursor();
|
||||
} else {
|
||||
// 遇到非Letter字符,输出词元
|
||||
Lexeme newLexeme = new Lexeme(context.getBufferOffset(), this.start, this.end - this.start
|
||||
+ 1, Lexeme.TYPE_LETTER);
|
||||
context.addLexeme(newLexeme);
|
||||
this.start = -1;
|
||||
this.end = -1;
|
||||
}
|
||||
}
|
||||
|
||||
// 判断缓冲区是否已经读完
|
||||
if (context.isBufferConsumed()) {
|
||||
if (this.start != -1 && this.end != -1) {
|
||||
// 缓冲以读完,输出词元
|
||||
Lexeme newLexeme = new Lexeme(context.getBufferOffset(), this.start, this.end - this.start
|
||||
+ 1, Lexeme.TYPE_LETTER);
|
||||
context.addLexeme(newLexeme);
|
||||
this.start = -1;
|
||||
this.end = -1;
|
||||
}
|
||||
}
|
||||
|
||||
// 判断是否锁定缓冲区
|
||||
if (this.start == -1 && this.end == -1) {
|
||||
// 对缓冲区解锁
|
||||
needLock = false;
|
||||
} else {
|
||||
needLock = true;
|
||||
}
|
||||
return needLock;
|
||||
}
|
||||
|
||||
/**
|
||||
* 处理纯英文字母输出
|
||||
* @param context
|
||||
* @return
|
||||
*/
|
||||
private boolean processEnglishLetter(AnalyzeContext context) {
|
||||
boolean needLock = false;
|
||||
|
||||
if (this.englishStart == -1) {// 当前的分词器尚未开始处理英文字符
|
||||
if (CharacterUtil.CHAR_ENGLISH == context.getCurrentCharType()) {
|
||||
// 记录起始指针的位置,标明分词器进入处理状态
|
||||
this.englishStart = context.getCursor();
|
||||
this.englishEnd = this.englishStart;
|
||||
}
|
||||
} else {// 当前的分词器正在处理英文字符
|
||||
if (CharacterUtil.CHAR_ENGLISH == context.getCurrentCharType()) {
|
||||
// 记录当前指针位置为结束位置
|
||||
this.englishEnd = context.getCursor();
|
||||
} else {
|
||||
// 遇到非English字符,输出词元
|
||||
Lexeme newLexeme = new Lexeme(context.getBufferOffset(), this.englishStart, this.englishEnd
|
||||
- this.englishStart + 1, Lexeme.TYPE_ENGLISH);
|
||||
context.addLexeme(newLexeme);
|
||||
this.englishStart = -1;
|
||||
this.englishEnd = -1;
|
||||
}
|
||||
}
|
||||
|
||||
// 判断缓冲区是否已经读完
|
||||
if (context.isBufferConsumed()) {
|
||||
if (this.englishStart != -1 && this.englishEnd != -1) {
|
||||
// 缓冲以读完,输出词元
|
||||
Lexeme newLexeme = new Lexeme(context.getBufferOffset(), this.englishStart, this.englishEnd
|
||||
- this.englishStart + 1, Lexeme.TYPE_ENGLISH);
|
||||
context.addLexeme(newLexeme);
|
||||
this.englishStart = -1;
|
||||
this.englishEnd = -1;
|
||||
}
|
||||
}
|
||||
|
||||
// 判断是否锁定缓冲区
|
||||
if (this.englishStart == -1 && this.englishEnd == -1) {
|
||||
// 对缓冲区解锁
|
||||
needLock = false;
|
||||
} else {
|
||||
needLock = true;
|
||||
}
|
||||
return needLock;
|
||||
}
|
||||
|
||||
/**
|
||||
* 处理阿拉伯数字输出
|
||||
* @param context
|
||||
* @return
|
||||
*/
|
||||
private boolean processArabicLetter(AnalyzeContext context) {
|
||||
boolean needLock = false;
|
||||
|
||||
if (this.arabicStart == -1) {// 当前的分词器尚未开始处理数字字符
|
||||
if (CharacterUtil.CHAR_ARABIC == context.getCurrentCharType()) {
|
||||
// 记录起始指针的位置,标明分词器进入处理状态
|
||||
this.arabicStart = context.getCursor();
|
||||
this.arabicEnd = this.arabicStart;
|
||||
}
|
||||
} else {// 当前的分词器正在处理数字字符
|
||||
if (CharacterUtil.CHAR_ARABIC == context.getCurrentCharType()) {
|
||||
// 记录当前指针位置为结束位置
|
||||
this.arabicEnd = context.getCursor();
|
||||
} else if (CharacterUtil.CHAR_USELESS == context.getCurrentCharType()
|
||||
&& this.isNumConnector(context.getCurrentChar())) {
|
||||
// 不输出数字,但不标记结束
|
||||
} else {
|
||||
// //遇到非Arabic字符,输出词元
|
||||
Lexeme newLexeme = new Lexeme(context.getBufferOffset(), this.arabicStart, this.arabicEnd
|
||||
- this.arabicStart + 1, Lexeme.TYPE_ARABIC);
|
||||
context.addLexeme(newLexeme);
|
||||
this.arabicStart = -1;
|
||||
this.arabicEnd = -1;
|
||||
}
|
||||
}
|
||||
|
||||
// 判断缓冲区是否已经读完
|
||||
if (context.isBufferConsumed()) {
|
||||
if (this.arabicStart != -1 && this.arabicEnd != -1) {
|
||||
// 生成已切分的词元
|
||||
Lexeme newLexeme = new Lexeme(context.getBufferOffset(), this.arabicStart, this.arabicEnd
|
||||
- this.arabicStart + 1, Lexeme.TYPE_ARABIC);
|
||||
context.addLexeme(newLexeme);
|
||||
this.arabicStart = -1;
|
||||
this.arabicEnd = -1;
|
||||
}
|
||||
}
|
||||
|
||||
// 判断是否锁定缓冲区
|
||||
if (this.arabicStart == -1 && this.arabicEnd == -1) {
|
||||
// 对缓冲区解锁
|
||||
needLock = false;
|
||||
} else {
|
||||
needLock = true;
|
||||
}
|
||||
return needLock;
|
||||
}
|
||||
|
||||
/**
|
||||
* 判断是否是字母连接符号
|
||||
* @param input
|
||||
* @return
|
||||
*/
|
||||
private boolean isLetterConnector(char input) {
|
||||
int index = Arrays.binarySearch(Letter_Connector, input);
|
||||
return index >= 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* 判断是否是数字连接符号
|
||||
* @param input
|
||||
* @return
|
||||
*/
|
||||
private boolean isNumConnector(char input) {
|
||||
int index = Arrays.binarySearch(Num_Connector, input);
|
||||
return index >= 0;
|
||||
}
|
||||
}
|
279
src/main/java/com/rymcu/forest/lucene/core/Lexeme.java
Executable file
279
src/main/java/com/rymcu/forest/lucene/core/Lexeme.java
Executable file
@ -0,0 +1,279 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.core;
|
||||
|
||||
/**
|
||||
* IK词元对象
|
||||
*/
|
||||
public class Lexeme implements Comparable<Lexeme> {
|
||||
// lexemeType常量
|
||||
// 未知
|
||||
public static final int TYPE_UNKNOWN = 0;
|
||||
// 英文
|
||||
public static final int TYPE_ENGLISH = 1;
|
||||
// 数字
|
||||
public static final int TYPE_ARABIC = 2;
|
||||
// 英文数字混合
|
||||
public static final int TYPE_LETTER = 3;
|
||||
// 中文词元
|
||||
public static final int TYPE_CNWORD = 4;
|
||||
// 中文单字
|
||||
public static final int TYPE_CNCHAR = 64;
|
||||
// 日韩文字
|
||||
public static final int TYPE_OTHER_CJK = 8;
|
||||
// 中文数词
|
||||
public static final int TYPE_CNUM = 16;
|
||||
// 中文量词
|
||||
public static final int TYPE_COUNT = 32;
|
||||
// 中文数量词
|
||||
public static final int TYPE_CQUAN = 48;
|
||||
|
||||
// 词元的起始位移
|
||||
private int offset;
|
||||
// 词元的相对起始位置
|
||||
private int begin;
|
||||
// 词元的长度
|
||||
private int length;
|
||||
// 词元文本
|
||||
private String lexemeText;
|
||||
// 词元类型
|
||||
private int lexemeType;
|
||||
|
||||
public Lexeme(int offset, int begin, int length, int lexemeType) {
|
||||
this.offset = offset;
|
||||
this.begin = begin;
|
||||
if (length < 0) {
|
||||
throw new IllegalArgumentException("length < 0");
|
||||
}
|
||||
this.length = length;
|
||||
this.lexemeType = lexemeType;
|
||||
}
|
||||
|
||||
/*
|
||||
* 判断词元相等算法 起始位置偏移、起始位置、终止位置相同
|
||||
* @see java.lang.Object#equals(Object o)
|
||||
*/
|
||||
public boolean equals(Object o) {
|
||||
if (o == null) {
|
||||
return false;
|
||||
}
|
||||
|
||||
if (this == o) {
|
||||
return true;
|
||||
}
|
||||
|
||||
if (o instanceof Lexeme) {
|
||||
Lexeme other = (Lexeme) o;
|
||||
if (this.offset == other.getOffset() && this.begin == other.getBegin()
|
||||
&& this.length == other.getLength()) {
|
||||
return true;
|
||||
} else {
|
||||
return false;
|
||||
}
|
||||
} else {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* 词元哈希编码算法
|
||||
* @see java.lang.Object#hashCode()
|
||||
*/
|
||||
public int hashCode() {
|
||||
int absBegin = getBeginPosition();
|
||||
int absEnd = getEndPosition();
|
||||
return (absBegin * 37) + (absEnd * 31) + ((absBegin * absEnd) % getLength()) * 11;
|
||||
}
|
||||
|
||||
/*
|
||||
* 词元在排序集合中的比较算法
|
||||
* @see java.lang.Comparable#compareTo(java.lang.Object)
|
||||
*/
|
||||
public int compareTo(Lexeme other) {
|
||||
// 起始位置优先
|
||||
if (this.begin < other.getBegin()) {
|
||||
return -1;
|
||||
} else if (this.begin == other.getBegin()) {
|
||||
// 词元长度优先
|
||||
if (this.length > other.getLength()) {
|
||||
return -1;
|
||||
} else if (this.length == other.getLength()) {
|
||||
return 0;
|
||||
} else {// this.length < other.getLength()
|
||||
return 1;
|
||||
}
|
||||
|
||||
} else {// this.begin > other.getBegin()
|
||||
return 1;
|
||||
}
|
||||
}
|
||||
|
||||
public int getOffset() {
|
||||
return offset;
|
||||
}
|
||||
|
||||
public void setOffset(int offset) {
|
||||
this.offset = offset;
|
||||
}
|
||||
|
||||
public int getBegin() {
|
||||
return begin;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取词元在文本中的起始位置
|
||||
* @return int
|
||||
*/
|
||||
public int getBeginPosition() {
|
||||
return offset + begin;
|
||||
}
|
||||
|
||||
public void setBegin(int begin) {
|
||||
this.begin = begin;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取词元在文本中的结束位置
|
||||
* @return int
|
||||
*/
|
||||
public int getEndPosition() {
|
||||
return offset + begin + length;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取词元的字符长度
|
||||
* @return int
|
||||
*/
|
||||
public int getLength() {
|
||||
return this.length;
|
||||
}
|
||||
|
||||
public void setLength(int length) {
|
||||
if (this.length < 0) {
|
||||
throw new IllegalArgumentException("length < 0");
|
||||
}
|
||||
this.length = length;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取词元的文本内容
|
||||
* @return String
|
||||
*/
|
||||
public String getLexemeText() {
|
||||
if (lexemeText == null) {
|
||||
return "";
|
||||
}
|
||||
return lexemeText;
|
||||
}
|
||||
|
||||
public void setLexemeText(String lexemeText) {
|
||||
if (lexemeText == null) {
|
||||
this.lexemeText = "";
|
||||
this.length = 0;
|
||||
} else {
|
||||
this.lexemeText = lexemeText;
|
||||
this.length = lexemeText.length();
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取词元类型
|
||||
* @return int
|
||||
*/
|
||||
public int getLexemeType() {
|
||||
return lexemeType;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取词元类型标示字符串
|
||||
* @return String
|
||||
*/
|
||||
public String getLexemeTypeString() {
|
||||
switch (lexemeType) {
|
||||
|
||||
case TYPE_ENGLISH:
|
||||
return "ENGLISH";
|
||||
|
||||
case TYPE_ARABIC:
|
||||
return "ARABIC";
|
||||
|
||||
case TYPE_LETTER:
|
||||
return "LETTER";
|
||||
|
||||
case TYPE_CNWORD:
|
||||
return "CN_WORD";
|
||||
|
||||
case TYPE_CNCHAR:
|
||||
return "CN_CHAR";
|
||||
|
||||
case TYPE_OTHER_CJK:
|
||||
return "OTHER_CJK";
|
||||
|
||||
case TYPE_COUNT:
|
||||
return "COUNT";
|
||||
|
||||
case TYPE_CNUM:
|
||||
return "TYPE_CNUM";
|
||||
|
||||
case TYPE_CQUAN:
|
||||
return "TYPE_CQUAN";
|
||||
|
||||
default:
|
||||
return "UNKONW";
|
||||
}
|
||||
}
|
||||
|
||||
public void setLexemeType(int lexemeType) {
|
||||
this.lexemeType = lexemeType;
|
||||
}
|
||||
|
||||
/**
|
||||
* 合并两个相邻的词元
|
||||
* @param l
|
||||
* @param lexemeType
|
||||
* @return boolean 词元是否成功合并
|
||||
*/
|
||||
public boolean append(Lexeme l, int lexemeType) {
|
||||
if (l != null && this.getEndPosition() == l.getBeginPosition()) {
|
||||
this.length += l.getLength();
|
||||
this.lexemeType = lexemeType;
|
||||
return true;
|
||||
} else {
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
*
|
||||
*/
|
||||
public String toString() {
|
||||
StringBuffer strbuf = new StringBuffer();
|
||||
strbuf.append(this.getBeginPosition()).append("-").append(this.getEndPosition());
|
||||
strbuf.append(" : ").append(this.lexemeText).append(" : \t");
|
||||
strbuf.append(this.getLexemeTypeString());
|
||||
return strbuf.toString();
|
||||
}
|
||||
|
||||
}
|
255
src/main/java/com/rymcu/forest/lucene/core/LexemePath.java
Executable file
255
src/main/java/com/rymcu/forest/lucene/core/LexemePath.java
Executable file
@ -0,0 +1,255 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.core;
|
||||
|
||||
/**
|
||||
* Lexeme链(路径)
|
||||
*/
|
||||
class LexemePath extends QuickSortSet implements Comparable<LexemePath> {
|
||||
|
||||
// 起始位置
|
||||
private int pathBegin;
|
||||
// 结束
|
||||
private int pathEnd;
|
||||
// 词元链的有效字符长度
|
||||
private int payloadLength;
|
||||
|
||||
LexemePath() {
|
||||
this.pathBegin = -1;
|
||||
this.pathEnd = -1;
|
||||
this.payloadLength = 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* 向LexemePath追加相交的Lexeme
|
||||
* @param lexeme
|
||||
* @return
|
||||
*/
|
||||
boolean addCrossLexeme(Lexeme lexeme) {
|
||||
if (this.isEmpty()) {
|
||||
this.addLexeme(lexeme);
|
||||
this.pathBegin = lexeme.getBegin();
|
||||
this.pathEnd = lexeme.getBegin() + lexeme.getLength();
|
||||
this.payloadLength += lexeme.getLength();
|
||||
return true;
|
||||
|
||||
} else if (this.checkCross(lexeme)) {
|
||||
this.addLexeme(lexeme);
|
||||
if (lexeme.getBegin() + lexeme.getLength() > this.pathEnd) {
|
||||
this.pathEnd = lexeme.getBegin() + lexeme.getLength();
|
||||
}
|
||||
this.payloadLength = this.pathEnd - this.pathBegin;
|
||||
return true;
|
||||
|
||||
} else {
|
||||
return false;
|
||||
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 向LexemePath追加不相交的Lexeme
|
||||
* @param lexeme
|
||||
* @return
|
||||
*/
|
||||
boolean addNotCrossLexeme(Lexeme lexeme) {
|
||||
if (this.isEmpty()) {
|
||||
this.addLexeme(lexeme);
|
||||
this.pathBegin = lexeme.getBegin();
|
||||
this.pathEnd = lexeme.getBegin() + lexeme.getLength();
|
||||
this.payloadLength += lexeme.getLength();
|
||||
return true;
|
||||
|
||||
} else if (this.checkCross(lexeme)) {
|
||||
return false;
|
||||
|
||||
} else {
|
||||
this.addLexeme(lexeme);
|
||||
this.payloadLength += lexeme.getLength();
|
||||
Lexeme head = this.peekFirst();
|
||||
this.pathBegin = head.getBegin();
|
||||
Lexeme tail = this.peekLast();
|
||||
this.pathEnd = tail.getBegin() + tail.getLength();
|
||||
return true;
|
||||
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 移除尾部的Lexeme
|
||||
* @return
|
||||
*/
|
||||
Lexeme removeTail() {
|
||||
Lexeme tail = this.pollLast();
|
||||
if (this.isEmpty()) {
|
||||
this.pathBegin = -1;
|
||||
this.pathEnd = -1;
|
||||
this.payloadLength = 0;
|
||||
} else {
|
||||
this.payloadLength -= tail.getLength();
|
||||
Lexeme newTail = this.peekLast();
|
||||
this.pathEnd = newTail.getBegin() + newTail.getLength();
|
||||
}
|
||||
return tail;
|
||||
}
|
||||
|
||||
/**
|
||||
* 检测词元位置交叉(有歧义的切分)
|
||||
* @param lexeme
|
||||
* @return
|
||||
*/
|
||||
boolean checkCross(Lexeme lexeme) {
|
||||
return (lexeme.getBegin() >= this.pathBegin && lexeme.getBegin() < this.pathEnd)
|
||||
|| (this.pathBegin >= lexeme.getBegin() && this.pathBegin < lexeme.getBegin()
|
||||
+ lexeme.getLength());
|
||||
}
|
||||
|
||||
int getPathBegin() {
|
||||
return pathBegin;
|
||||
}
|
||||
|
||||
int getPathEnd() {
|
||||
return pathEnd;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取Path的有效词长
|
||||
* @return
|
||||
*/
|
||||
int getPayloadLength() {
|
||||
return this.payloadLength;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取LexemePath的路径长度
|
||||
* @return
|
||||
*/
|
||||
int getPathLength() {
|
||||
return this.pathEnd - this.pathBegin;
|
||||
}
|
||||
|
||||
/**
|
||||
* X权重(词元长度积)
|
||||
* @return
|
||||
*/
|
||||
int getXWeight() {
|
||||
int product = 1;
|
||||
Cell c = this.getHead();
|
||||
while (c != null && c.getLexeme() != null) {
|
||||
product *= c.getLexeme().getLength();
|
||||
c = c.getNext();
|
||||
}
|
||||
return product;
|
||||
}
|
||||
|
||||
/**
|
||||
* 词元位置权重
|
||||
* @return
|
||||
*/
|
||||
int getPWeight() {
|
||||
int pWeight = 0;
|
||||
int p = 0;
|
||||
Cell c = this.getHead();
|
||||
while (c != null && c.getLexeme() != null) {
|
||||
p++;
|
||||
pWeight += p * c.getLexeme().getLength();
|
||||
c = c.getNext();
|
||||
}
|
||||
return pWeight;
|
||||
}
|
||||
|
||||
LexemePath copy() {
|
||||
LexemePath theCopy = new LexemePath();
|
||||
theCopy.pathBegin = this.pathBegin;
|
||||
theCopy.pathEnd = this.pathEnd;
|
||||
theCopy.payloadLength = this.payloadLength;
|
||||
Cell c = this.getHead();
|
||||
while (c != null && c.getLexeme() != null) {
|
||||
theCopy.addLexeme(c.getLexeme());
|
||||
c = c.getNext();
|
||||
}
|
||||
return theCopy;
|
||||
}
|
||||
|
||||
public int compareTo(LexemePath o) {
|
||||
// 比较有效文本长度
|
||||
if (this.payloadLength > o.payloadLength) {
|
||||
return -1;
|
||||
} else if (this.payloadLength < o.payloadLength) {
|
||||
return 1;
|
||||
} else {
|
||||
// 比较词元个数,越少越好
|
||||
if (this.size() < o.size()) {
|
||||
return -1;
|
||||
} else if (this.size() > o.size()) {
|
||||
return 1;
|
||||
} else {
|
||||
// 路径跨度越大越好
|
||||
if (this.getPathLength() > o.getPathLength()) {
|
||||
return -1;
|
||||
} else if (this.getPathLength() < o.getPathLength()) {
|
||||
return 1;
|
||||
} else {
|
||||
// 根据统计学结论,逆向切分概率高于正向切分,因此位置越靠后的优先
|
||||
if (this.pathEnd > o.pathEnd) {
|
||||
return -1;
|
||||
} else if (pathEnd < o.pathEnd) {
|
||||
return 1;
|
||||
} else {
|
||||
// 词长越平均越好
|
||||
if (this.getXWeight() > o.getXWeight()) {
|
||||
return -1;
|
||||
} else if (this.getXWeight() < o.getXWeight()) {
|
||||
return 1;
|
||||
} else {
|
||||
// 词元位置权重比较
|
||||
if (this.getPWeight() > o.getPWeight()) {
|
||||
return -1;
|
||||
} else if (this.getPWeight() < o.getPWeight()) {
|
||||
return 1;
|
||||
}
|
||||
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
public String toString() {
|
||||
StringBuffer sb = new StringBuffer();
|
||||
sb.append("pathBegin : ").append(pathBegin).append("\r\n");
|
||||
sb.append("pathEnd : ").append(pathEnd).append("\r\n");
|
||||
sb.append("payloadLength : ").append(payloadLength).append("\r\n");
|
||||
Cell head = this.getHead();
|
||||
while (head != null) {
|
||||
sb.append("lexeme : ").append(head.getLexeme()).append("\r\n");
|
||||
head = head.getNext();
|
||||
}
|
||||
return sb.toString();
|
||||
}
|
||||
|
||||
}
|
239
src/main/java/com/rymcu/forest/lucene/core/QuickSortSet.java
Executable file
239
src/main/java/com/rymcu/forest/lucene/core/QuickSortSet.java
Executable file
@ -0,0 +1,239 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.core;
|
||||
|
||||
/**
|
||||
* IK分词器专用的Lexem快速排序集合
|
||||
*/
|
||||
class QuickSortSet {
|
||||
// 链表头
|
||||
private Cell head;
|
||||
// 链表尾
|
||||
private Cell tail;
|
||||
// 链表的实际大小
|
||||
private int size;
|
||||
|
||||
QuickSortSet() {
|
||||
this.size = 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* 向链表集合添加词元
|
||||
* @param lexeme
|
||||
*/
|
||||
boolean addLexeme(Lexeme lexeme) {
|
||||
Cell newCell = new Cell(lexeme);
|
||||
if (this.size == 0) {
|
||||
this.head = newCell;
|
||||
this.tail = newCell;
|
||||
this.size++;
|
||||
return true;
|
||||
|
||||
} else {
|
||||
if (this.tail.compareTo(newCell) == 0) {// 词元与尾部词元相同,不放入集合
|
||||
return false;
|
||||
|
||||
} else if (this.tail.compareTo(newCell) < 0) {// 词元接入链表尾部
|
||||
this.tail.next = newCell;
|
||||
newCell.prev = this.tail;
|
||||
this.tail = newCell;
|
||||
this.size++;
|
||||
return true;
|
||||
|
||||
} else if (this.head.compareTo(newCell) > 0) {// 词元接入链表头部
|
||||
this.head.prev = newCell;
|
||||
newCell.next = this.head;
|
||||
this.head = newCell;
|
||||
this.size++;
|
||||
return true;
|
||||
|
||||
} else {
|
||||
// 从尾部上逆
|
||||
Cell index = this.tail;
|
||||
while (index != null && index.compareTo(newCell) > 0) {
|
||||
index = index.prev;
|
||||
}
|
||||
if (index.compareTo(newCell) == 0) {// 词元与集合中的词元重复,不放入集合
|
||||
return false;
|
||||
|
||||
} else if (index.compareTo(newCell) < 0) {// 词元插入链表中的某个位置
|
||||
newCell.prev = index;
|
||||
newCell.next = index.next;
|
||||
index.next.prev = newCell;
|
||||
index.next = newCell;
|
||||
this.size++;
|
||||
return true;
|
||||
}
|
||||
}
|
||||
}
|
||||
return false;
|
||||
}
|
||||
|
||||
/**
|
||||
* 返回链表头部元素
|
||||
* @return
|
||||
*/
|
||||
Lexeme peekFirst() {
|
||||
if (this.head != null) {
|
||||
return this.head.lexeme;
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
||||
/**
|
||||
* 取出链表集合的第一个元素
|
||||
* @return Lexeme
|
||||
*/
|
||||
Lexeme pollFirst() {
|
||||
if (this.size == 1) {
|
||||
Lexeme first = this.head.lexeme;
|
||||
this.head = null;
|
||||
this.tail = null;
|
||||
this.size--;
|
||||
return first;
|
||||
} else if (this.size > 1) {
|
||||
Lexeme first = this.head.lexeme;
|
||||
this.head = this.head.next;
|
||||
this.size--;
|
||||
return first;
|
||||
} else {
|
||||
return null;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 返回链表尾部元素
|
||||
* @return
|
||||
*/
|
||||
Lexeme peekLast() {
|
||||
if (this.tail != null) {
|
||||
return this.tail.lexeme;
|
||||
}
|
||||
return null;
|
||||
}
|
||||
|
||||
/**
|
||||
* 取出链表集合的最后一个元素
|
||||
* @return Lexeme
|
||||
*/
|
||||
Lexeme pollLast() {
|
||||
if (this.size == 1) {
|
||||
Lexeme last = this.head.lexeme;
|
||||
this.head = null;
|
||||
this.tail = null;
|
||||
this.size--;
|
||||
return last;
|
||||
|
||||
} else if (this.size > 1) {
|
||||
Lexeme last = this.tail.lexeme;
|
||||
this.tail = this.tail.prev;
|
||||
this.size--;
|
||||
return last;
|
||||
|
||||
} else {
|
||||
return null;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 返回集合大小
|
||||
* @return
|
||||
*/
|
||||
int size() {
|
||||
return this.size;
|
||||
}
|
||||
|
||||
/**
|
||||
* 判断集合是否为空
|
||||
* @return
|
||||
*/
|
||||
boolean isEmpty() {
|
||||
return this.size == 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* 返回lexeme链的头部
|
||||
* @return
|
||||
*/
|
||||
Cell getHead() {
|
||||
return this.head;
|
||||
}
|
||||
|
||||
/**
|
||||
*
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
* QuickSortSet集合单元
|
||||
*
|
||||
*/
|
||||
class Cell implements Comparable<Cell> {
|
||||
private Cell prev;
|
||||
private Cell next;
|
||||
private Lexeme lexeme;
|
||||
|
||||
Cell(Lexeme lexeme) {
|
||||
if (lexeme == null) {
|
||||
throw new IllegalArgumentException("lexeme must not be null");
|
||||
}
|
||||
this.lexeme = lexeme;
|
||||
}
|
||||
|
||||
public int compareTo(Cell o) {
|
||||
return this.lexeme.compareTo(o.lexeme);
|
||||
}
|
||||
|
||||
public Cell getPrev() {
|
||||
return this.prev;
|
||||
}
|
||||
|
||||
public Cell getNext() {
|
||||
return this.next;
|
||||
}
|
||||
|
||||
public Lexeme getLexeme() {
|
||||
return this.lexeme;
|
||||
}
|
||||
}
|
||||
}
|
328
src/main/java/com/rymcu/forest/lucene/dic/DictSegment.java
Executable file
328
src/main/java/com/rymcu/forest/lucene/dic/DictSegment.java
Executable file
@ -0,0 +1,328 @@
|
||||
/**
|
||||
*
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.dic;
|
||||
|
||||
import java.util.Arrays;
|
||||
import java.util.HashMap;
|
||||
import java.util.Map;
|
||||
|
||||
/**
|
||||
* 词典树分段,表示词典树的一个分枝
|
||||
*/
|
||||
class DictSegment implements Comparable<DictSegment> {
|
||||
|
||||
// 公用字典表,存储汉字
|
||||
private static final Map<Character, Character> charMap = new HashMap<Character, Character>(16,
|
||||
0.95f);
|
||||
// 数组大小上限
|
||||
private static final int ARRAY_LENGTH_LIMIT = 3;
|
||||
|
||||
// Map存储结构
|
||||
private Map<Character, DictSegment> childrenMap;
|
||||
// 数组方式存储结构
|
||||
private DictSegment[] childrenArray;
|
||||
|
||||
// 当前节点上存储的字符
|
||||
private Character nodeChar;
|
||||
// 当前节点存储的Segment数目
|
||||
// storeSize <=ARRAY_LENGTH_LIMIT ,使用数组存储, storeSize >ARRAY_LENGTH_LIMIT ,则使用Map存储
|
||||
private int storeSize = 0;
|
||||
// 当前DictSegment状态 ,默认 0 , 1表示从根节点到当前节点的路径表示一个词
|
||||
private int nodeState = 0;
|
||||
|
||||
DictSegment(Character nodeChar) {
|
||||
if (nodeChar == null) {
|
||||
throw new IllegalArgumentException("参数为空异常,字符不能为空");
|
||||
}
|
||||
this.nodeChar = nodeChar;
|
||||
}
|
||||
|
||||
Character getNodeChar() {
|
||||
return nodeChar;
|
||||
}
|
||||
|
||||
/*
|
||||
* 判断是否有下一个节点
|
||||
*/
|
||||
boolean hasNextNode() {
|
||||
return this.storeSize > 0;
|
||||
}
|
||||
|
||||
/**
|
||||
* 匹配词段
|
||||
* @param charArray
|
||||
* @return Hit
|
||||
*/
|
||||
Hit match(char[] charArray) {
|
||||
return this.match(charArray, 0, charArray.length, null);
|
||||
}
|
||||
|
||||
/**
|
||||
* 匹配词段
|
||||
* @param charArray
|
||||
* @param begin
|
||||
* @param length
|
||||
* @return Hit
|
||||
*/
|
||||
Hit match(char[] charArray, int begin, int length) {
|
||||
return this.match(charArray, begin, length, null);
|
||||
}
|
||||
|
||||
/**
|
||||
* 匹配词段
|
||||
* @param charArray
|
||||
* @param begin
|
||||
* @param length
|
||||
* @param searchHit
|
||||
* @return Hit
|
||||
*/
|
||||
Hit match(char[] charArray, int begin, int length, Hit searchHit) {
|
||||
|
||||
if (searchHit == null) {
|
||||
// 如果hit为空,新建
|
||||
searchHit = new Hit();
|
||||
// 设置hit的其实文本位置
|
||||
searchHit.setBegin(begin);
|
||||
} else {
|
||||
// 否则要将HIT状态重置
|
||||
searchHit.setUnmatch();
|
||||
}
|
||||
// 设置hit的当前处理位置
|
||||
searchHit.setEnd(begin);
|
||||
|
||||
Character keyChar = new Character(charArray[begin]);
|
||||
DictSegment ds = null;
|
||||
|
||||
// 引用实例变量为本地变量,避免查询时遇到更新的同步问题
|
||||
DictSegment[] segmentArray = this.childrenArray;
|
||||
Map<Character, DictSegment> segmentMap = this.childrenMap;
|
||||
|
||||
// STEP1 在节点中查找keyChar对应的DictSegment
|
||||
if (segmentArray != null) {
|
||||
// 在数组中查找
|
||||
DictSegment keySegment = new DictSegment(keyChar);
|
||||
int position = Arrays.binarySearch(segmentArray, 0, this.storeSize, keySegment);
|
||||
if (position >= 0) {
|
||||
ds = segmentArray[position];
|
||||
}
|
||||
|
||||
} else if (segmentMap != null) {
|
||||
// 在map中查找
|
||||
ds = (DictSegment) segmentMap.get(keyChar);
|
||||
}
|
||||
|
||||
// STEP2 找到DictSegment,判断词的匹配状态,是否继续递归,还是返回结果
|
||||
if (ds != null) {
|
||||
if (length > 1) {
|
||||
// 词未匹配完,继续往下搜索
|
||||
return ds.match(charArray, begin + 1, length - 1, searchHit);
|
||||
} else if (length == 1) {
|
||||
|
||||
// 搜索最后一个char
|
||||
if (ds.nodeState == 1) {
|
||||
// 添加HIT状态为完全匹配
|
||||
searchHit.setMatch();
|
||||
}
|
||||
if (ds.hasNextNode()) {
|
||||
// 添加HIT状态为前缀匹配
|
||||
searchHit.setPrefix();
|
||||
// 记录当前位置的DictSegment
|
||||
searchHit.setMatchedDictSegment(ds);
|
||||
}
|
||||
return searchHit;
|
||||
}
|
||||
|
||||
}
|
||||
// STEP3 没有找到DictSegment, 将HIT设置为不匹配
|
||||
return searchHit;
|
||||
}
|
||||
|
||||
/**
|
||||
* 加载填充词典片段
|
||||
* @param charArray
|
||||
*/
|
||||
void fillSegment(char[] charArray) {
|
||||
this.fillSegment(charArray, 0, charArray.length, 1);
|
||||
}
|
||||
|
||||
/**
|
||||
* 屏蔽词典中的一个词
|
||||
* @param charArray
|
||||
*/
|
||||
void disableSegment(char[] charArray) {
|
||||
this.fillSegment(charArray, 0, charArray.length, 0);
|
||||
}
|
||||
|
||||
/**
|
||||
* 加载填充词典片段
|
||||
* @param charArray
|
||||
* @param begin
|
||||
* @param length
|
||||
* @param enabled
|
||||
*/
|
||||
private synchronized void fillSegment(char[] charArray, int begin, int length, int enabled) {
|
||||
// 获取字典表中的汉字对象
|
||||
Character beginChar = new Character(charArray[begin]);
|
||||
Character keyChar = charMap.get(beginChar);
|
||||
// 字典中没有该字,则将其添加入字典
|
||||
if (keyChar == null) {
|
||||
charMap.put(beginChar, beginChar);
|
||||
keyChar = beginChar;
|
||||
}
|
||||
|
||||
// 搜索当前节点的存储,查询对应keyChar的keyChar,如果没有则创建
|
||||
DictSegment ds = lookforSegment(keyChar, enabled);
|
||||
if (ds != null) {
|
||||
// 处理keyChar对应的segment
|
||||
if (length > 1) {
|
||||
// 词元还没有完全加入词典树
|
||||
ds.fillSegment(charArray, begin + 1, length - 1, enabled);
|
||||
} else if (length == 1) {
|
||||
// 已经是词元的最后一个char,设置当前节点状态为enabled,
|
||||
// enabled=1表明一个完整的词,enabled=0表示从词典中屏蔽当前词
|
||||
ds.nodeState = enabled;
|
||||
}
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
/**
|
||||
* 查找本节点下对应的keyChar的segment *
|
||||
* @param keyChar
|
||||
* @param create =1如果没有找到,则创建新的segment ; =0如果没有找到,不创建,返回null
|
||||
* @return
|
||||
*/
|
||||
private DictSegment lookforSegment(Character keyChar, int create) {
|
||||
|
||||
DictSegment ds = null;
|
||||
|
||||
if (this.storeSize <= ARRAY_LENGTH_LIMIT) {
|
||||
// 获取数组容器,如果数组未创建则创建数组
|
||||
DictSegment[] segmentArray = getChildrenArray();
|
||||
// 搜寻数组
|
||||
DictSegment keySegment = new DictSegment(keyChar);
|
||||
int position = Arrays.binarySearch(segmentArray, 0, this.storeSize, keySegment);
|
||||
if (position >= 0) {
|
||||
ds = segmentArray[position];
|
||||
}
|
||||
|
||||
// 遍历数组后没有找到对应的segment
|
||||
if (ds == null && create == 1) {
|
||||
ds = keySegment;
|
||||
if (this.storeSize < ARRAY_LENGTH_LIMIT) {
|
||||
// 数组容量未满,使用数组存储
|
||||
segmentArray[this.storeSize] = ds;
|
||||
// segment数目+1
|
||||
this.storeSize++;
|
||||
Arrays.sort(segmentArray, 0, this.storeSize);
|
||||
|
||||
} else {
|
||||
// 数组容量已满,切换Map存储
|
||||
// 获取Map容器,如果Map未创建,则创建Map
|
||||
Map<Character, DictSegment> segmentMap = getChildrenMap();
|
||||
// 将数组中的segment迁移到Map中
|
||||
migrate(segmentArray, segmentMap);
|
||||
// 存储新的segment
|
||||
segmentMap.put(keyChar, ds);
|
||||
// segment数目+1 , 必须在释放数组前执行storeSize++ , 确保极端情况下,不会取到空的数组
|
||||
this.storeSize++;
|
||||
// 释放当前的数组引用
|
||||
this.childrenArray = null;
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
} else {
|
||||
// 获取Map容器,如果Map未创建,则创建Map
|
||||
Map<Character, DictSegment> segmentMap = getChildrenMap();
|
||||
// 搜索Map
|
||||
ds = (DictSegment) segmentMap.get(keyChar);
|
||||
if (ds == null && create == 1) {
|
||||
// 构造新的segment
|
||||
ds = new DictSegment(keyChar);
|
||||
segmentMap.put(keyChar, ds);
|
||||
// 当前节点存储segment数目+1
|
||||
this.storeSize++;
|
||||
}
|
||||
}
|
||||
|
||||
return ds;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取数组容器
|
||||
* 线程同步方法
|
||||
*/
|
||||
private DictSegment[] getChildrenArray() {
|
||||
if (this.childrenArray == null) {
|
||||
synchronized (this) {
|
||||
if (this.childrenArray == null) {
|
||||
this.childrenArray = new DictSegment[ARRAY_LENGTH_LIMIT];
|
||||
}
|
||||
}
|
||||
}
|
||||
return this.childrenArray;
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取Map容器
|
||||
* 线程同步方法
|
||||
*/
|
||||
private Map<Character, DictSegment> getChildrenMap() {
|
||||
if (this.childrenMap == null) {
|
||||
synchronized (this) {
|
||||
if (this.childrenMap == null) {
|
||||
this.childrenMap = new HashMap<Character, DictSegment>(ARRAY_LENGTH_LIMIT * 2, 0.8f);
|
||||
}
|
||||
}
|
||||
}
|
||||
return this.childrenMap;
|
||||
}
|
||||
|
||||
/**
|
||||
* 将数组中的segment迁移到Map中
|
||||
* @param segmentArray
|
||||
*/
|
||||
private void migrate(DictSegment[] segmentArray, Map<Character, DictSegment> segmentMap) {
|
||||
for (DictSegment segment : segmentArray) {
|
||||
if (segment != null) {
|
||||
segmentMap.put(segment.nodeChar, segment);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 实现Comparable接口
|
||||
* @param o
|
||||
* @return int
|
||||
*/
|
||||
@Override
|
||||
public int compareTo(DictSegment o) {
|
||||
// 对当前节点存储的char进行比较
|
||||
return this.nodeChar.compareTo(o.nodeChar);
|
||||
}
|
||||
|
||||
}
|
333
src/main/java/com/rymcu/forest/lucene/dic/Dictionary.java
Executable file
333
src/main/java/com/rymcu/forest/lucene/dic/Dictionary.java
Executable file
@ -0,0 +1,333 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0 IK Analyzer release 5.0
|
||||
*
|
||||
* <p>Licensed to the Apache Software Foundation (ASF) under one or more contributor license
|
||||
* agreements. See the NOTICE file distributed with this work for additional information regarding
|
||||
* copyright ownership. The ASF licenses this file to You under the Apache License, Version 2.0 (the
|
||||
* "License"); you may not use this file except in compliance with the License. You may obtain a
|
||||
* copy of the License at
|
||||
*
|
||||
* <p>http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* <p>Unless required by applicable law or agreed to in writing, software distributed under the
|
||||
* License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either
|
||||
* express or implied. See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* <p>源代码由林良益(linliangyi2005@gmail.com)提供 版权声明 2012,乌龙茶工作室 provided by Linliangyi and copyright 2012
|
||||
* by Oolong studio
|
||||
*/
|
||||
package com.rymcu.forest.lucene.dic;
|
||||
|
||||
import com.rymcu.forest.lucene.cfg.Configuration;
|
||||
import org.springframework.core.io.ClassPathResource;
|
||||
import org.springframework.core.io.Resource;
|
||||
|
||||
import java.io.BufferedReader;
|
||||
import java.io.IOException;
|
||||
import java.io.InputStream;
|
||||
import java.io.InputStreamReader;
|
||||
import java.nio.charset.StandardCharsets;
|
||||
import java.util.Collection;
|
||||
import java.util.List;
|
||||
|
||||
/** 词典管理类,单子模式 */
|
||||
public class Dictionary {
|
||||
|
||||
/*
|
||||
* 词典单子实例
|
||||
*/
|
||||
private static Dictionary singleton;
|
||||
|
||||
/*
|
||||
* 主词典对象
|
||||
*/
|
||||
private DictSegment _MainDict;
|
||||
|
||||
/*
|
||||
* 停止词词典
|
||||
*/
|
||||
private DictSegment _StopWordDict;
|
||||
/*
|
||||
* 量词词典
|
||||
*/
|
||||
private DictSegment _QuantifierDict;
|
||||
|
||||
/** 配置对象 */
|
||||
private Configuration cfg;
|
||||
|
||||
private Dictionary(Configuration cfg) {
|
||||
this.cfg = cfg;
|
||||
this.loadMainDict();
|
||||
this.loadStopWordDict();
|
||||
this.loadQuantifierDict();
|
||||
}
|
||||
|
||||
/**
|
||||
* 词典初始化 由于IK Analyzer的词典采用Dictionary类的静态方法进行词典初始化 只有当Dictionary类被实际调用时,才会开始载入词典, 这将延长首次分词操作的时间
|
||||
* 该方法提供了一个在应用加载阶段就初始化字典的手段
|
||||
*/
|
||||
public static void initial(Configuration cfg) {
|
||||
if (singleton == null) {
|
||||
synchronized (Dictionary.class) {
|
||||
if (singleton == null) {
|
||||
singleton = new Dictionary(cfg);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 获取词典单子实例
|
||||
*
|
||||
* @return Dictionary 单例对象
|
||||
*/
|
||||
public static Dictionary getSingleton() {
|
||||
if (singleton == null) {
|
||||
throw new IllegalStateException("词典尚未初始化,请先调用initial方法");
|
||||
}
|
||||
return singleton;
|
||||
}
|
||||
|
||||
/**
|
||||
* 批量加载新词条
|
||||
*
|
||||
* @param words Collection<String>词条列表
|
||||
*/
|
||||
public void addWords(Collection<String> words) {
|
||||
if (words != null) {
|
||||
for (String word : words) {
|
||||
if (word != null) {
|
||||
// 批量加载词条到主内存词典中
|
||||
singleton._MainDict.fillSegment(word.trim().toLowerCase().toCharArray());
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 批量移除(屏蔽)词条
|
||||
*
|
||||
* @param words
|
||||
*/
|
||||
public void disableWords(Collection<String> words) {
|
||||
if (words != null) {
|
||||
for (String word : words) {
|
||||
if (word != null) {
|
||||
// 批量屏蔽词条
|
||||
singleton._MainDict.disableSegment(word.trim().toLowerCase().toCharArray());
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 检索匹配主词典
|
||||
*
|
||||
* @param charArray
|
||||
* @return Hit 匹配结果描述
|
||||
*/
|
||||
public Hit matchInMainDict(char[] charArray) {
|
||||
return singleton._MainDict.match(charArray);
|
||||
}
|
||||
|
||||
/**
|
||||
* 检索匹配主词典
|
||||
*
|
||||
* @param charArray
|
||||
* @param begin
|
||||
* @param length
|
||||
* @return Hit 匹配结果描述
|
||||
*/
|
||||
public Hit matchInMainDict(char[] charArray, int begin, int length) {
|
||||
return singleton._MainDict.match(charArray, begin, length);
|
||||
}
|
||||
|
||||
/**
|
||||
* 检索匹配量词词典
|
||||
*
|
||||
* @param charArray
|
||||
* @param begin
|
||||
* @param length
|
||||
* @return Hit 匹配结果描述
|
||||
*/
|
||||
public Hit matchInQuantifierDict(char[] charArray, int begin, int length) {
|
||||
return singleton._QuantifierDict.match(charArray, begin, length);
|
||||
}
|
||||
|
||||
/**
|
||||
* 从已匹配的Hit中直接取出DictSegment,继续向下匹配
|
||||
*
|
||||
* @param charArray
|
||||
* @param currentIndex
|
||||
* @param matchedHit
|
||||
* @return Hit
|
||||
*/
|
||||
public Hit matchWithHit(char[] charArray, int currentIndex, Hit matchedHit) {
|
||||
DictSegment ds = matchedHit.getMatchedDictSegment();
|
||||
return ds.match(charArray, currentIndex, 1, matchedHit);
|
||||
}
|
||||
|
||||
/**
|
||||
* 判断是否是停止词
|
||||
*
|
||||
* @param charArray
|
||||
* @param begin
|
||||
* @param length
|
||||
* @return boolean
|
||||
*/
|
||||
public boolean isStopWord(char[] charArray, int begin, int length) {
|
||||
return singleton._StopWordDict.match(charArray, begin, length).isMatch();
|
||||
}
|
||||
|
||||
/** 加载主词典及扩展词典 */
|
||||
private void loadMainDict() {
|
||||
// 建立一个主词典实例
|
||||
_MainDict = new DictSegment((char) 0);
|
||||
// 读取主词典文件
|
||||
Resource resource = new ClassPathResource(cfg.getMainDictionary());
|
||||
try {
|
||||
InputStream is = resource.getInputStream();
|
||||
BufferedReader br =
|
||||
new BufferedReader(new InputStreamReader(is, StandardCharsets.UTF_8), 512);
|
||||
String theWord = null;
|
||||
do {
|
||||
theWord = br.readLine();
|
||||
if (theWord != null && !"".equals(theWord.trim())) {
|
||||
_MainDict.fillSegment(theWord.trim().toLowerCase().toCharArray());
|
||||
}
|
||||
} while (theWord != null);
|
||||
} catch (IOException e) {
|
||||
System.err.println("Main Dictionary loading exception.");
|
||||
e.printStackTrace();
|
||||
}
|
||||
// 加载扩展词典
|
||||
this.loadExtDict();
|
||||
}
|
||||
|
||||
/** 加载用户配置的扩展词典到主词库表 */
|
||||
private void loadExtDict() {
|
||||
// 加载扩展词典配置
|
||||
List<String> extDictFiles = cfg.getExtDictionary();
|
||||
if (extDictFiles != null) {
|
||||
InputStream is = null;
|
||||
for (String extDictName : extDictFiles) {
|
||||
// 读取扩展词典文件
|
||||
System.out.println("加载扩展词典:" + extDictName);
|
||||
is = this.getClass().getClassLoader().getResourceAsStream(extDictName);
|
||||
// 如果找不到扩展的字典,则忽略
|
||||
if (is == null) {
|
||||
continue;
|
||||
}
|
||||
try {
|
||||
BufferedReader br = new BufferedReader(new InputStreamReader(is, "UTF-8"), 512);
|
||||
String theWord = null;
|
||||
do {
|
||||
theWord = br.readLine();
|
||||
if (theWord != null && !"".equals(theWord.trim())) {
|
||||
// 加载扩展词典数据到主内存词典中
|
||||
// System.out.println(theWord);
|
||||
_MainDict.fillSegment(theWord.trim().toLowerCase().toCharArray());
|
||||
}
|
||||
} while (theWord != null);
|
||||
|
||||
} catch (IOException ioe) {
|
||||
System.err.println("Extension Dictionary loading exception.");
|
||||
ioe.printStackTrace();
|
||||
|
||||
} finally {
|
||||
try {
|
||||
if (is != null) {
|
||||
is.close();
|
||||
is = null;
|
||||
}
|
||||
} catch (IOException e) {
|
||||
e.printStackTrace();
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/** 加载用户扩展的停止词词典 */
|
||||
private void loadStopWordDict() {
|
||||
// 建立一个主词典实例
|
||||
_StopWordDict = new DictSegment((char) 0);
|
||||
// 加载扩展停止词典
|
||||
List<String> extStopWordDictFiles = cfg.getExtStopWordDictionary();
|
||||
if (extStopWordDictFiles != null) {
|
||||
InputStream is = null;
|
||||
for (String extStopWordDictName : extStopWordDictFiles) {
|
||||
System.out.println("加载扩展停止词典:" + extStopWordDictName);
|
||||
// 读取扩展词典文件
|
||||
is = this.getClass().getClassLoader().getResourceAsStream(extStopWordDictName);
|
||||
// 如果找不到扩展的字典,则忽略
|
||||
if (is == null) {
|
||||
continue;
|
||||
}
|
||||
try {
|
||||
BufferedReader br = new BufferedReader(new InputStreamReader(is, "UTF-8"), 512);
|
||||
String theWord = null;
|
||||
do {
|
||||
theWord = br.readLine();
|
||||
if (theWord != null && !"".equals(theWord.trim())) {
|
||||
// System.out.println(theWord);
|
||||
// 加载扩展停止词典数据到内存中
|
||||
_StopWordDict.fillSegment(theWord.trim().toLowerCase().toCharArray());
|
||||
}
|
||||
} while (theWord != null);
|
||||
|
||||
} catch (IOException ioe) {
|
||||
System.err.println("Extension Stop word Dictionary loading exception.");
|
||||
ioe.printStackTrace();
|
||||
|
||||
} finally {
|
||||
try {
|
||||
if (is != null) {
|
||||
is.close();
|
||||
is = null;
|
||||
}
|
||||
} catch (IOException e) {
|
||||
e.printStackTrace();
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/** 加载量词词典 */
|
||||
private void loadQuantifierDict() {
|
||||
// 建立一个量词典实例
|
||||
_QuantifierDict = new DictSegment((char) 0);
|
||||
// 读取量词词典文件
|
||||
InputStream is =
|
||||
this.getClass().getClassLoader().getResourceAsStream(cfg.getQuantifierDictionary());
|
||||
if (is == null) {
|
||||
throw new RuntimeException("Quantifier Dictionary not found!!!");
|
||||
}
|
||||
try {
|
||||
BufferedReader br = new BufferedReader(new InputStreamReader(is, "UTF-8"), 512);
|
||||
String theWord = null;
|
||||
do {
|
||||
theWord = br.readLine();
|
||||
if (theWord != null && !"".equals(theWord.trim())) {
|
||||
_QuantifierDict.fillSegment(theWord.trim().toLowerCase().toCharArray());
|
||||
}
|
||||
} while (theWord != null);
|
||||
|
||||
} catch (IOException ioe) {
|
||||
System.err.println("Quantifier Dictionary loading exception.");
|
||||
ioe.printStackTrace();
|
||||
|
||||
} finally {
|
||||
try {
|
||||
if (is != null) {
|
||||
is.close();
|
||||
is = null;
|
||||
}
|
||||
} catch (IOException e) {
|
||||
e.printStackTrace();
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
119
src/main/java/com/rymcu/forest/lucene/dic/Hit.java
Executable file
119
src/main/java/com/rymcu/forest/lucene/dic/Hit.java
Executable file
@ -0,0 +1,119 @@
|
||||
/**
|
||||
*
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.dic;
|
||||
|
||||
/**
|
||||
* 表示一次词典匹配的命中
|
||||
*/
|
||||
public class Hit {
|
||||
// Hit不匹配
|
||||
private static final int UNMATCH = 0x00000000;
|
||||
// Hit完全匹配
|
||||
private static final int MATCH = 0x00000001;
|
||||
// Hit前缀匹配
|
||||
private static final int PREFIX = 0x00000010;
|
||||
|
||||
// 该HIT当前状态,默认未匹配
|
||||
private int hitState = UNMATCH;
|
||||
|
||||
// 记录词典匹配过程中,当前匹配到的词典分支节点
|
||||
private DictSegment matchedDictSegment;
|
||||
/*
|
||||
* 词段开始位置
|
||||
*/
|
||||
private int begin;
|
||||
/*
|
||||
* 词段的结束位置
|
||||
*/
|
||||
private int end;
|
||||
|
||||
/**
|
||||
* 判断是否完全匹配
|
||||
*/
|
||||
public boolean isMatch() {
|
||||
return (this.hitState & MATCH) > 0;
|
||||
}
|
||||
|
||||
/**
|
||||
*
|
||||
*/
|
||||
public void setMatch() {
|
||||
this.hitState = this.hitState | MATCH;
|
||||
}
|
||||
|
||||
/**
|
||||
* 判断是否是词的前缀
|
||||
*/
|
||||
public boolean isPrefix() {
|
||||
return (this.hitState & PREFIX) > 0;
|
||||
}
|
||||
|
||||
/**
|
||||
*
|
||||
*/
|
||||
public void setPrefix() {
|
||||
this.hitState = this.hitState | PREFIX;
|
||||
}
|
||||
|
||||
/**
|
||||
* 判断是否是不匹配
|
||||
*/
|
||||
public boolean isUnmatch() {
|
||||
return this.hitState == UNMATCH;
|
||||
}
|
||||
|
||||
/**
|
||||
*
|
||||
*/
|
||||
public void setUnmatch() {
|
||||
this.hitState = UNMATCH;
|
||||
}
|
||||
|
||||
public DictSegment getMatchedDictSegment() {
|
||||
return matchedDictSegment;
|
||||
}
|
||||
|
||||
public void setMatchedDictSegment(DictSegment matchedDictSegment) {
|
||||
this.matchedDictSegment = matchedDictSegment;
|
||||
}
|
||||
|
||||
public int getBegin() {
|
||||
return begin;
|
||||
}
|
||||
|
||||
public void setBegin(int begin) {
|
||||
this.begin = begin;
|
||||
}
|
||||
|
||||
public int getEnd() {
|
||||
return end;
|
||||
}
|
||||
|
||||
public void setEnd(int end) {
|
||||
this.end = end;
|
||||
}
|
||||
|
||||
}
|
@ -0,0 +1,50 @@
|
||||
package com.rymcu.forest.lucene.lucene;
|
||||
|
||||
|
||||
import com.rymcu.forest.dto.ArticleDTO;
|
||||
import com.rymcu.forest.lucene.model.Baike;
|
||||
import org.apache.lucene.document.Document;
|
||||
import org.apache.lucene.document.Field;
|
||||
import org.apache.lucene.document.TextField;
|
||||
import org.apache.lucene.index.IndexWriter;
|
||||
import org.apache.lucene.index.IndexWriterConfig;
|
||||
import org.apache.lucene.index.Term;
|
||||
|
||||
import java.util.List;
|
||||
import java.util.concurrent.CountDownLatch;
|
||||
|
||||
/**
|
||||
* BaiKeBeanIndex
|
||||
*
|
||||
* @author suwen
|
||||
* @date 2021/2/2 14:10
|
||||
*/
|
||||
public class ArticleBeanIndex extends BaseIndex<ArticleDTO>{
|
||||
|
||||
public ArticleBeanIndex(IndexWriter writer, CountDownLatch countDownLatch1,
|
||||
CountDownLatch countDownLatch2, List<ArticleDTO> list) {
|
||||
super(writer, countDownLatch1, countDownLatch2, list);
|
||||
}
|
||||
public ArticleBeanIndex(String parentIndexPath, int subIndex, CountDownLatch countDownLatch1,
|
||||
CountDownLatch countDownLatch2, List<ArticleDTO> list) {
|
||||
super(parentIndexPath, subIndex, countDownLatch1, countDownLatch2, list);
|
||||
}
|
||||
@Override
|
||||
public void indexDoc(IndexWriter writer, ArticleDTO t) throws Exception {
|
||||
Document doc = new Document();
|
||||
Field id = new Field("id", t.getIdArticle()+"", TextField.TYPE_STORED);
|
||||
Field title = new Field("title", t.getArticleTitle(), TextField.TYPE_STORED);
|
||||
Field summary = new Field("summary", t.getArticleContent(), TextField.TYPE_STORED);
|
||||
//添加到Document中
|
||||
doc.add(id);
|
||||
doc.add(title);
|
||||
doc.add(summary);
|
||||
if (writer.getConfig().getOpenMode() == IndexWriterConfig.OpenMode.CREATE){
|
||||
writer.addDocument(doc);
|
||||
}else{
|
||||
writer.updateDocument(new Term("id", t.getIdArticle()+""), doc);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
}
|
@ -0,0 +1,49 @@
|
||||
package com.rymcu.forest.lucene.lucene;
|
||||
|
||||
|
||||
import com.rymcu.forest.lucene.model.Baike;
|
||||
import org.apache.lucene.document.Document;
|
||||
import org.apache.lucene.document.Field;
|
||||
import org.apache.lucene.document.TextField;
|
||||
import org.apache.lucene.index.IndexWriter;
|
||||
import org.apache.lucene.index.IndexWriterConfig;
|
||||
import org.apache.lucene.index.Term;
|
||||
|
||||
import java.util.List;
|
||||
import java.util.concurrent.CountDownLatch;
|
||||
|
||||
/**
|
||||
* BaiKeBeanIndex
|
||||
*
|
||||
* @author suwen
|
||||
* @date 2021/2/2 14:10
|
||||
*/
|
||||
public class BaiKeBeanIndex extends BaseIndex<Baike>{
|
||||
|
||||
public BaiKeBeanIndex(IndexWriter writer, CountDownLatch countDownLatch1,
|
||||
CountDownLatch countDownLatch2, List<Baike> list) {
|
||||
super(writer, countDownLatch1, countDownLatch2, list);
|
||||
}
|
||||
public BaiKeBeanIndex(String parentIndexPath, int subIndex, CountDownLatch countDownLatch1,
|
||||
CountDownLatch countDownLatch2, List<Baike> list) {
|
||||
super(parentIndexPath, subIndex, countDownLatch1, countDownLatch2, list);
|
||||
}
|
||||
@Override
|
||||
public void indexDoc(IndexWriter writer, Baike t) throws Exception {
|
||||
Document doc = new Document();
|
||||
Field id = new Field("id", t.getId()+"", TextField.TYPE_STORED);
|
||||
Field title = new Field("title", t.getTitle(), TextField.TYPE_STORED);
|
||||
Field summary = new Field("summary", t.getSummary(), TextField.TYPE_STORED);
|
||||
//添加到Document中
|
||||
doc.add(id);
|
||||
doc.add(title);
|
||||
doc.add(summary);
|
||||
if (writer.getConfig().getOpenMode() == IndexWriterConfig.OpenMode.CREATE){
|
||||
writer.addDocument(doc);
|
||||
}else{
|
||||
writer.updateDocument(new Term("id", t.getId()+""), doc);
|
||||
}
|
||||
}
|
||||
|
||||
|
||||
}
|
127
src/main/java/com/rymcu/forest/lucene/lucene/BaseIndex.java
Normal file
127
src/main/java/com/rymcu/forest/lucene/lucene/BaseIndex.java
Normal file
@ -0,0 +1,127 @@
|
||||
package com.rymcu.forest.lucene.lucene;
|
||||
|
||||
import com.rymcu.forest.lucene.util.IndexUtil;
|
||||
import org.apache.lucene.index.IndexWriter;
|
||||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
import java.text.ParseException;
|
||||
import java.util.List;
|
||||
import java.util.concurrent.CountDownLatch;
|
||||
|
||||
|
||||
/**
|
||||
* BaseIndex
|
||||
*
|
||||
* @author suwen
|
||||
* @date 2021/2/2 14:14
|
||||
*/
|
||||
public abstract class BaseIndex<T> implements Runnable{
|
||||
/**
|
||||
* 父级索引路径
|
||||
*/
|
||||
private String parentIndexPath;
|
||||
/**
|
||||
* 索引编写器
|
||||
*/
|
||||
private IndexWriter writer;
|
||||
private int subIndex;
|
||||
/**
|
||||
* 主线程
|
||||
*/
|
||||
private final CountDownLatch countDownLatch1;
|
||||
/**
|
||||
*工作线程
|
||||
*/
|
||||
private final CountDownLatch countDownLatch2;
|
||||
/**
|
||||
* 对象列表
|
||||
*/
|
||||
private List<T> list;
|
||||
public BaseIndex(IndexWriter writer,CountDownLatch countDownLatch1, CountDownLatch countDownLatch2,
|
||||
List<T> list){
|
||||
super();
|
||||
this.writer = writer;
|
||||
this.countDownLatch1 = countDownLatch1;
|
||||
this.countDownLatch2 = countDownLatch2;
|
||||
this.list = list;
|
||||
}
|
||||
public BaseIndex(String parentIndexPath, int subIndex,
|
||||
CountDownLatch countDownLatch1, CountDownLatch countDownLatch2,
|
||||
List<T> list) {
|
||||
super();
|
||||
this.parentIndexPath = parentIndexPath;
|
||||
this.subIndex = subIndex;
|
||||
try {
|
||||
//多目录索引创建
|
||||
File file = new File(parentIndexPath+"/index"+subIndex);
|
||||
if(!file.exists()){
|
||||
file.mkdir();
|
||||
}
|
||||
this.writer = IndexUtil.getIndexWriter(parentIndexPath+"/index"+subIndex, true);
|
||||
} catch (IOException e) {
|
||||
e.printStackTrace();
|
||||
};
|
||||
this.subIndex = subIndex;
|
||||
this.countDownLatch1 = countDownLatch1;
|
||||
this.countDownLatch2 = countDownLatch2;
|
||||
this.list = list;
|
||||
}
|
||||
public BaseIndex(String path,CountDownLatch countDownLatch1, CountDownLatch countDownLatch2,
|
||||
List<T> list) {
|
||||
super();
|
||||
try {
|
||||
//单目录索引创建
|
||||
File file = new File(path);
|
||||
if(!file.exists()){
|
||||
file.mkdir();
|
||||
}
|
||||
this.writer = IndexUtil.getIndexWriter(path,true);
|
||||
} catch (IOException e) {
|
||||
e.printStackTrace();
|
||||
};
|
||||
this.countDownLatch1 = countDownLatch1;
|
||||
this.countDownLatch2 = countDownLatch2;
|
||||
this.list = list;
|
||||
}
|
||||
|
||||
/**创建索引
|
||||
* @param writer
|
||||
* @throws IOException
|
||||
* @throws ParseException
|
||||
*/
|
||||
public abstract void indexDoc(IndexWriter writer,T t) throws Exception;
|
||||
/**批量索引创建
|
||||
* @param writer
|
||||
* @param t
|
||||
* @throws Exception
|
||||
*/
|
||||
public void indexDocs(IndexWriter writer,List<T> t) throws Exception{
|
||||
for (T t2 : t) {
|
||||
indexDoc(writer,t2);
|
||||
}
|
||||
}
|
||||
|
||||
@Override
|
||||
public void run() {
|
||||
try {
|
||||
countDownLatch1.await();
|
||||
System.out.println(writer);
|
||||
indexDocs(writer,list);
|
||||
} catch (InterruptedException e) {
|
||||
e.printStackTrace();
|
||||
} catch (Exception e) {
|
||||
e.printStackTrace();
|
||||
}finally{
|
||||
countDownLatch2.countDown();
|
||||
try {
|
||||
writer.commit();
|
||||
writer.close();
|
||||
} catch (IOException e) {
|
||||
e.printStackTrace();
|
||||
}
|
||||
|
||||
}
|
||||
|
||||
}
|
||||
}
|
75
src/main/java/com/rymcu/forest/lucene/lucene/IKAnalyzer.java
Executable file
75
src/main/java/com/rymcu/forest/lucene/lucene/IKAnalyzer.java
Executable file
@ -0,0 +1,75 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0.1
|
||||
* IK Analyzer release 5.0.1
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.lucene;
|
||||
|
||||
import org.apache.lucene.analysis.Analyzer;
|
||||
import org.apache.lucene.analysis.Tokenizer;
|
||||
|
||||
/**
|
||||
* IK分词器,Lucene Analyzer接口实现
|
||||
*/
|
||||
public final class IKAnalyzer extends Analyzer {
|
||||
|
||||
private boolean useSmart;
|
||||
|
||||
public boolean useSmart() {
|
||||
return useSmart;
|
||||
}
|
||||
|
||||
public void setUseSmart(boolean useSmart) {
|
||||
this.useSmart = useSmart;
|
||||
}
|
||||
|
||||
/**
|
||||
* IK分词器Lucene Analyzer接口实现类
|
||||
*
|
||||
* 默认细粒度切分算法
|
||||
*/
|
||||
public IKAnalyzer() {
|
||||
this(false);
|
||||
}
|
||||
|
||||
/**
|
||||
* IK分词器Lucene Analyzer接口实现类
|
||||
*
|
||||
* @param useSmart 当为true时,分词器进行智能切分
|
||||
*/
|
||||
public IKAnalyzer(boolean useSmart) {
|
||||
super();
|
||||
this.useSmart = useSmart;
|
||||
}
|
||||
|
||||
|
||||
/**
|
||||
* lucene 6.0
|
||||
* 重载Analyzer接口,构造分词组件
|
||||
*/
|
||||
@Override
|
||||
protected TokenStreamComponents createComponents(String fieldName) {
|
||||
Tokenizer IKTokenizer = new IKTokenizer(this.useSmart());
|
||||
return new Analyzer.TokenStreamComponents(IKTokenizer);
|
||||
}
|
||||
|
||||
}
|
121
src/main/java/com/rymcu/forest/lucene/lucene/IKTokenizer.java
Executable file
121
src/main/java/com/rymcu/forest/lucene/lucene/IKTokenizer.java
Executable file
@ -0,0 +1,121 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0.1
|
||||
* IK Analyzer release 5.0.1
|
||||
* <p>
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
* <p>
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
* <p>
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
* <p>
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*/
|
||||
package com.rymcu.forest.lucene.lucene;
|
||||
|
||||
import com.rymcu.forest.lucene.core.IKSegmenter;
|
||||
import com.rymcu.forest.lucene.core.Lexeme;
|
||||
import org.apache.lucene.analysis.Tokenizer;
|
||||
import org.apache.lucene.analysis.tokenattributes.CharTermAttribute;
|
||||
import org.apache.lucene.analysis.tokenattributes.OffsetAttribute;
|
||||
import org.apache.lucene.analysis.tokenattributes.TypeAttribute;
|
||||
import org.apache.lucene.util.AttributeFactory;
|
||||
|
||||
import java.io.IOException;
|
||||
|
||||
|
||||
/**
|
||||
* IK分词器 Lucene Tokenizer适配器类
|
||||
* 兼容Lucene 6.0版本
|
||||
*/
|
||||
public final class IKTokenizer extends Tokenizer {
|
||||
// IK分词器实现
|
||||
private IKSegmenter _IKImplement;
|
||||
|
||||
// 词元文本属性
|
||||
private final CharTermAttribute termAtt;
|
||||
// 词元位移属性
|
||||
private final OffsetAttribute offsetAtt;
|
||||
// 词元分类属性(该属性分类参考org.wltea.analyzer.core.Lexeme中的分类常量)
|
||||
private final TypeAttribute typeAtt;
|
||||
// 记录最后一个词元的结束位置
|
||||
private int endPosition;
|
||||
|
||||
private Tokenizer delegate;
|
||||
|
||||
/**
|
||||
* Lucene 6.0 Tokenizer适配器类构造函数
|
||||
* @param useSmart
|
||||
*/
|
||||
public IKTokenizer(boolean useSmart) {
|
||||
super();
|
||||
offsetAtt = addAttribute(OffsetAttribute.class);
|
||||
termAtt = addAttribute(CharTermAttribute.class);
|
||||
typeAtt = addAttribute(TypeAttribute.class);
|
||||
//传入 IKSegmenter 的 input Reader 流,会被 父类 Tokenizer 类的无参构造器
|
||||
//初始化为 this.input = ILLEGAL_STATE_READER;
|
||||
_IKImplement = new IKSegmenter(input, useSmart);
|
||||
}
|
||||
|
||||
/**
|
||||
* lucene 6.0 新增
|
||||
* 方便创建 工厂类
|
||||
* @param factory
|
||||
* @param useSmart
|
||||
*/
|
||||
public IKTokenizer(AttributeFactory factory, boolean useSmart) {
|
||||
super(factory);
|
||||
offsetAtt = addAttribute(OffsetAttribute.class);
|
||||
termAtt = addAttribute(CharTermAttribute.class);
|
||||
typeAtt = addAttribute(TypeAttribute.class);
|
||||
_IKImplement = new IKSegmenter(input, useSmart);
|
||||
}
|
||||
|
||||
|
||||
@Override
|
||||
public boolean incrementToken() throws IOException {
|
||||
// 清除所有的词元属性
|
||||
clearAttributes();
|
||||
Lexeme nextLexeme = _IKImplement.next();
|
||||
if (nextLexeme != null) {
|
||||
// 将Lexeme转成Attributes
|
||||
// 设置词元文本
|
||||
termAtt.append(nextLexeme.getLexemeText());
|
||||
// 设置词元长度
|
||||
termAtt.setLength(nextLexeme.getLength());
|
||||
// 设置词元位移
|
||||
offsetAtt.setOffset(nextLexeme.getBeginPosition(), nextLexeme.getEndPosition());
|
||||
// 记录分词的最后位置
|
||||
endPosition = nextLexeme.getEndPosition();
|
||||
// 记录词元分类
|
||||
typeAtt.setType(nextLexeme.getLexemeTypeString());
|
||||
// 返会true告知还有下个词元
|
||||
return true;
|
||||
}
|
||||
// 返会false告知词元输出完毕
|
||||
return false;
|
||||
}
|
||||
|
||||
@Override
|
||||
public void reset() throws IOException {
|
||||
super.reset();
|
||||
_IKImplement.reset(input);
|
||||
}
|
||||
|
||||
@Override
|
||||
public final void end() {
|
||||
// set final offset
|
||||
int finalOffset = correctOffset(this.endPosition);
|
||||
offsetAtt.setOffset(finalOffset, finalOffset);
|
||||
}
|
||||
}
|
18
src/main/java/com/rymcu/forest/lucene/mapper/BaikeMapper.java
Executable file
18
src/main/java/com/rymcu/forest/lucene/mapper/BaikeMapper.java
Executable file
@ -0,0 +1,18 @@
|
||||
package com.rymcu.forest.lucene.mapper;
|
||||
|
||||
import com.rymcu.forest.lucene.model.Baike;
|
||||
import org.apache.ibatis.annotations.Mapper;
|
||||
import org.apache.ibatis.annotations.Param;
|
||||
|
||||
import java.util.List;
|
||||
|
||||
/**
|
||||
* BaikeMapper
|
||||
*
|
||||
* @author suwen
|
||||
* @date 2021/2/2 14:10
|
||||
*/
|
||||
@Mapper
|
||||
public interface BaikeMapper {
|
||||
List<Baike> getAllBaike(@Param("limit") int limit, @Param("offset") int offset);
|
||||
}
|
45
src/main/java/com/rymcu/forest/lucene/model/Baike.java
Executable file
45
src/main/java/com/rymcu/forest/lucene/model/Baike.java
Executable file
@ -0,0 +1,45 @@
|
||||
package com.rymcu.forest.lucene.model;
|
||||
|
||||
import lombok.Data;
|
||||
|
||||
import javax.persistence.Table;
|
||||
|
||||
@Data
|
||||
@Table(name = "lucene_baike")
|
||||
public class Baike {
|
||||
private Integer id;
|
||||
|
||||
private String title;
|
||||
|
||||
private String summary;
|
||||
|
||||
public Integer getId() {
|
||||
return id;
|
||||
}
|
||||
|
||||
public void setId(Integer id) {
|
||||
this.id = id;
|
||||
}
|
||||
|
||||
public String getTitle() {
|
||||
if(title == null){
|
||||
title = "";
|
||||
}
|
||||
return title;
|
||||
}
|
||||
|
||||
public void setTitle(String title) {
|
||||
this.title = title == null ? null : title.trim();
|
||||
}
|
||||
|
||||
public String getSummary() {
|
||||
if(summary == null){
|
||||
summary = "";
|
||||
}
|
||||
return summary;
|
||||
}
|
||||
|
||||
public void setSummary(String summary) {
|
||||
this.summary = summary == null ? null : summary.trim();
|
||||
}
|
||||
}
|
728
src/main/java/com/rymcu/forest/lucene/query/IKQueryExpressionParser.java
Executable file
728
src/main/java/com/rymcu/forest/lucene/query/IKQueryExpressionParser.java
Executable file
@ -0,0 +1,728 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
* <p>
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
* <p>
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
* <p>
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
* <p>
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*/
|
||||
package com.rymcu.forest.lucene.query;
|
||||
|
||||
import org.apache.lucene.index.Term;
|
||||
import org.apache.lucene.search.*;
|
||||
import org.apache.lucene.search.BooleanClause.Occur;
|
||||
import org.apache.lucene.search.BooleanQuery.Builder;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.util.BytesRef;
|
||||
|
||||
import java.util.*;
|
||||
|
||||
/**
|
||||
* IK简易查询表达式解析
|
||||
* 结合SWMCQuery算法
|
||||
*
|
||||
* 表达式例子 :
|
||||
* (id='1231231' && title:'monkey') || (content:'你好吗' || ulr='www.ik.com') - name:'helloword'
|
||||
* @author linliangyi
|
||||
*/
|
||||
public class IKQueryExpressionParser {
|
||||
|
||||
private List<Element> elements = new ArrayList<Element>();
|
||||
|
||||
private Stack<Query> querys = new Stack<Query>();
|
||||
|
||||
private Stack<Element> operates = new Stack<Element>();
|
||||
|
||||
/**
|
||||
* 解析查询表达式,生成Lucene Query对象
|
||||
*
|
||||
* @param expression
|
||||
* @param quickMode
|
||||
* @return Lucene query
|
||||
*/
|
||||
public Query parseExp(String expression, boolean quickMode) {
|
||||
Query lucenceQuery = null;
|
||||
if (expression != null && !"".equals(expression.trim())) {
|
||||
try {
|
||||
//文法解析
|
||||
this.splitElements(expression);
|
||||
//语法解析
|
||||
this.parseSyntax(quickMode);
|
||||
if (this.querys.size() == 1) {
|
||||
lucenceQuery = this.querys.pop();
|
||||
} else {
|
||||
throw new IllegalStateException("表达式异常: 缺少逻辑操作符 或 括号缺失");
|
||||
}
|
||||
} finally {
|
||||
elements.clear();
|
||||
querys.clear();
|
||||
operates.clear();
|
||||
}
|
||||
}
|
||||
return lucenceQuery;
|
||||
}
|
||||
|
||||
/**
|
||||
* 表达式文法解析
|
||||
* @param expression
|
||||
*/
|
||||
private void splitElements(String expression) {
|
||||
|
||||
if (expression == null) {
|
||||
return;
|
||||
}
|
||||
Element curretElement = null;
|
||||
|
||||
char[] expChars = expression.toCharArray();
|
||||
for (int i = 0; i < expChars.length; i++) {
|
||||
switch (expChars[i]) {
|
||||
case '&':
|
||||
if (curretElement == null) {
|
||||
curretElement = new Element();
|
||||
curretElement.type = '&';
|
||||
curretElement.append(expChars[i]);
|
||||
} else if (curretElement.type == '&') {
|
||||
curretElement.append(expChars[i]);
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
} else if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
curretElement = new Element();
|
||||
curretElement.type = '&';
|
||||
curretElement.append(expChars[i]);
|
||||
}
|
||||
break;
|
||||
|
||||
case '|':
|
||||
if (curretElement == null) {
|
||||
curretElement = new Element();
|
||||
curretElement.type = '|';
|
||||
curretElement.append(expChars[i]);
|
||||
} else if (curretElement.type == '|') {
|
||||
curretElement.append(expChars[i]);
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
} else if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
curretElement = new Element();
|
||||
curretElement.type = '|';
|
||||
curretElement.append(expChars[i]);
|
||||
}
|
||||
break;
|
||||
|
||||
case '-':
|
||||
if (curretElement != null) {
|
||||
if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
continue;
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
}
|
||||
}
|
||||
curretElement = new Element();
|
||||
curretElement.type = '-';
|
||||
curretElement.append(expChars[i]);
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
break;
|
||||
|
||||
case '(':
|
||||
if (curretElement != null) {
|
||||
if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
continue;
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
}
|
||||
}
|
||||
curretElement = new Element();
|
||||
curretElement.type = '(';
|
||||
curretElement.append(expChars[i]);
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
break;
|
||||
|
||||
case ')':
|
||||
if (curretElement != null) {
|
||||
if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
continue;
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
}
|
||||
}
|
||||
curretElement = new Element();
|
||||
curretElement.type = ')';
|
||||
curretElement.append(expChars[i]);
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
break;
|
||||
|
||||
case ':':
|
||||
if (curretElement != null) {
|
||||
if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
continue;
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
}
|
||||
}
|
||||
curretElement = new Element();
|
||||
curretElement.type = ':';
|
||||
curretElement.append(expChars[i]);
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
break;
|
||||
|
||||
case '=':
|
||||
if (curretElement != null) {
|
||||
if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
continue;
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
}
|
||||
}
|
||||
curretElement = new Element();
|
||||
curretElement.type = '=';
|
||||
curretElement.append(expChars[i]);
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
break;
|
||||
|
||||
case ' ':
|
||||
if (curretElement != null) {
|
||||
if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
}
|
||||
}
|
||||
|
||||
break;
|
||||
|
||||
case '\'':
|
||||
if (curretElement == null) {
|
||||
curretElement = new Element();
|
||||
curretElement.type = '\'';
|
||||
|
||||
} else if (curretElement.type == '\'') {
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
curretElement = new Element();
|
||||
curretElement.type = '\'';
|
||||
|
||||
}
|
||||
break;
|
||||
|
||||
case '[':
|
||||
if (curretElement != null) {
|
||||
if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
continue;
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
}
|
||||
}
|
||||
curretElement = new Element();
|
||||
curretElement.type = '[';
|
||||
curretElement.append(expChars[i]);
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
break;
|
||||
|
||||
case ']':
|
||||
if (curretElement != null) {
|
||||
if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
continue;
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
}
|
||||
}
|
||||
curretElement = new Element();
|
||||
curretElement.type = ']';
|
||||
curretElement.append(expChars[i]);
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
|
||||
break;
|
||||
|
||||
case '{':
|
||||
if (curretElement != null) {
|
||||
if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
continue;
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
}
|
||||
}
|
||||
curretElement = new Element();
|
||||
curretElement.type = '{';
|
||||
curretElement.append(expChars[i]);
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
break;
|
||||
|
||||
case '}':
|
||||
if (curretElement != null) {
|
||||
if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
continue;
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
}
|
||||
}
|
||||
curretElement = new Element();
|
||||
curretElement.type = '}';
|
||||
curretElement.append(expChars[i]);
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
|
||||
break;
|
||||
case ',':
|
||||
if (curretElement != null) {
|
||||
if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
continue;
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
}
|
||||
}
|
||||
curretElement = new Element();
|
||||
curretElement.type = ',';
|
||||
curretElement.append(expChars[i]);
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
|
||||
break;
|
||||
|
||||
default:
|
||||
if (curretElement == null) {
|
||||
curretElement = new Element();
|
||||
curretElement.type = 'F';
|
||||
curretElement.append(expChars[i]);
|
||||
|
||||
} else if (curretElement.type == 'F') {
|
||||
curretElement.append(expChars[i]);
|
||||
|
||||
} else if (curretElement.type == '\'') {
|
||||
curretElement.append(expChars[i]);
|
||||
|
||||
} else {
|
||||
this.elements.add(curretElement);
|
||||
curretElement = new Element();
|
||||
curretElement.type = 'F';
|
||||
curretElement.append(expChars[i]);
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
if (curretElement != null) {
|
||||
this.elements.add(curretElement);
|
||||
curretElement = null;
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 语法解析
|
||||
*
|
||||
*/
|
||||
private void parseSyntax(boolean quickMode) {
|
||||
for (int i = 0; i < this.elements.size(); i++) {
|
||||
Element e = this.elements.get(i);
|
||||
if ('F' == e.type) {
|
||||
Element e2 = this.elements.get(i + 1);
|
||||
if ('=' != e2.type && ':' != e2.type) {
|
||||
throw new IllegalStateException("表达式异常: = 或 : 号丢失");
|
||||
}
|
||||
Element e3 = this.elements.get(i + 2);
|
||||
//处理 = 和 : 运算
|
||||
if ('\'' == e3.type) {
|
||||
i += 2;
|
||||
if ('=' == e2.type) {
|
||||
TermQuery tQuery = new TermQuery(new Term(e.toString(), e3.toString()));
|
||||
this.querys.push(tQuery);
|
||||
} else if (':' == e2.type) {
|
||||
String keyword = e3.toString();
|
||||
//SWMCQuery Here
|
||||
Query _SWMCQuery = SWMCQueryBuilder.create(e.toString(), keyword, quickMode);
|
||||
this.querys.push(_SWMCQuery);
|
||||
}
|
||||
|
||||
} else if ('[' == e3.type || '{' == e3.type) {
|
||||
i += 2;
|
||||
//处理 [] 和 {}
|
||||
LinkedList<Element> eQueue = new LinkedList<Element>();
|
||||
eQueue.add(e3);
|
||||
for (i++; i < this.elements.size(); i++) {
|
||||
Element eN = this.elements.get(i);
|
||||
eQueue.add(eN);
|
||||
if (']' == eN.type || '}' == eN.type) {
|
||||
break;
|
||||
}
|
||||
}
|
||||
//翻译RangeQuery
|
||||
Query rangeQuery = this.toTermRangeQuery(e, eQueue);
|
||||
this.querys.push(rangeQuery);
|
||||
} else {
|
||||
throw new IllegalStateException("表达式异常:匹配值丢失");
|
||||
}
|
||||
|
||||
} else if ('(' == e.type) {
|
||||
this.operates.push(e);
|
||||
|
||||
} else if (')' == e.type) {
|
||||
boolean doPop = true;
|
||||
while (doPop && !this.operates.empty()) {
|
||||
Element op = this.operates.pop();
|
||||
if ('(' == op.type) {
|
||||
doPop = false;
|
||||
} else {
|
||||
Query q = toBooleanQuery(op);
|
||||
this.querys.push(q);
|
||||
}
|
||||
|
||||
}
|
||||
} else {
|
||||
|
||||
if (this.operates.isEmpty()) {
|
||||
this.operates.push(e);
|
||||
} else {
|
||||
boolean doPeek = true;
|
||||
while (doPeek && !this.operates.isEmpty()) {
|
||||
Element eleOnTop = this.operates.peek();
|
||||
if ('(' == eleOnTop.type) {
|
||||
doPeek = false;
|
||||
this.operates.push(e);
|
||||
} else if (compare(e, eleOnTop) == 1) {
|
||||
this.operates.push(e);
|
||||
doPeek = false;
|
||||
} else if (compare(e, eleOnTop) == 0) {
|
||||
Query q = toBooleanQuery(eleOnTop);
|
||||
this.operates.pop();
|
||||
this.querys.push(q);
|
||||
} else {
|
||||
Query q = toBooleanQuery(eleOnTop);
|
||||
this.operates.pop();
|
||||
this.querys.push(q);
|
||||
}
|
||||
}
|
||||
|
||||
if (doPeek && this.operates.empty()) {
|
||||
this.operates.push(e);
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
while (!this.operates.isEmpty()) {
|
||||
Element eleOnTop = this.operates.pop();
|
||||
Query q = toBooleanQuery(eleOnTop);
|
||||
this.querys.push(q);
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 根据逻辑操作符,生成BooleanQuery
|
||||
* @param op
|
||||
* @return
|
||||
*/
|
||||
private Query toBooleanQuery(Element op) {
|
||||
if (this.querys.size() == 0) {
|
||||
return null;
|
||||
}
|
||||
|
||||
// BooleanQuery resultQuery = new BooleanQuery.Builder().build();
|
||||
Builder resultQuery = new Builder();
|
||||
|
||||
if (this.querys.size() == 1) {
|
||||
return this.querys.get(0);
|
||||
}
|
||||
|
||||
Query q2 = this.querys.pop();
|
||||
Query q1 = this.querys.pop();
|
||||
if ('&' == op.type) {
|
||||
if (q1 != null) {
|
||||
if (q1 instanceof BooleanQuery) {
|
||||
Iterator<BooleanClause> clauses = ((BooleanQuery) q1).iterator();
|
||||
while (clauses.hasNext()) {
|
||||
BooleanClause clause = clauses.next();
|
||||
if (clause.getOccur() == Occur.MUST) {
|
||||
resultQuery.add(clause);
|
||||
} else {
|
||||
resultQuery.add(q1, Occur.MUST);
|
||||
}
|
||||
}
|
||||
|
||||
/*BooleanClause[] clauses = ((BooleanQuery)q1).getClauses();
|
||||
if(clauses.length > 0
|
||||
&& clauses[0].getOccur() == Occur.MUST){
|
||||
for(BooleanClause c : clauses){
|
||||
resultQuery.add(c);
|
||||
}
|
||||
}else{
|
||||
resultQuery.add(q1,Occur.MUST);
|
||||
}*/
|
||||
|
||||
} else {
|
||||
//q1 instanceof TermQuery
|
||||
//q1 instanceof TermRangeQuery
|
||||
//q1 instanceof PhraseQuery
|
||||
//others
|
||||
resultQuery.add(q1, Occur.MUST);
|
||||
}
|
||||
}
|
||||
|
||||
if (q2 != null) {
|
||||
if (q2 instanceof BooleanQuery) {
|
||||
Iterator<BooleanClause> clauses = ((BooleanQuery) q2).iterator();
|
||||
while (clauses.hasNext()) {
|
||||
BooleanClause clause = clauses.next();
|
||||
if (clause.getOccur() == Occur.MUST) {
|
||||
resultQuery.add(clause);
|
||||
} else {
|
||||
resultQuery.add(q2, Occur.MUST);
|
||||
}
|
||||
}
|
||||
/*BooleanClause[] clauses = ((BooleanQuery)q2).getClauses();
|
||||
if(clauses.length > 0
|
||||
&& clauses[0].getOccur() == Occur.MUST){
|
||||
for(BooleanClause c : clauses){
|
||||
resultQuery.add(c);
|
||||
}
|
||||
}else{
|
||||
resultQuery.add(q2,Occur.MUST);
|
||||
}*/
|
||||
|
||||
} else {
|
||||
//q1 instanceof TermQuery
|
||||
//q1 instanceof TermRangeQuery
|
||||
//q1 instanceof PhraseQuery
|
||||
//others
|
||||
resultQuery.add(q2, Occur.MUST);
|
||||
}
|
||||
}
|
||||
|
||||
} else if ('|' == op.type) {
|
||||
if (q1 != null) {
|
||||
if (q1 instanceof BooleanQuery) {
|
||||
Iterator<BooleanClause> clauses = ((BooleanQuery) q1).iterator();
|
||||
while (clauses.hasNext()) {
|
||||
BooleanClause clause = clauses.next();
|
||||
if (clause.getOccur() == Occur.SHOULD) {
|
||||
resultQuery.add(clause);
|
||||
} else {
|
||||
resultQuery.add(q1, Occur.SHOULD);
|
||||
}
|
||||
}
|
||||
/*BooleanClause[] clauses = ((BooleanQuery)q1).getClauses();
|
||||
if(clauses.length > 0
|
||||
&& clauses[0].getOccur() == Occur.SHOULD){
|
||||
for(BooleanClause c : clauses){
|
||||
resultQuery.add(c);
|
||||
}
|
||||
}else{
|
||||
resultQuery.add(q1,Occur.SHOULD);
|
||||
}*/
|
||||
|
||||
} else {
|
||||
//q1 instanceof TermQuery
|
||||
//q1 instanceof TermRangeQuery
|
||||
//q1 instanceof PhraseQuery
|
||||
//others
|
||||
resultQuery.add(q1, Occur.SHOULD);
|
||||
}
|
||||
}
|
||||
|
||||
if (q2 != null) {
|
||||
if (q2 instanceof BooleanQuery) {
|
||||
Iterator<BooleanClause> clauses = ((BooleanQuery) q1).iterator();
|
||||
while (clauses.hasNext()) {
|
||||
BooleanClause clause = clauses.next();
|
||||
if (clause.getOccur() == Occur.SHOULD) {
|
||||
resultQuery.add(clause);
|
||||
} else {
|
||||
resultQuery.add(q1, Occur.SHOULD);
|
||||
}
|
||||
}
|
||||
/*BooleanClause[] clauses = ((BooleanQuery)q2).getClauses();
|
||||
if(clauses.length > 0
|
||||
&& clauses[0].getOccur() == Occur.SHOULD){
|
||||
for(BooleanClause c : clauses){
|
||||
resultQuery.add(c);
|
||||
}
|
||||
}else{
|
||||
resultQuery.add(q2,Occur.SHOULD);
|
||||
}*/
|
||||
} else {
|
||||
//q2 instanceof TermQuery
|
||||
//q2 instanceof TermRangeQuery
|
||||
//q2 instanceof PhraseQuery
|
||||
//others
|
||||
resultQuery.add(q2, Occur.SHOULD);
|
||||
|
||||
}
|
||||
}
|
||||
|
||||
} else if ('-' == op.type) {
|
||||
if (q1 == null || q2 == null) {
|
||||
throw new IllegalStateException("表达式异常:SubQuery 个数不匹配");
|
||||
}
|
||||
|
||||
if (q1 instanceof BooleanQuery) {
|
||||
Iterator<BooleanClause> clauses = ((BooleanQuery) q1).iterator();
|
||||
while (clauses.hasNext()) {
|
||||
BooleanClause clause = clauses.next();
|
||||
if (clause.getOccur() == Occur.MUST) {
|
||||
resultQuery.add(clause);
|
||||
} else {
|
||||
resultQuery.add(q1, Occur.MUST);
|
||||
}
|
||||
}
|
||||
} else {
|
||||
resultQuery.add(q1, Occur.MUST);
|
||||
}
|
||||
|
||||
resultQuery.add(q2, Occur.MUST_NOT);
|
||||
}
|
||||
return resultQuery.build();
|
||||
}
|
||||
|
||||
/**
|
||||
* 组装TermRangeQuery
|
||||
* @param elements
|
||||
* @return
|
||||
*/
|
||||
private TermRangeQuery toTermRangeQuery(Element fieldNameEle, LinkedList<Element> elements) {
|
||||
|
||||
boolean includeFirst = false;
|
||||
boolean includeLast = false;
|
||||
String firstValue = null;
|
||||
String lastValue = null;
|
||||
//检查第一个元素是否是[或者{
|
||||
Element first = elements.getFirst();
|
||||
if ('[' == first.type) {
|
||||
includeFirst = true;
|
||||
} else if ('{' == first.type) {
|
||||
} else {
|
||||
throw new IllegalStateException("表达式异常");
|
||||
}
|
||||
//检查最后一个元素是否是]或者}
|
||||
Element last = elements.getLast();
|
||||
if (']' == last.type) {
|
||||
includeLast = true;
|
||||
}else {
|
||||
throw new IllegalStateException("表达式异常, RangeQuery缺少结束括号");
|
||||
}
|
||||
if (elements.size() < 4 || elements.size() > 5) {
|
||||
throw new IllegalStateException("表达式异常, RangeQuery 错误");
|
||||
}
|
||||
//读出中间部分
|
||||
Element e2 = elements.get(1);
|
||||
if ('\'' == e2.type) {
|
||||
firstValue = e2.toString();
|
||||
//
|
||||
Element e3 = elements.get(2);
|
||||
if (',' != e3.type) {
|
||||
throw new IllegalStateException("表达式异常, RangeQuery缺少逗号分隔");
|
||||
}
|
||||
//
|
||||
Element e4 = elements.get(3);
|
||||
if ('\'' == e4.type) {
|
||||
lastValue = e4.toString();
|
||||
} else if (e4 != last) {
|
||||
throw new IllegalStateException("表达式异常,RangeQuery格式错误");
|
||||
}
|
||||
} else if (',' == e2.type) {
|
||||
Element e3 = elements.get(2);
|
||||
if ('\'' == e3.type) {
|
||||
lastValue = e3.toString();
|
||||
} else {
|
||||
throw new IllegalStateException("表达式异常,RangeQuery格式错误");
|
||||
}
|
||||
|
||||
} else {
|
||||
throw new IllegalStateException("表达式异常, RangeQuery格式错误");
|
||||
}
|
||||
|
||||
return new TermRangeQuery(fieldNameEle.toString(), new BytesRef(firstValue), new BytesRef(lastValue), includeFirst, includeLast);
|
||||
}
|
||||
|
||||
/**
|
||||
* 比较操作符优先级
|
||||
* @param e1
|
||||
* @param e2
|
||||
* @return
|
||||
*/
|
||||
private int compare(Element e1, Element e2) {
|
||||
if ('&' == e1.type) {
|
||||
if ('&' == e2.type) {
|
||||
return 0;
|
||||
} else {
|
||||
return 1;
|
||||
}
|
||||
} else if ('|' == e1.type) {
|
||||
if ('&' == e2.type) {
|
||||
return -1;
|
||||
} else if ('|' == e2.type) {
|
||||
return 0;
|
||||
} else {
|
||||
return 1;
|
||||
}
|
||||
} else {
|
||||
if ('-' == e2.type) {
|
||||
return 0;
|
||||
} else {
|
||||
return -1;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 表达式元素(操作符、FieldName、FieldValue)
|
||||
* @author linliangyi
|
||||
* May 20, 2010
|
||||
*/
|
||||
private static class Element {
|
||||
char type = 0;
|
||||
StringBuffer eleTextBuff;
|
||||
|
||||
public Element() {
|
||||
eleTextBuff = new StringBuffer();
|
||||
}
|
||||
|
||||
public void append(char c) {
|
||||
this.eleTextBuff.append(c);
|
||||
}
|
||||
|
||||
@Override
|
||||
public String toString() {
|
||||
return this.eleTextBuff.toString();
|
||||
}
|
||||
}
|
||||
|
||||
}
|
146
src/main/java/com/rymcu/forest/lucene/query/SWMCQueryBuilder.java
Executable file
146
src/main/java/com/rymcu/forest/lucene/query/SWMCQueryBuilder.java
Executable file
@ -0,0 +1,146 @@
|
||||
/**
|
||||
* IK 中文分词 版本 5.0
|
||||
* IK Analyzer release 5.0
|
||||
*
|
||||
* Licensed to the Apache Software Foundation (ASF) under one or more
|
||||
* contributor license agreements. See the NOTICE file distributed with
|
||||
* this work for additional information regarding copyright ownership.
|
||||
* The ASF licenses this file to You under the Apache License, Version 2.0
|
||||
* (the "License"); you may not use this file except in compliance with
|
||||
* the License. You may obtain a copy of the License at
|
||||
*
|
||||
* http://www.apache.org/licenses/LICENSE-2.0
|
||||
*
|
||||
* Unless required by applicable law or agreed to in writing, software
|
||||
* distributed under the License is distributed on an "AS IS" BASIS,
|
||||
* WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
|
||||
* See the License for the specific language governing permissions and
|
||||
* limitations under the License.
|
||||
*
|
||||
* 源代码由林良益(linliangyi2005@gmail.com)提供
|
||||
* 版权声明 2012,乌龙茶工作室
|
||||
* provided by Linliangyi and copyright 2012 by Oolong studio
|
||||
*
|
||||
*/
|
||||
package com.rymcu.forest.lucene.query;
|
||||
|
||||
import com.rymcu.forest.lucene.core.IKSegmenter;
|
||||
import com.rymcu.forest.lucene.core.Lexeme;
|
||||
import org.apache.lucene.analysis.standard.StandardAnalyzer;
|
||||
import org.apache.lucene.queryparser.classic.ParseException;
|
||||
import org.apache.lucene.queryparser.classic.QueryParser;
|
||||
import org.apache.lucene.search.Query;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.io.StringReader;
|
||||
import java.util.ArrayList;
|
||||
import java.util.List;
|
||||
|
||||
/**
|
||||
* Single Word Multi Char Query Builder
|
||||
*
|
||||
* IK分词算法专用
|
||||
* @author linliangyi
|
||||
*/
|
||||
public class SWMCQueryBuilder {
|
||||
|
||||
/**
|
||||
* 生成SWMCQuery
|
||||
* @param fieldName
|
||||
* @param keywords
|
||||
* @param quickMode
|
||||
* @return Lucene Query
|
||||
*/
|
||||
public static Query create(String fieldName, String keywords, boolean quickMode) {
|
||||
if (fieldName == null || keywords == null) {
|
||||
throw new IllegalArgumentException("参数 fieldName 、 keywords 不能为null.");
|
||||
}
|
||||
// 1.对keywords进行分词处理
|
||||
List<Lexeme> lexemes = doAnalyze(keywords);
|
||||
// 2.根据分词结果,生成SWMCQuery
|
||||
Query _SWMCQuery = getSWMCQuery(fieldName, lexemes, quickMode);
|
||||
return _SWMCQuery;
|
||||
}
|
||||
|
||||
/**
|
||||
* 分词切分,并返回结链表
|
||||
* @param keywords
|
||||
* @return
|
||||
*/
|
||||
private static List<Lexeme> doAnalyze(String keywords) {
|
||||
List<Lexeme> lexemes = new ArrayList<Lexeme>();
|
||||
IKSegmenter ikSeg = new IKSegmenter(new StringReader(keywords), true);
|
||||
try {
|
||||
Lexeme l = null;
|
||||
while ((l = ikSeg.next()) != null) {
|
||||
lexemes.add(l);
|
||||
}
|
||||
} catch (IOException e) {
|
||||
e.printStackTrace();
|
||||
}
|
||||
return lexemes;
|
||||
}
|
||||
|
||||
/**
|
||||
* 根据分词结果生成SWMC搜索
|
||||
* @param fieldName
|
||||
* @param lexemes
|
||||
* @param quickMode
|
||||
* @return
|
||||
*/
|
||||
private static Query getSWMCQuery(String fieldName, List<Lexeme> lexemes, boolean quickMode) {
|
||||
// 构造SWMC的查询表达式
|
||||
StringBuffer keywordBuffer = new StringBuffer();
|
||||
// 精简的SWMC的查询表达式
|
||||
StringBuffer keywordBuffer_Short = new StringBuffer();
|
||||
// 记录最后词元长度
|
||||
int lastLexemeLength = 0;
|
||||
// 记录最后词元结束位置
|
||||
int lastLexemeEnd = -1;
|
||||
|
||||
int shortCount = 0;
|
||||
int totalCount = 0;
|
||||
for (Lexeme l : lexemes) {
|
||||
totalCount += l.getLength();
|
||||
// 精简表达式
|
||||
if (l.getLength() > 1) {
|
||||
keywordBuffer_Short.append(' ').append(l.getLexemeText());
|
||||
shortCount += l.getLength();
|
||||
}
|
||||
|
||||
if (lastLexemeLength == 0) {
|
||||
keywordBuffer.append(l.getLexemeText());
|
||||
} else if (lastLexemeLength == 1 && l.getLength() == 1
|
||||
&& lastLexemeEnd == l.getBeginPosition()) {// 单字位置相邻,长度为一,合并)
|
||||
keywordBuffer.append(l.getLexemeText());
|
||||
} else {
|
||||
keywordBuffer.append(' ').append(l.getLexemeText());
|
||||
|
||||
}
|
||||
lastLexemeLength = l.getLength();
|
||||
lastLexemeEnd = l.getEndPosition();
|
||||
}
|
||||
|
||||
QueryParser qp = new QueryParser(fieldName, new StandardAnalyzer());
|
||||
qp.setDefaultOperator(QueryParser.AND_OPERATOR);
|
||||
qp.setAutoGeneratePhraseQueries(true);
|
||||
|
||||
if (quickMode && (shortCount * 1.0f / totalCount) > 0.5f) {
|
||||
try {
|
||||
return qp.parse(keywordBuffer_Short.toString());
|
||||
} catch (ParseException e) {
|
||||
e.printStackTrace();
|
||||
}
|
||||
|
||||
} else {
|
||||
if (keywordBuffer.length() > 0) {
|
||||
try {
|
||||
return qp.parse(keywordBuffer.toString());
|
||||
} catch (ParseException e) {
|
||||
e.printStackTrace();
|
||||
}
|
||||
}
|
||||
}
|
||||
return null;
|
||||
}
|
||||
}
|
281
src/main/java/com/rymcu/forest/lucene/service/SearchService.java
Normal file
281
src/main/java/com/rymcu/forest/lucene/service/SearchService.java
Normal file
@ -0,0 +1,281 @@
|
||||
package com.rymcu.forest.lucene.service;
|
||||
|
||||
import com.rymcu.forest.dto.ArticleDTO;
|
||||
import com.rymcu.forest.lucene.lucene.ArticleBeanIndex;
|
||||
import com.rymcu.forest.lucene.lucene.BaiKeBeanIndex;
|
||||
import com.rymcu.forest.lucene.lucene.IKAnalyzer;
|
||||
import com.rymcu.forest.lucene.model.Baike;
|
||||
import com.rymcu.forest.lucene.util.SearchUtil;
|
||||
import org.apache.lucene.analysis.Analyzer;
|
||||
import org.apache.lucene.analysis.TokenStream;
|
||||
import org.apache.lucene.document.Document;
|
||||
import org.apache.lucene.queryparser.classic.MultiFieldQueryParser;
|
||||
import org.apache.lucene.search.IndexSearcher;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.ScoreDoc;
|
||||
import org.apache.lucene.search.TopDocs;
|
||||
import org.apache.lucene.search.highlight.*;
|
||||
import org.springframework.stereotype.Service;
|
||||
|
||||
import java.io.BufferedReader;
|
||||
import java.io.IOException;
|
||||
import java.io.InputStreamReader;
|
||||
import java.nio.charset.StandardCharsets;
|
||||
import java.util.ArrayList;
|
||||
import java.util.HashMap;
|
||||
import java.util.List;
|
||||
import java.util.Map;
|
||||
import java.util.concurrent.CountDownLatch;
|
||||
import java.util.concurrent.ExecutorService;
|
||||
import java.util.concurrent.Executors;
|
||||
|
||||
/**
|
||||
* SearchService
|
||||
*
|
||||
* @author suwen
|
||||
* @date 2021/2/2 14:01
|
||||
*/
|
||||
@Service
|
||||
public class SearchService {
|
||||
|
||||
/** Lucene索引文件路径 */
|
||||
private final String indexPath = System.getProperty("user.dir") + "/index";
|
||||
|
||||
/**
|
||||
* 封裝一个方法,用于将数据库中的数据解析为一个个关键字词存储到索引文件中
|
||||
*
|
||||
* @param baikes
|
||||
*/
|
||||
public void write(List<Baike> baikes) {
|
||||
try {
|
||||
int totalCount = baikes.size();
|
||||
int perThreadCount = 3000;
|
||||
int threadCount = totalCount / perThreadCount + (totalCount % perThreadCount == 0 ? 0 : 1);
|
||||
ExecutorService pool = Executors.newFixedThreadPool(threadCount);
|
||||
CountDownLatch countDownLatch1 = new CountDownLatch(1);
|
||||
CountDownLatch countDownLatch2 = new CountDownLatch(threadCount);
|
||||
|
||||
for (int i = 0; i < threadCount; i++) {
|
||||
int start = i * perThreadCount;
|
||||
int end = Math.min((i + 1) * perThreadCount, totalCount);
|
||||
List<Baike> subList = baikes.subList(start, end);
|
||||
Runnable runnable =
|
||||
new BaiKeBeanIndex("index", i, countDownLatch1, countDownLatch2, subList);
|
||||
// 子线程交给线程池管理
|
||||
pool.execute(runnable);
|
||||
}
|
||||
countDownLatch1.countDown();
|
||||
System.out.println("开始创建索引");
|
||||
// 等待所有线程都完成
|
||||
countDownLatch2.await();
|
||||
// 线程全部完成工作
|
||||
System.out.println("所有线程都创建索引完毕");
|
||||
// 释放线程池资源
|
||||
pool.shutdown();
|
||||
} catch (Exception e) {
|
||||
e.printStackTrace();
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 封裝一个方法,用于将数据库中的数据解析为一个个关键字词存储到索引文件中
|
||||
*
|
||||
* @param list
|
||||
*/
|
||||
public void writeArticle(List<ArticleDTO> list) {
|
||||
try {
|
||||
int totalCount = list.size();
|
||||
int perThreadCount = 3000;
|
||||
int threadCount = totalCount / perThreadCount + (totalCount % perThreadCount == 0 ? 0 : 1);
|
||||
ExecutorService pool = Executors.newFixedThreadPool(threadCount);
|
||||
CountDownLatch countDownLatch1 = new CountDownLatch(1);
|
||||
CountDownLatch countDownLatch2 = new CountDownLatch(threadCount);
|
||||
|
||||
for (int i = 0; i < threadCount; i++) {
|
||||
int start = i * perThreadCount;
|
||||
int end = Math.min((i + 1) * perThreadCount, totalCount);
|
||||
List<ArticleDTO> subList = list.subList(start, end);
|
||||
Runnable runnable =
|
||||
new ArticleBeanIndex("articlesIndex", i, countDownLatch1, countDownLatch2, subList);
|
||||
// 子线程交给线程池管理
|
||||
pool.execute(runnable);
|
||||
}
|
||||
countDownLatch1.countDown();
|
||||
System.out.println("开始创建索引");
|
||||
// 等待所有线程都完成
|
||||
countDownLatch2.await();
|
||||
// 线程全部完成工作
|
||||
System.out.println("所有线程都创建索引完毕");
|
||||
// 释放线程池资源
|
||||
pool.shutdown();
|
||||
} catch (Exception e) {
|
||||
e.printStackTrace();
|
||||
}
|
||||
}
|
||||
|
||||
/**
|
||||
* 搜索
|
||||
*
|
||||
* @param value
|
||||
* @return
|
||||
* @throws Exception
|
||||
*/
|
||||
public List<Map<String, String>> search(String value) throws Exception {
|
||||
List<Map<String, String>> list = new ArrayList<>();
|
||||
ExecutorService service = Executors.newCachedThreadPool();
|
||||
// 定义分词器
|
||||
Analyzer analyzer = new IKAnalyzer();
|
||||
try {
|
||||
IndexSearcher searcher = SearchUtil.getIndexSearcherByParentPath(indexPath, service);
|
||||
String[] fields = {"title", "summary"};
|
||||
// 构造Query对象
|
||||
MultiFieldQueryParser parser = new MultiFieldQueryParser(fields, analyzer);
|
||||
|
||||
BufferedReader in =
|
||||
new BufferedReader(new InputStreamReader(System.in, StandardCharsets.UTF_8));
|
||||
String line = value != null ? value : in.readLine();
|
||||
Query query = parser.parse(line);
|
||||
// 最终被分词后添加的前缀和后缀处理器,默认是粗体<B></B>
|
||||
SimpleHTMLFormatter htmlFormatter =
|
||||
new SimpleHTMLFormatter("<font color=" + "\"" + "red" + "\"" + ">", "</font>");
|
||||
// 高亮搜索的词添加到高亮处理器中
|
||||
Highlighter highlighter = new Highlighter(htmlFormatter, new QueryScorer(query));
|
||||
|
||||
// 获取搜索的结果,指定返回document返回的个数
|
||||
// 默认搜索结果为显示第一页,1000 条,可以优化
|
||||
TopDocs results = SearchUtil.getScoreDocsByPerPage(1, 100, searcher, query);
|
||||
ScoreDoc[] hits = results.scoreDocs;
|
||||
|
||||
// 遍历,输出
|
||||
for (ScoreDoc hit : hits) {
|
||||
int id = hit.doc;
|
||||
float score = hit.score;
|
||||
Document hitDoc = searcher.doc(hit.doc);
|
||||
Map<String, String> map = new HashMap<>();
|
||||
map.put("id", hitDoc.get("id"));
|
||||
|
||||
// 获取到summary
|
||||
String name = hitDoc.get("summary");
|
||||
// 将查询的词和搜索词匹配,匹配到添加前缀和后缀
|
||||
TokenStream tokenStream =
|
||||
TokenSources.getAnyTokenStream(searcher.getIndexReader(), id, "summary", analyzer);
|
||||
// 传入的第二个参数是查询的值
|
||||
TextFragment[] frag = highlighter.getBestTextFragments(tokenStream, name, false, 10);
|
||||
StringBuilder baikeValue = new StringBuilder();
|
||||
for (TextFragment textFragment : frag) {
|
||||
if ((textFragment != null) && (textFragment.getScore() > 0)) {
|
||||
// if ((frag[j] != null)) {
|
||||
// 获取 summary 的值
|
||||
baikeValue.append(textFragment.toString());
|
||||
}
|
||||
}
|
||||
|
||||
// 获取到title
|
||||
String title = hitDoc.get("title");
|
||||
TokenStream titleTokenStream =
|
||||
TokenSources.getAnyTokenStream(searcher.getIndexReader(), id, "title", analyzer);
|
||||
TextFragment[] titleFrag =
|
||||
highlighter.getBestTextFragments(titleTokenStream, title, false, 10);
|
||||
StringBuilder titleValue = new StringBuilder();
|
||||
for (int j = 0; j < titleFrag.length; j++) {
|
||||
if ((frag[j] != null)) {
|
||||
titleValue.append(titleFrag[j].toString());
|
||||
}
|
||||
}
|
||||
map.put("title", titleValue.toString());
|
||||
map.put("summary", baikeValue.toString());
|
||||
map.put("score", String.valueOf(score));
|
||||
list.add(map);
|
||||
}
|
||||
} catch (IOException e) {
|
||||
e.printStackTrace();
|
||||
} finally {
|
||||
service.shutdownNow();
|
||||
}
|
||||
return list;
|
||||
}
|
||||
|
||||
/**
|
||||
* 搜索
|
||||
*
|
||||
* @param value
|
||||
* @return
|
||||
* @throws Exception
|
||||
*/
|
||||
public List<Map<String, String>> searchArticle(String value) throws Exception {
|
||||
List<Map<String, String>> list = new ArrayList<>();
|
||||
ExecutorService service = Executors.newCachedThreadPool();
|
||||
// 定义分词器
|
||||
Analyzer analyzer = new IKAnalyzer();
|
||||
try {
|
||||
IndexSearcher searcher =
|
||||
SearchUtil.getIndexSearcherByParentPath(
|
||||
System.getProperty("user.dir") + "/articlesIndex", service);
|
||||
String[] fields = {"title", "summary"};
|
||||
// 构造Query对象
|
||||
MultiFieldQueryParser parser = new MultiFieldQueryParser(fields, analyzer);
|
||||
|
||||
BufferedReader in =
|
||||
new BufferedReader(new InputStreamReader(System.in, StandardCharsets.UTF_8));
|
||||
String line = value != null ? value : in.readLine();
|
||||
Query query = parser.parse(line);
|
||||
// 最终被分词后添加的前缀和后缀处理器,默认是粗体<B></B>
|
||||
SimpleHTMLFormatter htmlFormatter =
|
||||
new SimpleHTMLFormatter("<font color=" + "\"" + "red" + "\"" + ">", "</font>");
|
||||
// 高亮搜索的词添加到高亮处理器中
|
||||
Highlighter highlighter = new Highlighter(htmlFormatter, new QueryScorer(query));
|
||||
|
||||
// 获取搜索的结果,指定返回document返回的个数
|
||||
// 默认搜索结果为显示第一页,1000 条,可以优化
|
||||
TopDocs results = SearchUtil.getScoreDocsByPerPage(1, 100, searcher, query);
|
||||
ScoreDoc[] hits = results.scoreDocs;
|
||||
|
||||
// 遍历,输出
|
||||
for (ScoreDoc hit : hits) {
|
||||
int id = hit.doc;
|
||||
float score = hit.score;
|
||||
Document hitDoc = searcher.doc(hit.doc);
|
||||
Map<String, String> map = new HashMap<>();
|
||||
map.put("id", hitDoc.get("id"));
|
||||
|
||||
// 获取到summary
|
||||
String name = hitDoc.get("summary");
|
||||
// 将查询的词和搜索词匹配,匹配到添加前缀和后缀
|
||||
TokenStream tokenStream =
|
||||
TokenSources.getAnyTokenStream(searcher.getIndexReader(), id, "summary", analyzer);
|
||||
// 传入的第二个参数是查询的值
|
||||
TextFragment[] frag = highlighter.getBestTextFragments(tokenStream, name, false, 10);
|
||||
StringBuilder baikeValue = new StringBuilder();
|
||||
for (TextFragment textFragment : frag) {
|
||||
if ((textFragment != null) && (textFragment.getScore() > 0)) {
|
||||
// if ((frag[j] != null)) {
|
||||
// 获取 summary 的值
|
||||
baikeValue.append(textFragment.toString());
|
||||
}
|
||||
}
|
||||
|
||||
// 获取到title
|
||||
String title = hitDoc.get("title");
|
||||
TokenStream titleTokenStream =
|
||||
TokenSources.getAnyTokenStream(searcher.getIndexReader(), id, "title", analyzer);
|
||||
TextFragment[] titleFrag =
|
||||
highlighter.getBestTextFragments(titleTokenStream, title, false, 10);
|
||||
StringBuilder titleValue = new StringBuilder();
|
||||
for (int j = 0; j < titleFrag.length; j++) {
|
||||
if ((frag[j] != null)) {
|
||||
titleValue.append(titleFrag[j].toString());
|
||||
}
|
||||
}
|
||||
map.put("title", titleValue.toString());
|
||||
map.put("summary", baikeValue.toString());
|
||||
map.put("score", String.valueOf(score));
|
||||
list.add(map);
|
||||
}
|
||||
} catch (IOException e) {
|
||||
e.printStackTrace();
|
||||
} finally {
|
||||
service.shutdownNow();
|
||||
}
|
||||
return list;
|
||||
}
|
||||
}
|
50
src/main/java/com/rymcu/forest/lucene/util/IndexUtil.java
Normal file
50
src/main/java/com/rymcu/forest/lucene/util/IndexUtil.java
Normal file
@ -0,0 +1,50 @@
|
||||
package com.rymcu.forest.lucene.util;
|
||||
|
||||
import com.rymcu.forest.lucene.lucene.IKAnalyzer;
|
||||
import org.apache.lucene.analysis.Analyzer;
|
||||
import org.apache.lucene.index.IndexWriter;
|
||||
import org.apache.lucene.index.IndexWriterConfig;
|
||||
import org.apache.lucene.index.LogByteSizeMergePolicy;
|
||||
import org.apache.lucene.index.LogMergePolicy;
|
||||
import org.apache.lucene.store.Directory;
|
||||
import org.apache.lucene.store.FSDirectory;
|
||||
|
||||
import java.io.IOException;
|
||||
import java.nio.file.Paths;
|
||||
|
||||
|
||||
/**
|
||||
* IndexUtil
|
||||
*
|
||||
* @author suwen
|
||||
* @date 2021/2/2 14:03
|
||||
*/
|
||||
public class IndexUtil {
|
||||
/**
|
||||
* 创建索引写入器
|
||||
* @param indexPath
|
||||
* @param create
|
||||
* @return
|
||||
* @throws IOException
|
||||
*/
|
||||
public static IndexWriter getIndexWriter(String indexPath,boolean create) throws IOException{
|
||||
Directory dir = FSDirectory.open(Paths.get(indexPath, new String[0]));
|
||||
Analyzer analyzer = new IKAnalyzer();
|
||||
IndexWriterConfig iwc = new IndexWriterConfig(analyzer);
|
||||
LogMergePolicy mergePolicy = new LogByteSizeMergePolicy();
|
||||
//设置segment添加文档(Document)时的合并频率
|
||||
// 值较小,建立索引的速度就较慢
|
||||
// 值较大,建立索引的速度就较快,>10适合批量建立索引
|
||||
mergePolicy.setMergeFactor(50);
|
||||
//设置segment最大合并文档(Document)数
|
||||
//值较小有利于追加索引的速度
|
||||
//值较大,适合批量建立索引和更快的搜索
|
||||
mergePolicy.setMaxMergeDocs(5000);
|
||||
if (create){
|
||||
iwc.setOpenMode(IndexWriterConfig.OpenMode.CREATE);
|
||||
}else {
|
||||
iwc.setOpenMode(IndexWriterConfig.OpenMode.CREATE_OR_APPEND);
|
||||
}
|
||||
return new IndexWriter(dir, iwc);
|
||||
}
|
||||
}
|
157
src/main/java/com/rymcu/forest/lucene/util/SearchUtil.java
Normal file
157
src/main/java/com/rymcu/forest/lucene/util/SearchUtil.java
Normal file
@ -0,0 +1,157 @@
|
||||
package com.rymcu.forest.lucene.util;
|
||||
|
||||
import org.apache.lucene.document.Document;
|
||||
import org.apache.lucene.index.DirectoryReader;
|
||||
import org.apache.lucene.index.IndexReader;
|
||||
import org.apache.lucene.index.MultiReader;
|
||||
import org.apache.lucene.search.IndexSearcher;
|
||||
import org.apache.lucene.search.Query;
|
||||
import org.apache.lucene.search.ScoreDoc;
|
||||
import org.apache.lucene.search.TopDocs;
|
||||
import org.apache.lucene.store.FSDirectory;
|
||||
|
||||
import java.io.File;
|
||||
import java.io.IOException;
|
||||
import java.nio.file.Paths;
|
||||
import java.util.Set;
|
||||
import java.util.concurrent.ExecutorService;
|
||||
|
||||
/**
|
||||
* SearchUtil lucene索引查询工具类
|
||||
*
|
||||
* @author suwen
|
||||
* @date 2021/2/2 14:04
|
||||
*/
|
||||
public class SearchUtil {
|
||||
/**
|
||||
* 获取IndexSearcher对象
|
||||
*
|
||||
* @param service
|
||||
* @return
|
||||
* @throws IOException
|
||||
*/
|
||||
public static IndexSearcher getIndexSearcherByParentPath(
|
||||
String parentPath, ExecutorService service) {
|
||||
MultiReader reader = null;
|
||||
// 设置
|
||||
try {
|
||||
File[] files = new File(parentPath).listFiles();
|
||||
IndexReader[] readers = new IndexReader[files.length];
|
||||
for (int i = 0; i < files.length; i++) {
|
||||
readers[i] =
|
||||
DirectoryReader.open(FSDirectory.open(Paths.get(files[i].getPath(), new String[0])));
|
||||
}
|
||||
reader = new MultiReader(readers);
|
||||
} catch (IOException e) {
|
||||
e.printStackTrace();
|
||||
}
|
||||
return new IndexSearcher(reader, service);
|
||||
}
|
||||
/**
|
||||
* 根据索引路径获取IndexReader
|
||||
*
|
||||
* @param indexPath
|
||||
* @return
|
||||
* @throws IOException
|
||||
*/
|
||||
public static DirectoryReader getIndexReader(String indexPath) throws IOException {
|
||||
return DirectoryReader.open(FSDirectory.open(Paths.get(indexPath, new String[0])));
|
||||
}
|
||||
/**
|
||||
* 根据索引路径获取IndexSearcher
|
||||
*
|
||||
* @param indexPath
|
||||
* @param service
|
||||
* @return
|
||||
* @throws IOException
|
||||
*/
|
||||
public static IndexSearcher getIndexSearcherByIndexPath(String indexPath, ExecutorService service)
|
||||
throws IOException {
|
||||
IndexReader reader = getIndexReader(indexPath);
|
||||
return new IndexSearcher(reader, service);
|
||||
}
|
||||
|
||||
/**
|
||||
* 如果索引目录会有变更用此方法获取新的IndexSearcher这种方式会占用较少的资源
|
||||
*
|
||||
* @param oldSearcher
|
||||
* @param service
|
||||
* @return
|
||||
* @throws IOException
|
||||
*/
|
||||
public static IndexSearcher getIndexSearcherOpenIfChanged(
|
||||
IndexSearcher oldSearcher, ExecutorService service) throws IOException {
|
||||
DirectoryReader reader = (DirectoryReader) oldSearcher.getIndexReader();
|
||||
DirectoryReader newReader = DirectoryReader.openIfChanged(reader);
|
||||
return new IndexSearcher(newReader, service);
|
||||
}
|
||||
|
||||
/**
|
||||
* 根据IndexSearcher和docID获取默认的document
|
||||
*
|
||||
* @param searcher
|
||||
* @param docID
|
||||
* @return
|
||||
* @throws IOException
|
||||
*/
|
||||
public static Document getDefaultFullDocument(IndexSearcher searcher, int docID)
|
||||
throws IOException {
|
||||
return searcher.doc(docID);
|
||||
}
|
||||
/**
|
||||
* 根据IndexSearcher和docID
|
||||
*
|
||||
* @param searcher
|
||||
* @param docID
|
||||
* @param listField
|
||||
* @return
|
||||
* @throws IOException
|
||||
*/
|
||||
public static Document getDocumentByListField(
|
||||
IndexSearcher searcher, int docID, Set<String> listField) throws IOException {
|
||||
return searcher.doc(docID, listField);
|
||||
}
|
||||
|
||||
/**
|
||||
* 分页查询
|
||||
*
|
||||
* @param page 当前页数
|
||||
* @param perPage 每页显示条数
|
||||
* @param searcher searcher查询器
|
||||
* @param query 查询条件
|
||||
* @return
|
||||
* @throws IOException
|
||||
*/
|
||||
public static TopDocs getScoreDocsByPerPage(
|
||||
int page, int perPage, IndexSearcher searcher, Query query) throws IOException {
|
||||
TopDocs result = null;
|
||||
if (query == null) {
|
||||
System.out.println(" Query is null return null ");
|
||||
return null;
|
||||
}
|
||||
ScoreDoc before = null;
|
||||
if (page != 1) {
|
||||
TopDocs docsBefore = searcher.search(query, (page - 1) * perPage);
|
||||
ScoreDoc[] scoreDocs = docsBefore.scoreDocs;
|
||||
if (scoreDocs.length > 0) {
|
||||
before = scoreDocs[scoreDocs.length - 1];
|
||||
}
|
||||
}
|
||||
result = searcher.searchAfter(before, query, perPage);
|
||||
return result;
|
||||
}
|
||||
|
||||
public static TopDocs getScoreDocs(IndexSearcher searcher, Query query) throws IOException {
|
||||
TopDocs docs = searcher.search(query, getMaxDocId(searcher));
|
||||
return docs;
|
||||
}
|
||||
/**
|
||||
* 统计document的数量,此方法等同于matchAllDocsQuery查询
|
||||
*
|
||||
* @param searcher
|
||||
* @return
|
||||
*/
|
||||
public static int getMaxDocId(IndexSearcher searcher) {
|
||||
return searcher.getIndexReader().maxDoc();
|
||||
}
|
||||
}
|
21
src/main/java/mapper/BaikeMapper.xml
Executable file
21
src/main/java/mapper/BaikeMapper.xml
Executable file
@ -0,0 +1,21 @@
|
||||
<?xml version="1.0" encoding="UTF-8" ?>
|
||||
<!DOCTYPE mapper PUBLIC "-//mybatis.org//DTD Mapper 3.0//EN" "http://mybatis.org/dtd/mybatis-3-mapper.dtd" >
|
||||
<mapper namespace="com.rymcu.forest.lucene.mapper.BaikeMapper" >
|
||||
<resultMap id="BaseResultMap" type="com.rymcu.forest.lucene.model.Baike" >
|
||||
<id column="id" property="id" jdbcType="INTEGER" />
|
||||
<result column="title" property="title" jdbcType="VARCHAR" />
|
||||
</resultMap>
|
||||
<resultMap id="ResultMapWithBLOBs" type="com.rymcu.forest.lucene.model.Baike" extends="BaseResultMap" >
|
||||
<result column="summary" property="summary" jdbcType="LONGVARCHAR" />
|
||||
</resultMap>
|
||||
|
||||
<sql id="Base_Column_List" >
|
||||
id, title
|
||||
</sql>
|
||||
<sql id="Blob_Column_List" >
|
||||
summary
|
||||
</sql>
|
||||
<select id="getAllBaike" resultType="com.rymcu.forest.lucene.model.Baike">
|
||||
select * from lucene_baike limit #{limit} offset #{offset};
|
||||
</select>
|
||||
</mapper>
|
8
src/main/resources/lucene/dictionary.dic
Normal file
8
src/main/resources/lucene/dictionary.dic
Normal file
@ -0,0 +1,8 @@
|
||||
屌丝男士
|
||||
吊死男士
|
||||
diaosinanshi
|
||||
男
|
||||
吊
|
||||
男屌丝
|
||||
男吊丝士
|
||||
吊男丝士
|
4
src/main/resources/lucene/ext.dic
Executable file
4
src/main/resources/lucene/ext.dic
Executable file
@ -0,0 +1,4 @@
|
||||
全文搜索
|
||||
苏雄
|
||||
杨跃进
|
||||
癜风
|
275716
src/main/resources/lucene/main2012.dic
Executable file
275716
src/main/resources/lucene/main2012.dic
Executable file
File diff suppressed because it is too large
Load Diff
316
src/main/resources/lucene/quantifier.dic
Executable file
316
src/main/resources/lucene/quantifier.dic
Executable file
@ -0,0 +1,316 @@
|
||||
丈
|
||||
下
|
||||
世
|
||||
世纪
|
||||
两
|
||||
个
|
||||
中
|
||||
串
|
||||
亩
|
||||
人
|
||||
介
|
||||
付
|
||||
代
|
||||
件
|
||||
任
|
||||
份
|
||||
伏
|
||||
伙
|
||||
位
|
||||
位数
|
||||
例
|
||||
倍
|
||||
像素
|
||||
元
|
||||
克
|
||||
克拉
|
||||
公亩
|
||||
公克
|
||||
公分
|
||||
公升
|
||||
公尺
|
||||
公担
|
||||
公斤
|
||||
公里
|
||||
公顷
|
||||
具
|
||||
册
|
||||
出
|
||||
刀
|
||||
分
|
||||
分钟
|
||||
分米
|
||||
划
|
||||
列
|
||||
则
|
||||
刻
|
||||
剂
|
||||
剑
|
||||
副
|
||||
加仑
|
||||
勺
|
||||
包
|
||||
匙
|
||||
匹
|
||||
区
|
||||
千克
|
||||
千米
|
||||
升
|
||||
卷
|
||||
厅
|
||||
厘
|
||||
厘米
|
||||
双
|
||||
发
|
||||
口
|
||||
句
|
||||
只
|
||||
台
|
||||
叶
|
||||
号
|
||||
名
|
||||
吨
|
||||
听
|
||||
员
|
||||
周
|
||||
周年
|
||||
品
|
||||
回
|
||||
团
|
||||
圆
|
||||
圈
|
||||
地
|
||||
场
|
||||
块
|
||||
坪
|
||||
堆
|
||||
声
|
||||
壶
|
||||
处
|
||||
夜
|
||||
大
|
||||
天
|
||||
头
|
||||
套
|
||||
女
|
||||
孔
|
||||
字
|
||||
宗
|
||||
室
|
||||
家
|
||||
寸
|
||||
对
|
||||
封
|
||||
尊
|
||||
小时
|
||||
尺
|
||||
尾
|
||||
局
|
||||
层
|
||||
届
|
||||
岁
|
||||
师
|
||||
帧
|
||||
幅
|
||||
幕
|
||||
幢
|
||||
平方
|
||||
平方公尺
|
||||
平方公里
|
||||
平方分米
|
||||
平方厘米
|
||||
平方码
|
||||
平方米
|
||||
平方英寸
|
||||
平方英尺
|
||||
平方英里
|
||||
平米
|
||||
年
|
||||
年代
|
||||
年级
|
||||
度
|
||||
座
|
||||
式
|
||||
引
|
||||
张
|
||||
成
|
||||
战
|
||||
截
|
||||
户
|
||||
房
|
||||
所
|
||||
扇
|
||||
手
|
||||
打
|
||||
批
|
||||
把
|
||||
折
|
||||
担
|
||||
拍
|
||||
招
|
||||
拨
|
||||
拳
|
||||
指
|
||||
掌
|
||||
排
|
||||
撮
|
||||
支
|
||||
文
|
||||
斗
|
||||
斤
|
||||
方
|
||||
族
|
||||
日
|
||||
时
|
||||
曲
|
||||
月
|
||||
月份
|
||||
期
|
||||
本
|
||||
朵
|
||||
村
|
||||
束
|
||||
条
|
||||
来
|
||||
杯
|
||||
枚
|
||||
枝
|
||||
枪
|
||||
架
|
||||
柄
|
||||
柜
|
||||
栋
|
||||
栏
|
||||
株
|
||||
样
|
||||
根
|
||||
格
|
||||
案
|
||||
桌
|
||||
档
|
||||
桩
|
||||
桶
|
||||
梯
|
||||
棵
|
||||
楼
|
||||
次
|
||||
款
|
||||
步
|
||||
段
|
||||
毛
|
||||
毫
|
||||
毫升
|
||||
毫米
|
||||
毫克
|
||||
池
|
||||
洲
|
||||
派
|
||||
海里
|
||||
滴
|
||||
炮
|
||||
点
|
||||
点钟
|
||||
片
|
||||
版
|
||||
环
|
||||
班
|
||||
瓣
|
||||
瓶
|
||||
生
|
||||
男
|
||||
画
|
||||
界
|
||||
盆
|
||||
盎司
|
||||
盏
|
||||
盒
|
||||
盘
|
||||
相
|
||||
眼
|
||||
石
|
||||
码
|
||||
碗
|
||||
碟
|
||||
磅
|
||||
种
|
||||
科
|
||||
秒
|
||||
秒钟
|
||||
窝
|
||||
立方公尺
|
||||
立方分米
|
||||
立方厘米
|
||||
立方码
|
||||
立方米
|
||||
立方英寸
|
||||
立方英尺
|
||||
站
|
||||
章
|
||||
笔
|
||||
等
|
||||
筐
|
||||
筒
|
||||
箱
|
||||
篇
|
||||
篓
|
||||
篮
|
||||
簇
|
||||
米
|
||||
类
|
||||
粒
|
||||
级
|
||||
组
|
||||
维
|
||||
缕
|
||||
缸
|
||||
罐
|
||||
网
|
||||
群
|
||||
股
|
||||
脚
|
||||
船
|
||||
艇
|
||||
艘
|
||||
色
|
||||
节
|
||||
英亩
|
||||
英寸
|
||||
英尺
|
||||
英里
|
||||
行
|
||||
袋
|
||||
角
|
||||
言
|
||||
课
|
||||
起
|
||||
趟
|
||||
路
|
||||
车
|
||||
转
|
||||
轮
|
||||
辆
|
||||
辈
|
||||
连
|
||||
通
|
||||
遍
|
||||
部
|
||||
里
|
||||
重
|
||||
针
|
||||
钟
|
||||
钱
|
||||
锅
|
||||
门
|
||||
间
|
||||
队
|
||||
阶段
|
||||
隅
|
||||
集
|
||||
页
|
||||
顶
|
||||
顷
|
||||
项
|
||||
顿
|
||||
颗
|
||||
餐
|
||||
首
|
10231
src/main/resources/lucene/sensitive.dic
Executable file
10231
src/main/resources/lucene/sensitive.dic
Executable file
File diff suppressed because it is too large
Load Diff
64
src/main/resources/lucene/stopword.dic
Executable file
64
src/main/resources/lucene/stopword.dic
Executable file
@ -0,0 +1,64 @@
|
||||
a
|
||||
an
|
||||
and
|
||||
are
|
||||
as
|
||||
at
|
||||
be
|
||||
but
|
||||
by
|
||||
for
|
||||
if
|
||||
in
|
||||
into
|
||||
is
|
||||
it
|
||||
no
|
||||
not
|
||||
of
|
||||
on
|
||||
or
|
||||
such
|
||||
that
|
||||
the
|
||||
their
|
||||
then
|
||||
there
|
||||
these
|
||||
they
|
||||
this
|
||||
to
|
||||
was
|
||||
will
|
||||
with
|
||||
也
|
||||
了
|
||||
仍
|
||||
从
|
||||
以
|
||||
使
|
||||
则
|
||||
却
|
||||
又
|
||||
及
|
||||
对
|
||||
就
|
||||
并
|
||||
很
|
||||
或
|
||||
把
|
||||
是
|
||||
的
|
||||
着
|
||||
给
|
||||
而
|
||||
被
|
||||
让
|
||||
在
|
||||
还
|
||||
比
|
||||
等
|
||||
当
|
||||
与
|
||||
于
|
||||
但
|
2108
src/main/resources/lucene/web.dic
Executable file
2108
src/main/resources/lucene/web.dic
Executable file
File diff suppressed because it is too large
Load Diff
Loading…
x
Reference in New Issue
Block a user