es配置中文和拼音分词器

版权声明:本文为博主原创文章,未经博主允许不得转载。 https://blog.csdn.net/u013219624/article/details/83868620

1.简介

es默认使用standard分词器 es还有其他分词器比如simple  writespace language

2.配置中文分词器(需先安装git maven unzip)

git clone https://github.com/medcl/elasticsearch-analysis-ik.git
cd elasticsearch-analysis-ik
mvn package

mkdir /usr/local/elasticsearch/plugins/ik
mv target/releases/elasticsearch-analysis-ik-6.4.0.zip /usr/local/elasticsearch/plugins/ik
cd /usr/local/elasticsearch/plugins/ik
unzip elasticsearch-analysis-ik-6.4.0.zip

rm -rf elasticsearch-analysis-ik-6.4.0.zip

chown -R es:gp ik
./bin/elasticsearch -d

3.测试中文分词器

# ik_max_word尽可能多切分单词
GET _analyze
{
  "analyzer":"ik_max_word",
  "text":"阿里巴巴"
}

ik_smart尽可能少切分单词
GET _analyze
{
  "analyzer":"ik_smart",
  "text":"阿里巴巴"
}

4.配置拼音分词器

git clone https://github.com/medcl/elasticsearch-analysis-pinyin.git
cd elasticsearch-analysis-pinyin
mvn package

mkdir /usr/local/elasticsearch/plugins/pinyin
mv target/releases/elasticsearch-analysis-pinyin-6.4.0.zip /usr/local/elasticsearch/plugins/pinyin
cd /usr/local/elasticsearch/plugins/pinyin
unzip elasticsearch-analysis-pinyin-6.4.0.zip

rm -rf elasticsearch-analysis-pinyin-6.4.0.zip

chown -R es:gp pinyin
./bin/elasticsearch -d

5.测试拼音分词器

GET _analyze
{
  "text": "阿里巴巴",
  "analyzer": "pinyin"
}

猜你喜欢

转载自blog.csdn.net/u013219624/article/details/83868620