代码拉取完成,页面将自动刷新
7.17.x
分支,支持es 7.17.0
,JDK版本:11.0.7
, gradle版本:7.6
8.4.1
分支,支持es 8.4.1
,JDK版本:18.0.2.1
, gradle版本:7.6
gradle wrapper --gradle-version 7.6
jieba analysis plugin for elasticsearch: 7.7.0, 7.4.2, 7.3.0, 7.0.0, 6.4.0, 6.0.0 , 5.4.0, 5.3.0, 5.2.2, 5.2.1, 5.2.0, 5.1.2, 5.1.1
分支 | tag | elasticsearch版本 | Release Link |
---|---|---|---|
7.7.0 | tag v7.7.1 | v7.7.0 | Download: v7.7.0 |
7.4.2 | tag v7.4.2 | v7.4.2 | Download: v7.4.2 |
7.3.0 | tag v7.3.0 | v7.3.0 | Download: v7.3.0 |
7.0.0 | tag v7.0.0 | v7.0.0 | Download: v7.0.0 |
6.4.0 | tag v6.4.1 | v6.4.0 | Download: v6.4.1 |
6.4.0 | tag v6.4.0 | v6.4.0 | Download: v6.4.0 |
6.0.0 | tag v6.0.0 | v6.0.0 | Download: v6.0.1 |
5.4.0 | tag v5.4.0 | v5.4.0 | Download: v5.4.0 |
5.3.0 | tag v5.3.0 | v5.3.0 | Download: v5.3.0 |
5.2.2 | tag v5.2.2 | v5.2.2 | Download: v5.2.2 |
5.2.1 | tag v5.2.1 | v5.2.1 | Download: v5.2.1 |
5.2 | tag v5.2.0 | v5.2.0 | Download: v5.2.0 |
5.1.2 | tag v5.1.2 | v5.1.2 | Download: v5.1.2 |
5.1.1 | tag v5.1.1 | v5.1.1 | Download: v5.1.1 |
git clone https://github.com/sing1ee/elasticsearch-jieba-plugin.git --recursive
./gradlew clean pz
cp build/distributions/elasticsearch-jieba-plugin-5.1.2.zip ${path.home}/plugins
unzip elasticsearch-jieba-plugin-5.1.2.zip
rm elasticsearch-jieba-plugin-5.1.2.zip
./bin/elasticsearch
Just put you dict file with suffix .dict into ${path.home}/plugins/jieba/dic. Your dict file should like this:
小清新 3
百搭 3
显瘦 3
隨身碟 100
your_word word_freq
mkdir -p {path.home}/config/stopwords
cp ${path.home}/plugins/jieba/dic/stopwords.txt {path.home}/config/stopwords
PUT http://localhost:9200/jieba_index
{
"settings": {
"analysis": {
"filter": {
"jieba_stop": {
"type": "stop",
"stopwords_path": "stopwords/stopwords.txt"
},
"jieba_synonym": {
"type": "synonym",
"synonyms_path": "synonyms/synonyms.txt"
}
},
"analyzer": {
"my_ana": {
"tokenizer": "jieba_index",
"filter": [
"lowercase",
"jieba_stop",
"jieba_synonym"
]
}
}
}
}
}
PUT http://localhost:9200/jieba_index/_analyze
{
"analyzer" : "my_ana",
"text" : "黄河之水天上来"
}
Response as follow:
{
"tokens": [
{
"token": "黄河",
"start_offset": 0,
"end_offset": 2,
"type": "word",
"position": 0
},
{
"token": "黄河之水天上来",
"start_offset": 0,
"end_offset": 7,
"type": "word",
"position": 0
},
{
"token": "之水",
"start_offset": 2,
"end_offset": 4,
"type": "word",
"position": 1
},
{
"token": "天上",
"start_offset": 4,
"end_offset": 6,
"type": "word",
"position": 2
},
{
"token": "上来",
"start_offset": 5,
"end_offset": 7,
"type": "word",
"position": 2
}
]
}
migrate from jieba-solr
I will add more analyzer support:
If you have some ideas, you should create an issue. Then, we will do it together.
此处可能存在不合适展示的内容,页面不予展示。您可通过相关编辑功能自查并修改。
如您确认内容无涉及 不当用语 / 纯广告导流 / 暴力 / 低俗色情 / 侵权 / 盗版 / 虚假 / 无价值内容或违法国家有关法律法规的内容,可点击提交进行申诉,我们将尽快为您处理。