"Stutter" Chinese word segmentation: Be the best PHP Chinese word segmentation and Chinese word segmentation component. The current translation version is jieba-0.33 version. In the future, it will be upgraded slowly and the performance needs to be improved. Interested developers are asked to join the development together! If you want to use the Python version, please go to fxsjy/jieba
Traditional Chinese can now be supported! Just switch the dictionary to big mode!
"Jieba" (Chinese for "to stutter") Chinese text segmentation: built to be the best PHP Chinese word segmentation module.
Scroll down for English documentation.
Supports three word participle modes:
1) Default precision mode, trying to cut sentences most accurately, suitable for text analysis;
2) Full pattern, scan out all words in the sentence that can be word-based, but cannot resolve the ambiguity. (Add sufficient dictionary is required)
Support traditional verb verb
Supports custom dictionary
Code Example
composer require fukuball/jieba-php:dev-masterCode Example
require_once " /path/to/your/vendor/autoload.php " ;Code Example
require_once " /path/to/your/vendor/multi-array/MultiArray.php " ;
require_once " /path/to/your/vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once " /path/to/your/class/Jieba.php " ;
require_once " /path/to/your/class/Finalseg.php " ;cut method accepts an input parameter: 1) The first parameter is the string that requires word segmentation 2) The cut_all parameter is used to control word segmentation patterncutForSearch method accepts a parameter: a string that requires word segmentation. This method is suitable for search engines to build inverted indexed word segmentation, with a relatively fine granularitycut and cutForSearch is an iterable arrayCode example (Tutorial)
ini_set ( ' memory_limit ' , ' 1024M ' );
require_once " /path/to/your/vendor/multi-array/MultiArray.php " ;
require_once " /path/to/your/vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once " /path/to/your/class/Jieba.php " ;
require_once " /path/to/your/class/Finalseg.php " ;
use Fukuball Jieba Jieba ;
use Fukuball Jieba Finalseg ;
Jieba:: init ();
Finalseg:: init ();
$ seg_list = Jieba:: cut ( "怜香惜玉也得要看对象啊! " );
var_dump ( $ seg_list );
$ seg_list = Jieba:: cut ( "我来到北京清华大学" , true );
var_dump ( $ seg_list ); #全模式
$ seg_list = Jieba:: cut ( "我来到北京清华大学" , false );
var_dump ( $ seg_list ); #默認精確模式
$ seg_list = Jieba:: cut ( "他来到了网易杭研大厦" );
var_dump ( $ seg_list );
$ seg_list = Jieba:: cutForSearch ( "小明硕士毕业于中国科学院计算所,后在日本京都大学深造" ); #搜索引擎模式
var_dump ( $ seg_list );Output:
array ( 7 ) {
[ 0 ]=>
string( 12 ) "怜香惜玉"
[ 1 ]=>
string( 3 ) "也"
[ 2 ]=>
string( 3 ) "得"
[ 3 ]=>
string( 3 ) "要"
[ 4 ]=>
string( 3 ) "看"
[ 5 ]=>
string( 6 ) "对象"
[ 6 ]=>
string( 3 ) "啊"
}
Full Mode:
array ( 15 ) {
[ 0 ]=>
string( 3 ) "我"
[ 1 ]=>
string( 3 ) "来"
[ 2 ]=>
string( 6 ) "来到"
[ 3 ]=>
string( 3 ) "到"
[ 4 ]=>
string( 3 ) "北"
[ 5 ]=>
string( 6 ) "北京"
[ 6 ]=>
string( 3 ) "京"
[ 7 ]=>
string( 3 ) "清"
[ 8 ]=>
string( 6 ) "清华"
[ 9 ]=>
string( 12 ) "清华大学"
[ 10 ]=>
string( 3 ) "华"
[ 11 ]=>
string( 6 ) "华大"
[ 12 ]=>
string( 3 ) "大"
[ 13 ]=>
string( 6 ) "大学"
[ 14 ]=>
string( 3 ) "学"
}
Default Mode:
array ( 4 ) {
[ 0 ]=>
string( 3 ) "我"
[ 1 ]=>
string( 6 ) "来到"
[ 2 ]=>
string( 6 ) "北京"
[ 3 ]=>
string( 12 ) "清华大学"
}
array ( 6 ) {
[ 0 ]=>
string( 3 ) "他"
[ 1 ]=>
string( 6 ) "来到"
[ 2 ]=>
string( 3 ) "了"
[ 3 ]=>
string( 6 ) "网易"
[ 4 ]=>
string( 6 ) "杭研"
[ 5 ]=>
string( 6 ) "大厦"
}
(此處,“杭研“並沒有在詞典中,但是也被 Viterbi 算法識別出來了)
Search Engine Mode:
array ( 18 ) {
[ 0 ]=>
string( 6 ) "小明"
[ 1 ]=>
string( 6 ) "硕士"
[ 2 ]=>
string( 6 ) "毕业"
[ 3 ]=>
string( 3 ) "于"
[ 4 ]=>
string( 6 ) "中国"
[ 5 ]=>
string( 6 ) "科学"
[ 6 ]=>
string( 6 ) "学院"
[ 7 ]=>
string( 9 ) "科学院"
[ 8 ]=>
string( 15 ) "中国科学院"
[ 9 ]=>
string( 6 ) "计算"
[ 10 ]=>
string( 9 ) "计算所"
[ 11 ]=>
string( 3 ) "后"
[ 12 ]=>
string( 3 ) "在"
[ 13 ]=>
string( 6 ) "日本"
[ 14 ]=>
string( 6 ) "京都"
[ 15 ]=>
string( 6 ) "大学"
[ 16 ]=>
string( 18 ) "日本京都大学"
[ 17 ]=>
string( 6 ) "深造"
}Developers can specify their own custom dictionary to include words that are not in the jieba vocabulary. Although jieba has the ability to recognize new words, adding new words on your own can ensure higher accuracy
Usage: Jieba::loadUserDict(file_name) # file_name is the absolute path to a custom dictionary
The dictionary format is the same as dict.txt, and a word occupies one line; each line is divided into three parts, one part of the word, one part of the word, and one part of the word, separated by spaces.
example:
Cloud computing 5 n Li Xiaofu 2 n Innovation office 3 n
Previously: Li Xiaofu/Yes/Innovation/Office/Director/Yes/Yes/Cloud/Computing/Evidence/After loading the custom thesaurus: Li Xiaofu/Yes/Innovation/Office/Director/Yes/Yes/Yes/Cloud/Evidence/
Description: "Enhance ambiguity error correction through user-defined dictionary" --- fxsjy/jieba#14
Code example (keyword extraction)
ini_set ( ' memory_limit ' , ' 600M ' );
require_once " /path/to/your/vendor/multi-array/MultiArray.php " ;
require_once " /path/to/your/vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once " /path/to/your/class/Jieba.php " ;
require_once " /path/to/your/class/Finalseg.php " ;
require_once " /path/to/your/class/JiebaAnalyse.php " ;
use Fukuball Jieba Jieba ;
use Fukuball Jieba Finalseg ;
use Fukuball Jieba JiebaAnalyse ;
Jieba:: init ( array ( ' mode ' => ' test ' , ' dict ' => ' small ' ));
Finalseg:: init ();
JiebaAnalyse:: init ();
$ top_k = 10 ;
$ content = file_get_contents ( " /path/to/your/dict/lyric.txt " , " r " );
$ tags = JiebaAnalyse:: extractTags ( $ content , $ top_k );
var_dump ( $ tags );
JiebaAnalyse:: setStopWords ( ' /path/to/your/dict/stop_words.txt ' );
$ tags = JiebaAnalyse:: extractTags ( $ content , $ top_k );
var_dump ( $ tags );Output:
array ( 10 ) {
'沒有' =>
double( 1.0592831964595 )
'所謂' =>
double( 0.90795702553671 )
'是否' =>
double( 0.66385043195443 )
'一般' =>
double( 0.54607060161899 )
'雖然' =>
double( 0.30265234184557 )
'來說' =>
double( 0.30265234184557 )
'肌迫' =>
double( 0.30265234184557 )
'退縮' =>
double( 0.30265234184557 )
'矯作' =>
double( 0.30265234184557 )
'怯懦' =>
double( 0.24364586159392 )
}
array ( 10 ) {
'所謂' =>
double( 1.1569129841516 )
'一般' =>
double( 0.69579963754677 )
'矯作' =>
double( 0.38563766138387 )
'來說' =>
double( 0.38563766138387 )
'退縮' =>
double( 0.38563766138387 )
'雖然' =>
double( 0.38563766138387 )
'肌迫' =>
double( 0.38563766138387 )
'怯懦' =>
double( 0.31045198493419 )
'隨便說說' =>
double( 0.19281883069194 )
'一場' =>
double( 0.19281883069194 )
}Code example (Tutorial)
ini_set ( ' memory_limit ' , ' 600M ' );
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/MultiArray.php " ;
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Jieba.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Finalseg.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Posseg.php " ;
use Fukuball Jieba Jieba ;
use Fukuball Jieba Finalseg ;
use Fukuball Jieba Posseg ;
Jieba:: init ();
Finalseg:: init ();
Posseg:: init ();
$ seg_list = Posseg:: cut ( "这是一个伸手不见五指的黑夜。我叫孙悟空,我爱北京,我爱Python和C++。 " );
var_dump ( $ seg_list );Output:
array ( 21 ) {
[ 0 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "这"
[ " tag " ]=>
string( 1 ) " r "
}
[ 1 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "是"
[ " tag " ]=>
string( 1 ) " v "
}
[ 2 ]=>
array ( 2 ) {
[ " word " ]=>
string( 6 ) "一个"
[ " tag " ]=>
string( 1 ) " m "
}
[ 3 ]=>
array ( 2 ) {
[ " word " ]=>
string( 18 ) "伸手不见五指"
[ " tag " ]=>
string( 1 ) " i "
}
[ 4 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "的"
[ " tag " ]=>
string( 2 ) " uj "
}
[ 5 ]=>
array ( 2 ) {
[ " word " ]=>
string( 6 ) "黑夜"
[ " tag " ]=>
string( 1 ) " n "
}
[ 6 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) " 。 "
[ " tag " ]=>
string( 1 ) " x "
}
[ 7 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "我"
[ " tag " ]=>
string( 1 ) " r "
}
[ 8 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "叫"
[ " tag " ]=>
string( 1 ) " v "
}
[ 9 ]=>
array ( 2 ) {
[ " word " ]=>
string( 9 ) "孙悟空"
[ " tag " ]=>
string( 2 ) " nr "
}
[ 10 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) " , "
[ " tag " ]=>
string( 1 ) " x "
}
[ 11 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "我"
[ " tag " ]=>
string( 1 ) " r "
}
[ 12 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "爱"
[ " tag " ]=>
string( 1 ) " v "
}
[ 13 ]=>
array ( 2 ) {
[ " word " ]=>
string( 6 ) "北京"
[ " tag " ]=>
string( 2 ) " ns "
}
[ 14 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) " , "
[ " tag " ]=>
string( 1 ) " x "
}
[ 15 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "我"
[ " tag " ]=>
string( 1 ) " r "
}
[ 16 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "爱"
[ " tag " ]=>
string( 1 ) " v "
}
[ 17 ]=>
array ( 2 ) {
[ " word " ]=>
string( 6 ) " Python "
[ " tag " ]=>
string( 3 ) " eng "
}
[ 18 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "和"
[ " tag " ]=>
string( 1 ) " c "
}
[ 19 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) " C++ "
[ " tag " ]=>
string( 3 ) " eng "
}
[ 20 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) " 。 "
[ " tag " ]=>
string( 1 ) " x "
}
}Code example (Tutorial)
ini_set ( ' memory_limit ' , ' 1024M ' );
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/MultiArray.php " ;
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Jieba.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Finalseg.php " ;
use Fukuball Jieba Jieba ;
use Fukuball Jieba Finalseg ;
Jieba:: init ( array ( ' mode ' => ' default ' , ' dict ' => ' big ' ));
Finalseg:: init ();
$ seg_list = Jieba:: cut ( "怜香惜玉也得要看对象啊! " );
var_dump ( $ seg_list );
$ seg_list = Jieba:: cut ( "憐香惜玉也得要看對象啊! " );
var_dump ( $ seg_list );Output:
array ( 7 ) {
[ 0 ]=>
string( 12 ) "怜香惜玉"
[ 1 ]=>
string( 3 ) "也"
[ 2 ]=>
string( 3 ) "得"
[ 3 ]=>
string( 3 ) "要"
[ 4 ]=>
string( 3 ) "看"
[ 5 ]=>
string( 6 ) "对象"
[ 6 ]=>
string( 3 ) "啊"
}
array ( 7 ) {
[ 0 ]=>
string( 12 ) "憐香惜玉"
[ 1 ]=>
string( 3 ) "也"
[ 2 ]=>
string( 3 ) "得"
[ 3 ]=>
string( 3 ) "要"
[ 4 ]=>
string( 3 ) "看"
[ 5 ]=>
string( 6 ) "對象"
[ 6 ]=>
string( 3 ) "啊"
}Code example (Tutorial)
ini_set ( ' memory_limit ' , ' 1024M ' );
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/MultiArray.php " ;
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Jieba.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Finalseg.php " ;
use Fukuball Jieba Jieba ;
use Fukuball Jieba Finalseg ;
Jieba:: init ( array ( ' mode ' => ' default ' , ' dict ' => ' big ' ));
Finalseg:: init ();
$ seg_list = Jieba:: cut ( "怜香惜玉也得要看对象啊! " );
var_dump ( $ seg_list );
$ seg_list = Jieba:: cut ( "憐香惜玉也得要看對象啊! " );
var_dump ( $ seg_list );Output:
array ( 7 ) {
[ 0 ]=>
string( 12 ) "怜香惜玉"
[ 1 ]=>
string( 3 ) "也"
[ 2 ]=>
string( 3 ) "得"
[ 3 ]=>
string( 3 ) "要"
[ 4 ]=>
string( 3 ) "看"
[ 5 ]=>
string( 6 ) "对象"
[ 6 ]=>
string( 3 ) "啊"
}
array ( 7 ) {
[ 0 ]=>
string( 12 ) "憐香惜玉"
[ 1 ]=>
string( 3 ) "也"
[ 2 ]=>
string( 3 ) "得"
[ 3 ]=>
string( 3 ) "要"
[ 4 ]=>
string( 3 ) "看"
[ 5 ]=>
string( 6 ) "對象"
[ 6 ]=>
string( 3 ) "啊"
}Code example (Tutorial)
ini_set ( ' memory_limit ' , ' 1024M ' );
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/MultiArray.php " ;
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Jieba.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Finalseg.php " ;
use Fukuball Jieba Jieba ;
use Fukuball Jieba Finalseg ;
Jieba:: init ( array ( ' cjk ' => ' all ' ));
Finalseg:: init ();
$ seg_list = Jieba:: cut ( " 한국어 또는 조선말은 제주특별자치도를 제외한 한반도 및 그 부속 도서와 한민족 거주 지역에서 쓰이는 언어로 " );
var_dump ( $ seg_list );
$ seg_list = Jieba:: cut ( "日本語は、主に日本国内や日本人同士の間で使われている言語である。 " );
var_dump ( $ seg_list );
// 加载日语词库可以对日语进行简单的分词
Jieba:: loadUserDict ( " /path/to/your/japanese/dict.txt " );
$ seg_list = Jieba:: cut ( "日本語は、主に日本国内や日本人同士の間で使われている言語である。 " );
var_dump ( $ seg_list );Output:
array ( 15 ) {
[ 0 ]=>
string( 9 ) " 한국어 "
[ 1 ]=>
string( 6 ) " 또는 "
[ 2 ]=>
string( 12 ) " 조선말은 "
[ 3 ]=>
string( 24 ) " 제주특별자치도를 "
[ 4 ]=>
string( 9 ) " 제외한 "
[ 5 ]=>
string( 9 ) " 한반도 "
[ 6 ]=>
string( 3 ) " 및 "
[ 7 ]=>
string( 3 ) " 그 "
[ 8 ]=>
string( 6 ) " 부속 "
[ 9 ]=>
string( 9 ) " 도서와 "
[ 10 ]=>
string( 9 ) " 한민족 "
[ 11 ]=>
string( 6 ) " 거주 "
[ 12 ]=>
string( 12 ) " 지역에서 "
[ 13 ]=>
string( 9 ) " 쓰이는 "
[ 14 ]=>
string( 9 ) " 언어로 "
}
array ( 21 ) {
[ 0 ]=>
string( 6 ) "日本"
[ 1 ]=>
string( 3 ) "語"
[ 2 ]=>
string( 3 ) "は"
[ 3 ]=>
string( 3 ) "主"
[ 4 ]=>
string( 3 ) "に"
[ 5 ]=>
string( 6 ) "日本"
[ 6 ]=>
string( 6 ) "国内"
[ 7 ]=>
string( 3 ) "や"
[ 8 ]=>
string( 6 ) "日本"
[ 9 ]=>
string( 3 ) "人"
[ 10 ]=>
string( 6 ) "同士"
[ 11 ]=>
string( 3 ) "の"
[ 12 ]=>
string( 3 ) "間"
[ 13 ]=>
string( 3 ) "で"
[ 14 ]=>
string( 3 ) "使"
[ 15 ]=>
string( 3 ) "わ"
[ 16 ]=>
string( 6 ) "れて"
[ 17 ]=>
string( 6 ) "いる"
[ 18 ]=>
string( 6 ) "言語"
[ 19 ]=>
string( 3 ) "で"
[ 20 ]=>
string( 6 ) "ある"
}
array ( 17 ) {
[ 0 ]=>
string( 9 ) "日本語"
[ 1 ]=>
string( 3 ) "は"
[ 2 ]=>
string( 6 ) "主に"
[ 3 ]=>
string( 9 ) "日本国"
[ 4 ]=>
string( 3 ) "内"
[ 5 ]=>
string( 3 ) "や"
[ 6 ]=>
string( 9 ) "日本人"
[ 7 ]=>
string( 6 ) "同士"
[ 8 ]=>
string( 3 ) "の"
[ 9 ]=>
string( 3 ) "間"
[ 10 ]=>
string( 3 ) "で"
[ 11 ]=>
string( 3 ) "使"
[ 12 ]=>
string( 3 ) "わ"
[ 13 ]=>
string( 6 ) "れて"
[ 14 ]=>
string( 6 ) "いる"
[ 15 ]=>
string( 6 ) "言語"
[ 16 ]=>
string( 9 ) "である"
}Code example (Tutorial)
ini_set ( ' memory_limit ' , ' 1024M ' );
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/MultiArray.php " ;
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Jieba.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Finalseg.php " ;
use Fukuball Jieba Jieba ;
use Fukuball Jieba Finalseg ;
Jieba:: init ( array ( ' mode ' => ' test ' , ' dict ' => ' big ' ));
Finalseg:: init ();
$ seg_list = Jieba:: tokenize ( "永和服装饰品有限公司" );
var_dump ( $ seg_list );Output:
array ( 4 ) {
[ 0 ] =>
array ( 3 ) {
' word ' =>
string( 6 ) "永和"
' start ' =>
int( 0 )
' end ' =>
int( 2 )
}
[ 1 ] =>
array ( 3 ) {
' word ' =>
string( 6 ) "服装"
' start ' =>
int( 2 )
' end ' =>
int( 4 )
}
[ 2 ] =>
array ( 3 ) {
' word ' =>
string( 6 ) "饰品"
' start ' =>
int( 4 )
' end ' =>
int( 6 )
}
[ 3 ] =>
array ( 3 ) {
' word ' =>
string( 12 ) "有限公司"
' start ' =>
int( 6 )
' end ' =>
int( 10 )
}
}Dictionary with smaller occupancy https://github.com/fukuball/jieba-php/blob/master/src/dict/dict.small.txt
Dictionary that supports traditional Chinese verb verb https://github.com/fukuball/jieba-php/blob/master/src/dict/dict.big.txt
cut method accepts two parameters: 1) first parameter is the string to segmentation 2) the second parameter cut_all to control segmentation mode.cutForSearch accpets only on parameter: the string that requires segmentation, and it will cut the sentence into short wordscut and cutForSearch return an segmented array.Example (Tutorial)
ini_set ( ' memory_limit ' , ' 1024M ' );
require_once " /path/to/your/vendor/multi-array/MultiArray.php " ;
require_once " /path/to/your/vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once " /path/to/your/class/Jieba.php " ;
require_once " /path/to/your/class/Finalseg.php " ;
use Fukuball Jieba Jieba ;
use Fukuball Jieba Finalseg ;
Jieba:: init ();
Finalseg:: init ();
$ seg_list = Jieba:: cut ( "怜香惜玉也得要看对象啊! " );
var_dump ( $ seg_list );
seg_list = jieba. cut ( "我来到北京清华大学" , true )
var_dump ( $ seg_list ); #全模式
seg_list = jieba. cut ( "我来到北京清华大学" , false )
var_dump ( $ seg_list ); #默認精確模式
seg_list = jieba. cut ( "他来到了网易杭研大厦" )
var_dump ( $ seg_list );
seg_list = jieba. cut_for_search ( "小明硕士毕业于中国科学院计算所,后在日本京都大学深造" ) #搜索引擎模式
var_dump ( $ seg_list );Output:
array ( 7 ) {
[ 0 ]=>
string( 12 ) "怜香惜玉"
[ 1 ]=>
string( 3 ) "也"
[ 2 ]=>
string( 3 ) "得"
[ 3 ]=>
string( 3 ) "要"
[ 4 ]=>
string( 3 ) "看"
[ 5 ]=>
string( 6 ) "对象"
[ 6 ]=>
string( 3 ) "啊"
}
Full Mode:
array ( 15 ) {
[ 0 ]=>
string( 3 ) "我"
[ 1 ]=>
string( 3 ) "来"
[ 2 ]=>
string( 6 ) "来到"
[ 3 ]=>
string( 3 ) "到"
[ 4 ]=>
string( 3 ) "北"
[ 5 ]=>
string( 6 ) "北京"
[ 6 ]=>
string( 3 ) "京"
[ 7 ]=>
string( 3 ) "清"
[ 8 ]=>
string( 6 ) "清华"
[ 9 ]=>
string( 12 ) "清华大学"
[ 10 ]=>
string( 3 ) "华"
[ 11 ]=>
string( 6 ) "华大"
[ 12 ]=>
string( 3 ) "大"
[ 13 ]=>
string( 6 ) "大学"
[ 14 ]=>
string( 3 ) "学"
}
Default Mode:
array ( 4 ) {
[ 0 ]=>
string( 3 ) "我"
[ 1 ]=>
string( 6 ) "来到"
[ 2 ]=>
string( 6 ) "北京"
[ 3 ]=>
string( 12 ) "清华大学"
}
array ( 6 ) {
[ 0 ]=>
string( 3 ) "他"
[ 1 ]=>
string( 6 ) "来到"
[ 2 ]=>
string( 3 ) "了"
[ 3 ]=>
string( 6 ) "网易"
[ 4 ]=>
string( 6 ) "杭研"
[ 5 ]=>
string( 6 ) "大厦"
}
(此處,“杭研“並沒有在詞典中,但是也被 Viterbi 算法識別出來了)
Search Engine Mode:
array ( 18 ) {
[ 0 ]=>
string( 6 ) "小明"
[ 1 ]=>
string( 6 ) "硕士"
[ 2 ]=>
string( 6 ) "毕业"
[ 3 ]=>
string( 3 ) "于"
[ 4 ]=>
string( 6 ) "中国"
[ 5 ]=>
string( 6 ) "科学"
[ 6 ]=>
string( 6 ) "学院"
[ 7 ]=>
string( 9 ) "科学院"
[ 8 ]=>
string( 15 ) "中国科学院"
[ 9 ]=>
string( 6 ) "计算"
[ 10 ]=>
string( 9 ) "计算所"
[ 11 ]=>
string( 3 ) "后"
[ 12 ]=>
string( 3 ) "在"
[ 13 ]=>
string( 6 ) "日本"
[ 14 ]=>
string( 6 ) "京都"
[ 15 ]=>
string( 6 ) "大学"
[ 16 ]=>
string( 18 ) "日本京都大学"
[ 17 ]=>
string( 6 ) "深造"
}Developers can specify their own custom dictionary to include in the jieba thesaurus. jieba has the ability to identify new words, but adding your own new words can ensure a higher rate of correct segmentation.
Usage: Jieba::loadUserDict(file_name) # file_name is a custom dictionary path.
The dictionary format is the same as that of dict.txt : one word per line; each line is divided into two parts, the first is the word itself, the other is the word frequency, separated by a space.
Example:
Cloud Computing 5 Li Xiaofu 2 Innovation Office 3
Previously: Li Xiaofu/Yes/Innovation/Office/Director/Yes/Yes/Cloud/Computing/Evidence/After loading the custom thesaurus: Li Xiaofu/Yes/Innovation/Office/Director/Yes/Yes/Yes/Cloud/Evidence/
Example (keyword extraction)
ini_set ( ' memory_limit ' , ' 600M ' );
require_once " /path/to/your/vendor/multi-array/MultiArray.php " ;
require_once " /path/to/your/vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once " /path/to/your/class/Jieba.php " ;
require_once " /path/to/your/class/Finalseg.php " ;
require_once " /path/to/your/class/JiebaAnalyse.php " ;
use Fukuball Jieba Jieba ;
use Fukuball Jieba Finalseg ;
use Fukuball Jieba JiebaAnalyse ;
Jieba:: init ( array ( ' mode ' => ' test ' , ' dict ' => ' small ' ));
Finalseg:: init ();
JiebaAnalyse:: init ();
$ top_k = 10 ;
$ content = file_get_contents ( " /path/to/your/dict/lyric.txt " , " r " );
$ tags = JiebaAnalyse:: extractTags ( $ content , $ top_k );
var_dump ( $ tags );Output:
array ( 10 ) {
[ "是否" ]=>
float( 1.2196321889395 )
[ "一般" ]=>
float( 1.0032459890209 )
[ "肌迫" ]=>
float( 0.64654314660465 )
[ "怯懦" ]=>
float( 0.44762844339349 )
[ "藉口" ]=>
float( 0.32327157330233 )
[ "逼不得已" ]=>
float( 0.32327157330233 )
[ "不安全感" ]=>
float( 0.26548304656279 )
[ "同感" ]=>
float( 0.23929673812326 )
[ "有把握" ]=>
float( 0.21043366018744 )
[ "空洞" ]=>
float( 0.20598261709442 )
}Example (word tagging)
ini_set ( ' memory_limit ' , ' 600M ' );
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/MultiArray.php " ;
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Jieba.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Finalseg.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Posseg.php " ;
use Fukuball Jieba Jieba ;
use Fukuball Jieba Finalseg ;
use Fukuball Jieba Posseg ;
Jieba:: init ();
Finalseg:: init ();
Posseg:: init ();
$ seg_list = Posseg:: cut ( "这是一个伸手不见五指的黑夜。我叫孙悟空,我爱北京,我爱Python和C++。 " );
var_dump ( $ seg_list );Output:
array ( 21 ) {
[ 0 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "这"
[ " tag " ]=>
string( 1 ) " r "
}
[ 1 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "是"
[ " tag " ]=>
string( 1 ) " v "
}
[ 2 ]=>
array ( 2 ) {
[ " word " ]=>
string( 6 ) "一个"
[ " tag " ]=>
string( 1 ) " m "
}
[ 3 ]=>
array ( 2 ) {
[ " word " ]=>
string( 18 ) "伸手不见五指"
[ " tag " ]=>
string( 1 ) " i "
}
[ 4 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "的"
[ " tag " ]=>
string( 2 ) " uj "
}
[ 5 ]=>
array ( 2 ) {
[ " word " ]=>
string( 6 ) "黑夜"
[ " tag " ]=>
string( 1 ) " n "
}
[ 6 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) " 。 "
[ " tag " ]=>
string( 1 ) " w "
}
[ 7 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "我"
[ " tag " ]=>
string( 1 ) " r "
}
[ 8 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "叫"
[ " tag " ]=>
string( 1 ) " v "
}
[ 9 ]=>
array ( 2 ) {
[ " word " ]=>
string( 9 ) "孙悟空"
[ " tag " ]=>
string( 2 ) " nr "
}
[ 10 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) " , "
[ " tag " ]=>
string( 1 ) " w "
}
[ 11 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "我"
[ " tag " ]=>
string( 1 ) " r "
}
[ 12 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "爱"
[ " tag " ]=>
string( 1 ) " v "
}
[ 13 ]=>
array ( 2 ) {
[ " word " ]=>
string( 6 ) "北京"
[ " tag " ]=>
string( 2 ) " ns "
}
[ 14 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) " , "
[ " tag " ]=>
string( 1 ) " w "
}
[ 15 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "我"
[ " tag " ]=>
string( 1 ) " r "
}
[ 16 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "爱"
[ " tag " ]=>
string( 1 ) " v "
}
[ 17 ]=>
array ( 2 ) {
[ " word " ]=>
string( 6 ) " Python "
[ " tag " ]=>
string( 3 ) " eng "
}
[ 18 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) "和"
[ " tag " ]=>
string( 1 ) " c "
}
[ 19 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) " C++ "
[ " tag " ]=>
string( 3 ) " eng "
}
[ 20 ]=>
array ( 2 ) {
[ " word " ]=>
string( 3 ) " 。 "
[ " tag " ]=>
string( 1 ) " w "
}
}Example (Tutorial)
ini_set ( ' memory_limit ' , ' 1024M ' );
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/MultiArray.php " ;
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Jieba.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Finalseg.php " ;
use Fukuball Jieba Jieba ;
use Fukuball Jieba Finalseg ;
Jieba:: init ( array ( ' mode ' => ' default ' , ' dict ' => ' big ' ));
Finalseg:: init ();
$ seg_list = Jieba:: cut ( "怜香惜玉也得要看对象啊! " );
var_dump ( $ seg_list );
$ seg_list = Jieba:: cut ( "憐香惜玉也得要看對象啊! " );
var_dump ( $ seg_list );Output:
array ( 7 ) {
[ 0 ]=>
string( 12 ) "怜香惜玉"
[ 1 ]=>
string( 3 ) "也"
[ 2 ]=>
string( 3 ) "得"
[ 3 ]=>
string( 3 ) "要"
[ 4 ]=>
string( 3 ) "看"
[ 5 ]=>
string( 6 ) "对象"
[ 6 ]=>
string( 3 ) "啊"
}
array ( 7 ) {
[ 0 ]=>
string( 12 ) "憐香惜玉"
[ 1 ]=>
string( 3 ) "也"
[ 2 ]=>
string( 3 ) "得"
[ 3 ]=>
string( 3 ) "要"
[ 4 ]=>
string( 3 ) "看"
[ 5 ]=>
string( 6 ) "對象"
[ 6 ]=>
string( 3 ) "啊"
}Example (Tutorial)
ini_set ( ' memory_limit ' , ' 1024M ' );
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/MultiArray.php " ;
require_once dirname ( dirname ( __FILE__ )). " /vendor/multi-array/Factory/MultiArrayFactory.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Jieba.php " ;
require_once dirname ( dirname ( __FILE__ )). " /class/Finalseg.php " ;
use Fukuball Jieba Jieba ;
use Fukuball Jieba Finalseg ;
Jieba:: init ( array ( ' cjk ' => ' all ' ));
Finalseg:: init ();
$ seg_list = Jieba:: cut ( " 한국어 또는 조선말은 제주특별자치도를 제외한 한반도 및 그 부속 도서와 한민족 거주 지역에서 쓰이는 언어로 " );
var_dump ( $ seg_list );
$ seg_list = Jieba:: cut ( "日本語は、主に日本国内や日本人同士の間で使われている言語である。 " );
var_dump ( $ seg_list );
// Loading custom Japanese dictionary can do a simple word segmentation
Jieba:: loadUserDict ( " /path/to/your/japanese/dict.txt " );
$ seg_list = Jieba:: cut ( "日本語は、主に日本国内や日本人同士の間で使われている言語である。 " );
var_dump ( $ seg_list );Output:
array ( 15 ) {
[ 0 ]=>
string( 9 ) " 한국어 "
[ 1 ]=>
string( 6 ) " 또는 "
[ 2 ]=>
string( 12 ) " 조선말은 "
[ 3 ]=>
string( 24 ) " 제주특별자치도를 "
[ 4 ]=>
string( 9 ) " 제외한 "
[ 5 ]=>
string( 9 ) " 한반도 "
[ 6 ]=>
string( 3 ) " 및 "
[ 7 ]=>
string( 3 ) " 그 "
[ 8 ]=>
string( 6 ) " 부속 "
[ 9 ]=>
string( 9 ) " 도서와 "
[ 10 ]=>
string( 9 ) " 한민족 "
[ 11 ]=>
string( 6 ) " 거주 "
[ 12 ]=>
string( 12 ) " 지역에서 "
[ 13 ]=>
string( 9 ) " 쓰이는 "
[ 14 ]=>
string( 9 ) " 언어로 "
}
array ( 21 ) {
[ 0 ]=>
string( 6 ) "日本"
[ 1 ]=>
string( 3 ) "語"
[ 2 ]=>
string( 3 ) "は"
[ 3 ]=>
string( 3 ) "主"
[ 4 ]=>
string( 3 ) "に"
[ 5 ]=>
string( 6 ) "日本"
[ 6 ]=>
string( 6 ) "国内"
[ 7 ]=>
string( 3 ) "や"
[ 8 ]=>
string( 6 ) "日本"
[ 9 ]=>
string( 3 ) "人"
[ 10 ]=>
string( 6 ) "同士"
[ 11 ]=>
string( 3 ) "の"
[ 12 ]=>
string( 3 ) "間"
[ 13 ]=>
string( 3 ) "で"
[ 14 ]=>
string( 3 ) "使"
[ 15 ]=>
string( 3 ) "わ"
[ 16 ]=>
string( 6 ) "れて"
[ 17 ]=>
string( 6 ) "いる"
[ 18 ]=>
string( 6 ) "言語"
[ 19 ]=>
string( 3 ) "で"
[ 20 ]=>
string( 6 ) "ある"
}
array ( 17 ) {
[ 0 ]=>
string( 9 ) "日本語"
[ 1 ]=>
string( 3 ) "は"
[ 2 ]=>
string( 6 ) "主に"
[ 3 ]=>
string( 9 ) "日本国"
[ 4 ]=>
string( 3 ) "内"
[ 5 ]=>
string( 3 ) "や"
[ 6 ]=>
string( 9 ) "日本人"
[ 7 ]=>
string( 6 ) "同士"
[ 8 ]=>
string( 3 ) "の"
[ 9 ]=>
string( 3 ) "間"
[ 10 ]=>
string( 3 ) "で"
[ 11 ]=>
string( 3 ) "使"
[ 12 ]=>
string( 3 ) "わ"
[ 13 ]=>
string( 6 ) "れて"
[ 14 ]=>
string( 6 ) "いる"
[ 15 ]=>
string( 6 ) "言語"
[ 16 ]=>
string( 9 ) "である"
} a 形容词 (取英语形容词 adjective 的第 1 个字母。)
ad 副形词 (直接作状语的形容词,形容词代码 a 和副词代码 d 并在一起。)
ag 形容词性语素 (形容词性语素,形容词代码为 a,语素代码 g 前面置以 a。)
an 名形词 (具有名词功能的形容词,形容词代码 a 和名词代码 n 并在一起。)
b 区别词 (取汉字「别」的声母。)
c 连词 (取英语连词 conjunction 的第 1 个字母。)
d 副词 (取 adverb 的第 2 个字母,因其第 1 个字母已用于形容词。)
df 副词*
dg 副语素 (副词性语素,副词代码为 d,语素代码 g 前面置以 d。)
e 叹词 (取英语叹词 exclamation 的第 1 个字母。)
eng 外语
f 方位词 (取汉字「方」的声母。)
g 语素 (绝大多数语素都能作为合成词的「词根」,取汉字「根」的声母。)
h 前接成分 (取英语 head 的第 1 个字母。)
i 成语 (取英语成语 idiom 的第 1 个字母。)
j 简称略语 (取汉字「简」的声母。)
k 后接成分
l 习用语 (习用语尚未成为成语,有点「临时性」,取「临」的声母。)
m 数词 (取英语 numeral 的第 3 个字母,n,u 已有他用。)
mg 数语素
mq 数词*
n 名词 (取英语名词 noun 的第 1 个字母。)
ng 名语素 (名词性语素,名词代码为 n,语素代码 g 前面置以 n。)
nr 人名 (名词代码n和「人(ren)」的声母并在一起。)
nrfg 名词*
nrt 名词*
ns 地名 (名词代码 n 和处所词代码 s 并在一起。)
nt 机构团体 (「团」的声母为 t,名词代码 n 和 t 并在一起。)
nz 其他专名 (「专」的声母的第 1 个字母为 z,名词代码 n 和 z 并在一起。)
o 拟声词 (取英语拟声词 onomatopoeia 的第 1 个字母。)
p 介词 (取英语介词 prepositional 的第 1 个字母。)
q 量词 (取英语 quantity 的第 1 个字母。)
r 代词 (取英语代词 pronoun的 第 2 个字母,因 p 已用于介词。)
rg 代词语素
rr 代词*
rz 代词*
s 处所词 (取英语 space 的第 1 个字母。)
t 时间词 (取英语 time 的第 1 个字母。)
tg 时语素 (时间词性语素,时间词代码为 t,在语素的代码 g 前面置以 t。)
u 助词 (取英语助词 auxiliary 的第 2 个字母,因 a 已用于形容词。)
ud 助词*
ug 助词*
uj 助词*
ul 助词*
uv 助词*
uz 助词*
v 动词 (取英语动词 verb 的第一个字母。)
vd 副动词 (直接作状语的动词,动词和副词的代码并在一起。)
vg 动语素
vi 动词*
vn 名动词 (指具有名词功能的动词,动词和名词的代码并在一起。)
vq 动词*
w 标点符号
x 非语素字 (非语素字只是一个符号,字母 x 通常用于代表未知数、符号。)
y 语气词 (取汉字「语」的声母。)
z 状态词 (取汉字「状」的声母的前一个字母。)
zg 状态词*
If you find fuku-ml useful, please consider a donation. Thank you!
The MIT License (MIT)
Copyright (c) 2015 fukuball
Permission is hereby granted, free of charge, to any person obtaining a copy of this software and associated documentation files (the "Software"), to deal in the Software without restrictions, including without limitation the rights to use, copy, modify, merge, publish, distribution, sublicense, and/or sell copies of the Software, and to permit persons to whom the Software is furnished to do so, subject to the following conditions:
The above copyright notice and this permission notice shall be included in all copies or substantial portions of the Software.
THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE.