Use external program to split pages into words
An external tokenizer inserts extra spaces to mark words in the input text. The text is sent through STDIN and STDOUT file handles. A good choice for Chinese and Japanese is MeCab. http://sourceforge.net/projects/mecab/ With the command line 'mecab -O wakati'
Showing
- conf/dokuwiki.php 2 additions, 0 deletionsconf/dokuwiki.php
- inc/indexer.php 16 additions, 16 deletionsinc/indexer.php
- lib/plugins/config/lang/en/lang.php 2 additions, 0 deletionslib/plugins/config/lang/en/lang.php
- lib/plugins/config/settings/config.metadata.php 2 additions, 0 deletionslib/plugins/config/settings/config.metadata.php
Please register or sign in to comment