WARNING: Version 6.1 of Elasticsearch has passed its EOL date.
This documentation is no longer being maintained and may be removed. If you are running this version, we strongly advise you to upgrade. For the latest information, see the current release documentation.
Thai Tokenizer
editThai Tokenizer
editThe thai
tokenizer segments Thai text into words, using the Thai
segmentation algorithm included with Java. Text in other languages in general
will be treated the same as the
standard
tokenizer.
This tokenizer may not be supported by all JREs. It is known to work with Sun/Oracle and OpenJDK. If your application needs to be fully portable, consider using the ICU Tokenizer instead.
Example output
editPOST _analyze { "tokenizer": "thai", "text": "การที่ได้ต้องแสดงว่างานดี" }
The above sentence would produce the following terms:
[ การ, ที่, ได้, ต้อง, แสดง, ว่า, งาน, ดี ]
Configuration
editThe thai
tokenizer is not configurable.