![]() Segments are created from this amount of data using tokenization. ![]() These segments make it possible to search for terms (mostly single words). The simplest way for tokenization to work is with the white space strategy: a term ends when a space occurs. However, this does not apply if fixed terms consist of several words, such as “Christmas Eve.” Additional dictionaries are used for this, which can also be implemented in the Lucene code. Lucene also performs a normalization when analyzing the data of which tokenization is a part. This means that the terms are written in a standardized form e.g. all capital letters are written in lower case. As a user, you probably want to get the most relevant or latest results first – the search engine’s algorithms enable this.įor users to find anything at all, they must enter a search term in a text line. The term or terms are called query in the Lucene context. The word “request” indicates that the input must not only consist of one or more words, but can also contain modifiers such as AND, OR, or + and – as well as placeholders. ![]() The QueryParser – a class within the program library – translated the input into a specific search request for the search engine. Developers can also make settings for the QueryParser.
0 Comments
Leave a Reply. |