Book Image

Mastering Elasticsearch 5.x - Third Edition

By : Bharvi Dixit
Book Image

Mastering Elasticsearch 5.x - Third Edition

By: Bharvi Dixit

Overview of this book

Elasticsearch is a modern, fast, distributed, scalable, fault tolerant, and open source search and analytics engine. Elasticsearch leverages the capabilities of Apache Lucene, and provides a new level of control over how you can index and search even huge sets of data. This book will give you a brief recap of the basics and also introduce you to the new features of Elasticsearch 5. We will guide you through the intermediate and advanced functionalities of Elasticsearch, such as querying, indexing, searching, and modifying data. We’ll also explore advanced concepts, including aggregation, index control, sharding, replication, and clustering. We’ll show you the modules of monitoring and administration available in Elasticsearch, and will also cover backup and recovery. You will get an understanding of how you can scale your Elasticsearch cluster to contextualize it and improve its performance. We’ll also show you how you can create your own analysis plugin in Elasticsearch. By the end of the book, you will have all the knowledge necessary to master Elasticsearch and put it to efficient use.
Table of Contents (13 chapters)

Correcting user spelling mistakes

One of the simplest ways to improve the user search experience is to correct spelling mistakes either automatically, or by just showing the correct query phrase and allowing the user to use it. For example, this is what Google shows us when we type in elasticsaerch instead of Elasticsearch:

Correcting user spelling mistakes

Starting from version 0.90.0 Beta1, Elasticsearch allows us to use the suggest API to correct user spelling mistakes. With the newer versions of Elasticsearch, the API was changed, bringing new features and becoming more and more powerful. In this section, we will try to bring you a comprehensive guide on how to use the suggest API provided by Elasticsearch, both in simple use cases and in ones that require more configuration.

Testing data

For the purposes of this section, we need a bit more data than a few documents. In order to get the data, we have decided to index some news documents from Wikipedia dump.

Let's follow the given steps for downloading and importing...