From 5cedae98fa31b86aa5a66617cacd612dd441d5a4 Mon Sep 17 00:00:00 2001 From: James Rodewig Date: Wed, 1 Jul 2020 18:41:07 -0400 Subject: [PATCH] [DOCS] Fix headings for simple analyzer docs --- .../analyzers/simple-analyzer.asciidoc | 52 ++++++++----------- 1 file changed, 23 insertions(+), 29 deletions(-) diff --git a/docs/reference/analysis/analyzers/simple-analyzer.asciidoc b/docs/reference/analysis/analyzers/simple-analyzer.asciidoc index ee87aaa0260d8..6de529113a3bc 100644 --- a/docs/reference/analysis/analyzers/simple-analyzer.asciidoc +++ b/docs/reference/analysis/analyzers/simple-analyzer.asciidoc @@ -4,25 +4,25 @@ Simple ++++ -The `simple` analyzer breaks text into terms at any non-letter character, such +The `simple` analyzer breaks text into tokens at any non-letter character, such as numbers, spaces, hyphens and apostrophes, discards non-letter characters, and changes uppercase to lowercase. +[[analysis-simple-analyzer-ex]] ==== Example [source,console] ---------------------------- +---- POST _analyze { "analyzer": "simple", "text": "The 2 QUICK Brown-Foxes jumped over the lazy dog's bone." } ---------------------------- - -///////////////////// +---- +//// [source,console-result] ----------------------------- +---- { "tokens": [ { @@ -104,53 +104,47 @@ POST _analyze } ] } ----------------------------- - -///////////////////// - +---- +//// The `simple` analyzer parses the sentence and produces the following -terms: +tokens: [source,text] ---------------------------- +---- [ the, quick, brown, foxes, jumped, over, the, lazy, dog, s, bone ] ---------------------------- - -==== Configure parameters - -The `simple` analyzer does not contain configurable parameters. +---- -==== Customize +[[analysis-simple-analyzer-definition]] +==== Definition The `simple` analyzer is defined by one tokenizer: Tokenizer:: -* <> +* <> + +[[analysis-simple-analyzer-customize]] +==== Customize To customize the `simple` analyzer, duplicate it to create the basis for -a `custom` analyzer. The new analyzer can be modified as required, usually by +a custom analyzer. This custom analyzer can be modified as required, usually by adding token filters. -===== Example - [source,console] ----------------------------------------------------- -PUT /simple_example +---- +PUT /my_index { "settings": { "analysis": { "analyzer": { - "rebuilt_simple": { + "my_custom_simple_analyzer": { "tokenizer": "lowercase", - "filter": [ <1> + "filter": [ <1> ] } } } } } ----------------------------------------------------- -// TEST[s/\n$/\nstartyaml\n - compare_analyzers: {index: simple_example, first: simple, second: rebuilt_simple}\nendyaml\n/] - +---- <1> Add token filters here.