Skip to content

Commit 78fd33e

Browse files
authored
[DOCS] Reformat Plugin snippets to use two-space indents (#59895) (#59918)
1 parent dca8fa0 commit 78fd33e

File tree

5 files changed

+476
-476
lines changed

5 files changed

+476
-476
lines changed

docs/plugins/analysis-icu.asciidoc

Lines changed: 16 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -153,24 +153,24 @@ Then create an analyzer to use this rule file as follows:
153153
--------------------------------------------------
154154
PUT icu_sample
155155
{
156-
"settings": {
157-
"index":{
158-
"analysis":{
159-
"tokenizer" : {
160-
"icu_user_file" : {
161-
"type" : "icu_tokenizer",
162-
"rule_files" : "Latn:KeywordTokenizer.rbbi"
163-
}
164-
},
165-
"analyzer" : {
166-
"my_analyzer" : {
167-
"type" : "custom",
168-
"tokenizer" : "icu_user_file"
169-
}
170-
}
171-
}
156+
"settings": {
157+
"index": {
158+
"analysis": {
159+
"tokenizer": {
160+
"icu_user_file": {
161+
"type": "icu_tokenizer",
162+
"rule_files": "Latn:KeywordTokenizer.rbbi"
163+
}
164+
},
165+
"analyzer": {
166+
"my_analyzer": {
167+
"type": "custom",
168+
"tokenizer": "icu_user_file"
169+
}
172170
}
171+
}
173172
}
173+
}
174174
}
175175
176176
GET icu_sample/_analyze

docs/plugins/analysis-kuromoji.asciidoc

Lines changed: 24 additions & 24 deletions
Original file line numberDiff line numberDiff line change
@@ -348,32 +348,32 @@ katakana reading form:
348348
--------------------------------------------------
349349
PUT kuromoji_sample
350350
{
351-
"settings": {
352-
"index":{
353-
"analysis":{
354-
"analyzer" : {
355-
"romaji_analyzer" : {
356-
"tokenizer" : "kuromoji_tokenizer",
357-
"filter" : ["romaji_readingform"]
358-
},
359-
"katakana_analyzer" : {
360-
"tokenizer" : "kuromoji_tokenizer",
361-
"filter" : ["katakana_readingform"]
362-
}
363-
},
364-
"filter" : {
365-
"romaji_readingform" : {
366-
"type" : "kuromoji_readingform",
367-
"use_romaji" : true
368-
},
369-
"katakana_readingform" : {
370-
"type" : "kuromoji_readingform",
371-
"use_romaji" : false
372-
}
373-
}
374-
}
351+
"settings": {
352+
"index": {
353+
"analysis": {
354+
"analyzer": {
355+
"romaji_analyzer": {
356+
"tokenizer": "kuromoji_tokenizer",
357+
"filter": [ "romaji_readingform" ]
358+
},
359+
"katakana_analyzer": {
360+
"tokenizer": "kuromoji_tokenizer",
361+
"filter": [ "katakana_readingform" ]
362+
}
363+
},
364+
"filter": {
365+
"romaji_readingform": {
366+
"type": "kuromoji_readingform",
367+
"use_romaji": true
368+
},
369+
"katakana_readingform": {
370+
"type": "kuromoji_readingform",
371+
"use_romaji": false
372+
}
375373
}
374+
}
376375
}
376+
}
377377
}
378378
379379
GET kuromoji_sample/_analyze

docs/plugins/analysis-nori.asciidoc

Lines changed: 91 additions & 91 deletions
Original file line numberDiff line numberDiff line change
@@ -214,88 +214,88 @@ Which responds with:
214214
[source,console-result]
215215
--------------------------------------------------
216216
{
217-
"detail": {
218-
"custom_analyzer": true,
219-
"charfilters": [],
220-
"tokenizer": {
221-
"name": "nori_tokenizer",
222-
"tokens": [
223-
{
224-
"token": "뿌리",
225-
"start_offset": 0,
226-
"end_offset": 2,
227-
"type": "word",
228-
"position": 0,
229-
"leftPOS": "NNG(General Noun)",
230-
"morphemes": null,
231-
"posType": "MORPHEME",
232-
"reading": null,
233-
"rightPOS": "NNG(General Noun)"
234-
},
235-
{
236-
"token": "가",
237-
"start_offset": 2,
238-
"end_offset": 3,
239-
"type": "word",
240-
"position": 1,
241-
"leftPOS": "J(Ending Particle)",
242-
"morphemes": null,
243-
"posType": "MORPHEME",
244-
"reading": null,
245-
"rightPOS": "J(Ending Particle)"
246-
},
247-
{
248-
"token": "깊",
249-
"start_offset": 4,
250-
"end_offset": 5,
251-
"type": "word",
252-
"position": 2,
253-
"leftPOS": "VA(Adjective)",
254-
"morphemes": null,
255-
"posType": "MORPHEME",
256-
"reading": null,
257-
"rightPOS": "VA(Adjective)"
258-
},
259-
{
260-
"token": "은",
261-
"start_offset": 5,
262-
"end_offset": 6,
263-
"type": "word",
264-
"position": 3,
265-
"leftPOS": "E(Verbal endings)",
266-
"morphemes": null,
267-
"posType": "MORPHEME",
268-
"reading": null,
269-
"rightPOS": "E(Verbal endings)"
270-
},
271-
{
272-
"token": "나무",
273-
"start_offset": 7,
274-
"end_offset": 9,
275-
"type": "word",
276-
"position": 4,
277-
"leftPOS": "NNG(General Noun)",
278-
"morphemes": null,
279-
"posType": "MORPHEME",
280-
"reading": null,
281-
"rightPOS": "NNG(General Noun)"
282-
},
283-
{
284-
"token": "는",
285-
"start_offset": 9,
286-
"end_offset": 10,
287-
"type": "word",
288-
"position": 5,
289-
"leftPOS": "J(Ending Particle)",
290-
"morphemes": null,
291-
"posType": "MORPHEME",
292-
"reading": null,
293-
"rightPOS": "J(Ending Particle)"
294-
}
295-
]
217+
"detail": {
218+
"custom_analyzer": true,
219+
"charfilters": [],
220+
"tokenizer": {
221+
"name": "nori_tokenizer",
222+
"tokens": [
223+
{
224+
"token": "뿌리",
225+
"start_offset": 0,
226+
"end_offset": 2,
227+
"type": "word",
228+
"position": 0,
229+
"leftPOS": "NNG(General Noun)",
230+
"morphemes": null,
231+
"posType": "MORPHEME",
232+
"reading": null,
233+
"rightPOS": "NNG(General Noun)"
296234
},
297-
"tokenfilters": []
298-
}
235+
{
236+
"token": "가",
237+
"start_offset": 2,
238+
"end_offset": 3,
239+
"type": "word",
240+
"position": 1,
241+
"leftPOS": "J(Ending Particle)",
242+
"morphemes": null,
243+
"posType": "MORPHEME",
244+
"reading": null,
245+
"rightPOS": "J(Ending Particle)"
246+
},
247+
{
248+
"token": "깊",
249+
"start_offset": 4,
250+
"end_offset": 5,
251+
"type": "word",
252+
"position": 2,
253+
"leftPOS": "VA(Adjective)",
254+
"morphemes": null,
255+
"posType": "MORPHEME",
256+
"reading": null,
257+
"rightPOS": "VA(Adjective)"
258+
},
259+
{
260+
"token": "은",
261+
"start_offset": 5,
262+
"end_offset": 6,
263+
"type": "word",
264+
"position": 3,
265+
"leftPOS": "E(Verbal endings)",
266+
"morphemes": null,
267+
"posType": "MORPHEME",
268+
"reading": null,
269+
"rightPOS": "E(Verbal endings)"
270+
},
271+
{
272+
"token": "나무",
273+
"start_offset": 7,
274+
"end_offset": 9,
275+
"type": "word",
276+
"position": 4,
277+
"leftPOS": "NNG(General Noun)",
278+
"morphemes": null,
279+
"posType": "MORPHEME",
280+
"reading": null,
281+
"rightPOS": "NNG(General Noun)"
282+
},
283+
{
284+
"token": "는",
285+
"start_offset": 9,
286+
"end_offset": 10,
287+
"type": "word",
288+
"position": 5,
289+
"leftPOS": "J(Ending Particle)",
290+
"morphemes": null,
291+
"posType": "MORPHEME",
292+
"reading": null,
293+
"rightPOS": "J(Ending Particle)"
294+
}
295+
]
296+
},
297+
"tokenfilters": []
298+
}
299299
}
300300
--------------------------------------------------
301301

@@ -400,18 +400,18 @@ The `nori_readingform` token filter rewrites tokens written in Hanja to their Ha
400400
--------------------------------------------------
401401
PUT nori_sample
402402
{
403-
"settings": {
404-
"index":{
405-
"analysis":{
406-
"analyzer" : {
407-
"my_analyzer" : {
408-
"tokenizer" : "nori_tokenizer",
409-
"filter" : ["nori_readingform"]
410-
}
411-
}
412-
}
403+
"settings": {
404+
"index": {
405+
"analysis": {
406+
"analyzer": {
407+
"my_analyzer": {
408+
"tokenizer": "nori_tokenizer",
409+
"filter": [ "nori_readingform" ]
410+
}
413411
}
412+
}
414413
}
414+
}
415415
}
416416
417417
GET nori_sample/_analyze

0 commit comments

Comments
 (0)