Skip to content
Closed
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,7 @@
import org.elasticsearch.Version;
import org.elasticsearch.cluster.ClusterModule;
import org.elasticsearch.cluster.metadata.IndexMetadata;
import org.elasticsearch.cluster.metadata.MappingMetadata;
import org.elasticsearch.common.compress.CompressedXContent;
import org.elasticsearch.common.settings.Settings;
import org.elasticsearch.common.xcontent.LoggingDeprecationHandler;
Expand Down Expand Up @@ -72,7 +73,7 @@ public <T> T compile(Script script, ScriptContext<T> scriptContext) {
);

try {
mapperService.merge("_doc", new CompressedXContent(mappings), MapperService.MergeReason.MAPPING_UPDATE);
mapperService.merge(new MappingMetadata(new CompressedXContent(mappings)), MapperService.MergeReason.MAPPING_UPDATE);
return mapperService;
} catch (IOException e) {
throw new UncheckedIOException(e);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -20,6 +20,7 @@
import org.elasticsearch.Version;
import org.elasticsearch.cluster.ClusterModule;
import org.elasticsearch.cluster.metadata.IndexMetadata;
import org.elasticsearch.cluster.metadata.MappingMetadata;
import org.elasticsearch.common.bytes.BytesArray;
import org.elasticsearch.common.compress.CompressedXContent;
import org.elasticsearch.common.io.stream.NamedWriteableRegistry;
Expand Down Expand Up @@ -193,7 +194,7 @@ public <T> T compile(Script script, ScriptContext<T> scriptContext) {
);

try {
mapperService.merge("_doc", new CompressedXContent(mappings), MapperService.MergeReason.MAPPING_UPDATE);
mapperService.merge(new MappingMetadata(new CompressedXContent(mappings)), MapperService.MergeReason.MAPPING_UPDATE);
return mapperService;
} catch (IOException e) {
throw new UncheckedIOException(e);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -10,6 +10,7 @@
import org.elasticsearch.Version;
import org.elasticsearch.cluster.metadata.DataStream;
import org.elasticsearch.cluster.metadata.IndexMetadata;
import org.elasticsearch.cluster.metadata.MappingMetadata;
import org.elasticsearch.cluster.metadata.Metadata;
import org.elasticsearch.common.UUIDs;
import org.elasticsearch.common.compress.CompressedXContent;
Expand Down Expand Up @@ -159,7 +160,7 @@ private List<String> findRoutingPaths(String indexName, Settings allSettings, Li
// Create MapperService just to extract keyword dimension fields:
try (var mapperService = mapperServiceFactory.apply(tmpIndexMetadata.build())) {
for (var mapping : combinedTemplateMappings) {
mapperService.merge(MapperService.SINGLE_MAPPING_NAME, mapping, MapperService.MergeReason.INDEX_TEMPLATE);
mapperService.merge(new MappingMetadata(mapping), MapperService.MergeReason.INDEX_TEMPLATE);
}

List<String> routingPaths = new ArrayList<>();
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -8,6 +8,7 @@

package org.elasticsearch.index.mapper.extras;

import org.elasticsearch.cluster.metadata.MappingMetadata;
import org.elasticsearch.common.Strings;
import org.elasticsearch.common.bytes.BytesReference;
import org.elasticsearch.common.compress.CompressedXContent;
Expand All @@ -18,6 +19,7 @@
import org.elasticsearch.index.mapper.Mapping;
import org.elasticsearch.index.mapper.SourceToParse;
import org.elasticsearch.plugins.Plugin;
import org.elasticsearch.xcontent.XContentBuilder;
import org.elasticsearch.xcontent.XContentFactory;
import org.elasticsearch.xcontent.XContentType;
import org.hamcrest.CoreMatchers;
Expand All @@ -33,20 +35,18 @@ protected Collection<? extends Plugin> getPlugins() {
}

public void testBasics() throws Exception {
String mapping = Strings.toString(
XContentFactory.jsonBuilder()
.startObject()
.startObject("type")
.startObject("properties")
.startObject("field")
.field("type", "rank_feature")
.endObject()
.endObject()
.endObject()
.endObject()
);
XContentBuilder mapping = XContentFactory.jsonBuilder()
.startObject()
.startObject("type")
.startObject("properties")
.startObject("field")
.field("type", "rank_feature")
.endObject()
.endObject()
.endObject()
.endObject();

Mapping parsedMapping = createMapperService(mapping).parseMapping("type", new CompressedXContent(mapping));
Mapping parsedMapping = createMapperService(mapping).parseMapping(toMappingMetadata(mapping));
assertEquals(mapping, parsedMapping.toCompressedXContent().toString());
assertNotNull(parsedMapping.getMetadataMapperByClass(RankFeatureMetaFieldMapper.class));
}
Expand All @@ -58,8 +58,7 @@ public void testBasics() throws Exception {
public void testDocumentParsingFailsOnMetaField() throws Exception {
String mapping = Strings.toString(XContentFactory.jsonBuilder().startObject().startObject("_doc").endObject().endObject());
DocumentMapper mapper = createMapperService(mapping).merge(
"_doc",
new CompressedXContent(mapping),
new MappingMetadata(new CompressedXContent(mapping)),
MapperService.MergeReason.MAPPING_UPDATE
);
String rfMetaField = RankFeatureMetaFieldMapper.CONTENT_TYPE;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -12,6 +12,7 @@
import org.apache.lucene.search.MatchNoDocsQuery;
import org.apache.lucene.search.Query;
import org.elasticsearch.action.admin.indices.mapping.put.PutMappingRequest;
import org.elasticsearch.cluster.metadata.MappingMetadata;
import org.elasticsearch.common.Strings;
import org.elasticsearch.common.compress.CompressedXContent;
import org.elasticsearch.index.mapper.MapperService;
Expand All @@ -34,16 +35,17 @@ public class RankFeatureQueryBuilderTests extends AbstractQueryTestCase<RankFeat
@Override
protected void initializeAdditionalMappings(MapperService mapperService) throws IOException {
mapperService.merge(
"_doc",
new CompressedXContent(
Strings.toString(
PutMappingRequest.simpleMapping(
"my_feature_field",
"type=rank_feature",
"my_negative_feature_field",
"type=rank_feature,positive_score_impact=false",
"my_feature_vector_field",
"type=rank_features"
new MappingMetadata(
new CompressedXContent(
Strings.toString(
PutMappingRequest.simpleMapping(
"my_feature_field",
"type=rank_feature",
"my_negative_feature_field",
"type=rank_feature,positive_score_impact=false",
"my_feature_vector_field",
"type=rank_features"
)
)
)
),
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,8 +23,6 @@
import org.apache.lucene.search.similarities.Similarity;
import org.elasticsearch.ElasticsearchException;
import org.elasticsearch.Version;
import org.elasticsearch.common.Strings;
import org.elasticsearch.common.compress.CompressedXContent;
import org.elasticsearch.common.settings.Settings;
import org.elasticsearch.index.mapper.IdFieldMapper;
import org.elasticsearch.index.mapper.MapperService;
Expand Down Expand Up @@ -66,7 +64,6 @@

public class HasChildQueryBuilderTests extends AbstractQueryTestCase<HasChildQueryBuilder> {

private static final String TYPE = "_doc";
private static final String PARENT_DOC = "parent";
private static final String CHILD_DOC = "child";

Expand Down Expand Up @@ -120,7 +117,7 @@ protected void initializeAdditionalMappings(MapperService mapperService) throws
.endObject()
.endObject();

mapperService.merge(TYPE, new CompressedXContent(Strings.toString(mapping)), MapperService.MergeReason.MAPPING_UPDATE);
mapperService.merge(toMappingMetadata(mapping), MapperService.MergeReason.MAPPING_UPDATE);
}

/**
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -13,8 +13,6 @@
import org.apache.lucene.search.join.ScoreMode;
import org.elasticsearch.ElasticsearchException;
import org.elasticsearch.Version;
import org.elasticsearch.common.Strings;
import org.elasticsearch.common.compress.CompressedXContent;
import org.elasticsearch.index.mapper.MapperService;
import org.elasticsearch.index.query.IdsQueryBuilder;
import org.elasticsearch.index.query.InnerHitBuilder;
Expand Down Expand Up @@ -51,7 +49,7 @@
import static org.mockito.Mockito.when;

public class HasParentQueryBuilderTests extends AbstractQueryTestCase<HasParentQueryBuilder> {
private static final String TYPE = "_doc";

private static final String PARENT_DOC = "parent";
private static final String CHILD_DOC = "child";

Expand Down Expand Up @@ -98,7 +96,7 @@ protected void initializeAdditionalMappings(MapperService mapperService) throws
.endObject()
.endObject();

mapperService.merge(TYPE, new CompressedXContent(Strings.toString(mapping)), MapperService.MergeReason.MAPPING_UPDATE);
mapperService.merge(toMappingMetadata(mapping), MapperService.MergeReason.MAPPING_UPDATE);
}

/**
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -15,8 +15,6 @@
import org.apache.lucene.search.Query;
import org.apache.lucene.search.TermQuery;
import org.elasticsearch.ElasticsearchException;
import org.elasticsearch.common.Strings;
import org.elasticsearch.common.compress.CompressedXContent;
import org.elasticsearch.index.mapper.MapperService;
import org.elasticsearch.index.query.QueryShardException;
import org.elasticsearch.index.query.SearchExecutionContext;
Expand All @@ -40,7 +38,6 @@

public class ParentIdQueryBuilderTests extends AbstractQueryTestCase<ParentIdQueryBuilder> {

private static final String TYPE = "_doc";
private static final String JOIN_FIELD_NAME = "join_field";
private static final String PARENT_NAME = "parent";
private static final String CHILD_NAME = "child";
Expand Down Expand Up @@ -86,7 +83,7 @@ protected void initializeAdditionalMappings(MapperService mapperService) throws
.endObject()
.endObject();

mapperService.merge(TYPE, new CompressedXContent(Strings.toString(mapping)), MapperService.MergeReason.MAPPING_UPDATE);
mapperService.merge(toMappingMetadata(mapping), MapperService.MergeReason.MAPPING_UPDATE);
}

@Override
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -66,10 +66,8 @@
import org.apache.lucene.store.Directory;
import org.apache.lucene.util.BytesRef;
import org.elasticsearch.Version;
import org.elasticsearch.common.Strings;
import org.elasticsearch.common.bytes.BytesArray;
import org.elasticsearch.common.bytes.BytesReference;
import org.elasticsearch.common.compress.CompressedXContent;
import org.elasticsearch.common.geo.ShapeRelation;
import org.elasticsearch.common.lucene.search.function.FunctionScoreQuery;
import org.elasticsearch.common.settings.Settings;
Expand All @@ -86,6 +84,7 @@
import org.elasticsearch.plugins.Plugin;
import org.elasticsearch.test.ESSingleNodeTestCase;
import org.elasticsearch.test.VersionUtils;
import org.elasticsearch.xcontent.XContentBuilder;
import org.elasticsearch.xcontent.XContentFactory;
import org.junit.After;
import org.junit.Before;
Expand Down Expand Up @@ -136,52 +135,48 @@ public void init() throws Exception {
indexService = createIndex(indexName, Settings.EMPTY);
mapperService = indexService.mapperService();

String mapper = Strings.toString(
XContentFactory.jsonBuilder()
.startObject()
.startObject("type")
.startObject("properties")
.startObject("int_field")
.field("type", "integer")
.endObject()
.startObject("long_field")
.field("type", "long")
.endObject()
.startObject("half_float_field")
.field("type", "half_float")
.endObject()
.startObject("float_field")
.field("type", "float")
.endObject()
.startObject("double_field")
.field("type", "double")
.endObject()
.startObject("ip_field")
.field("type", "ip")
.endObject()
.startObject("field")
.field("type", "keyword")
.endObject()
.endObject()
.endObject()
.endObject()
);
mapperService.merge("type", new CompressedXContent(mapper), MapperService.MergeReason.MAPPING_UPDATE);
XContentBuilder mappings = XContentFactory.jsonBuilder()
.startObject()
.startObject("type")
.startObject("properties")
.startObject("int_field")
.field("type", "integer")
.endObject()
.startObject("long_field")
.field("type", "long")
.endObject()
.startObject("half_float_field")
.field("type", "half_float")
.endObject()
.startObject("float_field")
.field("type", "float")
.endObject()
.startObject("double_field")
.field("type", "double")
.endObject()
.startObject("ip_field")
.field("type", "ip")
.endObject()
.startObject("field")
.field("type", "keyword")
.endObject()
.endObject()
.endObject()
.endObject();
mapperService.merge(toMappingMetadata(mappings), MapperService.MergeReason.MAPPING_UPDATE);

String queryField = "query_field";
String percolatorMapper = Strings.toString(
XContentFactory.jsonBuilder()
.startObject()
.startObject("type")
.startObject("properties")
.startObject(queryField)
.field("type", "percolator")
.endObject()
.endObject()
.endObject()
.endObject()
);
mapperService.merge("type", new CompressedXContent(percolatorMapper), MapperService.MergeReason.MAPPING_UPDATE);
mappings = XContentFactory.jsonBuilder()
.startObject()
.startObject("type")
.startObject("properties")
.startObject(queryField)
.field("type", "percolator")
.endObject()
.endObject()
.endObject()
.endObject();
mapperService.merge(toMappingMetadata(mappings), MapperService.MergeReason.MAPPING_UPDATE);
fieldMapper = (PercolatorFieldMapper) mapperService.documentMapper().mappers().getMapper(queryField);
fieldType = (PercolatorFieldMapper.PercolatorFieldType) fieldMapper.fieldType();

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,6 @@
import org.elasticsearch.common.Strings;
import org.elasticsearch.common.bytes.BytesArray;
import org.elasticsearch.common.bytes.BytesReference;
import org.elasticsearch.common.compress.CompressedXContent;
import org.elasticsearch.common.io.stream.BytesStreamOutput;
import org.elasticsearch.common.lucene.uid.Versions;
import org.elasticsearch.core.RestApiVersion;
Expand Down Expand Up @@ -86,17 +85,11 @@ protected void initializeAdditionalMappings(MapperService mapperService) throws

docType = "_doc";
mapperService.merge(
docType,
new CompressedXContent(
Strings.toString(
PutMappingRequest.simpleMapping(queryField, "type=percolator", aliasField, "type=alias,path=" + queryField)
)
),
toMappingMetadata(PutMappingRequest.simpleMapping(queryField, "type=percolator", aliasField, "type=alias,path=" + queryField)),
MapperService.MergeReason.MAPPING_UPDATE
);
mapperService.merge(
docType,
new CompressedXContent(Strings.toString(PutMappingRequest.simpleMapping(TEXT_FIELD_NAME, "type=text"))),
toMappingMetadata(PutMappingRequest.simpleMapping(TEXT_FIELD_NAME, "type=text")),
MapperService.MergeReason.MAPPING_UPDATE
);
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -9,9 +9,7 @@
package org.elasticsearch.percolator;

import org.elasticsearch.action.admin.indices.mapping.put.PutMappingRequest;
import org.elasticsearch.common.Strings;
import org.elasticsearch.common.bytes.BytesArray;
import org.elasticsearch.common.compress.CompressedXContent;
import org.elasticsearch.index.mapper.MapperService;
import org.elasticsearch.index.query.QueryBuilder;
import org.elasticsearch.index.query.SearchExecutionContext;
Expand All @@ -25,8 +23,7 @@ public class PercolateWithNestedQueryBuilderTests extends PercolateQueryBuilderT
protected void initializeAdditionalMappings(MapperService mapperService) throws IOException {
super.initializeAdditionalMappings(mapperService);
mapperService.merge(
"_doc",
new CompressedXContent(Strings.toString(PutMappingRequest.simpleMapping("some_nested_object", "type=nested"))),
toMappingMetadata(PutMappingRequest.simpleMapping("some_nested_object", "type=nested")),
MapperService.MergeReason.MAPPING_UPDATE
);
}
Expand Down
Loading