diff options
Diffstat (limited to 'core/src/test/java/org/elasticsearch/search/aggregations/bucket/significant/SignificanceHeuristicTests.java')
-rw-r--r-- | core/src/test/java/org/elasticsearch/search/aggregations/bucket/significant/SignificanceHeuristicTests.java | 25 |
1 files changed, 13 insertions, 12 deletions
diff --git a/core/src/test/java/org/elasticsearch/search/aggregations/bucket/significant/SignificanceHeuristicTests.java b/core/src/test/java/org/elasticsearch/search/aggregations/bucket/significant/SignificanceHeuristicTests.java index 0fe9113e8f..59c110e774 100644 --- a/core/src/test/java/org/elasticsearch/search/aggregations/bucket/significant/SignificanceHeuristicTests.java +++ b/core/src/test/java/org/elasticsearch/search/aggregations/bucket/significant/SignificanceHeuristicTests.java @@ -27,6 +27,7 @@ import org.elasticsearch.common.xcontent.XContentBuilder; import org.elasticsearch.common.xcontent.XContentFactory; import org.elasticsearch.common.xcontent.XContentParser; import org.elasticsearch.common.xcontent.json.JsonXContent; +import org.elasticsearch.index.Index; import org.elasticsearch.search.SearchShardTarget; import org.elasticsearch.search.aggregations.InternalAggregation; import org.elasticsearch.search.aggregations.InternalAggregations; @@ -75,7 +76,7 @@ public class SignificanceHeuristicTests extends ESTestCase { @Override public SearchShardTarget shardTarget() { - return new SearchShardTarget("no node, this is a unit test", "no index, this is a unit test", 0); + return new SearchShardTarget("no node, this is a unit test", new Index("no index, this is a unit test", "_na_"), 0); } } @@ -102,8 +103,8 @@ public class SignificanceHeuristicTests extends ESTestCase { assertThat(originalBucket.getKeyAsString(), equalTo(streamedBucket.getKeyAsString())); assertThat(originalBucket.getSupersetDf(), equalTo(streamedBucket.getSupersetDf())); assertThat(originalBucket.getSubsetDf(), equalTo(streamedBucket.getSubsetDf())); - assertThat(streamedBucket.getSubsetSize(), equalTo(10l)); - assertThat(streamedBucket.getSupersetSize(), equalTo(20l)); + assertThat(streamedBucket.getSubsetSize(), equalTo(10L)); + assertThat(streamedBucket.getSupersetSize(), equalTo(20L)); } InternalSignificantTerms[] getRandomSignificantTerms(SignificanceHeuristic heuristic) { @@ -138,14 +139,14 @@ public class SignificanceHeuristicTests extends ESTestCase { List<InternalAggregation> aggs = createInternalAggregations(); SignificantTerms reducedAgg = (SignificantTerms) aggs.get(0).doReduce(aggs, null); assertThat(reducedAgg.getBuckets().size(), equalTo(2)); - assertThat(reducedAgg.getBuckets().get(0).getSubsetDf(), equalTo(8l)); - assertThat(reducedAgg.getBuckets().get(0).getSubsetSize(), equalTo(16l)); - assertThat(reducedAgg.getBuckets().get(0).getSupersetDf(), equalTo(10l)); - assertThat(reducedAgg.getBuckets().get(0).getSupersetSize(), equalTo(30l)); - assertThat(reducedAgg.getBuckets().get(1).getSubsetDf(), equalTo(8l)); - assertThat(reducedAgg.getBuckets().get(1).getSubsetSize(), equalTo(16l)); - assertThat(reducedAgg.getBuckets().get(1).getSupersetDf(), equalTo(10l)); - assertThat(reducedAgg.getBuckets().get(1).getSupersetSize(), equalTo(30l)); + assertThat(reducedAgg.getBuckets().get(0).getSubsetDf(), equalTo(8L)); + assertThat(reducedAgg.getBuckets().get(0).getSubsetSize(), equalTo(16L)); + assertThat(reducedAgg.getBuckets().get(0).getSupersetDf(), equalTo(10L)); + assertThat(reducedAgg.getBuckets().get(0).getSupersetSize(), equalTo(30L)); + assertThat(reducedAgg.getBuckets().get(1).getSubsetDf(), equalTo(8L)); + assertThat(reducedAgg.getBuckets().get(1).getSubsetSize(), equalTo(16L)); + assertThat(reducedAgg.getBuckets().get(1).getSupersetDf(), equalTo(10L)); + assertThat(reducedAgg.getBuckets().get(1).getSupersetSize(), equalTo(30L)); } // Create aggregations as they might come from three different shards and return as list. @@ -251,7 +252,7 @@ public class SignificanceHeuristicTests extends ESTestCase { stParser.nextToken(); SignificantTermsAggregatorFactory aggregatorFactory = (SignificantTermsAggregatorFactory) new SignificantTermsParser(heuristicParserMapper).parse("testagg", stParser, searchContext); stParser.nextToken(); - assertThat(aggregatorFactory.getBucketCountThresholds().getMinDocCount(), equalTo(200l)); + assertThat(aggregatorFactory.getBucketCountThresholds().getMinDocCount(), equalTo(200L)); assertThat(stParser.currentToken(), equalTo(null)); stParser.close(); return aggregatorFactory.getSignificanceHeuristic(); |