diff options
author | Robert Muir <rmuir@apache.org> | 2015-09-21 23:35:32 -0400 |
---|---|---|
committer | Robert Muir <rmuir@apache.org> | 2015-09-21 23:35:32 -0400 |
commit | 2f67cacaa3002d1f81eb341bab00252d275e9b0a (patch) | |
tree | 75cf226259d0dcabec9bd59042ff243393c6838e /core/src/main/java/org/elasticsearch/index | |
parent | eedb39a5a950f2685fd63524ce58c4bb321ecb36 (diff) |
Fix all javadocs issues, re-enable compiler warnings (but disable on java9 where maven is broken)
Diffstat (limited to 'core/src/main/java/org/elasticsearch/index')
61 files changed, 116 insertions, 150 deletions
diff --git a/core/src/main/java/org/elasticsearch/index/aliases/IndexAliasesService.java b/core/src/main/java/org/elasticsearch/index/aliases/IndexAliasesService.java index fc93153b50..2b519c2b84 100644 --- a/core/src/main/java/org/elasticsearch/index/aliases/IndexAliasesService.java +++ b/core/src/main/java/org/elasticsearch/index/aliases/IndexAliasesService.java @@ -56,8 +56,8 @@ public class IndexAliasesService extends AbstractIndexComponent { /** * Returns the filter associated with listed filtering aliases. - * <p/> - * <p>The list of filtering aliases should be obtained by calling MetaData.filteringAliases. + * <p> + * The list of filtering aliases should be obtained by calling MetaData.filteringAliases. * Returns <tt>null</tt> if no filtering is required.</p> */ public Query aliasFilter(String... aliasNames) { diff --git a/core/src/main/java/org/elasticsearch/index/analysis/AnalysisSettingsRequired.java b/core/src/main/java/org/elasticsearch/index/analysis/AnalysisSettingsRequired.java index 155b78a11d..847752c514 100644 --- a/core/src/main/java/org/elasticsearch/index/analysis/AnalysisSettingsRequired.java +++ b/core/src/main/java/org/elasticsearch/index/analysis/AnalysisSettingsRequired.java @@ -22,11 +22,11 @@ import java.lang.annotation.*; /** * A marker annotation on {@link CharFilterFactory}, {@link AnalyzerProvider}, {@link TokenFilterFactory}, - * or {@link @TokenizerFactory} which will cause the provider/factory to only be created when explicit settings + * or {@link TokenizerFactory} which will cause the provider/factory to only be created when explicit settings * are provided. */ @Target({ElementType.TYPE}) @Retention(RetentionPolicy.RUNTIME) @Documented public @interface AnalysisSettingsRequired { -}
\ No newline at end of file +} diff --git a/core/src/main/java/org/elasticsearch/index/analysis/CJKBigramFilterFactory.java b/core/src/main/java/org/elasticsearch/index/analysis/CJKBigramFilterFactory.java index 4b3a1e3bcb..7221c4b12e 100644 --- a/core/src/main/java/org/elasticsearch/index/analysis/CJKBigramFilterFactory.java +++ b/core/src/main/java/org/elasticsearch/index/analysis/CJKBigramFilterFactory.java @@ -34,15 +34,15 @@ import java.util.Set; /** * Factory that creates a {@link CJKBigramFilter} to form bigrams of CJK terms * that are generated from StandardTokenizer or ICUTokenizer. - * <p/> + * <p> * CJK types are set by these tokenizers, but you can also use flags to * explicitly control which of the CJK scripts are turned into bigrams. - * <p/> + * <p> * By default, when a CJK character has no adjacent characters to form a bigram, * it is output in unigram form. If you want to always output both unigrams and * bigrams, set the <code>outputUnigrams</code> flag. This can be used for a * combined unigram+bigram approach. - * <p/> + * <p> * In all cases, all non-CJK input is passed thru unmodified. */ public final class CJKBigramFilterFactory extends AbstractTokenFilterFactory { diff --git a/core/src/main/java/org/elasticsearch/index/analysis/KeepTypesFilterFactory.java b/core/src/main/java/org/elasticsearch/index/analysis/KeepTypesFilterFactory.java index 6b8e81f495..9f17dadfeb 100644 --- a/core/src/main/java/org/elasticsearch/index/analysis/KeepTypesFilterFactory.java +++ b/core/src/main/java/org/elasticsearch/index/analysis/KeepTypesFilterFactory.java @@ -33,12 +33,11 @@ import java.util.HashSet; import java.util.Set; /** - * A {@link TokenFilterFactory} for {@link TypeFilter}. This filter only + * A {@link TokenFilterFactory} for {@link TypeTokenFilter}. This filter only * keep tokens that are contained in the set configured via * {@value #KEEP_TYPES_KEY} setting. - * <p/> + * <p> * Configuration options: - * <p/> * <ul> * <li>{@value #KEEP_TYPES_KEY} the array of words / tokens to keep.</li> * </ul> diff --git a/core/src/main/java/org/elasticsearch/index/analysis/KeepWordFilterFactory.java b/core/src/main/java/org/elasticsearch/index/analysis/KeepWordFilterFactory.java index a92ade2467..2c082288ed 100644 --- a/core/src/main/java/org/elasticsearch/index/analysis/KeepWordFilterFactory.java +++ b/core/src/main/java/org/elasticsearch/index/analysis/KeepWordFilterFactory.java @@ -36,20 +36,16 @@ import org.elasticsearch.index.settings.IndexSettings; * keep tokens that are contained in the term set configured via * {@value #KEEP_WORDS_KEY} setting. This filter acts like an inverse stop * filter. - * <p/> + * <p> * Configuration options: - * <p/> * <ul> * <li>{@value #KEEP_WORDS_KEY} the array of words / tokens to keep.</li> - * <p/> * <li>{@value #KEEP_WORDS_PATH_KEY} an reference to a file containing the words * / tokens to keep. Note: this is an alternative to {@value #KEEP_WORDS_KEY} if * both are set an exception will be thrown.</li> - * <p/> * <li>{@value #ENABLE_POS_INC_KEY} <code>true</code> iff the filter should * maintain position increments for dropped tokens. The default is * <code>true</code>.</li> - * <p/> * <li>{@value #KEEP_WORDS_CASE_KEY} to use case sensitive keep words. The * default is <code>false</code> which corresponds to case-sensitive.</li> * </ul> diff --git a/core/src/main/java/org/elasticsearch/index/analysis/SnowballAnalyzer.java b/core/src/main/java/org/elasticsearch/index/analysis/SnowballAnalyzer.java index 76bf8d4265..4ce0bee7a2 100644 --- a/core/src/main/java/org/elasticsearch/index/analysis/SnowballAnalyzer.java +++ b/core/src/main/java/org/elasticsearch/index/analysis/SnowballAnalyzer.java @@ -28,6 +28,7 @@ import org.apache.lucene.analysis.core.LowerCaseFilter; import org.apache.lucene.analysis.core.StopFilter; import org.apache.lucene.analysis.snowball.SnowballFilter; import org.apache.lucene.analysis.standard.StandardTokenizer; +import org.apache.lucene.analysis.standard.StandardFilter; import org.apache.lucene.analysis.standard.std40.StandardTokenizer40; import org.apache.lucene.analysis.util.CharArraySet; import org.apache.lucene.util.Version; @@ -83,4 +84,4 @@ public final class SnowballAnalyzer extends Analyzer { result = new SnowballFilter(result, name); return new TokenStreamComponents(tokenizer, result); } -}
\ No newline at end of file +} diff --git a/core/src/main/java/org/elasticsearch/index/analysis/SnowballAnalyzerProvider.java b/core/src/main/java/org/elasticsearch/index/analysis/SnowballAnalyzerProvider.java index 1d1a9ebb70..39cf56b406 100644 --- a/core/src/main/java/org/elasticsearch/index/analysis/SnowballAnalyzerProvider.java +++ b/core/src/main/java/org/elasticsearch/index/analysis/SnowballAnalyzerProvider.java @@ -39,7 +39,7 @@ import org.elasticsearch.index.settings.IndexSettings; * Stemmer, use them directly with the SnowballFilter and a CustomAnalyzer. * Configuration of language is done with the "language" attribute or the analyzer. * Also supports additional stopwords via "stopwords" attribute - * <p/> + * <p> * The SnowballAnalyzer comes with a StandardFilter, LowerCaseFilter, StopFilter * and the SnowballFilter. * @@ -73,4 +73,4 @@ public class SnowballAnalyzerProvider extends AbstractIndexAnalyzerProvider<Snow public SnowballAnalyzer get() { return this.analyzer; } -}
\ No newline at end of file +} diff --git a/core/src/main/java/org/elasticsearch/index/analysis/StandardHtmlStripAnalyzer.java b/core/src/main/java/org/elasticsearch/index/analysis/StandardHtmlStripAnalyzer.java index 7202f1f94a..a6cfe91cbf 100644 --- a/core/src/main/java/org/elasticsearch/index/analysis/StandardHtmlStripAnalyzer.java +++ b/core/src/main/java/org/elasticsearch/index/analysis/StandardHtmlStripAnalyzer.java @@ -34,8 +34,7 @@ import org.apache.lucene.util.Version; public class StandardHtmlStripAnalyzer extends StopwordAnalyzerBase { /** - * @deprecated use {@link StandardHtmlStripAnalyzer#StandardHtmlStripAnalyzer(org.apache.lucene.util.Version, - * org.apache.lucene.analysis.util.CharArraySet)} instead + * @deprecated use {@link StandardHtmlStripAnalyzer#StandardHtmlStripAnalyzer(CharArraySet)} instead */ @Deprecated public StandardHtmlStripAnalyzer() { @@ -62,4 +61,4 @@ public class StandardHtmlStripAnalyzer extends StopwordAnalyzerBase { return new TokenStreamComponents(src, tok); } -}
\ No newline at end of file +} diff --git a/core/src/main/java/org/elasticsearch/index/cache/bitset/BitsetFilterCache.java b/core/src/main/java/org/elasticsearch/index/cache/bitset/BitsetFilterCache.java index 58768cd08d..30c0905dc0 100644 --- a/core/src/main/java/org/elasticsearch/index/cache/bitset/BitsetFilterCache.java +++ b/core/src/main/java/org/elasticsearch/index/cache/bitset/BitsetFilterCache.java @@ -65,7 +65,7 @@ import java.util.concurrent.*; /** * This is a cache for {@link BitDocIdSet} based filters and is unbounded by size or time. - * <p/> + * <p> * Use this cache with care, only components that require that a filter is to be materialized as a {@link BitDocIdSet} * and require that it should always be around should use this cache, otherwise the * {@link org.elasticsearch.index.cache.query.QueryCache} should be used instead. diff --git a/core/src/main/java/org/elasticsearch/index/cache/query/QueryCacheStats.java b/core/src/main/java/org/elasticsearch/index/cache/query/QueryCacheStats.java index aaf2730a31..947968deab 100644 --- a/core/src/main/java/org/elasticsearch/index/cache/query/QueryCacheStats.java +++ b/core/src/main/java/org/elasticsearch/index/cache/query/QueryCacheStats.java @@ -27,6 +27,8 @@ import org.elasticsearch.common.xcontent.ToXContent; import org.elasticsearch.common.xcontent.XContentBuilder; import org.elasticsearch.common.xcontent.XContentBuilderString; +import org.apache.lucene.search.DocIdSet; + import java.io.IOException; /** diff --git a/core/src/main/java/org/elasticsearch/index/codec/CodecService.java b/core/src/main/java/org/elasticsearch/index/codec/CodecService.java index aa29f79ba7..77d8319b55 100644 --- a/core/src/main/java/org/elasticsearch/index/codec/CodecService.java +++ b/core/src/main/java/org/elasticsearch/index/codec/CodecService.java @@ -34,12 +34,10 @@ import org.elasticsearch.index.settings.IndexSettings; /** * Since Lucene 4.0 low level index segments are read and written through a - * codec layer that allows to use use-case specific file formats & + * codec layer that allows to use use-case specific file formats & * data-structures per field. Elasticsearch exposes the full * {@link Codec} capabilities through this {@link CodecService}. * - * @see PostingsFormatService - * @see DocValuesFormatService */ public class CodecService extends AbstractIndexComponent { diff --git a/core/src/main/java/org/elasticsearch/index/codec/postingsformat/BloomFilterPostingsFormat.java b/core/src/main/java/org/elasticsearch/index/codec/postingsformat/BloomFilterPostingsFormat.java index 9b29c9cc81..71a52a7fbd 100644 --- a/core/src/main/java/org/elasticsearch/index/codec/postingsformat/BloomFilterPostingsFormat.java +++ b/core/src/main/java/org/elasticsearch/index/codec/postingsformat/BloomFilterPostingsFormat.java @@ -39,8 +39,7 @@ import java.util.Map.Entry; * </p> * <p> * This is a special bloom filter version, based on {@link org.elasticsearch.common.util.BloomFilter} and inspired - * by Lucene {@link org.apache.lucene.codecs.bloom.BloomFilteringPostingsFormat}. - * </p> + * by Lucene {@code org.apache.lucene.codecs.bloom.BloomFilteringPostingsFormat}. * @deprecated only for reading old segments */ @Deprecated @@ -67,7 +66,7 @@ public class BloomFilterPostingsFormat extends PostingsFormat { * * @param delegatePostingsFormat The PostingsFormat that records all the non-bloom filter data i.e. * postings info. - * @param bloomFilterFactory The {@link BloomFilter.Factory} responsible for sizing BloomFilters + * @param bloomFilterFactory The {@link org.elasticsearch.common.util.BloomFilter.Factory} responsible for sizing BloomFilters * appropriately */ public BloomFilterPostingsFormat(PostingsFormat delegatePostingsFormat, diff --git a/core/src/main/java/org/elasticsearch/index/engine/Engine.java b/core/src/main/java/org/elasticsearch/index/engine/Engine.java index 92434d340c..0c66b5148c 100644 --- a/core/src/main/java/org/elasticsearch/index/engine/Engine.java +++ b/core/src/main/java/org/elasticsearch/index/engine/Engine.java @@ -1131,7 +1131,6 @@ public abstract class Engine implements Closeable { /** * Returns <code>true</code> the internal writer has any uncommitted changes. Otherwise <code>false</code> - * @return */ public abstract boolean hasUncommittedChanges(); diff --git a/core/src/main/java/org/elasticsearch/index/engine/EngineClosedException.java b/core/src/main/java/org/elasticsearch/index/engine/EngineClosedException.java index ef55708622..062e79eac1 100644 --- a/core/src/main/java/org/elasticsearch/index/engine/EngineClosedException.java +++ b/core/src/main/java/org/elasticsearch/index/engine/EngineClosedException.java @@ -27,8 +27,8 @@ import java.io.IOException; /** * An engine is already closed. - * <p/> - * <p>Note, the relationship between shard and engine indicates that engine closed is shard closed, and + * <p> + * Note, the relationship between shard and engine indicates that engine closed is shard closed, and * we might get something slipping through the the shard and into the engine while the shard is closing. * * diff --git a/core/src/main/java/org/elasticsearch/index/fielddata/FieldData.java b/core/src/main/java/org/elasticsearch/index/fielddata/FieldData.java index 0e286217ab..97750cf069 100644 --- a/core/src/main/java/org/elasticsearch/index/fielddata/FieldData.java +++ b/core/src/main/java/org/elasticsearch/index/fielddata/FieldData.java @@ -218,9 +218,9 @@ public enum FieldData { /** * Returns a single-valued view of the {@link SortedNumericDoubleValues}, - * if it was previously wrapped with {@link #singleton(NumericDocValues, Bits)}, + * if it was previously wrapped with {@link DocValues#singleton(NumericDocValues, Bits)}, * or null. - * @see #unwrapSingletonBits(SortedNumericDocValues) + * @see DocValues#unwrapSingletonBits(SortedNumericDocValues) */ public static NumericDoubleValues unwrapSingleton(SortedNumericDoubleValues values) { if (values instanceof SingletonSortedNumericDoubleValues) { diff --git a/core/src/main/java/org/elasticsearch/index/fielddata/MultiGeoPointValues.java b/core/src/main/java/org/elasticsearch/index/fielddata/MultiGeoPointValues.java index 1d79077abc..6fa9c799dd 100644 --- a/core/src/main/java/org/elasticsearch/index/fielddata/MultiGeoPointValues.java +++ b/core/src/main/java/org/elasticsearch/index/fielddata/MultiGeoPointValues.java @@ -27,7 +27,7 @@ import org.elasticsearch.common.geo.GeoPoint; * GeoPointValues values = ..; * values.setDocId(docId); * final int numValues = values.count(); - * for (int i = 0; i < numValues; i++) { + * for (int i = 0; i < numValues; i++) { * GeoPoint value = values.valueAt(i); * // process value * } diff --git a/core/src/main/java/org/elasticsearch/index/fielddata/ordinals/OrdinalsBuilder.java b/core/src/main/java/org/elasticsearch/index/fielddata/ordinals/OrdinalsBuilder.java index fa7eef6e6b..3b66adfee9 100644 --- a/core/src/main/java/org/elasticsearch/index/fielddata/ordinals/OrdinalsBuilder.java +++ b/core/src/main/java/org/elasticsearch/index/fielddata/ordinals/OrdinalsBuilder.java @@ -34,7 +34,7 @@ import java.io.IOException; import java.util.Arrays; /** - * Simple class to build document ID <-> ordinal mapping. Note: Ordinals are + * Simple class to build document ID <-> ordinal mapping. Note: Ordinals are * <tt>1</tt> based monotonically increasing positive integers. <tt>0</tt> * donates the missing value in this context. */ @@ -75,7 +75,7 @@ public final class OrdinalsBuilder implements Closeable { * with document 2: it has 2 more ordinals on level 1: 3 and 4 and its next level index is 1 meaning that there are remaining * ordinals on the next level. On level 2 at index 1, we can read [5 0 0 0] meaning that 5 is an ordinal as well, but the * fact that it is followed by zeros means that there are no more ordinals. In the end, document 2 has 2, 3, 4 and 5 as ordinals. - * <p/> + * <p> * In addition to these structures, there is another array which stores the current position (level + slice + offset in the slice) * in order to be able to append data in constant time. */ @@ -300,7 +300,7 @@ public final class OrdinalsBuilder implements Closeable { } /** - * Return a {@link PackedInts.Reader} instance mapping every doc ID to its first ordinal + 1 if it exists and 0 otherwise. + * Return a {@link org.apache.lucene.util.packed.PackedInts.Reader} instance mapping every doc ID to its first ordinal + 1 if it exists and 0 otherwise. */ public PackedInts.Reader getFirstOrdinals() { return ordinals.firstOrdinals; @@ -419,7 +419,7 @@ public final class OrdinalsBuilder implements Closeable { /** * A {@link TermsEnum} that iterates only full precision prefix coded 64 bit values. * - * @see #buildFromTerms(TermsEnum, Bits) + * @see #buildFromTerms(TermsEnum) */ public static TermsEnum wrapNumeric64Bit(TermsEnum termsEnum) { return new FilteredTermsEnum(termsEnum, false) { @@ -434,7 +434,7 @@ public final class OrdinalsBuilder implements Closeable { /** * A {@link TermsEnum} that iterates only full precision prefix coded 32 bit values. * - * @see #buildFromTerms(TermsEnum, Bits) + * @see #buildFromTerms(TermsEnum) */ public static TermsEnum wrapNumeric32Bit(TermsEnum termsEnum) { return new FilteredTermsEnum(termsEnum, false) { diff --git a/core/src/main/java/org/elasticsearch/index/fielddata/plain/AbstractIndexFieldData.java b/core/src/main/java/org/elasticsearch/index/fielddata/plain/AbstractIndexFieldData.java index 1601edc7ad..4cd172fd22 100644 --- a/core/src/main/java/org/elasticsearch/index/fielddata/plain/AbstractIndexFieldData.java +++ b/core/src/main/java/org/elasticsearch/index/fielddata/plain/AbstractIndexFieldData.java @@ -99,7 +99,7 @@ public abstract class AbstractIndexFieldData<FD extends AtomicFieldData> extends * the memory overhead for loading the data. Each field data * implementation should implement its own {@code PerValueEstimator} if it * intends to take advantage of the MemoryCircuitBreaker. - * <p/> + * <p> * Note that the .beforeLoad(...) and .afterLoad(...) methods must be * manually called. */ @@ -118,7 +118,6 @@ public abstract class AbstractIndexFieldData<FD extends AtomicFieldData> extends * * @param terms terms to be estimated * @return A TermsEnum for the given terms - * @throws IOException */ public TermsEnum beforeLoad(Terms terms) throws IOException; diff --git a/core/src/main/java/org/elasticsearch/index/fielddata/plain/PackedArrayIndexFieldData.java b/core/src/main/java/org/elasticsearch/index/fielddata/plain/PackedArrayIndexFieldData.java index 1b54e386c0..5f95d99f85 100644 --- a/core/src/main/java/org/elasticsearch/index/fielddata/plain/PackedArrayIndexFieldData.java +++ b/core/src/main/java/org/elasticsearch/index/fielddata/plain/PackedArrayIndexFieldData.java @@ -418,7 +418,6 @@ public class PackedArrayIndexFieldData extends AbstractIndexFieldData<AtomicNume /** * @return A TermsEnum wrapped in a RamAccountingTermsEnum - * @throws IOException */ @Override public TermsEnum beforeLoad(Terms terms) throws IOException { diff --git a/core/src/main/java/org/elasticsearch/index/fielddata/plain/PagedBytesIndexFieldData.java b/core/src/main/java/org/elasticsearch/index/fielddata/plain/PagedBytesIndexFieldData.java index dce5e403e2..48c954394e 100644 --- a/core/src/main/java/org/elasticsearch/index/fielddata/plain/PagedBytesIndexFieldData.java +++ b/core/src/main/java/org/elasticsearch/index/fielddata/plain/PagedBytesIndexFieldData.java @@ -189,7 +189,6 @@ public class PagedBytesIndexFieldData extends AbstractIndexOrdinalsFieldData { * * @param terms terms to be estimated * @return A possibly wrapped TermsEnum for the terms - * @throws IOException */ @Override public TermsEnum beforeLoad(Terms terms) throws IOException { diff --git a/core/src/main/java/org/elasticsearch/index/fielddata/plain/SortedNumericDVIndexFieldData.java b/core/src/main/java/org/elasticsearch/index/fielddata/plain/SortedNumericDVIndexFieldData.java index 354ad43a0c..0c78d2e5f1 100644 --- a/core/src/main/java/org/elasticsearch/index/fielddata/plain/SortedNumericDVIndexFieldData.java +++ b/core/src/main/java/org/elasticsearch/index/fielddata/plain/SortedNumericDVIndexFieldData.java @@ -37,7 +37,7 @@ import java.util.Collections; /** * FieldData backed by {@link LeafReader#getSortedNumericDocValues(String)} - * @see FieldInfo.DocValuesType#SORTED_NUMERIC + * @see DocValuesType#SORTED_NUMERIC */ public class SortedNumericDVIndexFieldData extends DocValuesIndexFieldData implements IndexNumericFieldData { private final NumericType numericType; @@ -132,9 +132,7 @@ public class SortedNumericDVIndexFieldData extends DocValuesIndexFieldData imple * Order of values within a document is consistent with * {@link Float#compareTo(Float)}, hence the following reversible * transformation is applied at both index and search: - * {code} - * bits ^ (bits >> 31) & 0x7fffffff - * {code} + * {@code bits ^ (bits >> 31) & 0x7fffffff} * <p> * Although the API is multi-valued, most codecs in Lucene specialize * for the case where documents have at most one value. In this case @@ -223,9 +221,7 @@ public class SortedNumericDVIndexFieldData extends DocValuesIndexFieldData imple * Order of values within a document is consistent with * {@link Double#compareTo(Double)}, hence the following reversible * transformation is applied at both index and search: - * {code} - * bits ^ (bits >> 63) & 0x7fffffffffffffffL - * {code} + * {@code bits ^ (bits >> 63) & 0x7fffffffffffffffL} * <p> * Although the API is multi-valued, most codecs in Lucene specialize * for the case where documents have at most one value. In this case diff --git a/core/src/main/java/org/elasticsearch/index/fieldvisitor/FieldsVisitor.java b/core/src/main/java/org/elasticsearch/index/fieldvisitor/FieldsVisitor.java index 5048baec12..31b0a08022 100644 --- a/core/src/main/java/org/elasticsearch/index/fieldvisitor/FieldsVisitor.java +++ b/core/src/main/java/org/elasticsearch/index/fieldvisitor/FieldsVisitor.java @@ -46,8 +46,10 @@ import java.util.List; import java.util.Map; import java.util.Set; +import org.apache.lucene.index.StoredFieldVisitor; + /** - * Base {@link StoredFieldsVisitor} that retrieves all non-redundant metadata. + * Base {@link StoredFieldVisitor} that retrieves all non-redundant metadata. */ public class FieldsVisitor extends StoredFieldVisitor { diff --git a/core/src/main/java/org/elasticsearch/index/get/ShardGetService.java b/core/src/main/java/org/elasticsearch/index/get/ShardGetService.java index 212362c394..1cf68ad383 100644 --- a/core/src/main/java/org/elasticsearch/index/get/ShardGetService.java +++ b/core/src/main/java/org/elasticsearch/index/get/ShardGetService.java @@ -102,10 +102,10 @@ public final class ShardGetService extends AbstractIndexShardComponent { } /** - * Returns {@link GetResult} based on the specified {@link Engine.GetResult} argument. + * Returns {@link GetResult} based on the specified {@link org.elasticsearch.index.engine.Engine.GetResult} argument. * This method basically loads specified fields for the associated document in the engineGetResult. * This method load the fields from the Lucene index and not from transaction log and therefore isn't realtime. - * <p/> + * <p> * Note: Call <b>must</b> release engine searcher associated with engineGetResult! */ public GetResult get(Engine.GetResult engineGetResult, String id, String type, String[] fields, FetchSourceContext fetchSourceContext, boolean ignoreErrorsOnGeneratedFields) { diff --git a/core/src/main/java/org/elasticsearch/index/indexing/IndexingOperationListener.java b/core/src/main/java/org/elasticsearch/index/indexing/IndexingOperationListener.java index bb4c109e6a..858453fcba 100644 --- a/core/src/main/java/org/elasticsearch/index/indexing/IndexingOperationListener.java +++ b/core/src/main/java/org/elasticsearch/index/indexing/IndexingOperationListener.java @@ -35,7 +35,7 @@ public abstract class IndexingOperationListener { /** * Called after the indexing occurs, under a locking scheme to maintain * concurrent updates to the same doc. - * <p/> + * <p> * Note, long operations should not occur under this callback. */ public void postCreateUnderLock(Engine.Create create) { @@ -66,7 +66,7 @@ public abstract class IndexingOperationListener { /** * Called after the indexing occurs, under a locking scheme to maintain * concurrent updates to the same doc. - * <p/> + * <p> * Note, long operations should not occur under this callback. */ public void postIndexUnderLock(Engine.Index index) { @@ -97,7 +97,7 @@ public abstract class IndexingOperationListener { /** * Called after the delete occurs, under a locking scheme to maintain * concurrent updates to the same doc. - * <p/> + * <p> * Note, long operations should not occur under this callback. */ public void postDeleteUnderLock(Engine.Delete delete) { diff --git a/core/src/main/java/org/elasticsearch/index/indexing/IndexingStats.java b/core/src/main/java/org/elasticsearch/index/indexing/IndexingStats.java index 5bda1de73b..af4add9132 100644 --- a/core/src/main/java/org/elasticsearch/index/indexing/IndexingStats.java +++ b/core/src/main/java/org/elasticsearch/index/indexing/IndexingStats.java @@ -113,7 +113,6 @@ public class IndexingStats implements Streamable, ToXContent { /** * Returns if the index is under merge throttling control - * @return */ public boolean isThrottled() { return isThrottled; @@ -121,7 +120,6 @@ public class IndexingStats implements Streamable, ToXContent { /** * Gets the amount of time in milliseconds that the index has been under merge throttling control - * @return */ public long getThrottleTimeInMillis() { return throttleTimeInMillis; @@ -129,7 +127,6 @@ public class IndexingStats implements Streamable, ToXContent { /** * Gets the amount of time in a TimeValue that the index has been under merge throttling control - * @return */ public TimeValue getThrottleTime() { return new TimeValue(throttleTimeInMillis); diff --git a/core/src/main/java/org/elasticsearch/index/mapper/geo/GeoPointFieldMapper.java b/core/src/main/java/org/elasticsearch/index/mapper/geo/GeoPointFieldMapper.java index 863d71f30b..222d2623c5 100644 --- a/core/src/main/java/org/elasticsearch/index/mapper/geo/GeoPointFieldMapper.java +++ b/core/src/main/java/org/elasticsearch/index/mapper/geo/GeoPointFieldMapper.java @@ -69,7 +69,7 @@ import static org.elasticsearch.index.mapper.core.TypeParsers.parsePathType; /** * Parsing: We handle: - * <p/> + * <p> * - "field" : "geo_hash" * - "field" : "lat,lon" * - "field" : { diff --git a/core/src/main/java/org/elasticsearch/index/mapper/geo/GeoShapeFieldMapper.java b/core/src/main/java/org/elasticsearch/index/mapper/geo/GeoShapeFieldMapper.java index 417950808a..770df633b9 100644 --- a/core/src/main/java/org/elasticsearch/index/mapper/geo/GeoShapeFieldMapper.java +++ b/core/src/main/java/org/elasticsearch/index/mapper/geo/GeoShapeFieldMapper.java @@ -61,13 +61,13 @@ import static org.elasticsearch.index.mapper.MapperBuilders.geoShapeField; /** * FieldMapper for indexing {@link com.spatial4j.core.shape.Shape}s. - * <p/> + * <p> * Currently Shapes can only be indexed and can only be queried using * {@link org.elasticsearch.index.query.GeoShapeQueryParser}, consequently * a lot of behavior in this Mapper is disabled. - * <p/> + * <p> * Format supported: - * <p/> + * <p> * "field" : { * "type" : "polygon", * "coordinates" : [ diff --git a/core/src/main/java/org/elasticsearch/index/percolator/PercolatorQueriesRegistry.java b/core/src/main/java/org/elasticsearch/index/percolator/PercolatorQueriesRegistry.java index 9283a9b033..9740054c18 100644 --- a/core/src/main/java/org/elasticsearch/index/percolator/PercolatorQueriesRegistry.java +++ b/core/src/main/java/org/elasticsearch/index/percolator/PercolatorQueriesRegistry.java @@ -60,7 +60,7 @@ import java.util.concurrent.atomic.AtomicBoolean; /** * Each shard will have a percolator registry even if there isn't a {@link PercolatorService#TYPE_NAME} document type in the index. * For shards with indices that have no {@link PercolatorService#TYPE_NAME} document type, this will hold no percolate queries. - * <p/> + * <p> * Once a document type has been created, the real-time percolator will start to listen to write events and update the * this registry with queries in real time. */ diff --git a/core/src/main/java/org/elasticsearch/index/query/BoolQueryBuilder.java b/core/src/main/java/org/elasticsearch/index/query/BoolQueryBuilder.java index c377667310..d3fa929945 100644 --- a/core/src/main/java/org/elasticsearch/index/query/BoolQueryBuilder.java +++ b/core/src/main/java/org/elasticsearch/index/query/BoolQueryBuilder.java @@ -108,12 +108,12 @@ public class BoolQueryBuilder extends QueryBuilder implements BoostableQueryBuil /** * Specifies a minimum number of the optional (should) boolean clauses which must be satisfied. - * <p/> - * <p>By default no optional clauses are necessary for a match + * <p> + * By default no optional clauses are necessary for a match * (unless there are no required clauses). If this method is used, * then the specified number of clauses is required. - * <p/> - * <p>Use of this method is totally independent of specifying that + * <p> + * Use of this method is totally independent of specifying that * any specific clauses are required (or prohibited). This number will * only be compared against the number of matching optional clauses. * diff --git a/core/src/main/java/org/elasticsearch/index/query/BoostingQueryBuilder.java b/core/src/main/java/org/elasticsearch/index/query/BoostingQueryBuilder.java index 9d67469deb..1e2f9c4d00 100644 --- a/core/src/main/java/org/elasticsearch/index/query/BoostingQueryBuilder.java +++ b/core/src/main/java/org/elasticsearch/index/query/BoostingQueryBuilder.java @@ -27,7 +27,7 @@ import java.io.IOException; * The BoostingQuery class can be used to effectively demote results that match a given query. * Unlike the "NOT" clause, this still selects documents that contain undesirable terms, * but reduces their overall score: - * <p/> + * <p> * Query balancedQuery = new BoostingQuery(positiveQuery, negativeQuery, 0.01f); * In this scenario the positiveQuery contains the mandatory, desirable criteria which is used to * select all matching documents, and the negativeQuery contains the undesirable elements which diff --git a/core/src/main/java/org/elasticsearch/index/query/CommonTermsQueryBuilder.java b/core/src/main/java/org/elasticsearch/index/query/CommonTermsQueryBuilder.java index ae9c10d295..32b74d0c09 100644 --- a/core/src/main/java/org/elasticsearch/index/query/CommonTermsQueryBuilder.java +++ b/core/src/main/java/org/elasticsearch/index/query/CommonTermsQueryBuilder.java @@ -29,7 +29,7 @@ import java.io.IOException; /** * CommonTermsQuery query is a query that executes high-frequency terms in a * optional sub-query to prevent slow queries due to "common" terms like - * stopwords. This query basically builds 2 queries off the {@link #add(Term) + * stopwords. This query basically builds 2 queries off the {@code #add(Term) * added} terms where low-frequency terms are added to a required boolean clause * and high-frequency terms are added to an optional boolean clause. The * optional clause is only executed if the required "low-frequency' clause @@ -40,7 +40,6 @@ import java.io.IOException; * significantly contribute to the document score unless at least one of the * low-frequency terms are matched such that this query can improve query * execution times significantly if applicable. - * <p> */ public class CommonTermsQueryBuilder extends QueryBuilder implements BoostableQueryBuilder<CommonTermsQueryBuilder> { @@ -123,7 +122,7 @@ public class CommonTermsQueryBuilder extends QueryBuilder implements BoostableQu /** * Sets the cutoff document frequency for high / low frequent terms. A value - * in [0..1] (or absolute number >=1) representing the maximum threshold of + * in [0..1] (or absolute number >=1) representing the maximum threshold of * a terms document frequency to be considered a low frequency term. * Defaults to * <tt>{@value CommonTermsQueryParser#DEFAULT_MAX_TERM_DOC_FREQ}</tt> diff --git a/core/src/main/java/org/elasticsearch/index/query/GeoPolygonQueryBuilder.java b/core/src/main/java/org/elasticsearch/index/query/GeoPolygonQueryBuilder.java index da1dac2c11..2d486e05a1 100644 --- a/core/src/main/java/org/elasticsearch/index/query/GeoPolygonQueryBuilder.java +++ b/core/src/main/java/org/elasticsearch/index/query/GeoPolygonQueryBuilder.java @@ -49,7 +49,6 @@ public class GeoPolygonQueryBuilder extends QueryBuilder { * * @param lat The latitude * @param lon The longitude - * @return */ public GeoPolygonQueryBuilder addPoint(double lat, double lon) { return addPoint(new GeoPoint(lat, lon)); diff --git a/core/src/main/java/org/elasticsearch/index/query/GeohashCellQuery.java b/core/src/main/java/org/elasticsearch/index/query/GeohashCellQuery.java index 6067bc675b..84d3857812 100644 --- a/core/src/main/java/org/elasticsearch/index/query/GeohashCellQuery.java +++ b/core/src/main/java/org/elasticsearch/index/query/GeohashCellQuery.java @@ -44,7 +44,7 @@ import java.util.List; * Geohash prefix is defined by the filter and all geohashes that are matching this * prefix will be returned. The <code>neighbors</code> flag allows to filter * geohashes that surround the given geohash. In general the neighborhood of a - * geohash is defined by its eight adjacent cells.<br /> + * geohash is defined by its eight adjacent cells.<br> * The structure of the {@link GeohashCellQuery} is defined as: * <pre> * "geohash_bbox" { diff --git a/core/src/main/java/org/elasticsearch/index/query/MatchQueryBuilder.java b/core/src/main/java/org/elasticsearch/index/query/MatchQueryBuilder.java index 6f73f08dc0..c7c530b1d5 100644 --- a/core/src/main/java/org/elasticsearch/index/query/MatchQueryBuilder.java +++ b/core/src/main/java/org/elasticsearch/index/query/MatchQueryBuilder.java @@ -163,7 +163,7 @@ public class MatchQueryBuilder extends QueryBuilder implements BoostableQueryBui } /** - * Set a cutoff value in [0..1] (or absolute number >=1) representing the + * Set a cutoff value in [0..1] (or absolute number >=1) representing the * maximum threshold of a terms document frequency to be considered a low * frequency term. */ @@ -266,4 +266,4 @@ public class MatchQueryBuilder extends QueryBuilder implements BoostableQueryBui builder.endObject(); builder.endObject(); } -}
\ No newline at end of file +} diff --git a/core/src/main/java/org/elasticsearch/index/query/MoreLikeThisQueryBuilder.java b/core/src/main/java/org/elasticsearch/index/query/MoreLikeThisQueryBuilder.java index 4994070fd7..5c7e24b53c 100644 --- a/core/src/main/java/org/elasticsearch/index/query/MoreLikeThisQueryBuilder.java +++ b/core/src/main/java/org/elasticsearch/index/query/MoreLikeThisQueryBuilder.java @@ -512,8 +512,8 @@ public class MoreLikeThisQueryBuilder extends QueryBuilder implements BoostableQ /** * Set the set of stopwords. - * <p/> - * <p>Any word in this set is considered "uninteresting" and ignored. Even if your Analyzer allows stopwords, you + * <p> + * Any word in this set is considered "uninteresting" and ignored. Even if your Analyzer allows stopwords, you * might want to tell the MoreLikeThis code to ignore them, as for the purposes of document similarity it seems * reasonable to assume that "a stop word is never interesting". */ diff --git a/core/src/main/java/org/elasticsearch/index/query/MultiMatchQueryBuilder.java b/core/src/main/java/org/elasticsearch/index/query/MultiMatchQueryBuilder.java index b0949159b2..d42f0c786c 100644 --- a/core/src/main/java/org/elasticsearch/index/query/MultiMatchQueryBuilder.java +++ b/core/src/main/java/org/elasticsearch/index/query/MultiMatchQueryBuilder.java @@ -302,7 +302,7 @@ public class MultiMatchQueryBuilder extends QueryBuilder implements BoostableQue /** - * Set a cutoff value in [0..1] (or absolute number >=1) representing the + * Set a cutoff value in [0..1] (or absolute number >=1) representing the * maximum threshold of a terms document frequency to be considered a low * frequency term. */ diff --git a/core/src/main/java/org/elasticsearch/index/query/QueryBuilders.java b/core/src/main/java/org/elasticsearch/index/query/QueryBuilders.java index aa2dd1d242..f042056b27 100644 --- a/core/src/main/java/org/elasticsearch/index/query/QueryBuilders.java +++ b/core/src/main/java/org/elasticsearch/index/query/QueryBuilders.java @@ -340,7 +340,6 @@ public abstract class QueryBuilders { * * @param multiTermQueryBuilder The {@link MultiTermQueryBuilder} that * backs the created builder. - * @return */ public static SpanMultiTermQueryBuilder spanMultiTermQueryBuilder(MultiTermQueryBuilder multiTermQueryBuilder) { diff --git a/core/src/main/java/org/elasticsearch/index/query/QueryParser.java b/core/src/main/java/org/elasticsearch/index/query/QueryParser.java index d561a0a42b..9553d93bce 100644 --- a/core/src/main/java/org/elasticsearch/index/query/QueryParser.java +++ b/core/src/main/java/org/elasticsearch/index/query/QueryParser.java @@ -38,7 +38,7 @@ public interface QueryParser { /** * Parses the into a query from the current parser location. Will be at "START_OBJECT" location, * and should end when the token is at the matching "END_OBJECT". - * <p/> + * <p> * Returns <tt>null</tt> if this query should be ignored in the context of the DSL. */ @Nullable diff --git a/core/src/main/java/org/elasticsearch/index/query/QueryStringQueryBuilder.java b/core/src/main/java/org/elasticsearch/index/query/QueryStringQueryBuilder.java index c7a297e319..78bfac73f9 100644 --- a/core/src/main/java/org/elasticsearch/index/query/QueryStringQueryBuilder.java +++ b/core/src/main/java/org/elasticsearch/index/query/QueryStringQueryBuilder.java @@ -34,7 +34,6 @@ import java.util.Locale; * will use the {@link #defaultField(String)} set. The second, when one or more fields are added * (using {@link #field(String)}), will run the parsed query against the provided fields, and combine * them either using DisMax or a plain boolean query (see {@link #useDisMax(boolean)}). - * <p/> */ public class QueryStringQueryBuilder extends QueryBuilder implements BoostableQueryBuilder<QueryStringQueryBuilder> { @@ -157,12 +156,12 @@ public class QueryStringQueryBuilder extends QueryBuilder implements BoostableQu /** * Sets the boolean operator of the query parser used to parse the query string. - * <p/> - * <p>In default mode ({@link Operator#OR}) terms without any modifiers + * <p> + * In default mode ({@link Operator#OR}) terms without any modifiers * are considered optional: for example <code>capital of Hungary</code> is equal to * <code>capital OR of OR Hungary</code>. - * <p/> - * <p>In {@link Operator#AND} mode terms are considered to be in conjunction: the + * <p> + * In {@link Operator#AND} mode terms are considered to be in conjunction: the * above mentioned query is parsed as <code>capital AND of AND Hungary</code> */ public QueryStringQueryBuilder defaultOperator(Operator defaultOperator) { @@ -194,7 +193,7 @@ public class QueryStringQueryBuilder extends QueryBuilder implements BoostableQu * when the analyzer returns more than one term from whitespace * delimited text. * NOTE: this behavior may not be suitable for all languages. - * <p/> + * <p> * Set to false if phrase queries should only be generated when * surrounded by double quotes. */ @@ -231,8 +230,8 @@ public class QueryStringQueryBuilder extends QueryBuilder implements BoostableQu /** * Set to <tt>true</tt> to enable position increments in result query. Defaults to * <tt>true</tt>. - * <p/> - * <p>When set, result phrase and multi-phrase queries will be aware of position increments. + * <p> + * When set, result phrase and multi-phrase queries will be aware of position increments. * Useful when e.g. a StopFilter increases the position increment of the token that follows an omitted token. */ public QueryStringQueryBuilder enablePositionIncrements(boolean enablePositionIncrements) { diff --git a/core/src/main/java/org/elasticsearch/index/query/RegexpFlag.java b/core/src/main/java/org/elasticsearch/index/query/RegexpFlag.java index 0af6b86db3..45f58c47da 100644 --- a/core/src/main/java/org/elasticsearch/index/query/RegexpFlag.java +++ b/core/src/main/java/org/elasticsearch/index/query/RegexpFlag.java @@ -89,9 +89,9 @@ public enum RegexpFlag { /** * Resolves the combined OR'ed value for the given list of regular expression flags. The given flags must follow the * following syntax: - * <p/> + * <p> * <tt>flag_name</tt>(|<tt>flag_name</tt>)* - * <p/> + * <p> * Where <tt>flag_name</tt> is one of the following: * <ul> * <li>INTERSECTION</li> @@ -102,10 +102,10 @@ public enum RegexpFlag { * <li>NONE</li> * <li>ALL</li> * </ul> - * <p/> + * <p> * Example: <tt>INTERSECTION|COMPLEMENT|EMPTY</tt> * - * @param flags A string representing a list of regualr expression flags + * @param flags A string representing a list of regular expression flags * @return The combined OR'ed value for all the flags */ static int resolveValue(String flags) { diff --git a/core/src/main/java/org/elasticsearch/index/query/SimpleQueryStringParser.java b/core/src/main/java/org/elasticsearch/index/query/SimpleQueryStringParser.java index a81ffbf348..4207f93fa7 100644 --- a/core/src/main/java/org/elasticsearch/index/query/SimpleQueryStringParser.java +++ b/core/src/main/java/org/elasticsearch/index/query/SimpleQueryStringParser.java @@ -42,7 +42,6 @@ import java.util.Map; * SimpleQueryStringParser is a query parser that acts similar to a query_string * query, but won't throw exceptions for any weird string syntax. It supports * the following: - * <p/> * <ul> * <li>'{@code +}' specifies {@code AND} operation: <tt>token1+token2</tt> * <li>'{@code |}' specifies {@code OR} operation: <tt>token1|token2</tt> @@ -53,14 +52,14 @@ import java.util.Map; * <li>'{@code ~}N' at the end of terms specifies fuzzy query: <tt>term~1</tt> * <li>'{@code ~}N' at the end of phrases specifies near/slop query: <tt>"term1 term2"~5</tt> * </ul> - * <p/> + * <p> * See: {@link SimpleQueryParser} for more information. - * <p/> + * <p> * This query supports these options: - * <p/> + * <p> * Required: * {@code query} - query text to be converted into other queries - * <p/> + * <p> * Optional: * {@code analyzer} - anaylzer to be used for analyzing tokens to determine * which kind of query they should be converted into, defaults to "standard" diff --git a/core/src/main/java/org/elasticsearch/index/query/WrapperQueryBuilder.java b/core/src/main/java/org/elasticsearch/index/query/WrapperQueryBuilder.java index a6e8e23e00..e7de5fd048 100644 --- a/core/src/main/java/org/elasticsearch/index/query/WrapperQueryBuilder.java +++ b/core/src/main/java/org/elasticsearch/index/query/WrapperQueryBuilder.java @@ -29,14 +29,14 @@ import java.io.IOException; * A Query builder which allows building a query given JSON string or binary data provided as input. This is useful when you want * to use the Java Builder API but still have JSON query strings at hand that you want to combine with other * query builders. - * <p/> + * <p> * Example usage in a boolean query : * <pre> - * {@code + * <code> * BoolQueryBuilder bool = new BoolQueryBuilder(); * bool.must(new WrapperQueryBuilder("{\"term\": {\"field\":\"value\"}}"); * bool.must(new TermQueryBuilder("field2","value2"); - * } + * </code> * </pre> */ public class WrapperQueryBuilder extends QueryBuilder { diff --git a/core/src/main/java/org/elasticsearch/index/query/functionscore/DecayFunctionParser.java b/core/src/main/java/org/elasticsearch/index/query/functionscore/DecayFunctionParser.java index bfe988a268..6e74959ab3 100644 --- a/core/src/main/java/org/elasticsearch/index/query/functionscore/DecayFunctionParser.java +++ b/core/src/main/java/org/elasticsearch/index/query/functionscore/DecayFunctionParser.java @@ -60,14 +60,14 @@ import java.util.Locale; * This parser parses this kind of input * * <pre> - * {@code} + * <code> * { * "fieldname1" : { * "origin" = "someValue", * "scale" = "someValue" * } * - * } + * </code> * </pre> * * "origin" here refers to the reference point and "scale" to the level of @@ -106,7 +106,7 @@ public abstract class DecayFunctionParser implements ScoreFunctionParser { * Parses bodies of the kind * * <pre> - * {@code} + * <code> * { * "fieldname1" : { * "origin" = "someValue", @@ -114,6 +114,7 @@ public abstract class DecayFunctionParser implements ScoreFunctionParser { * } * * } + * </code> * </pre> * * */ diff --git a/core/src/main/java/org/elasticsearch/index/query/functionscore/random/RandomScoreFunctionBuilder.java b/core/src/main/java/org/elasticsearch/index/query/functionscore/random/RandomScoreFunctionBuilder.java index ea2293c55a..22285f8cc6 100644 --- a/core/src/main/java/org/elasticsearch/index/query/functionscore/random/RandomScoreFunctionBuilder.java +++ b/core/src/main/java/org/elasticsearch/index/query/functionscore/random/RandomScoreFunctionBuilder.java @@ -51,7 +51,7 @@ public class RandomScoreFunctionBuilder extends ScoreFunctionBuilder { /** * seed variant taking a long value. - * @see {@link #seed(int)} + * @see #seed(int) */ public RandomScoreFunctionBuilder seed(long seed) { this.seed = seed; @@ -60,7 +60,7 @@ public class RandomScoreFunctionBuilder extends ScoreFunctionBuilder { /** * seed variant taking a String value. - * @see {@link #seed(int)} + * @see #seed(int) */ public RandomScoreFunctionBuilder seed(String seed) { this.seed = seed; @@ -78,4 +78,4 @@ public class RandomScoreFunctionBuilder extends ScoreFunctionBuilder { builder.endObject(); } -}
\ No newline at end of file +} diff --git a/core/src/main/java/org/elasticsearch/index/shard/MergePolicyConfig.java b/core/src/main/java/org/elasticsearch/index/shard/MergePolicyConfig.java index 3895bbed2c..c664d3a379 100644 --- a/core/src/main/java/org/elasticsearch/index/shard/MergePolicyConfig.java +++ b/core/src/main/java/org/elasticsearch/index/shard/MergePolicyConfig.java @@ -79,7 +79,7 @@ import org.elasticsearch.index.settings.IndexSettingsService; * * Sets the allowed number of segments per tier. Smaller values mean more * merging but fewer segments. Default is <code>10</code>. Note, this value needs to be - * >= than the <code>max_merge_at_once</code> otherwise you'll force too many merges to + * >= than the <code>max_merge_at_once</code> otherwise you'll force too many merges to * occur. * * <li><code>index.merge.policy.reclaim_deletes_weight</code>: diff --git a/core/src/main/java/org/elasticsearch/index/similarity/BM25SimilarityProvider.java b/core/src/main/java/org/elasticsearch/index/similarity/BM25SimilarityProvider.java index ca24985516..1983c4e8ec 100644 --- a/core/src/main/java/org/elasticsearch/index/similarity/BM25SimilarityProvider.java +++ b/core/src/main/java/org/elasticsearch/index/similarity/BM25SimilarityProvider.java @@ -27,7 +27,7 @@ import org.elasticsearch.common.settings.Settings; /** * {@link SimilarityProvider} for the {@link BM25Similarity}. - * <p/> + * <p> * Configuration options available: * <ul> * <li>k1</li> diff --git a/core/src/main/java/org/elasticsearch/index/similarity/DFRSimilarityProvider.java b/core/src/main/java/org/elasticsearch/index/similarity/DFRSimilarityProvider.java index 6d30e81c09..b5a5cdcfef 100644 --- a/core/src/main/java/org/elasticsearch/index/similarity/DFRSimilarityProvider.java +++ b/core/src/main/java/org/elasticsearch/index/similarity/DFRSimilarityProvider.java @@ -28,7 +28,7 @@ import org.elasticsearch.common.settings.Settings; /** * {@link SimilarityProvider} for {@link DFRSimilarity}. - * <p/> + * <p> * Configuration options available: * <ul> * <li>basic_model</li> diff --git a/core/src/main/java/org/elasticsearch/index/similarity/DefaultSimilarityProvider.java b/core/src/main/java/org/elasticsearch/index/similarity/DefaultSimilarityProvider.java index a6e04434b3..0f9feba952 100644 --- a/core/src/main/java/org/elasticsearch/index/similarity/DefaultSimilarityProvider.java +++ b/core/src/main/java/org/elasticsearch/index/similarity/DefaultSimilarityProvider.java @@ -26,7 +26,7 @@ import org.elasticsearch.common.settings.Settings; /** * {@link SimilarityProvider} for {@link DefaultSimilarity}. - * <p/> + * <p> * Configuration options available: * <ul> * <li>discount_overlaps</li> diff --git a/core/src/main/java/org/elasticsearch/index/similarity/IBSimilarityProvider.java b/core/src/main/java/org/elasticsearch/index/similarity/IBSimilarityProvider.java index 4741247080..161ca9c10e 100644 --- a/core/src/main/java/org/elasticsearch/index/similarity/IBSimilarityProvider.java +++ b/core/src/main/java/org/elasticsearch/index/similarity/IBSimilarityProvider.java @@ -28,7 +28,7 @@ import org.elasticsearch.common.settings.Settings; /** * {@link SimilarityProvider} for {@link IBSimilarity}. - * <p/> + * <p> * Configuration options available: * <ul> * <li>distribution</li> diff --git a/core/src/main/java/org/elasticsearch/index/similarity/LMDirichletSimilarityProvider.java b/core/src/main/java/org/elasticsearch/index/similarity/LMDirichletSimilarityProvider.java index 797ce6417e..efea285639 100644 --- a/core/src/main/java/org/elasticsearch/index/similarity/LMDirichletSimilarityProvider.java +++ b/core/src/main/java/org/elasticsearch/index/similarity/LMDirichletSimilarityProvider.java @@ -27,7 +27,7 @@ import org.elasticsearch.common.settings.Settings; /** * {@link SimilarityProvider} for {@link LMDirichletSimilarity}. - * <p/> + * <p> * Configuration options available: * <ul> * <li>mu</li> diff --git a/core/src/main/java/org/elasticsearch/index/similarity/LMJelinekMercerSimilarityProvider.java b/core/src/main/java/org/elasticsearch/index/similarity/LMJelinekMercerSimilarityProvider.java index 9be02366b6..5d30b300d5 100644 --- a/core/src/main/java/org/elasticsearch/index/similarity/LMJelinekMercerSimilarityProvider.java +++ b/core/src/main/java/org/elasticsearch/index/similarity/LMJelinekMercerSimilarityProvider.java @@ -27,7 +27,7 @@ import org.elasticsearch.common.settings.Settings; /** * {@link SimilarityProvider} for {@link LMJelinekMercerSimilarity}. - * <p/> + * <p> * Configuration options available: * <ul> * <li>lambda</li> diff --git a/core/src/main/java/org/elasticsearch/index/similarity/SimilarityLookupService.java b/core/src/main/java/org/elasticsearch/index/similarity/SimilarityLookupService.java index 49e3df989f..8b8b68812c 100644 --- a/core/src/main/java/org/elasticsearch/index/similarity/SimilarityLookupService.java +++ b/core/src/main/java/org/elasticsearch/index/similarity/SimilarityLookupService.java @@ -31,7 +31,7 @@ import java.util.Map; /** * Service for looking up configured {@link SimilarityProvider} implementations by name. - * <p/> + * <p> * The service instantiates the Providers through their Factories using configuration * values found with the {@link SimilarityModule#SIMILARITY_SETTINGS_PREFIX} prefix. */ diff --git a/core/src/main/java/org/elasticsearch/index/snapshots/IndexShardRepository.java b/core/src/main/java/org/elasticsearch/index/snapshots/IndexShardRepository.java index 7c77884692..8ce487fe14 100644 --- a/core/src/main/java/org/elasticsearch/index/snapshots/IndexShardRepository.java +++ b/core/src/main/java/org/elasticsearch/index/snapshots/IndexShardRepository.java @@ -27,7 +27,7 @@ import org.elasticsearch.indices.recovery.RecoveryState; /** * Shard-level snapshot repository - * <p/> + * <p> * IndexShardRepository is used on data node to create snapshots of individual shards. See {@link org.elasticsearch.repositories.Repository} * for more information. */ @@ -35,10 +35,10 @@ public interface IndexShardRepository { /** * Creates a snapshot of the shard based on the index commit point. - * <p/> + * <p> * The index commit point can be obtained by using {@link org.elasticsearch.index.engine.Engine#snapshotIndex} method. * IndexShardRepository implementations shouldn't release the snapshot index commit point. It is done by the method caller. - * <p/> + * <p> * As snapshot process progresses, implementation of this method should update {@link IndexShardSnapshotStatus} object and check * {@link IndexShardSnapshotStatus#aborted()} to see if the snapshot process should be aborted. * @@ -51,7 +51,7 @@ public interface IndexShardRepository { /** * Restores snapshot of the shard. - * <p/> + * <p> * The index can be renamed on restore, hence different {@code shardId} and {@code snapshotShardId} are supplied. * * @param snapshotId snapshot id diff --git a/core/src/main/java/org/elasticsearch/index/snapshots/blobstore/BlobStoreIndexShardRepository.java b/core/src/main/java/org/elasticsearch/index/snapshots/blobstore/BlobStoreIndexShardRepository.java index 947a36a6c1..912be76fb8 100644 --- a/core/src/main/java/org/elasticsearch/index/snapshots/blobstore/BlobStoreIndexShardRepository.java +++ b/core/src/main/java/org/elasticsearch/index/snapshots/blobstore/BlobStoreIndexShardRepository.java @@ -623,12 +623,11 @@ public class BlobStoreIndexShardRepository extends AbstractComponent implements /** * Snapshot individual file - * <p/> + * <p> * This is asynchronous method. Upon completion of the operation latch is getting counted down and any failures are * added to the {@code failures} list * * @param fileInfo file to be snapshotted - * @throws IOException */ private void snapshotFile(final BlobStoreIndexShardSnapshot.FileInfo fileInfo) throws IOException { final String file = fileInfo.physicalName(); diff --git a/core/src/main/java/org/elasticsearch/index/snapshots/blobstore/BlobStoreIndexShardSnapshot.java b/core/src/main/java/org/elasticsearch/index/snapshots/blobstore/BlobStoreIndexShardSnapshot.java index 808f13ba23..56d9882063 100644 --- a/core/src/main/java/org/elasticsearch/index/snapshots/blobstore/BlobStoreIndexShardSnapshot.java +++ b/core/src/main/java/org/elasticsearch/index/snapshots/blobstore/BlobStoreIndexShardSnapshot.java @@ -228,7 +228,6 @@ public class BlobStoreIndexShardSnapshot implements ToXContent, FromXContentBuil * @param file file info * @param builder XContent builder * @param params parameters - * @throws IOException */ public static void toXContent(FileInfo file, XContentBuilder builder, ToXContent.Params params) throws IOException { builder.startObject(); @@ -257,7 +256,6 @@ public class BlobStoreIndexShardSnapshot implements ToXContent, FromXContentBuil * * @param parser parser * @return file info - * @throws IOException */ public static FileInfo fromXContent(XContentParser parser) throws IOException { XContentParser.Token token = parser.currentToken(); @@ -446,7 +444,6 @@ public class BlobStoreIndexShardSnapshot implements ToXContent, FromXContentBuil * * @param builder XContent builder * @param params parameters - * @throws IOException */ @Override public XContentBuilder toXContent(XContentBuilder builder, Params params) throws IOException { @@ -469,7 +466,6 @@ public class BlobStoreIndexShardSnapshot implements ToXContent, FromXContentBuil * * @param parser parser * @return shard snapshot metadata - * @throws IOException */ public BlobStoreIndexShardSnapshot fromXContent(XContentParser parser, ParseFieldMatcher parseFieldMatcher) throws IOException { diff --git a/core/src/main/java/org/elasticsearch/index/snapshots/blobstore/BlobStoreIndexShardSnapshots.java b/core/src/main/java/org/elasticsearch/index/snapshots/blobstore/BlobStoreIndexShardSnapshots.java index 268e681e93..bf8c1a83c9 100644 --- a/core/src/main/java/org/elasticsearch/index/snapshots/blobstore/BlobStoreIndexShardSnapshots.java +++ b/core/src/main/java/org/elasticsearch/index/snapshots/blobstore/BlobStoreIndexShardSnapshots.java @@ -40,7 +40,7 @@ import java.util.Map; /** * Contains information about all snapshot for the given shard in repository - * <p/> + * <p> * This class is used to find files that were already snapshoted and clear out files that no longer referenced by any * snapshots */ @@ -160,7 +160,7 @@ public class BlobStoreIndexShardSnapshots implements Iterable<SnapshotFiles>, To /** * Writes index file for the shard in the following format. * <pre> - * {@code + * <code> * { * "files": [{ * "name": "__3", @@ -206,6 +206,7 @@ public class BlobStoreIndexShardSnapshots implements Iterable<SnapshotFiles>, To * } * } * } + * </code> * </pre> */ @Override diff --git a/core/src/main/java/org/elasticsearch/index/store/Store.java b/core/src/main/java/org/elasticsearch/index/store/Store.java index 2b5051b412..09f8ec23ad 100644 --- a/core/src/main/java/org/elasticsearch/index/store/Store.java +++ b/core/src/main/java/org/elasticsearch/index/store/Store.java @@ -70,7 +70,7 @@ import java.util.zip.Checksum; * This class also provides access to metadata information like checksums for committed files. A committed * file is a file that belongs to a segment written by a Lucene commit. Files that have not been committed * ie. created during a merge or a shard refresh / NRT reopen are not considered in the MetadataSnapshot. - * <p/> + * <p> * Note: If you use a store it's reference count should be increased before using it by calling #incRef and a * corresponding #decRef must be called in a try/finally block to release the store again ie.: * <pre> @@ -299,7 +299,7 @@ public class Store extends AbstractIndexShardComponent implements Closeable, Ref * corresponding {@link #decRef}, in a finally clause; otherwise the store may never be closed. Note that * {@link #close} simply calls decRef(), which means that the Store will not really be closed until {@link * #decRef} has been called for all outstanding references. - * <p/> + * <p> * Note: Close can safely be called multiple times. * * @throws AlreadyClosedException iff the reference counter can not be incremented. @@ -318,7 +318,7 @@ public class Store extends AbstractIndexShardComponent implements Closeable, Ref * corresponding {@link #decRef}, in a finally clause; otherwise the store may never be closed. Note that * {@link #close} simply calls decRef(), which means that the Store will not really be closed until {@link * #decRef} has been called for all outstanding references. - * <p/> + * <p> * Note: Close can safely be called multiple times. * * @see #decRef() @@ -413,7 +413,7 @@ public class Store extends AbstractIndexShardComponent implements Closeable, Ref * The returned IndexOutput might validate the files checksum if the file has been written with a newer lucene version * and the metadata holds the necessary information to detect that it was been written by Lucene 4.8 or newer. If it has only * a legacy checksum, returned IndexOutput will not verify the checksum. - * <p/> + * <p> * Note: Checksums are calculated nevertheless since lucene does it by default sicne version 4.8.0. This method only adds the * verification against the checksum in the given metadata and does not add any significant overhead. */ @@ -723,7 +723,7 @@ public class Store extends AbstractIndexShardComponent implements Closeable, Ref * Only files that are part of the last commit are considered in this datastrucutre. * For backwards compatibility the snapshot might include legacy checksums that * are derived from a dedicated checksum file written by older elasticsearch version pre 1.3 - * <p/> + * <p> * Note: This class will ignore the <tt>segments.gen</tt> file since it's optional and might * change concurrently for safety reasons. * @@ -867,13 +867,12 @@ public class Store extends AbstractIndexShardComponent implements Closeable, Ref /** * Reads legacy checksum files found in the directory. - * <p/> + * <p> * Files are expected to start with _checksums- prefix * followed by long file version. Only file with the highest version is read, all other files are ignored. * * @param directory the directory to read checksums from * @return a map of file checksums and the checksum file version - * @throws IOException */ static Tuple<Map<String, String>, Long> readLegacyChecksums(Directory directory) throws IOException { synchronized (directory) { @@ -902,7 +901,6 @@ public class Store extends AbstractIndexShardComponent implements Closeable, Ref * * @param directory the directory to clean * @param newVersion the latest checksum file version - * @throws IOException */ static void cleanLegacyChecksums(Directory directory, long newVersion) throws IOException { synchronized (directory) { @@ -949,7 +947,7 @@ public class Store extends AbstractIndexShardComponent implements Closeable, Ref } /** - * Computes a strong hash value for small files. Note that this method should only be used for files < 1MB + * Computes a strong hash value for small files. Note that this method should only be used for files < 1MB */ public static BytesRef hashFile(Directory directory, String file) throws IOException { final BytesRefBuilder fileHash = new BytesRefBuilder(); @@ -961,7 +959,7 @@ public class Store extends AbstractIndexShardComponent implements Closeable, Ref /** - * Computes a strong hash value for small files. Note that this method should only be used for files < 1MB + * Computes a strong hash value for small files. Note that this method should only be used for files < 1MB */ public static void hashFile(BytesRefBuilder fileHash, InputStream in, long size) throws IOException { final int len = (int) Math.min(1024 * 1024, size); // for safety we limit this to 1MB @@ -1006,10 +1004,10 @@ public class Store extends AbstractIndexShardComponent implements Closeable, Ref * <li>all files in this segment have the same length</li> * <li>the segments <tt>.si</tt> files hashes are byte-identical Note: This is a using a perfect hash function, The metadata transfers the <tt>.si</tt> file content as it's hash</li> * </ul> - * <p/> + * <p> * The <tt>.si</tt> file contains a lot of diagnostics including a timestamp etc. in the future there might be * unique segment identifiers in there hardening this method further. - * <p/> + * <p> * The per-commit files handles very similar. A commit is composed of the <tt>segments_N</tt> files as well as generational files like * deletes (<tt>_x_y.del</tt>) or field-info (<tt>_x_y.fnm</tt>) files. On a per-commit level files for a commit are treated * as identical iff: @@ -1018,7 +1016,7 @@ public class Store extends AbstractIndexShardComponent implements Closeable, Ref * <li>all files belonging to this commit have the same length</li> * <li>the segments file <tt>segments_N</tt> files hashes are byte-identical Note: This is a using a perfect hash function, The metadata transfers the <tt>segments_N</tt> file content as it's hash</li> * </ul> - * <p/> + * <p> * NOTE: this diff will not contain the <tt>segments.gen</tt> file. This file is omitted on recovery. */ public RecoveryDiff recoveryDiff(MetadataSnapshot recoveryTargetSnapshot) { @@ -1314,7 +1312,7 @@ public class Store extends AbstractIndexShardComponent implements Closeable, Ref /** * Index input that calculates checksum as data is read from the input. - * <p/> + * <p> * This class supports random access (it is possible to seek backward and forward) in order to accommodate retry * mechanism that is used in some repository plugins (S3 for example). However, the checksum is only calculated on * the first read. All consecutive reads of the same data are not used to calculate the checksum. diff --git a/core/src/main/java/org/elasticsearch/index/store/StoreFileMetaData.java b/core/src/main/java/org/elasticsearch/index/store/StoreFileMetaData.java index 9aaea73efb..31ad068a58 100644 --- a/core/src/main/java/org/elasticsearch/index/store/StoreFileMetaData.java +++ b/core/src/main/java/org/elasticsearch/index/store/StoreFileMetaData.java @@ -87,7 +87,6 @@ public class StoreFileMetaData implements Streamable { * Returns a string representation of the files checksum. Since Lucene 4.8 this is a CRC32 checksum written * by lucene. Previously we use Adler32 on top of Lucene as the checksum algorithm, if {@link #hasLegacyChecksum()} returns * <code>true</code> this is a Adler32 checksum. - * @return */ @Nullable public String checksum() { diff --git a/core/src/main/java/org/elasticsearch/index/translog/LegacyTranslogReader.java b/core/src/main/java/org/elasticsearch/index/translog/LegacyTranslogReader.java index 7a131f9d0d..d1cd3b1efd 100644 --- a/core/src/main/java/org/elasticsearch/index/translog/LegacyTranslogReader.java +++ b/core/src/main/java/org/elasticsearch/index/translog/LegacyTranslogReader.java @@ -32,9 +32,6 @@ public final class LegacyTranslogReader extends LegacyTranslogReaderBase { /** * Create a snapshot of translog file channel. The length parameter should be consistent with totalOperations and point * at the end of the last operation in this snapshot. - * - * @param generation - * @param channelReference */ LegacyTranslogReader(long generation, ChannelReference channelReference, long fileLength) { super(generation, channelReference, 0, fileLength); diff --git a/core/src/main/java/org/elasticsearch/index/translog/Translog.java b/core/src/main/java/org/elasticsearch/index/translog/Translog.java index 9b1913e585..5084895151 100644 --- a/core/src/main/java/org/elasticsearch/index/translog/Translog.java +++ b/core/src/main/java/org/elasticsearch/index/translog/Translog.java @@ -1764,8 +1764,6 @@ public class Translog extends AbstractIndexShardComponent implements IndexShardC /** * Returns <code>true</code> iff the given generation is the current gbeneration of this translog - * @param generation - * @return */ public boolean isCurrent(TranslogGeneration generation) { try (ReleasableLock lock = writeLock.acquire()) { diff --git a/core/src/main/java/org/elasticsearch/index/translog/TranslogReader.java b/core/src/main/java/org/elasticsearch/index/translog/TranslogReader.java index c7feb8d19c..590bc31905 100644 --- a/core/src/main/java/org/elasticsearch/index/translog/TranslogReader.java +++ b/core/src/main/java/org/elasticsearch/index/translog/TranslogReader.java @@ -170,9 +170,6 @@ public abstract class TranslogReader implements Closeable, Comparable<TranslogRe * optionally-existing header in the file. If the file does not exist, or * has zero length, returns the latest version. If the header does not * exist, assumes Version 0 of the translog file format. - * <p/> - * - * @throws IOException */ public static ImmutableTranslogReader open(ChannelReference channelReference, Checkpoint checkpoint, String translogUUID) throws IOException { final FileChannel channel = channelReference.getChannel(); |