diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/BufferedMutatorParams.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/BufferedMutatorParams.java index 9044cdbb36ff..e71a1d3b2744 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/BufferedMutatorParams.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/BufferedMutatorParams.java @@ -76,7 +76,7 @@ public BufferedMutatorParams operationTimeout(final int operationTimeout) { } /** - * @deprecated Since 2.3.0, will be removed in 4.0.0. Use {@link #operationTimeout()} + * @deprecated Since 2.3.0, will be removed in 4.0.0. Use {@link #operationTimeout(int)} */ @Deprecated public BufferedMutatorParams opertationTimeout(final int operationTimeout) { diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/SimpleRequestController.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/SimpleRequestController.java index 4981d62a605b..d87014428c3b 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/SimpleRequestController.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/SimpleRequestController.java @@ -431,9 +431,10 @@ private void waitForRegion() throws InterruptedIOException { * regions. 3) check the total concurrent tasks. 4) check the concurrent * tasks for server. * - * @param loc - * @param heapSizeOfRow - * @return either Include {@link ReturnCode} or Skip {@link ReturnCode} + * @param loc the destination of data + * @param heapSizeOfRow the data size + * @return either Include {@link RequestController.ReturnCode} or skip + * {@link RequestController.ReturnCode} */ @Override public ReturnCode canTakeOperation(HRegionLocation loc, long heapSizeOfRow) { diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Table.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Table.java index def977446267..41b0e47c0758 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Table.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/Table.java @@ -176,9 +176,9 @@ default Result get(Get get) throws IOException { * @throws IOException if a remote or network exception occurs. * @since 0.90.0 * @apiNote {@link #put(List)} runs pre-flight validations on the input list on client. - * Currently {@link #get(List)} doesn't run any validations on the client-side, currently there - * is no need, but this may change in the future. An - * {@link IllegalArgumentException} will be thrown in this case. + * Currently {@link #get(List)} doesn't run any validations on the client-side, + * currently there is no need, but this may change in the future. An + * {@link IllegalArgumentException} will be thrown in this case. */ default Result[] get(List gets) throws IOException { throw new NotImplementedException("Add an implementation!"); @@ -284,10 +284,10 @@ default void delete(Delete delete) throws IOException { * that have not be successfully applied. * @since 0.20.1 * @apiNote In 3.0.0 version, the input list {@code deletes} will no longer be modified. Also, - * {@link #put(List)} runs pre-flight validations on the input list on client. Currently - * {@link #delete(List)} doesn't run validations on the client, there is no need currently, - * but this may change in the future. An * {@link IllegalArgumentException} will be thrown - * in this case. + * {@link #put(List)} runs pre-flight validations on the input list on client. Currently + * {@link #delete(List)} doesn't run validations on the client, there is no need + * currently, but this may change in the future. An * {@link IllegalArgumentException} + * will be thrown in this case. */ default void delete(List deletes) throws IOException { throw new NotImplementedException("Add an implementation!"); diff --git a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Batch.java b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Batch.java index 82b005d018c1..c3defda5b4d3 100644 --- a/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Batch.java +++ b/hbase-client/src/main/java/org/apache/hadoop/hbase/client/coprocessor/Batch.java @@ -39,7 +39,7 @@ public abstract class Batch { * the implementations {@link Batch.Call#call(Object)} method will be invoked * with a proxy to each region's coprocessor {@link com.google.protobuf.Service} implementation. *

- * @see org.apache.hadoop.hbase.client.coprocessor + * @see org.apache.hadoop.hbase.client.coprocessor.Batch * @see org.apache.hadoop.hbase.client.Table#coprocessorService(byte[]) * @see org.apache.hadoop.hbase.client.Table#coprocessorService(Class, byte[], byte[], * org.apache.hadoop.hbase.client.coprocessor.Batch.Call) diff --git a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java index 6907a8817cc5..4993feea223e 100644 --- a/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java +++ b/hbase-mapreduce/src/main/java/org/apache/hadoop/hbase/snapshot/ExportSnapshot.java @@ -267,9 +267,10 @@ private Path getOutputPath(final SnapshotFileInfo inputInfo) throws IOException return new Path(outputArchive, path); } + @SuppressWarnings("checkstyle:linelength") /** * Used by TestExportSnapshot to test for retries when failures happen. - * Failure is injected in {@link #copyFile(Context, SnapshotFileInfo, Path)}. + * Failure is injected in {@link #copyFile(Mapper.Context, org.apache.hadoop.hbase.shaded.protobuf.generated.SnapshotProtos.SnapshotFileInfo, Path)}. */ private void injectTestFailure(final Context context, final SnapshotFileInfo inputInfo) throws IOException { diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.java index 82767e963c74..4a6b006c8607 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/io/hfile/bucket/BucketCache.java @@ -1529,11 +1529,11 @@ static class RAMCache { /** * Defined the map as {@link ConcurrentHashMap} explicitly here, because in * {@link RAMCache#get(BlockCacheKey)} and - * {@link RAMCache#putIfAbsent(BlockCacheKey, RAMQueueEntry)} , we need to guarantee the - * atomicity of map#computeIfPresent(key, func) and map#putIfAbsent(key, func). Besides, the - * func method can execute exactly once only when the key is present(or absent) and under the - * lock context. Otherwise, the reference count of block will be messed up. Notice that the - * {@link java.util.concurrent.ConcurrentSkipListMap} can not guarantee that. + * {@link RAMCache#putIfAbsent(BlockCacheKey, BucketCache.RAMQueueEntry)} , we need to + * guarantee the atomicity of map#computeIfPresent(key, func) and map#putIfAbsent(key, func). + * Besides, the func method can execute exactly once only when the key is present(or absent) + * and under the lock context. Otherwise, the reference count of block will be messed up. + * Notice that the {@link java.util.concurrent.ConcurrentSkipListMap} can not guarantee that. */ final ConcurrentHashMap delegate = new ConcurrentHashMap<>(); diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStateNode.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStateNode.java index 7dbdbee42ac5..12eb0a0fe4ff 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStateNode.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/RegionStateNode.java @@ -99,7 +99,7 @@ public AssignmentProcedureEvent(final RegionInfo regionInfo) { /** * Updated whenever a call to {@link #setRegionLocation(ServerName)} or - * {@link #setState(State, State...)}. + * {@link #setState(RegionState.State, RegionState.State...)}. */ private volatile long lastUpdate = 0; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java index 9f31522d7644..f3bf271f7007 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/regionserver/HRegion.java @@ -4047,7 +4047,7 @@ public OperationStatus[] batchReplay(MutationReplay[] mutations, long replaySeqI * a batch are stored with highest durability specified of for all operations in a batch, * except for {@link Durability#SKIP_WAL}. * - *

This function is called from {@link #batchReplay(MutationReplay[], long)} with + *

This function is called from {@link #batchReplay(WALSplitUtil.MutationReplay[], long)} with * {@link ReplayBatchOperation} instance and {@link #batchMutate(Mutation[], long, long)} with * {@link MutationBatchOperation} instance as an argument. As the processing of replay batch * and mutation batch is very similar, lot of code is shared by providing generic methods in @@ -4058,7 +4058,7 @@ public OperationStatus[] batchReplay(MutationReplay[] mutations, long replaySeqI * @param batchOp contains the list of mutations * @return an array of OperationStatus which internally contains the * OperationStatusCode and the exceptionMessage if any. - * @throws IOException + * @throws IOException if an IO problem is encountered */ OperationStatus[] batchMutate(BatchOperation batchOp) throws IOException { boolean initialized = false; diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.java index abf1fd831f9d..46a0669e5c52 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/tool/BulkLoadHFilesTool.java @@ -453,7 +453,8 @@ private boolean checkHFilesCountPerRegionPerFamily( } /** - * @param table the table to load into + * @param conn the HBase cluster connection + * @param tableName the table name of the table to load into * @param pool the ExecutorService * @param queue the queue for LoadQueueItem * @param startEndKeys start and end keys diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java index 09f93e9edaa5..973cfb0a85e6 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java @@ -2039,16 +2039,16 @@ public HRegion createLocalHRegion(RegionInfo info, TableDescriptor desc, WAL wal } /** - * @param tableName - * @param startKey - * @param stopKey - * @param callingMethod - * @param conf - * @param isReadOnly - * @param families - * @throws IOException - * @return A region on which you must call - {@link HBaseTestingUtility#closeRegionAndWAL(HRegion)} when done. + * @param tableName the name of the table + * @param startKey the start key of the region + * @param stopKey the stop key of the region + * @param callingMethod the name of the calling method probably a test method + * @param conf the configuration to use + * @param isReadOnly {@code true} if the table is read only, {@code false} otherwise + * @param families the column families to use + * @throws IOException if an IO problem is encountered + * @return A region on which you must call {@link HBaseTestingUtility#closeRegionAndWAL(HRegion)} + * when done. * @deprecated since 2.0.0 and will be removed in 3.0.0. Use * {@link #createLocalHRegion(TableName, byte[], byte[], boolean, Durability, WAL, byte[]...)} * instead. @@ -3292,10 +3292,9 @@ public boolean visit(Result r) throws IOException { * Waits for a table to be 'enabled'. Enabled means that table is set as 'enabled' and the * regions have been all assigned. Will timeout after default period (30 seconds) * Tolerates nonexistent table. - * @param table Table to wait on. - * @param table - * @throws InterruptedException - * @throws IOException + * @param table the table to wait on. + * @throws InterruptedException if interrupted while waiting + * @throws IOException if an IO problem is encountered */ public void waitTableEnabled(TableName table) throws InterruptedException, IOException { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java index 211efc0183cc..be0ece469549 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/MockRegionServer.java @@ -153,8 +153,8 @@ class MockRegionServer implements AdminProtos.AdminService.BlockingInterface, private final Random random = new Random(); /** - * Map of regions to map of rows and {@link Result}. Used as data source when - * {@link #get(RpcController, GetRequest)} is called. Because we have a byte + * Map of regions to map of rows and {@link Result}. Used as data source when + * {@link #get(RpcController, ClientProtos.GetRequest)} is called. Because we have a byte * key, need to use TreeMap and provide a Comparator. Use * {@link #setGetResult(byte[], byte[], Result)} filling this map. */ @@ -205,10 +205,11 @@ int getThenIncrement() { } /** - * Use this method filling the backing data source used by {@link #get(RpcController, GetRequest)} - * @param regionName - * @param row - * @param r + * Use this method filling the backing data source used by + * {@link #get(RpcController, ClientProtos.GetRequest)} + * @param regionName the region name to assign + * @param row the row key + * @param r the single row result */ void setGetResult(final byte [] regionName, final byte [] row, final Result r) { Map value = this.gets.get(regionName); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java index 1d81dd757d6c..df03313f4461 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/snapshot/TestFlushSnapshotFromClient.java @@ -284,7 +284,7 @@ public void testSnapshotFailsOnNonExistantTable() throws Exception { * complete on the server by repeatedly checking the master. * @param master the master running the snapshot * @param snapshot the snapshot to check - * @param sleep amount to sleep between checks to see if the snapshot is done + * @param timeoutNanos the timeout in nano between checks to see if the snapshot is done */ private static void waitForSnapshotToComplete(HMaster master, SnapshotProtos.SnapshotDescription snapshot, long timeoutNanos) throws Exception { diff --git a/pom.xml b/pom.xml index b00f9bbfc734..72c59bb48aec 100755 --- a/pom.xml +++ b/pom.xml @@ -3761,6 +3761,15 @@ org.apache.maven.plugins maven-javadoc-plugin + + + + apiNote + a + API Note: + + +