Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
The table of contents is too big for display.
Diff view
Diff view
  •  
  •  
  •  
Original file line number Diff line number Diff line change
Expand Up @@ -368,7 +368,7 @@ private static class BuilderPayloadSetter {
* Create a ByteString from byte array without copying (wrap), and then set it as the payload
* for the builder.
* @param builder builder for HDFS DataTransferEncryptorMessage.
* @param payload byte array of payload. n
* @param payload byte array of payload.
*/
static void wrapAndSetPayload(DataTransferEncryptorMessageProto.Builder builder,
byte[] payload) throws IOException {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -165,7 +165,7 @@ public static boolean isBackupEnabled(Configuration conf) {
}

/**
* Get configuration n
* Get configuration
*/
Configuration getConf() {
return conf;
Expand All @@ -192,7 +192,8 @@ public void close() {
* @param tableList table list
* @param targetRootDir root dir
* @param workers number of parallel workers
* @param bandwidth bandwidth per worker in MB per sec n * @throws BackupException exception
* @param bandwidth bandwidth per worker in MB per sec
* @throws BackupException exception
*/
public BackupInfo createBackupInfo(String backupId, BackupType type, List<TableName> tableList,
String targetRootDir, int workers, long bandwidth) throws BackupException {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -184,9 +184,7 @@ public static ServerName[] getFavoredNodesList(byte[] favoredNodes) throws IOExc
return servers;
}

/**
* n * @return PB'ed bytes of {@link FavoredNodes} generated by the server list.
*/
/** Returns PB'ed bytes of {@link FavoredNodes} generated by the server list. */
public static byte[] getFavoredNodes(List<ServerName> serverAddrList) {
FavoredNodes.Builder f = FavoredNodes.newBuilder();
for (ServerName s : serverAddrList) {
Expand Down Expand Up @@ -319,8 +317,8 @@ public ServerName[] getSecondaryAndTertiary(RegionInfo regionInfo, ServerName pr

/**
* For regions that share the primary, avoid placing the secondary and tertiary on a same RS. Used
* for generating new assignments for the primary/secondary/tertiary RegionServers n * @return the
* map of regions to the servers the region-files should be hosted on
* for generating new assignments for the primary/secondary/tertiary RegionServers
* @return the map of regions to the servers the region-files should be hosted on
*/
public Map<RegionInfo, ServerName[]>
placeSecondaryAndTertiaryWithRestrictions(Map<RegionInfo, ServerName> primaryRSMap) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -80,7 +80,7 @@ public List<ServerName> getFavoredNodes(RegionInfo region) {

/**
* Return the position of the server in the favoredNodes list. Assumes the favoredNodes list is of
* size 3. n
* size 3.
*/
public static Position getFavoredServerPosition(List<ServerName> favoredNodes,
ServerName server) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -294,7 +294,7 @@ public void fillUp(TableName tableName, SnapshotOfRegionAssignmentFromMeta snaps
}

/**
* Use this to project the dispersion scores nnn
* Use this to project the dispersion scores
*/
public void fillUpDispersion(TableName tableName, SnapshotOfRegionAssignmentFromMeta snapshot,
FavoredNodesPlan newPlan) {
Expand Down Expand Up @@ -566,7 +566,8 @@ int getTotalFavoredAssignments() {

/**
* Return the number of regions based on the position (primary/secondary/ tertiary) assigned to
* their favored nodes n * @return the number of regions
* their favored nodes
* @return the number of regions
*/
int getNumRegionsOnFavoredNodeByPosition(FavoredNodesPlan.Position position) {
return favoredNodes[position.ordinal()];
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -53,7 +53,8 @@ public byte[] toByteArray() {
/**
* Parse the serialized representation of the {@link ClusterId}
* @param bytes A pb serialized {@link ClusterId} instance with pb magic prefix
* @return An instance of {@link ClusterId} made from <code>bytes</code> n * @see #toByteArray()
* @return An instance of {@link ClusterId} made from <code>bytes</code>
* @see #toByteArray()
*/
public static ClusterId parseFrom(final byte[] bytes) throws DeserializationException {
if (ProtobufUtil.isPBMagicPrefix(bytes)) {
Expand All @@ -79,9 +80,7 @@ public ClusterIdProtos.ClusterId convert() {
return builder.setClusterId(this.id).build();
}

/**
* n * @return A {@link ClusterId} made from the passed in <code>cid</code>
*/
/** Returns A {@link ClusterId} made from the passed in <code>cid</code> */
public static ClusterId convert(final ClusterIdProtos.ClusterId cid) {
return new ClusterId(cid.getClusterId());
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -80,9 +80,7 @@ public int hashCode() {
return this.serverName.hashCode();
}

/**
* n
*/
/** Returns regionInfo */
public RegionInfo getRegion() {
return regionInfo;
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -34,8 +34,7 @@ public NotAllMetaRegionsOnlineException() {
}

/**
* n
*/
* */
public NotAllMetaRegionsOnlineException(String message) {
super(message);
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -76,7 +76,7 @@ default String getVersion() {
Map<String, List<ReplicationLoadSource>> getReplicationLoadSourceMap();

/**
* Call directly from client such as hbase shell n
* Call directly from client such as hbase shell
*/
@Nullable
ReplicationLoadSink getReplicationLoadSink();
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -373,9 +373,10 @@ default void enableTable(TableName tableName) throws IOException {
* Disable table and wait on completion. May timeout eventually. Use
* {@link #disableTableAsync(org.apache.hadoop.hbase.TableName)} and
* {@link #isTableDisabled(org.apache.hadoop.hbase.TableName)} instead. The table has to be in
* enabled state for it to be disabled. n * @throws IOException There could be couple types of
* IOException TableNotFoundException means the table doesn't exist. TableNotEnabledException
* means the table isn't in enabled state.
* enabled state for it to be disabled.
* @throws IOException There could be couple types of IOException TableNotFoundException means the
* table doesn't exist. TableNotEnabledException means the table isn't in
* enabled state.
*/
default void disableTable(TableName tableName) throws IOException {
get(disableTableAsync(tableName), getSyncWaitTimeout(), TimeUnit.MILLISECONDS);
Expand Down Expand Up @@ -598,7 +599,7 @@ Future<Void> modifyColumnFamilyStoreFileTrackerAsync(TableName tableName, byte[]
* then it returns. It does not wait on the completion of Compaction (it can take a while).
* @param tableName table to compact
* @param compactType {@link org.apache.hadoop.hbase.client.CompactType}
* @throws IOException if a remote or network exception occurs n
* @throws IOException if a remote or network exception occurs
*/
void compact(TableName tableName, CompactType compactType)
throws IOException, InterruptedException;
Expand All @@ -610,7 +611,7 @@ void compact(TableName tableName, CompactType compactType)
* @param tableName table to compact
* @param columnFamily column family within a table
* @param compactType {@link org.apache.hadoop.hbase.client.CompactType}
* @throws IOException if not a mob column family or if a remote or network exception occurs n
* @throws IOException if not a mob column family or if a remote or network exception occurs
*/
void compact(TableName tableName, byte[] columnFamily, CompactType compactType)
throws IOException, InterruptedException;
Expand Down Expand Up @@ -659,7 +660,7 @@ void compact(TableName tableName, byte[] columnFamily, CompactType compactType)
* while).
* @param tableName table to compact
* @param compactType {@link org.apache.hadoop.hbase.client.CompactType}
* @throws IOException if a remote or network exception occurs n
* @throws IOException if a remote or network exception occurs
*/
void majorCompact(TableName tableName, CompactType compactType)
throws IOException, InterruptedException;
Expand All @@ -671,7 +672,7 @@ void majorCompact(TableName tableName, CompactType compactType)
* @param tableName table to compact
* @param columnFamily column family within a table
* @param compactType {@link org.apache.hadoop.hbase.client.CompactType}
* @throws IOException if not a mob column family or if a remote or network exception occurs n
* @throws IOException if not a mob column family or if a remote or network exception occurs
*/
void majorCompact(TableName tableName, byte[] columnFamily, CompactType compactType)
throws IOException, InterruptedException;
Expand Down Expand Up @@ -1880,7 +1881,7 @@ default int getMasterInfoPort() throws IOException {

/**
* Return the set of supported security capabilities.
* @throws IOException if a remote or network exception occurs n
* @throws IOException if a remote or network exception occurs
*/
List<SecurityCapability> getSecurityCapabilities() throws IOException;

Expand Down Expand Up @@ -2215,7 +2216,7 @@ void recommissionRegionServer(ServerName server, List<byte[]> encodedRegionNames
* Clear compacting queues on a regionserver.
* @param serverName the region server name
* @param queues the set of queue name
* @throws IOException if a remote or network exception occurs n
* @throws IOException if a remote or network exception occurs
*/
void clearCompactionQueues(ServerName serverName, Set<String> queues)
throws IOException, InterruptedException;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -63,15 +63,15 @@ public class Append extends Mutation {
* <p>
* This range is used as [minStamp, maxStamp).
* @param minStamp minimum timestamp value, inclusive
* @param maxStamp maximum timestamp value, exclusive n
* @param maxStamp maximum timestamp value, exclusive
*/
public Append setTimeRange(long minStamp, long maxStamp) {
tr = TimeRange.between(minStamp, maxStamp);
return this;
}

/**
* Gets the TimeRange used for this append. n
* Gets the TimeRange used for this append.
*/
public TimeRange getTimeRange() {
return this.tr;
Expand All @@ -83,7 +83,7 @@ protected long extraHeapSize() {
}

/**
* n * True (default) if the append operation should return the results. A client that is not
* True (default) if the append operation should return the results. A client that is not
* interested in the result can save network bandwidth setting this to false.
*/
@Override
Expand Down Expand Up @@ -122,7 +122,7 @@ public Append(Append appendToCopy) {
* Create a Append operation for the specified row.
* <p>
* At least one column must be appended to.
* @param rowArray Makes a copy out of this buffer. nn
* @param rowArray Makes a copy out of this buffer.
*/
public Append(final byte[] rowArray, final int rowOffset, final int rowLength) {
checkRow(rowArray, rowOffset, rowLength);
Expand All @@ -144,7 +144,7 @@ public Append(byte[] row, long ts, NavigableMap<byte[], List<Cell>> familyMap) {
* Add the specified column and value to this Append operation.
* @param family family name
* @param qualifier column qualifier
* @param value value to append to specified column n
* @param value value to append to specified column
*/
public Append addColumn(byte[] family, byte[] qualifier, byte[] value) {
KeyValue kv = new KeyValue(this.row, family, qualifier, this.ts, KeyValue.Type.Put, value);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -211,7 +211,7 @@ CompletableFuture<Void> createTable(TableDescriptor desc, byte[] startKey, byte[
CompletableFuture<Void> enableTable(TableName tableName);

/**
* Disable a table. The table has to be in enabled state for it to be disabled. n
* Disable a table. The table has to be in enabled state for it to be disabled.
*/
CompletableFuture<Void> disableTable(TableName tableName);

Expand Down Expand Up @@ -1156,7 +1156,7 @@ default CompletableFuture<Integer> getMasterInfoPort() {
CompletableFuture<Void> stopMaster();

/**
* Stop the designated regionserver. n
* Stop the designated regionserver.
*/
CompletableFuture<Void> stopRegionServer(ServerName serverName);

Expand Down Expand Up @@ -1365,8 +1365,8 @@ default CompletableFuture<Boolean> normalize() {
CompletableFuture<Boolean> normalize(NormalizeTableFilterParams ntfp);

/**
* Turn the cleaner chore on/off. n * @return Previous cleaner state wrapped by a
* {@link CompletableFuture}
* Turn the cleaner chore on/off.
* @return Previous cleaner state wrapped by a {@link CompletableFuture}
*/
CompletableFuture<Boolean> cleanerChoreSwitch(boolean on);

Expand All @@ -1385,8 +1385,8 @@ default CompletableFuture<Boolean> normalize() {
CompletableFuture<Boolean> runCleanerChore();

/**
* Turn the catalog janitor on/off. n * @return the previous state wrapped by a
* {@link CompletableFuture}
* Turn the catalog janitor on/off.
* @return the previous state wrapped by a {@link CompletableFuture}
*/
CompletableFuture<Boolean> catalogJanitorSwitch(boolean on);

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -185,7 +185,7 @@ private void spawnRenewalChore(final UserGroupInformation user) {
}

/**
* If choreService has not been created yet, create the ChoreService. n
* If choreService has not been created yet, create the ChoreService.
*/
synchronized ChoreService getChoreService() {
if (isClosed()) {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -133,7 +133,7 @@ public interface ColumnFamilyDescriptor {
int getMinVersions();

/**
* Get the mob compact partition policy for this family n
* Get the mob compact partition policy for this family
*/
MobCompactPartitionPolicy getMobCompactPartitionPolicy();

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -861,8 +861,8 @@ public ModifyableColumnFamilyDescriptor setIndexBlockEncoding(IndexBlockEncoding

/**
* Set whether the tags should be compressed along with DataBlockEncoding. When no
* DataBlockEncoding is been used, this is having no effect. n * @return this (for chained
* invocation)
* DataBlockEncoding is been used, this is having no effect.
* @return this (for chained invocation)
*/
public ModifyableColumnFamilyDescriptor setCompressTags(boolean compressTags) {
return setValue(COMPRESS_TAGS_BYTES, String.valueOf(compressTags));
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -97,7 +97,7 @@ public Delete(byte[] row, long timestamp) {
* <p>
* This timestamp is ONLY used for a delete row operation. If specifying families or columns, you
* must specify each timestamp individually.
* @param row We make a local copy of this passed in row. nn
* @param row We make a local copy of this passed in row.
*/
public Delete(final byte[] row, final int rowOffset, final int rowLength) {
this(row, rowOffset, rowLength, HConstants.LATEST_TIMESTAMP);
Expand Down Expand Up @@ -141,7 +141,7 @@ public Delete(byte[] row, long ts, NavigableMap<byte[], List<Cell>> familyMap) {
/**
* Add an existing delete marker to this Delete object.
* @param cell An existing cell of type "delete".
* @return this for invocation chaining n
* @return this for invocation chaining
*/
@Override
public Delete add(Cell cell) throws IOException {
Expand Down
Loading