From e176909987e786d3701741254ee01aaaba6d0b75 Mon Sep 17 00:00:00 2001 From: Lokesh Khurana Date: Tue, 25 Feb 2020 00:11:04 +0530 Subject: [PATCH 1/7] Moving classes out of test for downstreamers --- hbase-it/pom.xml | 19 +++++++++++++++++++ .../apache/hadoop/hbase/ClusterManager.java | 4 ++-- .../hadoop/hbase/DistributedHBaseCluster.java | 2 +- .../hadoop/hbase/HBaseClusterManager.java | 5 +++-- .../hbase/IntegrationTestingUtility.java | 2 ++ .../hadoop/hbase/chaos/actions/Action.java | 2 ++ .../hbase/chaos/actions/AddCPULoadAction.java | 2 ++ .../hbase/chaos/actions/AddColumnAction.java | 2 ++ .../chaos/actions/BatchRestartRsAction.java | 2 ++ .../actions/ChangeBloomFilterAction.java | 2 ++ .../actions/ChangeCompressionAction.java | 2 ++ .../chaos/actions/ChangeEncodingAction.java | 2 ++ .../actions/ChangeSplitPolicyAction.java | 2 ++ .../chaos/actions/ChangeVersionsAction.java | 2 ++ .../hbase/chaos/actions/CompactMobAction.java | 2 ++ .../CompactRandomRegionOfTableAction.java | 2 ++ .../chaos/actions/CompactTableAction.java | 2 ++ .../chaos/actions/CorruptDataFilesAction.java | 2 ++ .../actions/CorruptPacketsCommandAction.java | 2 ++ .../actions/DecreaseMaxHFileSizeAction.java | 2 ++ .../actions/DelayPacketsCommandAction.java | 2 ++ .../chaos/actions/DeleteDataFilesAction.java | 2 ++ .../actions/DumpClusterStatusAction.java | 2 ++ .../DuplicatePacketsCommandAction.java | 2 ++ .../chaos/actions/FillDiskCommandAction.java | 2 ++ .../FlushRandomRegionOfTableAction.java | 2 ++ .../hbase/chaos/actions/FlushTableAction.java | 2 ++ .../chaos/actions/ForceBalancerAction.java | 2 ++ .../GracefulRollingRestartRsAction.java | 2 ++ .../actions/LosePacketsCommandAction.java | 2 ++ ...rgeRandomAdjacentRegionsOfTableAction.java | 2 ++ .../MoveRandomRegionOfTableAction.java | 2 ++ .../actions/MoveRegionsOfTableAction.java | 2 ++ .../chaos/actions/RemoveColumnAction.java | 2 ++ .../actions/ReorderPacketsCommandAction.java | 2 ++ .../actions/RestartActionBaseAction.java | 2 ++ .../actions/RestartActiveMasterAction.java | 2 ++ .../actions/RestartActiveNameNodeAction.java | 2 ++ .../actions/RestartRandomDataNodeAction.java | 2 ++ .../chaos/actions/RestartRandomRsAction.java | 2 ++ .../RestartRandomRsExceptMetaAction.java | 2 ++ .../actions/RestartRandomZKNodeAction.java | 2 ++ .../actions/RestartRsHoldingMetaAction.java | 2 ++ .../actions/RestartRsHoldingTableAction.java | 2 ++ .../actions/RollingBatchRestartRsAction.java | 2 ++ ...RollingBatchRestartRsExceptMetaAction.java | 2 ++ .../RollingBatchSuspendResumeRsAction.java | 2 ++ .../chaos/actions/SnapshotTableAction.java | 2 ++ .../actions/SplitAllRegionOfTableAction.java | 3 ++- .../SplitRandomRegionOfTableAction.java | 2 ++ .../chaos/actions/SudoCommandAction.java | 2 ++ .../hbase/chaos/actions/TCCommandAction.java | 3 +++ .../chaos/actions/TruncateTableAction.java | 2 ++ .../UnbalanceKillAndRebalanceAction.java | 2 ++ .../chaos/actions/UnbalanceRegionsAction.java | 2 ++ .../chaos/factories/CalmMonkeyFactory.java | 3 ++- .../factories/DataIssuesMonkeyFactory.java | 2 ++ .../DistributedIssuesMonkeyFactory.java | 2 ++ .../factories/MasterKillingMonkeyFactory.java | 2 ++ .../factories/MobNoKillMonkeyFactory.java | 2 ++ .../MobSlowDeterministicMonkeyFactory.java | 2 ++ .../chaos/factories/MonkeyConstants.java | 2 ++ .../hbase/chaos/factories/MonkeyFactory.java | 2 ++ .../chaos/factories/NoKillMonkeyFactory.java | 2 ++ ...erAndDependenciesKillingMonkeyFactory.java | 2 ++ .../factories/ServerKillingMonkeyFactory.java | 2 ++ .../SlowDeterministicMonkeyFactory.java | 2 ++ .../StressAssignmentManagerMonkeyFactory.java | 2 ++ .../factories/UnbalanceMonkeyFactory.java | 2 ++ .../hbase/chaos/monkies/CalmChaosMonkey.java | 3 +++ .../hbase/chaos/monkies/ChaosMonkey.java | 2 ++ .../chaos/monkies/PolicyBasedChaosMonkey.java | 2 ++ .../policies/CompositeSequentialPolicy.java | 2 ++ .../chaos/policies/DoActionsOncePolicy.java | 2 ++ .../hbase/chaos/policies/PeriodicPolicy.java | 2 ++ .../policies/PeriodicRandomActionPolicy.java | 2 ++ .../hadoop/hbase/chaos/policies/Policy.java | 2 ++ .../policies/TwoConcurrentActionPolicy.java | 2 ++ .../hbase/chaos/util/ChaosMonkeyRunner.java | 2 ++ .../hadoop/hbase/chaos/util/Monkeys.java | 2 ++ 80 files changed, 179 insertions(+), 7 deletions(-) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/ClusterManager.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java (99%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/HBaseClusterManager.java (99%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/IntegrationTestingUtility.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/Action.java (99%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/AddCPULoadAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/AddColumnAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/ChangeBloomFilterAction.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/ChangeCompressionAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/ChangeEncodingAction.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/ChangeSplitPolicyAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/ChangeVersionsAction.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/CompactMobAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/CompactRandomRegionOfTableAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/CompactTableAction.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/CorruptDataFilesAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/CorruptPacketsCommandAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/DecreaseMaxHFileSizeAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/DelayPacketsCommandAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/DeleteDataFilesAction.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/DumpClusterStatusAction.java (94%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/DuplicatePacketsCommandAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/FillDiskCommandAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/FlushRandomRegionOfTableAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/FlushTableAction.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/ForceBalancerAction.java (94%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/GracefulRollingRestartRsAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/LosePacketsCommandAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/MergeRandomAdjacentRegionsOfTableAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/MoveRandomRegionOfTableAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/MoveRegionsOfTableAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RemoveColumnAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/ReorderPacketsCommandAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RestartActionBaseAction.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveMasterAction.java (95%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveNameNodeAction.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsAction.java (95%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsExceptMetaAction.java (95%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomZKNodeAction.java (95%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingMetaAction.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingTableAction.java (95%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java (95%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchSuspendResumeRsAction.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/SnapshotTableAction.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/SplitAllRegionOfTableAction.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/SplitRandomRegionOfTableAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/SudoCommandAction.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/TCCommandAction.java (93%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/TruncateTableAction.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/CalmMonkeyFactory.java (93%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/DataIssuesMonkeyFactory.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/DistributedIssuesMonkeyFactory.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/MasterKillingMonkeyFactory.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/MobNoKillMonkeyFactory.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/MobSlowDeterministicMonkeyFactory.java (99%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/MonkeyFactory.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/NoKillMonkeyFactory.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/SlowDeterministicMonkeyFactory.java (99%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/StressAssignmentManagerMonkeyFactory.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/factories/UnbalanceMonkeyFactory.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/monkies/CalmChaosMonkey.java (94%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/policies/CompositeSequentialPolicy.java (95%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/policies/PeriodicPolicy.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/policies/Policy.java (96%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/policies/TwoConcurrentActionPolicy.java (97%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/util/ChaosMonkeyRunner.java (98%) rename hbase-it/src/{test => main}/java/org/apache/hadoop/hbase/chaos/util/Monkeys.java (97%) diff --git a/hbase-it/pom.xml b/hbase-it/pom.xml index 0fe578a99f9a..48241eb77937 100644 --- a/hbase-it/pom.xml +++ b/hbase-it/pom.xml @@ -154,10 +154,29 @@ test-jar test + + org.apache.hbase + hbase-server + test-jar + compile + + + org.apache.hbase + hbase-zookeeper + test-jar + compile + org.apache.hbase hbase-common jar + compile + + + org.apache.hbase + hbase-common + test-jar + compile org.apache.hbase diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/ClusterManager.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/ClusterManager.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/ClusterManager.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/ClusterManager.java index 6356345aa1aa..d6917c1ebdf8 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/ClusterManager.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/ClusterManager.java @@ -29,8 +29,8 @@ * for starting / stopping / killing Hadoop/HBase daemons. Concrete implementations provide actual * functionality for carrying out deployment-specific tasks. */ -@InterfaceAudience.Private -interface ClusterManager extends Configurable { +@InterfaceAudience.Public +public interface ClusterManager extends Configurable { /** * Type of the service daemon */ diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java similarity index 99% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java index 796bc1f27e00..865c5e2191ab 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java @@ -41,7 +41,7 @@ * Manages the interactions with an already deployed distributed cluster (as opposed to * a pseudo-distributed, or mini/local cluster). This is used by integration and system tests. */ -@InterfaceAudience.Private +@InterfaceAudience.Public public class DistributedHBaseCluster extends HBaseCluster { private Admin admin; private final Connection connection; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/HBaseClusterManager.java similarity index 99% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/HBaseClusterManager.java index 2f75c731bffa..ed09d17af60b 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/HBaseClusterManager.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/HBaseClusterManager.java @@ -43,8 +43,9 @@ * servers on the remote machines (for example, the test user could be the same user as the * user the daemon is running as) */ -@InterfaceAudience.Private -public class HBaseClusterManager extends Configured implements ClusterManager { +@InterfaceAudience.Public +public class HBaseClusterManager extends Configured implements + org.apache.hadoop.hbase.ClusterManager { private static final String SIGKILL = "SIGKILL"; private static final String SIGSTOP = "SIGSTOP"; private static final String SIGCONT = "SIGCONT"; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestingUtility.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/IntegrationTestingUtility.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestingUtility.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/IntegrationTestingUtility.java index a9e555eaec95..4862e72a96a9 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/IntegrationTestingUtility.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/IntegrationTestingUtility.java @@ -21,6 +21,7 @@ import java.io.IOException; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.util.ReflectionUtils; +import org.apache.yetus.audience.InterfaceAudience; /** * Facility for integration/system tests. This extends {@link HBaseTestingUtility} @@ -39,6 +40,7 @@ * via {@link #initializeCluster(int)}. Individual tests should not directly call * {@link #setUseDistributedCluster(Configuration)}. */ +@InterfaceAudience.Public public class IntegrationTestingUtility extends HBaseTestingUtility { public IntegrationTestingUtility() { diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/Action.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/Action.java similarity index 99% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/Action.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/Action.java index 98ad2dc85ea2..d6e693ae0a59 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/Action.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/Action.java @@ -46,12 +46,14 @@ import org.apache.hadoop.hbase.client.TableDescriptor; import org.apache.hadoop.hbase.client.TableDescriptorBuilder; import org.apache.hadoop.hbase.util.Bytes; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * A (possibly mischievous) action that the ChaosMonkey can perform. */ +@InterfaceAudience.Public public class Action { public static final String KILL_MASTER_TIMEOUT_KEY = diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/AddCPULoadAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddCPULoadAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/AddCPULoadAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddCPULoadAction.java index 9d6437e431ba..3e89b51df049 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/AddCPULoadAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddCPULoadAction.java @@ -22,12 +22,14 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that adds high cpu load to a random regionserver for a given duration */ +@InterfaceAudience.Public public class AddCPULoadAction extends SudoCommandAction { protected static final Logger LOG = LoggerFactory.getLogger(AddCPULoadAction.class); private static final String CPU_LOAD_COMMAND = diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/AddColumnAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddColumnAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/AddColumnAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddColumnAction.java index a01222aee945..6e3d65cbe3e9 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/AddColumnAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddColumnAction.java @@ -27,12 +27,14 @@ import org.apache.hadoop.hbase.client.ColumnFamilyDescriptorBuilder; import org.apache.hadoop.hbase.client.TableDescriptor; import org.apache.hadoop.hbase.client.TableDescriptorBuilder; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action the adds a column family to a table. */ +@InterfaceAudience.Public public class AddColumnAction extends Action { private final TableName tableName; private Admin admin; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java index 2026a913a525..c54cd3a847e0 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java @@ -24,12 +24,14 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Restarts a ratio of the running regionservers at the same time */ +@InterfaceAudience.Public public class BatchRestartRsAction extends RestartActionBaseAction { float ratio; //ratio of regionservers to restart private static final Logger LOG = diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeBloomFilterAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeBloomFilterAction.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeBloomFilterAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeBloomFilterAction.java index e76675aa4b1c..cb56b2fac27c 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeBloomFilterAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeBloomFilterAction.java @@ -23,6 +23,7 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.regionserver.BloomType; import org.apache.hadoop.hbase.util.BloomFilterUtil; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -30,6 +31,7 @@ * Action that tries to adjust the bloom filter setting on all the columns of a * table */ +@InterfaceAudience.Public public class ChangeBloomFilterAction extends Action { private final long sleepTime; private final TableName tableName; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeCompressionAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeCompressionAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeCompressionAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeCompressionAction.java index 533028428f67..3c1d4d60c3cc 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeCompressionAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeCompressionAction.java @@ -24,12 +24,14 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.io.compress.Compression.Algorithm; import org.apache.hadoop.io.compress.Compressor; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that changes the compression algorithm on a column family from a list of tables. */ +@InterfaceAudience.Public public class ChangeCompressionAction extends Action { private final TableName tableName; private final Random random; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeEncodingAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeEncodingAction.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeEncodingAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeEncodingAction.java index 55a308ce3868..4d9cfb4db861 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeEncodingAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeEncodingAction.java @@ -23,12 +23,14 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that changes the encoding on a column family from a list of tables. */ +@InterfaceAudience.Public public class ChangeEncodingAction extends Action { private final TableName tableName; private final Random random; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeSplitPolicyAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeSplitPolicyAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeSplitPolicyAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeSplitPolicyAction.java index 72ff50fc0929..604f6ba9f798 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeSplitPolicyAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeSplitPolicyAction.java @@ -27,9 +27,11 @@ import org.apache.hadoop.hbase.regionserver.ConstantSizeRegionSplitPolicy; import org.apache.hadoop.hbase.regionserver.DisabledRegionSplitPolicy; import org.apache.hadoop.hbase.regionserver.IncreasingToUpperBoundRegionSplitPolicy; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; +@InterfaceAudience.Public public class ChangeSplitPolicyAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(ChangeSplitPolicyAction.class); private final TableName tableName; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeVersionsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeVersionsAction.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeVersionsAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeVersionsAction.java index 18dba94a3031..25617f452d22 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ChangeVersionsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeVersionsAction.java @@ -22,6 +22,7 @@ import java.util.Random; import org.apache.hadoop.hbase.TableName; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -30,6 +31,7 @@ * * Always keeps at least 1 as the number of versions. */ +@InterfaceAudience.Public public class ChangeVersionsAction extends Action { private final TableName tableName; private static final Logger LOG = LoggerFactory.getLogger(ChangeVersionsAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CompactMobAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactMobAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CompactMobAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactMobAction.java index 6ac71d5739a0..b7f1604174b7 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CompactMobAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactMobAction.java @@ -23,12 +23,14 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.CompactType; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that queues a table compaction. */ +@InterfaceAudience.Public public class CompactMobAction extends Action { private final TableName tableName; private final int majorRatio; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CompactRandomRegionOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactRandomRegionOfTableAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CompactRandomRegionOfTableAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactRandomRegionOfTableAction.java index 9a46147014c6..cba460b18f3f 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CompactRandomRegionOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactRandomRegionOfTableAction.java @@ -26,12 +26,14 @@ import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionInfo; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Region that queues a compaction of a random region from the table. */ +@InterfaceAudience.Public public class CompactRandomRegionOfTableAction extends Action { private final int majorRatio; private final long sleepTime; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CompactTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactTableAction.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CompactTableAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactTableAction.java index c05d43b89f57..53b9d9889035 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CompactTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactTableAction.java @@ -22,12 +22,14 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that queues a table compaction. */ +@InterfaceAudience.Public public class CompactTableAction extends Action { private final TableName tableName; private final int majorRatio; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CorruptDataFilesAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptDataFilesAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CorruptDataFilesAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptDataFilesAction.java index 83e8fe08a495..c53a720d5256 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CorruptDataFilesAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptDataFilesAction.java @@ -26,12 +26,14 @@ import org.apache.hadoop.fs.RemoteIterator; import org.apache.hadoop.hbase.io.hfile.HFile; import org.apache.hadoop.hbase.util.CommonFSUtils; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action corrupts HFiles with a certain chance. */ +@InterfaceAudience.Public public class CorruptDataFilesAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(CorruptDataFilesAction.class); private float chance; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CorruptPacketsCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptPacketsCommandAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CorruptPacketsCommandAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptPacketsCommandAction.java index 965bcbb6d417..d2927d76a1c3 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/CorruptPacketsCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptPacketsCommandAction.java @@ -22,6 +22,7 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -29,6 +30,7 @@ * * Corrupt network packets on a random regionserver. */ +@InterfaceAudience.Public public class CorruptPacketsCommandAction extends TCCommandAction { private static final Logger LOG = LoggerFactory.getLogger(CorruptPacketsCommandAction.class); private float ratio; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DecreaseMaxHFileSizeAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DecreaseMaxHFileSizeAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DecreaseMaxHFileSizeAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DecreaseMaxHFileSizeAction.java index 4610ef0de7de..de762d1e45d5 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DecreaseMaxHFileSizeAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DecreaseMaxHFileSizeAction.java @@ -26,7 +26,9 @@ import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.TableDescriptor; import org.apache.hadoop.hbase.client.TableDescriptorBuilder; +import org.apache.yetus.audience.InterfaceAudience; +@InterfaceAudience.Public public class DecreaseMaxHFileSizeAction extends Action { private static final long minFileSize = 1 * 1024 * 1024 * 1024L; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DelayPacketsCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DelayPacketsCommandAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DelayPacketsCommandAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DelayPacketsCommandAction.java index 38578d8690e3..d4e5b5a74295 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DelayPacketsCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DelayPacketsCommandAction.java @@ -22,12 +22,14 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action adds latency to communication on a random regionserver. */ +@InterfaceAudience.Public public class DelayPacketsCommandAction extends TCCommandAction { private static final Logger LOG = LoggerFactory.getLogger(DelayPacketsCommandAction.class); private long delay; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DeleteDataFilesAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DeleteDataFilesAction.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DeleteDataFilesAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DeleteDataFilesAction.java index 4919adce490c..b71788d8063b 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DeleteDataFilesAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DeleteDataFilesAction.java @@ -25,12 +25,14 @@ import org.apache.hadoop.fs.RemoteIterator; import org.apache.hadoop.hbase.io.hfile.HFile; import org.apache.hadoop.hbase.util.CommonFSUtils; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action deletes HFiles with a certain chance. */ +@InterfaceAudience.Public public class DeleteDataFilesAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(DeleteDataFilesAction.class); private float chance; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DumpClusterStatusAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DumpClusterStatusAction.java similarity index 94% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DumpClusterStatusAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DumpClusterStatusAction.java index 3c70af0349c0..4cd0e5371175 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DumpClusterStatusAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DumpClusterStatusAction.java @@ -20,12 +20,14 @@ import java.io.IOException; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action to dump the cluster status. */ +@InterfaceAudience.Public public class DumpClusterStatusAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(DumpClusterStatusAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DuplicatePacketsCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DuplicatePacketsCommandAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DuplicatePacketsCommandAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DuplicatePacketsCommandAction.java index 7ed74ec4cde4..a2ad9bebf32f 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/DuplicatePacketsCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DuplicatePacketsCommandAction.java @@ -22,6 +22,7 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -29,6 +30,7 @@ * * Duplicate network packets on a random regionserver. */ +@InterfaceAudience.Public public class DuplicatePacketsCommandAction extends TCCommandAction { private static final Logger LOG = LoggerFactory.getLogger(DuplicatePacketsCommandAction.class); private float ratio; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/FillDiskCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FillDiskCommandAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/FillDiskCommandAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FillDiskCommandAction.java index f5ca1d7f1944..8179d70ae822 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/FillDiskCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FillDiskCommandAction.java @@ -22,6 +22,7 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -29,6 +30,7 @@ * * Fill the disk on a random regionserver. */ +@InterfaceAudience.Public public class FillDiskCommandAction extends SudoCommandAction { private static final Logger LOG = LoggerFactory.getLogger(FillDiskCommandAction.class); private long size; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/FlushRandomRegionOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushRandomRegionOfTableAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/FlushRandomRegionOfTableAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushRandomRegionOfTableAction.java index 083423c4ca99..e5f18cbc506b 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/FlushRandomRegionOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushRandomRegionOfTableAction.java @@ -24,12 +24,14 @@ import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionInfo; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to flush a random region of a table. */ +@InterfaceAudience.Public public class FlushRandomRegionOfTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(FlushRandomRegionOfTableAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/FlushTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushTableAction.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/FlushTableAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushTableAction.java index 93f7347ac93d..1778f9454595 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/FlushTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushTableAction.java @@ -21,12 +21,14 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to flush a table. */ +@InterfaceAudience.Public public class FlushTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(FlushTableAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ForceBalancerAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ForceBalancerAction.java similarity index 94% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ForceBalancerAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ForceBalancerAction.java index d75475432a12..bcb16296c9a1 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ForceBalancerAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ForceBalancerAction.java @@ -18,12 +18,14 @@ package org.apache.hadoop.hbase.chaos.actions; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to force a balancer run. */ +@InterfaceAudience.Public public class ForceBalancerAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(ForceBalancerAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/GracefulRollingRestartRsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/GracefulRollingRestartRsAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/GracefulRollingRestartRsAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/GracefulRollingRestartRsAction.java index 82005bbbd4e7..5bdc3199ddd9 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/GracefulRollingRestartRsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/GracefulRollingRestartRsAction.java @@ -25,6 +25,7 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.util.RegionMover; import org.apache.hadoop.util.Shell; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -32,6 +33,7 @@ * Gracefully restarts every regionserver in a rolling fashion. At each step, it unloads, * restarts the loads every rs server sleeping randomly (0-sleepTime) in between servers. */ +@InterfaceAudience.Public public class GracefulRollingRestartRsAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(GracefulRollingRestartRsAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/LosePacketsCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/LosePacketsCommandAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/LosePacketsCommandAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/LosePacketsCommandAction.java index 8b09daf4cf88..b5c9a9cd4c03 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/LosePacketsCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/LosePacketsCommandAction.java @@ -22,6 +22,7 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -29,6 +30,7 @@ * * Lose network packets on a random regionserver. */ +@InterfaceAudience.Public public class LosePacketsCommandAction extends TCCommandAction { private static final Logger LOG = LoggerFactory.getLogger(LosePacketsCommandAction.class); private float ratio; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/MergeRandomAdjacentRegionsOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MergeRandomAdjacentRegionsOfTableAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/MergeRandomAdjacentRegionsOfTableAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MergeRandomAdjacentRegionsOfTableAction.java index 085413d6e6fa..bbdfe743b979 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/MergeRandomAdjacentRegionsOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MergeRandomAdjacentRegionsOfTableAction.java @@ -24,12 +24,14 @@ import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionInfo; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action to merge regions of a table. */ +@InterfaceAudience.Public public class MergeRandomAdjacentRegionsOfTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(MergeRandomAdjacentRegionsOfTableAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/MoveRandomRegionOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRandomRegionOfTableAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/MoveRandomRegionOfTableAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRandomRegionOfTableAction.java index cdd2f20aed82..8ebcd3caf84e 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/MoveRandomRegionOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRandomRegionOfTableAction.java @@ -25,12 +25,14 @@ import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionInfo; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to move a random region of a table. */ +@InterfaceAudience.Public public class MoveRandomRegionOfTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(MoveRandomRegionOfTableAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/MoveRegionsOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRegionsOfTableAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/MoveRegionsOfTableAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRegionsOfTableAction.java index 1c1f07b3df58..27869de28258 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/MoveRegionsOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRegionsOfTableAction.java @@ -29,12 +29,14 @@ import org.apache.hadoop.hbase.chaos.factories.MonkeyConstants; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionInfo; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to move every region of a table. */ +@InterfaceAudience.Public public class MoveRegionsOfTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(MoveRegionsOfTableAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RemoveColumnAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RemoveColumnAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RemoveColumnAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RemoveColumnAction.java index 37e540104d68..56f87062628d 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RemoveColumnAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RemoveColumnAction.java @@ -28,12 +28,14 @@ import org.apache.hadoop.hbase.client.TableDescriptor; import org.apache.hadoop.hbase.client.TableDescriptorBuilder; import org.apache.hadoop.hbase.util.Bytes; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that removes a column family. */ +@InterfaceAudience.Public public class RemoveColumnAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(RemoveColumnAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ReorderPacketsCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ReorderPacketsCommandAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ReorderPacketsCommandAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ReorderPacketsCommandAction.java index 1bb7969bb707..9316c58b57a2 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/ReorderPacketsCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ReorderPacketsCommandAction.java @@ -22,6 +22,7 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -29,6 +30,7 @@ * * Reorder network packets on a random regionserver. */ +@InterfaceAudience.Public public class ReorderPacketsCommandAction extends TCCommandAction { private static final Logger LOG = LoggerFactory.getLogger(ReorderPacketsCommandAction.class); private float ratio; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartActionBaseAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActionBaseAction.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartActionBaseAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActionBaseAction.java index c53de9082261..b86332dc3bc7 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartActionBaseAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActionBaseAction.java @@ -22,12 +22,14 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.util.Threads; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Base class for restarting HBaseServer's */ +@InterfaceAudience.Public public class RestartActionBaseAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(RestartActionBaseAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveMasterAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveMasterAction.java similarity index 95% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveMasterAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveMasterAction.java index ddd719a6e67a..be1da5ce1886 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveMasterAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveMasterAction.java @@ -19,12 +19,14 @@ package org.apache.hadoop.hbase.chaos.actions; import org.apache.hadoop.hbase.ServerName; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to restart the active master. */ +@InterfaceAudience.Public public class RestartActiveMasterAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartActionBaseAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveNameNodeAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveNameNodeAction.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveNameNodeAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveNameNodeAction.java index 2fc64e68a2cf..5ceaa619a4b3 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveNameNodeAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveNameNodeAction.java @@ -30,12 +30,14 @@ import org.apache.hadoop.hdfs.DFSUtil; import org.apache.hadoop.hdfs.HAUtil; import org.apache.hadoop.hdfs.server.namenode.ha.proto.HAZKInfoProtos.ActiveNodeInfo; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to restart the active namenode. */ +@InterfaceAudience.Public public class RestartActiveNameNodeAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartActiveNameNodeAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java index 5824a42fca8c..ae7f6135da2c 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java @@ -29,12 +29,14 @@ import org.apache.hadoop.hdfs.DistributedFileSystem; import org.apache.hadoop.hdfs.protocol.DatanodeInfo; import org.apache.hadoop.hdfs.protocol.HdfsConstants; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that restarts a random datanode. */ +@InterfaceAudience.Public public class RestartRandomDataNodeAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartRandomDataNodeAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsAction.java similarity index 95% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsAction.java index 48458b68dcf4..eca3a003ad5f 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsAction.java @@ -20,12 +20,14 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that restarts a random HRegionServer */ +@InterfaceAudience.Public public class RestartRandomRsAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartRandomRsAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsExceptMetaAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsExceptMetaAction.java similarity index 95% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsExceptMetaAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsExceptMetaAction.java index b78144a1ca2f..538cb63db3e9 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsExceptMetaAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsExceptMetaAction.java @@ -20,7 +20,9 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; +import org.apache.yetus.audience.InterfaceAudience; +@InterfaceAudience.Public public class RestartRandomRsExceptMetaAction extends RestartRandomRsAction { public RestartRandomRsExceptMetaAction(long sleepTime) { super(sleepTime); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomZKNodeAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomZKNodeAction.java similarity index 95% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomZKNodeAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomZKNodeAction.java index 3ed7a0d92379..aa923614688a 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomZKNodeAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomZKNodeAction.java @@ -21,12 +21,14 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.zookeeper.ZKServerTool; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that restarts a random zookeeper node. */ +@InterfaceAudience.Public public class RestartRandomZKNodeAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartRandomZKNodeAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingMetaAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingMetaAction.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingMetaAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingMetaAction.java index aeecf0a891fe..0f167ac7827c 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingMetaAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingMetaAction.java @@ -20,12 +20,14 @@ import org.apache.hadoop.hbase.ClusterMetrics; import org.apache.hadoop.hbase.ServerName; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to restart the HRegionServer holding Meta. */ +@InterfaceAudience.Public public class RestartRsHoldingMetaAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartRsHoldingMetaAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingTableAction.java similarity index 95% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingTableAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingTableAction.java index a63c672db39e..c06902464da5 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingTableAction.java @@ -23,12 +23,14 @@ import org.apache.commons.lang3.RandomUtils; import org.apache.hadoop.hbase.HRegionLocation; import org.apache.hadoop.hbase.client.RegionLocator; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that restarts an HRegionServer holding one of the regions of the table. */ +@InterfaceAudience.Public public class RestartRsHoldingTableAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartRsHoldingTableAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java index 53f952083437..fa7f5171f2c6 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java @@ -27,6 +27,7 @@ import org.apache.commons.lang3.RandomUtils; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -36,6 +37,7 @@ * The parameter maxDeadServers limits the maximum number of servers that * can be down at the same time during rolling restarts. */ +@InterfaceAudience.Public public class RollingBatchRestartRsAction extends BatchRestartRsAction { private static final Logger LOG = LoggerFactory.getLogger(RollingBatchRestartRsAction.class); protected int maxDeadServers; // number of maximum dead servers at any given time. Defaults to 5 diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java similarity index 95% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java index d0d0fe561d4f..a575d1e37296 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java @@ -21,11 +21,13 @@ import java.util.List; import org.apache.hadoop.hbase.ServerName; +import org.apache.yetus.audience.InterfaceAudience; /** * Same as in {@link RollingBatchRestartRsAction} except that this action * does not restart the region server holding the META table. */ +@InterfaceAudience.Public public class RollingBatchRestartRsExceptMetaAction extends RollingBatchRestartRsAction { public RollingBatchRestartRsExceptMetaAction(long sleepTime, float ratio, int maxDeadServers) { diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchSuspendResumeRsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchSuspendResumeRsAction.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchSuspendResumeRsAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchSuspendResumeRsAction.java index d4ad3e40b5f8..2720578dccfd 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchSuspendResumeRsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchSuspendResumeRsAction.java @@ -28,6 +28,7 @@ import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.util.Threads; import org.apache.hadoop.util.Shell; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -37,6 +38,7 @@ * maxSuspendedServers limits the maximum number of servers that can be down at the same time * during rolling restarts. */ +@InterfaceAudience.Public public class RollingBatchSuspendResumeRsAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(RollingBatchSuspendResumeRsAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/SnapshotTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SnapshotTableAction.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/SnapshotTableAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SnapshotTableAction.java index 6ee9b57b8a7c..1ebf42037beb 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/SnapshotTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SnapshotTableAction.java @@ -21,12 +21,14 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to take a snapshot of a table. */ +@InterfaceAudience.Public public class SnapshotTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(SnapshotTableAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/SplitAllRegionOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitAllRegionOfTableAction.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/SplitAllRegionOfTableAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitAllRegionOfTableAction.java index 10d544051601..aa9e8f0ed569 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/SplitAllRegionOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitAllRegionOfTableAction.java @@ -23,10 +23,11 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; - +@InterfaceAudience.Public public class SplitAllRegionOfTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(SplitAllRegionOfTableAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/SplitRandomRegionOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitRandomRegionOfTableAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/SplitRandomRegionOfTableAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitRandomRegionOfTableAction.java index b4873753bb35..d2f7090dc4bc 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/SplitRandomRegionOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitRandomRegionOfTableAction.java @@ -25,12 +25,14 @@ import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionInfo; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to split a random region of a table. */ +@InterfaceAudience.Public public class SplitRandomRegionOfTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(SplitRandomRegionOfTableAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/SudoCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SudoCommandAction.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/SudoCommandAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SudoCommandAction.java index 6092a5dbbc6a..a8cbd87a3bf3 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/SudoCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SudoCommandAction.java @@ -23,12 +23,14 @@ import org.apache.hadoop.hbase.DistributedHBaseCluster; import org.apache.hadoop.hbase.HBaseCluster; import org.apache.hadoop.hbase.HBaseClusterManager; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Base class for performing Actions based on linux commands requiring sudo privileges */ +@InterfaceAudience.Public abstract public class SudoCommandAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(SudoCommandAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/TCCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TCCommandAction.java similarity index 93% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/TCCommandAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TCCommandAction.java index 9444f876f72a..4754cf534a61 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/TCCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TCCommandAction.java @@ -18,9 +18,12 @@ package org.apache.hadoop.hbase.chaos.actions; +import org.apache.yetus.audience.InterfaceAudience; + /** * Base class for tc command actions */ +@InterfaceAudience.Public abstract public class TCCommandAction extends SudoCommandAction { protected static final String ADD = "add"; protected static final String DELETE = "del"; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/TruncateTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TruncateTableAction.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/TruncateTableAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TruncateTableAction.java index b79dafa14e82..f8634f63c354 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/TruncateTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TruncateTableAction.java @@ -23,12 +23,14 @@ import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to truncate of a table. */ +@InterfaceAudience.Public public class TruncateTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(TruncateTableAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java index 5e5504af10ee..021e94c05c72 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java @@ -27,11 +27,13 @@ import org.apache.commons.lang3.RandomUtils; import org.apache.hadoop.hbase.ClusterMetrics; import org.apache.hadoop.hbase.ServerName; +import org.apache.yetus.audience.InterfaceAudience; import org.junit.Assert; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** This action is too specific to put in ChaosMonkey; put it here */ +@InterfaceAudience.Public public class UnbalanceKillAndRebalanceAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(UnbalanceKillAndRebalanceAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java index 3f2a3a1b518a..65b2417901aa 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java @@ -25,12 +25,14 @@ import org.apache.commons.lang3.RandomUtils; import org.apache.hadoop.hbase.ClusterMetrics; import org.apache.hadoop.hbase.ServerName; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to unbalance the regions of a cluster. */ +@InterfaceAudience.Public public class UnbalanceRegionsAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(UnbalanceRegionsAction.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/CalmMonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/CalmMonkeyFactory.java similarity index 93% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/CalmMonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/CalmMonkeyFactory.java index e10958572274..35c08bdb0e1b 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/CalmMonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/CalmMonkeyFactory.java @@ -20,11 +20,12 @@ import org.apache.hadoop.hbase.chaos.monkies.CalmChaosMonkey; import org.apache.hadoop.hbase.chaos.monkies.ChaosMonkey; - +import org.apache.yetus.audience.InterfaceAudience; /** * Factory to create a calm ChaosMonkey. */ +@InterfaceAudience.Private public class CalmMonkeyFactory extends MonkeyFactory { @Override public ChaosMonkey build() { diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/DataIssuesMonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/DataIssuesMonkeyFactory.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/DataIssuesMonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/DataIssuesMonkeyFactory.java index a06a9779e414..0b575945d50a 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/DataIssuesMonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/DataIssuesMonkeyFactory.java @@ -25,12 +25,14 @@ import org.apache.hadoop.hbase.chaos.monkies.ChaosMonkey; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.chaos.policies.PeriodicRandomActionPolicy; +import org.apache.yetus.audience.InterfaceAudience; /** * A chaos monkey to delete and corrupt regionserver data, requires a user with * passwordless ssh access to the cluster and sudo privileges. * Highly destructive */ +@InterfaceAudience.Private public class DataIssuesMonkeyFactory extends MonkeyFactory { private long action1Period; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/DistributedIssuesMonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/DistributedIssuesMonkeyFactory.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/DistributedIssuesMonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/DistributedIssuesMonkeyFactory.java index 3091fa5557b6..03dc0e6c9ee1 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/DistributedIssuesMonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/DistributedIssuesMonkeyFactory.java @@ -30,11 +30,13 @@ import org.apache.hadoop.hbase.chaos.monkies.ChaosMonkey; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.chaos.policies.PeriodicRandomActionPolicy; +import org.apache.yetus.audience.InterfaceAudience; /** * A chaos monkey to create distributed cluster related issues, requires a user with * passwordless ssh access to the cluster and sudo privileges. */ +@InterfaceAudience.Private public class DistributedIssuesMonkeyFactory extends MonkeyFactory { private long action1Period; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MasterKillingMonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MasterKillingMonkeyFactory.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MasterKillingMonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MasterKillingMonkeyFactory.java index 52dec3beb3c1..dc39b83c98cd 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MasterKillingMonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MasterKillingMonkeyFactory.java @@ -24,11 +24,13 @@ import org.apache.hadoop.hbase.chaos.monkies.ChaosMonkey; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.chaos.policies.PeriodicRandomActionPolicy; +import org.apache.yetus.audience.InterfaceAudience; /** * A chaos monkey to kill the active master periodically. Can be run in single master * or multi master setup. */ +@InterfaceAudience.Private public class MasterKillingMonkeyFactory extends MonkeyFactory { private long action1Period; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MobNoKillMonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MobNoKillMonkeyFactory.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MobNoKillMonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MobNoKillMonkeyFactory.java index 639937820f31..42df04d257bf 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MobNoKillMonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MobNoKillMonkeyFactory.java @@ -39,10 +39,12 @@ import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.chaos.policies.PeriodicRandomActionPolicy; import org.apache.hadoop.hbase.chaos.policies.TwoConcurrentActionPolicy; +import org.apache.yetus.audience.InterfaceAudience; /** * This is a copy of NoKillMonkeyFactory that also does mob compactions. */ +@InterfaceAudience.Private public class MobNoKillMonkeyFactory extends MonkeyFactory { @Override public ChaosMonkey build() { Action[] actions1 = new Action[] { diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MobSlowDeterministicMonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MobSlowDeterministicMonkeyFactory.java similarity index 99% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MobSlowDeterministicMonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MobSlowDeterministicMonkeyFactory.java index c59bfd66607f..a7b7555a5109 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MobSlowDeterministicMonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MobSlowDeterministicMonkeyFactory.java @@ -45,10 +45,12 @@ import org.apache.hadoop.hbase.chaos.policies.CompositeSequentialPolicy; import org.apache.hadoop.hbase.chaos.policies.DoActionsOncePolicy; import org.apache.hadoop.hbase.chaos.policies.PeriodicRandomActionPolicy; +import org.apache.yetus.audience.InterfaceAudience; /** * This is a copy of SlowDeterministicMonkeyFactory that also does mob compactions. */ +@InterfaceAudience.Private public class MobSlowDeterministicMonkeyFactory extends MonkeyFactory { private long action1Period; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java index 7fbdd1ff8324..aa7864748e00 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java @@ -17,10 +17,12 @@ */ package org.apache.hadoop.hbase.chaos.factories; +import org.apache.yetus.audience.InterfaceAudience; import java.util.Arrays; import java.util.HashSet; import java.util.Set; +@InterfaceAudience.Private public interface MonkeyConstants { String PERIODIC_ACTION1_PERIOD = "sdm.action1.period"; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyFactory.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyFactory.java index 73f696824403..1a6e5b234f9a 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/MonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyFactory.java @@ -28,12 +28,14 @@ import org.apache.hbase.thirdparty.com.google.common.collect.ImmutableMap; import org.apache.hadoop.hbase.util.ReflectionUtils; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Base class of the factory that will create a ChaosMonkey. */ +@InterfaceAudience.Public public abstract class MonkeyFactory { private static final Logger LOG = LoggerFactory.getLogger(MonkeyFactory.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/NoKillMonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/NoKillMonkeyFactory.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/NoKillMonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/NoKillMonkeyFactory.java index def9df0bfa80..5e77347e96cc 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/NoKillMonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/NoKillMonkeyFactory.java @@ -39,11 +39,13 @@ import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.chaos.policies.PeriodicRandomActionPolicy; import org.apache.hadoop.hbase.chaos.policies.TwoConcurrentActionPolicy; +import org.apache.yetus.audience.InterfaceAudience; /** * Monkey factory to create a ChaosMonkey that will not need access to ssh. It will not * kill any services and it will not perform any restarts. */ +@InterfaceAudience.Private public class NoKillMonkeyFactory extends MonkeyFactory { @Override public ChaosMonkey build() { Action[] actions1 = new Action[] { diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java index 5cb2d7f7b8a8..db4ff7f9699f 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/ServerAndDependenciesKillingMonkeyFactory.java @@ -33,11 +33,13 @@ import org.apache.hadoop.hbase.chaos.policies.CompositeSequentialPolicy; import org.apache.hadoop.hbase.chaos.policies.DoActionsOncePolicy; import org.apache.hadoop.hbase.chaos.policies.PeriodicRandomActionPolicy; +import org.apache.yetus.audience.InterfaceAudience; /** * Creates ChaosMonkeys for doing server restart actions, but not * flush / compact / snapshot kind of actions. */ +@InterfaceAudience.Private public class ServerAndDependenciesKillingMonkeyFactory extends MonkeyFactory { private long gracefulRollingRestartTSSLeepTime; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java index 3f2edcc9f8fd..ba3ab9f3b71b 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/ServerKillingMonkeyFactory.java @@ -31,11 +31,13 @@ import org.apache.hadoop.hbase.chaos.policies.CompositeSequentialPolicy; import org.apache.hadoop.hbase.chaos.policies.DoActionsOncePolicy; import org.apache.hadoop.hbase.chaos.policies.PeriodicRandomActionPolicy; +import org.apache.yetus.audience.InterfaceAudience; /** * Creates ChaosMonkeys for doing server restart actions, but not * flush / compact / snapshot kind of actions. */ +@InterfaceAudience.Private public class ServerKillingMonkeyFactory extends MonkeyFactory { private long gracefulRollingRestartTSSLeepTime; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/SlowDeterministicMonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/SlowDeterministicMonkeyFactory.java similarity index 99% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/SlowDeterministicMonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/SlowDeterministicMonkeyFactory.java index deaf25640c8d..f162d2b2d9a5 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/SlowDeterministicMonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/SlowDeterministicMonkeyFactory.java @@ -49,7 +49,9 @@ import org.apache.hadoop.hbase.chaos.policies.CompositeSequentialPolicy; import org.apache.hadoop.hbase.chaos.policies.DoActionsOncePolicy; import org.apache.hadoop.hbase.chaos.policies.PeriodicRandomActionPolicy; +import org.apache.yetus.audience.InterfaceAudience; +@InterfaceAudience.Private public class SlowDeterministicMonkeyFactory extends MonkeyFactory { private long action1Period; private long action2Period; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/StressAssignmentManagerMonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/StressAssignmentManagerMonkeyFactory.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/StressAssignmentManagerMonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/StressAssignmentManagerMonkeyFactory.java index b25bef7a334f..dbd737c52996 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/StressAssignmentManagerMonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/StressAssignmentManagerMonkeyFactory.java @@ -43,7 +43,9 @@ import org.apache.hadoop.hbase.chaos.policies.CompositeSequentialPolicy; import org.apache.hadoop.hbase.chaos.policies.DoActionsOncePolicy; import org.apache.hadoop.hbase.chaos.policies.PeriodicRandomActionPolicy; +import org.apache.yetus.audience.InterfaceAudience; +@InterfaceAudience.Private public class StressAssignmentManagerMonkeyFactory extends MonkeyFactory { private long gracefulRollingRestartTSSLeepTime; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/UnbalanceMonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/UnbalanceMonkeyFactory.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/UnbalanceMonkeyFactory.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/UnbalanceMonkeyFactory.java index 28ab7a03bed6..162ea45bda72 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/factories/UnbalanceMonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/UnbalanceMonkeyFactory.java @@ -23,7 +23,9 @@ import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.chaos.policies.PeriodicRandomActionPolicy; import org.apache.hadoop.hbase.chaos.policies.Policy; +import org.apache.yetus.audience.InterfaceAudience; +@InterfaceAudience.Private public class UnbalanceMonkeyFactory extends MonkeyFactory { /** How often to introduce the chaos. If too frequent, sequence of kills on minicluster * can cause test to fail when Put runs out of retries. */ diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/CalmChaosMonkey.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/CalmChaosMonkey.java similarity index 94% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/CalmChaosMonkey.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/CalmChaosMonkey.java index bcf7d85ee75a..7a45a366bd91 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/CalmChaosMonkey.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/CalmChaosMonkey.java @@ -18,9 +18,12 @@ package org.apache.hadoop.hbase.chaos.monkies; +import org.apache.yetus.audience.InterfaceAudience; + /** * Chaos Monkey that does nothing. */ +@InterfaceAudience.Public public class CalmChaosMonkey extends ChaosMonkey { @Override public void start() throws Exception { diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java index 86cbfe5c612d..682a0902f59c 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java @@ -19,6 +19,7 @@ package org.apache.hadoop.hbase.chaos.monkies; import org.apache.hadoop.hbase.Stoppable; +import org.apache.yetus.audience.InterfaceAudience; /** * A utility to injects faults in a running cluster. @@ -41,6 +42,7 @@ * ChaosMonkey class is indeed inspired by the Netflix's same-named tool: * http://techblog.netflix.com/2012/07/chaos-monkey-released-into-wild.html */ +@InterfaceAudience.Public public abstract class ChaosMonkey implements Stoppable { public abstract void start() throws Exception; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java index fed51491af3f..d5ebe21f9412 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java @@ -27,12 +27,14 @@ import org.apache.hadoop.hbase.IntegrationTestingUtility; import org.apache.hadoop.hbase.chaos.policies.Policy; import org.apache.hadoop.hbase.util.Pair; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Chaos monkey that given multiple policies will run actions against the cluster. */ +@InterfaceAudience.Public public class PolicyBasedChaosMonkey extends ChaosMonkey { private static final Logger LOG = LoggerFactory.getLogger(PolicyBasedChaosMonkey.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/CompositeSequentialPolicy.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/CompositeSequentialPolicy.java similarity index 95% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/CompositeSequentialPolicy.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/CompositeSequentialPolicy.java index 4fefca802a17..c8eadbfb27a2 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/CompositeSequentialPolicy.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/CompositeSequentialPolicy.java @@ -18,10 +18,12 @@ package org.apache.hadoop.hbase.chaos.policies; +import org.apache.yetus.audience.InterfaceAudience; import java.util.Arrays; import java.util.List; /** A policy that runs multiple other policies one after the other */ +@InterfaceAudience.Public public class CompositeSequentialPolicy extends Policy { private List policies; public CompositeSequentialPolicy(Policy... policies) { diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java index 0d8564d355bb..03a92e5e66b2 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java @@ -24,8 +24,10 @@ import org.apache.hadoop.hbase.chaos.actions.Action; import org.apache.hadoop.util.StringUtils; +import org.apache.yetus.audience.InterfaceAudience; /** A policy which performs a sequence of actions deterministically. */ +@InterfaceAudience.Public public class DoActionsOncePolicy extends PeriodicPolicy { private List actions; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/PeriodicPolicy.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicPolicy.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/PeriodicPolicy.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicPolicy.java index 9ffef16a1a96..a699b3036054 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/PeriodicPolicy.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicPolicy.java @@ -20,8 +20,10 @@ import org.apache.commons.lang3.RandomUtils; import org.apache.hadoop.hbase.util.Threads; +import org.apache.yetus.audience.InterfaceAudience; /** A policy which does stuff every time interval. */ +@InterfaceAudience.Public public abstract class PeriodicPolicy extends Policy { private long periodMs; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java index a0dc89e0192d..7f1c40fc69c1 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java @@ -26,11 +26,13 @@ import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.util.Pair; import org.apache.hadoop.util.StringUtils; +import org.apache.yetus.audience.InterfaceAudience; /** * A policy, which picks a random action according to the given weights, * and performs it every configurable period. */ +@InterfaceAudience.Public public class PeriodicRandomActionPolicy extends PeriodicPolicy { private List> actions; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/Policy.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/Policy.java similarity index 96% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/Policy.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/Policy.java index 81267a656882..12a6c652dae4 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/Policy.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/Policy.java @@ -21,12 +21,14 @@ import org.apache.hadoop.hbase.IntegrationTestingUtility; import org.apache.hadoop.hbase.chaos.actions.Action; import org.apache.hadoop.hbase.util.StoppableImplementation; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * A policy to introduce chaos to the cluster */ +@InterfaceAudience.Public public abstract class Policy extends StoppableImplementation implements Runnable { protected static final Logger LOG = LoggerFactory.getLogger(Policy.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/TwoConcurrentActionPolicy.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/TwoConcurrentActionPolicy.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/TwoConcurrentActionPolicy.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/TwoConcurrentActionPolicy.java index d801ba81b601..98c7d62189c9 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/policies/TwoConcurrentActionPolicy.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/TwoConcurrentActionPolicy.java @@ -22,6 +22,7 @@ import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.util.Threads; import org.apache.hadoop.util.StringUtils; +import org.apache.yetus.audience.InterfaceAudience; import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutorService; @@ -32,6 +33,7 @@ * Chaos Monkey policy that will run two different actions at the same time. * A random action from each array of actions will be chosen and then run in parallel. */ +@InterfaceAudience.Public public class TwoConcurrentActionPolicy extends PeriodicPolicy { private final Action[] actionsOne; private final Action[] actionsTwo; diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/util/ChaosMonkeyRunner.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/ChaosMonkeyRunner.java similarity index 98% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/util/ChaosMonkeyRunner.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/ChaosMonkeyRunner.java index 504bd628dc0a..d9c7455697ed 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/util/ChaosMonkeyRunner.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/ChaosMonkeyRunner.java @@ -31,11 +31,13 @@ import org.apache.hadoop.hbase.chaos.monkies.ChaosMonkey; import org.apache.hadoop.hbase.util.AbstractHBaseTool; import org.apache.hadoop.util.ToolRunner; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.hbase.thirdparty.com.google.common.collect.Sets; import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine; +@InterfaceAudience.Public public class ChaosMonkeyRunner extends AbstractHBaseTool { private static final Logger LOG = LoggerFactory.getLogger(ChaosMonkeyRunner.class); diff --git a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/util/Monkeys.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/Monkeys.java similarity index 97% rename from hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/util/Monkeys.java rename to hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/Monkeys.java index acdd7b083757..470ebd7aca4d 100644 --- a/hbase-it/src/test/java/org/apache/hadoop/hbase/chaos/util/Monkeys.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/Monkeys.java @@ -27,6 +27,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.IntegrationTestingUtility; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -36,6 +37,7 @@ /** * This class can be used to control chaos monkeys life cycle. */ +@InterfaceAudience.Public public class Monkeys implements Closeable { private static final Logger LOG = LoggerFactory.getLogger(Monkeys.class); From 83a83f23db127041c6bbc132da933cecf344432b Mon Sep 17 00:00:00 2001 From: Lokesh Khurana Date: Tue, 25 Feb 2020 19:19:43 +0530 Subject: [PATCH 2/7] Making MonkeyConstants Public --- .../main/java/org/apache/hadoop/hbase/HBaseClusterManager.java | 3 +-- .../apache/hadoop/hbase/chaos/factories/MonkeyConstants.java | 2 +- 2 files changed, 2 insertions(+), 3 deletions(-) diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/HBaseClusterManager.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/HBaseClusterManager.java index ed09d17af60b..1d361d52d14c 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/HBaseClusterManager.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/HBaseClusterManager.java @@ -44,8 +44,7 @@ * user the daemon is running as) */ @InterfaceAudience.Public -public class HBaseClusterManager extends Configured implements - org.apache.hadoop.hbase.ClusterManager { +public class HBaseClusterManager extends Configured implements ClusterManager { private static final String SIGKILL = "SIGKILL"; private static final String SIGSTOP = "SIGSTOP"; private static final String SIGCONT = "SIGCONT"; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java index aa7864748e00..b8e69b40d637 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java @@ -22,7 +22,7 @@ import java.util.HashSet; import java.util.Set; -@InterfaceAudience.Private +@InterfaceAudience.Public public interface MonkeyConstants { String PERIODIC_ACTION1_PERIOD = "sdm.action1.period"; From 2b796866af42583d284b807c502ab7fe5d156535 Mon Sep 17 00:00:00 2001 From: Lokesh Khurana Date: Mon, 2 Mar 2020 23:47:31 +0530 Subject: [PATCH 3/7] Making required classes LimitedPrivate with HIA.CHAOS and IS.Evolving --- .../org/apache/hadoop/hbase/HBaseInterfaceAudience.java | 5 +++++ .../main/java/org/apache/hadoop/hbase/ClusterManager.java | 5 +++-- .../org/apache/hadoop/hbase/DistributedHBaseCluster.java | 4 +++- .../java/org/apache/hadoop/hbase/HBaseClusterManager.java | 4 +++- .../org/apache/hadoop/hbase/IntegrationTestingUtility.java | 4 +++- .../java/org/apache/hadoop/hbase/chaos/actions/Action.java | 5 ++++- .../hadoop/hbase/chaos/actions/AddCPULoadAction.java | 5 ++++- .../apache/hadoop/hbase/chaos/actions/AddColumnAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/BatchRestartRsAction.java | 5 ++++- .../hbase/chaos/actions/ChangeBloomFilterAction.java | 5 ++++- .../hbase/chaos/actions/ChangeCompressionAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/ChangeEncodingAction.java | 5 ++++- .../hbase/chaos/actions/ChangeSplitPolicyAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/ChangeVersionsAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/CompactMobAction.java | 5 ++++- .../chaos/actions/CompactRandomRegionOfTableAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/CompactTableAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/CorruptDataFilesAction.java | 5 ++++- .../hbase/chaos/actions/CorruptPacketsCommandAction.java | 5 ++++- .../hbase/chaos/actions/DecreaseMaxHFileSizeAction.java | 5 ++++- .../hbase/chaos/actions/DelayPacketsCommandAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/DeleteDataFilesAction.java | 5 ++++- .../hbase/chaos/actions/DumpClusterStatusAction.java | 5 ++++- .../hbase/chaos/actions/DuplicatePacketsCommandAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/FillDiskCommandAction.java | 5 ++++- .../chaos/actions/FlushRandomRegionOfTableAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/FlushTableAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/ForceBalancerAction.java | 5 ++++- .../chaos/actions/GracefulRollingRestartRsAction.java | 5 ++++- .../hbase/chaos/actions/LosePacketsCommandAction.java | 5 ++++- .../actions/MergeRandomAdjacentRegionsOfTableAction.java | 5 ++++- .../hbase/chaos/actions/MoveRandomRegionOfTableAction.java | 5 ++++- .../hbase/chaos/actions/MoveRegionsOfTableAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/RemoveColumnAction.java | 5 ++++- .../hbase/chaos/actions/ReorderPacketsCommandAction.java | 5 ++++- .../hbase/chaos/actions/RestartActionBaseAction.java | 5 ++++- .../hbase/chaos/actions/RestartActiveMasterAction.java | 5 ++++- .../hbase/chaos/actions/RestartActiveNameNodeAction.java | 5 ++++- .../hbase/chaos/actions/RestartRandomDataNodeAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/RestartRandomRsAction.java | 5 ++++- .../chaos/actions/RestartRandomRsExceptMetaAction.java | 5 ++++- .../hbase/chaos/actions/RestartRandomZKNodeAction.java | 5 ++++- .../hbase/chaos/actions/RestartRsHoldingMetaAction.java | 5 ++++- .../hbase/chaos/actions/RestartRsHoldingTableAction.java | 5 ++++- .../hbase/chaos/actions/RollingBatchRestartRsAction.java | 5 ++++- .../actions/RollingBatchRestartRsExceptMetaAction.java | 5 ++++- .../chaos/actions/RollingBatchSuspendResumeRsAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/SnapshotTableAction.java | 5 ++++- .../hbase/chaos/actions/SplitAllRegionOfTableAction.java | 5 ++++- .../chaos/actions/SplitRandomRegionOfTableAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/SudoCommandAction.java | 5 ++++- .../apache/hadoop/hbase/chaos/actions/TCCommandAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/TruncateTableAction.java | 5 ++++- .../chaos/actions/UnbalanceKillAndRebalanceAction.java | 5 ++++- .../hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java | 5 ++++- .../hadoop/hbase/chaos/factories/MonkeyConstants.java | 7 +++++-- .../apache/hadoop/hbase/chaos/factories/MonkeyFactory.java | 5 ++++- .../apache/hadoop/hbase/chaos/monkies/CalmChaosMonkey.java | 5 ++++- .../org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java | 5 ++++- .../hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java | 5 ++++- .../hbase/chaos/policies/CompositeSequentialPolicy.java | 7 +++++-- .../hadoop/hbase/chaos/policies/DoActionsOncePolicy.java | 5 ++++- .../apache/hadoop/hbase/chaos/policies/PeriodicPolicy.java | 5 ++++- .../hbase/chaos/policies/PeriodicRandomActionPolicy.java | 5 ++++- .../org/apache/hadoop/hbase/chaos/policies/Policy.java | 5 ++++- .../hbase/chaos/policies/TwoConcurrentActionPolicy.java | 5 ++++- .../apache/hadoop/hbase/chaos/util/ChaosMonkeyRunner.java | 5 ++++- .../java/org/apache/hadoop/hbase/chaos/util/Monkeys.java | 5 ++++- 68 files changed, 271 insertions(+), 70 deletions(-) diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseInterfaceAudience.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseInterfaceAudience.java index 12b8398362ae..4117e0d9febd 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseInterfaceAudience.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseInterfaceAudience.java @@ -58,4 +58,9 @@ private HBaseInterfaceAudience(){} * Denotes classes that can be used to build custom authentication solutions. */ public static final String AUTHENTICATION = "Authentication"; + + /** + * Denotes classes that are being used to build Custom Chaos Monkey. + */ + public static final String CHAOS = "ChaosMonkey"; } diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/ClusterManager.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/ClusterManager.java index d6917c1ebdf8..878268b519d1 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/ClusterManager.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/ClusterManager.java @@ -22,14 +22,15 @@ import org.apache.hadoop.conf.Configurable; import org.apache.yetus.audience.InterfaceAudience; - +import org.apache.yetus.audience.InterfaceStability; /** * ClusterManager is an api to manage servers in a distributed environment. It provides services * for starting / stopping / killing Hadoop/HBase daemons. Concrete implementations provide actual * functionality for carrying out deployment-specific tasks. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public interface ClusterManager extends Configurable { /** * Type of the service daemon diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java index 865c5e2191ab..550db91dda29 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/DistributedHBaseCluster.java @@ -36,12 +36,14 @@ import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.util.Threads; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; /** * Manages the interactions with an already deployed distributed cluster (as opposed to * a pseudo-distributed, or mini/local cluster). This is used by integration and system tests. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class DistributedHBaseCluster extends HBaseCluster { private Admin admin; private final Connection connection; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/HBaseClusterManager.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/HBaseClusterManager.java index 1d361d52d14c..701648a0354c 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/HBaseClusterManager.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/HBaseClusterManager.java @@ -33,6 +33,7 @@ import org.apache.hadoop.hbase.util.RetryCounterFactory; import org.apache.hadoop.util.Shell; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -43,7 +44,8 @@ * servers on the remote machines (for example, the test user could be the same user as the * user the daemon is running as) */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class HBaseClusterManager extends Configured implements ClusterManager { private static final String SIGKILL = "SIGKILL"; private static final String SIGSTOP = "SIGSTOP"; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/IntegrationTestingUtility.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/IntegrationTestingUtility.java index 4862e72a96a9..ff175ea4a483 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/IntegrationTestingUtility.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/IntegrationTestingUtility.java @@ -22,6 +22,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.util.ReflectionUtils; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; /** * Facility for integration/system tests. This extends {@link HBaseTestingUtility} @@ -40,7 +41,8 @@ * via {@link #initializeCluster(int)}. Individual tests should not directly call * {@link #setUseDistributedCluster(Configuration)}. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class IntegrationTestingUtility extends HBaseTestingUtility { public IntegrationTestingUtility() { diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/Action.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/Action.java index d6e693ae0a59..151ddd21b83d 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/Action.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/Action.java @@ -32,6 +32,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.ClusterMetrics; import org.apache.hadoop.hbase.HBaseCluster; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.HRegionInfo; import org.apache.hadoop.hbase.IntegrationTestingUtility; @@ -47,13 +48,15 @@ import org.apache.hadoop.hbase.client.TableDescriptorBuilder; import org.apache.hadoop.hbase.util.Bytes; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * A (possibly mischievous) action that the ChaosMonkey can perform. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class Action { public static final String KILL_MASTER_TIMEOUT_KEY = diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddCPULoadAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddCPULoadAction.java index 3e89b51df049..f619a3518cf8 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddCPULoadAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddCPULoadAction.java @@ -20,16 +20,19 @@ import java.io.IOException; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that adds high cpu load to a random regionserver for a given duration */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class AddCPULoadAction extends SudoCommandAction { protected static final Logger LOG = LoggerFactory.getLogger(AddCPULoadAction.class); private static final String CPU_LOAD_COMMAND = diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddColumnAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddColumnAction.java index 6e3d65cbe3e9..a4a2cf3b6813 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddColumnAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/AddColumnAction.java @@ -21,6 +21,7 @@ import java.io.IOException; import org.apache.commons.lang3.RandomStringUtils; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; @@ -28,13 +29,15 @@ import org.apache.hadoop.hbase.client.TableDescriptor; import org.apache.hadoop.hbase.client.TableDescriptorBuilder; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action the adds a column family to a table. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class AddColumnAction extends Action { private final TableName tableName; private Admin admin; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java index c54cd3a847e0..beead72aa30a 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/BatchRestartRsAction.java @@ -22,16 +22,19 @@ import java.util.List; import java.util.Set; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Restarts a ratio of the running regionservers at the same time */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class BatchRestartRsAction extends RestartActionBaseAction { float ratio; //ratio of regionservers to restart private static final Logger LOG = diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeBloomFilterAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeBloomFilterAction.java index cb56b2fac27c..74adcb077d2d 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeBloomFilterAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeBloomFilterAction.java @@ -20,10 +20,12 @@ import java.util.Random; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.regionserver.BloomType; import org.apache.hadoop.hbase.util.BloomFilterUtil; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -31,7 +33,8 @@ * Action that tries to adjust the bloom filter setting on all the columns of a * table */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class ChangeBloomFilterAction extends Action { private final long sleepTime; private final TableName tableName; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeCompressionAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeCompressionAction.java index 3c1d4d60c3cc..afac8af23e1b 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeCompressionAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeCompressionAction.java @@ -21,17 +21,20 @@ import java.io.IOException; import java.util.Random; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.io.compress.Compression.Algorithm; import org.apache.hadoop.io.compress.Compressor; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that changes the compression algorithm on a column family from a list of tables. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class ChangeCompressionAction extends Action { private final TableName tableName; private final Random random; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeEncodingAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeEncodingAction.java index 4d9cfb4db861..6b32b30249b1 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeEncodingAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeEncodingAction.java @@ -21,16 +21,19 @@ import java.io.IOException; import java.util.Random; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.io.encoding.DataBlockEncoding; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that changes the encoding on a column family from a list of tables. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class ChangeEncodingAction extends Action { private final TableName tableName; private final Random random; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeSplitPolicyAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeSplitPolicyAction.java index 604f6ba9f798..d3558bbe6fe2 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeSplitPolicyAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeSplitPolicyAction.java @@ -19,6 +19,7 @@ import java.util.Random; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; @@ -28,10 +29,12 @@ import org.apache.hadoop.hbase.regionserver.DisabledRegionSplitPolicy; import org.apache.hadoop.hbase.regionserver.IncreasingToUpperBoundRegionSplitPolicy; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class ChangeSplitPolicyAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(ChangeSplitPolicyAction.class); private final TableName tableName; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeVersionsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeVersionsAction.java index 25617f452d22..7af5dd7e1fc0 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeVersionsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ChangeVersionsAction.java @@ -21,8 +21,10 @@ import java.io.IOException; import java.util.Random; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.TableName; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -31,7 +33,8 @@ * * Always keeps at least 1 as the number of versions. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class ChangeVersionsAction extends Action { private final TableName tableName; private static final Logger LOG = LoggerFactory.getLogger(ChangeVersionsAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactMobAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactMobAction.java index b7f1604174b7..f5ce538fc150 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactMobAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactMobAction.java @@ -19,18 +19,21 @@ package org.apache.hadoop.hbase.chaos.actions; import org.apache.commons.lang3.RandomUtils; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.CompactType; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that queues a table compaction. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class CompactMobAction extends Action { private final TableName tableName; private final int majorRatio; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactRandomRegionOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactRandomRegionOfTableAction.java index cba460b18f3f..48aacd4b23e4 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactRandomRegionOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactRandomRegionOfTableAction.java @@ -21,19 +21,22 @@ import java.util.List; import org.apache.commons.lang3.RandomUtils; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionInfo; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Region that queues a compaction of a random region from the table. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class CompactRandomRegionOfTableAction extends Action { private final int majorRatio; private final long sleepTime; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactTableAction.java index 53b9d9889035..abaeac577f15 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CompactTableAction.java @@ -19,17 +19,20 @@ package org.apache.hadoop.hbase.chaos.actions; import org.apache.commons.lang3.RandomUtils; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that queues a table compaction. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class CompactTableAction extends Action { private final TableName tableName; private final int majorRatio; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptDataFilesAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptDataFilesAction.java index c53a720d5256..2246a24946b3 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptDataFilesAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptDataFilesAction.java @@ -24,16 +24,19 @@ import org.apache.hadoop.fs.LocatedFileStatus; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.RemoteIterator; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.io.hfile.HFile; import org.apache.hadoop.hbase.util.CommonFSUtils; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action corrupts HFiles with a certain chance. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class CorruptDataFilesAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(CorruptDataFilesAction.class); private float chance; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptPacketsCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptPacketsCommandAction.java index d2927d76a1c3..da68f4adf291 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptPacketsCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/CorruptPacketsCommandAction.java @@ -20,9 +20,11 @@ import java.io.IOException; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -30,7 +32,8 @@ * * Corrupt network packets on a random regionserver. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class CorruptPacketsCommandAction extends TCCommandAction { private static final Logger LOG = LoggerFactory.getLogger(CorruptPacketsCommandAction.class); private float ratio; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DecreaseMaxHFileSizeAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DecreaseMaxHFileSizeAction.java index de762d1e45d5..85970c9e8544 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DecreaseMaxHFileSizeAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DecreaseMaxHFileSizeAction.java @@ -21,14 +21,17 @@ import java.io.IOException; import java.util.Random; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.TableDescriptor; import org.apache.hadoop.hbase.client.TableDescriptorBuilder; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class DecreaseMaxHFileSizeAction extends Action { private static final long minFileSize = 1 * 1024 * 1024 * 1024L; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DelayPacketsCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DelayPacketsCommandAction.java index d4e5b5a74295..e11c1e4c73f5 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DelayPacketsCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DelayPacketsCommandAction.java @@ -20,16 +20,19 @@ import java.io.IOException; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action adds latency to communication on a random regionserver. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class DelayPacketsCommandAction extends TCCommandAction { private static final Logger LOG = LoggerFactory.getLogger(DelayPacketsCommandAction.class); private long delay; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DeleteDataFilesAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DeleteDataFilesAction.java index b71788d8063b..f3cc138187fa 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DeleteDataFilesAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DeleteDataFilesAction.java @@ -23,16 +23,19 @@ import org.apache.hadoop.fs.LocatedFileStatus; import org.apache.hadoop.fs.Path; import org.apache.hadoop.fs.RemoteIterator; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.io.hfile.HFile; import org.apache.hadoop.hbase.util.CommonFSUtils; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action deletes HFiles with a certain chance. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class DeleteDataFilesAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(DeleteDataFilesAction.class); private float chance; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DumpClusterStatusAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DumpClusterStatusAction.java index 4cd0e5371175..640a0d51a06f 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DumpClusterStatusAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DumpClusterStatusAction.java @@ -20,14 +20,17 @@ import java.io.IOException; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action to dump the cluster status. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class DumpClusterStatusAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(DumpClusterStatusAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DuplicatePacketsCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DuplicatePacketsCommandAction.java index a2ad9bebf32f..37dbe6e0364c 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DuplicatePacketsCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/DuplicatePacketsCommandAction.java @@ -20,9 +20,11 @@ import java.io.IOException; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -30,7 +32,8 @@ * * Duplicate network packets on a random regionserver. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class DuplicatePacketsCommandAction extends TCCommandAction { private static final Logger LOG = LoggerFactory.getLogger(DuplicatePacketsCommandAction.class); private float ratio; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FillDiskCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FillDiskCommandAction.java index 8179d70ae822..7707e4523711 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FillDiskCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FillDiskCommandAction.java @@ -20,9 +20,11 @@ import java.io.IOException; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -30,7 +32,8 @@ * * Fill the disk on a random regionserver. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class FillDiskCommandAction extends SudoCommandAction { private static final Logger LOG = LoggerFactory.getLogger(FillDiskCommandAction.class); private long size; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushRandomRegionOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushRandomRegionOfTableAction.java index e5f18cbc506b..045038746dea 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushRandomRegionOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushRandomRegionOfTableAction.java @@ -19,19 +19,22 @@ package org.apache.hadoop.hbase.chaos.actions; import java.util.List; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionInfo; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to flush a random region of a table. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class FlushRandomRegionOfTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(FlushRandomRegionOfTableAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushTableAction.java index 1778f9454595..053243f2aab2 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/FlushTableAction.java @@ -18,17 +18,20 @@ package org.apache.hadoop.hbase.chaos.actions; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to flush a table. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class FlushTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(FlushTableAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ForceBalancerAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ForceBalancerAction.java index bcb16296c9a1..51870de1ae62 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ForceBalancerAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ForceBalancerAction.java @@ -18,14 +18,17 @@ package org.apache.hadoop.hbase.chaos.actions; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to force a balancer run. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class ForceBalancerAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(ForceBalancerAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/GracefulRollingRestartRsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/GracefulRollingRestartRsAction.java index 5bdc3199ddd9..224cd3eb844e 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/GracefulRollingRestartRsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/GracefulRollingRestartRsAction.java @@ -22,10 +22,12 @@ import java.util.Arrays; import java.util.List; import org.apache.commons.lang3.RandomUtils; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.util.RegionMover; import org.apache.hadoop.util.Shell; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -33,7 +35,8 @@ * Gracefully restarts every regionserver in a rolling fashion. At each step, it unloads, * restarts the loads every rs server sleeping randomly (0-sleepTime) in between servers. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class GracefulRollingRestartRsAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(GracefulRollingRestartRsAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/LosePacketsCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/LosePacketsCommandAction.java index b5c9a9cd4c03..7a982aace66f 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/LosePacketsCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/LosePacketsCommandAction.java @@ -20,9 +20,11 @@ import java.io.IOException; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -30,7 +32,8 @@ * * Lose network packets on a random regionserver. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class LosePacketsCommandAction extends TCCommandAction { private static final Logger LOG = LoggerFactory.getLogger(LosePacketsCommandAction.class); private float ratio; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MergeRandomAdjacentRegionsOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MergeRandomAdjacentRegionsOfTableAction.java index bbdfe743b979..f784c4e04bf1 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MergeRandomAdjacentRegionsOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MergeRandomAdjacentRegionsOfTableAction.java @@ -20,18 +20,21 @@ import java.util.List; import org.apache.commons.lang3.RandomUtils; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionInfo; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action to merge regions of a table. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class MergeRandomAdjacentRegionsOfTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(MergeRandomAdjacentRegionsOfTableAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRandomRegionOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRandomRegionOfTableAction.java index 8ebcd3caf84e..ba140cf6ceae 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRandomRegionOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRandomRegionOfTableAction.java @@ -20,19 +20,22 @@ import java.util.List; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionInfo; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to move a random region of a table. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class MoveRandomRegionOfTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(MoveRandomRegionOfTableAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRegionsOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRegionsOfTableAction.java index 27869de28258..1aff0f88f317 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRegionsOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/MoveRegionsOfTableAction.java @@ -24,19 +24,22 @@ import java.util.List; import org.apache.commons.lang3.RandomUtils; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.chaos.factories.MonkeyConstants; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionInfo; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to move every region of a table. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class MoveRegionsOfTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(MoveRegionsOfTableAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RemoveColumnAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RemoveColumnAction.java index 56f87062628d..3adbda90d314 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RemoveColumnAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RemoveColumnAction.java @@ -22,6 +22,7 @@ import java.util.Random; import java.util.Set; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.ColumnFamilyDescriptor; @@ -29,13 +30,15 @@ import org.apache.hadoop.hbase.client.TableDescriptorBuilder; import org.apache.hadoop.hbase.util.Bytes; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that removes a column family. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RemoveColumnAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(RemoveColumnAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ReorderPacketsCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ReorderPacketsCommandAction.java index 9316c58b57a2..afe7b8f74357 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ReorderPacketsCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/ReorderPacketsCommandAction.java @@ -20,9 +20,11 @@ import java.io.IOException; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -30,7 +32,8 @@ * * Reorder network packets on a random regionserver. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class ReorderPacketsCommandAction extends TCCommandAction { private static final Logger LOG = LoggerFactory.getLogger(ReorderPacketsCommandAction.class); private float ratio; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActionBaseAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActionBaseAction.java index b86332dc3bc7..da891a30bab3 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActionBaseAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActionBaseAction.java @@ -20,16 +20,19 @@ import java.io.IOException; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.util.Threads; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Base class for restarting HBaseServer's */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RestartActionBaseAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(RestartActionBaseAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveMasterAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveMasterAction.java index be1da5ce1886..bfe2802e831d 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveMasterAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveMasterAction.java @@ -18,15 +18,18 @@ package org.apache.hadoop.hbase.chaos.actions; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to restart the active master. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RestartActiveMasterAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartActionBaseAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveNameNodeAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveNameNodeAction.java index 5ceaa619a4b3..f1c141afb7ec 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveNameNodeAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartActiveNameNodeAction.java @@ -21,6 +21,7 @@ import java.util.List; import org.apache.hadoop.conf.Configuration; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.util.FSUtils; import org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper; @@ -31,13 +32,15 @@ import org.apache.hadoop.hdfs.HAUtil; import org.apache.hadoop.hdfs.server.namenode.ha.proto.HAZKInfoProtos.ActiveNodeInfo; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to restart the active namenode. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RestartActiveNameNodeAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartActiveNameNodeAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java index ae7f6135da2c..9a2a173db898 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomDataNodeAction.java @@ -22,6 +22,7 @@ import java.util.LinkedList; import java.util.List; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.util.FSUtils; @@ -30,13 +31,15 @@ import org.apache.hadoop.hdfs.protocol.DatanodeInfo; import org.apache.hadoop.hdfs.protocol.HdfsConstants; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that restarts a random datanode. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RestartRandomDataNodeAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartRandomDataNodeAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsAction.java index eca3a003ad5f..1bc6223c674b 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsAction.java @@ -18,16 +18,19 @@ package org.apache.hadoop.hbase.chaos.actions; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that restarts a random HRegionServer */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RestartRandomRsAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartRandomRsAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsExceptMetaAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsExceptMetaAction.java index 538cb63db3e9..13ac37eaabca 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsExceptMetaAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomRsExceptMetaAction.java @@ -18,11 +18,14 @@ package org.apache.hadoop.hbase.chaos.actions; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RestartRandomRsExceptMetaAction extends RestartRandomRsAction { public RestartRandomRsExceptMetaAction(long sleepTime) { super(sleepTime); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomZKNodeAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomZKNodeAction.java index aa923614688a..560ebe3ac0ee 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomZKNodeAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRandomZKNodeAction.java @@ -18,17 +18,20 @@ package org.apache.hadoop.hbase.chaos.actions; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.zookeeper.ZKServerTool; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that restarts a random zookeeper node. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RestartRandomZKNodeAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartRandomZKNodeAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingMetaAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingMetaAction.java index 0f167ac7827c..97adc059492f 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingMetaAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingMetaAction.java @@ -19,15 +19,18 @@ package org.apache.hadoop.hbase.chaos.actions; import org.apache.hadoop.hbase.ClusterMetrics; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to restart the HRegionServer holding Meta. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RestartRsHoldingMetaAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartRsHoldingMetaAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingTableAction.java index c06902464da5..80b4c5a81373 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RestartRsHoldingTableAction.java @@ -21,16 +21,19 @@ import java.util.List; import org.apache.commons.lang3.RandomUtils; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HRegionLocation; import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that restarts an HRegionServer holding one of the regions of the table. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RestartRsHoldingTableAction extends RestartActionBaseAction { private static final Logger LOG = LoggerFactory.getLogger(RestartRsHoldingTableAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java index fa7f5171f2c6..c8a54571c817 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsAction.java @@ -25,9 +25,11 @@ import java.util.Queue; import org.apache.commons.lang3.RandomUtils; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -37,7 +39,8 @@ * The parameter maxDeadServers limits the maximum number of servers that * can be down at the same time during rolling restarts. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RollingBatchRestartRsAction extends BatchRestartRsAction { private static final Logger LOG = LoggerFactory.getLogger(RollingBatchRestartRsAction.class); protected int maxDeadServers; // number of maximum dead servers at any given time. Defaults to 5 diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java index a575d1e37296..3666c05bf681 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchRestartRsExceptMetaAction.java @@ -20,14 +20,17 @@ import java.util.List; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; /** * Same as in {@link RollingBatchRestartRsAction} except that this action * does not restart the region server holding the META table. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RollingBatchRestartRsExceptMetaAction extends RollingBatchRestartRsAction { public RollingBatchRestartRsExceptMetaAction(long sleepTime, float ratio, int maxDeadServers) { diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchSuspendResumeRsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchSuspendResumeRsAction.java index 2720578dccfd..81f38b7ea6a7 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchSuspendResumeRsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/RollingBatchSuspendResumeRsAction.java @@ -24,11 +24,13 @@ import java.util.Queue; import org.apache.commons.lang3.RandomUtils; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.util.Threads; import org.apache.hadoop.util.Shell; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -38,7 +40,8 @@ * maxSuspendedServers limits the maximum number of servers that can be down at the same time * during rolling restarts. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class RollingBatchSuspendResumeRsAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(RollingBatchSuspendResumeRsAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SnapshotTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SnapshotTableAction.java index 1ebf42037beb..119ea7b61ffd 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SnapshotTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SnapshotTableAction.java @@ -18,17 +18,20 @@ package org.apache.hadoop.hbase.chaos.actions; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to take a snapshot of a table. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class SnapshotTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(SnapshotTableAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitAllRegionOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitAllRegionOfTableAction.java index aa9e8f0ed569..456ec2c2d35e 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitAllRegionOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitAllRegionOfTableAction.java @@ -20,14 +20,17 @@ import java.io.IOException; import java.util.concurrent.ThreadLocalRandom; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class SplitAllRegionOfTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(SplitAllRegionOfTableAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitRandomRegionOfTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitRandomRegionOfTableAction.java index d2f7090dc4bc..a4c9edbce8dd 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitRandomRegionOfTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SplitRandomRegionOfTableAction.java @@ -20,19 +20,22 @@ import java.util.List; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.client.Admin; import org.apache.hadoop.hbase.client.RegionInfo; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to split a random region of a table. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class SplitRandomRegionOfTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(SplitRandomRegionOfTableAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SudoCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SudoCommandAction.java index a8cbd87a3bf3..10d100908557 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SudoCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/SudoCommandAction.java @@ -23,14 +23,17 @@ import org.apache.hadoop.hbase.DistributedHBaseCluster; import org.apache.hadoop.hbase.HBaseCluster; import org.apache.hadoop.hbase.HBaseClusterManager; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Base class for performing Actions based on linux commands requiring sudo privileges */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving abstract public class SudoCommandAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(SudoCommandAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TCCommandAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TCCommandAction.java index 4754cf534a61..dc8affe26396 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TCCommandAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TCCommandAction.java @@ -18,12 +18,15 @@ package org.apache.hadoop.hbase.chaos.actions; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; /** * Base class for tc command actions */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving abstract public class TCCommandAction extends SudoCommandAction { protected static final String ADD = "add"; protected static final String DELETE = "del"; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TruncateTableAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TruncateTableAction.java index f8634f63c354..5322f2839438 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TruncateTableAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/TruncateTableAction.java @@ -20,17 +20,20 @@ import java.util.Random; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HBaseTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.client.Admin; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to truncate of a table. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class TruncateTableAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(TruncateTableAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java index 021e94c05c72..61f1fcf317ab 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceKillAndRebalanceAction.java @@ -26,14 +26,17 @@ import org.apache.commons.lang3.RandomUtils; import org.apache.hadoop.hbase.ClusterMetrics; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.junit.Assert; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** This action is too specific to put in ChaosMonkey; put it here */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class UnbalanceKillAndRebalanceAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(UnbalanceKillAndRebalanceAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java index 65b2417901aa..35779965c28d 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/actions/UnbalanceRegionsAction.java @@ -24,15 +24,18 @@ import org.apache.commons.lang3.RandomUtils; import org.apache.hadoop.hbase.ClusterMetrics; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.ServerName; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Action that tries to unbalance the regions of a cluster. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class UnbalanceRegionsAction extends Action { private static final Logger LOG = LoggerFactory.getLogger(UnbalanceRegionsAction.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java index b8e69b40d637..ed94d615aa52 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyConstants.java @@ -17,12 +17,15 @@ */ package org.apache.hadoop.hbase.chaos.factories; -import org.apache.yetus.audience.InterfaceAudience; import java.util.Arrays; import java.util.HashSet; import java.util.Set; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; +import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public interface MonkeyConstants { String PERIODIC_ACTION1_PERIOD = "sdm.action1.period"; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyFactory.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyFactory.java index 1a6e5b234f9a..796c7cdf8d3e 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyFactory.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/factories/MonkeyFactory.java @@ -22,6 +22,7 @@ import java.util.Properties; import java.util.Set; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.IntegrationTestingUtility; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.chaos.monkies.ChaosMonkey; @@ -29,13 +30,15 @@ import org.apache.hbase.thirdparty.com.google.common.collect.ImmutableMap; import org.apache.hadoop.hbase.util.ReflectionUtils; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Base class of the factory that will create a ChaosMonkey. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public abstract class MonkeyFactory { private static final Logger LOG = LoggerFactory.getLogger(MonkeyFactory.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/CalmChaosMonkey.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/CalmChaosMonkey.java index 7a45a366bd91..89b7579969db 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/CalmChaosMonkey.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/CalmChaosMonkey.java @@ -18,12 +18,15 @@ package org.apache.hadoop.hbase.chaos.monkies; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; /** * Chaos Monkey that does nothing. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class CalmChaosMonkey extends ChaosMonkey { @Override public void start() throws Exception { diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java index 682a0902f59c..02d2d95a2114 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java @@ -19,7 +19,9 @@ package org.apache.hadoop.hbase.chaos.monkies; import org.apache.hadoop.hbase.Stoppable; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; /** * A utility to injects faults in a running cluster. @@ -42,7 +44,8 @@ * ChaosMonkey class is indeed inspired by the Netflix's same-named tool: * http://techblog.netflix.com/2012/07/chaos-monkey-released-into-wild.html */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public abstract class ChaosMonkey implements Stoppable { public abstract void start() throws Exception; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java index d5ebe21f9412..2393af52103f 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/PolicyBasedChaosMonkey.java @@ -24,17 +24,20 @@ import java.util.List; import org.apache.commons.lang3.RandomUtils; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.IntegrationTestingUtility; import org.apache.hadoop.hbase.chaos.policies.Policy; import org.apache.hadoop.hbase.util.Pair; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * Chaos monkey that given multiple policies will run actions against the cluster. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class PolicyBasedChaosMonkey extends ChaosMonkey { private static final Logger LOG = LoggerFactory.getLogger(PolicyBasedChaosMonkey.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/CompositeSequentialPolicy.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/CompositeSequentialPolicy.java index c8eadbfb27a2..77f878dcf933 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/CompositeSequentialPolicy.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/CompositeSequentialPolicy.java @@ -18,12 +18,15 @@ package org.apache.hadoop.hbase.chaos.policies; -import org.apache.yetus.audience.InterfaceAudience; import java.util.Arrays; import java.util.List; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; +import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; /** A policy that runs multiple other policies one after the other */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class CompositeSequentialPolicy extends Policy { private List policies; public CompositeSequentialPolicy(Policy... policies) { diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java index 03a92e5e66b2..8076485fa1cb 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/DoActionsOncePolicy.java @@ -22,12 +22,15 @@ import java.util.Arrays; import java.util.List; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.chaos.actions.Action; import org.apache.hadoop.util.StringUtils; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; /** A policy which performs a sequence of actions deterministically. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class DoActionsOncePolicy extends PeriodicPolicy { private List actions; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicPolicy.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicPolicy.java index a699b3036054..2c271246d513 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicPolicy.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicPolicy.java @@ -19,11 +19,14 @@ package org.apache.hadoop.hbase.chaos.policies; import org.apache.commons.lang3.RandomUtils; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.util.Threads; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; /** A policy which does stuff every time interval. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public abstract class PeriodicPolicy extends Policy { private long periodMs; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java index 7f1c40fc69c1..988d648a2c1c 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/PeriodicRandomActionPolicy.java @@ -22,17 +22,20 @@ import java.util.Arrays; import java.util.List; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.chaos.actions.Action; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.util.Pair; import org.apache.hadoop.util.StringUtils; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; /** * A policy, which picks a random action according to the given weights, * and performs it every configurable period. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class PeriodicRandomActionPolicy extends PeriodicPolicy { private List> actions; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/Policy.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/Policy.java index 12a6c652dae4..5bac3e9ec251 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/Policy.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/Policy.java @@ -18,17 +18,20 @@ package org.apache.hadoop.hbase.chaos.policies; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.IntegrationTestingUtility; import org.apache.hadoop.hbase.chaos.actions.Action; import org.apache.hadoop.hbase.util.StoppableImplementation; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; /** * A policy to introduce chaos to the cluster */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public abstract class Policy extends StoppableImplementation implements Runnable { protected static final Logger LOG = LoggerFactory.getLogger(Policy.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/TwoConcurrentActionPolicy.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/TwoConcurrentActionPolicy.java index 98c7d62189c9..9661e6b9522b 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/TwoConcurrentActionPolicy.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/policies/TwoConcurrentActionPolicy.java @@ -18,11 +18,13 @@ package org.apache.hadoop.hbase.chaos.policies; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.chaos.actions.Action; import org.apache.hadoop.hbase.chaos.monkies.PolicyBasedChaosMonkey; import org.apache.hadoop.hbase.util.Threads; import org.apache.hadoop.util.StringUtils; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import java.util.concurrent.ExecutionException; import java.util.concurrent.ExecutorService; @@ -33,7 +35,8 @@ * Chaos Monkey policy that will run two different actions at the same time. * A random action from each array of actions will be chosen and then run in parallel. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class TwoConcurrentActionPolicy extends PeriodicPolicy { private final Action[] actionsOne; private final Action[] actionsTwo; diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/ChaosMonkeyRunner.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/ChaosMonkeyRunner.java index d9c7455697ed..a1ebe6375b44 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/ChaosMonkeyRunner.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/ChaosMonkeyRunner.java @@ -24,6 +24,7 @@ import org.apache.commons.lang3.StringUtils; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.HConstants; import org.apache.hadoop.hbase.IntegrationTestingUtility; import org.apache.hadoop.hbase.TableName; @@ -32,12 +33,14 @@ import org.apache.hadoop.hbase.util.AbstractHBaseTool; import org.apache.hadoop.util.ToolRunner; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.hbase.thirdparty.com.google.common.collect.Sets; import org.apache.hbase.thirdparty.org.apache.commons.cli.CommandLine; -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class ChaosMonkeyRunner extends AbstractHBaseTool { private static final Logger LOG = LoggerFactory.getLogger(ChaosMonkeyRunner.class); diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/Monkeys.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/Monkeys.java index 470ebd7aca4d..debf33abf4f5 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/Monkeys.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/util/Monkeys.java @@ -26,8 +26,10 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; +import org.apache.hadoop.hbase.HBaseInterfaceAudience; import org.apache.hadoop.hbase.IntegrationTestingUtility; import org.apache.yetus.audience.InterfaceAudience; +import org.apache.yetus.audience.InterfaceStability; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -37,7 +39,8 @@ /** * This class can be used to control chaos monkeys life cycle. */ -@InterfaceAudience.Public +@InterfaceAudience.LimitedPrivate(HBaseInterfaceAudience.CHAOS) +@InterfaceStability.Evolving public class Monkeys implements Closeable { private static final Logger LOG = LoggerFactory.getLogger(Monkeys.class); From a1dd3ec88789c421f953da989063eb55531d97a7 Mon Sep 17 00:00:00 2001 From: Lokesh Khurana Date: Tue, 21 Apr 2020 23:47:42 +0530 Subject: [PATCH 4/7] moving all classes from root to src --- hbase-common/pom.xml | 4 + .../hbase/HBaseCommonTestingUtility.java | 0 .../java/org/apache/hadoop/hbase/Waiter.java | 4 +- .../hadoop/hbase}/MetricsAssertHelper.java | 4 +- .../hbase/test/MetricsAssertHelperImpl.java | 1 + ...g.apache.hadoop.hbase.MetricsAssertHelper} | 0 hbase-http/pom.xml | 6 + .../hbase/http/ssl/KeyStoreTestUtil.java | 2 + hbase-it/pom.xml | 7 +- .../src/main/resources/META-INF/LICENSE.vm | 2 +- .../hbase/rest/TestGetAndPutResource.java | 2 +- hbase-server/pom.xml | 27 ++++ .../org/apache/hadoop/hbase/HBaseCluster.java | 2 +- .../hadoop/hbase/HBaseTestingUtility.java | 137 +++++++++--------- .../apache/hadoop/hbase/MiniHBaseCluster.java | 1 - .../hbase/MockRegionServerServices.java | 2 + .../hadoop/hbase/StartMiniClusterOption.java | 0 .../hbase/mapreduce/MapreduceTestingShim.java | 2 + .../assignment/AssignmentTestingUtil.java | 0 .../hbase/security/HBaseKerberosUtils.java | 2 +- .../hbase/util/StoppableImplementation.java | 0 .../hbase/client/TestMultiRespectsLimits.java | 2 +- .../apache/hadoop/hbase/io/TestMetricsIO.java | 2 +- .../hadoop/hbase/ipc/TestRpcMetrics.java | 2 +- .../master/TestAssignmentManagerMetrics.java | 2 +- .../hbase/master/TestMasterMetrics.java | 2 +- .../hbase/regionserver/TestHRegion.java | 2 +- .../TestMetricsHeapMemoryManager.java | 2 +- .../hbase/regionserver/TestMetricsRegion.java | 2 +- .../regionserver/TestMetricsRegionServer.java | 2 +- .../TestMetricsTableAggregate.java | 2 +- .../TestMetricsTableLatencies.java | 2 +- .../TestMetricsUserAggregate.java | 2 +- .../regionserver/TestRegionServerMetrics.java | 2 +- .../regionserver/TestRemoveRegionMetrics.java | 2 +- .../hadoop/hbase/thrift/TestCallQueue.java | 2 +- .../hadoop/hbase/thrift/TestThriftServer.java | 2 +- .../TestThriftHBaseServiceHandler.java | 2 +- .../hadoop/hbase/HBaseZKTestingUtility.java | 4 +- 39 files changed, 142 insertions(+), 101 deletions(-) rename hbase-common/src/{test => main}/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java (100%) rename hbase-common/src/{test => main}/java/org/apache/hadoop/hbase/Waiter.java (99%) rename hbase-hadoop-compat/src/{test/java/org/apache/hadoop/hbase/test => main/java/org/apache/hadoop/hbase}/MetricsAssertHelper.java (98%) rename hbase-hadoop2-compat/src/test/resources/META-INF/services/{org.apache.hadoop.hbase.test.MetricsAssertHelper => org.apache.hadoop.hbase.MetricsAssertHelper} (100%) rename hbase-http/src/{test => main}/java/org/apache/hadoop/hbase/http/ssl/KeyStoreTestUtil.java (99%) rename hbase-server/src/{test => main}/java/org/apache/hadoop/hbase/HBaseCluster.java (99%) rename hbase-server/src/{test => main}/java/org/apache/hadoop/hbase/HBaseTestingUtility.java (96%) rename hbase-server/src/{test => main}/java/org/apache/hadoop/hbase/MiniHBaseCluster.java (99%) rename hbase-server/src/{test => main}/java/org/apache/hadoop/hbase/MockRegionServerServices.java (99%) rename hbase-server/src/{test => main}/java/org/apache/hadoop/hbase/StartMiniClusterOption.java (100%) rename hbase-server/src/{test => main}/java/org/apache/hadoop/hbase/mapreduce/MapreduceTestingShim.java (98%) rename hbase-server/src/{test => main}/java/org/apache/hadoop/hbase/master/assignment/AssignmentTestingUtil.java (100%) rename hbase-server/src/{test => main}/java/org/apache/hadoop/hbase/security/HBaseKerberosUtils.java (100%) rename hbase-server/src/{test => main}/java/org/apache/hadoop/hbase/util/StoppableImplementation.java (100%) rename hbase-zookeeper/src/{test => main}/java/org/apache/hadoop/hbase/HBaseZKTestingUtility.java (96%) diff --git a/hbase-common/pom.xml b/hbase-common/pom.xml index f9d039a5ba53..d11bf7905048 100644 --- a/hbase-common/pom.xml +++ b/hbase-common/pom.xml @@ -216,6 +216,10 @@ log4j log4j + + junit + junit + diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java similarity index 100% rename from hbase-common/src/test/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java rename to hbase-common/src/main/java/org/apache/hadoop/hbase/HBaseCommonTestingUtility.java diff --git a/hbase-common/src/test/java/org/apache/hadoop/hbase/Waiter.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/Waiter.java similarity index 99% rename from hbase-common/src/test/java/org/apache/hadoop/hbase/Waiter.java rename to hbase-common/src/main/java/org/apache/hadoop/hbase/Waiter.java index 5302093b66a0..de6c0bfe95d9 100644 --- a/hbase-common/src/test/java/org/apache/hadoop/hbase/Waiter.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/Waiter.java @@ -18,12 +18,12 @@ */ package org.apache.hadoop.hbase; -import static org.junit.Assert.fail; import java.text.MessageFormat; import org.apache.hadoop.conf.Configuration; import org.apache.yetus.audience.InterfaceAudience; +import org.junit.Assert; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -200,7 +200,7 @@ public static long waitFor(Configuration conf, long timeou System.currentTimeMillis() - started)); } else if (failIfTimeout) { String msg = getExplanation(predicate); - fail(MessageFormat + Assert.fail(MessageFormat .format("Waiting timed out after [{0}] msec", adjustedTimeout) + msg); } else { String msg = getExplanation(predicate); diff --git a/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelper.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/MetricsAssertHelper.java similarity index 98% rename from hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelper.java rename to hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/MetricsAssertHelper.java index 52e0d091279f..e4d5f20f745d 100644 --- a/hbase-hadoop-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelper.java +++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/MetricsAssertHelper.java @@ -16,11 +16,13 @@ * limitations under the License. */ -package org.apache.hadoop.hbase.test; +package org.apache.hadoop.hbase; import org.apache.hadoop.hbase.metrics.BaseSource; +import org.apache.yetus.audience.InterfaceAudience; /** Interface of a class to make assertions about metrics values. */ +@InterfaceAudience.Private public interface MetricsAssertHelper { /** diff --git a/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java b/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java index afe31f166671..523eae0d9cae 100644 --- a/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java +++ b/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java @@ -26,6 +26,7 @@ import java.util.Locale; import java.util.Map; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.metrics.BaseSource; import org.apache.hadoop.metrics2.AbstractMetric; import org.apache.hadoop.metrics2.MetricsCollector; diff --git a/hbase-hadoop2-compat/src/test/resources/META-INF/services/org.apache.hadoop.hbase.test.MetricsAssertHelper b/hbase-hadoop2-compat/src/test/resources/META-INF/services/org.apache.hadoop.hbase.MetricsAssertHelper similarity index 100% rename from hbase-hadoop2-compat/src/test/resources/META-INF/services/org.apache.hadoop.hbase.test.MetricsAssertHelper rename to hbase-hadoop2-compat/src/test/resources/META-INF/services/org.apache.hadoop.hbase.MetricsAssertHelper diff --git a/hbase-http/pom.xml b/hbase-http/pom.xml index 0bf21d292708..e375929eeb96 100644 --- a/hbase-http/pom.xml +++ b/hbase-http/pom.xml @@ -256,6 +256,12 @@ hadoop-minikdc test + + bcprov-jdk15 + bouncycastle + 140 + compile + diff --git a/hbase-http/src/test/java/org/apache/hadoop/hbase/http/ssl/KeyStoreTestUtil.java b/hbase-http/src/main/java/org/apache/hadoop/hbase/http/ssl/KeyStoreTestUtil.java similarity index 99% rename from hbase-http/src/test/java/org/apache/hadoop/hbase/http/ssl/KeyStoreTestUtil.java rename to hbase-http/src/main/java/org/apache/hadoop/hbase/http/ssl/KeyStoreTestUtil.java index c8abd9c6cebc..cc08e4e8c97b 100644 --- a/hbase-http/src/test/java/org/apache/hadoop/hbase/http/ssl/KeyStoreTestUtil.java +++ b/hbase-http/src/main/java/org/apache/hadoop/hbase/http/ssl/KeyStoreTestUtil.java @@ -46,8 +46,10 @@ import org.apache.hadoop.hbase.HBaseCommonTestingUtility; import org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory; import org.apache.hadoop.security.ssl.SSLFactory; +import org.apache.yetus.audience.InterfaceAudience; import org.bouncycastle.x509.X509V1CertificateGenerator; +@InterfaceAudience.Private public final class KeyStoreTestUtil { private KeyStoreTestUtil() { } diff --git a/hbase-it/pom.xml b/hbase-it/pom.xml index 8e6f3e40fa4c..50749a994da9 100644 --- a/hbase-it/pom.xml +++ b/hbase-it/pom.xml @@ -158,25 +158,24 @@ org.apache.hbase hbase-server test-jar - compile + test org.apache.hbase hbase-zookeeper test-jar - compile + test org.apache.hbase hbase-common jar - compile org.apache.hbase hbase-common test-jar - compile + test org.apache.hbase diff --git a/hbase-resource-bundle/src/main/resources/META-INF/LICENSE.vm b/hbase-resource-bundle/src/main/resources/META-INF/LICENSE.vm index 760208f4b863..a0cfeb6484b0 100644 --- a/hbase-resource-bundle/src/main/resources/META-INF/LICENSE.vm +++ b/hbase-resource-bundle/src/main/resources/META-INF/LICENSE.vm @@ -1343,7 +1343,7 @@ You can redistribute it and/or modify it under either the terms of the ## See this FAQ link for justifications: https://www.apache.org/legal/resolved.html ## ## NB: This list is later compared as lower-case. New entries must also be all lower-case -#set($non_aggregate_fine = [ 'public domain', 'new bsd license', 'bsd license', 'bsd', 'bsd 2-clause license', 'mozilla public license version 1.1', 'mozilla public license version 2.0', 'creative commons attribution license, version 2.5' ]) +#set($non_aggregate_fine = [ 'public domain', 'new bsd license', 'bsd license', 'bsd', 'bsd 2-clause license', 'mozilla public license version 1.1', 'mozilla public license version 2.0', 'creative commons attribution license, version 2.5', 'bouncy castle license' ]) ## include LICENSE sections for anything not under ASL2.0 #foreach( ${dep} in ${projects} ) ## if there are no licenses we'll fail the build later, so diff --git a/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestGetAndPutResource.java b/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestGetAndPutResource.java index e1dec900d491..de1da94b46e7 100644 --- a/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestGetAndPutResource.java +++ b/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestGetAndPutResource.java @@ -35,7 +35,7 @@ import org.apache.hadoop.hbase.rest.model.CellSetModel; import org.apache.hadoop.hbase.rest.model.RowModel; import org.apache.hadoop.hbase.security.UserProvider; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.RestTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/pom.xml b/hbase-server/pom.xml index 05813f5b0673..1c6159901c0c 100644 --- a/hbase-server/pom.xml +++ b/hbase-server/pom.xml @@ -515,6 +515,27 @@ mockito-core test + + junit + junit + + + org.apache.hadoop + hadoop-mapreduce-client-jobclient + ${hadoop.version} + tests + + + com.google.code.findbugs + jsr305 + + + + + org.apache.hadoop + hadoop-minikdc + compile + @@ -660,6 +681,12 @@ test-jar test + + org.apache.hadoop + hadoop-hdfs + test-jar + compile + org.apache.hadoop hadoop-minicluster diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseCluster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseCluster.java similarity index 99% rename from hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseCluster.java rename to hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseCluster.java index 85dff357ca62..fdfdb86d4d96 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseCluster.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseCluster.java @@ -51,7 +51,7 @@ * In that sense, this class does not abstract away every interface that * MiniHBaseCluster or DistributedHBaseCluster provide. */ -@InterfaceAudience.Private +@InterfaceAudience.Public public abstract class HBaseCluster implements Closeable, Configurable { // Log is being used in DistributedHBaseCluster class, hence keeping it as package scope static final Logger LOG = LoggerFactory.getLogger(HBaseCluster.class.getName()); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseTestingUtility.java similarity index 96% rename from hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java rename to hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseTestingUtility.java index 834fbac77c20..e8c476e1acae 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/HBaseTestingUtility.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseTestingUtility.java @@ -17,11 +17,8 @@ */ package org.apache.hadoop.hbase; -import static org.junit.Assert.assertEquals; -import static org.junit.Assert.assertTrue; -import static org.junit.Assert.fail; - import edu.umd.cs.findbugs.annotations.Nullable; + import java.io.File; import java.io.IOException; import java.io.OutputStream; @@ -50,6 +47,7 @@ import java.util.TreeSet; import java.util.concurrent.TimeUnit; import java.util.concurrent.atomic.AtomicReference; + import org.apache.commons.io.FileUtils; import org.apache.commons.lang3.RandomStringUtils; import org.apache.commons.logging.impl.Jdk14Logger; @@ -147,12 +145,11 @@ import org.apache.zookeeper.WatchedEvent; import org.apache.zookeeper.ZooKeeper; import org.apache.zookeeper.ZooKeeper.States; +import org.junit.Assert; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.slf4j.impl.Log4jLoggerAdapter; - import org.apache.hbase.thirdparty.com.google.common.io.Closeables; - import org.apache.hadoop.hbase.shaded.protobuf.ProtobufUtil; /** @@ -335,7 +332,7 @@ public HBaseTestingUtility(@Nullable Configuration conf) { String dataTestDir = getDataTestDir().toString(); this.conf.set("fs.defaultFS","file:///"); this.conf.set(HConstants.HBASE_DIR, "file://" + dataTestDir); - LOG.debug("Setting {} to {}", HConstants.HBASE_DIR, dataTestDir); + HBaseCommonTestingUtility.LOG.debug("Setting {} to {}", HConstants.HBASE_DIR, dataTestDir); this.conf.setBoolean(CommonFSUtils.UNSAFE_STREAM_CAPABILITY_ENFORCE,false); // If the value for random ports isn't set set it to true, thus making // tests opt-out for random port assignment @@ -429,11 +426,11 @@ private void createSubDirAndSystemProperty( if (sysValue != null) { // There is already a value set. So we do nothing but hope // that there will be no conflicts - LOG.info("System.getProperty(\""+propertyName+"\") already set to: "+ + HBaseCommonTestingUtility.LOG.info("System.getProperty(\""+propertyName+"\") already set to: "+ sysValue + " so I do NOT create it in " + parent); String confValue = conf.get(propertyName); if (confValue != null && !confValue.endsWith(sysValue)){ - LOG.warn( + HBaseCommonTestingUtility.LOG.warn( propertyName + " property value differs in configuration and system: "+ "Configuration="+confValue+" while System="+sysValue+ " Erasing configuration value by system value." @@ -489,7 +486,7 @@ public Path getDataTestDirOnTestFS(final String subdirName) throws IOException { */ private void setupDataTestDirOnTestFS() throws IOException { if (dataTestDirOnTestFS != null) { - LOG.warn("Data test on test fs dir already setup in " + HBaseCommonTestingUtility.LOG.warn("Data test on test fs dir already setup in " + dataTestDirOnTestFS.toString()); return; } @@ -506,7 +503,7 @@ private Path getNewDataTestDirOnTestFS() throws IOException { //the working directory, and create a unique sub dir there FileSystem fs = getTestFileSystem(); Path newDataTestDir; - String randomStr = getRandomUUID().toString(); + String randomStr = HBaseCommonTestingUtility.getRandomUUID().toString(); if (fs.getUri().getScheme().equals(FileSystem.getLocal(conf).getUri().getScheme())) { newDataTestDir = new Path(getDataTestDir(), randomStr); File dataTestDir = new File(newDataTestDir.toString()); @@ -588,7 +585,7 @@ public MiniDFSCluster startMiniDFSCluster(int servers, final String hosts[]) private void setFs() throws IOException { if(this.dfsCluster == null){ - LOG.info("Skipping setting fs because dfsCluster is null"); + HBaseCommonTestingUtility.LOG.info("Skipping setting fs because dfsCluster is null"); return; } FileSystem fs = this.dfsCluster.getFileSystem(); @@ -624,7 +621,7 @@ public MiniDFSCluster startMiniDFSCluster(int servers, final String racks[], St dataTestDirOnTestFS = null; String dataTestDir = getDataTestDir().toString(); conf.set(HConstants.HBASE_DIR, dataTestDir); - LOG.debug("Setting {} to {}", HConstants.HBASE_DIR, dataTestDir); + HBaseCommonTestingUtility.LOG.debug("Setting {} to {}", HConstants.HBASE_DIR, dataTestDir); return this.dfsCluster; } @@ -693,7 +690,7 @@ public boolean isReadShortCircuitOn(){ private void enableShortCircuit() { if (isReadShortCircuitOn()) { String curUser = System.getProperty("user.name"); - LOG.info("read short circuit is ON for user " + curUser); + HBaseCommonTestingUtility.LOG.info("read short circuit is ON for user " + curUser); // read short circuit, for hdfs conf.set("dfs.block.local-path-access.user", curUser); // read short circuit, for hbase @@ -701,7 +698,7 @@ private void enableShortCircuit() { // Skip checking checksum, for the hdfs client and the datanode conf.setBoolean("dfs.client.read.shortcircuit.skip.checksum", true); } else { - LOG.info("read short circuit is OFF"); + HBaseCommonTestingUtility.LOG.info("read short circuit is OFF"); } } @@ -710,7 +707,7 @@ private String createDirAndSetProperty(final String relPath, String property) { System.setProperty(property, path); conf.set(property, path); new File(path).mkdirs(); - LOG.info("Setting " + property + " to " + path + " in system properties and HBase conf"); + HBaseCommonTestingUtility.LOG.info("Setting " + property + " to " + path + " in system properties and HBase conf"); return path; } @@ -1021,7 +1018,7 @@ public MiniHBaseCluster startMiniCluster() throws Exception { * @see #shutdownMiniDFSCluster() */ public MiniHBaseCluster startMiniCluster(StartMiniClusterOption option) throws Exception { - LOG.info("Starting up minicluster with option: {}", option); + HBaseCommonTestingUtility.LOG.info("Starting up minicluster with option: {}", option); // If we already put up a cluster, fail. if (miniClusterRunning) { @@ -1035,10 +1032,10 @@ public MiniHBaseCluster startMiniCluster(StartMiniClusterOption option) throws E // Bring up mini dfs cluster. This spews a bunch of warnings about missing // scheme. Complaints are 'Scheme is undefined for build/test/data/dfs/name1'. if (dfsCluster == null) { - LOG.info("STARTING DFS"); + HBaseCommonTestingUtility.LOG.info("STARTING DFS"); dfsCluster = startMiniDFSCluster(option.getNumDataNodes(), option.getDataNodeHosts()); } else { - LOG.info("NOT STARTING DFS"); + HBaseCommonTestingUtility.LOG.info("NOT STARTING DFS"); } // Start up a zk cluster. @@ -1096,7 +1093,7 @@ public MiniHBaseCluster startMiniHBaseCluster(StartMiniClusterOption option) t.close(); getAdmin(); // create immediately the hbaseAdmin - LOG.info("Minicluster is up; activeMaster={}", getHBaseCluster().getMaster()); + HBaseCommonTestingUtility.LOG.info("Minicluster is up; activeMaster={}", getHBaseCluster().getMaster()); return (MiniHBaseCluster) hbaseCluster; } @@ -1222,7 +1219,7 @@ public void restartHBaseCluster(StartMiniClusterOption option) while (s.next() != null) { // do nothing } - LOG.info("HBase has been restarted"); + HBaseCommonTestingUtility.LOG.info("HBase has been restarted"); s.close(); t.close(); conn.close(); @@ -1246,14 +1243,14 @@ public MiniHBaseCluster getMiniHBaseCluster() { * @see #startMiniCluster(int) */ public void shutdownMiniCluster() throws IOException { - LOG.info("Shutting down minicluster"); + HBaseCommonTestingUtility.LOG.info("Shutting down minicluster"); shutdownMiniHBaseCluster(); shutdownMiniDFSCluster(); shutdownMiniZKCluster(); cleanupTestDir(); miniClusterRunning = false; - LOG.info("Minicluster is down"); + HBaseCommonTestingUtility.LOG.info("Minicluster is down"); } /** @@ -1377,9 +1374,9 @@ private void setHBaseFsTmpDir() throws IOException { String hbaseFsTmpDirInString = this.conf.get("hbase.fs.tmp.dir"); if (hbaseFsTmpDirInString == null) { this.conf.set("hbase.fs.tmp.dir", getDataTestDirOnTestFS("hbase-staging").toString()); - LOG.info("Setting hbase.fs.tmp.dir to " + this.conf.get("hbase.fs.tmp.dir")); + HBaseCommonTestingUtility.LOG.info("Setting hbase.fs.tmp.dir to " + this.conf.get("hbase.fs.tmp.dir")); } else { - LOG.info("The hbase.fs.tmp.dir is set to " + hbaseFsTmpDirInString); + HBaseCommonTestingUtility.LOG.info("The hbase.fs.tmp.dir is set to " + hbaseFsTmpDirInString); } } @@ -1871,7 +1868,7 @@ public void deleteTable(TableName tableName) throws IOException { try { getAdmin().disableTable(tableName); } catch (TableNotEnabledException e) { - LOG.debug("Table: " + tableName + " already disabled, so just deleting it."); + HBaseCommonTestingUtility.LOG.debug("Table: " + tableName + " already disabled, so just deleting it."); } getAdmin().deleteTable(tableName); } @@ -2322,10 +2319,10 @@ public void verifyNumericRows(Table table, final byte[] f, int startRow, int end get.setReplicaId(replicaId); get.setConsistency(Consistency.TIMELINE); Result result = table.get(get); - assertTrue(failMsg, result.containsColumn(f, null)); - assertEquals(failMsg, 1, result.getColumnCells(f, null).size()); + Assert.assertTrue(failMsg, result.containsColumn(f, null)); + Assert.assertEquals(failMsg, 1, result.getColumnCells(f, null).size()); Cell cell = result.getColumnLatestCell(f, null); - assertTrue(failMsg, + Assert.assertTrue(failMsg, Bytes.equals(data, 0, data.length, cell.getValueArray(), cell.getValueOffset(), cell.getValueLength())); } @@ -2354,13 +2351,13 @@ public void verifyNumericRows(HRegion region, final byte[] f, int startRow, int Result result = region.get(new Get(data)); boolean hasResult = result != null && !result.isEmpty(); - assertEquals(failMsg + result, present, hasResult); + Assert.assertEquals(failMsg + result, present, hasResult); if (!present) continue; - assertTrue(failMsg, result.containsColumn(f, null)); - assertEquals(failMsg, 1, result.getColumnCells(f, null).size()); + Assert.assertTrue(failMsg, result.containsColumn(f, null)); + Assert.assertEquals(failMsg, 1, result.getColumnCells(f, null).size()); Cell cell = result.getColumnLatestCell(f, null); - assertTrue(failMsg, + Assert.assertTrue(failMsg, Bytes.equals(data, 0, data.length, cell.getValueArray(), cell.getValueOffset(), cell.getValueLength())); } @@ -2598,7 +2595,7 @@ public List getMetaTableRows() throws IOException { List rows = new ArrayList<>(); ResultScanner s = t.getScanner(new Scan()); for (Result result : s) { - LOG.info("getMetaTableRows: row -> " + + HBaseCommonTestingUtility.LOG.info("getMetaTableRows: row -> " + Bytes.toStringBinary(result.getRow())); rows.add(result.getRow()); } @@ -2620,13 +2617,13 @@ public List getMetaTableRows(TableName tableName) throws IOException { for (Result result : s) { RegionInfo info = MetaTableAccessor.getRegionInfo(result); if (info == null) { - LOG.error("No region info for row " + Bytes.toString(result.getRow())); + HBaseCommonTestingUtility.LOG.error("No region info for row " + Bytes.toString(result.getRow())); // TODO figure out what to do for this new hosed case. continue; } if (info.getTable().equals(tableName)) { - LOG.info("getMetaTableRows: row -> " + + HBaseCommonTestingUtility.LOG.info("getMetaTableRows: row -> " + Bytes.toStringBinary(result.getRow()) + info); rows.add(result.getRow()); } @@ -2678,7 +2675,7 @@ public HRegionServer getRSForFirstRegionInTable(TableName tableName) if (regions == null || regions.isEmpty()) { return null; } - LOG.debug("Found " + regions.size() + " regions for table " + + HBaseCommonTestingUtility.LOG.debug("Found " + regions.size() + " regions for table " + tableName); byte[] firstRegionName = regions.stream() @@ -2687,7 +2684,7 @@ public HRegionServer getRSForFirstRegionInTable(TableName tableName) .findFirst() .orElseThrow(() -> new IOException("online regions not found in table " + tableName)); - LOG.debug("firstRegionName=" + Bytes.toString(firstRegionName)); + HBaseCommonTestingUtility.LOG.debug("firstRegionName=" + Bytes.toString(firstRegionName)); long pause = getConfiguration().getLong(HConstants.HBASE_CLIENT_PAUSE, HConstants.DEFAULT_HBASE_CLIENT_PAUSE); int numRetries = getConfiguration().getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, @@ -2756,7 +2753,7 @@ private void startMiniMapReduceCluster(final int servers) throws IOException { if (mrCluster != null) { throw new IllegalStateException("MiniMRCluster is already running"); } - LOG.info("Starting mini mapreduce cluster..."); + HBaseCommonTestingUtility.LOG.info("Starting mini mapreduce cluster..."); setupClusterTestDir(); createDirsAndSetProperties(); @@ -2784,7 +2781,7 @@ private void startMiniMapReduceCluster(final int servers) throws IOException { jobConf.set("mapreduce.cluster.local.dir", conf.get("mapreduce.cluster.local.dir")); //Hadoop MiniMR overwrites this while it should not - LOG.info("Mini mapreduce cluster started"); + HBaseCommonTestingUtility.LOG.info("Mini mapreduce cluster started"); // In hadoop2, YARN/MR2 starts a mini cluster with its own conf instance and updates settings. // Our HBase MR jobs need several of these settings in order to properly run. So we copy the @@ -2823,10 +2820,10 @@ private void startMiniMapReduceCluster(final int servers) throws IOException { */ public void shutdownMiniMapReduceCluster() { if (mrCluster != null) { - LOG.info("Stopping mini mapreduce cluster..."); + HBaseCommonTestingUtility.LOG.info("Stopping mini mapreduce cluster..."); mrCluster.shutdown(); mrCluster = null; - LOG.info("Mini mapreduce cluster stopped"); + HBaseCommonTestingUtility.LOG.info("Mini mapreduce cluster stopped"); } // Restore configuration to point to local jobtracker conf.set("mapreduce.jobtracker.address", "local"); @@ -2950,7 +2947,7 @@ public void expireSession(ZKWatcher nodeZK, boolean checkStatus) 1000, new org.apache.zookeeper.Watcher(){ @Override public void process(WatchedEvent watchedEvent) { - LOG.info("Monitor ZKW received event="+watchedEvent); + HBaseCommonTestingUtility.LOG.info("Monitor ZKW received event="+watchedEvent); } } , sessionID, password); @@ -2966,7 +2963,7 @@ public void process(WatchedEvent watchedEvent) { Thread.sleep(1); } newZK.close(); - LOG.info("ZK Closed Session 0x" + Long.toHexString(sessionID)); + HBaseCommonTestingUtility.LOG.info("ZK Closed Session 0x" + Long.toHexString(sessionID)); // Now closing & waiting to be sure that the clients get it. monitor.close(); @@ -3019,7 +3016,7 @@ public void invalidateConnection() throws IOException { // Update the master addresses if they changed. final String masterConfigBefore = conf.get(HConstants.MASTER_ADDRS_KEY); final String masterConfAfter = getMiniHBaseCluster().conf.get(HConstants.MASTER_ADDRS_KEY); - LOG.info("Invalidated connection. Updating master addresses before: {} after: {}", + HBaseCommonTestingUtility.LOG.info("Invalidated connection. Updating master addresses before: {} after: {}", masterConfigBefore, masterConfAfter); conf.set(HConstants.MASTER_ADDRS_KEY, getMiniHBaseCluster().conf.get(HConstants.MASTER_ADDRS_KEY)); @@ -3142,7 +3139,7 @@ public HRegion getSplittableRegion(TableName tableName, int maxAttempts) { regions.get(idx).checkSplit(); return regions.get(idx); } catch (Exception ex) { - LOG.warn("Caught exception", ex); + HBaseCommonTestingUtility.LOG.warn("Caught exception", ex); attempted.add(idx); } } @@ -3338,7 +3335,7 @@ public boolean ensureSomeRegionServersAvailable(final int num) boolean startedServer = false; MiniHBaseCluster hbaseCluster = getMiniHBaseCluster(); for (int i=hbaseCluster.getLiveRegionServerThreads().size(); i() { @Override @@ -3582,14 +3579,14 @@ public boolean evaluate() throws IOException { } } if (!tableFound) { - LOG.warn("Didn't find the entries for table " + tableName + " in meta, already deleted?"); + HBaseCommonTestingUtility.LOG.warn("Didn't find the entries for table " + tableName + " in meta, already deleted?"); } return tableFound; } }); } } - LOG.info("All regions for table " + tableName + " assigned to meta. Checking AM states."); + HBaseCommonTestingUtility.LOG.info("All regions for table " + tableName + " assigned to meta. Checking AM states."); // check from the master state if we are using a mini cluster if (!getHBaseClusterInterface().isDistributedCluster()) { // So, all regions are in the meta table but make sure master knows of the assignments before @@ -3609,7 +3606,7 @@ public boolean evaluate() throws IOException { } }); } - LOG.info("All regions for table " + tableName + " assigned."); + HBaseCommonTestingUtility.LOG.info("All regions for table " + tableName + " assigned."); } /** @@ -3647,7 +3644,7 @@ public static List getFromStoreFile(HStore store, * @return resulting split keys */ public byte[][] getRegionSplitStartKeys(byte[] startKey, byte[] endKey, int numRegions){ - assertTrue(numRegions>3); + Assert.assertTrue(numRegions>3); byte [][] tmpSplitKeys = Bytes.split(startKey, endKey, numRegions - 3); byte [][] result = new byte[tmpSplitKeys.length+1][]; System.arraycopy(tmpSplitKeys, 0, result, 1, tmpSplitKeys.length); @@ -3722,7 +3719,7 @@ public Table createRandomTable(TableName tableName, final int numRowsPerFlush) throws IOException, InterruptedException { - LOG.info("\n\nCreating random table " + tableName + " with " + numRegions + + HBaseCommonTestingUtility.LOG.info("\n\nCreating random table " + tableName + " with " + numRegions + " regions, " + numFlushes + " storefiles per region, " + numRowsPerFlush + " rows per flush, maxVersions=" + maxVersions + "\n"); @@ -3786,7 +3783,7 @@ public Table createRandomTable(TableName tableName, mutator.mutate(del); } } - LOG.info("Initiating flush #" + iFlush + " for table " + tableName); + HBaseCommonTestingUtility.LOG.info("Initiating flush #" + iFlush + " for table " + tableName); mutator.flush(); if (hbaseCluster != null) { getMiniHBaseCluster().flushcache(table.getName()); @@ -3880,13 +3877,13 @@ public static void waitForHostPort(String host, int port) final int maxTimeMs = 10000; final int maxNumAttempts = maxTimeMs / HConstants.SOCKET_RETRY_WAIT_MS; IOException savedException = null; - LOG.info("Waiting for server at " + host + ":" + port); + HBaseCommonTestingUtility.LOG.info("Waiting for server at " + host + ":" + port); for (int attempt = 0; attempt < maxNumAttempts; ++attempt) { try { Socket sock = new Socket(InetAddress.getByName(host), port); sock.close(); savedException = null; - LOG.info("Server at " + host + ":" + port + " is available"); + HBaseCommonTestingUtility.LOG.info("Server at " + host + ":" + port + " is available"); break; } catch (UnknownHostException e) { throw new IOException("Failed to look up " + host, e); @@ -4022,7 +4019,7 @@ public static int createPreSplitLoadTestTable(Configuration conf, } totalNumberOfRegions = numberOfServers * numRegionsPerServer; - LOG.info("Number of live regionservers: " + numberOfServers + ", " + + HBaseCommonTestingUtility.LOG.info("Number of live regionservers: " + numberOfServers + ", " + "pre-splitting table into " + totalNumberOfRegions + " regions " + "(regions per server: " + numRegionsPerServer + ")"); @@ -4031,10 +4028,10 @@ public static int createPreSplitLoadTestTable(Configuration conf, admin.createTable(td, splits); } catch (MasterNotRunningException e) { - LOG.error("Master not running", e); + HBaseCommonTestingUtility.LOG.error("Master not running", e); throw new IOException(e); } catch (TableExistsException e) { - LOG.warn("Table " + td.getTableName() + + HBaseCommonTestingUtility.LOG.warn("Table " + td.getTableName() + " already exists, continuing"); } finally { admin.close(); @@ -4066,7 +4063,7 @@ public void assertRegionOnServer( if (now > timeoutTime) break; Thread.sleep(10); } - fail("Could not find region " + hri.getRegionNameAsString() + Assert.fail("Could not find region " + hri.getRegionNameAsString() + " on server " + server); } @@ -4090,7 +4087,7 @@ public void assertRegionOnlyOnServer( } Collection hrs = rs.getOnlineRegionsLocalContext(); for (HRegion r: hrs) { - assertTrue("Region should not be double assigned", + Assert.assertTrue("Region should not be double assigned", r.getRegionInfo().getRegionId() != hri.getRegionId()); } } @@ -4100,7 +4097,7 @@ public void assertRegionOnlyOnServer( if (now > timeoutTime) break; Thread.sleep(10); } - fail("Could not find region " + hri.getRegionNameAsString() + Assert.fail("Could not find region " + hri.getRegionNameAsString() + " on server " + server); } @@ -4365,10 +4362,10 @@ public MiniKdc setupMiniKdc(File keytabFile) throws Exception { FileUtils.deleteDirectory(dir); // clean directory numTries++; if (numTries == 3) { - LOG.error("Failed setting up MiniKDC. Tried " + numTries + " times."); + HBaseCommonTestingUtility.LOG.error("Failed setting up MiniKDC. Tried " + numTries + " times."); throw e; } - LOG.error("BindException encountered when setting up MiniKdc. Trying again."); + HBaseCommonTestingUtility.LOG.error("BindException encountered when setting up MiniKdc. Trying again."); bindException = true; } } while (bindException); @@ -4395,13 +4392,13 @@ public int getNumHFilesForRS(final HRegionServer rs, final TableName tableName, } public void verifyTableDescriptorIgnoreTableName(TableDescriptor ltd, TableDescriptor rtd) { - assertEquals(ltd.getValues().hashCode(), rtd.getValues().hashCode()); + Assert.assertEquals(ltd.getValues().hashCode(), rtd.getValues().hashCode()); Collection ltdFamilies = Arrays.asList(ltd.getColumnFamilies()); Collection rtdFamilies = Arrays.asList(rtd.getColumnFamilies()); - assertEquals(ltdFamilies.size(), rtdFamilies.size()); + Assert.assertEquals(ltdFamilies.size(), rtdFamilies.size()); for (Iterator it = ltdFamilies.iterator(), it2 = rtdFamilies.iterator(); it.hasNext();) { - assertEquals(0, + Assert.assertEquals(0, ColumnFamilyDescriptor.COMPARATOR.compare(it.next(), it2.next())); } } diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/MiniHBaseCluster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/MiniHBaseCluster.java similarity index 99% rename from hbase-server/src/test/java/org/apache/hadoop/hbase/MiniHBaseCluster.java rename to hbase-server/src/main/java/org/apache/hadoop/hbase/MiniHBaseCluster.java index 53c590b74e1d..103960340b57 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/MiniHBaseCluster.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/MiniHBaseCluster.java @@ -32,7 +32,6 @@ import org.apache.hadoop.hbase.regionserver.HRegionServer; import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.security.User; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.util.JVMClusterUtil; import org.apache.hadoop.hbase.util.JVMClusterUtil.MasterThread; import org.apache.hadoop.hbase.util.JVMClusterUtil.RegionServerThread; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/MockRegionServerServices.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/MockRegionServerServices.java similarity index 99% rename from hbase-server/src/test/java/org/apache/hadoop/hbase/MockRegionServerServices.java rename to hbase-server/src/main/java/org/apache/hadoop/hbase/MockRegionServerServices.java index f0a85b580d4b..e8111026e5b8 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/MockRegionServerServices.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/MockRegionServerServices.java @@ -60,6 +60,7 @@ import org.apache.hadoop.hbase.util.Bytes; import org.apache.hadoop.hbase.wal.WAL; import org.apache.hadoop.hbase.zookeeper.ZKWatcher; +import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -70,6 +71,7 @@ /** * Basic mock region server services. Should only be instantiated by HBaseTestingUtility.b */ +@InterfaceAudience.Private public class MockRegionServerServices implements RegionServerServices { protected static final Logger LOG = LoggerFactory.getLogger(MockRegionServerServices.class); private final Map regions = new HashMap<>(); diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/StartMiniClusterOption.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/StartMiniClusterOption.java similarity index 100% rename from hbase-server/src/test/java/org/apache/hadoop/hbase/StartMiniClusterOption.java rename to hbase-server/src/main/java/org/apache/hadoop/hbase/StartMiniClusterOption.java diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/MapreduceTestingShim.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MapreduceTestingShim.java similarity index 98% rename from hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/MapreduceTestingShim.java rename to hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MapreduceTestingShim.java index 69f738741eb1..db53890871d3 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/mapreduce/MapreduceTestingShim.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/mapreduce/MapreduceTestingShim.java @@ -28,6 +28,7 @@ import org.apache.hadoop.mapreduce.Job; import org.apache.hadoop.mapreduce.JobContext; import org.apache.hadoop.mapreduce.JobID; +import org.apache.yetus.audience.InterfaceAudience; /** * This class provides shims for HBase to interact with the Hadoop 1.0.x and the @@ -35,6 +36,7 @@ * * NOTE: No testing done against 0.22.x, or 0.21.x. */ +@InterfaceAudience.Private abstract public class MapreduceTestingShim { private static MapreduceTestingShim instance; private static Class[] emptyParam = new Class[] {}; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/assignment/AssignmentTestingUtil.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentTestingUtil.java similarity index 100% rename from hbase-server/src/test/java/org/apache/hadoop/hbase/master/assignment/AssignmentTestingUtil.java rename to hbase-server/src/main/java/org/apache/hadoop/hbase/master/assignment/AssignmentTestingUtil.java diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/HBaseKerberosUtils.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/HBaseKerberosUtils.java similarity index 100% rename from hbase-server/src/test/java/org/apache/hadoop/hbase/security/HBaseKerberosUtils.java rename to hbase-server/src/main/java/org/apache/hadoop/hbase/security/HBaseKerberosUtils.java index f80e5f48919b..d73411e01ea6 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/security/HBaseKerberosUtils.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/security/HBaseKerberosUtils.java @@ -29,12 +29,12 @@ import org.apache.hadoop.hbase.http.ssl.KeyStoreTestUtil; import org.apache.hadoop.hdfs.DFSConfigKeys; import org.apache.hadoop.http.HttpConfig; +import org.apache.hadoop.security.UserGroupInformation; import org.apache.hadoop.yarn.conf.YarnConfiguration; import org.apache.yetus.audience.InterfaceAudience; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.apache.hbase.thirdparty.com.google.common.base.Strings; -import org.apache.hadoop.security.UserGroupInformation; @InterfaceAudience.Private public class HBaseKerberosUtils { diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/util/StoppableImplementation.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/util/StoppableImplementation.java similarity index 100% rename from hbase-server/src/test/java/org/apache/hadoop/hbase/util/StoppableImplementation.java rename to hbase-server/src/main/java/org/apache/hadoop/hbase/util/StoppableImplementation.java diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiRespectsLimits.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiRespectsLimits.java index dec9f65716f7..0b5db840e66f 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiRespectsLimits.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiRespectsLimits.java @@ -35,7 +35,7 @@ import org.apache.hadoop.hbase.ipc.RpcServerInterface; import org.apache.hadoop.hbase.metrics.BaseSource; import org.apache.hadoop.hbase.regionserver.HRegionServer; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.ClientTests; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/TestMetricsIO.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/TestMetricsIO.java index c6062044a928..18ce091a6c14 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/TestMetricsIO.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/TestMetricsIO.java @@ -19,7 +19,7 @@ import org.apache.hadoop.hbase.CompatibilityFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.ClassRule; import org.junit.Test; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestRpcMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestRpcMetrics.java index 2f99d2b65929..7a10317ce4a4 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestRpcMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestRpcMetrics.java @@ -26,7 +26,7 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException; import org.apache.hadoop.hbase.exceptions.RegionMovedException; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.RPCTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.ClassRule; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerMetrics.java index 6103de075af2..62f5ab6d5557 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerMetrics.java @@ -35,7 +35,7 @@ import org.apache.hadoop.hbase.client.TableDescriptorBuilder; import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; import org.apache.hadoop.hbase.master.assignment.AssignmentManager; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.MasterTests; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterMetrics.java index 75d9ee199b96..5331976417a3 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterMetrics.java @@ -25,7 +25,7 @@ import org.apache.hadoop.hbase.MiniHBaseCluster; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.StartMiniClusterOption; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.MasterTests; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.zookeeper.KeeperException; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java index 423208e7c8f0..e038f9e54e59 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java @@ -143,7 +143,7 @@ import org.apache.hadoop.hbase.regionserver.wal.WALUtil; import org.apache.hadoop.hbase.replication.regionserver.ReplicationObserver; import org.apache.hadoop.hbase.security.User; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.VerySlowRegionServerTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsHeapMemoryManager.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsHeapMemoryManager.java index 9958ca664bd4..21645d278fa6 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsHeapMemoryManager.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsHeapMemoryManager.java @@ -21,7 +21,7 @@ import org.apache.hadoop.hbase.CompatibilitySingletonFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.Before; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java index 00031130604c..30464f2b1e44 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java @@ -20,7 +20,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.CompatibilityFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.ClassRule; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegionServer.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegionServer.java index 8215ea13ee77..b4ad2ba7d79d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegionServer.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegionServer.java @@ -22,7 +22,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.CompatibilityFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.apache.hadoop.hbase.util.JvmPauseMonitor; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableAggregate.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableAggregate.java index 98a412cf59bf..a02868534a12 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableAggregate.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableAggregate.java @@ -27,7 +27,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.CompatibilityFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.junit.Before; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableLatencies.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableLatencies.java index f16086e49118..6405f7ced35d 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableLatencies.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableLatencies.java @@ -25,7 +25,7 @@ import org.apache.hadoop.hbase.CompatibilitySingletonFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.ClassRule; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsUserAggregate.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsUserAggregate.java index 645c7f2e66ba..8b7432d3ee65 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsUserAggregate.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsUserAggregate.java @@ -27,7 +27,7 @@ import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.security.User; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.junit.Before; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerMetrics.java index 833e78ba3b0e..7a4d141c32c1 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerMetrics.java @@ -59,7 +59,7 @@ import org.apache.hadoop.hbase.regionserver.compactions.CompactionLifeCycleTracker; import org.apache.hadoop.hbase.regionserver.throttle.NoLimitThroughputController; import org.apache.hadoop.hbase.security.User; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRemoveRegionMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRemoveRegionMetrics.java index acc3f9dd1eee..54c6df860950 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRemoveRegionMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRemoveRegionMetrics.java @@ -31,7 +31,7 @@ import org.apache.hadoop.hbase.client.RegionInfo; import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.hadoop.hbase.client.Table; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestCallQueue.java b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestCallQueue.java index 955b4ca9c99e..8d4cf76d45c5 100644 --- a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestCallQueue.java +++ b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestCallQueue.java @@ -24,7 +24,7 @@ import org.apache.hadoop.hbase.CompatibilitySingletonFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; import org.apache.hadoop.hbase.HBaseTestingUtility; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.ClientTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.ClassRule; diff --git a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestThriftServer.java b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestThriftServer.java index 5cda76af1684..7a7de2abc937 100644 --- a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestThriftServer.java +++ b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestThriftServer.java @@ -47,7 +47,7 @@ import org.apache.hadoop.hbase.client.TableDescriptorBuilder; import org.apache.hadoop.hbase.filter.ParseFilter; import org.apache.hadoop.hbase.security.UserProvider; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.ClientTests; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType; diff --git a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift2/TestThriftHBaseServiceHandler.java b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift2/TestThriftHBaseServiceHandler.java index 0a6b42b69c34..273d7e0a1f8d 100644 --- a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift2/TestThriftHBaseServiceHandler.java +++ b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift2/TestThriftHBaseServiceHandler.java @@ -74,7 +74,7 @@ import org.apache.hadoop.hbase.coprocessor.RegionObserver; import org.apache.hadoop.hbase.filter.ParseFilter; import org.apache.hadoop.hbase.security.UserProvider; -import org.apache.hadoop.hbase.test.MetricsAssertHelper; +import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.ClientTests; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.thrift.ErrorThrowingGetObserver; diff --git a/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/HBaseZKTestingUtility.java b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/HBaseZKTestingUtility.java similarity index 96% rename from hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/HBaseZKTestingUtility.java rename to hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/HBaseZKTestingUtility.java index dd8b3909d6d2..42ce1f9a29c7 100644 --- a/hbase-zookeeper/src/test/java/org/apache/hadoop/hbase/HBaseZKTestingUtility.java +++ b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/HBaseZKTestingUtility.java @@ -73,14 +73,14 @@ protected void setupClusterTestDir() { // Using randomUUID ensures that multiple clusters can be launched by // a same test, if it stops & starts them - Path testDir = getDataTestDir("cluster_" + getRandomUUID().toString()); + Path testDir = getDataTestDir("cluster_" + HBaseCommonTestingUtility.getRandomUUID().toString()); clusterTestDir = new File(testDir.toString()).getAbsoluteFile(); // Have it cleaned up on exit boolean b = deleteOnExit(); if (b) { clusterTestDir.deleteOnExit(); } - LOG.info("Created new mini-cluster data directory: " + clusterTestDir + ", deleteOnExit=" + b); + HBaseCommonTestingUtility.LOG.info("Created new mini-cluster data directory: " + clusterTestDir + ", deleteOnExit=" + b); } /** From 31c281e6f46c1f0c0958fad0e1c93d9b80d227e1 Mon Sep 17 00:00:00 2001 From: Lokesh Khurana Date: Wed, 22 Apr 2020 17:58:30 +0530 Subject: [PATCH 5/7] Removing LICENSE.vm changes --- hbase-common/pom.xml | 4 +++ .../java/org/apache/hadoop/hbase/Waiter.java | 4 +-- .../hbase/{ => test}/MetricsAssertHelper.java | 2 +- .../hbase/test/MetricsAssertHelperImpl.java | 1 - ...che.hadoop.hbase.test.MetricsAssertHelper} | 0 hbase-http/pom.xml | 11 +++---- .../src/main/resources/META-INF/LICENSE.vm | 2 +- .../hbase/rest/TestGetAndPutResource.java | 2 +- hbase-server/pom.xml | 12 +++---- .../hadoop/hbase/HBaseTestingUtility.java | 33 ++++++++++--------- .../apache/hadoop/hbase/MiniHBaseCluster.java | 1 + .../hbase/client/TestMultiRespectsLimits.java | 2 +- .../apache/hadoop/hbase/io/TestMetricsIO.java | 2 +- .../hadoop/hbase/ipc/TestRpcMetrics.java | 2 +- .../master/TestAssignmentManagerMetrics.java | 2 +- .../hbase/master/TestMasterMetrics.java | 2 +- .../hbase/regionserver/TestHRegion.java | 2 +- .../TestMetricsHeapMemoryManager.java | 2 +- .../hbase/regionserver/TestMetricsRegion.java | 2 +- .../regionserver/TestMetricsRegionServer.java | 2 +- .../TestMetricsTableAggregate.java | 2 +- .../TestMetricsTableLatencies.java | 2 +- .../TestMetricsUserAggregate.java | 2 +- .../regionserver/TestRegionServerMetrics.java | 2 +- .../regionserver/TestRemoveRegionMetrics.java | 2 +- .../hadoop/hbase/thrift/TestCallQueue.java | 2 +- .../hadoop/hbase/thrift/TestThriftServer.java | 2 +- .../TestThriftHBaseServiceHandler.java | 2 +- 28 files changed, 56 insertions(+), 50 deletions(-) rename hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/{ => test}/MetricsAssertHelper.java (99%) rename hbase-hadoop2-compat/src/test/resources/META-INF/services/{org.apache.hadoop.hbase.MetricsAssertHelper => org.apache.hadoop.hbase.test.MetricsAssertHelper} (100%) diff --git a/hbase-common/pom.xml b/hbase-common/pom.xml index d11bf7905048..22c04e3e8587 100644 --- a/hbase-common/pom.xml +++ b/hbase-common/pom.xml @@ -220,6 +220,10 @@ junit junit + + junit + junit + diff --git a/hbase-common/src/main/java/org/apache/hadoop/hbase/Waiter.java b/hbase-common/src/main/java/org/apache/hadoop/hbase/Waiter.java index de6c0bfe95d9..5302093b66a0 100644 --- a/hbase-common/src/main/java/org/apache/hadoop/hbase/Waiter.java +++ b/hbase-common/src/main/java/org/apache/hadoop/hbase/Waiter.java @@ -18,12 +18,12 @@ */ package org.apache.hadoop.hbase; +import static org.junit.Assert.fail; import java.text.MessageFormat; import org.apache.hadoop.conf.Configuration; import org.apache.yetus.audience.InterfaceAudience; -import org.junit.Assert; import org.slf4j.Logger; import org.slf4j.LoggerFactory; @@ -200,7 +200,7 @@ public static long waitFor(Configuration conf, long timeou System.currentTimeMillis() - started)); } else if (failIfTimeout) { String msg = getExplanation(predicate); - Assert.fail(MessageFormat + fail(MessageFormat .format("Waiting timed out after [{0}] msec", adjustedTimeout) + msg); } else { String msg = getExplanation(predicate); diff --git a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/MetricsAssertHelper.java b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/test/MetricsAssertHelper.java similarity index 99% rename from hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/MetricsAssertHelper.java rename to hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/test/MetricsAssertHelper.java index e4d5f20f745d..967dd95a32de 100644 --- a/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/MetricsAssertHelper.java +++ b/hbase-hadoop-compat/src/main/java/org/apache/hadoop/hbase/test/MetricsAssertHelper.java @@ -16,7 +16,7 @@ * limitations under the License. */ -package org.apache.hadoop.hbase; +package org.apache.hadoop.hbase.test; import org.apache.hadoop.hbase.metrics.BaseSource; import org.apache.yetus.audience.InterfaceAudience; diff --git a/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java b/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java index 523eae0d9cae..afe31f166671 100644 --- a/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java +++ b/hbase-hadoop2-compat/src/test/java/org/apache/hadoop/hbase/test/MetricsAssertHelperImpl.java @@ -26,7 +26,6 @@ import java.util.Locale; import java.util.Map; -import org.apache.hadoop.hbase.MetricsAssertHelper; import org.apache.hadoop.hbase.metrics.BaseSource; import org.apache.hadoop.metrics2.AbstractMetric; import org.apache.hadoop.metrics2.MetricsCollector; diff --git a/hbase-hadoop2-compat/src/test/resources/META-INF/services/org.apache.hadoop.hbase.MetricsAssertHelper b/hbase-hadoop2-compat/src/test/resources/META-INF/services/org.apache.hadoop.hbase.test.MetricsAssertHelper similarity index 100% rename from hbase-hadoop2-compat/src/test/resources/META-INF/services/org.apache.hadoop.hbase.MetricsAssertHelper rename to hbase-hadoop2-compat/src/test/resources/META-INF/services/org.apache.hadoop.hbase.test.MetricsAssertHelper diff --git a/hbase-http/pom.xml b/hbase-http/pom.xml index e375929eeb96..0d059e5d3b3b 100644 --- a/hbase-http/pom.xml +++ b/hbase-http/pom.xml @@ -226,6 +226,11 @@ bcprov-jdk15on test + + org.bouncycastle + bcprov-jdk15on + compile + org.apache.kerby kerb-core @@ -256,12 +261,6 @@ hadoop-minikdc test - - bcprov-jdk15 - bouncycastle - 140 - compile - diff --git a/hbase-resource-bundle/src/main/resources/META-INF/LICENSE.vm b/hbase-resource-bundle/src/main/resources/META-INF/LICENSE.vm index a0cfeb6484b0..760208f4b863 100644 --- a/hbase-resource-bundle/src/main/resources/META-INF/LICENSE.vm +++ b/hbase-resource-bundle/src/main/resources/META-INF/LICENSE.vm @@ -1343,7 +1343,7 @@ You can redistribute it and/or modify it under either the terms of the ## See this FAQ link for justifications: https://www.apache.org/legal/resolved.html ## ## NB: This list is later compared as lower-case. New entries must also be all lower-case -#set($non_aggregate_fine = [ 'public domain', 'new bsd license', 'bsd license', 'bsd', 'bsd 2-clause license', 'mozilla public license version 1.1', 'mozilla public license version 2.0', 'creative commons attribution license, version 2.5', 'bouncy castle license' ]) +#set($non_aggregate_fine = [ 'public domain', 'new bsd license', 'bsd license', 'bsd', 'bsd 2-clause license', 'mozilla public license version 1.1', 'mozilla public license version 2.0', 'creative commons attribution license, version 2.5' ]) ## include LICENSE sections for anything not under ASL2.0 #foreach( ${dep} in ${projects} ) ## if there are no licenses we'll fail the build later, so diff --git a/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestGetAndPutResource.java b/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestGetAndPutResource.java index de1da94b46e7..e1dec900d491 100644 --- a/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestGetAndPutResource.java +++ b/hbase-rest/src/test/java/org/apache/hadoop/hbase/rest/TestGetAndPutResource.java @@ -35,7 +35,7 @@ import org.apache.hadoop.hbase.rest.model.CellSetModel; import org.apache.hadoop.hbase.rest.model.RowModel; import org.apache.hadoop.hbase.security.UserProvider; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.RestTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/pom.xml b/hbase-server/pom.xml index 1c6159901c0c..638ad70627ca 100644 --- a/hbase-server/pom.xml +++ b/hbase-server/pom.xml @@ -536,6 +536,12 @@ hadoop-minikdc compile + + org.apache.hadoop + hadoop-hdfs + ${hadoop.version} + tests + @@ -681,12 +687,6 @@ test-jar test - - org.apache.hadoop - hadoop-hdfs - test-jar - compile - org.apache.hadoop hadoop-minicluster diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseTestingUtility.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseTestingUtility.java index e8c476e1acae..e9c1cccfc85b 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseTestingUtility.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseTestingUtility.java @@ -17,6 +17,10 @@ */ package org.apache.hadoop.hbase; +import static org.junit.Assert.assertEquals; +import static org.junit.Assert.assertTrue; +import static org.junit.Assert.fail; + import edu.umd.cs.findbugs.annotations.Nullable; import java.io.File; @@ -145,7 +149,6 @@ import org.apache.zookeeper.WatchedEvent; import org.apache.zookeeper.ZooKeeper; import org.apache.zookeeper.ZooKeeper.States; -import org.junit.Assert; import org.slf4j.Logger; import org.slf4j.LoggerFactory; import org.slf4j.impl.Log4jLoggerAdapter; @@ -2319,10 +2322,10 @@ public void verifyNumericRows(Table table, final byte[] f, int startRow, int end get.setReplicaId(replicaId); get.setConsistency(Consistency.TIMELINE); Result result = table.get(get); - Assert.assertTrue(failMsg, result.containsColumn(f, null)); - Assert.assertEquals(failMsg, 1, result.getColumnCells(f, null).size()); + assertTrue(failMsg, result.containsColumn(f, null)); + assertEquals(failMsg, 1, result.getColumnCells(f, null).size()); Cell cell = result.getColumnLatestCell(f, null); - Assert.assertTrue(failMsg, + assertTrue(failMsg, Bytes.equals(data, 0, data.length, cell.getValueArray(), cell.getValueOffset(), cell.getValueLength())); } @@ -2351,13 +2354,13 @@ public void verifyNumericRows(HRegion region, final byte[] f, int startRow, int Result result = region.get(new Get(data)); boolean hasResult = result != null && !result.isEmpty(); - Assert.assertEquals(failMsg + result, present, hasResult); + assertEquals(failMsg + result, present, hasResult); if (!present) continue; - Assert.assertTrue(failMsg, result.containsColumn(f, null)); - Assert.assertEquals(failMsg, 1, result.getColumnCells(f, null).size()); + assertTrue(failMsg, result.containsColumn(f, null)); + assertEquals(failMsg, 1, result.getColumnCells(f, null).size()); Cell cell = result.getColumnLatestCell(f, null); - Assert.assertTrue(failMsg, + assertTrue(failMsg, Bytes.equals(data, 0, data.length, cell.getValueArray(), cell.getValueOffset(), cell.getValueLength())); } @@ -3644,7 +3647,7 @@ public static List getFromStoreFile(HStore store, * @return resulting split keys */ public byte[][] getRegionSplitStartKeys(byte[] startKey, byte[] endKey, int numRegions){ - Assert.assertTrue(numRegions>3); + assertTrue(numRegions>3); byte [][] tmpSplitKeys = Bytes.split(startKey, endKey, numRegions - 3); byte [][] result = new byte[tmpSplitKeys.length+1][]; System.arraycopy(tmpSplitKeys, 0, result, 1, tmpSplitKeys.length); @@ -4063,7 +4066,7 @@ public void assertRegionOnServer( if (now > timeoutTime) break; Thread.sleep(10); } - Assert.fail("Could not find region " + hri.getRegionNameAsString() + fail("Could not find region " + hri.getRegionNameAsString() + " on server " + server); } @@ -4087,7 +4090,7 @@ public void assertRegionOnlyOnServer( } Collection hrs = rs.getOnlineRegionsLocalContext(); for (HRegion r: hrs) { - Assert.assertTrue("Region should not be double assigned", + assertTrue("Region should not be double assigned", r.getRegionInfo().getRegionId() != hri.getRegionId()); } } @@ -4097,7 +4100,7 @@ public void assertRegionOnlyOnServer( if (now > timeoutTime) break; Thread.sleep(10); } - Assert.fail("Could not find region " + hri.getRegionNameAsString() + fail("Could not find region " + hri.getRegionNameAsString() + " on server " + server); } @@ -4392,13 +4395,13 @@ public int getNumHFilesForRS(final HRegionServer rs, final TableName tableName, } public void verifyTableDescriptorIgnoreTableName(TableDescriptor ltd, TableDescriptor rtd) { - Assert.assertEquals(ltd.getValues().hashCode(), rtd.getValues().hashCode()); + assertEquals(ltd.getValues().hashCode(), rtd.getValues().hashCode()); Collection ltdFamilies = Arrays.asList(ltd.getColumnFamilies()); Collection rtdFamilies = Arrays.asList(rtd.getColumnFamilies()); - Assert.assertEquals(ltdFamilies.size(), rtdFamilies.size()); + assertEquals(ltdFamilies.size(), rtdFamilies.size()); for (Iterator it = ltdFamilies.iterator(), it2 = rtdFamilies.iterator(); it.hasNext();) { - Assert.assertEquals(0, + assertEquals(0, ColumnFamilyDescriptor.COMPARATOR.compare(it.next(), it2.next())); } } diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/MiniHBaseCluster.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/MiniHBaseCluster.java index 103960340b57..53c590b74e1d 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/MiniHBaseCluster.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/MiniHBaseCluster.java @@ -32,6 +32,7 @@ import org.apache.hadoop.hbase.regionserver.HRegionServer; import org.apache.hadoop.hbase.regionserver.Region; import org.apache.hadoop.hbase.security.User; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.util.JVMClusterUtil; import org.apache.hadoop.hbase.util.JVMClusterUtil.MasterThread; import org.apache.hadoop.hbase.util.JVMClusterUtil.RegionServerThread; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiRespectsLimits.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiRespectsLimits.java index 0b5db840e66f..dec9f65716f7 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiRespectsLimits.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/client/TestMultiRespectsLimits.java @@ -35,7 +35,7 @@ import org.apache.hadoop.hbase.ipc.RpcServerInterface; import org.apache.hadoop.hbase.metrics.BaseSource; import org.apache.hadoop.hbase.regionserver.HRegionServer; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.ClientTests; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/TestMetricsIO.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/TestMetricsIO.java index 18ce091a6c14..c6062044a928 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/io/TestMetricsIO.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/io/TestMetricsIO.java @@ -19,7 +19,7 @@ import org.apache.hadoop.hbase.CompatibilityFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.ClassRule; import org.junit.Test; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestRpcMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestRpcMetrics.java index 7a10317ce4a4..2f99d2b65929 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestRpcMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/ipc/TestRpcMetrics.java @@ -26,7 +26,7 @@ import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.exceptions.OutOfOrderScannerNextException; import org.apache.hadoop.hbase.exceptions.RegionMovedException; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.RPCTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.ClassRule; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerMetrics.java index 62f5ab6d5557..6103de075af2 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestAssignmentManagerMetrics.java @@ -35,7 +35,7 @@ import org.apache.hadoop.hbase.client.TableDescriptorBuilder; import org.apache.hadoop.hbase.coprocessor.CoprocessorHost; import org.apache.hadoop.hbase.master.assignment.AssignmentManager; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.MasterTests; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterMetrics.java index 5331976417a3..75d9ee199b96 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/master/TestMasterMetrics.java @@ -25,7 +25,7 @@ import org.apache.hadoop.hbase.MiniHBaseCluster; import org.apache.hadoop.hbase.ServerName; import org.apache.hadoop.hbase.StartMiniClusterOption; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.MasterTests; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.zookeeper.KeeperException; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java index e038f9e54e59..423208e7c8f0 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestHRegion.java @@ -143,7 +143,7 @@ import org.apache.hadoop.hbase.regionserver.wal.WALUtil; import org.apache.hadoop.hbase.replication.regionserver.ReplicationObserver; import org.apache.hadoop.hbase.security.User; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.VerySlowRegionServerTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsHeapMemoryManager.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsHeapMemoryManager.java index 21645d278fa6..9958ca664bd4 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsHeapMemoryManager.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsHeapMemoryManager.java @@ -21,7 +21,7 @@ import org.apache.hadoop.hbase.CompatibilitySingletonFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.Before; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java index 30464f2b1e44..00031130604c 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegion.java @@ -20,7 +20,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.CompatibilityFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.ClassRule; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegionServer.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegionServer.java index b4ad2ba7d79d..8215ea13ee77 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegionServer.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsRegionServer.java @@ -22,7 +22,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.CompatibilityFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.apache.hadoop.hbase.util.JvmPauseMonitor; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableAggregate.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableAggregate.java index a02868534a12..98a412cf59bf 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableAggregate.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableAggregate.java @@ -27,7 +27,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.CompatibilityFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.junit.Before; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableLatencies.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableLatencies.java index 6405f7ced35d..f16086e49118 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableLatencies.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsTableLatencies.java @@ -25,7 +25,7 @@ import org.apache.hadoop.hbase.CompatibilitySingletonFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; import org.apache.hadoop.hbase.TableName; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.ClassRule; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsUserAggregate.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsUserAggregate.java index 8b7432d3ee65..645c7f2e66ba 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsUserAggregate.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestMetricsUserAggregate.java @@ -27,7 +27,7 @@ import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.security.User; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.junit.Before; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerMetrics.java index 7a4d141c32c1..833e78ba3b0e 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRegionServerMetrics.java @@ -59,7 +59,7 @@ import org.apache.hadoop.hbase.regionserver.compactions.CompactionLifeCycleTracker; import org.apache.hadoop.hbase.regionserver.throttle.NoLimitThroughputController; import org.apache.hadoop.hbase.security.User; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRemoveRegionMetrics.java b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRemoveRegionMetrics.java index 54c6df860950..acc3f9dd1eee 100644 --- a/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRemoveRegionMetrics.java +++ b/hbase-server/src/test/java/org/apache/hadoop/hbase/regionserver/TestRemoveRegionMetrics.java @@ -31,7 +31,7 @@ import org.apache.hadoop.hbase.client.RegionInfo; import org.apache.hadoop.hbase.client.RegionLocator; import org.apache.hadoop.hbase.client.Table; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.testclassification.RegionServerTests; import org.apache.hadoop.hbase.util.Bytes; diff --git a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestCallQueue.java b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestCallQueue.java index 8d4cf76d45c5..955b4ca9c99e 100644 --- a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestCallQueue.java +++ b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestCallQueue.java @@ -24,7 +24,7 @@ import org.apache.hadoop.hbase.CompatibilitySingletonFactory; import org.apache.hadoop.hbase.HBaseClassTestRule; import org.apache.hadoop.hbase.HBaseTestingUtility; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.ClientTests; import org.apache.hadoop.hbase.testclassification.SmallTests; import org.junit.ClassRule; diff --git a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestThriftServer.java b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestThriftServer.java index 7a7de2abc937..5cda76af1684 100644 --- a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestThriftServer.java +++ b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift/TestThriftServer.java @@ -47,7 +47,7 @@ import org.apache.hadoop.hbase.client.TableDescriptorBuilder; import org.apache.hadoop.hbase.filter.ParseFilter; import org.apache.hadoop.hbase.security.UserProvider; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.ClientTests; import org.apache.hadoop.hbase.testclassification.LargeTests; import org.apache.hadoop.hbase.thrift.ThriftMetrics.ThriftServerType; diff --git a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift2/TestThriftHBaseServiceHandler.java b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift2/TestThriftHBaseServiceHandler.java index 273d7e0a1f8d..0a6b42b69c34 100644 --- a/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift2/TestThriftHBaseServiceHandler.java +++ b/hbase-thrift/src/test/java/org/apache/hadoop/hbase/thrift2/TestThriftHBaseServiceHandler.java @@ -74,7 +74,7 @@ import org.apache.hadoop.hbase.coprocessor.RegionObserver; import org.apache.hadoop.hbase.filter.ParseFilter; import org.apache.hadoop.hbase.security.UserProvider; -import org.apache.hadoop.hbase.MetricsAssertHelper; +import org.apache.hadoop.hbase.test.MetricsAssertHelper; import org.apache.hadoop.hbase.testclassification.ClientTests; import org.apache.hadoop.hbase.testclassification.MediumTests; import org.apache.hadoop.hbase.thrift.ErrorThrowingGetObserver; From fc7c306c49e584a7a66fa243a6a606fd146305c3 Mon Sep 17 00:00:00 2001 From: Lokesh Khurana Date: Thu, 23 Apr 2020 20:08:54 +0530 Subject: [PATCH 6/7] Making pom changes --- hbase-common/pom.xml | 4 - hbase-it/pom.xml | 18 ---- .../hadoop/hbase/HBaseTestingUtility.java | 100 +++++++++--------- 3 files changed, 50 insertions(+), 72 deletions(-) diff --git a/hbase-common/pom.xml b/hbase-common/pom.xml index 22c04e3e8587..d11bf7905048 100644 --- a/hbase-common/pom.xml +++ b/hbase-common/pom.xml @@ -220,10 +220,6 @@ junit junit - - junit - junit - diff --git a/hbase-it/pom.xml b/hbase-it/pom.xml index 50749a994da9..6b173f0ffd67 100644 --- a/hbase-it/pom.xml +++ b/hbase-it/pom.xml @@ -154,29 +154,11 @@ test-jar test - - org.apache.hbase - hbase-server - test-jar - test - - - org.apache.hbase - hbase-zookeeper - test-jar - test - org.apache.hbase hbase-common jar - - org.apache.hbase - hbase-common - test-jar - test - org.apache.hbase hbase-protocol diff --git a/hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseTestingUtility.java b/hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseTestingUtility.java index e9c1cccfc85b..d312d69035d4 100644 --- a/hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseTestingUtility.java +++ b/hbase-server/src/main/java/org/apache/hadoop/hbase/HBaseTestingUtility.java @@ -335,7 +335,7 @@ public HBaseTestingUtility(@Nullable Configuration conf) { String dataTestDir = getDataTestDir().toString(); this.conf.set("fs.defaultFS","file:///"); this.conf.set(HConstants.HBASE_DIR, "file://" + dataTestDir); - HBaseCommonTestingUtility.LOG.debug("Setting {} to {}", HConstants.HBASE_DIR, dataTestDir); + LOG.debug("Setting {} to {}", HConstants.HBASE_DIR, dataTestDir); this.conf.setBoolean(CommonFSUtils.UNSAFE_STREAM_CAPABILITY_ENFORCE,false); // If the value for random ports isn't set set it to true, thus making // tests opt-out for random port assignment @@ -429,11 +429,11 @@ private void createSubDirAndSystemProperty( if (sysValue != null) { // There is already a value set. So we do nothing but hope // that there will be no conflicts - HBaseCommonTestingUtility.LOG.info("System.getProperty(\""+propertyName+"\") already set to: "+ + LOG.info("System.getProperty(\""+propertyName+"\") already set to: "+ sysValue + " so I do NOT create it in " + parent); String confValue = conf.get(propertyName); if (confValue != null && !confValue.endsWith(sysValue)){ - HBaseCommonTestingUtility.LOG.warn( + LOG.warn( propertyName + " property value differs in configuration and system: "+ "Configuration="+confValue+" while System="+sysValue+ " Erasing configuration value by system value." @@ -489,7 +489,7 @@ public Path getDataTestDirOnTestFS(final String subdirName) throws IOException { */ private void setupDataTestDirOnTestFS() throws IOException { if (dataTestDirOnTestFS != null) { - HBaseCommonTestingUtility.LOG.warn("Data test on test fs dir already setup in " + LOG.warn("Data test on test fs dir already setup in " + dataTestDirOnTestFS.toString()); return; } @@ -506,7 +506,7 @@ private Path getNewDataTestDirOnTestFS() throws IOException { //the working directory, and create a unique sub dir there FileSystem fs = getTestFileSystem(); Path newDataTestDir; - String randomStr = HBaseCommonTestingUtility.getRandomUUID().toString(); + String randomStr = getRandomUUID().toString(); if (fs.getUri().getScheme().equals(FileSystem.getLocal(conf).getUri().getScheme())) { newDataTestDir = new Path(getDataTestDir(), randomStr); File dataTestDir = new File(newDataTestDir.toString()); @@ -588,7 +588,7 @@ public MiniDFSCluster startMiniDFSCluster(int servers, final String hosts[]) private void setFs() throws IOException { if(this.dfsCluster == null){ - HBaseCommonTestingUtility.LOG.info("Skipping setting fs because dfsCluster is null"); + LOG.info("Skipping setting fs because dfsCluster is null"); return; } FileSystem fs = this.dfsCluster.getFileSystem(); @@ -624,7 +624,7 @@ public MiniDFSCluster startMiniDFSCluster(int servers, final String racks[], St dataTestDirOnTestFS = null; String dataTestDir = getDataTestDir().toString(); conf.set(HConstants.HBASE_DIR, dataTestDir); - HBaseCommonTestingUtility.LOG.debug("Setting {} to {}", HConstants.HBASE_DIR, dataTestDir); + LOG.debug("Setting {} to {}", HConstants.HBASE_DIR, dataTestDir); return this.dfsCluster; } @@ -693,7 +693,7 @@ public boolean isReadShortCircuitOn(){ private void enableShortCircuit() { if (isReadShortCircuitOn()) { String curUser = System.getProperty("user.name"); - HBaseCommonTestingUtility.LOG.info("read short circuit is ON for user " + curUser); + LOG.info("read short circuit is ON for user " + curUser); // read short circuit, for hdfs conf.set("dfs.block.local-path-access.user", curUser); // read short circuit, for hbase @@ -701,7 +701,7 @@ private void enableShortCircuit() { // Skip checking checksum, for the hdfs client and the datanode conf.setBoolean("dfs.client.read.shortcircuit.skip.checksum", true); } else { - HBaseCommonTestingUtility.LOG.info("read short circuit is OFF"); + LOG.info("read short circuit is OFF"); } } @@ -710,7 +710,7 @@ private String createDirAndSetProperty(final String relPath, String property) { System.setProperty(property, path); conf.set(property, path); new File(path).mkdirs(); - HBaseCommonTestingUtility.LOG.info("Setting " + property + " to " + path + " in system properties and HBase conf"); + LOG.info("Setting " + property + " to " + path + " in system properties and HBase conf"); return path; } @@ -1021,7 +1021,7 @@ public MiniHBaseCluster startMiniCluster() throws Exception { * @see #shutdownMiniDFSCluster() */ public MiniHBaseCluster startMiniCluster(StartMiniClusterOption option) throws Exception { - HBaseCommonTestingUtility.LOG.info("Starting up minicluster with option: {}", option); + LOG.info("Starting up minicluster with option: {}", option); // If we already put up a cluster, fail. if (miniClusterRunning) { @@ -1035,10 +1035,10 @@ public MiniHBaseCluster startMiniCluster(StartMiniClusterOption option) throws E // Bring up mini dfs cluster. This spews a bunch of warnings about missing // scheme. Complaints are 'Scheme is undefined for build/test/data/dfs/name1'. if (dfsCluster == null) { - HBaseCommonTestingUtility.LOG.info("STARTING DFS"); + LOG.info("STARTING DFS"); dfsCluster = startMiniDFSCluster(option.getNumDataNodes(), option.getDataNodeHosts()); } else { - HBaseCommonTestingUtility.LOG.info("NOT STARTING DFS"); + LOG.info("NOT STARTING DFS"); } // Start up a zk cluster. @@ -1096,7 +1096,7 @@ public MiniHBaseCluster startMiniHBaseCluster(StartMiniClusterOption option) t.close(); getAdmin(); // create immediately the hbaseAdmin - HBaseCommonTestingUtility.LOG.info("Minicluster is up; activeMaster={}", getHBaseCluster().getMaster()); + LOG.info("Minicluster is up; activeMaster={}", getHBaseCluster().getMaster()); return (MiniHBaseCluster) hbaseCluster; } @@ -1222,7 +1222,7 @@ public void restartHBaseCluster(StartMiniClusterOption option) while (s.next() != null) { // do nothing } - HBaseCommonTestingUtility.LOG.info("HBase has been restarted"); + LOG.info("HBase has been restarted"); s.close(); t.close(); conn.close(); @@ -1246,14 +1246,14 @@ public MiniHBaseCluster getMiniHBaseCluster() { * @see #startMiniCluster(int) */ public void shutdownMiniCluster() throws IOException { - HBaseCommonTestingUtility.LOG.info("Shutting down minicluster"); + LOG.info("Shutting down minicluster"); shutdownMiniHBaseCluster(); shutdownMiniDFSCluster(); shutdownMiniZKCluster(); cleanupTestDir(); miniClusterRunning = false; - HBaseCommonTestingUtility.LOG.info("Minicluster is down"); + LOG.info("Minicluster is down"); } /** @@ -1377,9 +1377,9 @@ private void setHBaseFsTmpDir() throws IOException { String hbaseFsTmpDirInString = this.conf.get("hbase.fs.tmp.dir"); if (hbaseFsTmpDirInString == null) { this.conf.set("hbase.fs.tmp.dir", getDataTestDirOnTestFS("hbase-staging").toString()); - HBaseCommonTestingUtility.LOG.info("Setting hbase.fs.tmp.dir to " + this.conf.get("hbase.fs.tmp.dir")); + LOG.info("Setting hbase.fs.tmp.dir to " + this.conf.get("hbase.fs.tmp.dir")); } else { - HBaseCommonTestingUtility.LOG.info("The hbase.fs.tmp.dir is set to " + hbaseFsTmpDirInString); + LOG.info("The hbase.fs.tmp.dir is set to " + hbaseFsTmpDirInString); } } @@ -1871,7 +1871,7 @@ public void deleteTable(TableName tableName) throws IOException { try { getAdmin().disableTable(tableName); } catch (TableNotEnabledException e) { - HBaseCommonTestingUtility.LOG.debug("Table: " + tableName + " already disabled, so just deleting it."); + LOG.debug("Table: " + tableName + " already disabled, so just deleting it."); } getAdmin().deleteTable(tableName); } @@ -2598,7 +2598,7 @@ public List getMetaTableRows() throws IOException { List rows = new ArrayList<>(); ResultScanner s = t.getScanner(new Scan()); for (Result result : s) { - HBaseCommonTestingUtility.LOG.info("getMetaTableRows: row -> " + + LOG.info("getMetaTableRows: row -> " + Bytes.toStringBinary(result.getRow())); rows.add(result.getRow()); } @@ -2620,13 +2620,13 @@ public List getMetaTableRows(TableName tableName) throws IOException { for (Result result : s) { RegionInfo info = MetaTableAccessor.getRegionInfo(result); if (info == null) { - HBaseCommonTestingUtility.LOG.error("No region info for row " + Bytes.toString(result.getRow())); + LOG.error("No region info for row " + Bytes.toString(result.getRow())); // TODO figure out what to do for this new hosed case. continue; } if (info.getTable().equals(tableName)) { - HBaseCommonTestingUtility.LOG.info("getMetaTableRows: row -> " + + LOG.info("getMetaTableRows: row -> " + Bytes.toStringBinary(result.getRow()) + info); rows.add(result.getRow()); } @@ -2678,7 +2678,7 @@ public HRegionServer getRSForFirstRegionInTable(TableName tableName) if (regions == null || regions.isEmpty()) { return null; } - HBaseCommonTestingUtility.LOG.debug("Found " + regions.size() + " regions for table " + + LOG.debug("Found " + regions.size() + " regions for table " + tableName); byte[] firstRegionName = regions.stream() @@ -2687,7 +2687,7 @@ public HRegionServer getRSForFirstRegionInTable(TableName tableName) .findFirst() .orElseThrow(() -> new IOException("online regions not found in table " + tableName)); - HBaseCommonTestingUtility.LOG.debug("firstRegionName=" + Bytes.toString(firstRegionName)); + LOG.debug("firstRegionName=" + Bytes.toString(firstRegionName)); long pause = getConfiguration().getLong(HConstants.HBASE_CLIENT_PAUSE, HConstants.DEFAULT_HBASE_CLIENT_PAUSE); int numRetries = getConfiguration().getInt(HConstants.HBASE_CLIENT_RETRIES_NUMBER, @@ -2756,7 +2756,7 @@ private void startMiniMapReduceCluster(final int servers) throws IOException { if (mrCluster != null) { throw new IllegalStateException("MiniMRCluster is already running"); } - HBaseCommonTestingUtility.LOG.info("Starting mini mapreduce cluster..."); + LOG.info("Starting mini mapreduce cluster..."); setupClusterTestDir(); createDirsAndSetProperties(); @@ -2784,7 +2784,7 @@ private void startMiniMapReduceCluster(final int servers) throws IOException { jobConf.set("mapreduce.cluster.local.dir", conf.get("mapreduce.cluster.local.dir")); //Hadoop MiniMR overwrites this while it should not - HBaseCommonTestingUtility.LOG.info("Mini mapreduce cluster started"); + LOG.info("Mini mapreduce cluster started"); // In hadoop2, YARN/MR2 starts a mini cluster with its own conf instance and updates settings. // Our HBase MR jobs need several of these settings in order to properly run. So we copy the @@ -2823,10 +2823,10 @@ private void startMiniMapReduceCluster(final int servers) throws IOException { */ public void shutdownMiniMapReduceCluster() { if (mrCluster != null) { - HBaseCommonTestingUtility.LOG.info("Stopping mini mapreduce cluster..."); + LOG.info("Stopping mini mapreduce cluster..."); mrCluster.shutdown(); mrCluster = null; - HBaseCommonTestingUtility.LOG.info("Mini mapreduce cluster stopped"); + LOG.info("Mini mapreduce cluster stopped"); } // Restore configuration to point to local jobtracker conf.set("mapreduce.jobtracker.address", "local"); @@ -2950,7 +2950,7 @@ public void expireSession(ZKWatcher nodeZK, boolean checkStatus) 1000, new org.apache.zookeeper.Watcher(){ @Override public void process(WatchedEvent watchedEvent) { - HBaseCommonTestingUtility.LOG.info("Monitor ZKW received event="+watchedEvent); + LOG.info("Monitor ZKW received event="+watchedEvent); } } , sessionID, password); @@ -2966,7 +2966,7 @@ public void process(WatchedEvent watchedEvent) { Thread.sleep(1); } newZK.close(); - HBaseCommonTestingUtility.LOG.info("ZK Closed Session 0x" + Long.toHexString(sessionID)); + LOG.info("ZK Closed Session 0x" + Long.toHexString(sessionID)); // Now closing & waiting to be sure that the clients get it. monitor.close(); @@ -3019,7 +3019,7 @@ public void invalidateConnection() throws IOException { // Update the master addresses if they changed. final String masterConfigBefore = conf.get(HConstants.MASTER_ADDRS_KEY); final String masterConfAfter = getMiniHBaseCluster().conf.get(HConstants.MASTER_ADDRS_KEY); - HBaseCommonTestingUtility.LOG.info("Invalidated connection. Updating master addresses before: {} after: {}", + LOG.info("Invalidated connection. Updating master addresses before: {} after: {}", masterConfigBefore, masterConfAfter); conf.set(HConstants.MASTER_ADDRS_KEY, getMiniHBaseCluster().conf.get(HConstants.MASTER_ADDRS_KEY)); @@ -3142,7 +3142,7 @@ public HRegion getSplittableRegion(TableName tableName, int maxAttempts) { regions.get(idx).checkSplit(); return regions.get(idx); } catch (Exception ex) { - HBaseCommonTestingUtility.LOG.warn("Caught exception", ex); + LOG.warn("Caught exception", ex); attempted.add(idx); } } @@ -3338,7 +3338,7 @@ public boolean ensureSomeRegionServersAvailable(final int num) boolean startedServer = false; MiniHBaseCluster hbaseCluster = getMiniHBaseCluster(); for (int i=hbaseCluster.getLiveRegionServerThreads().size(); i() { @Override @@ -3582,14 +3582,14 @@ public boolean evaluate() throws IOException { } } if (!tableFound) { - HBaseCommonTestingUtility.LOG.warn("Didn't find the entries for table " + tableName + " in meta, already deleted?"); + LOG.warn("Didn't find the entries for table " + tableName + " in meta, already deleted?"); } return tableFound; } }); } } - HBaseCommonTestingUtility.LOG.info("All regions for table " + tableName + " assigned to meta. Checking AM states."); + LOG.info("All regions for table " + tableName + " assigned to meta. Checking AM states."); // check from the master state if we are using a mini cluster if (!getHBaseClusterInterface().isDistributedCluster()) { // So, all regions are in the meta table but make sure master knows of the assignments before @@ -3609,7 +3609,7 @@ public boolean evaluate() throws IOException { } }); } - HBaseCommonTestingUtility.LOG.info("All regions for table " + tableName + " assigned."); + LOG.info("All regions for table " + tableName + " assigned."); } /** @@ -3722,7 +3722,7 @@ public Table createRandomTable(TableName tableName, final int numRowsPerFlush) throws IOException, InterruptedException { - HBaseCommonTestingUtility.LOG.info("\n\nCreating random table " + tableName + " with " + numRegions + + LOG.info("\n\nCreating random table " + tableName + " with " + numRegions + " regions, " + numFlushes + " storefiles per region, " + numRowsPerFlush + " rows per flush, maxVersions=" + maxVersions + "\n"); @@ -3786,7 +3786,7 @@ public Table createRandomTable(TableName tableName, mutator.mutate(del); } } - HBaseCommonTestingUtility.LOG.info("Initiating flush #" + iFlush + " for table " + tableName); + LOG.info("Initiating flush #" + iFlush + " for table " + tableName); mutator.flush(); if (hbaseCluster != null) { getMiniHBaseCluster().flushcache(table.getName()); @@ -3880,13 +3880,13 @@ public static void waitForHostPort(String host, int port) final int maxTimeMs = 10000; final int maxNumAttempts = maxTimeMs / HConstants.SOCKET_RETRY_WAIT_MS; IOException savedException = null; - HBaseCommonTestingUtility.LOG.info("Waiting for server at " + host + ":" + port); + LOG.info("Waiting for server at " + host + ":" + port); for (int attempt = 0; attempt < maxNumAttempts; ++attempt) { try { Socket sock = new Socket(InetAddress.getByName(host), port); sock.close(); savedException = null; - HBaseCommonTestingUtility.LOG.info("Server at " + host + ":" + port + " is available"); + LOG.info("Server at " + host + ":" + port + " is available"); break; } catch (UnknownHostException e) { throw new IOException("Failed to look up " + host, e); @@ -4022,7 +4022,7 @@ public static int createPreSplitLoadTestTable(Configuration conf, } totalNumberOfRegions = numberOfServers * numRegionsPerServer; - HBaseCommonTestingUtility.LOG.info("Number of live regionservers: " + numberOfServers + ", " + + LOG.info("Number of live regionservers: " + numberOfServers + ", " + "pre-splitting table into " + totalNumberOfRegions + " regions " + "(regions per server: " + numRegionsPerServer + ")"); @@ -4031,10 +4031,10 @@ public static int createPreSplitLoadTestTable(Configuration conf, admin.createTable(td, splits); } catch (MasterNotRunningException e) { - HBaseCommonTestingUtility.LOG.error("Master not running", e); + LOG.error("Master not running", e); throw new IOException(e); } catch (TableExistsException e) { - HBaseCommonTestingUtility.LOG.warn("Table " + td.getTableName() + + LOG.warn("Table " + td.getTableName() + " already exists, continuing"); } finally { admin.close(); @@ -4365,10 +4365,10 @@ public MiniKdc setupMiniKdc(File keytabFile) throws Exception { FileUtils.deleteDirectory(dir); // clean directory numTries++; if (numTries == 3) { - HBaseCommonTestingUtility.LOG.error("Failed setting up MiniKDC. Tried " + numTries + " times."); + LOG.error("Failed setting up MiniKDC. Tried " + numTries + " times."); throw e; } - HBaseCommonTestingUtility.LOG.error("BindException encountered when setting up MiniKdc. Trying again."); + LOG.error("BindException encountered when setting up MiniKdc. Trying again."); bindException = true; } } while (bindException); From c133bbd197bd39738b4090388573fb76c63b93aa Mon Sep 17 00:00:00 2001 From: Lokesh Khurana Date: Thu, 23 Apr 2020 20:13:52 +0530 Subject: [PATCH 7/7] refactoring --- .../org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java | 2 +- .../java/org/apache/hadoop/hbase/HBaseZKTestingUtility.java | 4 ++-- 2 files changed, 3 insertions(+), 3 deletions(-) diff --git a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java index d983dd4710c1..082254bcb6ac 100644 --- a/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java +++ b/hbase-it/src/main/java/org/apache/hadoop/hbase/chaos/monkies/ChaosMonkey.java @@ -38,7 +38,7 @@ * policy can be active at any time. *

* Chaos monkey can be run from the command line, or can be invoked from integration tests. - * See {org.apache.hadoop.hbase.IntegrationTestIngest in hbase-it /test} or other integration tests that use + * See {@link org.apache.hadoop.hbase.IntegrationTestIngest} or other integration tests that use * chaos monkey for code examples. *

* ChaosMonkey class is indeed inspired by the Netflix's same-named tool: diff --git a/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/HBaseZKTestingUtility.java b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/HBaseZKTestingUtility.java index 42ce1f9a29c7..dd8b3909d6d2 100644 --- a/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/HBaseZKTestingUtility.java +++ b/hbase-zookeeper/src/main/java/org/apache/hadoop/hbase/HBaseZKTestingUtility.java @@ -73,14 +73,14 @@ protected void setupClusterTestDir() { // Using randomUUID ensures that multiple clusters can be launched by // a same test, if it stops & starts them - Path testDir = getDataTestDir("cluster_" + HBaseCommonTestingUtility.getRandomUUID().toString()); + Path testDir = getDataTestDir("cluster_" + getRandomUUID().toString()); clusterTestDir = new File(testDir.toString()).getAbsoluteFile(); // Have it cleaned up on exit boolean b = deleteOnExit(); if (b) { clusterTestDir.deleteOnExit(); } - HBaseCommonTestingUtility.LOG.info("Created new mini-cluster data directory: " + clusterTestDir + ", deleteOnExit=" + b); + LOG.info("Created new mini-cluster data directory: " + clusterTestDir + ", deleteOnExit=" + b); } /**