diff --git a/bigtop-manager-agent/pom.xml b/bigtop-manager-agent/pom.xml
index 6a6d7035..08fdcfcf 100644
--- a/bigtop-manager-agent/pom.xml
+++ b/bigtop-manager-agent/pom.xml
@@ -63,11 +63,6 @@
bigtop-manager-stack-bigtop
-
- org.apache.bigtop
- bigtop-manager-stack-nop
-
-
org.springframework.boot
spring-boot-starter-web
diff --git a/bigtop-manager-common/src/main/java/org/apache/bigtop/manager/common/utils/Environments.java b/bigtop-manager-common/src/main/java/org/apache/bigtop/manager/common/utils/Environments.java
index f344cfb7..e3b3cb12 100644
--- a/bigtop-manager-common/src/main/java/org/apache/bigtop/manager/common/utils/Environments.java
+++ b/bigtop-manager-common/src/main/java/org/apache/bigtop/manager/common/utils/Environments.java
@@ -23,8 +23,9 @@
public class Environments {
/**
- * Indicates whether the application is running in development mode, which is disabled by default.
- * In development mode, only NOP stacks are available and no real shell commands will be executed on the agent side.
+ * Indicates whether the application is running in development mode.
+ * In development mode, most tasks run on agent side will be proxied and return success by default.
+ * This should help developers test framework functions without depending on the existence of big data components.
*/
public static Boolean isDevMode() {
String devMode = System.getenv("DEV_MODE");
diff --git a/bigtop-manager-server/src/main/java/org/apache/bigtop/manager/server/utils/StackUtils.java b/bigtop-manager-server/src/main/java/org/apache/bigtop/manager/server/utils/StackUtils.java
index b178b33e..86b7525c 100644
--- a/bigtop-manager-server/src/main/java/org/apache/bigtop/manager/server/utils/StackUtils.java
+++ b/bigtop-manager-server/src/main/java/org/apache/bigtop/manager/server/utils/StackUtils.java
@@ -19,7 +19,6 @@
package org.apache.bigtop.manager.server.utils;
import org.apache.bigtop.manager.common.enums.Command;
-import org.apache.bigtop.manager.common.utils.Environments;
import org.apache.bigtop.manager.common.utils.JsonUtils;
import org.apache.bigtop.manager.server.exception.ApiException;
import org.apache.bigtop.manager.server.exception.ServerException;
@@ -77,8 +76,6 @@ public class StackUtils {
private static final String DEPENDENCY_FILE_NAME = "order.json";
- private static final String NOP_STACK = "nop";
-
private static final Map>> STACK_DEPENDENCY_MAP = new HashMap<>();
private static final Map>> STACK_CONFIG_MAP = new HashMap<>();
@@ -202,13 +199,6 @@ public static Map> stackList() {
for (File stackFolder : stackFolders) {
String stackName = stackFolder.getName();
-
- // If in dev mode, only parse nop stack
- // If not in dev mode, skip nop stack
- if (Environments.isDevMode() != stackName.equals(NOP_STACK)) {
- continue;
- }
-
File[] versionFolders = Optional.ofNullable(stackFolder.listFiles()).orElse(new File[0]);
for (File versionFolder : versionFolders) {
diff --git a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/metainfo.xml b/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/metainfo.xml
deleted file mode 100644
index c87d7b7d..00000000
--- a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/metainfo.xml
+++ /dev/null
@@ -1,65 +0,0 @@
-
-
-
-
-
- nop
- 1.0.0
- hadoop
- /opt
- curl,wget
-
-baseurl=${baseUrl}
-<#else>
-mirrorlist=${mirrorList}
-#if>
-path=/
-enabled=1
-gpgcheck=0
-]]>
-
-
-
- NOP-1.0.0
- NOP-1.0.0 for CentOS-7 x86_64
- centos7
- x86_64
- http://123.56.2.244/nop/1.0.0/centos7/x86_64/
-
-
- NOP-1.0.0
- NOP-1.0.0 for CentOS-7 aarch64
- centos7
- aarch64
- https://bigtop-snapshot.s3.amazonaws.com/centos-7/$basearch
-
-
- NOP-1.0.0
- NOP-1.0.0 for RockyLinux-8 x86_64
- rocky8
- x86_64
- http://123.56.2.244/nop/1.0.0/rocky8/x86_64/
-
-
-
-
\ No newline at end of file
diff --git a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/configuration/kafka-broker.xml b/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/configuration/kafka-broker.xml
deleted file mode 100644
index 1a4a5a75..00000000
--- a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/configuration/kafka-broker.xml
+++ /dev/null
@@ -1,537 +0,0 @@
-
-
-
-
-
- log.dirs
- Log directories
- /kafka-logs
-
- A comma-separated list of one or more directories in which Kafka data is stored.
- Each new partition that is created will be placed in the directory which currently has the fewest
- partitions.
-
-
-
- zookeeper.connect
- <#list zkHostList as host>${host}:2181<#sep>,#sep>#list><#else>localhost:2181#if>]]>
-
- Zookeeper also allows you to add a "chroot" path which will make all kafka data for this cluster appear
- under a particular path.
- This is a way to setup multiple Kafka clusters or other applications on the same zookeeper cluster.
- To do this give a connection string in the form hostname1:port1,hostname2:port2,hostname3:port3/chroot/path
- which would put all this cluster's data under the path /chroot/path. Note that consumers must use the same
- connection string.
-
-
-
- listeners
- ${host}:9092<#else>localhost:9092#if>]]>
-
- host and port where kafka broker will be accepting connections. localhost will be substituted with hostname.
-
-
-
- advertised.listeners
- ${host}:9092<#else>localhost:9092#if>]]>
-
- Listeners to publish to ZooKeeper for clients to use, if different than the listeners config property.
-
-
-
- message.max.bytes
- 1000000
-
- The maximum size of a message that the server can receive.
- It is important that this property be in sync with the maximum fetch size your consumers use or
- else an unruly producer will be able to publish messages too large for consumers to consume.
-
-
-
- num.network.threads
- 3
-
- The number of network threads that the server uses for handling network requests.
- You probably don't need to change this.
-
-
-
- num.io.threads
- 8
-
- The number of I/O threads that the server uses for executing requests. You should have at least as many
- threads as you have disks.
-
-
-
- queued.max.requests
- 500
-
- The number of requests that can be queued up for processing by the I/O threads before the network threads
- stop reading in new requests.
-
-
-
- socket.send.buffer.bytes
- 102400
- The SO_SNDBUFF buffer the server prefers for socket connections.
-
-
- socket.receive.buffer.bytes
- 102400
- The SO_RCVBUFF buffer the server prefers for socket connections.
-
-
- socket.request.max.bytes
- 104857600
-
- The maximum request size the server will allow. This prevents the server from running out of memory and
- should be smaller than the Java heap size.
-
-
-
- num.partitions
- 1
- The default number of partitions per topic.
-
-
- log.segment.bytes
- 1073741824
-
- The maximum request size the server will allow.
- This prevents the server from running out of memory and should be smaller than the Java heap size.
-
-
-
- log.roll.hours
- 168
-
- This setting will force Kafka to roll a new log segment even if the log.segment.bytes size has not been
- reached.
-
-
-
- log.retention.bytes
- -1
-
- The amount of data to retain in the log for each topic-partitions. Note that this is the limit per-partition
- so multiply by the number of partitions to get the total data retained for the topic.
- Also note that if both log.retention.hours and log.retention.bytes are both set we delete a segment when
- either limit is exceeded.
-
-
-
- log.retention.hours
- 168
-
- The number of hours to keep a log segment before it is deleted, i.e. the default data retention window for
- all topics.
- Note that if both log.retention.hours and log.retention.bytes are both set we delete a segment when either
- limit is exceeded.
-
-
-
- log.cleanup.interval.mins
- 10
-
- The frequency in minutes that the log cleaner checks whether any log segment is eligible for deletion to
- meet the retention policies.
-
-
-
- log.retention.check.interval.ms
- 600000
-
- The frequency in milliseconds that the log cleaner checks whether any log segment is eligible for deletion
- to meet the retention policies.
-
-
-
- log.index.size.max.bytes
- 10485760
-
- The maximum size in bytes we allow for the offset index for each log segment.
- Note that we will always pre-allocate a sparse file with this much space and shrink it down when the log
- rolls.
- If the index fills up we will roll a new log segment even if we haven't reached the log.segment.bytes limit.
-
-
-
- log.index.interval.bytes
- 4096
-
- The byte interval at which we add an entry to the offset index.
- When executing a fetch request the server must do a linear scan for up to this many bytes to find the
- correct position in the log to begin and end the fetch.
- So setting this value to be larger will mean larger index files (and a bit more memory usage) but less
- scanning.
- However the server will never add more than one index entry per log append (even if more than
- log.index.interval worth of messages are appended).
- In general you probably don't need to mess with this value.
-
-
-
- auto.create.topics.enable
- true
-
- Enable auto creation of topic on the server.
- If this is set to true then attempts to produce, consume, or fetch metadata for a non-existent topic will
- automatically create it with the default replication factor and number of partitions.
-
-
-
- controller.socket.timeout.ms
- 30000
-
- The socket timeout for commands from the partition management controller to the replicas.
-
-
-
- controller.message.queue.size
- 10
- The buffer size for controller-to-broker-channels
-
-
- default.replication.factor
- 1
- The default replication factor for automatically created topics.
-
-
- replica.lag.time.max.ms
- 10000
-
- If a follower hasn't sent any fetch requests for this window of time, the leader will remove the follower
- from ISR (in-sync replicas) and treat it as dead.
-
-
-
- replica.lag.max.messages
- 4000
-
- If a replica falls more than this many messages behind the leader, the leader will remove the follower from
- ISR and treat it as dead.
-
-
-
- replica.socket.timeout.ms
- 30000
- The socket timeout for network requests to the leader for replicating data.
-
-
- replica.socket.receive.buffer.bytes
- 65536
- The socket receive buffer for network requests to the leader for replicating data.
-
-
- replica.fetch.max.bytes
- 1048576
-
- The number of byes of messages to attempt to fetch for each partition in the fetch requests the replicas
- send to the leader.
-
-
-
- replica.fetch.wait.max.ms
- 500
-
- The maximum amount of time to wait time for data to arrive on the leader in the fetch requests sent by the
- replicas to the leader.
-
-
-
- replica.fetch.min.bytes
- 1
-
- Minimum bytes expected for each fetch response for the fetch requests from the replica to the leader.
- If not enough bytes, wait up to replica.fetch.wait.max.ms for this many bytes to arrive.
-
-
-
- num.replica.fetchers
- 1
-
- Number of threads used to replicate messages from leaders.
- Increasing this value can increase the degree of I/O parallelism in the follower broker.
-
-
-
- replica.high.watermark.checkpoint.interval.ms
- 5000
-
- The frequency with which each replica saves its high watermark to disk to handle recovery.
-
-
-
- fetch.purgatory.purge.interval.requests
- 10000
- The purge interval (in number of requests) of the fetch request purgatory.
-
-
- producer.purgatory.purge.interval.requests
- 10000
- The purge interval (in number of requests) of the producer request purgatory.
-
-
- zookeeper.session.timeout.ms
- 30000
-
- Zookeeper session timeout.
- If the server fails to heartbeat to zookeeper within this period of time it is considered dead.
- If you set this too low the server may be falsely considered dead; if you set it too high it may take too
- long to recognize a truly dead server.
-
-
-
- zookeeper.connection.timeout.ms
- 25000
-
- The maximum amount of time that the client waits to establish a connection to zookeeper.
-
-
-
- zookeeper.sync.time.ms
- 2000
- How far a ZK follower can be behind a ZK leader.
-
-
- controlled.shutdown.max.retries
- 3
-
- Number of retries to complete the controlled shutdown successfully before executing an unclean shutdown.
-
-
-
- controlled.shutdown.retry.backoff.ms
- 5000
- Backoff time between shutdown retries.
-
-
- controlled.shutdown.enable
- true
-
- Enable controlled shutdown of the broker.
- If enabled, the broker will move all leaders on it to some other brokers before shutting itself down.
- This reduces the unavailability window during shutdown.
-
-
-
- auto.leader.rebalance.enable
- true
-
- Enables auto leader balancing.
- A background thread checks and triggers leader balance if required at regular intervals.
-
-
-
- num.recovery.threads.per.data.dir
- 1
-
- The number of threads per data directory to be used for log recovery at startup and flushing at shutdown
-
-
-
- min.insync.replicas
- 1
-
- define the minimum number of replicas in ISR needed to satisfy a produce request with required.acks=-1 (or
- all)
-
-
-
- leader.imbalance.per.broker.percentage
- 10
-
- The ratio of leader imbalance allowed per broker.
- The controller would trigger a leader balance if it goes above this value per broker.
- The value is specified in percentage.
-
-
-
- leader.imbalance.check.interval.seconds
- 300
-
- The frequency with which the partition re-balance check is triggered by the controller
-
-
-
- offset.metadata.max.bytes
- 4096
- The maximum size for a metadata entry associated with an offset commit
-
-
- offsets.load.buffer.size
- 5242880
-
- Batch size for reading from the offsets segments when loading offsets into the cache.
-
-
-
- offsets.topic.replication.factor
- 3
-
- The replication factor for the offsets topic (set higher to ensure availability).
- To ensure that the effective replication factor of the offsets topic is the configured
- value, the number of alive brokers has to be at least the replication factor at the time of the first
- request for the offsets topic.
- If not, either the offsets topic creation will fail or it will get a replication factor of min(alive
- brokers, configured replication factor).
-
-
-
- offsets.topic.num.partitions
- 50
-
- The number of partitions for the offset commit topic (should not change after deployment)
-
-
-
- offsets.topic.segment.bytes
- 104857600
-
- The offsets topic segment bytes should be kept relatively small in order to facilitate faster log compaction
- and cache loads
-
-
-
- offsets.topic.compression.codec
- 0
-
- Compression codec for the offsets topic - compression may be used to achieve \"atomic\" commits.
- Default is NoCompression.
- For Gzip add value 1 , SnappyCompression add value 2, LZ4CompressionCodec 3.
-
-
-
- offsets.retention.minutes
- 86400000
- Log retention window in minutes for offsets topic
-
-
- offsets.retention.check.interval.ms
- 600000
- Frequency at which to check for stale offsets
-
-
- offsets.commit.timeout.ms
- 5000
-
- Offset commit will be delayed until all replicas for the offsets topic receive the commit or this timeout is
- reached. This is similar to the producer request timeout.
-
-
-
- offsets.commit.required.acks
- -1
-
- The required acks before the commit can be accepted.
- In general, the default (-1) should not be overridden
-
-
-
- delete.topic.enable
- true
-
- Enables delete topic.
- Delete topic through the admin tool will have no effect if this config is turned off
-
-
-
- compression.type
- producer
-
- Specify the final compression type for a given topic.
- This configuration accepts the standard compression codecs ('gzip', 'snappy', lz4).
- It additionally accepts 'uncompressed' which is equivalent to no compression; and 'producer' which means
- retain the original compression codec set by the producer.
-
-
-
- external.kafka.metrics.exclude.prefix
- kafka.network.RequestMetrics,kafka.server.DelayedOperationPurgatory,kafka.server.BrokerTopicMetrics.BytesRejectedPerSec
- Exclude metrics starting with these prefixes from being collected.
-
-
- external.kafka.metrics.include.prefix
- kafka.network.RequestMetrics.ResponseQueueTimeMs.request.OffsetCommit.98percentile,kafka.network.RequestMetrics.ResponseQueueTimeMs.request.Offsets.95percentile,kafka.network.RequestMetrics.ResponseSendTimeMs.request.Fetch.95percentile,kafka.network.RequestMetrics.RequestsPerSec.request
-
- These metrics would be included even if the exclude prefix omits them.
-
-
-
- sasl.enabled.mechanisms
- GSSAPI
-
- The list of SASL mechanisms enabled in the Kafka server.
- The list may contain any mechanism for which a security provider is available.
- Only GSSAPI is enabled by default.
-
-
-
- security.inter.broker.protocol
- PLAINTEXT
-
- Security protocol used to communicate between brokers.
- Valid values are: PLAINTEXT, SSL, SASL_PLAINTEXT, SASL_SSL.
- It is an error to set this and inter.broker.listener.name properties at the same time.
-
-
-
- sasl.mechanism.inter.broker.protocol
- GSSAPI
- SASL mechanism used for inter-broker communication. Default is GSSAPI.
-
-
- ssl.client.auth
- none
- Configures kafka broker to request client authentication.
-
-
- ssl.key.password
-
- The password of private key in the key store file.
-
-
- ssl.keystore.location
-
- The location of key store file.
-
-
- ssl.keystore.password
-
- The store password for key store file.
-
-
- ssl.truststore.location
-
- The location of trust store file.
-
-
- ssl.truststore.password
-
-
- The password for trust store file.
- If a password is not set access to the truststore is still available, but integrity checking is disabled
-
-
-
- producer.metrics.enable
- false
-
-
\ No newline at end of file
diff --git a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/configuration/kafka-env.xml b/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/configuration/kafka-env.xml
deleted file mode 100644
index 794129d9..00000000
--- a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/configuration/kafka-env.xml
+++ /dev/null
@@ -1,71 +0,0 @@
-
-
-
-
-
- logDir
- Kafka Log directory
- /var/log/kafka
-
-
-
- pidDir
- /var/run/kafka
- Kafka PID dir
-
-
-
- kafkaUserNofileLimit
- 128000
- Max open files limit setting for KAFKA user.
-
-
- kafkaUserNprocLimit
- 65536
- Max number of processes limit setting for KAFKA user.
-
-
-
- content
- kafka-env template
- This is the freemarker template for kafka-env.sh file
-
-export KAFKA_OPTS="-Djavax.security.auth.useSubjectCredsOnly=false {{kafka_kerberos_params}}"
-<#else>
-export KAFKA_OPTS={{kafka_kerberos_params}}
-#if>
-export CLASSPATH=$CLASSPATH:${CONF_DIR}
-]]>
-
-
- longtext
-
-
-
\ No newline at end of file
diff --git a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/configuration/kafka-log4j.xml b/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/configuration/kafka-log4j.xml
deleted file mode 100644
index f6f76a09..00000000
--- a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/configuration/kafka-log4j.xml
+++ /dev/null
@@ -1,143 +0,0 @@
-
-
-
-
-
- kafkaLogMaxFileSize
- 256
- The maximum size of backup file before the log is rotated
- Kafka Log: backup file size
-
-
- kafkaLogMaxBackupIndex
- 20
- The number of backup files
- Kafka Log: # of backup files
-
-
- controllerLogMaxFileSize
- 256
- The maximum size of backup file before the log is rotated
- Kafka Controller Log: backup file size
-
-
- controllerLogMaxBackupIndex
- 20
- The number of backup files
- Kafka Controller Log: # of backup files
-
-
- content
- kafka-log4j template
- Custom log4j.properties
- log4j.appender.kafkaAppender.File=${kafka.logs.dir}/server.log#noparse>
-log4j.appender.kafkaAppender.layout=org.apache.log4j.PatternLayout
-log4j.appender.kafkaAppender.layout.ConversionPattern=[%d{ISO8601}] %p %m (%c)%n
-log4j.appender.kafkaAppender.MaxFileSize = ${kafkaLogMaxFileSize}MB
-log4j.appender.kafkaAppender.MaxBackupIndex = ${kafkaLogMaxBackupIndex}
-
-log4j.appender.stateChangeAppender=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.stateChangeAppender.DatePattern='.'yyyy-MM-dd-HH
-<#noparse>log4j.appender.stateChangeAppender.File=${kafka.logs.dir}/state-change.log#noparse>
-log4j.appender.stateChangeAppender.layout=org.apache.log4j.PatternLayout
-log4j.appender.stateChangeAppender.layout.ConversionPattern=[%d{ISO8601}] %p %m (%c)%n
-
-log4j.appender.requestAppender=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.requestAppender.DatePattern='.'yyyy-MM-dd-HH
-<#noparse>log4j.appender.requestAppender.File=${kafka.logs.dir}/kafka-request.log#noparse>
-log4j.appender.requestAppender.layout=org.apache.log4j.PatternLayout
-log4j.appender.requestAppender.layout.ConversionPattern=[%d{ISO8601}] %p %m (%c)%n
-
-log4j.appender.cleanerAppender=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.cleanerAppender.DatePattern='.'yyyy-MM-dd-HH
-<#noparse>log4j.appender.cleanerAppender.File=${kafka.logs.dir}/log-cleaner.log#noparse>
-log4j.appender.cleanerAppender.layout=org.apache.log4j.PatternLayout
-log4j.appender.cleanerAppender.layout.ConversionPattern=[%d{ISO8601}] %p %m (%c)%n
-
-log4j.appender.controllerAppender=org.apache.log4j.DailyRollingFileAppender
-log4j.appender.controllerAppender.DatePattern='.'yyyy-MM-dd-HH
-<#noparse>log4j.appender.controllerAppender.File=${kafka.logs.dir}/controller.log#noparse>
-log4j.appender.controllerAppender.layout=org.apache.log4j.PatternLayout
-log4j.appender.controllerAppender.layout.ConversionPattern=[%d{ISO8601}] %p %m (%c)%n
-log4j.appender.controllerAppender.MaxFileSize = ${controllerLogMaxFileSize}MB
-log4j.appender.controllerAppender.MaxBackupIndex = ${controllerLogMaxBackupIndex}
-# Turn on all our debugging info
-#log4j.logger.kafka.producer.async.DefaultEventHandler=DEBUG, kafkaAppender
-#log4j.logger.kafka.client.ClientUtils=DEBUG, kafkaAppender
-#log4j.logger.kafka.perf=DEBUG, kafkaAppender
-<#noparse>#log4j.logger.kafka.perf.ProducerPerformance$ProducerThread=DEBUG, kafkaAppender#noparse>
-#log4j.logger.org.I0Itec.zkclient.ZkClient=DEBUG
-log4j.logger.kafka=INFO, kafkaAppender
-<#noparse>log4j.logger.kafka.network.RequestChannel$=WARN, requestAppender#noparse>
-<#noparse>log4j.additivity.kafka.network.RequestChannel$=false#noparse>
-
-#log4j.logger.kafka.network.Processor=TRACE, requestAppender
-#log4j.logger.kafka.server.KafkaApis=TRACE, requestAppender
-#log4j.additivity.kafka.server.KafkaApis=false
-log4j.logger.kafka.request.logger=WARN, requestAppender
-log4j.additivity.kafka.request.logger=false
-
-log4j.logger.kafka.controller=TRACE, controllerAppender
-log4j.additivity.kafka.controller=false
-
-log4j.logger.kafka.log.LogCleaner=INFO, cleanerAppender
-log4j.additivity.kafka.log.LogCleaner=false
-
-log4j.logger.state.change.logger=TRACE, stateChangeAppender
-log4j.additivity.state.change.logger=false
-]]>
-
-
- longtext
-
-
-
\ No newline at end of file
diff --git a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/configuration/kafka.conf.xml b/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/configuration/kafka.conf.xml
deleted file mode 100644
index 3b9a1518..00000000
--- a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/configuration/kafka.conf.xml
+++ /dev/null
@@ -1,53 +0,0 @@
-
-
-
-
-
- content
- kafka.conf template
- This is the freemarker template for kafka-env.sh file
-
-
-
- longtext
-
-
-
\ No newline at end of file
diff --git a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/metainfo.xml b/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/metainfo.xml
deleted file mode 100644
index c2c8df6b..00000000
--- a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/metainfo.xml
+++ /dev/null
@@ -1,73 +0,0 @@
-
-
-
-
-
- kafka
- Kafka
-
- Apache Kafka is an open-source distributed event streaming platform used by thousands of companies for
- high-performance data pipelines, streaming analytics, data integration, and mission-critical
- applications.
-
- 2.8.2
- kafka
-
-
-
- kafka_broker
- Kafka Broker
- master
- 1+
-
- org.apache.bigtop.manager.stack.nop.v1_0_0.kafka.KafkaBrokerScript
- java
- 1200
-
-
-
- test
-
- org.apache.bigtop.manager.stack.nop.v1_0_0.kafka.KafkaBrokerScript
- java
- 600
-
-
-
-
-
-
-
-
-
- centos7
- rocky8
-
-
- x86_64
-
-
- kafka_1_0_0
-
-
-
-
-
-
diff --git a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/order.json b/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/order.json
deleted file mode 100644
index 78adc63d..00000000
--- a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/kafka/order.json
+++ /dev/null
@@ -1,14 +0,0 @@
-{
- "KAFKA_BROKER-INSTALL": [
- "ZOOKEEPER_SERVER-INSTALL"
- ],
- "KAFKA_BROKER-START": [
- "ZOOKEEPER_SERVER-START"
- ],
- "KAFKA_BROKER-RESTART": [
- "ZOOKEEPER_SERVER-RESTART"
- ],
- "ZOOKEEPER_SERVER-STOP": [
- "KAFKA_BROKER-STOP"
- ]
-}
\ No newline at end of file
diff --git a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/zookeeper/configuration/zoo.cfg.xml b/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/zookeeper/configuration/zoo.cfg.xml
deleted file mode 100644
index 26139101..00000000
--- a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/zookeeper/configuration/zoo.cfg.xml
+++ /dev/null
@@ -1,128 +0,0 @@
-
-
-
-
-
- tickTime
- 3000
- Length of single Tick
-
- The length of a single tick in milliseconds, which is the basic time unit used by ZooKeeper
-
-
-
- initLimit
- 10
- Ticks to allow for sync at Init
- Ticks to allow for sync at Init.
-
-
- syncLimit
- 5
- Ticks to allow for sync at Runtime
- Ticks to allow for sync at Runtime.
-
-
- clientPort
- 2181
- Port for running ZK Server
- Port for running ZK Server.
-
-
- dataDir
- /hadoop/zookeeper
- ZooKeeper directory
- Data directory for ZooKeeper.
-
-
- autopurge.snapRetainCount
- 30
-
- ZooKeeper purge feature retains the autopurge.
- snapRetainCount most recent snapshots and the corresponding transaction logs in the dataDir and dataLogDir
- respectively and deletes the rest.
-
-
-
- autopurge.purgeInterval
- 24
-
- The time interval in hours for which the purge task has to be triggered.
- Set to a positive integer (1 and above) to enable the auto purging.
-
-
-
- 4lw.commands.whitelist
- ruok
-
- A list of comma separated Four Letter Words commands that user wants to use.
- A valid Four Letter Words command must be put in this list else ZooKeeper server will not enable the
- command.
- By default the whitelist only contains "srvr" command which zkServer.sh uses.
- The rest of four letter word commands are disabled by default.
-
-
-
- admin.enableServer
- true
- Set to "false" to disable the AdminServer. By default the AdminServer is enabled.
-
-
- admin.serverPort
- 9393
- The port the embedded Jetty server listens on. Defaults to 8080.
-
-
- templateContent
- The port the embedded Jetty server listens on. Defaults to 8080.
-
-<#list model as key,value>
- <#if value??>
- <#if value?string == 'true'>
- ${key}=true
- <#elseif value?string == 'false'>
- ${key}=false
- <#else>
- ${key}=${value}
- #if>
- #if>
-#list>
-#compress>
-
-<#noparse>
-<#if zkServerStr?? >
-${zkServerStr}
-#if>
-
-<#if securityEnabled?? && securityEnabled >
-authProvider.1=org.apache.zookeeper.server.auth.SASLAuthenticationProvider
-jaasLoginRenew=3600000
-kerberos.removeHostFromPrincipal=true
-kerberos.removeRealmFromPrincipal=true
-#if>
-#noparse>
-]]>
-
-
- longtext
-
-
-
diff --git a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/zookeeper/configuration/zookeeper-env.xml b/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/zookeeper/configuration/zookeeper-env.xml
deleted file mode 100644
index b5d1209f..00000000
--- a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/zookeeper/configuration/zookeeper-env.xml
+++ /dev/null
@@ -1,58 +0,0 @@
-
-
-
-
-
- logDir
- /var/log/zookeeper
- ZooKeeper Log Dir
- ZooKeeper Log Dir
-
-
- pidDir
- /var/run/zookeeper
- ZooKeeper PID Dir
- ZooKeeper Pid Dir
-
-
-
- content
- zookeeper-env template
- This is the freemarker template for zookeeper-env.sh file
-
-export SERVER_JVMFLAGS="$SERVER_JVMFLAGS -Djava.security.auth.login.config=${zk_server_jaas_file!}"
-export CLIENT_JVMFLAGS="$CLIENT_JVMFLAGS -Djava.security.auth.login.config=${zk_client_jaas_file!} -Dzookeeper.sasl.client.username=${zk_principal_user!}"
-#if>
-]]>
-
-
- longtext
-
-
-
diff --git a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/zookeeper/metainfo.xml b/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/zookeeper/metainfo.xml
deleted file mode 100644
index 0151b689..00000000
--- a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/zookeeper/metainfo.xml
+++ /dev/null
@@ -1,79 +0,0 @@
-
-
-
-
-
- zookeeper
- ZooKeeper
-
- Apache ZooKeeper is an effort to develop and maintain an open-source server which enables highly
- reliable distributed coordination.
-
- 3.6.4
- zookeeper
-
-
-
- zookeeper_server
- ZooKeeper Server
- master
- 1+
-
- org.apache.bigtop.manager.stack.nop.v1_0_0.zookeeper.ZookeeperServerScript
- java
- 1200
-
-
- ZooKeeper UI(Test)
- admin.serverPort
- 9393
- admin.serverPort
- 9393
-
-
-
-
- zookeeper_client
- ZooKeeper Client
- client
- 1+
-
- org.apache.bigtop.manager.stack.nop.v1_0_0.zookeeper.ZookeeperClientScript
- java
-
-
-
-
-
-
-
- centos7
- rocky8
-
-
- x86_64
-
-
- zookeeper_1_0_0
-
-
-
-
-
\ No newline at end of file
diff --git a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/zookeeper/order.json b/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/zookeeper/order.json
deleted file mode 100644
index a0d75ea4..00000000
--- a/bigtop-manager-server/src/main/resources/stacks/nop/1.0.0/services/zookeeper/order.json
+++ /dev/null
@@ -1,5 +0,0 @@
-{
- "KAFKA_BROKER-INSTALL": [
- "ZOOKEEPER_SERVER-INSTALL"
- ]
-}
\ No newline at end of file
diff --git a/bigtop-manager-stack/bigtop-manager-stack-core/src/main/java/org/apache/bigtop/manager/stack/core/executor/StackExecutor.java b/bigtop-manager-stack/bigtop-manager-stack-core/src/main/java/org/apache/bigtop/manager/stack/core/executor/StackExecutor.java
index af23aeef..7ef3a423 100644
--- a/bigtop-manager-stack/bigtop-manager-stack-core/src/main/java/org/apache/bigtop/manager/stack/core/executor/StackExecutor.java
+++ b/bigtop-manager-stack/bigtop-manager-stack-core/src/main/java/org/apache/bigtop/manager/stack/core/executor/StackExecutor.java
@@ -69,10 +69,6 @@ private static Script getCustomScript(String customCommand, List
-
-
- 4.0.0
-
- org.apache.bigtop
- bigtop-manager-stack
- ${revision}
-
-
- bigtop-manager-stack-nop
- ${project.artifactId}
- Bigtop Manager Stack NOP
-
-
-
- org.apache.bigtop
- bigtop-manager-stack-core
-
-
-
-
diff --git a/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/kafka/KafkaBrokerScript.java b/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/kafka/KafkaBrokerScript.java
deleted file mode 100644
index d13f7569..00000000
--- a/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/kafka/KafkaBrokerScript.java
+++ /dev/null
@@ -1,67 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * https://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied. See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-package org.apache.bigtop.manager.stack.nop.v1_0_0.kafka;
-
-import org.apache.bigtop.manager.common.shell.ShellResult;
-import org.apache.bigtop.manager.stack.core.param.Params;
-import org.apache.bigtop.manager.stack.core.spi.script.AbstractServerScript;
-import org.apache.bigtop.manager.stack.core.spi.script.Script;
-
-import com.google.auto.service.AutoService;
-import lombok.extern.slf4j.Slf4j;
-
-@Slf4j
-@AutoService(Script.class)
-public class KafkaBrokerScript extends AbstractServerScript {
-
- @Override
- public ShellResult install(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-
- @Override
- public ShellResult configure(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-
- @Override
- public ShellResult start(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-
- @Override
- public ShellResult stop(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-
- @Override
- public ShellResult status(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-
- public ShellResult test(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-}
diff --git a/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/kafka/KafkaParams.java b/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/kafka/KafkaParams.java
deleted file mode 100644
index b2494f61..00000000
--- a/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/kafka/KafkaParams.java
+++ /dev/null
@@ -1,32 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * https://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied. See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-package org.apache.bigtop.manager.stack.nop.v1_0_0.kafka;
-
-import org.apache.bigtop.manager.common.message.entity.payload.CommandPayload;
-import org.apache.bigtop.manager.stack.core.param.BaseParams;
-
-import lombok.Getter;
-
-@Getter
-public class KafkaParams extends BaseParams {
-
- public KafkaParams(CommandPayload commandPayload) {
- super(commandPayload);
- }
-}
diff --git a/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/zookeeper/ZookeeperClientScript.java b/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/zookeeper/ZookeeperClientScript.java
deleted file mode 100644
index 3244afc6..00000000
--- a/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/zookeeper/ZookeeperClientScript.java
+++ /dev/null
@@ -1,44 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * https://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied. See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-package org.apache.bigtop.manager.stack.nop.v1_0_0.zookeeper;
-
-import org.apache.bigtop.manager.common.shell.ShellResult;
-import org.apache.bigtop.manager.stack.core.param.Params;
-import org.apache.bigtop.manager.stack.core.spi.script.AbstractClientScript;
-import org.apache.bigtop.manager.stack.core.spi.script.Script;
-
-import com.google.auto.service.AutoService;
-import lombok.extern.slf4j.Slf4j;
-
-@Slf4j
-@AutoService(Script.class)
-public class ZookeeperClientScript extends AbstractClientScript {
-
- @Override
- public ShellResult install(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-
- @Override
- public ShellResult configure(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-}
diff --git a/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/zookeeper/ZookeeperParams.java b/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/zookeeper/ZookeeperParams.java
deleted file mode 100644
index 76687688..00000000
--- a/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/zookeeper/ZookeeperParams.java
+++ /dev/null
@@ -1,32 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * https://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied. See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-package org.apache.bigtop.manager.stack.nop.v1_0_0.zookeeper;
-
-import org.apache.bigtop.manager.common.message.entity.payload.CommandPayload;
-import org.apache.bigtop.manager.stack.core.param.BaseParams;
-
-import lombok.Getter;
-
-@Getter
-public class ZookeeperParams extends BaseParams {
-
- public ZookeeperParams(CommandPayload commandPayload) {
- super(commandPayload);
- }
-}
diff --git a/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/zookeeper/ZookeeperServerScript.java b/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/zookeeper/ZookeeperServerScript.java
deleted file mode 100644
index 9da8e2ce..00000000
--- a/bigtop-manager-stack/bigtop-manager-stack-nop/src/main/java/org/apache/bigtop/manager/stack/nop/v1_0_0/zookeeper/ZookeeperServerScript.java
+++ /dev/null
@@ -1,62 +0,0 @@
-/*
- * Licensed to the Apache Software Foundation (ASF) under one
- * or more contributor license agreements. See the NOTICE file
- * distributed with this work for additional information
- * regarding copyright ownership. The ASF licenses this file
- * to you under the Apache License, Version 2.0 (the
- * "License"); you may not use this file except in compliance
- * with the License. You may obtain a copy of the License at
- *
- * https://www.apache.org/licenses/LICENSE-2.0
- *
- * Unless required by applicable law or agreed to in writing,
- * software distributed under the License is distributed on an
- * "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY
- * KIND, either express or implied. See the License for the
- * specific language governing permissions and limitations
- * under the License.
- */
-package org.apache.bigtop.manager.stack.nop.v1_0_0.zookeeper;
-
-import org.apache.bigtop.manager.common.shell.ShellResult;
-import org.apache.bigtop.manager.stack.core.param.Params;
-import org.apache.bigtop.manager.stack.core.spi.script.AbstractServerScript;
-import org.apache.bigtop.manager.stack.core.spi.script.Script;
-
-import com.google.auto.service.AutoService;
-import lombok.extern.slf4j.Slf4j;
-
-@Slf4j
-@AutoService(Script.class)
-public class ZookeeperServerScript extends AbstractServerScript {
-
- @Override
- public ShellResult install(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-
- @Override
- public ShellResult configure(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-
- @Override
- public ShellResult start(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-
- @Override
- public ShellResult stop(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-
- @Override
- public ShellResult status(Params params) {
- log.info("Default to success on dev mode");
- return ShellResult.success();
- }
-}
diff --git a/bigtop-manager-stack/pom.xml b/bigtop-manager-stack/pom.xml
index bc759239..265065f6 100644
--- a/bigtop-manager-stack/pom.xml
+++ b/bigtop-manager-stack/pom.xml
@@ -35,7 +35,6 @@
bigtop-manager-stack-core
bigtop-manager-stack-bigtop
- bigtop-manager-stack-nop
diff --git a/pom.xml b/pom.xml
index 173c2efd..94fbcd8d 100644
--- a/pom.xml
+++ b/pom.xml
@@ -105,12 +105,6 @@
${project.version}
-
- org.apache.bigtop
- bigtop-manager-stack-nop
- ${project.version}
-
-
org.apache.bigtop
bigtop-manager-ui