Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -53,6 +53,7 @@ public enum ApiExceptionEnum {
HOST_UNABLE_TO_CONNECT(12004, LocaleKeys.HOST_UNABLE_TO_CONNECT),
HOST_UNABLE_TO_EXEC_COMMAND(12005, LocaleKeys.HOST_UNABLE_TO_EXEC_COMMAND),
HOST_HAS_COMPONENTS(12006, LocaleKeys.HOST_HAS_COMPONENTS),
AGENT_STILL_RUNNING(12007, LocaleKeys.HOST_AGENT_STILL_RUNNING),

// Stack Exceptions -- 13000 ~ 13999
STACK_NOT_FOUND(13000, LocaleKeys.STACK_NOT_FOUND),
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -55,6 +55,7 @@ public enum LocaleKeys {
HOST_UNABLE_TO_CONNECT("host.unable.to.connect"),
HOST_UNABLE_TO_EXEC_COMMAND("host.unable.to.exec.command"),
HOST_HAS_COMPONENTS("host.has.components"),
HOST_AGENT_STILL_RUNNING("host.agent.still.running"),

STACK_NOT_FOUND("stack.not.found"),

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -31,8 +31,6 @@ public class PropertyDTO implements Serializable {

private String value;

private String displayName;

private String desc;

private AttrsDTO attrs;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,8 +32,6 @@ public class PropertyReq {

private String value;

private String displayName;

private String desc;

private AttrsReq attrs;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -27,8 +27,6 @@ public class PropertyVO {

private String value;

private String displayName;

private String desc;

private AttrsVO attrs;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -198,10 +198,15 @@ public List<ComponentVO> components(Long id) {
@Override
public Boolean batchRemove(List<Long> ids) {
for (Long id : ids) {
HostPO hostPO = hostDao.findById(id);
if (componentDao.countByHostId(id) > 0) {
HostPO hostPO = hostDao.findById(id);
throw new ApiException(ApiExceptionEnum.HOST_HAS_COMPONENTS, hostPO.getHostname());
}

// Check if agent is stopped before removing host
if (!HealthyStatusEnum.UNHEALTHY.getCode().equals(hostPO.getStatus())) {
throw new ApiException(ApiExceptionEnum.AGENT_STILL_RUNNING, hostPO.getHostname());
}
}

return hostDao.deleteByIds(ids);
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -32,9 +32,6 @@ public class PropertyModel {

private String value;

@XmlElement(name = "display-name")
private String displayName;

@XmlElement(name = "description")
private String desc;

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,6 @@ public static List<PropertyDTO> loadConfig(String fileName) {

private static PropertyDTO getPropertyDTO(PropertyModel propertyModel) {
PropertyDTO propertyDTO = new PropertyDTO();
propertyDTO.setDisplayName(propertyModel.getDisplayName());
propertyDTO.setDesc(StringUtils.strip(propertyModel.getDesc()));
propertyDTO.setName(propertyModel.getName());
propertyDTO.setValue(StringUtils.strip(propertyModel.getValue()));
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -49,6 +49,7 @@ host.unable.to.resolve=Unable to resolve to host [{0}]
host.unable.to.connect=Unable to connect to host [{0}]
host.unable.to.exec.command=Unable to execute command on host [{0}]
host.has.components=Host [{0}] still has components, please remove them first
host.agent.still.running=Agent on host [{0}] is still running, please stop it first

stack.not.found=Stack not exist

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -49,6 +49,7 @@ host.unable.to.resolve=无法解析主机 [{0}]
host.unable.to.connect=无法连接到主机 [{0}]
host.unable.to.exec.command=无法在主机 [{0}] 上执行命令
host.has.components=主机 [{0}] 上仍有组件,请先移除
host.agent.still.running=主机 [{0}] 的 Agent 仍在运行,请先停止

stack.not.found=组件栈不存在

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -42,7 +42,6 @@
<!-- flink-conf.yaml -->
<property>
<name>content</name>
<display-name>flink-conf template</display-name>
<description>This is the freemarker template for flink-conf.xml file</description>
<value><![CDATA[
#==============================================================================
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,12 +22,10 @@
<property>
<name>flink_log_dir</name>
<value>/var/log/flink</value>
<display-name>Flink Log Dir Prefix</display-name>
<description>Log Directories for Flink.</description>
</property>
<property>
<name>flink_pid_dir</name>
<display-name>Flink PID directory</display-name>
<value>/var/run/flink</value>
</property>
</configuration>
Original file line number Diff line number Diff line change
Expand Up @@ -22,25 +22,21 @@
<property>
<name>hadoop_root_logger</name>
<value>INFO,RFA</value>
<display-name>Hadoop Root Logger</display-name>
<description>Hadoop Root Logger</description>
</property>
<property>
<name>hadoop_heapsize_max</name>
<value>1024</value>
<description>Hadoop maximum Java heap size</description>
<display-name>Hadoop maximum Java heap size</display-name>
</property>
<property>
<name>hadoop_heapsize_min</name>
<value>1024</value>
<description>Hadoop minimum Java heap size</description>
<display-name>Hadoop minimum Java heap size</display-name>
</property>
<!-- hadoop-env.sh -->
<property>
<name>content</name>
<display-name>hadoop-env template</display-name>
<description>This is the freemarker template for hadoop-env.sh file</description>
<value><![CDATA[
#
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -31,7 +31,6 @@
</property>
<property>
<name>content</name>
<display-name>hadoop.conf template</display-name>
<description>This is the freemarker template for hadoop file</description>
<value><![CDATA[
#
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,29 +23,24 @@
<name>hadoop_log_maxfilesize</name>
<value>256</value>
<description>The maximum size of backup file before the log is rotated</description>
<display-name>Hadoop Security Log: backup file size</display-name>
</property>
<property>
<name>hadoop_log_maxbackupindex</name>
<value>20</value>
<description>The number of backup files</description>
<display-name>Hadoop Security Log: # of backup files</display-name>
</property>
<property>
<name>hadoop_security_log_maxfilesize</name>
<value>256</value>
<description>The maximum size of backup file before the log is rotated</description>
<display-name>Hadoop Log: backup file size</display-name>
</property>
<property>
<name>hadoop_security_log_maxbackupindex</name>
<value>20</value>
<description>The number of backup files</description>
<display-name>Hadoop Log: # of backup files</display-name>
</property>
<property>
<name>content</name>
<display-name>hdfs-log4j template</display-name>
<description>Custom log4j.properties</description>
<value><![CDATA[
# Licensed to the Apache Software Foundation (ASF) under one
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -24,7 +24,6 @@
<name>dfs.namenode.name.dir</name>
<!-- cluster variant -->
<value>/hadoop/dfs/name</value>
<display-name>NameNode directories</display-name>
<description>Determines where on the local filesystem the DFS name node
should store the name table. If this is a comma-delimited list
of directories then the name table is replicated in all of the
Expand All @@ -35,12 +34,10 @@
<name>dfs.datanode.failed.volumes.tolerated</name>
<value>0</value>
<description>Number of failed disks a DataNode would tolerate before it stops offering service</description>
<display-name>DataNode failed disk tolerance</display-name>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>/hadoop/dfs/data</value>
<display-name>DataNode directories</display-name>
<description>Determines where on the local filesystem an DFS data node
should store its blocks. If this is a comma-delimited
list of directories, then data will be stored in all named
Expand All @@ -60,7 +57,6 @@
<property>
<name>dfs.namenode.checkpoint.dir</name>
<value>/hadoop/dfs/namesecondary</value>
<display-name>SecondaryNameNode Checkpoint directories</display-name>
<description>Determines where on the local filesystem the DFS secondary
name node should store the temporary images to merge.
If this is a comma-delimited list of directories then the image is
Expand All @@ -69,7 +65,6 @@
</property>
<property>
<name>dfs.namenode.checkpoint.edits.dir</name>
<display-name>NameNode Checkpoint Edits directory</display-name>
<value>${dfs.namenode.checkpoint.dir}</value>
<description>Determines where on the local filesystem the DFS secondary
name node should store the temporary edits to merge.
Expand All @@ -81,7 +76,6 @@
<property>
<name>dfs.namenode.checkpoint.period</name>
<value>3600</value>
<display-name>HDFS Maximum Checkpoint Delay</display-name>
<description>The number of seconds between two periodic checkpoints.</description>
</property>
<property>
Expand All @@ -105,7 +99,6 @@
<property>
<name>dfs.replication</name>
<value>3</value>
<display-name>Block replication</display-name>
<description>Default block replication.</description>
</property>
<property>
Expand All @@ -122,7 +115,6 @@
Values less than or equal to 0 mean not to start in safe mode.
Values greater than 1 will make safe mode permanent.
</description>
<display-name>Minimum replicated blocks %</display-name>
</property>
<property>
<name>dfs.datanode.balance.bandwidthPerSec</name>
Expand Down Expand Up @@ -189,7 +181,6 @@
<name>dfs.datanode.du.reserved</name>
<!-- cluster variant -->
<value>1073741824</value>
<display-name>Reserved space for HDFS</display-name>
<description>Reserved space in bytes per volume. Always leave this much space free for non dfs use.
</description>
</property>
Expand All @@ -203,7 +194,6 @@
<value>4096</value>
<description>Specifies the maximum number of threads to use for transferring data in and out of the datanode.
</description>
<display-name>DataNode max data transfer threads</display-name>
</property>
<!-- Permissions configuration -->
<property>
Expand All @@ -226,7 +216,6 @@
<name>dfs.namenode.handler.count</name>
<value>100</value>
<description>Added to grow Queue size so that more client connections are allowed</description>
<display-name>NameNode Server threads</display-name>
</property>
<property>
<name>dfs.block.access.token.enable</name>
Expand All @@ -239,7 +228,6 @@
<property>
<name>dfs.datanode.data.dir.perm</name>
<value>700</value>
<display-name>DataNode directories permission</display-name>
<description>The permissions that should be there on dfs.datanode.data.dir
directories. The datanode will not come up if the permissions are
different on existing dfs.datanode.data.dir directories. If the directories
Expand All @@ -249,7 +237,6 @@
<property>
<name>dfs.namenode.accesstime.precision</name>
<value>0</value>
<display-name>Access time precision</display-name>
<description>The access time for HDFS file is precise up to this value.
The default value is 1 hour. Setting a value of 0 disables
access times for HDFS.
Expand Down Expand Up @@ -312,15 +299,13 @@
</property>
<property>
<name>dfs.journalnode.edits.dir</name>
<display-name>JournalNode Edits directory</display-name>
<value>/hadoop/dfs/journalnode</value>
<description>The path where the JournalNode daemon will store its local state.</description>
</property>
<!-- HDFS Short-Circuit Local Reads -->
<property>
<name>dfs.client.read.shortcircuit</name>
<value>true</value>
<display-name>HDFS Short-circuit read</display-name>
<description>
This configuration parameter turns on short-circuit local reads.
</description>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,13 +21,11 @@
<configuration>
<property>
<name>jobhistory_heapsize</name>
<display-name>History Server heap size</display-name>
<value>1024</value>
<description>Value for JobHistoryServer heap_size variable in hadoop-env.sh</description>
</property>
<property>
<name>content</name>
<display-name>mapred-env template</display-name>
<description>This is the freemarker template for mapred-env.sh file</description>
<value><![CDATA[
export HADOOP_JOB_HISTORYSERVER_HEAPSIZE=${jobhistory_heapsize}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -23,7 +23,6 @@
<property>
<name>mapreduce.task.io.sort.mb</name>
<value>512</value>
<display-name>Sort Allocation Memory</display-name>
<description>
The total amount of buffer memory to use while sorting files, in megabytes.
By default, gives each merge stream 1MB, which should minimize seeks.
Expand Down Expand Up @@ -145,13 +144,11 @@
<name>mapreduce.map.memory.mb</name>
<value>512</value>
<description>Virtual memory for single Map task</description>
<display-name>Map Memory</display-name>
</property>
<property>
<name>mapreduce.reduce.memory.mb</name>
<value>512</value>
<description>Virtual memory for single Reduce task</description>
<display-name>Reduce Memory</display-name>
</property>
<property>
<name>mapreduce.shuffle.port</name>
Expand All @@ -164,13 +161,11 @@
</property>
<property>
<name>mapreduce.jobhistory.intermediate-done-dir</name>
<display-name>Mapreduce JobHistory Intermediate Done directory</display-name>
<value>/apps/mapred/history/tmp</value>
<description>Directory where history files are written by MapReduce jobs.</description>
</property>
<property>
<name>mapreduce.jobhistory.done-dir</name>
<display-name>Mapreduce JobHistory Done directory</display-name>
<value>/apps/mapred/history/done</value>
<description>Directory where history files are managed by the MR JobHistory Server.</description>
</property>
Expand All @@ -193,13 +188,11 @@
</property>
<property>
<name>yarn.app.mapreduce.am.staging-dir</name>
<display-name>YARN App Mapreduce AM Staging directory</display-name>
<value>/apps/mapred/staging</value>
<description>The staging dir used while submitting jobs.</description>
</property>
<property>
<name>yarn.app.mapreduce.am.resource.mb</name>
<display-name>AppMaster Memory</display-name>
<value>512</value>
<description>The amount of memory the MR AppMaster needs.</description>
</property>
Expand All @@ -219,7 +212,6 @@
of LD_LIBRARY_PATH in the map / reduce JVM env using the mapreduce.map.env and
mapreduce.reduce.env config settings.
</description>
<display-name>MR AppMaster Java Heap Size</display-name>
</property>
<property>
<name>yarn.app.mapreduce.am.admin-command-opts</name>
Expand All @@ -234,7 +226,6 @@
of LD_LIBRARY_PATH in the map / reduce JVM env using the mapreduce.map.env and
mapreduce.reduce.env config settings.
</description>
<display-name>MR AppMaster Java Heap Size</display-name>
</property>
<property>
<name>yarn.app.mapreduce.am.log.level</name>
Expand Down Expand Up @@ -268,13 +259,11 @@
</property>
<property>
<name>mapreduce.map.java.opts</name>
<display-name>MR Map Java Heap Size</display-name>
<value>-Xmx1024m</value>
<description>Larger heap-size for child jvms of maps.</description>
</property>
<property>
<name>mapreduce.reduce.java.opts</name>
<display-name>MR Reduce Java Heap Size</display-name>
<value>-Xmx1024m</value>
<description>Larger heap-size for child jvms of reduces.</description>
</property>
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -21,7 +21,6 @@
<configuration>
<property>
<name>content</name>
<display-name>hdfs workers template</display-name>
<description>This is the freemarker template for hdfs file</description>
<value><![CDATA[
#
Expand Down
Loading
Loading