Cluster Deployment
Cluster Deployment
This section describes how to manually deploy an instance that includes 3 ConfigNodes and 3 DataNodes, commonly known as a 3C3D cluster.
Note
Before installation, ensure that the system is complete by referring to System configuration
It is recommended to prioritize using
hostname
for IP configuration during deployment, which can avoid the problem of modifying the host IP in the later stage and causing the database to fail to start. To set the host name, you need to configure /etc/hosts on the target server. For example, if the local IP is 192.168.1.3 and the host name is iotdb-1, you can use the following command to set the server's host name and configure thecn_internal_address
anddn_internal_address
of IoTDB using the host name.echo "192.168.1.3 iotdb-1" >> /etc/hosts
Some parameters cannot be modified after the first startup. Please refer to the "Parameter Configuration" section below for settings.
Whether in linux or windows, ensure that the IoTDB installation path does not contain Spaces and Chinese characters to avoid software exceptions.
Please note that when installing and deploying IoTDB (including activating and using software), it is necessary to use the same user for operations. You can:
- Using root user (recommended): Using root user can avoid issues such as permissions.
- Using a fixed non root user:
- Using the same user operation: Ensure that the same user is used for start, activation, stop, and other operations, and do not switch users.
- Avoid using sudo: Try to avoid using sudo commands as they execute commands with root privileges, which may cause confusion or security issues.
- It is recommended to deploy a monitoring panel, which can monitor important operational indicators and keep track of database operation status at any time. The monitoring panel can be obtained by contacting the business department,The steps for deploying a monitoring panel can refer to:Monitoring Panel Deployment
Preparation Steps
- Prepare the IoTDB database installation package: iotdb enterprise- {version}-bin.zip(The installation package can be obtained from:IoTDB-Package)
- Configure the operating system environment according to environmental requirements(The system environment configuration can be found in:Environment Requirements)
Installation Steps
Assuming there are three Linux servers now, the IP addresses and service roles are assigned as follows:
Node IP | Host Name | Service |
---|---|---|
192.168.1.3 | iotdb-1 | ConfigNode、DataNode |
192.168.1.4 | iotdb-2 | ConfigNode、DataNode |
192.168.1.5 | iotdb-3 | ConfigNode、DataNode |
Set Host Name
On three machines, configure the host names separately. To set the host names, configure /etc/hosts
on the target server. Use the following command:
echo "192.168.1.3 iotdb-1" >> /etc/hosts
echo "192.168.1.4 iotdb-2" >> /etc/hosts
echo "192.168.1.5 iotdb-3" >> /etc/hosts
Configuration
Unzip the installation package and enter the installation directory
unzip iotdb-enterprise-{version}-bin.zip
cd iotdb-enterprise-{version}-bin
Environment script configuration
./conf/confignode-env.sh
configurationConfiguration Description Default Recommended value Note MEMORY_SIZE The total amount of memory that IoTDB ConfigNode nodes can use - Can be filled in as needed, and the system will allocate memory based on the filled in values Restarting the service takes effect ./conf/datanode-env.sh
configurationConfiguration Description Default Recommended value Note MEMORY_SIZE The total amount of memory that IoTDB DataNode nodes can use - Can be filled in as needed, and the system will allocate memory based on the filled in values Restarting the service takes effect
General Configuration
Open the general configuration file ./conf/iotdb-common.properties
,The following parameters can be set according to the deployment method:
Configuration | Description | 192.168.1.3 | 192.168.1.4 | 192.168.1.5 |
---|---|---|---|---|
cluster_name | Cluster Name | defaultCluster | defaultCluster | defaultCluster |
schema_replication_factor | The number of metadata replicas, the number of DataNodes should not be less than this number | 3 | 3 | 3 |
data_replication_factor | The number of data replicas should not be less than this number of DataNodes | 2 | 2 | 2 |
ConfigNode Configuration
Open the ConfigNode configuration file ./conf/iotdb-confignode.properties
,Set the following parameters
Configuration | Description | Default | Recommended value | 192.168.1.3 | 192.168.1.4 | 192.168.1.5 | Note |
---|---|---|---|---|---|---|---|
cn_internal_address | The address used by ConfigNode for communication within the cluster | 127.0.0.1 | The IPV4 address or host name of the server where it is located, and it is recommended to use host name | iotdb-1 | iotdb-2 | iotdb-3 | Cannot be modified after initial startup |
cn_internal_port | The port used by ConfigNode for communication within the cluster | 10710 | 10710 | 10710 | 10710 | 10710 | Cannot be modified after initial startup |
cn_consensus_port | The port used for ConfigNode replica group consensus protocol communication | 10720 | 10720 | 10720 | 10720 | 10720 | Cannot be modified after initial startup |
cn_seed_config_node | The address of the ConfigNode that the node connects to when registering to join the cluster, cn_internal_address:cn_internal_port | 127.0.0.1:10710 | The first CongfigNode's cn_internal-address: cn_internal_port | iotdb-1:10710 | iotdb-1:10710 | iotdb-1:10710 | Cannot be modified after initial startup |
DataNode Configuration
Open DataNode Configuration File ./conf/iotdb-datanode.properties
,Set the following parameters:
Configuration | Description | Default | Recommended value | 192.168.1.3 | 192.168.1.4 | 192.168.1.5 | Note |
---|---|---|---|---|---|---|---|
dn_rpc_address | The address of the client RPC service | 127.0.0.1 | Recommend using the IPV4 address or hostname of the server where it is located | iotdb-1 | iotdb-2 | iotdb-3 | Restarting the service takes effect |
dn_rpc_port | The port of the client RPC service | 6667 | 6667 | 6667 | 6667 | 6667 | Restarting the service takes effect |
dn_internal_address | The address used by DataNode for communication within the cluster | 127.0.0.1 | The IPV4 address or host name of the server where it is located, and it is recommended to use host name | iotdb-1 | iotdb-2 | iotdb-3 | Cannot be modified after initial startup |
dn_internal_port | The port used by DataNode for communication within the cluster | 10730 | 10730 | 10730 | 10730 | 10730 | Cannot be modified after initial startup |
dn_mpp_data_exchange_port | The port used by DataNode to receive data streams | 10740 | 10740 | 10740 | 10740 | 10740 | Cannot be modified after initial startup |
dn_data_region_consensus_port | The port used by DataNode for data replica consensus protocol communication | 10750 | 10750 | 10750 | 10750 | 10750 | Cannot be modified after initial startup |
dn_schema_region_consensus_port | The port used by DataNode for metadata replica consensus protocol communication | 10760 | 10760 | 10760 | 10760 | 10760 | Cannot be modified after initial startup |
dn_seed_config_node | The addresses of the ConfigNode that the node connects to when registering to join the cluster, i.e. cn_internal-address: cn_internal_port | 127.0.0.1:10710 | The first CongfigNode's cn_internal-address: cn_internal_port | iotdb-1:10710 | iotdb-1:10710 | iotdb-1:10710 | Cannot be modified after initial startup |
❗️Attention: Editors such as VSCode Remote do not have automatic configuration saving function. Please ensure that the modified files are saved persistently, otherwise the configuration items will not take effect
Start ConfigNode
Start the first confignode of IoTDB-1 first, ensuring that the seed confignode node starts first, and then start the second and third confignode nodes in sequence
cd sbin
./start-confignode.sh -d #"- d" parameter will start in the background
If the startup fails, please refer to Common Questions.
Activate Database
Method 1: Activate file copy activation
- After starting three confignode nodes in sequence, copy the
activation
folder of each machine and thesystem_info
file of each machine to the Timecho staff; - The staff will return the license files for each ConfigNode node, where 3 license files will be returned;
- Put the three license files into the
activation
folder of the corresponding ConfigNode node;
Method 2: Activate Script Activation
Obtain the machine codes of three machines in sequence, enter the
sbin
directory of the installation directory, and execute the activation scriptstart activate.sh
:cd sbin ./start-activate.sh
The following information is displayed, where the machine code of one machine is displayed:
Please copy the system_info's content and send it to Timecho: Y17hFA0xRCE1TmkVxILuxxxxxxxxxxxxxxxxxxxxxxxxxxxxW5P52KCccFMVeHTc= Please enter license:
The other two nodes execute the activation script
start activate.sh
in sequence, and then copy the machine codes of the three machines obtained to the Timecho staffThe staff will return 3 activation codes, which normally correspond to the order of the provided 3 machine codes. Please paste each activation code into the previous command line prompt
Please enter license:
, as shown below:Please enter license: Jw+MmF+Atxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxxx5bAOXNeob5l+HO5fEMgzrW8OJPh26Vl6ljKUpCvpTiw== License has been stored to sbin/../activation/license Import completed. Please start cluster and excute 'show cluster' to verify activation status
Start DataNode
Enter the sbin
directory of iotdb and start three datanode nodes in sequence:
cd sbin
./start-datanode.sh -d #"- d" parameter will start in the background
Verify Deployment
Can be executed directly Cli startup script in ./sbin
directory:
./start-cli.sh -h ip(local IP or domain name) -p port(6667)
After successful startup, the following interface will appear displaying successful installation of IOTDB.
After the installation success interface appears, continue to check if the activation is successful and use the show cluster
command.
When you see the display of Activated
on the far right, it indicates successful activation.
The appearance of
ACTIVATED (W)
indicates passive activation, which means that this Configurable Node does not have a license file (or has not issued the latest license file with a timestamp), and its activation depends on other Activated Configurable Nodes in the cluster. At this point, it is recommended to check if the license file has been placed in the license folder. If not, please place the license file. If a license file already exists, it may be due to inconsistency between the license file of this node and the information of other nodes. Please contact Timecho staff to reapply.
Node Maintenance Steps
ConfigNode Node Maintenance
ConfigNode node maintenance is divided into two types of operations: adding and removing ConfigNodes, with two common use cases:
- Cluster expansion: For example, when there is only one ConfigNode in the cluster, and you want to increase the high availability of ConfigNode nodes, you can add two ConfigNodes, making a total of three ConfigNodes in the cluster.
- Cluster failure recovery: When the machine where a ConfigNode is located fails, making the ConfigNode unable to run normally, you can remove this ConfigNode and then add a new ConfigNode to the cluster.
❗️Note, after completing ConfigNode node maintenance, you need to ensure that there are 1 or 3 ConfigNodes running normally in the cluster. Two ConfigNodes do not have high availability, and more than three ConfigNodes will lead to performance loss.
Adding ConfigNode Nodes
Script command:
# Linux / MacOS
# First switch to the IoTDB root directory
sbin/start-confignode.sh
# Windows
# First switch to the IoTDB root directory
sbin/start-confignode.bat
Parameter introduction:
Parameter | Description | Is it required |
---|---|---|
-v | Show version information | No |
-f | Run the script in the foreground, do not put it in the background | No |
-d | Start in daemon mode, i.e. run in the background | No |
-p | Specify a file to store the process ID for process management | No |
-c | Specify the path to the configuration file folder, the script will load the configuration file from here | No |
-g | Print detailed garbage collection (GC) information | No |
-H | Specify the path of the Java heap dump file, used when JVM memory overflows | No |
-E | Specify the path of the JVM error log file | No |
-D | Define system properties, in the format key=value | No |
-X | Pass -XX parameters directly to the JVM | No |
-h | Help instruction | No |
Removing ConfigNode Nodes
First connect to the cluster through the CLI and confirm the internal address and port number of the ConfigNode you want to remove by using show confignodes
:
IoTDB> show confignodes
+------+-------+---------------+------------+--------+
|NodeID| Status|InternalAddress|InternalPort| Role|
+------+-------+---------------+------------+--------+
| 0|Running| 127.0.0.1| 10710| Leader|
| 1|Running| 127.0.0.1| 10711|Follower|
| 2|Running| 127.0.0.1| 10712|Follower|
+------+-------+---------------+------------+--------+
Total line number = 3
It costs 0.030s
Then use the script to remove the DataNode. Script command:
# Linux / MacOS
sbin/remove-confignode.sh [confignode_id]
or
./sbin/remove-confignode.sh [cn_internal_address:cn_internal_port]
#Windows
sbin/remove-confignode.bat [confignode_id]
or
./sbin/remove-confignode.bat [cn_internal_address:cn_internal_port]
DataNode Node Maintenance
There are two common scenarios for DataNode node maintenance:
- Cluster expansion: For the purpose of expanding cluster capabilities, add new DataNodes to the cluster
- Cluster failure recovery: When a machine where a DataNode is located fails, making the DataNode unable to run normally, you can remove this DataNode and add a new DataNode to the cluster
❗️Note, in order for the cluster to work normally, during the process of DataNode node maintenance and after the maintenance is completed, the total number of DataNodes running normally should not be less than the number of data replicas (usually 2), nor less than the number of metadata replicas (usually 3).
Adding DataNode Nodes
Script command:
# Linux / MacOS
# First switch to the IoTDB root directory
sbin/start-datanode.sh
# Windows
# First switch to the IoTDB root directory
sbin/start-datanode.bat
Parameter introduction:
Abbreviation | Description | Is it required |
---|---|---|
-v | Show version information | No |
-f | Run the script in the foreground, do not put it in the background | No |
-d | Start in daemon mode, i.e. run in the background | No |
-p | Specify a file to store the process ID for process management | No |
-c | Specify the path to the configuration file folder, the script will load the configuration file from here | No |
-g | Print detailed garbage collection (GC) information | No |
-H | Specify the path of the Java heap dump file, used when JVM memory overflows | No |
-E | Specify the path of the JVM error log file | No |
-D | Define system properties, in the format key=value | No |
-X | Pass -XX parameters directly to the JVM | No |
-h | Help instruction | No |
Note: After adding a DataNode, as new writes arrive (and old data expires, if TTL is set), the cluster load will gradually balance towards the new DataNode, eventually achieving a balance of storage and computation resources on all nodes.
Removing DataNode Nodes
First connect to the cluster through the CLI and confirm the RPC address and port number of the DataNode you want to remove with show datanodes
:
IoTDB> show datanodes
+------+-------+----------+-------+-------------+---------------+
|NodeID| Status|RpcAddress|RpcPort|DataRegionNum|SchemaRegionNum|
+------+-------+----------+-------+-------------+---------------+
| 1|Running| 0.0.0.0| 6667| 0| 0|
| 2|Running| 0.0.0.0| 6668| 1| 1|
| 3|Running| 0.0.0.0| 6669| 1| 0|
+------+-------+----------+-------+-------------+---------------+
Total line number = 3
It costs 0.110s
Then use the script to remove the DataNode. Script command:
# Linux / MacOS
sbin/remove-datanode.sh [dn_rpc_address:dn_rpc_port]
#Windows
sbin/remove-datanode.bat [dn_rpc_address:dn_rpc_port]
Common Questions
Multiple prompts indicating activation failure during deployment process
- Use the
ls -al
command: Use thels -al
command to check if the owner information of the installation package root directory is the current user. - Check activation directory: Check all files in the
./activation
directory and whether the owner information is the current user.
- Use the
Confignode failed to start
Step 1: Please check the startup log to see if any parameters that cannot be changed after the first startup have been modified.
Step 2: Please check the startup log for any other abnormalities. If there are any abnormal phenomena in the log, please contact Timecho Technical Support personnel for consultation on solutions.
Step 3: If it is the first deployment or data can be deleted, you can also clean up the environment according to the following steps, redeploy, and restart.
Clean up the environment:
Execute the following on all nodes:
- Terminate all ConfigNode Node and DataNode processes.
# 1. Stop the ConfigNode and DataNode services sbin/stop-standalone.sh # 2. Check for any remaining processes jps # Or ps -ef|gerp iotdb # 3. If there are any remaining processes, manually kill the kill -9 <pid> # If you are sure there is only one iotdb on the machine, you can use the following command to clean up residual processes ps -ef|grep iotdb|grep -v grep|tr -s ' ' ' ' |cut -d ' ' -f2|xargs kill -9
- Delete the data and logs directories.
Explanation: Deleting the data directory is necessary, deleting the logs directory is for clean logs and is not mandatory.
cd /data/iotdb rm -rf data logs