preface
Recently, in the communication with fans, it was said that the white whoring server has not been used yet. Here I provide a way to use it:
She inspired me by the concept of pseudo distributed Hadoop deployment.
Because my private work often needs to use Hadoop cluster, local startup has been criticized for slow startup speed, troublesome operation and memory occupation,
In view of this, why not deploy the cloud cluster and choose the pseudo distributed deployment method of Hadoop 3. X.
1. White whoring server
I previously sent a guide to white whoring server:
Alibaba cloud double eleven server registration process
But now the event is over. Of course, if you are a student, it is still very cheap to buy a server. It only costs 9 yuan / month,
Alibaba cloud developer growth plan
2. Server selection and configuration
The lightweight server is selected here. The system image and application image do not need to be changed. Just keep the default value (WordPress, CentOS 7.3)
You need to set the root permission and password here
After setting successfully, ssh is built through local terminal (MAC) or cmd (Windows)
ssh root@****
Then enter the password of the root permission set previously (Note: the password here will not be displayed)
If the above situation occurs, you need to clean up the previous key
ssh-keygen -R XX.XX.XX.XX
Then ssh again, and then yes
OK, here we go to the Alibaba cloud server
3. Let's start configuring the java environment
First, download the java jdk
wget https://download.java.net/openjdk/jdk8u41/ri/openjdk-8u41-b04-linux-x64-14_jan_2020.tar.gz
Then unzip
tar -zxvf openjdk-8u41-b04-linux-x64-14_jan_2020.tar.gz
Move location and configure java path
mv java-se-8u41-ri/ /usr/java8 echo 'export JAVA_HOME=/usr/java8' >> /etc/profile echo 'export PATH=$PATH:$JAVA_HOME/bin' >> /etc/profile source /etc/profile
Check whether the installation is successful
java -version
This is an ideal situation. If the installation is successful, the following results will appear
4. We install Hadoop
# Download Hadoop with Tsinghua source wget https://mirrors.tuna.tsinghua.edu.cn/apache/hadoop/common/hadoop-3.2.2/hadoop-3.2.2.tar.gz here is the image source of Tsinghua University. Domestic partners download faster
Unzip as usual
tar -zxvf hadoop-3.2.2.tar.gz -C /opt/ mv /opt/hadoop-3.2.2 /opt/hadoop
Configuration address
echo 'export HADOOP_HOME=/opt/hadoop/' >> /etc/profile echo 'export PATH=$PATH:$HADOOP_HOME/bin' >> /etc/profile echo 'export PATH=$PATH:$HADOOP_HOME/sbin' >> /etc/profile source /etc/profile
Configure yarn and hadoop
echo "export JAVA_HOME=/usr/java8" >> /opt/hadoop/etc/hadoop/yarn-env.sh echo "export JAVA_HOME=/usr/java8" >> /opt/hadoop/etc/hadoop/hadoop-env.sh
View Hadoop installation
hadoop version
If the above situation occurs, the installation is successful
5. Next, you need to use vim to operate core site and HDFS site
vim /opt/hadoop/etc/hadoop/core-site.xml
Enter vim environment
Press i (insert) to modify
Move the cursor between configuration and copy the following information
<property> <name>hadoop.tmp.dir</name> <value>file:/opt/hadoop/tmp</value> <description>location to store temporary files</description> </property> <property> <name>fs.defaultFS</name> <value>hdfs://localhost:9000</value> </property>
Then press esc to stop the modification, and then press ": wq" (actually no "") to exit vim modification
Similarly, operate HDFS site
vim /opt/hadoop/etc/hadoop/hdfs-site.xml
<property> <name>dfs.replication</name> <value>1</value> </property> <property> <name>dfs.namenode.name.dir</name> <value>file:/opt/hadoop/tmp/dfs/name</value> </property> <property> <name>dfs.datanode.data.dir</name> <value>file:/opt/hadoop/tmp/dfs/data</value> </property>
Configure the master and slave connections, run the following commands, and press enter until the following figure appears
ssh-keygen -t rsa
Run the following code
cd .ssh cat id_rsa.pub >> authorized_keys
Start Hadoop
hadoop namenode -format start-dfs.sh start-yarn.sh
Stepping pit:
ERROR: but there is no YARN_NODEMANAGER_USER defined. Aborting operation
There's a small hole here,
Solution:
https://blog.csdn.net/ystyaoshengting/article/details/103026872
Check whether the configuration is successful
jps
Success diagram
6. Note: next, you need to open your alicloud firewall port before you can access it on the browser, otherwise you can't access it anyway
Finally, you can enter XX.XX.XX.XX * *: 9870 or * * XX. XX. XX: 8088 in the browser to access your Hadoop
The renderings are as follows
as well as
Postscript
📢 Blog home page: https://manor.blog.csdn.net
📢 Welcome to praise 👍 Collection ⭐ Leaving a message. 📝 Please correct any errors!
📢 This article was originally written by manor and started on CSDN blog 🙉
About what problems are encountered in the deployment of Hadoop cluster, please pay attention to the official account number.