Website of Ubuntu Desktop
Download Link
Hadoop 3.2.1
Download Link
wget https://downloads.apache.org/hadoop/common/hadoop-3.2.1/hadoop-3.2.1.tar.gz
sudo tar -xvf hadoop-3.2.1.tar.gz
Spark 3.0.1
Download Link w/ hadoop 3.2 prebuild
wget https://downloads.apache.org/spark/spark-3.0.1/spark-3.0.1-bin-hadoop3.2.tgz
sudo tar -xvf spark-3.0.1-bin-hadoop3.2.tgz
testing:
java --version
sudo apt install openjdk-8-jre-headless
sudo apt install openjdk-8-jdk-headless
java place:
update-alternatives --display java
Download:
sudo apt install ssh
sudo apt install pdsh
Generate key:
try ssh:
ssh localhost
cd /home/master/cluster
pcdm@master:~$ –> pcdm@master:~/cluster$
pcdm@master:~/cluster$ sudo mv hadoop-3.2.1 hadoop
pcdm@master:~/cluster$ sudo mv spark-3.0.1-bin-hadoop3.2 spark
這裡小心,如果有用adduser 記得要用adduser後的名字
pcdm@master:~/cluster$ sudo chown -R pcdm:master hadoop
pcdm@master:~/cluster$ sudo chown -R pcdm:master spark
sudo gedit ~/.bashrc
python:
sudo ln -s /usr/bin/python3 /usr/bin/python
Dont forget:
source ~/.bashrc
core-site.xml:
cd ~/cluster/hadoop/etc/hadoop/
sudo gedit core-site.xml
hdfs-site.xml:
datafile
setting
cd ~/cluster/hadoop/etc/hadoop/
sudo gedit hdfs-site.xml
mapred-site.xml
yarn-site.xml
Format the filesystem
hdfs namenode -format
Start dfs, yarn:
start-dfs.sh
start-yarn.sh
or start-all.sh
Stop dfs, yarn:
stop-dfs.sh
stop-yarn.sh
or stop-all.sh
Master:
start-master.sh
Slaves:
start-slaves.sh
first,
cp conf/log4j.properties.template conf/log4j.properties
then
change
log4j.rootCategory=INFO, console
to
log4j.rootCategory=WARN, console