今天自學完了大數據方面生態系統的一些核心技術,在這裡整理一下學過的知識算是複習,也方便初學者參考 ...
- 版本:CentOS-6.8-x86_64-minimal,hadoop2.6.4,jdk1.7.0
- 首先把jdk、hadoop壓縮包下載發送到CentOS下並解壓
下載發送不多贅述,解壓命令tar -zxvf 壓縮包
mv 原文件名 新文件名
(註意空格)
- 先配置jdk
- 進入jdk目錄 pwd 查看當前目錄複製備用/apps/jdk1.7.0_25
- 配置環境變數
vi ~/.bash_profile
JAVA_HOME=/apps/jdk1.7.0_25
PATH=$PATH:$HOME/bin:$JAVA_HOME/bin
source ~/.bash_profile
- 關閉防火牆,做好ssh免密登錄
- 關閉防火牆
- service iptables stop
- chkconfig iptables off
- ssh-keygen -t rsa
- ssh免密登錄
- master: cat /root/.ssh/id_rsa.pub >>/root/.ssh/authorized_keys
- master:scp /root/.ssh/authorized_keys @slave1:/root/.ssh/authorized_keys
- slave1:cat /root/.ssh/id_rsa.pub >>/root/.ssh/authorized_keys
- slave1:scp /root/.ssh/authorized_keys @slave2:/root/.ssh/authorized_keys
- slave2:cat /root/.ssh/id_rsa.pub >>/root/.ssh/authorized_keys
- slave2:scp /root/.ssh/authorized_keys @master:/root/.ssh/authorized_keys
- master:scp /root/.ssh/authorized_keys @slave1:/root/.ssh/authorized_keys
- slave2:scp /root/.ssh/authorized_keys @slave2:/root/.ssh/authorized_keys
- 關閉防火牆
- 再配置hadoop
- 環境變數vi ~/.bash_profile
HADOOP_HOME=/apps/hadoop-2.6.4
PATH=$PATH:$HOME/bin:$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbinsource ~/.bash_profile
- 配置運行環境
- core-site.xml
-
<property>
<name>fs.defaultFS</name>
<value>hdfs://master:9000</value>
</property>
-
- hdfs-site.xml
-
<property>
<name>dfs.replication</name>
<value>3</value>
</property><property>
<name>dfs.namenode.name.dir</name>
<value>/app/hadoop/dfs/name</value>
</property><property>
<name>dfs.datanode.data.dir</name>
<value>/app/hadoop/dfs/data</value>
</property><property>
<name>dfs.secondary.http.address</name>
<value>slave2:50090</value>
</property><property>
<name>dfs.namenode.checkpoint.dir</name>
<value>/app/hadoop/dfs/namesecondary</value>
</property>
-
- hadoop-env.sh
- export JAVA_HOME=/apps/jdk1.7.0_25
- yarn-site.xml
-
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property><property>
<name>yarn.resourcemanager.hostname</name>
<value>master</value>
</property>
mapreduce-site.xml
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property> - slaves
- 將原有內容替換為從節點主機名
-
- core-site.xml
- 將配置完成的jdk、hadoop發送到另外兩台主機上
- master:scp -r apps/ @slave1:/apps/
- master:scp -r apps/ @slave2:/apps/
- master:scp ~/.bash_profile @slave1:~/.bash_profile
- master:scp ~/.bash_profile @slave2:~/.bash_profile
- slave1: source ~/.bash_profile
- slave2: source ~/.bash_profile
hadoop集群配置完成
格式化namenode
hadoop namenode -format
啟動所有節點
start-all.sh