Hadoop 2.2.0部署安装(笔记,单机安装)详解大数据
具体配置步骤:
◎ 在root根目录下创建.ssh目录 (必须root用户登录)
cd /root mkdir .ssh
chmod 700 .ssh cd .ssh
◎ 创建密码为空的 RSA 密钥对:
ssh-keygen -t rsa -P
◎ 在提示的对称密钥名称中输入 id_rsa将公钥添加至 authorized_keys 中:
cat id_rsa.pub authorized_keys
chmod 644 authorized_keys # 重要
◎ 编辑 sshd 配置文件 /etc/ssh/sshd_config ,把 #AuthorizedKeysFile .ssh/authorized_keys 前面的注释取消掉。
◎ 重启 sshd 服务:
service sshd restart
◎ 测试 SSH 连接。连接时会提示是否连接,按回车后会将此公钥加入至 knows_hosts 中:
ssh localhost# 输入用户名密码
Hadoop 2.2.0部署安装
具体步骤如下:
◎ 下载文件。
◎ 解压hadoop 配置环境。
#在root根目录下创建hadoop文件夹
mkdir hadoop;
cd hadoop;
#将hadoop 2.2.0 安装文件放置到hadoop目录文件夹下。
#解压hadoop 2.2.0 文件
tar -zxvf hadoop-2.2.0.tar.gz
#进入hadoop -2.2.0 文件夹
cd hadoop-2.2.0
#进入hadoop配置文件夹
cd etc/hadoop
#修改core-site.xml
vi core-site.xml 添加以下信息(hadoop.tmp.dir、fs.default.name):
value /root/hadoop/hdfs/namenode /value description Determineswhere on the local filesystem the DFS name node should store the name table. Ifthis is a comma-delimited list of directories then the name table is replicatedin all of the directories, for redundancy. /description final true /final /property property name dfs.datanode.data.dir /name value /root/hadoop/hdfs/datanode /value description Determineswhere on the local filesystem an DFS data node should store its blocks. If thisis a comma-delimited list of directories, then data will be stored in all nameddirectories, typically on different devices.Directories that do not exist areignored. /description final true /final /property property !-- 副本个数-- name dfs.replication /name value 1 /value /property property name dfs.permissions /name value false /value /property /configuration
#修改mapred-site.xml
添加 dfs.namenode.name.dir、dfs.datanode.data.dir、dfs.replication、dfs.permissions等参数信息
value localhost:54311 /value description The host and port that the MapReduce job tracker runs at. If "local", thenjobs are run in-process as a single map and reduce task. /description /property property name mapred.map.tasks /name value 10 /value description As a rule of thumb, use 10x the number of slaves(i.e., number of tasktrackers). /description /property property name mapred.reduce.tasks /name value 2 /value description As a rule of thumb, use 2x the number of slaveprocessors (i.e., number of tasktrackers). /description /property /configuration
◎ 设置java环境(接上述步骤)。
#修改hadoop-env.sh 设置java路径参数,export JAVA_HOME=/usr/local/jdk1.7。
export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_NAMENODE_OPTS" export HADOOP_DATANODE_OPTS="-Dhadoop.security.logger=ERROR,RFAS $HADOOP_DATANODE_OPTS" export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_SECONDARYNAMENODE_OPTS" # The following applies to multiple commands (fs, dfs, fsck, distcp etc) export HADOOP_CLIENT_OPTS="-Xmx512m $HADOOP_CLIENT_OPTS" #HADOOP_JAVA_PLATFORM_OPTS="-XX:-UsePerfData $HADOOP_JAVA_PLATFORM_OPTS" # On secure datanodes, user to run the datanode as after dropping privileges export HADOOP_SECURE_DN_USER=${HADOOP_SECURE_DN_USER} # Where log files are stored. $HADOOP_HOME/logs by default. #export HADOOP_LOG_DIR=${HADOOP_LOG_DIR}/$USER # Where log files are stored in the secure data environment. export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER} # The directory where pid files are stored. /tmp by default. # NOTE: this should be set to a directory that can only be written to by # the user that will run the hadoop daemons. Otherwise there is the # potential for a symlink attack. export HADOOP_PID_DIR=${HADOOP_PID_DIR} export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR} # A string representing this instance of hadoop. $USER by default. export HADOOP_IDENT_STRING=$USER
◎ 设置hadoop环境变量[HADOOP_HOME]。
vi /etc/profile 输入 export HADOOP_HOME=/root/hadoop/hadoop-2.2.0
source /etc/profile 让环境变量生效。
测试hadoop环境变量是否生效:
echo $HADOOP_HOME
/root/hadoop/hadoop-2.2.0
◎ 进入hadoop安装目录,进入bin目录,格式化hdfs。
./hadoop namenode –format
◎ 启动hadoop ,进入hadoop安装目录,进入sbin目录。
./start-all.sh
◎ 验证安装,登录 http://localhost:50070/ 。
文章转载请注明出处:http://www.cnblogs.com/likehua/p/3825810.html
相关推荐:
sqoop安装参考:http://www.cnblogs.com/likehua/p/3825489.html
hive安装参考:http://www.cnblogs.com/likehua/p/3825479.html
原创文章,作者:ItWorker,如若转载,请注明出处:https://blog.ytso.com/9598.html
分布式文件系统,分布式数据库区块链并行处理(MPP)数据库,数据挖掘开源大数据平台数据中台数据分析数据开发数据治理数据湖数据采集相关文章
- python安装dlib库_pycharm安装dlib失败
- CentOS 7 常用软件安装汇总
- 【说站】php安装扩展的几种方法
- 如果还有问ARKIME不会部署安装,你就把这篇丢给他!
- 大数据必知必会:Hadoop(4)高可用集群安装
- Porn Data Anaylize — Hadoop安装
- 从零到壹:10元~Mapping神器STAR的安装及用
- Hadoop伪分布安装详解(一)大数据
- 高可用Hadoop平台-Ganglia安装部署详解大数据
- Oracle安装配置指南(oracle安装完配置)
- MySQL与Hadoop:改善数据处理效率(mysql与hadoop)
- 实用指南:集群 Linux 系统的快速安装步骤(集群linux系统安装)
- 下载安装MSSQL:一步一步实现(安装mssql下载)
- Linux下安装QQ:快速又安全!(linux下载qq)
- 苹果Mac系统下安装Oracle的指南(mac系统oracle)