zl程序教程

您现在的位置是:首页 >  其他

当前栏目

Hadoop 2.2.0部署安装(笔记,单机安装)详解大数据

安装hadoop部署笔记数据 详解 单机 2.2
2023-06-13 09:20:27 时间

具体配置步骤:

◎ 在root根目录下创建.ssh目录 (必须root用户登录)

cd /root   mkdir .ssh

chmod 700 .ssh   cd .ssh

◎ 创建密码为空的 RSA 密钥对:

ssh-keygen -t rsa -P 

◎ 在提示的对称密钥名称中输入 id_rsa将公钥添加至 authorized_keys 中:

cat id_rsa.pub   authorized_keys

chmod 644 authorized_keys # 重要

◎ 编辑 sshd 配置文件 /etc/ssh/sshd_config ,把 #AuthorizedKeysFile .ssh/authorized_keys 前面的注释取消掉。

◎ 重启 sshd 服务:

service sshd restart

◎ 测试 SSH 连接。连接时会提示是否连接,按回车后会将此公钥加入至 knows_hosts 中:

ssh localhost# 输入用户名密码 

 

 Hadoop 2.2.0部署安装

具体步骤如下:

◎ 下载文件。

◎ 解压hadoop 配置环境。

#在root根目录下创建hadoop文件夹

mkdir hadoop;

cd hadoop; 

#将hadoop 2.2.0 安装文件放置到hadoop目录文件夹下。

#解压hadoop 2.2.0 文件 

tar -zxvf hadoop-2.2.0.tar.gz

#进入hadoop -2.2.0 文件夹

cd hadoop-2.2.0

#进入hadoop配置文件夹

cd etc/hadoop

#修改core-site.xml 

vi core-site.xml 添加以下信息(hadoop.tmp.dir、fs.default.name):


value /root/hadoop/hdfs/namenode /value description Determineswhere on the local filesystem the DFS name node should store the name table. Ifthis is a comma-delimited list of directories then the name table is replicatedin all of the directories, for redundancy. /description final true /final /property property name dfs.datanode.data.dir /name value /root/hadoop/hdfs/datanode /value description Determineswhere on the local filesystem an DFS data node should store its blocks. If thisis a comma-delimited list of directories, then data will be stored in all nameddirectories, typically on different devices.Directories that do not exist areignored. /description final true /final /property property !-- 副本个数-- name dfs.replication /name value 1 /value /property property name dfs.permissions /name value false /value /property /configuration

 

 

#修改mapred-site.xml

添加 dfs.namenode.name.dir、dfs.datanode.data.dir、dfs.replication、dfs.permissions等参数信息

 


value localhost:54311 /value description The host and port that the MapReduce job tracker runs at. If "local", thenjobs are run in-process as a single map and reduce task. /description /property property name mapred.map.tasks /name value 10 /value description As a rule of thumb, use 10x the number of slaves(i.e., number of tasktrackers). /description /property property name mapred.reduce.tasks /name value 2 /value description As a rule of thumb, use 2x the number of slaveprocessors (i.e., number of tasktrackers). /description /property /configuration

 

 

◎ 设置java环境(接上述步骤)。

#修改hadoop-env.sh 设置java路径参数,export JAVA_HOME=/usr/local/jdk1.7。


export HADOOP_NAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_NAMENODE_OPTS" export HADOOP_DATANODE_OPTS="-Dhadoop.security.logger=ERROR,RFAS $HADOOP_DATANODE_OPTS" export HADOOP_SECONDARYNAMENODE_OPTS="-Dhadoop.security.logger=${HADOOP_SECURITY_LOGGER:-INFO,RFAS} -Dhdfs.audit.logger=${HDFS_AUDIT_LOGGER:-INFO,NullAppender} $HADOOP_SECONDARYNAMENODE_OPTS" # The following applies to multiple commands (fs, dfs, fsck, distcp etc) export HADOOP_CLIENT_OPTS="-Xmx512m $HADOOP_CLIENT_OPTS" #HADOOP_JAVA_PLATFORM_OPTS="-XX:-UsePerfData $HADOOP_JAVA_PLATFORM_OPTS" # On secure datanodes, user to run the datanode as after dropping privileges export HADOOP_SECURE_DN_USER=${HADOOP_SECURE_DN_USER} # Where log files are stored. $HADOOP_HOME/logs by default. #export HADOOP_LOG_DIR=${HADOOP_LOG_DIR}/$USER # Where log files are stored in the secure data environment. export HADOOP_SECURE_DN_LOG_DIR=${HADOOP_LOG_DIR}/${HADOOP_HDFS_USER} # The directory where pid files are stored. /tmp by default. # NOTE: this should be set to a directory that can only be written to by # the user that will run the hadoop daemons. Otherwise there is the # potential for a symlink attack. export HADOOP_PID_DIR=${HADOOP_PID_DIR} export HADOOP_SECURE_DN_PID_DIR=${HADOOP_PID_DIR} # A string representing this instance of hadoop. $USER by default. export HADOOP_IDENT_STRING=$USER

 

 

◎ 设置hadoop环境变量[HADOOP_HOME]。

vi /etc/profile 输入 export HADOOP_HOME=/root/hadoop/hadoop-2.2.0

source /etc/profile 让环境变量生效。

测试hadoop环境变量是否生效:

echo $HADOOP_HOME 

/root/hadoop/hadoop-2.2.0 

◎ 进入hadoop安装目录,进入bin目录,格式化hdfs。

./hadoop namenode –format 

◎ 启动hadoop ,进入hadoop安装目录,进入sbin目录。

./start-all.sh 

◎ 验证安装,登录 http://localhost:50070/ 。

 文章转载请注明出处:http://www.cnblogs.com/likehua/p/3825810.html 

相关推荐:

sqoop安装参考:http://www.cnblogs.com/likehua/p/3825489.html
hive安装参考:http://www.cnblogs.com/likehua/p/3825479.html

 

原创文章,作者:ItWorker,如若转载,请注明出处:https://blog.ytso.com/9598.html

分布式文件系统,分布式数据库区块链并行处理(MPP)数据库,数据挖掘开源大数据平台数据中台数据分析数据开发数据治理数据湖数据采集