大数据学习环境搭建(CentOS6.9+Hadoop2.7.3+Hive1.2.1+Hbase1.3.1+Spark2.1.1)详解大数据
2023-06-13 09:20:28 时间
以前搭建过一套,带Federation,至少需4台机器,过于复杂,笔记本也吃不消
。现为了学习Spark2.0版本,决定去掉
Federation,
简化学习环境,不过还是完全分布式
[[email protected] ~]# cp /root/zookeeper-3.4.9/conf/zoo_sample.cfg /root/zookeeper-3.4.9/conf/zoo.cfg
[[email protected] ~]# vi /root/zookeeper-3.4.9/conf/log4j.properties
export PATH=.:$PATH:$JAVA_HOME/bin:$SCALA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:/root:$HIVE_HOME/bin:$HBASE_HOME/bin:$SPARK_HOME
export CLASSPATH=.:$JAVA_HOME/jre/lib/rt.jar:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar
value org.apache.hadoop.hdfs.server.namenode.ha.ConfiguredFailoverProxyProvider /value
libhadoop.so.1.0.0: ELF 64-bit LSB shared object, x86-64, version 1 (SYSV), dynamically linked, not stripped
[[email protected] ~]# /root/hadoop-2.7.3/sbin/hadoop-daemon.sh start journalnode
[[email protected] ~]# /root/hadoop-2.7.3/sbin/hadoop-daemon.sh start journalnode
[[email protected] ~]# /root/hadoop-2.7.3/sbin/hadoop-daemon.sh start namenode
[[email protected] ~]# /root/hadoop-2.7.3/bin/hdfs namenode -bootstrapStandby
[[email protected] ~]# /root/hadoop-2.7.3/sbin/hadoop-daemon.sh start namenode
[[email protected] ~]# /root/hadoop-2.7.3/sbin/hadoop-daemon.sh start datanode
[[email protected] ~]# /root/hadoop-2.7.3/sbin/hadoop-daemon.sh start datanode
[[email protected] ~]# /root/hadoop-2.7.3/sbin/hadoop-daemon.sh start datanode
[[email protected] ~]# /root/hadoop-2.7.3/sbin/yarn-daemon.sh start resourcemanager
[[email protected] ~]# /root/hadoop-2.7.3/sbin/yarn-daemon.sh start resourcemanager
[[email protected] ~]# /root/hadoop-2.7.3/sbin/yarn-daemon.sh start nodemanager
[[email protected] ~]# /root/hadoop-2.7.3/sbin/yarn-daemon.sh start nodemanager
[[email protected] ~]# /root/hadoop-2.7.3/sbin/yarn-daemon.sh start nodemanager
[[email protected] ~]# mysql -h localhost -u root -p
Enter password: AAAaaa111mysql GRANT ALL PRIVILEGES ON *.* TO root @ % IDENTIFIED BY AAAaaa111 WITH GRANT OPTION; mysql flush privileges;
由于官方提供的spark-2.1.1-bin-hadoop2.7.tgz包中集成的Hive是1.2.1,所以Hive版本选择1.2.1
[[email protected] ~]# cp /root/apache-hive-1.2.1-bin/conf/hive-env.sh.template /root/apache-hive-1.2.1-bin/conf/hive-env.sh
[[email protected] ~]# cp /root/apache-hive-1.2.1-bin/conf/hive-log4j.properties.template /root/apache-hive-1.2.1-bin/conf/hive-log4j.properties
value jdbc:mysql://node1:3306/hive?createDatabaseIfNotExist=true amp;characterEncoding=UTF-8 /value
/root/apache-hive-1.2.1-bin/bin/hive service metastore /dev/null 2 1
[[email protected] ~]# ln -s /etc/init.d/hive-metastore /etc/rc.d/rc3.d/S65hive-metastore
[[email protected] ~]# cp /root/hadoop-2.7.3/etc/hadoop/hdfs-site.xml /root/hadoop-2.7.3/etc/hadoop/core-site.xml /root/hbase-1.3.1/conf/
#export HBASE_REGIONSERVER_OPTS= $HBASE_REGIONSERVER_OPTS -XX:PermSize=128m -XX:MaxPermSize=128m
[[email protected] ~]# cp /root/spark-2.1.1-bin-hadoop2.7/conf/spark-env.sh.template /root/spark-2.1.1-bin-hadoop2.7/conf/spark-env.sh
[[email protected] ~]# vi /root/spark-2.1.1-bin-hadoop2.7/conf/spark-env.sh
export SPARK_DAEMON_JAVA_OPTS= -Dspark.deploy.recoveryMode=ZOOKEEPER -Dspark.deploy.zookeeper.url=node1:2181,node2:2181,node3:2181 -Dspark.deploy.zookeeper.dir=/spark
[[email protected] ~]# cp /root/spark-2.1.1-bin-hadoop2.7/conf/slaves.template /root/spark-2.1.1-bin-hadoop2.7/conf/slaves
ssh [email protected] export BASH_ENV=/etc/profile;/root/zookeeper-3.4.9/bin/zkServer.sh start
ssh root@node3 export BASH_ENV=/etc/profile;/root/zookeeper-3.4.9/bin/zkServer.sh start
#ssh root@node2 export BASH_ENV=/etc/profile;/root/hadoop-2.7.3/sbin/yarn-daemon.sh start resourcemanager
ssh root@node2 export BASH_ENV=/etc/profile;/root/hadoop-2.7.3/sbin/hadoop-daemon.sh start zkfc
ssh root@node3 export BASH_ENV=/etc/profile;/root/hadoop-2.7.3/sbin/hadoop-daemon.sh start zkfc
echo Y | ssh root@node1 export BASH_ENV=/etc/profile;/root/hadoop-2.7.3/bin/yarn rmadmin -transitionToActive forcemanual rm1
#ssh root@node2 export BASH_ENV=/etc/profile;/root/hbase-1.3.1/bin/hbase-daemon.sh start master
#ssh root@node2 export BASH_ENV=/etc/profile;/root/spark-2.1.1-bin-hadoop2.7/sbin/start-master.sh
#ssh root@node2 export BASH_ENV=/etc/profile;/root/hadoop-2.7.3/sbin/yarn-daemon.sh stop resourcemanager
ssh root@node2 export BASH_ENV=/etc/profile;/root/hadoop-2.7.3/sbin/hadoop-daemon.sh stop zkfc
ssh root@node2 export BASH_ENV=/etc/profile;/root/zookeeper-3.4.9/bin/zkServer.sh stop
ssh root@node3 export BASH_ENV=/etc/profile;/root/zookeeper-3.4.9/bin/zkServer.sh stop
相关文章
- 学习大数据需要掌握哪些Java技术
- (数据科学学习手札144)使用管道操作符高效书写Python代码
- 默克与Saama合作构建机器学习驱动的临床数据层,加强临床开发能力
- 用Python机器学习模型预测世界杯结果靠谱吗?附代码数据集
- 画【Python折线图】的一百个学习报告(三、自动生成单一数据折线图)
- Mybatis配置之别名(typeAliases)优化、设置(settings)优化、映射器(mappers)优化以及生命周期和作用域的学习和理解
- MySQL学习之数据更新操作详解
- 机器学习之线性回归的最小二乘法求解详解大数据
- Spark学习笔记4:数据读取与保存详解大数据
- Spark学习笔记6:Spark调优与调试详解大数据
- 区块链学习(7) 共识详解大数据
- 区块链学习(3)交易(一)详解大数据
- HBase学习之路 (六)过滤器详解大数据
- ZooKeeper学习之路 (六)ZooKeeper API的简单使用(二)级联删除与创建详解大数据
- ZooKeeper学习之路 (五)ZooKeeper API的简单使用 增删改查详解大数据
- redis基础学习二详解大数据
- 20本机器学习与数据科学必读书籍详解大数据
- TensorFlow学习笔记(四)图像识别与卷积神经网络详解大数据
- 从sklearn.preprocessing, sklearn.feature_selection学习特征工程之预处理详解大数据
- 机器学习之特征工程(二)详解大数据
- HBase-1.2.1之Get操作源码学习详解大数据
- HBase深入学习(1)详解大数据
- 深度 | 做到这些值4亿美元:看 Nervana 如何利用深度学习实现数据革命
- 从零开始学习大数据 Linux基础,打造数据时代的核心竞争力(大数据linux基础)
- 学习SQL Server语法,开启数据技术之旅(sqlserver语法)
- Redis面试学习如何利用AOF记录数据(redis面试题aof)
- MongoDB学习笔记(二)通过samus驱动实现基本数据操作
- Python学习笔记_数据排序方法