安裝環境
主機 | ip | 作業系統 |
---|---|---|
master | 10.30.45.239 | centos7.4 |
slave1 | 10.31.155.33 | |
slave2 | 10.81.233.67 |
阿裡雲3台ecs搭建一個小型的hbase分布式資料存儲(這裡的ip都是内網)。一台作為主master兩台作為從slave。
安裝步驟
- 安裝vim編輯器
- 安裝ssh
- 配置ssh免密碼登陸
- 安裝jdk
- 安裝hadoop
- 安裝zookeeper
- 安裝時間同步服務 NTP
- 安裝hbase
準備工作
版本
centos7.4 核心為3.10
Hadoop 2.8.3
Hbase 1.2.6
Zookeeper 3.4.10
Sqoop 1.4.7
Phoenix 4.13.1
修改hostname
# vim /etc/hostname
輸入自定義主機名 ,三台都要改
修改hosts檔案
分别登陸三台主機,修改hosts
vim /etc/hosts
--添加以下資訊(三台都要改)
10.30.45.239 master
10.31.155.33 slave1
10.81.233.67 slave2
登陸master,下載下傳所有安裝包
--下載下傳jdk8u31
wget --no-check-certificate --no-cookies --header "Cookie: oraclelicense=accept-securebackup-cookie" http://download.oracle.com/otn-pub/java/jdk/8u131-b11/d54c1d3a095b4ff2b6607d096fa80163/jdk-8u131-linux-x64.tar.gz
-- 下載下傳 hadoop-2.8.3
wget http://mirrors.tuna.tsinghua.edu.cn/apache/hadoop/common/hadoop-2.8.3/hadoop-2.8.3.tar.gz
--下載下傳hbase 1.2.6
wget http://mirrors.hust.edu.cn/apache/hbase/1.2.6/hbase-1.2.6-bin.tar.gz
--下載下傳zookeeper 3.4.10
wget http://apache.claz.org/zookeeper/zookeeper-3.4.10/zookeeper-3.4.10.tar.gz
--下載下傳phoenix4.13.1
wget http://mirror.bit.edu.cn/apache/phoenix/apache-phoenix-4.13.1-HBase-1.2/bin/apache-phoenix-4.13.1-HBase-1.2-bin.tar.gz
--下載下傳sqoop1.4.7
wget http://mirror.cogentco.com/pub/apache/sqoop/1.4.7/sqoop-1.4.7.bin__hadoop-2.6.0.tar.gz
寫這篇文章的時候是2018年4月。适當選擇穩定的版本安裝。下載下傳官網如下:
- jdk : http://www.oracle.com/technetwork/java/javase/downloads/index.html
- hadoop: http://hadoop.apache.org/releases.html
- hbase: http://www.apache.org/dyn/closer.lua/hbase/
- zookeeper: http://zookeeper.apache.org/releases.html
- sqoop: http://sqoop.apache.org
- phoenix: https://phoenix.apache.org/
設定3台ecs之間免密登陸
友善後期之間檔案複制。
# # slave1
# ssh-keygen -t rsa
-- 不用輸入密碼,一路回車就行
# cp ~/.ssh/id_rsa.pub ~/.ssh/slave1_id_rsa.pub
# scp ~/.ssh/slave1_id_rsa.pub master:~/.ssh/
# # slave2
# ssh-keygen -t rsa
# cp ~/.ssh/id_rsa.pub ~/.ssh/slave2_id_rsa.pub
# scp ~/.ssh/slave2_id_rsa.pub master:~/.ssh/
# # master
ssh-keygen -t rsa
# cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
# cat ~/.ssh/slave1_id_rsa.pub >> ~/.ssh/authorized_keys
# cat ~/.ssh/slave2_id_rsa.pub >> ~/.ssh/authorized_kyes
-- 拷貝檔案至slave1及slave2
# scp ~/.ssh/authorized_keys slave1:~/.ssh
# scp ~/.ssh/authorized_keys slave2:~/.ssh
關閉防火牆及SELINUX(master、slave1、slave2均需修改)
-- 關閉防火牆
# systemctl stop firewalld.service
# systemctl disable firewalld.service
-- 關閉SELINUX
# vim /etc/selinux/config
-- 注釋掉
#SELINUX=enforcing
#SELINUXTYPE=targeted
-- 添加
SELINUX=disabled
設定時間同步
這裡省略,第一我懶得配,第二阿裡雲有自動同步
登陸master
-- 解壓jdk安裝包
# mkdir /usr/java
# tar -zxvf jdk-8u131-linux-x64.tar.gz -C /usr/java
-- 拷貝jdk至slave1及slave2中
# scp -r /usr/java slave1:/usr
# scp -r /usr/java slave2:/usr
-- 設定jdk環境變量(master、slave1、slave2均需修改)
# vim /etc/environment
JAVA_HOME=/usr/java/jdk1.8.0_131
JRE_HOME=/usr/java/jdk1.8.0_131/jre
# vim /etc/profile
export JAVA_HOME=/usr/java/jdk1.8.0_131
export JRE_HOME=${JAVA_HOME}/jre
export CLASSPATH=.:$JAVA_HOME/jre/lib/rt.jar:$JAVA_HOME/lib/dt.jar:$JAVA_HOME/lib/tools.jar
export PATH=$PATH:$JAVA_HOME/bin
搭建Hadoop環境
解壓安裝
登陸master
# tar -zxvf hadoop-2.8.3-x64.tar.gz -C /usr
# cd /usr/hadoop-2.8.3
# mkdir tmp logs hdf hdf/data hdf/name
修改Hadoop配置檔案
-- 修改 slaves 檔案
# vim /usr/hadoop-2.8.3/etc/hadoop/slaves
-- 删除 localhost,添加
slave1
slave2
-- 修改 core-site.xml 檔案
# vim /usr/hadoop-2.8.3/etc/hadoop/core-site.xml
-- 在 configuration 節點中添加以下内容
<property>
<name>fs.default.name</name>
<value>hdfs://master:9000</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>file:/usr/hadoop-2.8.3/tmp</value>
</property>
-- 修改 hdfs-site.xml 檔案
# vim /usr/hadoop-2.8.3/etc/hadoop/hdfs-site.xml
-- 在 configuration 節點添加以下内容
<property>
<name>dfs.datanode.data.dir</name>
<value>/usr/hadoop-2.8.3/hdf/data</value>
<final>true</final>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>/usr/hadoop-2.8.3/hdf/name</value>
<final>true</final>
</property>
-- 修改 mapred-site.xml 檔案
# cp /usr/hadoop-2.8.3/etc/hadoop/mapred-site.xml.template /usr/hadoop-2.8.3/etc/hadoop/mapred-site.xml
# vim /usr/hadoop-2.8.3/etc/hadoop/mapred-site.xml
-- 在 configuration 節點添加以下内容
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
<property>
<name>mapreduce.jobhistory.address</name>
<value>master:10020</value>
</property>
<property>
<name>mapreduce.jobhistory.webapp.address</name>
<value>master:19888</value>
</property>
-- 修改 yarn-site.xml 檔案
# vim /usr/hadoop-2.8.3/etc/hadoop/yarn-site.xml
-- 在 configuration 節點添加以下内容
<property>
<name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
<value>org.apache.mapred.ShuffleHandler</value>
</property>
<property>
<name>yarn.resourcemanager.address</name>
<value>master:8032</value>
</property>
<property>
<name>yarn.resourcemanager.scheduler.address</name>
<value>master:8030</value>
</property>
<property>
<name>yarn.resourcemanager.resource-tracker.address</name>
<value>master:8031</value>
</property>
<property>
<name>yarn.resourcemanager.admin.address</name>
<value>master:8033</value>
</property>
<property>
<name>yarn.resourcemanager.webapp.address</name>
<value>master:8088</value>
</property>
複制hadoop到slave節點
# scp -r /usr/hadoop-2.8.3 slave1:/usr
# scp -r /usr/hadoop-2.8.3 slave2:/usr
配置 master 和 slave 的 hadoop 環境變量
# vim /etc/profile
-- 添加如下内容
export HADOOP_HOME=/usr/hadoop-2.8.3
export PATH=$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$PATH
export HADOOP_LOG_DIR=/usr/hadoop-2.8.3/logs
export YARN_LOG_DIR=$HADOOP_LOG_DIR
-- 儲存後執行
# source /etc/profile
# vim ~/.bashrc
-- 添加如下内容
export HADOOP_PREFIX=/usr/hadoop-2.8.3/
-- 儲存後執行
# source ~/.bashrc
格式化 namenode
# /usr/hadoop-2.8.3/sbin/hdfs namenode -format
啟動 hadoop(僅在master節點執行)
# ssh master
# /usr/hadoop-2.8.3/sbin/start-all.sh
到這一步已經成功完成了hadoop環境的搭建
Zookeeper環境搭建
解壓縮 zookeeper 安裝包到master,并建立基本目錄
# tar -zxvf zookeeper-3.4.10.tar.gz -C /usr
# mkdir /usr/zookeeper-3.4.10/data
修改master配置檔案
-- 複制配置檔案模闆
# cp /usr/zookeeper-3.4.10/conf/zoo-sample.cfg /usr/zookeeper-3.4.10/conf/zoo.cfg
-- 修改配置檔案
# vim /usr/zookeeper-3.4.10/conf/zoo.cfg
-- 添加如下内容
dataDir=/usr/zookeeper-3.4.10/data
server.1=master:2888:3888
server.2=slave1:2888:3888
server.3=slave2:2888:3888
複制到各個子節點
# scp -r /usr/zookeeper-3.4.10 slave1:/usr
# scp -r /usr/zookeeper-3.4.10 slave2:/usr
建立myid檔案
-- master節點添加myid檔案
# ssh master
# touch /usr/zookeeper-3.4.10/data/myid
# echo 1 > /usr/zookeeper-3.4.10/data/myid
-- slave1節點添加myid檔案
# ssh slave1
# touch /usr/zookeeper-3.4.10/data/myid
# echo 2 > /usr/zookeeper-3.4.10/data/myid
-- slave2節點添加myid檔案
# ssh slave2
# touch /usr/zookeeper-3.4.10/data/myid
# echo 3 > /usr/zookeeper-3.4.10/data/myid
啟動zookeeper(master、slave1、slave2均需執行)
-- 啟動master
# ssh master
# cd /usr/zookeeper-3.4.10/bin
# ./zkServer.sh start
-- 啟動slave1
# ssh slave1
# cd /usr/zookeeper-3.4.10/bin
# ./zkServer.sh start
-- 啟動slave2
# ssh slave2
# cd /usr/zookeeper-3.4.10/bin
# ./zkServer.sh start
到這一步完成了zookeeper環境的搭建
HBase環境搭建
解壓縮hbase安裝包
# tar -zxvf hbase-1.2.6-bin.star.gz -C /usr
# mkdir /usr/hbase-1.2.6-bin/logs
修改Hbase啟動時要使用的環境變量( hbase-env.sh )
-- 打開環境變量配置檔案
# vim /usr/hbase-1.2.6/conf/hbase-env.sh
-- 添加如下内容
-- 1、設定java安裝路徑
export JAVA_HOME=/usr/java/jdk1.8.0_131
-- 2、設定hbase的日志位址
export HBASE_LOG_DIR=${HBASE_HOME}/logs
-- 3、設定是否使用hbase管理zookeeper(因使用zookeeper管理的方式,故此參數設定為false)
export HBASE_MANAGES_ZK=false
-- 4、設定hbase的pid檔案存放路徑
export HBASE_PID_DIR=/var/hadoop/pids
添加所有的region伺服器到regionservers檔案中
– 打開regionservers配置檔案
# vim /usr/hbase-1.2.5/conf/regionservers
-- 删除localhost,新增如下内容
master
slave1
slave2
注:hbase在啟動或關閉時會依次疊代通路每一行來啟動或關閉所有的region伺服器程序
修改Hbase叢集的基本配置資訊(hbase-site.xml),該配置将會覆寫Hbase的預設配置
-- 打開配置檔案
# vim /usr/hbase-1.2.6/conf/hbase-site.xml
-- 在configuration節點下添加如下内容
<property>
<name>hbase.rootdir</name>
<value>hdfs://master:9000/hbase</value>
</property>
<property>
<name>hbase.cluster.distributed</name>
<value>true</value>
</property>
<property>
<name>hbase.zookeeper.quorum</name>
<value>master,slave1,slave2</value>
</property>
<property>
<name>hbase.zookeeper.property.dataDir</name>
<value>/usr/zookeeper-3.4.10/data</value>
</property>
<property>
<name>hbase.master</name>
<value>hdfs://master:60000</value>
</property>
複制hbase到slave中
# scp -r /usr/hbase-1.2.6 slave1:/usr
# scp -r /usr/hbase-1.2.6 slave2:/usr
啟動hbase(僅在master節點上執行即可)
# ssh master
# /usr/hbase-1.2.5/bin/start-hbase.sh
到這一步hbase環境搭建完成
本地測試
使用Habse shell連接配接Hbase
cd /usr/hbase-1.2.6/bin/
./hbase shell
https://link.jianshu.com?t=https%3A%2F%2Fupload-images.jianshu.io%2Fupload_images%2F2833665-b83198fdc33649ae.png%3FimageMogr2%2Fauto-orient%2Fstrip%257CimageView2%2F2%2Fw%2F1240 
建立user1表
create 'user1','personalinfo'
https://link.jianshu.com?t=https%3A%2F%2Fupload-images.jianshu.io%2Fupload_images%2F2833665-bfd93fe082148816.png%3FimageMogr2%2Fauto-orient%2Fstrip%257CimageView2%2F2%2Fw%2F1240 對user1簡單CURD
插入
put 'user1','roow1','personalinfo:name','zhangsan'
https://link.jianshu.com?t=https%3A%2F%2Fupload-images.jianshu.io%2Fupload_images%2F2833665-4bec156eb2297bec.png%3FimageMogr2%2Fauto-orient%2Fstrip%257CimageView2%2F2%2Fw%2F1240 查詢
get 'user1','row1','personalinfo:name'
https://link.jianshu.com?t=https%3A%2F%2Fupload-images.jianshu.io%2Fupload_images%2F2833665-ee9d6e9d1c3f25f8.png%3FimageMogr2%2Fauto-orient%2Fstrip%257CimageView2%2F2%2Fw%2F1240 其他指令 help查詢就好
hbase(main):001:0> help
HBase Shell, version 1.2.6, rUnknown, Mon May 29 02:25:32 CDT 2017
Type 'help "COMMAND"', (e.g. 'help "get"' -- the quotes are necessary) for help on a specific command.
Commands are grouped. Type 'help "COMMAND_GROUP"', (e.g. 'help "general"') for help on a command group.
COMMAND GROUPS:
Group name: general
Commands: status, table_help, version, whoami
Group name: ddl
Commands: alter, alter_async, alter_status, create, describe, disable, disable_all, drop, drop_all, enable, enable_all, exists, get_table, is_disabled, is_enabled, list, locate_region, show_filters
Group name: namespace
Commands: alter_namespace, create_namespace, describe_namespace, drop_namespace, list_namespace, list_namespace_tables
Group name: dml
Commands: append, count, delete, deleteall, get, get_counter, get_splits, incr, put, scan, truncate, truncate_preserve
Group name: tools
Commands: assign, balance_switch, balancer, balancer_enabled, catalogjanitor_enabled, catalogjanitor_run, catalogjanitor_switch, close_region, compact, compact_rs, flush, major_compact, merge_region, move, normalize, normalizer_enabled, normalizer_switch, split, trace, unassign, wal_roll, zk_dump
Group name: replication
Commands: add_peer, append_peer_tableCFs, disable_peer, disable_table_replication, enable_peer, enable_table_replication, list_peers, list_replicated_tables, remove_peer, remove_peer_tableCFs, set_peer_tableCFs, show_peer_tableCFs
Group name: snapshots
Commands: clone_snapshot, delete_all_snapshot, delete_snapshot, list_snapshots, restore_snapshot, snapshot
Group name: configuration
Commands: update_all_config, update_config
Group name: quotas
Commands: list_quotas, set_quota
Group name: security
Commands: grant, list_security_capabilities, revoke, user_permission
Group name: procedures
Commands: abort_procedure, list_procedures
Group name: visibility labels
Commands: add_labels, clear_auths, get_auths, list_labels, set_auths, set_visibility
SHELL USAGE:
Quote all names in HBase Shell such as table and column names. Commas delimit
command parameters. Type <RETURN> after entering a command to run it.
Dictionaries of configuration used in the creation and alteration of tables are
Ruby Hashes. They look like this:
{'key1' => 'value1', 'key2' => 'value2', ...}
and are opened and closed with curley-braces. Key/values are delimited by the
'=>' character combination. Usually keys are predefined constants such as
NAME, VERSIONS, COMPRESSION, etc. Constants do not need to be quoted. Type
'Object.constants' to see a (messy) list of all constants in the environment.
If you are using binary keys or values and need to enter them in the shell, use
double-quote'd hexadecimal representation. For example:
hbase> get 't1', "key\x03\x3f\xcd"
hbase> get 't1', "key\003\023\011"
hbase> put 't1', "test\xef\xff", 'f1:', "\x01\x33\x40"
The HBase shell is the (J)Ruby IRB with the above HBase-specific commands added.
For more on the HBase Shell, see http://hbase.apache.org/book.html