Requirements
-------------
- Java 1.5.x, preferably from Sun.
- Hadoop 0.16.x. This version of HBase will only run on Hadoop 0.16.x..
- ssh must be installed and sshd must be running to use Hadoop's scripts
to manage remote Hadoop daemons.
- HBase currently is a file handle hog. The usual default of 1024 on *nix systems
is insufficient if you are loading any significant amount of data into regionservers.
See the FAQ: Why do I see "java.io.IOException...(Too many open files)" in my logs? for how to up the limit.
[C] Installation of HBase
--------------------------
(1) Java 1.5.x
- installed java 1.6.x
- Not sure whether 1.6.x will work
(2) Unzipped the Hadoop archive file (ver 1.3) at ~/hbase
(3) ssh & sshd configuration (execution)
- sudo apt-get install openssh-server
- sudo /etc/init.d/ssh start (stop)
(4) edit the limit number of file handlers
- checked /etc/security/limits.conf but there is no option to edit for this
[C] HBase Configuration
------------------------
(1) ${HBASE_HOME} : the real location of hbase (/home/luran/hbase)
(2) modify JAVA_HOME information on ${HBASE_HOME}/conf/hbase-env.sh
(3) start HBase by ${HBASE_HOME}/bin/start_hbase.sh
(4) stop HBase by ${HBASE_HOME}/bin/stop-hbase.sh
'Development > Hadoop, NoSQL, BigData' 카테고리의 다른 글
Docker기반 Spark Cluster 설치하기 (6) | 2020.12.15 |
---|---|
Local AirFlow 설치하기 (0) | 2020.12.09 |
brew로 local zeppelin 설치하기 (0) | 2020.11.30 |
The Hadoop Distributed File System : Architecture and Design 요약 (0) | 2008.07.29 |
HBase에서 HQL 사용하기 (0) | 2008.07.25 |
HBase 설치/설정하기 (0) | 2008.07.25 |
HDFS 설정하기 (0) | 2008.07.25 |
Hadoop Installation on Ubuntu Linux 7.10 (0) | 2008.07.17 |