Hbase的僞分佈式安裝部署(使用三個進程來看成集羣)html
下載地址:http://mirror.bit.edu.cn/apache/hbase/1.2.3/hbase-1.2.3-bin.tar.gzjava
在這裏,下載的是1.2.3版本node
關於hbase和hadoop的版本對應信息,可參考官檔的說明python
http://hbase.apache.org/book/configuration.html#basic.prerequisitesweb
tar -zxvf hbase-1.2.6-bin.tar.gz -C /opt/soft/ cd /opt/soft/hbase-1.2.6/ vim conf/hbase-env.sh #在內部加入export JAVA_HOME=/usr/local/jdk1.8(或者source /etc/profile) #配置Hbase mkdir /opt/soft/hbase-1.2.6/data vim conf/hbase-site.xml #####下面是在hbase-site.xml添加的項##### <configuration> <property> <name>hbase.rootdir</name> <value>hdfs://yyhhdfs/hbase</value> </property> <property> <name>hbase.zookeeper.property.dataDir</name> <value>/opt/soft/hbase-1.2.6/data/zookeeper</value> </property> <property> <name>hbase.cluster.distributed</name> <value>true</value> </property> </configuration> #在regionservers內部寫入localhost echo "localhost" > conf/regionservers
啓動Hbasesql
[root@yyh4 ~]# /opt/soft/hbase-1.2.6/bin/hbase-daemon.sh start zookeeper starting zookeeper, logging to /opt/soft/hbase-1.2.6/logs/hbase-root-zookeeper-yyh4.out [root@yyh4 ~]# /opt/soft/hbase-1.2.6/bin/hbase-daemon.sh start master starting master, logging to /opt/soft/hbase-1.2.6/logs/hbase-root-master-yyh4.out Java HotSpot(TM) 64-Bit Server VM warning: ignoring option PermSize=128m; support was removed in 8.0 Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=128m; support was removed in 8.0 [root@yyh4 ~]# /opt/soft/hbase-1.2.6/bin/hbase-daemon.sh start regionserver starting regionserver, logging to /opt/soft/hbase-1.2.6/logs/hbase-root-regionserver-yyh4.out Java HotSpot(TM) 64-Bit Server VM warning: ignoring option PermSize=128m; support was removed in 8.0 Java HotSpot(TM) 64-Bit Server VM warning: ignoring option MaxPermSize=128m; support was removed in 8.0 [root@yyh4 ~]# jps 25536 HMaster 25417 HQuorumPeer 25757 HRegionServer 26014 Jps
能夠看出,新增了HQuorumPeer,HRegionServer和HMaster三個進程。shell
經過http://yyh4:16030/訪問Hbase的web頁面apache
至此,Hbase的僞分佈式集羣搭建完畢vim
1.下載並解壓最新版的phoenix-[version]-bin.tar包緩存
地址:http://apache.fayea.com/phoenix/apache-phoenix-4.13.1-HBase-1.2/bin/
理由:hbase是1.2.6版本的,因此選用4.13.1版本的phoenix,
官網:http://apache.fayea.com/phoenix/
將phoenix-[version]-server.jar放入服務端和master節點的HBase的lib目錄下
重啓HBase
將phoenix-[version]-client.jar添加到全部Phoenix客戶端的classpath
#將phoenix-[version]-server.jar放入服務端和master節點的HBase的lib目錄下 #將phoenix-[version]-client.jar添加到全部Phoenix客戶端的classpath [root@yyh4 /]# cd /opt/soft [root@yyh4 soft]# tar -zxvf apache-phoenix-4.13.1-HBase-1.2-bin.tar.gz [root@yyh4 soft]# cp apache-phoenix-4.13.1-HBase-1.2-bin/phoenix-4.13.1-HBase-1.2-server.jar hbase-1.2.6/lib/ [root@yyh4 soft]# cp apache-phoenix-4.13.1-HBase-1.2-bin/phoenix-4.13.1-HBase-1.2-client.jar hbase-1.2.6/lib/
若要在命令行執行交互式SQL語句:
運行過程 (要求python及yum install python-argparse)
在phoenix-{version}/bin 目錄下
$ /opt/soft/apache-phoenix-4.13.1-HBase-1.2-bin/bin/sqlline.py localhost
能夠進入命令行模式
0: jdbc:phoenix:localhost>
退出命令行的方式是執行 !quit
0: jdbc:phoenix:localhost>!quit 0: jdbc:phoenix:localhost> help
其他常見指令
!all Execute the specified SQL against all the current connections !autocommit Set autocommit mode on or off !batch Start or execute a batch of statements !brief Set verbose mode off !call Execute a callable statement !close Close the current connection to the database !closeall Close all current open connections !columns List all the columns for the specified table !commit Commit the current transaction (if autocommit is off) !connect Open a new connection to the database. !dbinfo Give metadata information about the database !describe Describe a table !dropall Drop all tables in the current database !exportedkeys List all the exported keys for the specified table !go Select the current connection !help Print a summary of command usage !history Display the command history !importedkeys List all the imported keys for the specified table !indexes List all the indexes for the specified table !isolation Set the transaction isolation for this connection !list List the current connections !manual Display the SQLLine manual !metadata Obtain metadata information !nativesql Show the native SQL for the specified statement !outputformat Set the output format for displaying results (table,vertical,csv,tsv,xmlattrs,xmlelements) !primarykeys List all the primary keys for the specified table !procedures List all the procedures !properties Connect to the database specified in the properties file(s) !quit Exits the program !reconnect Reconnect to the database !record Record all output to the specified file !rehash Fetch table and column names for command completion !rollback Roll back the current transaction (if autocommit is off) !run Run a script from the specified file !save Save the current variabes and aliases !scan Scan for installed JDBC drivers !script Start saving a script to a file !set Set a sqlline variable !sql Execute a SQL command !tables List all the tables in the database !typeinfo Display the type map for the current connection !verbose Set verbose mode on
若要在命令行執行SQL腳本
$ sqlline.py localhost ../examples/stock_symbol.sql
測試ph:
#建立表 CREATE TABLE yyh ( pk VARCHAR PRIMARY KEY,val VARCHAR ); #增長,修改表數據 upsert into yyh values ('1','Helldgfho'); #查詢表數據 select * from yyh; #刪除表(若是是創建表,則同時刪除phoenix與hbase表數據,若是是創建view,則不刪除hbase的數據) drop table test;
Apache Phoenix Storage Handler是一個插件,它使Apache Hive可以使用HiveQL從Apache Hive命令行訪問Phoenix表。
使hive提供phoenix-{version}-hive.jar:
#在phoenix節點: [root@yyh4 apache-phoenix-4.13.1-HBase-1.2-bin]# scp phoenix-4.13.1-HBase-1.2-hive.jar yyh3://opt/soft/apache-hive-1.2.1-bin/lib #第1步:在hive-server節點:hive-env.sh(這裏不須要,由於lib內部) #HIVE_AUX_JARS_PATH = <jar的路徑> #第2步:將屬性添加到hive-site.xml(也不須要),以便Hive MapReduce做業可使用.jar: #<屬性> # <名稱> hive.aux.jars.path </名稱> # <值>文件:// <路徑> </值> #</屬性>
Phoenix Storage Handler支持INTERNAL和EXTERNAL Hive表。
對於INTERNAL表,Hive管理表和數據的生命週期。建立Hive表時,也會建立相應的Phoenix表。一旦Hive表被刪除,鳳凰表也被刪除。
create table phoenix_yyh ( s1 string, s2 string ) STORED BY 'org.apache.phoenix.hive.PhoenixStorageHandler' TBLPROPERTIES ( "phoenix.table.name" = "yyh", "phoenix.zookeeper.quorum" = "yyh4", "phoenix.zookeeper.znode.parent" = "/hbase", "phoenix.zookeeper.client.port" = "2181", "phoenix.rowkeys" = "s1, i1", "phoenix.column.mapping" = "s1:s1, i1:i1, f1:f1, d1:d1", "phoenix.table.options" = "SALT_BUCKETS=10, DATA_BLOCK_ENCODING='DIFF'" );
對於EXTERNAL表,Hive與現有的Phoenix表一塊兒使用,僅管理Hive元數據。從Hive中刪除EXTERNAL表只會刪除Hive元數據,但不會刪除Phoenix表。
create external table ayyh (pk string, value string) STORED BY 'org.apache.phoenix.hive.PhoenixStorageHandler' TBLPROPERTIES ( "phoenix.table.name" = "yyh", "phoenix.zookeeper.quorum" = "yyh4", "phoenix.zookeeper.znode.parent" = "/hbase", "phoenix.column.mapping" = "pk:PK,value:VAL", "phoenix.rowkeys" = "pk", "phoenix.table.options" = "SALT_BUCKETS=10, DATA_BLOCK_ENCODING='DIFF'" );
數據提取能夠經過Hive和Phoenix支持的全部方式完成:
Hive:
insert into table T values (....); insert into table T select c1,c2,c3 from source_table;
Phoenix:
upsert into table T values (.....); Phoenix CSV BulkLoad tools
All delete and update operations should be performed on the Phoenix side. See Limitations for more details. #全部刪除和更新操做都應在鳳凰方面執行。請參閱限制瞭解更多詳情。
這些選項能夠在Hive命令行界面(CLI)環境中設置。
Parameter | Default Value | Description |
---|---|---|
phoenix.upsert.batch.size | 1000 | Batch size for upsert.批量大小 |
[phoenix-table-name].disable.wal | false | Temporarily sets the table attribute DISABLE_WAL to true. Sometimes used to improve performance#暫時將表格屬性DISABLE_WAL設置爲true。有時用於提升性能 |
[phoenix-table-name].auto.flush | false | When WAL is disabled and if this value is true, then MemStore is flushed to an HFile.#當WAL被禁用時,若是該值爲true,則MemStore被刷新爲HFile |
You can use HiveQL for querying data in a Phoenix table. A Hive query on a single table can be as fast as running the query in the Phoenix CLI with the following property settings: hive.fetch.task.conversion=more and hive.exec.parallel=true
#您可使用HiveQL查詢Phoenix表中的數據。單個表上的Hive查詢能夠像運行Phoenix CLI中的查詢同樣快,並具備如下屬性設置:hive.fetch.task.conversion = more和hive.exec.parallel = true
Parameter | Default Value | Description |
---|---|---|
hbase.scan.cache | 100 | Read row size for a unit request#讀取單位請求的行大小 |
hbase.scan.cacheblock | false | Whether or not cache block#是否緩存塊 |
split.by.stats | false | If true, mappers use table statistics. One mapper per guide post. |
[hive-table-name].reducer.count | 1 | Number of reducers. In Tez mode, this affects only single-table queries. See Limitations. |
[phoenix-table-name].query.hint | Hint for Phoenix query (for example, NO_INDEX) |
#注意這樣的操做,尤爲是大小寫的處理,其實更符合全部組件 #####虛擬集羣測試概要##### #hbase shell #habse節點 2列,3數據 create 'YINGGDD','info' put 'YINGGDD', 'row021','info:name','phoenix' put 'YINGGDD', 'row012','info:name','hbase' put 'YINGGDD', 'row012','info:sname','shbase' #phoenix 節點 #opt/soft/apache-phoenix-4.13.1-HBase-1.2-bin/bin/sqlline.py yyh4 create table "YINGGDD" ("id" varchar primary key, "info"."name" varchar, "info"."sname" varchar); select * from "YINGGDD"; #hive節點 #beeline -u jdbc:hive2://yyh3:10000 create external table yingggy ( id string, name string, sname string) STORED BY 'org.apache.phoenix.hive.PhoenixStorageHandler' TBLPROPERTIES ( "phoenix.table.name" = "YINGGDD" , "phoenix.zookeeper.quorum" = "yyh4" , "phoenix.zookeeper.znode.parent" = "/hbase", "phoenix.column.mapping" = "pk:id,name:name,sname:sname", "phoenix.rowkeys" = "id", "phoenix.table.options" = "SALT_BUCKETS=10, DATA_BLOCK_ENCODING='DIFF'" ); select * from yingggy;
############公司內部集羣測試############### #hive 111 --- ph 113 #ph表 beeline -u jdbc:hive2://192.168.1.111:10000/dm -f /usr/local/apache-hive-2.3.2-bin/bin/yyh.sql add jar /usr/local/apache-hive-2.3.2-bin/lib/phoenix-4.9.0-cdh5.9.1-hive.jar; add jar /usr/local/apache-hive-2.3.2-bin/lib/phoenix-4.9.0-HBase-1.2-hive.jar; add jar /usr/local/apache-hive-2.3.2-bin/lib/phoenix-4.9.0-HBase-1.2-client.jar; #driver#phoenix-4.9.0-HBase-1.2-hive.org.apache.phoenix.jdbc.PhoenixDriver create external table ayyh (pk string, value string) STORED BY 'org.apache.phoenix.hive.PhoenixStorageHandler' TBLPROPERTIES ( "phoenix.table.name" = "yyh", "phoenix.zookeeper.quorum" = "192.168.1.112", "phoenix.zookeeper.znode.parent" = "/hbase", "phoenix.column.mapping" = "pk:PK,value:VAL", "phoenix.rowkeys" = "pk", "phoenix.table.options" = "SALT_BUCKETS=10, DATA_BLOCK_ENCODING='DIFF'" ); #113做爲hive的客戶端,去創建表格 cd ~ beeline -u jdbc:hive2://192.168.1.111:10000/dm -u hive -f yyha.sql #117登錄hive查看錶格 beeline -u jdbc:hive2://192.168.1.111:10000/dm -e "select * from ayyh limit 3"
Error: Error while compiling statement: FAILED: SemanticException Cannot find class 'org.apache.phoenix.hive.PhoenixStorageHandler' (state=42000,code=40000) ##分析以下,缺乏jar包 #導入add jar解決
#所以,進入beeline之後,用add jar(公司集羣,不容許頻繁啓動): 0: jdbc:hive2://yyh3:10000> add jar /opt/soft/apache-hive-1.2.1-bin/lib/phoenix-4.13.1-HBase-1.2-hive.jar; INFO : Added [/opt/soft/apache-hive-1.2.1-bin/lib/phoenix-4.13.1-HBase-1.2-hive.jar] to class path INFO : Added resources: [/opt/soft/apache-hive-1.2.1-bin/lib/phoenix-4.13.1-HBase-1.2-hive.jar] 發現報新錯誤: Error: Error while processing statement: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:No suitable driver found for jdbc:phoenix:yyh4:2181:/hbase;) (state=08S01,code=1) 其實,這個問題是在查公司集羣時體現出的報錯,爲告終構清晰,單獨寫在下面
[import@slave113 phoenix-4.9.0-cdh5.9.1]$ beeline -u jdbc:hive2://192.168.1.111:10000/dm SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/usr/local/apache-hive-2.3.2-bin/lib/log4j-slf4j-impl-2.6.2.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/usr/local/apache-hive-2.3.2-bin/lib/phoenix-4.13.1-HBase-1.2-hive.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/usr/local/apache-hive-2.3.2-bin/lib/phoenix-4.9.0-cdh5.9.1-hive.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/opt/cloudera/parcels/CDH-5.9.0-1.cdh5.9.0.p0.23/jars/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory] Connecting to jdbc:hive2://192.168.1.111:10000/dm Connected to: Apache Hive (version 2.3.2) Driver: Hive JDBC (version 2.3.2) Transaction isolation: TRANSACTION_REPEATABLE_READ Beeline version 2.3.2 by Apache Hive 0: jdbc:hive2://192.168.1.111:10000/dm> add jar /usr/local/apache-hive-2.3.2-bin/lib/phoenix-4.13.1-HBase-1.2-hive.jar . . . . . . . . . . . . . . . . . . . > ; Error: Error while processing statement: /usr/local/apache-hive-2.3.2-bin/lib/phoenix-4.13.1-HBase-1.2-hive.jar does not exist (state=,code=1) 0: jdbc:hive2://192.168.1.111:10000/dm> add jar /usr/local/apache-hive-2.3.2-bin/lib/phoenix-4.13.1-HBase-1.2-hive.jar; No rows affected (0.051 seconds) 0: jdbc:hive2://192.168.1.111:10000/dm> create external table ayyh . . . . . . . . . . . . . . . . . . . > (pk string, . . . . . . . . . . . . . . . . . . . > value string) . . . . . . . . . . . . . . . . . . . > STORED BY 'org.apache.phoenix.hive.PhoenixStorageHandler' . . . . . . . . . . . . . . . . . . . > TBLPROPERTIES ( . . . . . . . . . . . . . . . . . . . > "phoenix.table.name" = "yyh", . . . . . . . . . . . . . . . . . . . > "phoenix.zookeeper.quorum" = "192.168.1.113:2181", . . . . . . . . . . . . . . . . . . . > "phoenix.zookeeper.znode.parent" = "/hbase", . . . . . . . . . . . . . . . . . . . > "phoenix.column.mapping" = "pk:PK,value:VAL", . . . . . . . . . . . . . . . . . . . > "phoenix.rowkeys" = "pk", . . . . . . . . . . . . . . . . . . . > "phoenix.table.options" = "SALT_BUCKETS=10, DATA_BLOCK_ENCODING='DIFF'" . . . . . . . . . . . . . . . . . . . > ); Error: org.apache.hive.service.cli.HiveSQLException: Error while processing statement: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:No suitable driver found for jdbc:phoenix:192.168.1.113:2181:2181:/hbase;) at org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:380) at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:257) at org.apache.hive.service.cli.operation.SQLOperation.access$800(SQLOperation.java:91) at org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork$1.run(SQLOperation.java:348) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1698) at org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork.run(SQLOperation.java:362) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Caused by: org.apache.hadoop.hive.ql.metadata.HiveException: MetaException(message:No suitable driver found for jdbc:phoenix:192.168.1.113:2181:2181:/hbase;) at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:862) at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:867) at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4356) at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:354) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2183) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1839) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1526) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1232) at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:255) ... 11 more Caused by: MetaException(message:No suitable driver found for jdbc:phoenix:192.168.1.113:2181:2181:/hbase;) at org.apache.phoenix.hive.PhoenixMetaHook.preCreateTable(PhoenixMetaHook.java:84) at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:747) at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:740) at sun.reflect.GeneratedMethodAccessor74.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:173) at com.sun.proxy.$Proxy34.createTable(Unknown Source) at sun.reflect.GeneratedMethodAccessor74.invoke(Unknown Source) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2330) at com.sun.proxy.$Proxy34.createTable(Unknown Source) at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:852) ... 22 more (state=08S01,code=1) 0: jdbc:hive2://192.168.1.111:10000/dm> Closing: 0: jdbc:hive2://192.168.1.111:10000/dm You have new mail in /var/spool/mail/import
以上問題是由於add jar並不能加載driver
因此加入jar包,重啓hive解決
因此加入jar包,重啓hive,又出現新的錯誤,懷疑是版本問題,目前在公司集羣有問題,可是在虛擬機沒有問題,因此在排查。
0: jdbc:hive2://192.168.1.111:10000/dm> create external table ayyh . . . . . . . . . . . . . . . . . . . > (pk string, . . . . . . . . . . . . . . . . . . . > value string) . . . . . . . . . . . . . . . . . . . > STORED BY 'org.apache.phoenix.hive.PhoenixStorageHandler' . . . . . . . . . . . . . . . . . . . > TBLPROPERTIES ( . . . . . . . . . . . . . . . . . . . > "phoenix.table.name" = "yyh", . . . . . . . . . . . . . . . . . . . > "phoenix.zookeeper.quorum" = "192.168.1.112", . . . . . . . . . . . . . . . . . . . > "phoenix.zookeeper.znode.parent" = "/hbase", . . . . . . . . . . . . . . . . . . . > "phoenix.column.mapping" = "pk:PK,value:VAL", . . . . . . . . . . . . . . . . . . . > "phoenix.rowkeys" = "pk", . . . . . . . . . . . . . . . . . . . > "phoenix.table.options" = "SALT_BUCKETS=10, DATA_BLOCK_ENCODING='DIFF'" . . . . . . . . . . . . . . . . . . . > ); Error: org.apache.hive.service.cli.HiveSQLException: Error while processing statement: FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. org.apache.hadoop.hbase.client.HBaseAdmin.<init>(Lorg/apache/hadoop/hbase/client/HConnection;)V at org.apache.hive.service.cli.operation.Operation.toSQLException(Operation.java:380) at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:257) at org.apache.hive.service.cli.operation.SQLOperation.access$800(SQLOperation.java:91) at org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork$1.run(SQLOperation.java:348) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1714) at org.apache.hive.service.cli.operation.SQLOperation$BackgroundWork.run(SQLOperation.java:362) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) Caused by: java.lang.NoSuchMethodError: org.apache.hadoop.hbase.client.HBaseAdmin.<init>(Lorg/apache/hadoop/hbase/client/HConnection;)V at org.apache.phoenix.query.ConnectionQueryServicesImpl.getAdmin(ConnectionQueryServicesImpl.java:3282) at org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2377) at org.apache.phoenix.query.ConnectionQueryServicesImpl$13.call(ConnectionQueryServicesImpl.java:2352) at org.apache.phoenix.util.PhoenixContextExecutor.call(PhoenixContextExecutor.java:76) at org.apache.phoenix.query.ConnectionQueryServicesImpl.init(ConnectionQueryServicesImpl.java:2352) at org.apache.phoenix.jdbc.PhoenixDriver.getConnectionQueryServices(PhoenixDriver.java:232) at org.apache.phoenix.jdbc.PhoenixEmbeddedDriver.createConnection(PhoenixEmbeddedDriver.java:147) at org.apache.phoenix.jdbc.PhoenixDriver.connect(PhoenixDriver.java:202) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.phoenix.hive.util.PhoenixConnectionUtil.getConnection(PhoenixConnectionUtil.java:81) at org.apache.phoenix.hive.PhoenixMetaHook.preCreateTable(PhoenixMetaHook.java:55) at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:747) at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:740) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:173) at com.sun.proxy.$Proxy34.createTable(Unknown Source) at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) at java.lang.reflect.Method.invoke(Method.java:498) at org.apache.hadoop.hive.metastore.HiveMetaStoreClient$SynchronizedHandler.invoke(HiveMetaStoreClient.java:2330) at com.sun.proxy.$Proxy34.createTable(Unknown Source) at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:852) at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:867) at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:4356) at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:354) at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:199) at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:100) at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:2183) at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1839) at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1526) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1237) at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1232) at org.apache.hive.service.cli.operation.SQLOperation.runQuery(SQLOperation.java:255) ... 11 more (state=08S01,code=1)
官網 http://phoenix.apache.org/index.html
Hive整合 http://phoenix.apache.org/hive_storage_handler.html