使用Java API访问hbase很慢的问题
 有没有人出现过这种情况,就是使用java对HBASE进行连接的时候可以连接上,但是所花的时间都是20秒左右(不管是连接Linux上的HBASE还是自己windows上的HBASE都要20秒左右),不知道是什么原因导致的,时间都花在连接数据库上了,Java代码运行的时候也没有报任何的错误
habse安装的是单机版的,版本是1.2.6.1,zookeeper使用的是自带的,hadoop刚开始没有安装,后面安装了,但还是没什么作用,hosts文件这些都配置好了。
    现在一个简单的增删改查都要花20秒钟,崩溃,被这个问题困扰了好几天了。。好难受,求大神帮忙解答啊。。

2个回答

补充一下下面是我的配置文件的配置
hbase-site.xml的配置


hbase.rootdir
file:///usr/local/app/hbase-1.2/data


hbase.zookeeper.property.dataDir
/usr/local/app/hbase-1.2/zookeerperData


hbase-env.sh的配置
export JAVA_HOME=/usr/local/jdk1.8

export HBASE_MANAGES_ZK=true

做个HBase连接池 ,能快些,网上能找到

Csdn user default icon
上传中...
上传图片
插入图片
抄袭、复制答案,以达到刷声望分或其他目的的行为,在CSDN问答是严格禁止的,一经发现立刻封号。是时候展现真正的技术了!
其他相关推荐
使用Java对hbase进行连接测试,连接不上,超时,
我在Linux上安装了hbase的单机版,hbase版本是1.2.6, zookeeper使用的是hbase自己的,Linux防火墙开放了16010和2181端口, 使用hbase shell 可以正常操作hbase,浏览器也可以正常访问hbase, 但是我使用Java对hbase进行连接测试的时候就是一直连接不上,然后当我关闭 Linux防火墙的时候java又可以正常的对hbase进行连接访问了,???,这是怎么回事,是因为Linux防火墙还需要开放其他端口吗
Java项目连接hbase时超时
本人小白一枚,现在有个项目是用Java项目连接hbase。我用的是windows下的Java项目连接linux虚拟机上的hbase,hbase开启之后连接提示连接超时(ps:主机之间可互通),求助。。。 log4j:WARN No appenders could be found for logger (org.apache.hadoop.util.Shell). log4j:WARN Please initialize the log4j system properly. SLF4J: Class path contains multiple SLF4J bindings. SLF4J: Found binding in [jar:file:/E:/apache-tomcat-7.0.85-windows-x64/apache-tomcat-7.0.85/webapps/car_hbase/WEB-INF/lib/slf4j-log4j12-1.5.8.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: Found binding in [jar:file:/E:/apache-tomcat-7.0.85-windows-x64/apache-tomcat-7.0.85/webapps/car_hbase/WEB-INF/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class] SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation. e785dc9437424bf8a7714f460293896c HBASE表创建失败! java.io.IOException: Failed to get result within timeout, timeout=60000ms at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:232) at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:58) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:219) at org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:277) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:438) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:312) at org.apache.hadoop.hbase.MetaTableAccessor.fullScan(MetaTableAccessor.java:604) at org.apache.hadoop.hbase.MetaTableAccessor.tableExists(MetaTableAccessor.java:366) at org.apache.hadoop.hbase.client.HBaseAdmin.tableExists(HBaseAdmin.java:410) at org.apache.hadoop.hbase.client.HBaseAdmin.tableExists(HBaseAdmin.java:420) at util.HBaseUtil.createTable(HBaseUtil.java:45) at util.HbaseDemo.createTable(HbaseDemo.java:55) at util.StartupListener.contextInitialized(StartupListener.java:31) at org.apache.catalina.core.StandardContext.listenerStart(StandardContext.java:5118) at org.apache.catalina.core.StandardContext.startInternal(StandardContext.java:5641) at org.apache.catalina.util.LifecycleBase.start(LifecycleBase.java:145) at org.apache.catalina.core.ContainerBase.addChildInternal(ContainerBase.java:1015) at org.apache.catalina.core.ContainerBase.addChild(ContainerBase.java:991) at org.apache.catalina.core.StandardHost.addChild(StandardHost.java:652) at org.apache.catalina.startup.HostConfig.deployDirectory(HostConfig.java:1296) at org.apache.catalina.startup.HostConfig$DeployDirectory.run(HostConfig.java:2038) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745)
关于java开发hbase的框架
刚接触hbase,需要java客户端的编程。目前之了解到hbase客户端的API的编写。就是get,put之类的层次。但以往开发关系型数据库的时候都有很多框架。请问,有经验的朋友是如何快速开发基于hbase的程序的?谢谢?比如是用什么,可以使用sql语言来查询之类的。谢谢
java开发hbase
刚接触hbase,需要java客户端的编程。目前之了解到hbase客户端的API的编写。就是get,put之类的层次。但以往开发关系型数据库的时候都有很多框架。请问,有经验的朋友是如何快速开发基于hbase的程序的?谢谢?比如是用什么,可以使用sql语言来查询之类的。谢谢
java连hbase永久等待问题
只有一个有价值信息就是,Session establishment complete on server tb86-es01/172.18.1.86:2181, sessionid = 0x15b79d49b158594, negotiated timeout = 40000 代码如下: public class extract_hbase_job { private static Configuration hconf = null; static{ Configuration conf = new Configuration(); conf.set("hbase.zookeeper.quorum", "tb86-es01,tb87-es02,tb88-es03"); conf.set("hbase.zookeeper.property.clientPort", "2181"); System.setProperty("hadoop.home.dir", "D:\\work\\apache\\hadoop\\hadoop-2.6.0"); hconf = HBaseConfiguration.create(conf); } public static void getRecord(String tableName,String rowKey) throws Exception{ HTable table = new HTable(hconf, tableName.getBytes()); Get get = new Get(rowKey.getBytes()); Result rs = table.get(get); for (KeyValue kv : rs.raw()) { System.out.println(new String(kv.getQualifier()) + " " ); } @Test public void test() throws Exception{ extract_hbase_job.getRecord("default:relation_graph", "00:global:cp:10878983"); } } hbase中也有报错: java.lang.RuntimeException: Failed construction of Master: class org.apache.hadoop.hbase.master.HMaster at org.apache.hadoop.hbase.master.HMaster.constructMaster(HMaster.java:1867) at org.apache.hadoop.hbase.master.HMasterCommandLine.startMaster(HMasterCommandLine.java:198) at org.apache.hadoop.hbase.master.HMasterCommandLine.run(HMasterCommandLine.java:139) at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:70) at org.apache.hadoop.hbase.util.ServerCommandLine.doMain(ServerCommandLine.java:126) at org.apache.hadoop.hbase.master.HMaster.main(HMaster.java:1881) ed by: java.net.BindException: 地址已在使用 at sun.nio.ch.Net.bind0(Native Method) at sun.nio.ch.Net.bind(Net.java:444) at sun.nio.ch.Net.bind(Net.java:436) at sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:214) at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74) at org.apache.hadoop.hbase.ipc.RpcServer.bind(RpcServer.java:2366) at org.apache.hadoop.hbase.ipc.RpcServer$Listener.<init>(RpcServer.java:524) at org.apache.hadoop.hbase.ipc.RpcServer.<init>(RpcServer.java:1896) at org.apache.hadoop.hbase.regionserver.RSRpcServices.<init>(RSRpcServices.java:788) at org.apache.hadoop.hbase.master.MasterRpcServices.<init>(MasterRpcServices.java:197) at org.apache.hadoop.hbase.master.HMaster.createRpcServices(HMaster.java:401) at org.apache.hadoop.hbase.regionserver.HRegionServer.<init>(HRegionServer.java:487) at org.apache.hadoop.hbase.master.HMaster.<init>(HMaster.java:271) at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.hadoop.hbase.master.HMaster.constructMaster(HMaster.java:1862) ... 5 more
java连接hbase Kerberos 24小时报错问题
java连接hbase kerberos 24小时过期问题。求指点 ``` try { String krbStr = Thread.currentThread().getContextClassLoader().getResource("krb5.ini").getFile(); String keyStr = Thread.currentThread().getContextClassLoader().getResource(keytab).getFile(); System.setProperty("java.security.krb5.conf", krbStr); UserGroupInformation.setConfiguration(conf); UserGroupInformation.loginUserFromKeytab(principal, keyStr); } catch (IOException e) { log.error(e); } try { HBaseAdmin admin = new HBaseAdmin(conf); if (!admin.tableExists(tableName)) { HTableDescriptor tableDescripter = new HTableDescriptor( tableName.getBytes()); tableDescripter.addFamily(new HColumnDescriptor("data")); admin.createTable(tableDescripter); } } catch (Exception e) { log.error( e); } ``` 以上代码刚运行时没问题,但是过了24小时左右admin.tableExists(tableName)这句就会报错,错误如下 org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=36, exceptions: Sat Jul 13 14:30:29 CST 2019, null, java.net.SocketTimeoutException: callTimeout=60000, callDuration=68090: row 'p_rsdisk,,' on table 'hbase:meta' at region=hbase:meta,,1.1588230740, hostname=xx-xx-xx-xx-xx.indata.com,16020,1560266314048, seqNum=0 at org.apache.hadoop.hbase.client.RpcRetryingCallerWithReadReplicas.throwEnrichedException(RpcRetryingCallerWithReadReplicas.java:271) at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:223) at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:61) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:200) at org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:320) at org.apache.hadoop.hbase.client.ClientScanner.nextScanner(ClientScanner.java:295) at org.apache.hadoop.hbase.client.ClientScanner.initializeScannerInConstruction(ClientScanner.java:160) at org.apache.hadoop.hbase.client.ClientScanner.<init>(ClientScanner.java:155) at org.apache.hadoop.hbase.client.HTable.getScanner(HTable.java:811) at org.apache.hadoop.hbase.MetaTableAccessor.fullScan(MetaTableAccessor.java:602) at org.apache.hadoop.hbase.MetaTableAccessor.tableExists(MetaTableAccessor.java:366) at org.apache.hadoop.hbase.client.HBaseAdmin.tableExists(HBaseAdmin.java:303) at org.apache.hadoop.hbase.client.HBaseAdmin.tableExists(HBaseAdmin.java:313) at javax.servlet.http.HttpServlet.service(HttpServlet.java:731) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:303) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208) at org.apache.tomcat.websocket.server.WsFilter.doFilter(WsFilter.java:52) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208) at org.apache.catalina.filters.CorsFilter.handleNonCORS(CorsFilter.java:436) at org.apache.catalina.filters.CorsFilter.doFilter(CorsFilter.java:177) at org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(ApplicationFilterChain.java:241) at org.apache.catalina.core.ApplicationFilterChain.doFilter(ApplicationFilterChain.java:208) at org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapperValve.java:220) at org.apache.catalina.core.StandardContextValve.invoke(StandardContextValve.java:122) at org.apache.catalina.authenticator.AuthenticatorBase.invoke(AuthenticatorBase.java:505) at org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.java:170) at org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.java:103) at org.apache.catalina.valves.AccessLogValve.invoke(AccessLogValve.java:956) at org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.java:116) at org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.java:423) at org.apache.coyote.http11.AbstractHttp11Processor.process(AbstractHttp11Processor.java:1079) at org.apache.coyote.AbstractProtocol$AbstractConnectionHandler.process(AbstractProtocol.java:625) at org.apache.tomcat.util.net.JIoEndpoint$SocketProcessor.run(JIoEndpoint.java:316) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at org.apache.tomcat.util.threads.TaskThread$WrappingRunnable.run(TaskThread.java:61) at java.lang.Thread.run(Thread.java:748) Caused by: java.net.SocketTimeoutException: callTimeout=60000, callDuration=68090: row 'p_rsdisk,,' on table 'hbase:meta' at region=hbase:meta,,1.1588230740, hostname=xx-xx-xx-xx-xx.indata.com,16020,1560266314048, seqNum=0 at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:159) at org.apache.hadoop.hbase.client.ResultBoundedCompletionService$QueueingFuture.run(ResultBoundedCompletionService.java:64) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) ... 1 more Caused by: java.io.IOException: Could not set up IO Streams to xx-xx-xx-xx-xx.indata.com/xx.xx.xx.xx:16020 at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:777) at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:885) at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:854) at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1180) at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213) at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287) at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.scan(ClientProtos.java:32651) at org.apache.hadoop.hbase.client.ScannerCallable.openScanner(ScannerCallable.java:372) at org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:199) at org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:62) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:200) at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas$RetryingRPC.call(ScannerCallableWithReplicas.java:369) at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas$RetryingRPC.call(ScannerCallableWithReplicas.java:343) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126) ... 4 more Caused by: java.lang.RuntimeException: SASL authentication failed. The most likely cause is missing or invalid credentials. Consider 'kinit'. at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection$1.run(RpcClientImpl.java:677) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.handleSaslConnectionFailure(RpcClientImpl.java:635) at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:743) ... 17 more Caused by: javax.security.sasl.SaslException: GSS initiate failed [Caused by GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt)] at com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge(GssKrb5Client.java:211) at org.apache.hadoop.hbase.security.HBaseSaslRpcClient.saslConnect(HBaseSaslRpcClient.java:179) at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupSaslConnection(RpcClientImpl.java:609) at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.access$600(RpcClientImpl.java:154) at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection$2.run(RpcClientImpl.java:735) at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection$2.run(RpcClientImpl.java:732) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1628) at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:732) ... 17 more Caused by: GSSException: No valid credentials provided (Mechanism level: Failed to find any Kerberos tgt) at sun.security.jgss.krb5.Krb5InitCredential.getInstance(Krb5InitCredential.java:147) at sun.security.jgss.krb5.Krb5MechFactory.getCredentialElement(Krb5MechFactory.java:122) at sun.security.jgss.krb5.Krb5MechFactory.getMechanismContext(Krb5MechFactory.java:187) at sun.security.jgss.GSSManagerImpl.getMechanismContext(GSSManagerImpl.java:224) at sun.security.jgss.GSSContextImpl.initSecContext(GSSContextImpl.java:212) at sun.security.jgss.GSSContextImpl.initSecContext(GSSContextImpl.java:179) at com.sun.security.sasl.gsskerb.GssKrb5Client.evaluateChallenge(GssKrb5Client.java:192) ... 26 more 求解原因,是因为票据过期吗?
java连接单机hbase操作数据
16/03/04 17:09:56 INFO support.ClassPathXmlApplicationContext: Refreshing org.springframework.context.support.ClassPathXmlApplicationContext@1f6ae4d: startup date [Fri Mar 04 17:09:56 CST 2016]; root of context hierarchy 16/03/04 17:09:56 INFO xml.XmlBeanDefinitionReader: Loading XML bean definitions from class path resource [applicationContest.xml] 16/03/04 17:09:57 INFO support.DefaultListableBeanFactory: Pre-instantiating singletons in org.springframework.beans.factory.support.DefaultListableBeanFactory@1d8fe20: defining beans [hbaseConfiguration,htemplate]; root of factory hierarchy 16/03/04 17:09:57 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable 16/03/04 17:09:57 INFO zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0xc4dc7c connecting to ZooKeeper ensemble=192.168.1.202:2181 16/03/04 17:09:57 INFO zookeeper.ZooKeeper: Client environment:zookeeper.version=3.4.5-cdh5.2.0--1, built on 10/11/2014 20:49 GMT 16/03/04 17:09:57 INFO zookeeper.ZooKeeper: Client environment:host.name=xiaoming 16/03/04 17:09:57 INFO zookeeper.ZooKeeper: Client environment:java.version=1.7.0_17 16/03/04 17:09:57 INFO zookeeper.ZooKeeper: Client environment:java.vendor=Oracle Corporation 16/03/04 17:09:57 INFO zookeeper.ZooKeeper: Client environment:java.home=C:\Program Files (x86)\Java\jdk1.7.0_17\jre 16/03/04 17:09:57 INFO zookeeper.ZooKeeper: Client environment:java.class.path=F:\mess\12Hadoop\workspace\myhbasetest\WebRoot\WEB-INF\classes;F:\mess\12Hadoop\workspace\myhbasetest\WebRoot\WEB-INF\lib\commons-codec-1.7.jar;F:\mess\12Hadoop\workspace\myhbasetest\WebRoot\WEB-INF\lib\commons-collections-3.2.1.jar;F:\mess\12Hadoop\workspace\myhbasetest\WebRoot\WEB-INF\lib\commons-configuration-1.6.jar;F:\mess\12Hadoop\workspace\myhbasetest\WebRoot\WEB-INF\lib\commons-lang-2.6.jar;F:\mess\12Hadoop\workspace\myhbasetest\WebRoot\WEB-INF\lib\commons-logging-1.1.1.jar;F:\mess\12Hadoop\workspace\myhbasetest\WebRoot\WEB-INF\lib\guava-12.0.1.jar;F:\mess\12Hadoop\workspace\myhbasetest\WebRoot\WEB-INF\lib\hadoop-auth-2.5.0-cdh5.2.0.jar;F:\mess\12Hadoop\workspace\myhbasetest\WebRoot\WEB-INF\lib\hadoop-common-2.5.0-cdh5.2.0.jar;F:\mess\12Hadoop\workspace\myhbasetest\WebRoot\WEB-INF\lib\hadoop-core-2.5.0-mr1-cdh5.2.0.jar;F:\mess\12Hadoop\workspace\myhbasetest\WebRoot\WEB-INF\lib\hbase-client-0.98.6-cdh5.2.0.jar;F:\mess\12Hadoop\workspace\myhbasetest\WebRoot\WEB-INF\lib\hbase-common-0.98.6-cdh5.2.0.jar;F:\mess\12Hadoop\workspace....... 16/03/04 17:09:57 INFO zookeeper.ClientCnxn: Opening socket connection to server 192.168.1.202/192.168.1.202:2181. Will not attempt to authenticate using SASL (unknown error) 16/03/04 17:09:57 INFO zookeeper.ClientCnxn: Socket connection established to 192.168.1.202/192.168.1.202:2181, initiating session 16/03/04 17:09:57 INFO zookeeper.ClientCnxn: Session establishment complete on server 192.168.1.202/192.168.1.202:2181, sessionid = 0x15341c9328d0012, negotiated timeout = 40000 16/03/04 17:17:50 WARN client.HConnectionManager$HConnectionImplementation: Encountered problems when prefetch hbase:meta table: org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=31, exceptions: .......Fri Mar 04 17:17:50 CST 2016, org.apache.hadoop.hbase.client.RpcRetryingCaller@b7b17a, java.net.UnknownHostException: unknown host: hbase ......at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:129) at org.apache.hadoop.hbase.client.HTable.getRowOrBefore(HTable.java:714) at org.apache.hadoop.hbase.client.MetaScanner.metaScan(MetaScanner.java:144) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.prefetchRegionCache(HConnectionManager.java:1140) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegionInMeta(HConnectionManager.java:1204) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1092) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.locateRegion(HConnectionManager.java:1049) at org.apache.hadoop.hbase.client.HConnectionManager$HConnectionImplementation.getRegionLocation(HConnectionManager.java:890) at org.apache.hadoop.hbase.client.RegionServerCallable.prepare(RegionServerCallable.java:72) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:113) at org.apache.hadoop.hbase.client.HTable.get(HTable.java:780) at org.springframework.data.hadoop.hbase.HbaseTemplate$2.doInTable(HbaseTemplate.java:182) at org.springframework.data.hadoop.hbase.HbaseTemplate.execute(HbaseTemplate.java:58) at org.springframework.data.hadoop.hbase.HbaseTemplate.get(HbaseTemplate.java:168) at org.springframework.data.hadoop.hbase.HbaseTemplate.get(HbaseTemplate.java:158) at com.bw.test.Mytest.get(Mytest.java:46) at com.bw.test.Mytest.main(Mytest.java:40) Caused by: java.net.UnknownHostException: unknown host: hbase at org.apache.hadoop.hbase.ipc.RpcClient$Connection.<init>(RpcClient.java:385) at org.apache.hadoop.hbase.ipc.RpcClient.createConnection(RpcClient.java:351) at org.apache.hadoop.hbase.ipc.RpcClient.getConnection(RpcClient.java:1530) at org.apache.hadoop.hbase.ipc.RpcClient.call(RpcClient.java:1442) at org.apache.hadoop.hbase.ipc.RpcClient.callBlockingMethod(RpcClient.java:1661) at org.apache.hadoop.hbase.ipc.RpcClient$BlockingRpcChannelImplementation.callBlockingMethod(RpcClient.java:1719) at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.get(ClientProtos.java:29966) at org.apache.hadoop.hbase.protobuf.ProtobufUtil.getRowOrBefore(ProtobufUtil.java:1562) at org.apache.hadoop.hbase.client.HTable$2.call(HTable.java:710) at org.apache.hadoop.hbase.client.HTable$2.call(HTable.java:708) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:114) ... 16 more
HBASE如何创建用户名密码?
hbase如何创建用户名密码?java api如何指定用户名密码访问hbase
【急!】java访问hbase出错,无明显异常,无法读取对应表的信息(需求就是从Hbase之中的某张表读取信息)
有点着急请哪位大神帮忙一下 ## 本地Hosts配置(对应的机上面的hosts的文件配置都一样): -------------------------- 127.0.0.1 localhost 192.168.0.25 Master.Hadoop 192.168.0.26 Slave1.Hadoop 192.168.0.27 Slave2.Hadoop 192.168.0.28 Slave3.Hadoop ## hbase的基本信息: -------------------------- http://master.hadoop:60010/master-status HBase Root Directory: hdfs://Master.Hadoop:9000/hbase ## 代码 -------------------------- Configuration configuration = HBaseConfiguration.create(); configuration = HBaseConfiguration.create(); configuration.set("hbase.zookeeper.property.clientPort", "2181"); configuration.set("hbase.zookeeper.quorum", "Master.Hadoop,Slave1.Hadoop,Slave2.Hadoop,Slave1.Hadoop"); configuration.set("hbase.master", "Master.Hadoop:60010"); //"Master.Hadoop:60010" System.setProperty("hadoop.home.dir", "C:\\Program Files\\hadoop\\hadoop-common-2.2.0-bin-master"); connection = ConnectionFactory.createConnection(configuration); System.out.println( connection ); Admin admin = connection.getAdmin(); System.out.println( admin ); TableName tableName1 = TableName.valueOf("hbase:meta"); System.out.println( tableName1 ); System.out.println( admin.tableExists(tableName1) ); Table table = connection.getTable(tableName1); HTableDescriptor a = table.getTableDescriptor(); System.out.println( a ); ## 日志信息: -------------------------- "C:\Program Files\Java\jdk1.8.0_181\bin\java.exe" "-javaagent:C:\Program Files\JetBrains\IntelliJ IDEA 2018.2.3\lib\idea_rt.jar=57269:C:\Program Files\JetBrains\IntelliJ IDEA 2018.2.3\bin" -Dfile.encoding=UTF-8 -classpath "C:\Program Files\Java\jdk1.8.0_181\jre\lib\charsets.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\deploy.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\access-bridge-64.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\cldrdata.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\dnsns.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\jaccess.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\jfxrt.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\localedata.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\nashorn.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\sunec.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\sunjce_provider.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\sunmscapi.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\sunpkcs11.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\zipfs.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\javaws.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\jce.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\jfr.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\jfxswt.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\jsse.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\management-agent.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\plugin.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\resources.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\rt.jar;C:\Users\Administrator\IdeaProjects\TestHbase\out\production\TestHbase;D:\lib\xz-1.0.jar;D:\lib\asm-3.1.jar;D:\lib\avro-1.7.4.jar;D:\lib\common-1.0.jar;D:\lib\domain-1.0.jar;D:\lib\jfinal-3.1.jar;D:\lib\joni-2.1.2.jar;D:\lib\noggit-0.6.jar;D:\lib\jsch-0.1.42.jar;D:\lib\service-1.0.jar;D:\lib\xmlenc-0.52.jar;D:\lib\druid-1.0.31.jar;D:\lib\guava-12.0.1.jar;D:\lib\jcifs-1.3.17.jar;D:\lib\jetty-6.1.26.jar;D:\lib\jsr305-1.3.9.jar;D:\lib\log4j-1.2.16.jar;D:\lib\cos-26Dec2008.jar;D:\lib\paranamer-2.3.jar;D:\lib\activation-1.1.jar;D:\lib\commons-el-1.0.jar;D:\lib\commons-io-2.4.jar;D:\lib\httpcore-4.4.4.jar;D:\lib\httpmime-4.4.1.jar;D:\lib\jaxb-api-2.2.2.jar;D:\lib\jcodings-1.0.8.jar;D:\lib\jsp-2.1-6.1.14.jar;D:\lib\stax-api-1.0-2.jar;D:\lib\cglib-nodep-3.1.jar;D:\lib\commons-cli-1.2.jar;D:\lib\commons-net-3.1.jar;D:\lib\disruptor-3.3.0.jar;D:\lib\fastjson-1.2.37.jar;D:\lib\jersey-core-1.9.jar;D:\lib\servlet-api-2.4.jar;D:\lib\slf4j-api-1.6.6.jar;D:\lib\stax2-api-3.1.4.jar;D:\lib\zookeeper-3.4.6.jar;D:\lib\commons-lang-2.6.jar;D:\lib\commons-math-2.2.jar;D:\lib\httpclient-4.5.2.jar;D:\lib\solr-solrj-6.1.0.jar;D:\lib\freemarker-2.3.23.jar;D:\lib\hadoop-auth-2.5.1.jar;D:\lib\hadoop-hdfs-2.7.4.jar;D:\lib\jersey-server-1.9.jar;D:\lib\jetty-util-6.1.26.jar;D:\lib\netty-3.6.2.Final.jar;D:\lib\api-util-1.0.0-M20.jar;D:\lib\commons-codec-1.11.jar;D:\lib\hbase-client-1.2.3.jar;D:\lib\hbase-common-1.2.3.jar;D:\lib\hbase-server-1.4.0.jar;D:\lib\jsp-api-2.1-6.1.14.jar;D:\lib\leveldbjni-all-1.8.jar;D:\lib\metrics-core-2.2.0.jar;D:\lib\metrics-core-3.1.2.jar;D:\lib\commons-logging-1.2.jar;D:\lib\commons-math3-3.1.1.jar;D:\lib\hadoop-client-2.7.4.jar;D:\lib\hadoop-common-2.5.1.jar;D:\lib\hbase-metrics-1.4.0.jar;D:\lib\jamon-runtime-2.4.1.jar;D:\lib\protobuf-java-2.5.0.jar;D:\lib\slf4j-log4j12-1.6.6.jar;D:\lib\snappy-java-1.0.4.1.jar;D:\lib\commons-digester-1.8.jar;D:\lib\hbase-protocol-1.2.3.jar;D:\lib\jackson-jaxrs-1.9.13.jar;D:\lib\jcl-over-slf4j-1.7.7.jar;D:\lib\commons-daemon-1.0.13.jar;D:\lib\hadoop-yarn-api-2.7.4.jar;D:\lib\hbase-procedure-1.4.0.jar;D:\lib\jasper-runtime-5.5.23.jar;D:\lib\api-asn1-api-1.0.0-M20.jar;D:\lib\commons-compress-1.4.1.jar;D:\lib\commons-httpclient-3.1.jar;D:\lib\jasper-compiler-5.5.23.jar;D:\lib\jetty-sslengine-6.1.26.jar;D:\lib\netty-all-4.0.23.Final.jar;D:\lib\servlet-api-2.5-6.1.14.jar;D:\lib\apacheds-i18n-2.0.0-M15.jar;D:\lib\commons-beanutils-1.7.0.jar;D:\lib\hbase-annotations-1.2.3.jar;D:\lib\hbase-metrics-api-1.4.0.jar;D:\lib\hbase-prefix-tree-1.4.0.jar;D:\lib\jackson-core-asl-1.9.13.jar;D:\lib\woodstox-core-asl-4.4.1.jar;D:\lib\hadoop-annotations-2.5.1.jar;D:\lib\hadoop-yarn-client-2.7.4.jar;D:\lib\hadoop-yarn-common-2.5.1.jar;D:\lib\hbase-common-1.4.0-tests.jar;D:\lib\commons-collections-3.2.2.jar;D:\lib\commons-configuration-1.6.jar;D:\lib\hbase-hadoop-compat-1.4.0.jar;D:\lib\jackson-mapper-asl-1.9.13.jar;D:\lib\hbase-hadoop2-compat-1.4.0.jar;D:\lib\mysql-connector-java-5.1.38.jar;D:\lib\commons-beanutils-core-1.8.0.jar;D:\lib\findbugs-annotations-1.3.9-1.jar;D:\lib\htrace-core-3.1.0-incubating.jar;D:\lib\hadoop-yarn-server-common-2.7.4.jar;D:\lib\apacheds-kerberos-codec-2.0.0-M15.jar;D:\lib\hadoop-mapreduce-client-app-2.7.4.jar;D:\lib\hadoop-mapreduce-client-core-2.5.1.jar;D:\lib\hadoop-mapreduce-client-common-2.7.4.jar;D:\lib\hadoop-mapreduce-client-shuffle-2.7.4.jar;D:\lib\hadoop-mapreduce-client-jobclient-2.7.4.jar" TestHbase [DEBUG] [10:23:53] org.apache.hadoop.security.Groups - Creating new Groups object [DEBUG] [10:23:53] org.apache.hadoop.util.NativeCodeLoader - Trying to load the custom-built native-hadoop library... [DEBUG] [10:23:53] org.apache.hadoop.util.NativeCodeLoader - Failed to load native-hadoop with error: java.lang.UnsatisfiedLinkError: no hadoop in java.library.path [DEBUG] [10:23:53] org.apache.hadoop.util.NativeCodeLoader - java.library.path=C:\Program Files\Java\jdk1.8.0_181\bin;C:\Windows\Sun\Java\bin;C:\Windows\system32;C:\Windows;C:\Program Files (x86)\Common Files\Oracle\Java\javapath;C:\ProgramData\Oracle\Java\javapath;C:\Program Files\Java\jdk1.7.0_51\bin;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem;C:\Windows\System32\WindowsPowerShell\v1.0\;. [WARN ] [10:23:53] org.apache.hadoop.util.NativeCodeLoader - Unable to load native-hadoop library for your platform... using builtin-java classes where applicable [DEBUG] [10:23:53] org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback - Falling back to shell based [DEBUG] [10:23:53] org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback - Group mapping impl=org.apache.hadoop.security.ShellBasedUnixGroupsMapping [DEBUG] [10:23:53] org.apache.hadoop.security.Groups - Group mapping impl=org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback; cacheTimeout=300000; warningDeltaMs=5000 [DEBUG] [10:23:53] org.apache.hadoop.metrics2.lib.MutableMetricsFactory - field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginSuccess with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, sampleName=Ops, always=false, type=DEFAULT, value=[Rate of successful kerberos logins and latency (milliseconds)], valueName=Time) [DEBUG] [10:23:53] org.apache.hadoop.metrics2.lib.MutableMetricsFactory - field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.loginFailure with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, sampleName=Ops, always=false, type=DEFAULT, value=[Rate of failed kerberos logins and latency (milliseconds)], valueName=Time) [DEBUG] [10:23:53] org.apache.hadoop.metrics2.lib.MutableMetricsFactory - field org.apache.hadoop.metrics2.lib.MutableRate org.apache.hadoop.security.UserGroupInformation$UgiMetrics.getGroups with annotation @org.apache.hadoop.metrics2.annotation.Metric(about=, sampleName=Ops, always=false, type=DEFAULT, value=[GetGroups], valueName=Time) [DEBUG] [10:23:53] org.apache.hadoop.metrics2.impl.MetricsSystemImpl - UgiMetrics, User and group related metrics [DEBUG] [10:23:54] org.apache.hadoop.security.authentication.util.KerberosName - Kerberos krb5 configuration not found, setting default realm to empty [DEBUG] [10:23:54] org.apache.hadoop.security.UserGroupInformation - hadoop login [DEBUG] [10:23:54] org.apache.hadoop.security.UserGroupInformation - hadoop login commit [DEBUG] [10:23:54] org.apache.hadoop.security.UserGroupInformation - using local user:NTUserPrincipal: admin [DEBUG] [10:23:54] org.apache.hadoop.security.UserGroupInformation - UGI loginUser:admin (auth:SIMPLE) [INFO ] [10:23:55] org.apache.hadoop.hbase.zookeeper.RecoverableZooKeeper - Process identifier=hconnection-0x17d0685f connecting to ZooKeeper ensemble=192.168.0.25:2181 [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:zookeeper.version=3.4.6-1569965, built on 02/20/2014 09:09 GMT [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:host.name=WIN-SSJFMH6ELVT [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:java.version=1.8.0_181 [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:java.vendor=Oracle Corporation [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:java.home=C:\Program Files\Java\jdk1.8.0_181\jre [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:java.class.path=C:\Program Files\Java\jdk1.8.0_181\jre\lib\charsets.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\deploy.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\access-bridge-64.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\cldrdata.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\dnsns.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\jaccess.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\jfxrt.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\localedata.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\nashorn.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\sunec.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\sunjce_provider.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\sunmscapi.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\sunpkcs11.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\ext\zipfs.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\javaws.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\jce.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\jfr.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\jfxswt.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\jsse.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\management-agent.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\plugin.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\resources.jar;C:\Program Files\Java\jdk1.8.0_181\jre\lib\rt.jar;C:\Users\Administrator\IdeaProjects\TestHbase\out\production\TestHbase;D:\lib\xz-1.0.jar;D:\lib\asm-3.1.jar;D:\lib\avro-1.7.4.jar;D:\lib\common-1.0.jar;D:\lib\domain-1.0.jar;D:\lib\jfinal-3.1.jar;D:\lib\joni-2.1.2.jar;D:\lib\noggit-0.6.jar;D:\lib\jsch-0.1.42.jar;D:\lib\service-1.0.jar;D:\lib\xmlenc-0.52.jar;D:\lib\druid-1.0.31.jar;D:\lib\guava-12.0.1.jar;D:\lib\jcifs-1.3.17.jar;D:\lib\jetty-6.1.26.jar;D:\lib\jsr305-1.3.9.jar;D:\lib\log4j-1.2.16.jar;D:\lib\cos-26Dec2008.jar;D:\lib\paranamer-2.3.jar;D:\lib\activation-1.1.jar;D:\lib\commons-el-1.0.jar;D:\lib\commons-io-2.4.jar;D:\lib\httpcore-4.4.4.jar;D:\lib\httpmime-4.4.1.jar;D:\lib\jaxb-api-2.2.2.jar;D:\lib\jcodings-1.0.8.jar;D:\lib\jsp-2.1-6.1.14.jar;D:\lib\stax-api-1.0-2.jar;D:\lib\cglib-nodep-3.1.jar;D:\lib\commons-cli-1.2.jar;D:\lib\commons-net-3.1.jar;D:\lib\disruptor-3.3.0.jar;D:\lib\fastjson-1.2.37.jar;D:\lib\jersey-core-1.9.jar;D:\lib\servlet-api-2.4.jar;D:\lib\slf4j-api-1.6.6.jar;D:\lib\stax2-api-3.1.4.jar;D:\lib\zookeeper-3.4.6.jar;D:\lib\commons-lang-2.6.jar;D:\lib\commons-math-2.2.jar;D:\lib\httpclient-4.5.2.jar;D:\lib\solr-solrj-6.1.0.jar;D:\lib\freemarker-2.3.23.jar;D:\lib\hadoop-auth-2.5.1.jar;D:\lib\hadoop-hdfs-2.7.4.jar;D:\lib\jersey-server-1.9.jar;D:\lib\jetty-util-6.1.26.jar;D:\lib\netty-3.6.2.Final.jar;D:\lib\api-util-1.0.0-M20.jar;D:\lib\commons-codec-1.11.jar;D:\lib\hbase-client-1.2.3.jar;D:\lib\hbase-common-1.2.3.jar;D:\lib\hbase-server-1.4.0.jar;D:\lib\jsp-api-2.1-6.1.14.jar;D:\lib\leveldbjni-all-1.8.jar;D:\lib\metrics-core-2.2.0.jar;D:\lib\metrics-core-3.1.2.jar;D:\lib\commons-logging-1.2.jar;D:\lib\commons-math3-3.1.1.jar;D:\lib\hadoop-client-2.7.4.jar;D:\lib\hadoop-common-2.5.1.jar;D:\lib\hbase-metrics-1.4.0.jar;D:\lib\jamon-runtime-2.4.1.jar;D:\lib\protobuf-java-2.5.0.jar;D:\lib\slf4j-log4j12-1.6.6.jar;D:\lib\snappy-java-1.0.4.1.jar;D:\lib\commons-digester-1.8.jar;D:\lib\hbase-protocol-1.2.3.jar;D:\lib\jackson-jaxrs-1.9.13.jar;D:\lib\jcl-over-slf4j-1.7.7.jar;D:\lib\commons-daemon-1.0.13.jar;D:\lib\hadoop-yarn-api-2.7.4.jar;D:\lib\hbase-procedure-1.4.0.jar;D:\lib\jasper-runtime-5.5.23.jar;D:\lib\api-asn1-api-1.0.0-M20.jar;D:\lib\commons-compress-1.4.1.jar;D:\lib\commons-httpclient-3.1.jar;D:\lib\jasper-compiler-5.5.23.jar;D:\lib\jetty-sslengine-6.1.26.jar;D:\lib\netty-all-4.0.23.Final.jar;D:\lib\servlet-api-2.5-6.1.14.jar;D:\lib\apacheds-i18n-2.0.0-M15.jar;D:\lib\commons-beanutils-1.7.0.jar;D:\lib\hbase-annotations-1.2.3.jar;D:\lib\hbase-metrics-api-1.4.0.jar;D:\lib\hbase-prefix-tree-1.4.0.jar;D:\lib\jackson-core-asl-1.9.13.jar;D:\lib\woodstox-core-asl-4.4.1.jar;D:\lib\hadoop-annotations-2.5.1.jar;D:\lib\hadoop-yarn-client-2.7.4.jar;D:\lib\hadoop-yarn-common-2.5.1.jar;D:\lib\hbase-common-1.4.0-tests.jar;D:\lib\commons-collections-3.2.2.jar;D:\lib\commons-configuration-1.6.jar;D:\lib\hbase-hadoop-compat-1.4.0.jar;D:\lib\jackson-mapper-asl-1.9.13.jar;D:\lib\hbase-hadoop2-compat-1.4.0.jar;D:\lib\mysql-connector-java-5.1.38.jar;D:\lib\commons-beanutils-core-1.8.0.jar;D:\lib\findbugs-annotations-1.3.9-1.jar;D:\lib\htrace-core-3.1.0-incubating.jar;D:\lib\hadoop-yarn-server-common-2.7.4.jar;D:\lib\apacheds-kerberos-codec-2.0.0-M15.jar;D:\lib\hadoop-mapreduce-client-app-2.7.4.jar;D:\lib\hadoop-mapreduce-client-core-2.5.1.jar;D:\lib\hadoop-mapreduce-client-common-2.7.4.jar;D:\lib\hadoop-mapreduce-client-shuffle-2.7.4.jar;D:\lib\hadoop-mapreduce-client-jobclient-2.7.4.jar;C:\Program Files\JetBrains\IntelliJ IDEA 2018.2.3\lib\idea_rt.jar [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:java.library.path=C:\Program Files\Java\jdk1.8.0_181\bin;C:\Windows\Sun\Java\bin;C:\Windows\system32;C:\Windows;C:\Program Files (x86)\Common Files\Oracle\Java\javapath;C:\ProgramData\Oracle\Java\javapath;C:\Program Files\Java\jdk1.7.0_51\bin;C:\Windows\system32;C:\Windows;C:\Windows\System32\Wbem;C:\Windows\System32\WindowsPowerShell\v1.0\;. [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:java.io.tmpdir=C:\Users\ADMINI~1\AppData\Local\Temp\1\ [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:java.compiler=<NA> [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:os.name=Windows Server 2008 R2 [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:os.arch=amd64 [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:os.version=6.1 [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:user.name=admin [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:user.home=C:\Users\Administrator [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Client environment:user.dir=C:\Users\Administrator\IdeaProjects\TestHbase [INFO ] [10:23:55] org.apache.zookeeper.ZooKeeper - Initiating client connection, connectString=192.168.0.25:2181 sessionTimeout=90000 watcher=hconnection-0x17d0685f0x0, quorum=192.168.0.25:2181, baseZNode=/hbase [DEBUG] [10:23:55] org.apache.zookeeper.ClientCnxn - zookeeper.disableAutoWatchReset is false [INFO ] [10:23:55] org.apache.zookeeper.ClientCnxn - Opening socket connection to server 192.168.0.25/192.168.0.25:2181. Will not attempt to authenticate using SASL (unknown error) [INFO ] [10:23:55] org.apache.zookeeper.ClientCnxn - Socket connection established to 192.168.0.25/192.168.0.25:2181, initiating session [DEBUG] [10:23:55] org.apache.zookeeper.ClientCnxn - Session establishment request sent on 192.168.0.25/192.168.0.25:2181 [INFO ] [10:23:56] org.apache.zookeeper.ClientCnxn - Session establishment complete on server 192.168.0.25/192.168.0.25:2181, sessionid = 0x165a4d303830022, negotiated timeout = 40000 [DEBUG] [10:23:56] org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher - hconnection-0x17d0685f0x0, quorum=192.168.0.25:2181, baseZNode=/hbase Received ZooKeeper Event, type=None, state=SyncConnected, path=null [DEBUG] [10:23:56] org.apache.hadoop.hbase.zookeeper.ZooKeeperWatcher - hconnection-0x17d0685f-0x165a4d303830022 connected [DEBUG] [10:23:56] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 1,3 replyHeader:: 1,184683596486,0 request:: '/hbase/hbaseid,F response:: s{4294967310,184683593733,1421399758722,1536068918058,938,0,0,0,60,0,4294967310} [DEBUG] [10:23:56] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 2,4 replyHeader:: 2,184683596486,0 request:: '/hbase/hbaseid,F response:: #ffffffff000133135313538404d61737465722e4861646f6f7066626162613563302d313737332d343731342d613630622d643233626232623865373831,s{4294967310,184683593733,1421399758722,1536068918058,938,0,0,0,60,0,4294967310} [DEBUG] [10:23:56] org.apache.hadoop.hbase.ipc.AbstractRpcClient - Codec=org.apache.hadoop.hbase.codec.KeyValueCodec@7c1e2a9e, compressor=null, tcpKeepAlive=true, tcpNoDelay=true, connectTO=10000, readTO=20000, writeTO=60000, minIdleTimeBeforeClose=120000, maxRetries=0, fallbackAllowed=false, bind address=null hconnection-0x17d0685f org.apache.hadoop.hbase.client.HBaseAdmin@272ed83b hbase:meta true [DEBUG] [10:23:56] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 3,3 replyHeader:: 3,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:23:56] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 4,4 replyHeader:: 4,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:23:56] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:23:56] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [DEBUG] [10:23:57] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 5,3 replyHeader:: 5,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:23:57] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 6,4 replyHeader:: 6,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:23:57] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:23:57] org.apache.hadoop.hbase.ipc.RpcClientImpl - Not trying to connect to Master.Hadoop/192.168.0.25:60000 this server is in the failed servers list [DEBUG] [10:23:57] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 7,3 replyHeader:: 7,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:23:57] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 8,4 replyHeader:: 8,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:23:57] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:23:57] org.apache.hadoop.hbase.ipc.RpcClientImpl - Not trying to connect to Master.Hadoop/192.168.0.25:60000 this server is in the failed servers list [DEBUG] [10:23:57] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 9,3 replyHeader:: 9,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:23:57] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 10,4 replyHeader:: 10,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:23:57] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:23:57] org.apache.hadoop.hbase.ipc.RpcClientImpl - Not trying to connect to Master.Hadoop/192.168.0.25:60000 this server is in the failed servers list [DEBUG] [10:23:58] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 11,3 replyHeader:: 11,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:23:58] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 12,4 replyHeader:: 12,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:23:58] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:23:58] org.apache.hadoop.hbase.ipc.RpcClientImpl - Not trying to connect to Master.Hadoop/192.168.0.25:60000 this server is in the failed servers list [DEBUG] [10:23:59] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 13,3 replyHeader:: 13,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:23:59] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 14,4 replyHeader:: 14,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:23:59] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:23:59] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [DEBUG] [10:24:01] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 15,3 replyHeader:: 15,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:24:01] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 16,4 replyHeader:: 16,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:24:01] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:24:01] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [DEBUG] [10:24:05] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 17,3 replyHeader:: 17,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:24:05] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 18,4 replyHeader:: 18,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:24:05] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:24:05] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [DEBUG] [10:24:15] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 19,3 replyHeader:: 19,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:24:15] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 11ms [DEBUG] [10:24:15] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 20,4 replyHeader:: 20,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:24:15] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:24:15] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [DEBUG] [10:24:25] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 21,3 replyHeader:: 21,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:24:25] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 3ms [DEBUG] [10:24:25] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 22,4 replyHeader:: 22,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:24:25] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:24:25] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [DEBUG] [10:24:35] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 23,3 replyHeader:: 23,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:24:35] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 1ms [DEBUG] [10:24:35] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 24,4 replyHeader:: 24,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:24:35] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:24:35] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [INFO ] [10:24:35] org.apache.hadoop.hbase.client.RpcRetryingCaller - Call exception, tries=10, retries=35, started=38894 ms ago, cancelled=false, msg= [DEBUG] [10:24:45] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 25,3 replyHeader:: 25,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:24:45] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 1ms [DEBUG] [10:24:45] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 26,4 replyHeader:: 26,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:24:45] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:24:45] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [INFO ] [10:24:45] org.apache.hadoop.hbase.client.RpcRetryingCaller - Call exception, tries=11, retries=35, started=48956 ms ago, cancelled=false, msg= [DEBUG] [10:24:58] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 1ms [DEBUG] [10:25:05] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 27,3 replyHeader:: 27,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:25:05] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 28,4 replyHeader:: 28,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:25:05] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:25:05] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [INFO ] [10:25:05] org.apache.hadoop.hbase.client.RpcRetryingCaller - Call exception, tries=12, retries=35, started=69113 ms ago, cancelled=false, msg= [DEBUG] [10:25:19] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 1ms [DEBUG] [10:25:25] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 29,3 replyHeader:: 29,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:25:25] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 30,4 replyHeader:: 30,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:25:25] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:25:25] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [INFO ] [10:25:25] org.apache.hadoop.hbase.client.RpcRetryingCaller - Call exception, tries=13, retries=35, started=89176 ms ago, cancelled=false, msg= [DEBUG] [10:25:39] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 0ms [DEBUG] [10:25:45] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 31,3 replyHeader:: 31,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:25:45] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 32,4 replyHeader:: 32,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:25:45] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:25:45] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [INFO ] [10:25:45] org.apache.hadoop.hbase.client.RpcRetryingCaller - Call exception, tries=14, retries=35, started=109299 ms ago, cancelled=false, msg= [DEBUG] [10:25:59] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 0ms [DEBUG] [10:26:05] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 33,3 replyHeader:: 33,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:26:05] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 34,4 replyHeader:: 34,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:26:05] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:26:05] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [INFO ] [10:26:05] org.apache.hadoop.hbase.client.RpcRetryingCaller - Call exception, tries=15, retries=35, started=129364 ms ago, cancelled=false, msg= [DEBUG] [10:26:19] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 0ms [DEBUG] [10:26:26] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 35,3 replyHeader:: 35,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:26:26] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 36,4 replyHeader:: 36,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:26:26] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:26:26] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [INFO ] [10:26:26] org.apache.hadoop.hbase.client.RpcRetryingCaller - Call exception, tries=16, retries=35, started=149567 ms ago, cancelled=false, msg= [DEBUG] [10:26:39] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 0ms [DEBUG] [10:26:46] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 37,3 replyHeader:: 37,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:26:46] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 38,4 replyHeader:: 38,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:26:46] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:26:46] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [INFO ] [10:26:46] org.apache.hadoop.hbase.client.RpcRetryingCaller - Call exception, tries=17, retries=35, started=169742 ms ago, cancelled=false, msg= [DEBUG] [10:26:59] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 1ms [DEBUG] [10:27:06] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 39,3 replyHeader:: 39,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:27:06] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 40,4 replyHeader:: 40,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:27:06] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:27:06] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [INFO ] [10:27:06] org.apache.hadoop.hbase.client.RpcRetryingCaller - Call exception, tries=18, retries=35, started=189961 ms ago, cancelled=false, msg= [DEBUG] [10:27:19] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 0ms [DEBUG] [10:27:26] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 41,3 replyHeader:: 41,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:27:26] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 42,4 replyHeader:: 42,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:27:26] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:27:26] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [INFO ] [10:27:26] org.apache.hadoop.hbase.client.RpcRetryingCaller - Call exception, tries=19, retries=35, started=210070 ms ago, cancelled=false, msg= [DEBUG] [10:27:39] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 2ms [DEBUG] [10:27:46] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 43,3 replyHeader:: 43,184683596486,0 request:: '/hbase,F response:: s{4294967298,4294967298,1421399756338,1421399756338,0,1956,0,0,0,12,184683593750} [DEBUG] [10:27:46] org.apache.zookeeper.ClientCnxn - Reading reply sessionid:0x165a4d303830022, packet:: clientPath:null serverPath:null finished:false header:: 44,4 replyHeader:: 44,184683596486,0 request:: '/hbase/master,F response:: #ffffffff000133135313538404d61737465722e4861646f6f70004d61737465722e4861646f6f702c36303030302c31353336303638393135363739,s{184683593731,184683593731,1536068917156,1536068917156,0,0,0,316840575388352512,59,0,184683593731} [DEBUG] [10:27:46] org.apache.hadoop.hbase.ipc.RpcClientImpl - Use SIMPLE authentication for service MasterService, sasl=false [DEBUG] [10:27:46] org.apache.hadoop.hbase.ipc.RpcClientImpl - Connecting to Master.Hadoop/192.168.0.25:60000 [INFO ] [10:27:46] org.apache.hadoop.hbase.client.RpcRetryingCaller - Call exception, tries=20, retries=35, started=230074 ms ago, cancelled=false, msg= [DEBUG] [10:28:00] org.apache.zookeeper.ClientCnxn - Got ping response for sessionid: 0x165a4d303830022 after 1ms
java调用Hbase java.net.SocketTimeoutException
hbase菜鸟,工作需要,请大神帮忙。java连接hbase出现的错误error如下: log4j:WARN No appenders could be found for logger (org.apache.hadoop.metrics2.lib.MutableMetricsFactory). log4j:WARN Please initialize the log4j system properly. log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info. org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=36, exceptions: Thu Apr 07 09:43:39 CST 2016, null, java.net.SocketTimeoutException: callTimeout=60000, callDuration=80156: row 'gpsinfo,,' on table 'hbase:meta' at region=hbase:meta,,1.1588230740, hostname=zpu-storage3,16020,1443175301611, seqNum=0 at org.apache.hadoop.hbase.client.RpcRetryingCallerWithReadReplicas.throwEnrichedException(RpcRetryingCallerWithReadReplicas.java:271) at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:207) at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas.call(ScannerCallableWithReplicas.java:60) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:200) at org.apache.hadoop.hbase.client.ClientScanner.call(ClientScanner.java:320) at org.apache.hadoop.hbase.client.ClientScanner.loadCache(ClientScanner.java:401) at org.apache.hadoop.hbase.client.ClientScanner.next(ClientScanner.java:364) at org.apache.hadoop.hbase.MetaTableAccessor.fullScan(MetaTableAccessor.java:604) at org.apache.hadoop.hbase.MetaTableAccessor.tableExists(MetaTableAccessor.java:366) at org.apache.hadoop.hbase.client.HBaseAdmin.tableExists(HBaseAdmin.java:392) at org.apache.hadoop.hbase.client.HBaseAdmin.tableExists(HBaseAdmin.java:402) at Test.main(Test.java:19) Caused by: java.net.SocketTimeoutException: callTimeout=60000, callDuration=80156: row 'gpsinfo,,' on table 'hbase:meta' at region=hbase:meta,,1.1588230740, hostname=zpu-storage3,16020,1443175301611, seqNum=0 at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:159) at org.apache.hadoop.hbase.client.ResultBoundedCompletionService$QueueingFuture.run(ResultBoundedCompletionService.java:65) at java.util.concurrent.ThreadPoolExecutor.runWorker(Unknown Source) at java.util.concurrent.ThreadPoolExecutor$Worker.run(Unknown Source) at java.lang.Thread.run(Unknown Source) Caused by: java.net.UnknownHostException: unknown host: zpu-storage3 at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.<init>(RpcClientImpl.java:303) at org.apache.hadoop.hbase.ipc.RpcClientImpl.createConnection(RpcClientImpl.java:133) at org.apache.hadoop.hbase.ipc.RpcClientImpl.getConnection(RpcClientImpl.java:1283) at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1191) at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:222) at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:323) at org.apache.hadoop.hbase.protobuf.generated.ClientProtos$ClientService$BlockingStub.scan(ClientProtos.java:32831) at org.apache.hadoop.hbase.client.ScannerCallable.openScanner(ScannerCallable.java:373) at org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:200) at org.apache.hadoop.hbase.client.ScannerCallable.call(ScannerCallable.java:62) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithoutRetries(RpcRetryingCaller.java:200) at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas$RetryingRPC.call(ScannerCallableWithReplicas.java:360) at org.apache.hadoop.hbase.client.ScannerCallableWithReplicas$RetryingRPC.call(ScannerCallableWithReplicas.java:334) at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126) ... 4 more
java 如何访问远程 linux 服务器下利用docker容器建立的hbase集群
在远程linux服务器上利用docker搭建伪分布式集群的Hbase,如何在本地windows主机利用java来访问并进行操作。 ``` conf = HBaseConfiguration.create(); //我认为这里应该是我服务器的ip,因为容器的ip并没有映射到windows中 conf.set("hbase.zookeeper.quorum","192.168.10.168"); //2281是master容器zk2181端口的映射 conf.set("hbase.zookeeper.property.clientPort", "2281"); conn= ConnectionFactory.createConnection(conf); admin=conn.getAdmin(); System.out.println("connection success"); Table mk=conn.getTable(TableName.valueOf("user")); Get g = new Get(Bytes.toBytes("1234")); Result result = mk.get(g); ``` 如果这样子操作的话,运行到Result报缺少master主机的错误
HBase存储图片细节问题
图片读成二进制是数组形式,Hbase在JAVA API中的put在每行中直接put数组会报错,请问该怎样解决 希望大神提供示例代码
有用java 编的程序 调用 hbase 接口的吗?
有用java 编的程序 调用 hbase 接口的吗? 最好具体一些,对这方面不太懂
想用java操作hbase,遇到了些问题
新手 在网上看教程,用java在hbase中创建表,输入代码之后问题很多 package hbase; import java.io.IOException; import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.hbase.HBaseConfiguration; import org.apache.hadoop.hbase.HColumnDescriptor; import org.apache.hadoop.hbase.HTableDescriptor; import org.apache.hadoop.hbase.MasterNotRunningException; import org.apache.hadoop.hbase.TableName; import org.apache.hadoop.hbase.ZooKeeperConnectionException; import org.apache.hadoop.hbase.client.HBaseAdmin; public class hbasetest1 { private static final String TABLE_NAME = "test1"; private static final String FAMILY_NAME = "cf1"; public static void main(String[] args) throws Exception { // TODO Auto-generated method stub Configuration conf = HBaseConfiguration.create(); conf.set("hbase.rootdir","hdfs://master:9000/hbase"); conf.set("hbase.zookeeper.quorum","master:2181"); HBaseAdmin hBaseAdmin = new HBaseAdmin(conf); HTableDescriptor htableDescriptor = new HTableDescriptor(TABLE_NAME); htableDescriptor.addFamily(new HColumnDescriptor(FAMILY_NAME)); hBaseAdmin.createTable(htableDescriptor); } } 直接编译运行的话: log4j:WARN No appenders could be found for logger (org.apache.hadoop.security.Groups). log4j:WARN Please initialize the log4j system properly. log4j:WARN See http://logging.apache.org/log4j/1.2/faq.html#noconfig for more info. Exception in thread "main" org.apache.hadoop.hbase.TableExistsException: test1 at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method) at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57) at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45) at java.lang.reflect.Constructor.newInstance(Constructor.java:526) at org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106) at org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:95) at org.apache.hadoop.hbase.util.ForeignExceptionUtil.toIOException(ForeignExceptionUtil.java:45) at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.convertResult(HBaseAdmin.java:4617) at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4575) at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4508) at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:674) at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:604) at hbase.hbasetest1.main(hbasetest1.java:29) Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.hbase.TableExistsException): test1 at org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.prepareCreate(CreateTableProcedure.java:287) at org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.executeFromState(CreateTableProcedure.java:108) at org.apache.hadoop.hbase.master.procedure.CreateTableProcedure.executeFromState(CreateTableProcedure.java:58) at org.apache.hadoop.hbase.procedure2.StateMachineProcedure.execute(StateMachineProcedure.java:119) at org.apache.hadoop.hbase.procedure2.Procedure.doExecute(Procedure.java:465) at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execProcedure(ProcedureExecutor.java:1061) at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:856) at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.execLoop(ProcedureExecutor.java:809) at org.apache.hadoop.hbase.procedure2.ProcedureExecutor.access$400(ProcedureExecutor.java:75) at org.apache.hadoop.hbase.procedure2.ProcedureExecutor$2.run(ProcedureExecutor.java:495) ant运行: Buildfile: /home/celine/workspace/test-hbase/build.xml clean: [delete] Deleting directory /home/celine/workspace/test-hbase/dist [delete] Deleting directory /home/celine/workspace/test-hbase/classes [echo] 清理完毕 init: [mkdir] Created dir: /home/celine/workspace/test-hbase/classes [mkdir] Created dir: /home/celine/workspace/test-hbase/dist compile: [javac] Compiling 1 source file to /home/celine/workspace/test-hbase/classes [javac] Note: /home/celine/workspace/test-hbase/src/hbase/hbasetest1.java uses or overrides a deprecated API. [javac] Note: Recompile with -Xlint:deprecation for details. jar: [jar] Building jar: /home/celine/workspace/test-hbase/dist/jar.jar ssh: [scp] Connecting to 192.168.124.128:22 [scp] done. sshexec: [sshexec] Connecting to 192.168.124.128:22 [sshexec] cmd : source /etc/profile; hadoop jar ~/jar.jar hbase.hbasetest1 [sshexec] Exception in thread "main" [sshexec] java.lang.NoClassDefFoundError: org/apache/hadoop/hbase/HBaseConfiguration [sshexec] at hbase.hbasetest1.main(Unknown Source) [sshexec] at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) [sshexec] at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57) [sshexec] at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) [sshexec] at java.lang.reflect.Method.invoke(Method.java:606) [sshexec] at org.apache.hadoop.util.RunJar.run(RunJar.java:221) [sshexec] at org.apache.hadoop.util.RunJar.main(RunJar.java:136) [sshexec] Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.hbase.HBaseConfiguration [sshexec] at java.net.URLClassLoader$1.run(URLClassLoader.java:366) [sshexec] at java.net.URLClassLoader$1.run(URLClassLoader.java:355) [sshexec] at java.security.AccessController.doPrivileged(Native Method) [sshexec] at java.net.URLClassLoader.findClass(URLClassLoader.java:354) [sshexec] at java.lang.ClassLoader.loadClass(ClassLoader.java:425) [sshexec] at java.lang.ClassLoader.loadClass(ClassLoader.java:358) [sshexec] ... 7 more BUILD FAILED /home/celine/workspace/test-hbase/build.xml:72: Remote command failed with exit status 1 每次都说远程有问题,但是我在build.xml里面配好了相关信息,不知道为什么sshexec会失败 TAT刚刚开始学,好难,好多问题……
JAVA连接Hbase集群
java连接Hbase,代码卡在 HBaseAdmin admin1 = new HBaseAdmin(conf1);处
java多线程读取hbase, 是单例一个client供所有线程使用好呢 ,还是每个线程自己创建线程好呢
java多线程读取hbase, 是单例一个client供所有线程使用好呢 ,还是每个线程自己创建线程好呢 我想明白 ,单例client 与 new 多个client 想比较, 的优缺点和用处
java如何修改hbase表数据
注意:是修改hbase表的数据,不是新增、删除、查询,是修改
求解用java调用hbase表的问题
要用java调用数据库中表,因为表的rowkey值前有一个不定的分区符,查询全表的话又比较慢,有没有什么方法能够设置在设定startrow与stoprow的时候忽略这个分区符
c通过jni调用 HBASE JAVA API
jvm.dll!6db71017() 未知 [下面的框架可能不正确和/或缺失,没有为 jvm.dll 加载符号] jvm.dll!6db71086() 未知 jvm.dll!6db710e0() 未知 jvm.dll!6dde9d1d() 未知 jvm.dll!6deaa65d() 未知 jvm.dll!6ddfd291() 未知 jvm.dll!6dd07d3d() 未知 jvm.dll!6dd11511() 未知 022003e4() 未知 jvm.dll!6dd07cb9() 未知 jvm.dll!6ddfd291() 未知 jvm.dll!6dd07e61() 未知 jvm.dll!6dd07ebb() 未知 > msvcr71.dll!_heap_alloc(unsigned int size) 行 212 C jvm.dll!6de87cb4() 未知 jvm.dll!6ddfcc3c() 未知 kernel32.dll!_HeapFree@12() 未知 msvcr71.dll!free(void * pBlock) 行 103 C msvcr71.dll!_threadstartex(void * ptd) 行 241 C kernel32.dll!@BaseThreadInitThunk@12() 未知 ntdll.dll!___RtlUserThreadStart@8() 未知 ntdll.dll!__RtlUserThreadStart@8() 未知 每次都是这个问题,运行着就中断,忽略中断,有时也能运行,求解
相见恨晚的超实用网站
相见恨晚的超实用网站 持续更新中。。。
Java学习的正确打开方式
在博主认为,对于入门级学习java的最佳学习方法莫过于视频+博客+书籍+总结,前三者博主将淋漓尽致地挥毫于这篇博客文章中,至于总结在于个人,实际上越到后面你会发现学习的最好方式就是阅读参考官方文档其次就是国内的书籍,博客次之,这又是一个层次了,这里暂时不提后面再谈。博主将为各位入门java保驾护航,各位只管冲鸭!!!上天是公平的,只要不辜负时间,时间自然不会辜负你。 何谓学习?博主所理解的学习,它是一个过程,是一个不断累积、不断沉淀、不断总结、善于传达自己的个人见解以及乐于分享的过程。
程序员必须掌握的核心算法有哪些?
由于我之前一直强调数据结构以及算法学习的重要性,所以就有一些读者经常问我,数据结构与算法应该要学习到哪个程度呢?,说实话,这个问题我不知道要怎么回答你,主要取决于你想学习到哪些程度,不过针对这个问题,我稍微总结一下我学过的算法知识点,以及我觉得值得学习的算法。这些算法与数据结构的学习大多数是零散的,并没有一本把他们全部覆盖的书籍。下面是我觉得值得学习的一些算法以及数据结构,当然,我也会整理一些看过...
java中的Static、final、Static final各种用法
前言 对Static、final、Static final这几个关键词熟悉又陌生?想说却又不知怎么准确说出口?好的,本篇博客文章将简短概要出他们之间的各自的使用,希望各位要是被你的面试官问到了,也能从容的回答… static 加载:static在类加载时初始化(加载)完成 含义:Static意为静态的,但凡被static 修饰说明属于类,不属于类的对象。 可修饰:Static 可以修饰 内部类、方...
各大公司在GitHub上开源投入排名分析
基于GitHub的数据进行分析各个公司在开源上的投入排名
大学四年自学走来,这些私藏的实用工具/学习网站我贡献出来了
大学四年,看课本是不可能一直看课本的了,对于学习,特别是自学,善于搜索网上的一些资源来辅助,还是非常有必要的,下面我就把这几年私藏的各种资源,网站贡献出来给你们。主要有:电子书搜索、实用工具、在线视频学习网站、非视频学习网站、软件下载、面试/求职必备网站。 注意:文中提到的所有资源,文末我都给你整理好了,你们只管拿去,如果觉得不错,转发、分享就是最大的支持了。 一、电子书搜索 对于大部分程序员...
linux系列之常用运维命令整理笔录
本博客记录工作中需要的linux运维命令,大学时候开始接触linux,会一些基本操作,可是都没有整理起来,加上是做开发,不做运维,有些命令忘记了,所以现在整理成博客,当然vi,文件操作等就不介绍了,慢慢积累一些其它拓展的命令,博客不定时更新 free -m 其中:m表示兆,也可以用g,注意都要小写 Men:表示物理内存统计 total:表示物理内存总数(total=used+free) use...
python学习方法总结(内附python全套学习资料)
不要再问我python好不好学了 我之前做过半年少儿编程老师,一个小学四年级的小孩子都能在我的教学下独立完成python游戏,植物大战僵尸简单版,如果要肯花时间,接下来的网络开发也不是问题,人工智能也可以学个调包也没啥问题。。。。。所以python真的是想学就一定能学会的!!!! --------------------华丽的分割线-------------------------------- ...
Spring Boot 开发微信公众号后台
Hello 各位小伙伴,松哥今天要和大家聊一个有意思的话题,就是使用 Spring Boot 开发微信公众号后台。 很多小伙伴可能注意到松哥的个人网站(http://www.javaboy.org)前一阵子上线了一个公众号内回复口令解锁网站文章的功能,还有之前就有的公众号内回复口令获取超 2TB 免费视频教程的功能(免费视频教程),这两个都是松哥基于 Spring Boot 来做的,最近松哥打算通...
兼职程序员一般可以从什么平台接私活?
这个问题我进行了系统性的总结,以下将进行言简意赅的说明和渠道提供,希望对各位小猿/小媛们有帮助~ 根据我们的经验,程序员兼职主要分为三种:兼职职位众包、项目整包和自由职业者驻场。 所谓的兼职职位众包,指的是需求方这边有自有工程师配合,只需要某个职位的工程师开发某个模块的项目。比如开发一个 app,后端接口有人开发,但是缺少 iOS 前端开发工程师,那么他们就会发布一个职位招聘前端,来配合公司一...
程序员接私活怎样防止做完了不给钱?
首先跟大家说明一点,我们做 IT 类的外包开发,是非标品开发,所以很有可能在开发过程中会有这样那样的需求修改,而这种需求修改很容易造成扯皮,进而影响到费用支付,甚至出现做完了项目收不到钱的情况。 那么,怎么保证自己的薪酬安全呢? 我们在开工前,一定要做好一些证据方面的准备(也就是“讨薪”的理论依据),这其中最重要的就是需求文档和验收标准。一定要让需求方提供这两个文档资料作为开发的基础。之后开发...
网页实现一个简单的音乐播放器(大佬别看。(⊙﹏⊙))
今天闲着无事,就想写点东西。然后听了下歌,就打算写个播放器。 于是乎用h5 audio的加上js简单的播放器完工了。 演示地点演示 html代码如下` music 这个年纪 七月的风 音乐 ` 然后就是css`*{ margin: 0; padding: 0; text-decoration: none; list-...
Python十大装B语法
Python 是一种代表简单思想的语言,其语法相对简单,很容易上手。不过,如果就此小视 Python 语法的精妙和深邃,那就大错特错了。本文精心筛选了最能展现 Python 语法之精妙的十个知识点,并附上详细的实例代码。如能在实战中融会贯通、灵活使用,必将使代码更为精炼、高效,同时也会极大提升代码B格,使之看上去更老练,读起来更优雅。
数据库优化 - SQL优化
以实际SQL入手,带你一步一步走上SQL优化之路!
2019年11月中国大陆编程语言排行榜
2019年11月2日,我统计了某招聘网站,获得有效程序员招聘数据9万条。针对招聘信息,提取编程语言关键字,并统计如下: 编程语言比例 rank pl_ percentage 1 java 33.62% 2 cpp 16.42% 3 c_sharp 12.82% 4 javascript 12.31% 5 python 7.93% 6 go 7.25% 7 p...
C++知识点 —— 整合(持续更新中)
本文记录自己在自学C++过程中不同于C的一些知识点,适合于有C语言基础的同学阅读。如果纰漏,欢迎回复指正 目录 第一部分 基础知识 一、HelloWorld与命名空间 二、引用和引用参数 2.1引用的定义 2.2 将引用用作函数参数 2.3 将引用用于类对象 2.4 引用和继承 2.5 何时使用引用参数 2.6 引用和指针的区别 三、内联函数 四、默认参数的...
《奇巧淫技》系列-python!!每天早上八点自动发送天气预报邮件到QQ邮箱
将代码部署服务器,每日早上定时获取到天气数据,并发送到邮箱。 也可以说是一个小型人工智障。 知识可以运用在不同地方,不一定非是天气预报。
js知识点汇总
1.本门课的作用(JavaScript的作用)所有基于Web的程序开发基础 2.一种计算机客户端脚本语言,主要在Web浏览器解释执行。 3.浏览器中Javascript,用于与用户交互,以及实现页面中各种动态特效 4.在HTML文件中,任何JavaScript代码必须放到&lt;script&gt;标签中 5.把JavaScript代码放到外部文件中,在HTML代码中直接引入该文件(外部文...
经典算法(5)杨辉三角
杨辉三角 是经典算法,这篇博客对它的算法思想进行了讲解,并有完整的代码实现。
Python实例大全(基于Python3.7.4)
博客说明: 这是自己写的有关python语言的一篇综合博客。 只作为知识广度和编程技巧学习,不过于追究学习深度,点到即止、会用即可。 主要是基础语句,如三大控制语句(顺序、分支、循环),随机数的生成,数据类型的区分和使用; 也会涉及常用的算法和数据结构,以及面试题相关经验; 主体部分是针对python的数据挖掘和数据分析,主要先攻爬虫方向:正则表达式匹配,常用数据清洗办法,scrapy及其他爬虫框架,数据存储方式及其实现; 最后还会粗略涉及人工智能领域,玩转大数据与云计算、进行相关的预测和分析。
面试官:你连RESTful都不知道我怎么敢要你?
干货,2019 RESTful最贱实践
论文读不懂怎么办?
王树义读完需要18分钟速读仅需6分钟悄悄告诉你几个窍门。1 痛苦做科研,不能不读论文。但是,我见过不少研究生,论文都读得愁眉苦脸的。这其中,自然有因为拖延的关系。例如教授布置了2周后讨论论文,你原本可以每天抽出时间,慢慢消化。可你非得拖到截止日期的前一天晚上才打开看,希望毕其功于一役……这显然是态度问题。但是,不可忽略的,有些同学,却真的不是态度问题,而是方法不当。他们喜欢跟论文死磕。读不懂,...
用Python解决数据结构与算法问题(三):线性数据结构之栈
python学习之路 - 从入门到精通到大师 文章目录[python学习之路 - 从入门到精通到大师](https://blog.csdn.net/TeFuirnever/article/details/90017382)〇、写在前面3.1.目标3.2.什么是线性数据结构3.3.什么是栈3.4.栈的抽象数据类型3.5.Python实现栈3.6.简单括号匹配3.7.符号匹配3.8.十进制转换成二进制...
刷了几千道算法题,这些我私藏的刷题网站都在这里了!
遥想当年,机缘巧合入了 ACM 的坑,周边巨擘林立,从此过上了"天天被虐似死狗"的生活… 然而我是谁,我可是死狗中的战斗鸡,智力不够那刷题来凑,开始了夜以继日哼哧哼哧刷题的日子,从此"读题与提交齐飞, AC 与 WA 一色 ",我惊喜的发现被题虐既刺激又有快感,那一刻我泪流满面。这么好的事儿作为一个正直的人绝不能自己独享,经过激烈的颅内斗争,我决定把我私藏的十几个 T 的,阿不,十几个刷题网...
Java面试官最爱问的volatile关键字
在Java的面试当中,面试官最爱问的就是volatile关键字相关的问题。经过多次面试之后,你是否思考过,为什么他们那么爱问volatile关键字相关的问题?而对于你,如果作为面试官,是否也会考虑采用volatile关键字作为切入点呢? 为什么爱问volatile关键字 爱问volatile关键字的面试官,大多数情况下都是有一定功底的,因为volatile作为切入点,往底层走可以切入Java内存模...
项目中的if else太多了,该怎么重构?
介绍 最近跟着公司的大佬开发了一款IM系统,类似QQ和微信哈,就是聊天软件。我们有一部分业务逻辑是这样的 if (msgType = "文本") { // dosomething } else if(msgType = "图片") { // doshomething } else if(msgType = "视频") { // doshomething } else { // doshom...
致 Python 初学者
欢迎来到“Python进阶”专栏!来到这里的每一位同学,应该大致上学习了很多 Python 的基础知识,正在努力成长的过程中。在此期间,一定遇到了很多的困惑,对未来的学习方向感到迷茫。我非常理解你们所面临的处境。我从2007年开始接触 python 这门编程语言,从2009年开始单一使用 python 应对所有的开发工作,直至今天。回顾自己的学习过程,也曾经遇到过无数的困难,也曾经迷茫过、困惑过。开办这个专栏,正是为了帮助像我当年一样困惑的 Python 初学者走出困境、快速成长。希望我的经验能真正帮到你
【C++100问】深入理解理解顶层const和底层const
专栏C++学习笔记 声明 1)该文章整理自网上的大牛和相关专家无私奉献的资料,具体引用的资料请看参考文献。 2)本文仅供学术交流,非商用。所以每一部分具体的参考资料并没有详细对应。如果某部分不小心侵犯了大家的利益,还望海涵,并联系博主删除。 3)博主才疏学浅,文中如有不当之处,请各位指出,共同进步,谢谢。 4)此属于第一版本,若有错误,还需继续修正与增删。还望大家多多指点。大家都共享一点点,一起为...
吐血推荐珍藏的Visual Studio Code插件
作为一名Java工程师,由于工作需要,最近一个月一直在写NodeJS,这种经历可以说是一部辛酸史了。好在有神器Visual Studio Code陪伴,让我的这段经历没有更加困难。眼看这段经历要告一段落了,今天就来给大家分享一下我常用的一些VSC的插件。 VSC的插件安装方法很简单,只需要点击左侧最下方的插件栏选项,然后就可以搜索你想要的插件了。 下面我们进入正题 Material Theme ...
《C++ Primer》学习笔记(五):循环、分支、跳转和异常处理语句
专栏C++学习笔记 《C++ Primer》学习笔记/习题答案 总目录 https://blog.csdn.net/TeFuirnever/article/details/100700212 —————————————————————————————————————————————————————— 《C++ Primer》习题参考答案:第5章 - 语句 文章目录专栏C++学习笔记语句1、...
相关热词 c#委托 逆变与协变 c#新建一个项目 c#获取dll文件路径 c#子窗体调用主窗体事件 c# 拷贝目录 c# 调用cef 网页填表c#源代码 c#部署端口监听项目、 c#接口中的属性使用方法 c# 昨天
立即提问