How to avoid InterruptedIOException in Hbase put o

2019-06-12 02:06发布

While putting data in Hbase, using the HTable.put method, I'll end up with the below exception occasionally. But the data has been actually written to Hbase when I checked the get operation for that particular rowkey.

For the same time I have searched for the logs in both HMaster and HRegionservers to identify the issue. But unable to find that.

Please help to fine tune Hbase Configurations in order to avoid InterruptedIOException.

Hadoop Distribution: Apache
Version: HBase 1.2.6
Cluster size: 12nodes



java.io.InterruptedIOException: #17209, interrupted. currentNumberOfTask=1
    at org.apache.hadoop.hbase.client.AsyncProcess.waitForMaximumCurrentTasks(AsyncProcess.java:1764)
    at org.apache.hadoop.hbase.client.AsyncProcess.waitForMaximumCurrentTasks(AsyncProcess.java:1734)
    at org.apache.hadoop.hbase.client.AsyncProcess.waitForAllPreviousOpsAndReset(AsyncProcess.java:1810)
    at org.apache.hadoop.hbase.client.BufferedMutatorImpl.backgroundFlushCommits(BufferedMutatorImpl.java:240)
    at org.apache.hadoop.hbase.client.BufferedMutatorImpl.flush(BufferedMutatorImpl.java:190)
    at org.apache.hadoop.hbase.client.HTable.flushCommits(HTable.java:1434)
    at org.apache.hadoop.hbase.client.HTable.put(HTable.java:1018)

Please help to solve it

The same exception has been faced by someone . But in that thread, there is no explanation about which are configurations need to be checked in order to avoid it

https://groups.google.com/forum/#!topic/nosql-databases/UxfrmWl_ZnM

0条回答
登录 后发表回答