While putting data in Hbase
, using the HTable.put
method, I'll end up with the below exception occasionally. But the data has been actually written to Hbase
when I checked the get operation for that particular rowkey
.
For the same time I have searched for the logs in both HMaster and HRegionservers
to identify the issue. But unable to find that.
Please help to fine tune Hbase
Configurations in order to avoid InterruptedIOException.
Hadoop Distribution: Apache
Version: HBase 1.2.6
Cluster size: 12nodes
java.io.InterruptedIOException: #17209, interrupted. currentNumberOfTask=1
at org.apache.hadoop.hbase.client.AsyncProcess.waitForMaximumCurrentTasks(AsyncProcess.java:1764)
at org.apache.hadoop.hbase.client.AsyncProcess.waitForMaximumCurrentTasks(AsyncProcess.java:1734)
at org.apache.hadoop.hbase.client.AsyncProcess.waitForAllPreviousOpsAndReset(AsyncProcess.java:1810)
at org.apache.hadoop.hbase.client.BufferedMutatorImpl.backgroundFlushCommits(BufferedMutatorImpl.java:240)
at org.apache.hadoop.hbase.client.BufferedMutatorImpl.flush(BufferedMutatorImpl.java:190)
at org.apache.hadoop.hbase.client.HTable.flushCommits(HTable.java:1434)
at org.apache.hadoop.hbase.client.HTable.put(HTable.java:1018)
Please help to solve it
The same exception has been faced by someone . But in that thread, there is no explanation about which are configurations need to be checked in order to avoid it
https://groups.google.com/forum/#!topic/nosql-databases/UxfrmWl_ZnM