I'm running MyBench in a new environment and it fails while loading data into Phoenix table using CsvBulkLoadTool utility.
WARN tool.LoadIncrementalHFiles: Attempt to bulk load region containing into table BENCH.USERVISITS with files [family:0 path:hdfs://bidev/tmp/386640ec-d49e-4760-8257-05858a409321/BENCH.USERVISITS/0/b467b5560eee4d61a42d4c9e6a78eb7e] failed. This is recoverable and they will be retried.
INFO tool.LoadIncrementalHFiles: Split occurred while grouping HFiles, retry attempt 100 with 1 files remaining to group or split
ERROR tool.LoadIncrementalHFiles: -------------------------------------------------
Bulk load aborted with some files not yet loaded:
After closer examination, I discovered that the error takes place while moving/renaming input file into HBase staging directory /apps/hbase/data/staging. In this cluster, the HBase data is encrypted and moving data between encrypted and normal zone is not possible.
java.io.IOException: Failed to move HFile: hdfs://bidev/apps/hbase/data/staging/ambari-qa__BENCH.USERVISITS__dbb5qdfppq1diggr0dmdbcb1ji74ol4b9jn9ee2dgp1ttn9n5i6llfih7101fi1d/0/3a7f2d612c034253ad375ae002cc6ade to hdfs://bidev/tmp/fc43e454-00b3-4db0-8bdd-8b475885ab49/BENCH.USERVISITS/0/3a7f2d612c034253ad375ae002cc6ade
at org.apache.hadoop.hbase.regionserver.SecureBulkLoadManager$SecureBulkLoadListener.failedBulkLoad(SecureBulkLoadManager.java:423)
at org.apache.hadoop.hbase.regionserver.HRegion.bulkLoadHFiles(HRegion.java:6035)
at org.apache.hadoop.hbase.regionserver.SecureBulkLoadManager$3.run(SecureBulkLoadManager.java:284)
The source code can be found
here.
if (!FSUtils.isSameHdfs(conf, srcFs, fs)) { |
| LOG.debug("Bulk-load file " + srcPath + " is on different filesystem than " + |
| "the destination filesystem. Copying file over to destination staging dir."); |
| FileUtil.copy(srcFs, p, fs, stageP, false, conf); |
| } else if (copyFile) { |
| LOG.debug("Bulk-load file " + srcPath + " is copied to destination staging dir."); |
| FileUtil.copy(srcFs, p, fs, stageP, false, conf); |
| } else { |
| LOG.debug("Moving " + p + " to " + stageP); |
| FileStatus origFileStatus = fs.getFileStatus(p); |
| origPermissions.put(srcPath, origFileStatus.getPermission()); |
| if(!fs.rename(p, stageP)) { |
| throw new IOException("Failed to move HFile: " + p + " to " + stageP); |
| } |
| } |
When data is moved between different file system, the copying is enforced but unfortunately, data movement between encrypted and decrypted zone is not covered here.
Another option is to make use of "copyFile" parameter which enforces copying. After analyzing the control flow I discovered that there exists "hbase-site.xml" parameter
always.copy.files which seems to be the solution to the problem. But after applying this parameter, nothing has changed.
Further examination with a little help of remote debugging unearthed a sad truth. CsvBulkLoadTool is passing the control to
LoadIncrementalHFiles.java and "doBulkLoad" function.
public Map<LoadQueueItem, ByteBuffer> doBulkLoad(Path hfofDir, final Admin admin, Table table, RegionLocator regionLocator) throws TableNotFoundException, IOException {
return doBulkLoad(hfofDir, admin, table, regionLocator, false, false);
}
Unfortunately, the "copyFiles" parameter is hardcoded as "false" although there is a sound and ready to use "isAlwaysCopyFiles()" function utilizing "hbase-site.xml" config file.
The only solution is manual fix and recreating the package from source files. But it does not go easy because one has to leverage different and outdated versions of HBase and Phoenix to create "Phoenix client" package matching HDP 3.1.
So two days spent without a solution.