Details
-
Bug
-
Status: Open
-
Major
-
Resolution: Unresolved
-
1.2.0
-
None
-
None
-
None
-
RHEL 6 running against Hive storage type
Description
I have a java app accessing Drill 1.2 via JDBC, which runs 100s of counts on various tables. No concurrency is being used. The JDBC URL uses the format:
jdbc:drill:drillbit=a-bits-hostname
Hanifi suggested I check for open file descriptors using:
lsof -a -p DRILL_PID | wc -l
which I did on the two nodes, I currently have running drill, both, before and
after restarting.
Node from JDBC connection string (which had been previously restarted):
Before: 396
After: 396
Other node:
Before: 14
After: 395
The error, "Too many open files", persists after restarting the bits.
Opened as a result of this thread:
http://mail-archives.apache.org/mod_mbox/drill-user/201601.mbox/browser