Technology, KeyWords:
WebLogic Server 10.3, Cluster, Exception; Too many open files, Socket, Linux, SLES
Error Message: <BEA-002616> <Failed to listen on channel "Default"…>
Problem:
<Critical> <Server> <BEA-002616> <Failed to listen on channel "Default" on IP_NR:Port, failure count: 1, failing for 0 seconds, java.net.SocketException: Too many open files>
Description
Weblogic declaration from: http://docs.oracle.com/cd/E23549_01/apirefs.1111/e14397/Server.html
BEA-002616
Critical: Failed to listen on channel "channel" on listenAddress:port, failure count: fails1, failing for secs2 seconds, e3
Description
The server listener will retry the listen after a short delay.
Cause
The server got an exception while trying to accept client connections. It will try to backoff to aid recovery.
Action
The OS limit for the number of open file descriptor (FD limit) needs to be increased. Tune OS parameters that might help the server to accept more client connections (e.g. TCP accept back log).
More detail and Background information:
“Tune OS parameters” depends on OS which you use. For Linux, kernel parameters need adjustment- exact details depend on the distribution. There are individual user limits for open files. You can use `ulimit -a` command to find out the Linux for the user that owns.
You can find information regarding the important parameters e.g. number of "open files" or similar entry. Please consider, in AIX, the user limits apply as well as a system OS configuration for the total number of open files allowed on the host.
Solution
Here is an example how can you check and isolate the position of error.
1- Check linux configuration (e.g. on host XXX):
> ulimit –a
core file size (blocks, -c) 1
data seg size (kbytes, -d) unlimited
scheduling priority (-e) 0
file size (blocks, -f) unlimited
pending signals (-i) 127575
max locked memory (kbytes, -l) 64
max memory size (kbytes, -m) 13887980
open files (-n) 65600 à OK
pipe size (512 bytes, -p) 8
POSIX message queues (bytes, -q) 819200
real-time priority (-r) 0
stack size (kbytes, -s) 8192
cpu time (seconds, -t) unlimited
max user processes (-u) 127575
virtual memory (kbytes, -v) 17749200
file locks (-x) unlimited
The open file descriptor limit is at 65600 as recommended by Oracle.
2- Check currently open files
a. Find PID of Managed Server
> ps -fea|grep myManagedServer
b. Check open files
Please use lsof (list open files) that lists information about files opened by processes
You see list of open files via (e.g.PID is here 1234)
>lsof -p 1234
In order to findour the number of open files:
> lsof -p 1234 | wc –l
In this case, we observed that application has 13 thousand connections on "CLOSE_WAIT" status. This usually happens when the application doesn't close the connections properly: