I am having a problem with the http_bind connection crashing out with ejabberd 2.1.10 x86_64. We get about 10 of these per minute on a very high volume site. There are 7 nodes in the cluster. Here is the crash report:
=CRASH REPORT==== 26-Mar-2012::11:22:42 ===
crasher:
pid: <0.27085.30>
registered_name: []
exception exit: {noproc,
{gen_fsm,sync_send_all_state_event,
[<9433.26826.7749>,
{http_put,161896132,
[{"xmlns",
"http://jabber.org/protocol/httpbind"},
{"rid","161896132"},
{"sid",
"de5ced2d0e3eb78bd1d4195b25328a28ab598e80"}],
[],114,1,[],
{{127,0,0,1},32334}},
30000]}}
in function gen_fsm:sync_send_all_state_event/3
in call from ejabberd_http_bind:http_put/7
in call from ejabberd_http_bind:handle_http_put/7
in call from ejabberd_http:process/2
in call from ejabberd_http:process_request/1
in call from ejabberd_http:process_header/2
in call from ejabberd_http:receive_headers/1
initial call: ejabberd_http:init({gen_tcp,#Port<0.1944479>},
[inet,captcha,http_bind,http_poll,
web_admin,
{ip,{0,0,0,0}}])
ancestors: [ejabberd_http_sup,ejabberd_sup,<0.36.0>]
messages: []
links: [#Port<0.1944479>,<0.360.0>]
dictionary: []
trap_exit: false
status: running
heap_size: 4181
stack_size: 23
reductions: 1384
neighbours:
=SUPERVISOR REPORT==== 26-Mar-2012::11:22:42 ===
Supervisor: {local,ejabberd_http_sup}
Context: child_terminated
Reason: {noproc,
{gen_fsm,sync_send_all_state_event,
[<9433.26826.7749>,
{http_put,161896132,
[{"xmlns","http://jabber.org/protocol/httpbind"},
{"rid","161896132"},
{"sid",
"de5ced2d0e3eb78bd1d4195b25328a28ab598e80"}],
[],114,1,[],
{{127,0,0,1},32334}},
30000]}}
Offender: [{pid,<0.27085.30>},
{name,undefined},
{mfa,{ejabberd_http,start_link,
[{gen_tcp,#Port<0.1944479>},
[inet,captcha,http_bind,http_poll,
web_admin,
{ip,{0,0,0,0}}]]}},
{restart_type,temporary},
{shutdown,brutal_kill},
{child_type,worker}]
Here are the Mnesia tables and status of them on node 1:
Name Storage Type Elements Memory
acl RAM AND DISC 2 356
anonymous RAM 0 283
caps_features DISC 0 5,752
captcha RAM 0 283
config RAM AND DISC 17 655
http_bind RAM 52,639 7,324,783
iq_response RAM 0 283
irc_custom RAM AND DISC 0 283
last_activity RAM AND DISC 0 283
local_config RAM AND DISC 8 1,001
mod_register_ip RAM 0 283
motd RAM AND DISC 0 283
motd_users RAM AND DISC 0 283
muc_online_room RAM 418 36,573
muc_registered RAM AND DISC 0 283
muc_room RAM AND DISC 403 179,648
offline_msg DISC 125 128,568
passwd RAM AND DISC 1 351
privacy RAM AND DISC 0 283
private_storage DISC 0 5,752
pubsub_index RAM AND DISC 1 295
pubsub_item DISC 0 5,752
pubsub_last_item RAM 0 283
pubsub_node RAM AND DISC 2 704
pubsub_state RAM AND DISC 2 407
pubsub_subscription RAM AND DISC 0 283
reg_users_counter RAM 0 283
roster RAM AND DISC 0 283
roster_version RAM AND DISC 0 283
route RAM 28 1,849
s2s RAM 0 283
schema RAM AND DISC 39 5,405
session RAM 7,387 1,534,303
session_counter RAM 1 322
sr_group RAM AND DISC 0 283
sr_user RAM AND DISC 0 283
temporarily_blocked REMOTE 0 0
vcard DISC 0 5,752
vcard_search RAM AND DISC 0 283
maybe server just got into
maybe server just got into the limit of maximum open file descriptors or connections?
Did you ever figure out what
Did you ever figure out what this was? We are seeing this as well
Did you ever figure out what
Did you ever figure out what this was? We are seeing this as well
were you able to get to the
were you able to get to the bottom of this? we are seeing this issue as well