[OpenSIPS-Users] segfault on 2.4.4

Bogdan-Andrei Iancu bogdan at opensips.org
Thu Jan 3 06:55:35 EST 2019


Hi Jennifer,

Do you still have the core file for investigation with GDB ?

Also, are you 100% sure your cluster is consistent when comes to the ACC 
settings, like *all* the nodes do have exactly the same "extra" and 
"legs" sets ?

Best regards and A Happy New Year,

Bogdan-Andrei Iancu

OpenSIPS Founder and Developer
   https://www.opensips-solutions.com
OpenSIPS Summit 2019
   https://www.opensips.org/events/Summit-2019Amsterdam/

On 12/21/2018 08:48 PM, Jennifer Hashimoto wrote:
> Hi guys, I just updated and i’m getting segfault not sure why, maybe 
> to do with dialog replication?
>
> Here are the details, let me know if you could use more 
> information.https://opensips.org/html/docs/modules/2.4.x/ldap.html
>
> Thanks,
> Jen
>
>
>
> version: opensips 2.4.4 (x86_64/linux)
> flags: STATS: On, DISABLE_NAGLE, USE_MCAST, SHM_MMAP, PKG_MALLOC, 
> F_MALLOC, FAST_LOCK-ADAPTIVE_WAIT
> ADAPTIVE_WAIT_LOOPS=1024, MAX_RECV_BUFFER_SIZE 262144, MAX_LISTEN 16, 
> MAX_URI_SIZE 1024, BUF_SIZE 65535
> poll method support: poll, epoll, sigio_rt, select.
> main.c compiled on  with gcc 4.9.2
>
> Dec 21 13:42:55 NFO:clusterer:handle_internal_msg: Node [2] is UP
> Dec 21 13:43:11 13CB78-2323 64.86.243.114 BYE from caz5400 
> sip:8198213679 at 64.86.243.116 -> sip:33232323434 at 209.58.46.142:5060 
> bye_from=caz5400 64.86.243.116
> Dec 21 13:43:31 NFO:load_balancer:set_dst_state_from_rplcode: disable 
> destination 2 <sip:10.10.10.16> after 408 reply on probe
> Dec 21 13:43:49 19C280-2672 64.86.243.114 BYE from caz5400 
> sip:4036409658 at 64.86.243.116 -> sip:36205873156 at 209.58.46.142:5060 
> bye_from=caz5400 64.86.243.116
> Dec 21 13:44:02 RITICAL:core:sig_usr: segfault in process pid: 3050, 
> id: 61
> Dec 21 13:44:03 RITICAL:core:handle_worker: dead child 61 (EOF 
> received), pid 3050
> Dec 21 13:44:03 RITICAL:core:handle_tcp_worker: dead tcp worker 0 (EOF 
> received), pid 3050
> Dec 21 13:44:05 NFO:core:handle_sigs: child process 3050 exited by a 
> signal 11
> Dec 21 13:44:05 NFO:core:handle_sigs: core was generated
> Dec 21 13:44:05 NFO:core:handle_sigs: terminating due to SIGCHLD
> Dec 21 13:44:05 NFO:core:sig_usr: signal 15 received
> Dec 21 13:44:05 NFO:core:sig_usr: signal 15 received
> Dec 21 13:44:05 NFO:core:sig_usr: signal 15 received
>
> sudo gdb /usr/sbin/opensips core
> GNU gdb (Debian 7.7.1+dfsg-5) 7.7.1
> Copyright (C) 2014 Free Software Foundation, Inc.
> License GPLv3+: GNU GPL version 3 or later 
> <http://gnu.org/licenses/gpl.html>
> This is free software: you are free to change and redistribute it.
> There is NO WARRANTY, to the extent permitted by law.  Type "show copying"
> and "show warranty" for details.
> This GDB was configured as "x86_64-linux-gnu".
> Type "show configuration" for configuration details.
> For bug reporting instructions, please see:
> <http://www.gnu.org/software/gdb/bugs/>.
> Find the GDB manual and other documentation resources online at:
> <http://www.gnu.org/software/gdb/documentation/>.
> For help, type "help".
> Type "apropos word" to search for commands related to "word"...
> Reading symbols from /usr/sbin/opensips...Reading symbols from 
> /usr/lib/debug/.build-id/4c/1b7823a23c3dbd2b5d0ee6392836a093740d2a.debug...done.
> done.
> [New LWP 3050]
> [Thread debugging using libthread_db enabled]
> Using host libthread_db library "/lib/x86_64-linux-gnu/libthread_db.so.1".
> Core was generated by `/usr/sbin/opensips -P 
> /run/opensips/opensips.pid -f /etc/opensips/opensips.cfg'.
> Program terminated with signal SIGSEGV, Segmentation fault.
> #0  0x00000000004bca7a in fm_free (qm=0x7f42afdc9000,
>     p=p at entry=0x3a6874676e654c2d) at mem/f_malloc.c:495
> 495mem/f_malloc.c: No such file or directory.
> (gdb) bt full
> #0  0x00000000004bca7a in fm_free (qm=0x7f42afdc9000,
>     p=p at entry=0x3a6874676e654c2d) at mem/f_malloc.c:495
>         f = 0x3a6874676e654c15
>         n = <optimized out>
> #1  0x00007f42aad4bb8a in shm_free (_p=0x3a6874676e654c2d)
>     at ../../evi/../mem/shm_mem.h:588
> No locals.
> #2  set_value_shm (pvt=pvt at entry=0x7ffe501b6e90, extra=0x7f42b2a94bd0)
>     at acc_vars.c:143
>         s = <optimized out>
>         __FUNCTION__ = "set_value_shm"
> #3  0x00007f42aad3a85b in restore_extra_from_str (
>     tags_len=<optimized out>, extra_s=extra_s at entry=0x7ffe501b6f10,
>     extra_len=<optimized out>) at acc.c:1618
>         i = <optimized out>
>         value = {rs = {s = 0x7f42bfe53f3f "oreB", len = 0},
>           ri = -1297517200, flags = 1}
>         values = 0x7f42b2a94a68
>         __FUNCTION__ = "restore_extra_from_str"
> #4  0x00007f42aad42691 in restore_extra (
>     type_str=0x7f42aaf560e0 <extra_str>, ctx=0x7f42b2a87ad0,
> ---Type <return> to continue, or q <return> to quit---
>     dlg=0x7f42b011c258) at acc.c:1653
>         extra_len = <optimized out>
>         buffer = {s = 0x7f42bfe53f3d "", len = 2}
> #5  restore_dlg_extra (dlg=0x7f42b011c258, ctx_p=0x7ffe501b6f80)
>     at acc.c:1729
>         ctx = 0x7f42b2a87ad0
>         __FUNCTION__ = "restore_dlg_extra"
> #6  0x00007f42aad458a3 in acc_loaded_callback (dlg=0x7f42b011c258,
>     type=1852132397, _params=0x1) at acc_logic.c:662
>         flags_s = {s = 0x7ffe501b6f90 "\003", len = 8}
>         ctx_s = {s = 0x7ffe501b6fd8 "\b", len = 0}
>         table_s = {s = 0x7ffe501b6ff0 "|\362\345\277B\177",
>           len = 1343975760}
>         created_s = {s = 0x7ffe501b6f88 "b4\035\\", len = 8}
>         ctx = 0x0
>         created = 1545417826
>         flags = 9570149209145347
>         __FUNCTION__ = "acc_loaded_callback"
> #7  0x00007f42aaadadcd in run_load_callback_per_dlg (
>     dlg=0x7f42afdc9000) at dlg_cb.c:212
>         cb = 0x7f42b01fab30
> ---Type <return> to continue, or q <return> to quit---
> #8  0x00007f42aab16d27 in dlg_replicated_create (
>     packet=0x7f42afdc9000, cell=0x7f42b011c258, ftag=0x0,
>     ttag=0x7ffe501b7110, safe=-265519414) at dlg_replication.c:271
>         dir = 2863757965
>         dst_leg = 32578
>         callid = {
>           s = 0x7f42bfe5eef9 
> "2EB67E0D-48711E9-8927CE48-FE93CD99 at 10.10.20.22 
> <mailto:2EB67E0D-48711E9-8927CE48-FE93CD99 at 10.10.20.22>\f", len = 46}
>         from_uri = {
>           s = 0x7f42bfe5ef46 "sip:6138245700 at 10.10.20.22\036 
> <sip:6138245700 at 10.10.20.22%5C036>",
>           len = 26}
>         to_uri = {
>           s = 0x7f42bfe5ef62 
> "sip:87441189781400 at 10.10.20.42f\235j\033r4\035\\\003 
> <sip:87441189781400 at 10.10.20.42f%5C235j%5C033r4%5C035%5C%5C%5C003>", 
> len = 30}
>         from_tag = {s = 0x7f42bfe5ef29 "B5B6DC0-2171\r", len = 12}
>         to_tag = {s = 0x7f42bfe5ef37 "KjvKKDN565gee\032", len = 13}
>         cseq1 = {s = 0x7f42bfe5efba "0\003", len = 1}
>         cseq2 = {s = 0x7f42bfe5efbd "101", len = 3}
>         contact1 = {
>           s = 0x7f42bfe5efeb "sip:6138245700 at 10.10.20.22:5060/",
>           len = 31}
> ---Type <return> to continue, or q <return> to quit---
>         contact2 = {
>           s = 0x7f42bfe5f00c 
> "sip:441189781400 at 10.10.20.28:5070;transport=udp", len = 47}
>         rroute1 = {s = 0x0, len = 0}
>         rroute2 = {
>           s = 0x7f42bfe5efc4 
> "<sip:10.10.20.39;lr;did=06c.2f38c332>\037", len = 37}
>         mangled_fu = {s = 0x0, len = 0}
>         mangled_tu = {
>           s = 0x7f42bfe5f03f 
> "sip:441189781400 at 10.10.20.42\t\002accX_table#acc_caztel|accX_created#b4\035\\ 
> <sip:441189781400 at 10.10.20.42%5Ct%5C002accX_table#acc_caztel%7CaccX_created#b4%5C035%5C%5C>", 
> len = 28}
>         sock = {s = 0x7f42bfe5efa4 "udp:10.10.20.42:5060\001",
>           len = 20}
>         vars = {
>           s = 0x7f42bfe5f05d 
> "accX_table#acc_caztel|accX_created#b4\035\\", len = 521}
>         profiles = {
>           s = 0x7f42bfe5f268 "ani#6138245700|dnis#441189781400|",
>           len = 33}
>         dlg = 0x10d30f2bf
>         callee_sock = 0x0
> ---Type <return> to continue, or q <return> to quit---
>         d_entry = 0x7f42b00f0948
>         __FUNCTION__ = "dlg_replicated_create"
> #9  0x00007f42aab1a76e in receive_dlg_repl (packet=0x7ffe501b7350)
>     at dlg_replication.c:802
>         rc = 0
>         __FUNCTION__ = "receive_dlg_repl"
> #10 0x00007f42aa8ac8b9 in bin_rcv_mod_packets (packet=0x7f42afdc9000,
>     packet_type=1852132397, ri=0x1, ptr=0x7f42bfe53f3f)
>     at clusterer.c:1972
>         cl_cap = 0x1
>         source_id = 2
>         dest_id = 1
>         cluster_id = 1
>         ev_actions_required = 0
>         __FUNCTION__ = "bin_rcv_mod_packets"
> #11 0x0000000000429197 in call_callbacks (
>     buffer=0x7ffe501b7350 "\330\356\345\277B\177", rcv=0x7f42b10e76a0)
>     at bin_interface.c:446
>         p = 0x7f42bfe4ca68
>         pkg_len = 3219442280
>         packet = {buffer = {s = 0x7f42bfe5eed8 "P4CK\325\003",
> ---Type <return> to continue, or q <return> to quit---
>             len = 969}, front_pointer = 0x7f42bfe5f2a1 "\001",
>           size = 1031, type = 1, next = 0x0, src_id = 2}
>         __FUNCTION__ = "call_callbacks"
> #12 0x00007f42a8e72a6c in bin_handle_req (
>     _max_msg_chunks=<optimized out>, con=0x7f42b10e7680,
>     req=0x7f42a9076280 <bin_current_req>) at proto_bin.c:672
>         size = <optimized out>
> #13 bin_read_req (con=0x7f42b10e7680, bytes_read=0x7ffe501b7450)
>     at proto_bin.c:827
>         bytes = <optimized out>
>         total_bytes = 981
>         req = 0x7f42a9076280 <bin_current_req>
>         __FUNCTION__ = "bin_read_req"
> #14 0x000000000051e5e7 in handle_io (fm=0x7f42bfe8f318, idx=0,
>     event_type=-1324452224) at net/net_tcp_proc.c:241
>         ret = 0
>         n = -1324452224
>         s = 7
>         resp = 0
>         response = {139924415084160, 1}
>         __FUNCTION__ = "handle_io"
> ---Type <return> to continue, or q <return> to quit---
> #15 0x000000000052085f in io_wait_loop_epoll (h=<optimized out>,
>     t=<optimized out>, repeat=<optimized out>)
>     at net/../io_wait_loop.h:280
>         ep_event = {events = 5640552, data = {ptr = 0x0, fd = 0,
>             u32 = 0, u64 = 0}}
>         r = 0
>         i = 0
> #16 tcp_worker_proc_loop () at net/net_tcp_proc.c:386
>         __FUNCTION__ = "tcp_worker_proc_loop"
> #17 0x000000000052911d in tcp_start_processes (
>     chd_rank=chd_rank at entry=0x7c9ea8 <chd_rank>,
>     startup_done=startup_done at entry=0x0) at net/net_tcp.c:1892
>         r = 0
>         reader_fd = {230, 231}
>         pid = <optimized out>
>         __FUNCTION__ = "tcp_start_processes"
> #18 0x000000000041c679 in main_loop () at main.c:788
>         startup_done = 0x0
>         chd_rank = 57
>         rc = <optimized out>
> #19 main (argc=<optimized out>, argv=<optimized out>) at main.c:1439
> ---Type <return> to continue, or q <return> to quit---
>         cfg_stream = <optimized out>
>         c = <optimized out>
>         r = <optimized out>
>         tmp = 0x7ffe501b8ec3 ""
>         tmp_len = <optimized out>
>         port = <optimized out>
>         proto = <optimized out>
>         protos_no = <optimized out>
>         options = 0x55efc0 "f:cCm:M:b:l:n:N:rRvdDFEVhw:t:u:g:P:G:W:o:"
>         ret = -1
>         seed = 3877236842
>         __FUNCTION__ = "main"
> (gdb)
> ---------------------------------------------------
> Jennifer Akemi Hashimoto
> Caztel Communications
> jennifer.hashimoto at caztel.com <mailto:jennifer.hashimoto at caztel.com>
> 905-836-5445
>
>
>
> _______________________________________________
> Users mailing list
> Users at lists.opensips.org
> http://lists.opensips.org/cgi-bin/mailman/listinfo/users

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.opensips.org/pipermail/users/attachments/20190103/45f79a44/attachment-0001.html>


More information about the Users mailing list