[OpenSIPS-Users] Chasing down unreleased memory

Răzvan Crainea razvan at opensips.org
Tue Oct 2 11:04:19 EDT 2018

Hi, Jock!

Are you using local_route in your script? If so, are you doing any R-URI 
manipulations inside it?

Best regards,

On 10/2/18 12:00 PM, Răzvan Crainea wrote:
> Hi, Jock!
> Can you confirm exactly the opensips version you are running? (opensips 
> -V).
> Best regards,
> Razvan
> On 10/2/18 11:41 AM, Răzvan Crainea wrote:
>> Hi, Jock!
>> Good news :) I've just figured out the pkg memleak in mc_compact, and 
>> I've already pushed a fix upstream. Please pull the latest version and 
>> run a few tests, and let us know how they go.
>> In the meantime, I will investigate the shm leak.
>> Best regards,
>> Razvan
>> On 10/2/18 10:13 AM, Răzvan Crainea wrote:
>>> Hi, Jock!
>>> I am pretty busy these days, so I couldn't pay too much attention to 
>>> this thread. Nevertheless, I will try to prioritize it this week and 
>>> get back to you.
>>> I have also tracked down the pkg mem leak to the mc_compact() module, 
>>> but I can't pinpoint for now exactly the issue. Although I have some 
>>> assumptions, I can't be sure until I run some tests.
>>> If I need extra info during my debug, I will definitely let you know, 
>>> so keep an eye on this thread.
>>> Best regards,
>>> Răzvan
>>> On 10/1/18 9:14 PM, Jock McKechnie wrote:
>>>> Greetings;
>>>> I should have done this last week but, to be honest, it just didn't
>>>> occur to me. I pulled the compression module out (and commented out
>>>> mc_compact()) and re-ran my tests and the memory usage was massively
>>>> lower (losing 100k of memory per minute vs 3-5MB/minute) and within a
>>>> few minutes of calling ceasing almost all of the memory was released
>>>> (went from 42MB down to 3.8MB within 6 minutes).
>>>> So it appears that there's something leaking in the compression module.
>>>> I'm going to poke around a few more tests (include compression.so
>>>> without actually calling it, etc) but I wanted to pass this on as
>>>> there appears to be a bug that needs addressing.
>>>> Is there anything I can provide that might be useful in debugging this?
>>>> As always, thank you for your help.
>>>>   - Jock
>>>> On Fri, Sep 21, 2018 at 8:12 AM, Jock McKechnie
>>>> <jock.mckechnie at gmail.com> wrote:
>>>>> Thank you very much, Răzvan;
>>>>> We are using the compression module for mc_compact(). We use SIP
>>>>> headers for passing around data (don't ask) and the compact headers
>>>>> are mandatory or else we hit MTU/fragmentation problems. We're on the
>>>>> cusp of moving to SIP/TCP but haven't gotten there quite yet.
>>>>> The mc_compact() lives in the main route[1] block and is simply 
>>>>> this, in whole:
>>>>>      mc_compact("P-Asserted-Identity|Remote-Party-ID|X-UCID");
>>>>> And my apologies for not including the shm dump segment - please 
>>>>> find it below.
>>>>> https://pastebin.com/itSTER7Y
>>>>> Let me know what else I can provide, this is fairly easy to replicate
>>>>> (although it takes about an hour of runtime pummeling it to get nice
>>>>> easy to read dumps). Thank you, again, very much.
>>>>>   - Jock
>>>>> On Fri, Sep 21, 2018 at 3:01 AM, Răzvan Crainea 
>>>>> <razvan at opensips.org> wrote:
>>>>>> Hi, Jock!
>>>>>> I think you are actually heating two memory leaks: one in private 
>>>>>> memory and
>>>>>> one in shared memory.
>>>>>> In the memory dump you sent I can spot a pkg mem related to (I 
>>>>>> believe) the
>>>>>> compression module. Can you tell me what functions of the 
>>>>>> compression module
>>>>>> you are using and what flags are you using for them?
>>>>>> Besides that, you claim there is a shared memory leak as well. Can 
>>>>>> you send
>>>>>> us the logs for the shared memory? They are dumped by the first 
>>>>>> process
>>>>>> (attendant) and should start with the string "Memory status 
>>>>>> (shm):". Can you
>>>>>> post those logs as well?
>>>>>> Best regards,
>>>>>> Razvan
>>>>>> On 9/20/18 9:43 PM, Jock McKechnie wrote:
>>>>>>> Greetings OpenSIPS gurus;
>>>>>>> I've run into an issue that I'm not making any real headway with 
>>>>>>> on my
>>>>>>> own. I have an OpenSIPS 2.4.2 running on CentOS7 that does not 
>>>>>>> appear
>>>>>>> to be releasing the bulk of its shared memory when calls die off.
>>>>>>> Eventually OpenSIPS runs itself out of shared memory and goes beserk
>>>>>>> when it can't reclaim memory for new transactions, or I've also seen
>>>>>>> the kernel oomkiller take it down.
>>>>>>> The setup is, as mentioned, a 2.4.2 which has a chunk of perl living
>>>>>>> on its back that handles some API directed call decisions. 
>>>>>>> Because of
>>>>>>> the inevitable delay in the API lookups I am running a large 
>>>>>>> number of
>>>>>>> threads (100) to ensure that I don't start dropping packets under 
>>>>>>> high
>>>>>>> levels of load (50 CPS+ - until I bumped children counts up I was
>>>>>>> seeing packets being ignored which got the End User very 
>>>>>>> excitable). I
>>>>>>> appreciate having lots of children amplifies memory usage but my
>>>>>>> belief is during regular load on our side (~6000 calls at ~30 - 
>>>>>>> 40CPS)
>>>>>>> I should have a full day's runtime easily w/o running into problems.
>>>>>>> The issue is that because the memory is not released overnight it
>>>>>>> starts the next day with half its shared memory gone, and then 
>>>>>>> things
>>>>>>> go to pot the next day when it runs through all its remaining shared
>>>>>>> memory.
>>>>>>> I have it configured for a very large (I think) amount of memory to
>>>>>>> attempt to compensate short-term - 2048MB shared memory, 128MB 
>>>>>>> package
>>>>>>> memory. This is not a well thought-out number, just me throwing 
>>>>>>> RAM at
>>>>>>> the problem. (For reference this is on a VM running with 4 cores and
>>>>>>> 4096MB of RAM)
>>>>>>> I have graphed the memory stats out of the FIFO, after enabling
>>>>>>> memstats and DBG_MALLOC/etc. The graphs can be found here:
>>>>>>> https://imgur.com/a/oapKYJW
>>>>>>> You'll see that the first graph, the main shmem:, shows the usage 
>>>>>>> gets
>>>>>>> tied up and remains after calls die off (these examples run for 20
>>>>>>> minutes after the calls stopped - but I can leave it overnight 
>>>>>>> and the
>>>>>>> shmem: usage won't drop). This being said, the other graphed modules
>>>>>>> show that it releases memory correctly after calls quit like they
>>>>>>> should. None of the other modules graphed (eg, _all_ of the 
>>>>>>> modules I
>>>>>>> was using in OpenSIPS) showed no significant usage and I've omitted
>>>>>>> them.
>>>>>>> Interesting that the 'perl' module shows no stats ever which 
>>>>>>> makes me
>>>>>>> wonder if the stats on this module works. I am not storing anything
>>>>>>> inside perl, besides lots of local session variables, and all 
>>>>>>> channel
>>>>>>> variables are pushed back into OpenSIPS AVPs, but I assume there 
>>>>>>> must
>>>>>>> be some memory overhead, although it may not significantly change.
>>>>>>> When I shut OpenSIPS down I get child dumps like the following: 
>>>>>>> (I've
>>>>>>> included only one for brevity, but they're all much of a muchness)
>>>>>>> https://pastebin.com/KG2qxxBa
>>>>>>> If I'm reading this right (and there's a strong possibility I'm
>>>>>>> misunderstanding what I'm looking at) the memory is tied up in 
>>>>>>> io_wait
>>>>>>> and the parser. I'm not sure what this signifies, however.
>>>>>>> So I guess what I'm asking is:
>>>>>>> Am I understanding what I'm looking at correctly and that 
>>>>>>> OpenSIPS is
>>>>>>> not releasing its shared memory which causes it to run out of memory
>>>>>>> quicker the following day? (Which lines up with the symptoms of
>>>>>>> shorter runtimes before failure)
>>>>>>> Based on the dumps where should I look for the source of this issue?
>>>>>>> Finally, and unrelated to the above, I attempted to mail the 
>>>>>>> official
>>>>>>> OpenSIPS peeps to talk about paid support but none of the 
>>>>>>> 'contact us'
>>>>>>> submits appear to work - when I click on the submit button it 
>>>>>>> changes
>>>>>>> shade but does not appear to act. Certainly I never received any
>>>>>>> response. If there was a direct eMail address I could send to that
>>>>>>> would be preferable.
>>>>>>> As always, my thanks for your time.
>>>>>>>    - Jock
>>>>>>> PS -
>>>>>>> List of modules I'm using in this config:loadmodule "signaling.so"
>>>>>>> loadmodule "sl.so"
>>>>>>> loadmodule "tm.so"
>>>>>>> loadmodule "rr.so"
>>>>>>> loadmodule "maxfwd.so"
>>>>>>> loadmodule "textops.so"
>>>>>>> loadmodule "uri.so"
>>>>>>> loadmodule "mi_fifo.so"
>>>>>>> loadmodule "siptrace.so"
>>>>>>> loadmodule "sipmsgops.so"
>>>>>>> loadmodule "dialog.so"
>>>>>>> loadmodule "uac_auth.so"
>>>>>>> loadmodule "uac.so"
>>>>>>> loadmodule "perl.so"
>>>>>>> loadmodule "acc.so"
>>>>>>> loadmodule "proto_hep.so"
>>>>>>> loadmodule "proto_udp.so"
>>>>>>> loadmodule "compression.so"
>>>>>>> _______________________________________________
>>>>>>> Users mailing list
>>>>>>> Users at lists.opensips.org
>>>>>>> http://lists.opensips.org/cgi-bin/mailman/listinfo/users
>>>>>> -- 
>>>>>> Răzvan Crainea
>>>>>> OpenSIPS Core Developer
>>>>>>    http://www.opensips-solutions.com
>>>>>> _______________________________________________
>>>>>> Users mailing list
>>>>>> Users at lists.opensips.org
>>>>>> http://lists.opensips.org/cgi-bin/mailman/listinfo/users
>>>> _______________________________________________
>>>> Users mailing list
>>>> Users at lists.opensips.org
>>>> http://lists.opensips.org/cgi-bin/mailman/listinfo/users

Răzvan Crainea
OpenSIPS Core Developer

More information about the Users mailing list