I've been running 2.0-905 for ~25-26 hours. The manager's memory usage has
slowly crept up to 13 GB.
One thing of note - I'm using the ElasticSearch log writer. I see 3
possible scenarios for this memleak:
1) There is indeed a leak in master, potentially only triggered by
specific traffic,
2) There is a leak in the ElasticSearch log writer,
3) My ElasticSearch server can't keep up with the load, and the manager is
receiving logs faster than it can send them to the writer, so they just
queue up.
Has anyone else tried the current code over an extended period on live
traffic? Also, if anyone has any ideas to try to figure out where this
leak is occurring, please let me know. I'm going to switch back to ASCII
logs for a bit, and see what that's like.
--Vlad
(Taking to bro-dev).
That's not a good sign for the manager ... It's possible that we have
a memory leak in there.
I just reran our leak tests and they didn't report anything (which is
good, but doesn't completely rule out any leaks).
I did see this though from valgrind:
Object at 0x94e3410 of 68 bytes from an IgnoreObject() has disappeared
Does anybody know what valgrind it trying to tell me with that? Is it
a problem?
Robin