[rrd-users] Re: Updating 1000's of rrd files
Jamie Lerner
jlerner at jjlabs.com
Tue Jul 6 02:26:10 MEST 2004
I too have the same problem, we are updating 40,000 - 120,000 rrd files
every five minutes and have found the following.
The systems become IO constrained quickly because each RRD file is opened,
edited then closed, which is very IO intensive for the OS. You can do the
following to help speed it up:
RAID 1 plus 0 is much faster than RAID 5
Faster drives help (15,000 RPMs helps)
Memory helps (we use between 6 - 8 GB) but if you are looping every 5
minutes, the system can never catch up and eventually crashes anyway.
The real solution is to build some type of RRD server that can manage the
file IO and cacheing like a database does. Is there any technology like
this for RRD files? If not, we may consider moving from RRDs to a database
in order to get faster IO...something I would prefer not to do.
Jamie.
----- Original Message -----
From: "Tobias Oetiker" <oetiker at ee.ethz.ch>
To: "Seth Hettich" <shettich at gmail.com>
Cc: "rrdtool-list" <rrd-users at list.ee.ethz.ch>
Sent: Sunday, July 04, 2004 10:49 PM
Subject: [rrd-users] Re: Updating 1000's of rrd files
Yesterday Seth Hettich wrote:
> So, I have a problem:
>
> I need to collect some data from ~2,000 to 3,000 devices (perhaps more)
> and put it into rrd files. about 100 files per device. The code to
> collect all the data can run in ~2-3 min. Also, code to update one
> rrd file ~5,000 times can run in ~20 sec.
>
> But, code to collect all the data and update the rrd files take over an
hour
>
> I assume this is due to the time to update so many files (in the test case
> it's just one file over and over, so it's very fast).
>
>
> I assume it would go much faster if I could put all of the data
> sources for each device in one file, but new data sources will get
> added all the time, and I see no way to add a DS to an rrd file.
>
> Anyone have any ideas for how to make this work?
In my tests when writing rrdtool I saw that the number of files you
can update efficiently is tighly coupled to two things:
a) it only works fast if you use local disk (nfs is normally very
slow for this due to its 'write to disk first' behaviour
b) the number of files is tied to the amount of memory available
for caching ... on our old sun server which I used for testing
then, the rate of update dropped more than 10 fold once I went over
700 files ... it went down to 20 updates a second ... These days
everything will be faster, but I assume the basic problem remains
....
some systems also exhibit problems with having to deal with many
files in a single directory, so you might want to split this out
...
cheers
tobi
cheers
tobi
> -Seth
>
> --
> Unsubscribe mailto:rrd-users-request at list.ee.ethz.ch?subject=unsubscribe
> Help mailto:rrd-users-request at list.ee.ethz.ch?subject=help
> Archive http://www.ee.ethz.ch/~slist/rrd-users
> WebAdmin http://www.ee.ethz.ch/~slist/lsg2.cgi
>
--
______ __ _
/_ __/_ / / (_) Oetiker @ ISG.EE, ETZ J97, ETH, CH-8092 Zurich
/ // _ \/ _ \/ / System Manager, Time Lord, Coder, Designer, Coach
/_/ \.__/_.__/_/ http://people.ee.ethz.ch/~oetiker +41(0)44-632-5286
--
Unsubscribe mailto:rrd-users-request at list.ee.ethz.ch?subject=unsubscribe
Help mailto:rrd-users-request at list.ee.ethz.ch?subject=help
Archive http://www.ee.ethz.ch/~slist/rrd-users
WebAdmin http://www.ee.ethz.ch/~slist/lsg2.cgi
--
Unsubscribe mailto:rrd-users-request at list.ee.ethz.ch?subject=unsubscribe
Help mailto:rrd-users-request at list.ee.ethz.ch?subject=help
Archive http://www.ee.ethz.ch/~slist/rrd-users
WebAdmin http://www.ee.ethz.ch/~slist/lsg2.cgi
More information about the rrd-users
mailing list