Skip to Content.
Sympa Menu

perfsonar-user - Re: [perfsonar-user] measurement archive disk (and system) sizing recommendations?

Subject: perfSONAR User Q&A and Other Discussion

List archive

Re: [perfsonar-user] measurement archive disk (and system) sizing recommendations?


Chronological Thread 
  • From: Andrew Lake <>
  • To: perfsonar-user <>, "Christopher J. Tengi" <>
  • Subject: Re: [perfsonar-user] measurement archive disk (and system) sizing recommendations?
  • Date: Tue, 12 Jan 2016 09:57:34 -0500

Hi,


The specs you mention(4 cores and 8GB) I *think* will be fine, though it may be worth bumping the RAM to 16GB if you can. It also depends on how many tests each host will be running. The same is true for the disk space. The latency tests are especially important since they are storing new data every minute, as opposed to throughput tests which store just a few data points a day. Doing some very rough math based on the current disk usage of a few of our hosts you are probably looking at around 5MB per month for each latency test, and about 1MB per month for each throughput test. Cassandra does a pretty good job compressing, so it’s not huge, but can add up obviously. It’s not uncommon for latency hosts to run a couple hundred tests, so if you are gonna do a full mesh or similar, the results could add up over time.

You may have seen this, but there’s some good info on running a central MA such as clustering and managing the data under this heading: http://docs.perfsonar.net/index.html#central-measurement-archive

Hope that helps,
Andy




On January 11, 2016 at 2:22:55 PM, Christopher J. Tengi () wrote:

Greetings,
We have been running a few toolkit hosts and a measurement archive for a few months now, and plan to scale things *way* up in the not-too-distant future. I have the MA on a fairly hefty Dell server, and would like to move it to a VM - it does no measurements of its own. The hardware that the MA currently lives on has 2, 8 core CPUs (2.4 GHz), 24 GB of RAM, and around 7 TB of disk. Kind of overkill for what we’re doing, but what I had not in use at the time. :-)

I’m looking for a more realistic configuration on a VM for a system that will act as the MA for around 200 (eventually) low-cost mesh nodes at building entry points, as well as a few higher-performance hosts in the data centers and on the Science DMZ, once we deploy one. The VM would also be running MaDDash for the on-campus mesh as well as one or two (or 3) wide-area meshes. Note that the members of the wide-are meshes would be using their own MA, rather than this VM. Based on what I’ve read so far, I’m assuming that 4 cores and 8 GB of RAM would be more than sufficient for these purposes, but I’m not sure about the disk requirements.

The Toolkit documentation talks about 100 GB of disk space, with the assumption that old data is not automatically cleaned from the local MA. If I had a 200 node on-campus mesh doing latency (10pps) and bandwidth measurements (20s every 2h), is there a rule of thumb I could use to determine how much disk to put on the MA if I wanted to keep the data for a year? Six months? Three months? As for the CPU and RAM, if I’ve drastically over-specified or under-specified, I’d like to know. What has your experience shown?

Thanks,
/Chris




Archive powered by MHonArc 2.6.16.

Top of Page