This archive contains answers to questions sent to Unidata support through mid-2025. Note that the archive is no longer being updated. We provide the archive for reference; many of the answers presented here remain technically correct, even if somewhat outdated. For the most up-to-date information on the use of NSF Unidata software and data services, please consult the Software Documentation first.
David Wojtowicz wrote: > > I went to the boss this morning and got permission to buy a replacement > server. Just ordered it now. Dual Athlon MP 1900+ processors, > 2GB of RAM, 100GB of fast disk. That'll hopefully help. :-) > > --david > That sounds good! Glad to hear it. Anne > >Hi David, > > > >David Wojtowicz wrote: > >> > >> Hi Anne, > >> > >> Actually, I was reporting a problem primarily with flood.atmos.uiuc.edu > >> which serves NMC2 and NNEXRAD|FNEXRAD. squall.atmos.uiuc.edu does > >> everything else. To answer some of your questions... > >> > > > >Excuse me for confusing flood with squall. > > > >> The load on flood can range from 3.0 to 10.0+ depending on the time > >> of day. There are presently 22 rpc.ldmd's running. If you look at > >> top there are usually 5-6 in the "runable" (R) state, rather than > >> sleeping. That means this many of them are awaiting service by the > >> processor. > >> > > > >This does indicate that the CPU isn't fast enough for the load. > > > >> There are no topology listings of NNEXRAD or NMC2 like there are with > >> MCIDAS or FOS so I'm not sure who all is actually using us at the moment > >> since not, everyone who has an allow line is using it at any given time. > >> > > > >Here are a few command lines to give you a close approximation of who is > >currently feeding from you. In the logs directory, do: > > > > grep feed ldmd.log | grep Exit | awk '{ print $5 }' | sort > exited > > grep feed ldmd.log | awk '{ print $5 }' | sort | uniq > feeding > > diff feeding exited > > > >The diff will report the processes that are still feeding from you. > > > >Here is similar info (cleaned up) from the stats you are sending us: > > > >TOPOLOGY > >flood.atmos.uiuc.edu redwood.atmos.albany.edu NMC2 > >flood.atmos.uiuc.edu ice.atmos.uiuc.edu NMC2 > >flood.atmos.uiuc.edu cirrus.atmos.uiuc.edu NNEXRAD|FNEXRAD|NMC2 > >flood.atmos.uiuc.edu measol.meas.ncsu.edu NNEXRAD|FNEXRAD > >flood.atmos.uiuc.edu waterspout.cst.cmich.edu NNEXRAD|FNEXRAD > >flood.atmos.uiuc.edu anvil.eas.purdue.edu NNEXRAD > >flood.atmos.uiuc.edu yin.engin.umich.edu NNEXRAD > >flood.atmos.uiuc.edu ldmdata.sws.uiuc.edu NNEXRAD > >flood.atmos.uiuc.edu twister.sbs.ohio-state.edu NNEXRAD|FNEXRAD > >flood.atmos.uiuc.edu aeolus.valpo.edu NNEXRAD > >flood.atmos.uiuc.edu zelgadis.geol.iastate.edu NNEXRAD > >flood.atmos.uiuc.edu squall.atmos.uiuc.edu NNEXRAD|FNEXRAD > >flood.atmos.uiuc.edu papagayo.unl.edu FNEXRAD > >flood.atmos.uiuc.edu data2.atmos.uiuc.edu FNEXRAD > >TOPOEND > > > >Looks like you are feeding CONDUIT to two local sites, as well as > >Albany. > > > >It would be interesting to know how much NEXRAD data your downstream > >sites are requesting. We can get some sense of that here from an > >analysis of the incoming stats. You can also grep through your logs for > >'Start' or 'RECLASS' to see what feed types and patterns sites are > >requesting (but if the connection has been stable and products timely > >for a while you might not see any of these). > > > > > >> I'm guessing that it is the bottleneck, especially at the very busy > >> 12Z run distibution via CONDUIT. > >> > >> I'm concerned both that we are not servicing our downstream sites properly > >> and are losing CONDUIT products ourselves when the latencies exceed 1hr. > >> > > > > > >I can see that redwood is only getting a small percent of all the > >CONDUIT products you're getting (are they asking for less?). And, the > >delay between your site and redwood's site is significant - I'm seeing > >10 to 15 minutes. > > > >Regarding NEXRAD, both umich and purdue are timely, although they must > >not be requesting a lot. > > > > > >> Again, the machine does only relay and nothing else (no pqact or > >> other significant processes) It is a 400Mhz PC machine running Linux. > >> Granted, this is "slow" now, but was state of the art when we bought it > >> for CONDUIT about two years ago. So I wondering what specs I need to > >> be looking for in a replacement or if that would even help if the machine > >> is not the bottleneck. > >> > >> Thanks. > >> > >> --david > >> > > > >Yes, your machine seems overloaded. And I suspect it is introducing > >further latencies into the CONDUIT feed. But, you are receiving and > >relaying *a lot* of data. > > > >Regarding a replacement, for a "relay only" machine I would first > >consider CPU speed and RAM. Having enough RAM to hold a sufficient > >queue is important. Disk speed is important for the "read/write > >through" aspect of the queue, so a fast disk certainly wouldn't hurt. > > > >For CONDUIT data, the biggest hour is currently 1.2Gbytes. NEXRAD is > >now averaging 80Mbytes/hour, so you need more to handle a maximum hour. > >So, a queue size of 1.4Gbytes should accomodate both. For this volume, > >I'd say a gigabyte of RAM or more would be best. Room to grow is always > >good. We have several sites succesfully running with nearly 2Gbyte > >queues. > > > >How big is your queue now? And how much RAM do you have? And, how old > >is the oldest product in your queue? (Use pqmon to determine the > >latter. The last field is the age of the oldest product in the queue in > >seconds.) > > > >Regarding your current configuration, Chiz said he could have redwood > >feed elsewhere if that's better for you. Or, could one of your local > >sites relay to the other to reduce the load on flood? Or, we could > >reduce the NEXRAD feeds, although I'm guessing that's not making a big > >impact on your machine, assuming the other NEXRAD sites are being > >judicious like Purdue and UMich. > > > >Anne > >-- > >*************************************************** > >Anne Wilson UCAR Unidata Program > >address@hidden P.O. Box 3000 > > Boulder, CO 80307 > >---------------------------------------------------- > >Unidata WWW server http://www.unidata.ucar.edu/ > >**************************************************** > > -- > | David Wojtowicz, Sr. Research Programmer > | Department of Atmospheric Sciences Computer Services > | University of Illinois at Urbana-Champaign > | email: address@hidden phone: (217)333-8390 -- *************************************************** Anne Wilson UCAR Unidata Program address@hidden P.O. Box 3000 Boulder, CO 80307 ---------------------------------------------------- Unidata WWW server http://www.unidata.ucar.edu/ ****************************************************