Flattening Out The Linux Cluster Learning Curve 89
editingwhiz writes "IT Manager's Journal has a good look at a forthcoming book, The Linux Enterprise Cluster, that explains in clear, concise language how to build a Linux enterprise cluster using open source tools. Writer Elizabeth Ferranini interviews author Karl Kopper in a Q&A. Is this more complicated than it appears to be? (IT Manager's Journal is part of OSTG.)"
don't you mean... (Score:3, Insightful)
Re:don't you mean... (Score:1, Insightful)
if they swapped the axis (Score:1)
Re:don't you mean... (Score:5, Funny)
Thanks to this book the learning curve has been flattened down to something more appreciable and amenable to those who have complained about the problem. The curve has been flattened far enough that it takes two years to learn that clustering "will likely require more than one computer to operate correctly" (Chapter 403 pg. 8729). I count this as a big win for society.
Ignore the anonymous coward who replied before me.
Steep prerequisite curve (Score:1)
"steep prerequisite" curve
i.e. each advancement to the learning process requires a much higher prerequisite.
If step 1. requires a high prerequisite then you would get the "running into a wall" effect.
This is the kind of book we need... (Score:5, Funny)
Re:This is the kind of book we need... (Score:2, Informative)
Re:This is the kind of book we need... (Score:1, Interesting)
Oh for fuck's sake spare the geeky overliteral bullshit and grow a sense of humor and perspective. Thank fuck it's not GEEKS who are it manager's but true manager's, heaven help us if someone asked a geek to ever look at the big picture in an organisation.
"I fail to see how a examining a large painting would enhance our productivity".
Turn your coder brain off once in a while.
Re:This is the kind of book we need... (Score:3, Interesting)
Looking for a good book on High-Availability clusters would be so much simpler
Re:This is the kind of book we need... (Score:2, Insightful)
They're the one's that can get funding and support for you to put one together.
Re:This is the kind of book we need... (Score:2)
I proposed a minor change to the products we subscribe to from our ISP that would save money and I'm still fighting to prevent it from going to a committee to decide if we should get quotes from other ISPs.
Re:This is the kind of book we need... (Score:1, Funny)
Re:This is the kind of book we need... (Score:2, Interesting)
Okay, the classic beowulf cluster is a 4x4 matrix of computers. Now, to have a beowulf of beowulfs, each of those computers on a cluster must be connected to its own 4x4 grid, so you now have a cluster of 256 computers, arranged somewhat suboptimally. Now, in order to communicate with these
OSTG? (Score:5, Informative)
Re:OSTG? (Score:2)
Slashdot -> VA Linux -> VA Software -> OSDN -> OSTG
Simpler times, those were.
Let me get this straight (Score:3, Insightful)
Dont think so.
Re:Let me get this straight (Score:1)
Re:Let me get this straight (Score:2)
Re:Let me get this straight (Score:1)
Very nice (Score:5, Informative)
A quick Google search [google.com] though reveals a lot of free papers and manuals on this very topic.
Re:Very nice (Score:2)
The problem with clustering in Linux... (Score:5, Interesting)
Lars Marowsky-Brée had a paper in the proceedings of OLS 2002 [linux.org.uk] describing the problem and a suggeted solution in his paper entitled "The Open Clustering Framework". I'm not sure how far standardized clustering has come since then. Anyone has any insight on the matter?
Re:The problem with clustering in Linux... (Score:4, Interesting)
What I gathered from my introductory Operating Systems class was that this was the next frontier and exciting market to keep an eye for.......that and creating applications for these setups was not as you said, standarized yet. Can Linux applications that normally run a single box setup migrate relatively easily to a cluster setup yet?
Re:The problem with clustering in Linux... (Score:2, Interesting)
I use a PXE boot system with Anaconda kickstarts to get the software installed. A poast install script then configures everything else on the machine. When it reboots, the machine appears in the cluster and is ready to use. I use the Torque batch scheduling system.
You don't need the cluster toolkits to setup a cluster! DHCP, TFTP and a configured kickstart file
Re:The problem with clustering in Linux... (Score:2, Interesting)
Some types of applications, it's easy to visualize how to get a dozen or a hundred computers to help with the problem (serving static web pages). Others, it's not (databases)
Re:The problem with clustering in Linux... (Score:1)
Databases may not be easy, but they have been done (http://www.teradata.com/ [teradata.com]). Not sure if I'd call it a "typical cluster", but I can't say that any cluster is typical. Current Teradata stuff doesn't do Linux, but they're going there [teradata.com] as I type.
Re:The problem with clustering in Linux... (Score:1)
I've got a left-handed crescent wrench to help fasten that ID-10T wire...
Re:The problem with clustering in Linux... (Score:1)
Still won't help them with they're nuts...
Re:The problem with clustering in Linux... (Score:1, Interesting)
OpenMosix is a clustering technology that allows you to use regular apps and benifit from the cluster.
It works by migrating proccesses from one computer to another. So it's like a SMP machine, the fastest any single thread can be done is limited by the fastest cpu, however the it allows you to do more at once.
For example with a 2 system cluster, you would be compiling the kernel. If you set it up only to use one thread at a time, then you get 100% of the original single mach
Re:The problem with clustering in Linux... (Score:2)
Re:The problem with clustering in Linux... (Score:1)
Microsoft is the problem.
Sick of gentoo zealots throwing plugs in completely unrelated topics? Me too! {QUOTE}
This style of comment is the other.
S.
Re:The problem with clustering in Linux... (Score:2)
Logistics gone digital? (Score:2, Interesting)
He says, "Because we couldn't print the forms for the warehouse people to select the products to be put on the truck, we'd have dozens of truck drivers sitting in the break room each day for more than 10 minutes.
I actually don't get it, most logistics got wireless for about a decade now...
and the truck driver has no right for a break...
Re:Logistics gone digital? (Score:2, Insightful)
If the system is taking a long time to do the work required to print a form how is wireless going to help?
and the truck driver has no right for a break...
What do you propose the driver does? Go drive around the block for 10 min until they are ready to load?
Publication, standardization, multiplication (Score:4, Insightful)
Publications like this play an important role in establishing best practices and community, two key enablers of standardization.
These in turn will lead to greater adoption, and more publications. A virtuous cycle.
VMS clusters (Score:4, Interesting)
And please, don't be put off by VMS because DCL = your first exposure to a VMS system - feels more awkward than bash (in many ways, it certainly is!). It's in the underlying architecture of the OS where the fruits of tight engineering are really demonstrated.
Re:VMS clusters (Score:3, Informative)
http://gnv.sourceforge.net/
includes a port of bash to VMS. Not sure how good it is.
Having used and programmed DCL, it's not that bad.
h
Re:VMS clusters (Score:1)
Clustering VMS/VAXen was straight-forward, reliable, fast and exceedingly well-supported by DEC (for a fee anyways).
VMS is dying .. yeah right (Score:2)
1) It's been ported to Itanium, and will ship on that platform very soon.
2) ISV support for VMS on Itanium is strong and its customers are very loyal.
3) You said you wish it would die and allow "better systems" to take over. What better systems? In the context of clustering (cos that's what this discussion is about) VMS Clustering for High Availability systems is still at the top of the food chain. Nothing exists on Linux to match it. Not yet anyway.
Tru64 Unix clustering is about the next best thing,
Unless i read this incorrectly. (Score:2, Informative)
Re:Unless i read this incorrectly. (Score:1)
All enterprise clusters I know of are in the 2 to 10 node range. The only reason there is a cluster is because of automated failover in the case of node or site failure. Performance and scalability is of no importance, they just buy a bigger box if necessary.
Some of the sites have hundreds of machines, all clustered in small, manageable units. These are high-end IBM AIX boxes, high-end fiber storage with two sites 50km apart.
Enterprises buy clusters for availability, not for scalability !
Markus
Re:Unless i read this incorrectly. (Score:2)
You don't need 20 nodes if 6 can do the job.
Re:Unless i read this incorrectly. (Score:1)
Editor needed (Score:1, Informative)
Re:Editor needed (Score:1)
Mandrake CLIC (Score:5, Informative)
However, last january we set up a small (six node) cluster with the help of CLIC [mandrakesoft.com]. Once we realized the link between a Mandrake and consective dead CD drives [newsforge.com], we installed the cluster in little time.
CLIC might focus a little too much on userfriendlyness and a little too little on flexibility, but for our purposes it was great. It sports ganglia, gexec, distcc and MPI (and probably more), and administration and deployment of nodes is a breeze.
I heartily recommend CLIC for student/test/proof-of-concept projects.
I had to read that twice before I got it (Score:1)
Beowulf Newbie Question (Score:3, Interesting)
Is there any plans to take beowulf in this direction? Is it already possible, but I was just reading the wrong FAQ?
Re:Beowulf Newbie Question (Score:1)
I'm by no means an expert, but I was under the impression that a cluster of yesterday's computers would easily be outperformed by a single top of the line computer. So, except for fun and learning, clustering with old computers is just a waste of time.
Re:Beowulf Newbie Question (Score:1)
If it wasn't so, then it wouldn't be a feature of Oracle, SQL Server, DB2, etc...
Re:Beowulf Newbie Question (Score:4, Informative)
Inevitably high-performance clusters require software designed to run on high-performance clusters. It is better not to think of such a cluster as a single system, but rather as a network of individual machines with a tight network connection. Some of the clustering add-ons for linux approach and even achieve certain aspects of a "Single system image" type of configuration, but it's never completely like a single system.
Back in 1997 or so I tried to get as close as I could to a true Single System Image by building off of the beowulf patchsets combined with patches for Distributed SysV IPC/SHM and a globally-shared root filesystem using CNFS (cluster-nfs, so that a few essential configfiles can have unique copies per cluster node). It was very daunting work to get those patches integrated together, and the end result was that without some kind of network-interconnect that was as high-speed and low-latency as a processor's FSB, there was always going to be a big performance hit doing things this way. Of course if an application happens to be perfect for simple HPC clusters (all cpu intensive, very little I/O, and the work is easily divisible without tons of IPC between the workers), then it runs fantastically on such a Single System Image cluster, but then again it would have run fantastically on a simple cluster that doesn't look like a Single System too. So what the Single System concept bought me really was a nice abstraction layer that made everything easy to deploy, configure and manage. But it came at a severe initial cost of human labour. It's not worth the trouble.
This is what you were looking for (Score:2)
OpenSSI (Single System Image) Clusters for Linux [openssi.org]
The main features are: single root and single init, cluster filesystems and DLM, single process space and process migration, load leveling, single and shared IPC space, device space and networking space, and single management space
Re:Beowulf Newbie Question (Score:3, Informative)
Re:Beowulf Newbie Question (Score:1)
Re:Beowulf Newbie Question (Score:2)
Another very important thing to remember is power consumption and cooling. You might be able to get fifty PI
SSSShHHHH!!! (Score:2)
Please stop misusing the term 'learning curve' (Score:5, Informative)
A flat learning curve is a bad thing.
The term "learning curve" was invented by the aerospace industry in the 1930s as a way to quantify improved efficiency from mass production (basically, the more you do a task, the easier it becomes). The term was later adopted by psychology and the social sciences, where most people first encounter it.
In both cases, the horizontal axis of a learning curve represents time or effort, and the vertical axis represents amount learned or productivity. Therefore something that is intuitively obvious in fact has a steep learning curve.
"Learning curve" was a technical term with a specific definition for decades before it was ever a (misused) marketing buzzword.
Thank you for your time :)
Bzzt (Score:1)
Aren't you confusing "learning curve" with "economy of scales" here?
Re:Bzzt (Score:2)
Nope. [nasa.gov]
Unless... (Score:2, Funny)
Re:Please stop misusing the term 'learning curve' (Score:1, Informative)
Re:Please stop misusing the term 'learning curve' (Score:2)
Wrong - the American psychologist E.L. Thorndike was using learning curves in 1911 or before
I did not know that, thanks!
wow, thanks (Score:2)
I can't wait to beat someone with my new knowledge tomorrow! thanks!
little advice (Score:2)
I'd want to do more than just loadbalance webservices with it.
It want shell accounts on this cluster that act as one maineframe. People would shell into their home directories (which I suppose would all be from one big NFS), and run processes and whatnot on the entire cluster.
Any ideas?
Love zaq
Re:little advice (Score:2)
Yeah - go to Google or some clustering forum/mailing list.
Re:little advice (Score:3, Informative)
The second part with shell acounts and home directories are all problems already solved by NIS/NFS. You could setup a pool of machines that all share the same NIS/NFS info so anywhere the user logged in they'd have the same files/passwords, and load balance it via ipvs or dns.
AFAIK the current state of clustering works well for custom code situations, where you write your app to run on the cluster, but doesn't transparently make your 4 b
apps not designed for cluster with lots of state? (Score:3, Interesting)
I don't know about NFS, but in the case of CIFS, the protocol spec has provisions for renegotiating locks if a connection is broken, but I don't know if there are bugs in win2k/XP clients with samba 3 servers. OpenAFS can have a sort of active/active setup, but the archatecture is such that there is only one server that handles the writes and the rest are read-only. In all of these you can have a semi active/active failover cluster if you move half of the active volumes to the backup server, but this adds a lot to the complexity of your fail-over system.
Those services have a low to moderate amount of state information kept on the server. In the case of a graphical (VNC) terminal server, I don't know of any open source projects that will allow gnome session to be on one server, have that server go down, another server take over its ethernet MAC and IP address and continue processing where it left off on the backup server. The best I can think of is OpenMosix or maybe OpenSSI which are two single system image type clustering systems. If anyone knows anything, please reply and let me know thanks.
Good idea (Score:1)
Comments From the Author (Score:2, Interesting)
First, about the book being a "definitive" guide. I cannot possibly claim to be an expert on every topic in the book--in fact, no one person can. The book is definitive, however, in that project leaders from each of the open source projects participated in editing and reviewing the material for the book.
It is an over broad statement to say it is the definitive guide for building any and all types of Linux Clusters. The book describes how to build a cluster