Slashdot is powered by your submissions, so send in your scoop

 



Forgot your password?
typodupeerror
×
Linux Software

Putting Linux Reliability to the Test 296

Frank writes "This paper documents the test results and analysis of the Linux kernel and other core OS components, including everything from libraries and device drivers to file systems and networking, all under some fairly adverse conditions, and over lengthy durations. The IBM Linux Technology Center has just finished this comprehensive testing over a period of more than three months and shares the results of their LTP (Linux Test Project) testing."
This discussion has been archived. No new comments can be posted.

Putting Linux Reliability to the Test

Comments Filter:
  • Linux Test? (Score:5, Funny)

    by Anonymous Coward on Saturday December 27, 2003 @01:07AM (#7816016)
    Just put a link to each box on /. and wait 24 hours.
  • Of course, we all knew this in our hearts, nice to see it in writing.

  • by Anonymous Coward on Saturday December 27, 2003 @01:12AM (#7816031)
    Anyone know if the test will be repeated with kernel 2.6.x?
  • USE BAD HARDWARE! (Score:5, Insightful)

    by superpulpsicle ( 533373 ) on Saturday December 27, 2003 @01:15AM (#7816039)
    You want to put any OS to the ultimate test, you should run cheap generic hardware. I swear it's an industry conspiracy that generic parts struggle a boat load. If your parts don't come from the big boys (DELL, gateway, etc), you are likely going to see issues down the line.

    Get some ECS motherboard, generic RAM... bang. You're in for the evening.
    • by spikev ( 698637 ) on Saturday December 27, 2003 @01:47AM (#7816123)
      IMHO, it's the big boys that have the conspiracy to sell crappy hardware. Try performance testing almost any (PC Chips mobos don't count) custom system against a Dell with a similar hardware configuration, and you'll see what I mean.

      I've done it with my ECS board with generic ram, and I came out on top.

      It's the big computer makers that sell the cheap generic hardware. Try getting anything that's essential and non-OEM, hardware or software, to work with one of those boxes.
      • Re:USE BAD HARDWARE! (Score:2, Interesting)

        by router ( 28432 )
        Exactly. How the fsck do you think they manage to have a whole corporation hidden in the razor thin margins that exist on commodity hardware? By cutting everything down to the wire. Dell et. al. are like the major automakers, saving a quarter on every piece is 3 million on the bottom line....

        andy
        • Before you roll something out into production, you hammer on it for a while. The crappy parts will fail or generate errors and you can have them replaced.

          In my experience, most of the no-moving-parts hardware will fail within the first week, or last for years and years.

          The stuff with moving parts will eventually fail. But that's harder to predict.
      • No conspiracy at all, just profit margins . I don't have to worry about that as an individual, so I do OK and still save with decent parts. The hour or so of labor to assemble it all isn't much to me.
        IOW, I agree - pick decent parts and get *exactly* what you want. I usually pick the previous generation CPU and get the biggest mobo I can for that from trusted brands. Then I stuff the mobo with the most it can handle, which is a *lot* nowdays. Of course, I get it all below retail from local OEM's, cash pai
      • The funny thing is....

        ECS motherboards == PC Chips motherboards

        lol

        Oh well, they still make half decent parts provided you select the correct chipset.
    • I used to own (back in the day) a really cheap motherboard + all shit integrated on it (vga, lan etc.).

      Really, it sucked, windows sometimes crashed on boot on the thing, reinstalling the thing didn't help at all. Reliability was gruesome... until I installed Linux (this machine was one of the reasons I switched). I never had one single crash while the machine was running Linux, running it on Windows (dual-boot) was still as unreliable as ever.

      Offcourse this is only one story.
      • Re:USE BAD HARDWARE! (Score:3, Informative)

        by bhtooefr ( 649901 )
        Try running Memtest86 on something like that. It might be the RAM, as Windows handles RAM differently from Linux, and can hit bad parts sooner than Linux (and vice versa).
        • I have to second this recommendation. I'm sitting in front of the only system on which I've ever seen Linux crash completely. I'm talking about seeing kernel oops and kernel panic messages regularly. Memtest86 showed that one of my SIMMs was half bad. Until I started using RAM about the 256MB mark, everything worked.
    • by Anonymous Coward
      But then you have a linux on crappy hardware test. Not an OS test. If you want to test the OS, you have to minimize the impact of other factors.
      Otherwise you may have bad OS stability because you have a bad hardware constellation.
    • Why whould you ever run a server on some random ECS motherboard (not that they are particualry bad but for arguments sake) As allways if you want reliable get a server motherboard and powersupply with some nice ecc ram and a raid 5 set. Generic PC's are just that junk now granted Linux works extreamly well with generic PC's proably because thats what it gets tested on most. Now MS and IBM general have programmers of workstation class boxes that are closer to servers in there design (Realy the only differe
    • Well, you'd end up with a different kind of test.

      The value of this test is for people considering Linux for mission critical applications. That is to say the kind of bet-the-company applications that people don't run on a surplus box snatched out of a basement storage room.

      That said, the kind of "hey kids lets put on a show" projects are where open source really shines. I'd swear that if we had to make the decision to pay money up front, my company still wouldn't have an email server, a web server, an i
      • Yes, I know what you mean. We dealt with a mail server application for years because it was really cheap, even after the company that wrote it went under and what little support we had disappeared. Even at its best this program *SUCKED* but hey, it was really cheap. The only reason I was able to push a new system was because the damn program timed out! Can you believe it? Even though it was bought and paid for it timed out anyway. The guy assigned to administer the system had to keep resetting the ser
    • Well see, thats the catch. If you ran just about anything on that quality hardware it would run reliably. As for your average Joe Linuxuser, they arent too likely to have a pSeries sitting around their house.

      So essentially this isnt a test of Linux, this is just a test of the pSeries hardware.

      Now IBM can breathe easier, knowing they can just use Linux instead of having to pay people to support AIX. This is a great day for the pointy-haired boss!

    • Get some ECS motherboard, generic RAM... bang. You're in for the evening.

      There's plenty of good hardware to be had from places like Newegg, Directron and Computer Geeks. Just to name a few. Get yourself an ASUS motherboard, RAM from Crucial or from a reputable manufacturer like Kingston ValueRAM or Viking or Mushkin or Corsair, get a video card from a good manufacturer, and you have a nice solid machine that can handle anything.

  • by davidstrauss ( 544062 ) <david.davidstrauss@net> on Saturday December 27, 2003 @01:18AM (#7816050)
    Why do you trust IBM's Linux Technology Center to evaluate Linux?
    • Because what would they have to gain by lying? The true power of opensource is that when someone does point out the weaknesses, they are fixed quickly. IBM knows that if they tell the opensource world "Hey, LINUX is pretty good but it kinda struggles in the (foo) area." that the opensource community will redouble their efforts to fix that. Microsoft is only trying to say "Windows rules, Linux sux. See, this evaluation we did proves it. Buy windows"
      • by davidstrauss ( 544062 ) <david.davidstrauss@net> on Saturday December 27, 2003 @01:50AM (#7816128)
        Because what would they have to gain by lying?

        They have much to gain: more corporate customers and more respect and funding by greater IBM. Just because IBM supports Linux doesn't mean its motives are pure (not financially driven). Another reason for bias is the division also stood to have huge setbacks if the tests were unfavorable. How could they justify expansion and better funding if their previous statements about Linux being enterprise-ready were unfounded?

        • by Surazal ( 729 ) on Saturday December 27, 2003 @02:02AM (#7816163) Homepage Journal
          So, ya reply to one point but ignore the rest? I think his (ultimate) point is valid. If the test was rigged, the folks involved with developing the kernel would catch on and take IBM to task for fudging the results. No, I'm not talking about the Slashdot/Fark crowd. I'm talking about REAL developers.

          Also, Linux has weathered some unfavorable (and honost!) critiques before. Linus Torvalds said it best when he said (and I paraphrase since I am too lazy ATM to look up the actual quote) that it doesn't matter if there's negative publicity in the press about Linux. It just meant he got his bug reports from the Wall Street Journal as opposed to the regular kernel mailng list.
          • So, ya reply to one point but ignore the rest?

            I think the rest of your post is either biased (assuming IBM only intends this article for a technical audience that won't use it for business decisions) or inflammatory (the comments on Microsoft that mostly have root in the attitude on Slashdot more than Microsoft's actual statements).

            ...and take IBM to task for fudging the results

            I need something more empirical than this assertion. I don't see kernel developers spending weeks duplicating IBM's results jus

    • Why? Here's why... (Score:5, Interesting)

      by Crypto Gnome ( 651401 ) on Saturday December 27, 2003 @01:57AM (#7816149) Homepage Journal
      • because the test methodologies are documented
      • because it's disclosed up-front that it's IBM Linux Team testing Linux (ie no hidden conflict of interest
      As opposed to the usual (ie in the Microsoft World)
      • ZDNet (and/or others) "testing" Microsoft Products (but only vaguely describing how things were configured)
      • Microsoft paying someone to "report" on the quality/performance of a Microsoft product, but the evaluation is worded in such a way as to convince the user that it's an independent review and the "funded by microsoft" fact is never mentioned anywhere in the evaluation
      • because the test methodologies are documented

        Yes, they are documented, but some of the evaluation criteria ("expected behavior") depend on the opinions of the team performing the evaluation.

        because it's disclosed up-front that it's IBM Linux Team testing Linux

        Yes, that's better than the Ziff-Davis test (which I'm familiar with and mention in another post on this thread). However, assuming no bias because of a disclosed possibility of bias is illogical. Such disclosure is a necessary but insufficient co

        • by Crypto Gnome ( 651401 ) on Saturday December 27, 2003 @03:05AM (#7816292) Homepage Journal
          OK, so the Reality Check in this equation amounts to:

          You should not trust this evaluation at all.
          • Go to the site [sourceforge.net]
          • download the testing tools yourself
          • read the test paper
          • use the test methodologies as documented
          • do your best to verify their test results yourself
          • go back to the site
          • post your results for everyone else to see
          (ie follow the good practices of basic science)

          After all... On the internet , nobody knows you're a dog.

          Any JimBOB can write a convinving paper, with all the right buzzwords, that sounds as if X+Y=Z, especially if that was logically a likely/expected outcome in the first place.

          As a well-known TV show once said (several times and loudly) Trust No-One.

          Remember people, YMMV.
    • Why do you trust IBM's Linux Technology Center to evaluate Linux?

      Because it is to IBM's advantage to find any weakness AND FIX IT before their customers run into the same problems. Any "insider knowledge" would be used to make the tests harder rather than easier. If it were "IBM Linux" rather that "SuSE Linux" that was being tested you'd have at least a chance of making a point.
    • Why do you trust IBM's Linux Technology Center to evaluate Linux?

      Because the goal of the test is to find out whether or not IBM's customers should feel comfortable using Linux instead of AIX or some other highly reliable OS for mission critical computing applications, and IBM will look really bad in front of its own loyal, big-money customer base if the test results don't hold up in the real world.

      More concisely: Because IBM will lose sales if Linux fails after IBM said it wouldn't.

    • Because it confirms daily experience?

      MS is running ads saying how windows XP is so reliable. It is kinda hard to believe when you hear the ad because you a getting a cup of coffee waiting for XP to reboot. Same with 2k3. It crashes. Not as often as XP same as XP doesn't crash as often as 98 and so on. But it still crashes.

      Now on to my linux machines. Wich don't crash. I only run in total about a dozen of them and not one of them has crashed.

      I also have had some experience with AIX. Typically on machines

  • by Quixote ( 154172 ) on Saturday December 27, 2003 @01:18AM (#7816051) Homepage Journal
    I skimmed over the article (heretic!), and was wondering: how do they distinguish between software failures (the purpose of the test) and hardware failures (for example, random bit errors in the memory that could be caused by higher temperatures due to the stress testing)?

    I seem to recall getting random crashes with cheapo memory, and it was a pain to track down the offending component. Of course, one would assume that IBM wouldn't go for cheapo components, but still: how does one point the finger at the software, instead of hardware? Is it just repeatability?

    • by Anonymous Coward on Saturday December 27, 2003 @01:33AM (#7816084)
      http://www.memtest86.com/ Freeware GPL bootable memory tester for PC platforms... highly recommeded for troubleshooting flaky RAM...
      • Minor quibble: I don't think memtest86 is going to do a whole lot of good on the systems that were being tested in this particular case, since the systems that were being tested were PPC, which doesn't run x86 software very well. But yes, it's a nice program to keep around for your x86 hardware. I got my copy from the BBC bootable linux business card CD.
        • Shame it crashes and burns on Athlon64 - anyone know a version that would run on the latest hardware?
          • Works here... There's even a native 64bit version I 'm told
            • Odd. The floppy-bootable Memtest86 3.0 certifiably instantly reboots the computer that uses MSI KT8 DELTA motherboard and Athlon64 3200+. It flashes the testing screen for about half a second and reboots the computer. Tested on three separate computers using same motherboard+CPU combination.

              Which is kinda teh suck when you try to figure out if you have a bad motherboard or just a bad stick of ram...

              So, could someone point me towards a version (CD/floppy-bootable pre-built one, if possible) that does work?
    • Hardware fails in random patterns (bits flipped by beta radiation, for example), software fails the same way in controlled instances (the same flawed logic fails the same way each time).

      So when you run a test 5 times, and you get 5 results, the hardware is broken. When you run the same test 5 times, and it gets to the exact same point before sig11ing, you have a software flaw.

      This is also why you do multiple tests to ensure you're getting an accurate picture of what's going on (flawed or not).
      • > When you run the same test 5 times, and it gets to the exact same point before sig11ing, you have a software flaw.

        Not necessarily: When uncompressing one of the XFree86 source tarballs, X430src-3.tgz, on my old k6 2-450, gzip would always die with a bad CRC. Nothing else at all seemed to go wrong with the machine, but I couldn't uncompress the file until I downed the memory clock to 66MHz, rather than 100.

        I found one other person with the same motherboard having the same problem in a google search,

        • A bad CRC on a really-big gzip file isn't a very granular test. A good way of stressing a system so that it will show you wether it fails reproducibly is to compile a couple different versions of the Linux kernel, and see it if fails in the same spot (or at all). If you ran memtest [memtest86.com] on it, it'd likely have shown an error. Whenever you aren't sure, run memtest :)
          • You consider multiple kernel compiles and a full blown memory diag proportional troubleshooting when you can't unzip a tarball?
      • by ImpTech ( 549794 ) on Saturday December 27, 2003 @03:17AM (#7816307)
        Bleh, thats not necessarily true at all. A good race condition in a many-threaded program can quite easily look very much like a hardware problem, in that it is difficult to reproduce reliably.
      • by jmv ( 93421 ) on Saturday December 27, 2003 @03:40AM (#7816351) Homepage
        software fails the same way in controlled instances

        That's true... in theory. In practice, there are many ways software can fail in random (in the weak sense) ways. Many of these are related to timing. For example if you have many threads and fail to lock things properly, the result will depend on when the tasks are preempted. You can also have different results because of the way the interrupts (disk, net, ...) happen. There's also the not-initialize type of problem where the behaviour depends on whatever was there in memory before. There are probably many other ways for software to fail at random, including obscure combinations of events.

        I'd say that the only kind of software that can't fail randomly is single-threaded and doesn't rely on any input other than regular files (and even then I'm not sure it's enough).
        • You can always go and find broken corner cases, but the vast majority of software is single threaded (or behaves that way), and in a testing situation will have normalized input. This leads to fairly regular errors. Obscure combinations of events don't usually happen with software that's mature, since it's been through many combinations and patched to follow them (IE: Apache won't die in a random fashion).

          I've put in many hours of debugging software (my own and others), and all the crashes and other prob
          • You can always go and find broken corner cases

            Right. That's where two bugs get together and you notice something.
            Remove either of the bugs and nobody will notice anything.
            Remove both of the bugs and maybe you get a bit closer to having something debugged.

            (IE: Apache won't die in a random fashion)
            One of us severely misunderstands Apache. My understanding is that it is quite possible to run Apache in production with some very broken modules.
            "Try it again. Maybe your browser has some problems." ;-)

            and v
      • So when you run a test 5 times, and you get 5 results, the hardware is broken. When you run the same test 5 times, and it gets to the exact same point before sig11ing, you have a software flaw.

        This isn't true. If you're running a program that uses a deterministic memory allocation algorithm (a compiler, for instance) and have a segment of bad memory, then you easily could crash at the exact same point (when a pointer in that segment is dereferenced, for instance).

        I know. It's happened to me. I've even
      • by MoogMan ( 442253 ) on Saturday December 27, 2003 @08:41AM (#7816805)
        You will find that a lot of the trickier bugs can depend on certain [eg. race-] conditions. Such things that are very hard to recreate, even under carefully controlled situations. Then you get the heisen-bug variety etc. Such errors could easily be passed off as hardware failiure. Im sure you can dream up your own examples. (I cant; Im still drunk)
      • If you run the test 5 times on 5 different machines. Then you can rule out software vs hardware errors. PRESUMING the hardware is known to be usually good. Remember the old pentiums and their error baked in? Good luck determining that one without some kind of reference material.

        And this is the real way to determine it. Run the suspect software on a piece of hardware known to be good or run a piece of software known to be good on suspect hardware. Testing anything with a single sample is meaningless. If you

    • The pSeries has hardware validation built in. It a hard shows failure, it is shutdown (if nessary) and hard tech is called.

      Yes, the box makes the hardware call. It is freak to come to work and have IBM service sitting at your door want into install a hard drive or I/O Card... and the machine is still running.

      And now most of the time, they change it while the machines is active.

  • Not bad (Score:5, Insightful)

    by changelingyahoo.com ( 543558 ) <<changeling> <at> <yahoo.com>> on Saturday December 27, 2003 @01:20AM (#7816056) Homepage Journal
    This is nice to hear, but it would be even more valuable if the same tests were performed on a variety of operating systems in order to compare the results.

    Brian
    • good point, and they have to be done under similar test conditions too.

      Stability of a stripped down linux boxen doing just what it was intended to it might be much better than that of windows XP (with all the bells and whistles) but i would really like to see both of them loaded with similar number of apps and look at how their performance match. From my experience, XP has been much better in that instance.
    • it would be even more valuable if the same tests were performed on a variety of operating systems

      How many operating systems run on IBM's pSeries machines? AIX and...?
  • Results... (Score:3, Interesting)

    by rmdir -r * ( 716956 ) on Saturday December 27, 2003 @01:23AM (#7816063)
    The Linux kernel and other core OS components -- including libraries, device drivers, file systems, networking, IPC, and memory management -- operated consistently and completed all the expected durations of runs with zero critical system failures. Every run generated a high success rate (over 95%), with a very small number of expected intermittent failures that were the result of the concurrent executions of tests that are designed to overload resources. How does that compare with other OS's?
  • by ducomputergeek ( 595742 ) on Saturday December 27, 2003 @01:38AM (#7816099)
    Well I have some Karma to burn tonight. I don't mind that its the 2.4 KERNAL even though 2.6 is ready. Why? We never put anything on our production server that hasn't been out for at least 6 months with exception of security upgrades.

    Second off, If this were M$ testing 2k3 and publishing the paper, everyone here would be crying foul. But because its, "Linux" it must be 100% unbais and true.

    I've been using Linux for 8 years now including under high stress enviroments, 3d graphics rendering mainly, and from experiance I have see very good things from Linux. We have had software glitches before, but the core software maybe has caused 3 - 5% of our downtime. Over 70% of our downtime involves human error and about 25% of failures are due to hardware giving out.

    Still what my customers are wanting to see isn't benchmarks as "So easy Grandma could use it" in Linux. While the people in the datacenters want to know how well Linux will bear under a load, most end-users and SMB's don't need to worry about it, they just need something easy to use that works.

    • Couple of small points to nitpick. First, you, as so many others, seem to think it's a novel idea to not implement the bleeding edge beta software on your critical hardware. No need to point this out; you presumably have a standard policy of testing all upgrades and patches on development hardware before moving it to your production equipment (or should, if you can afford it).

      Second, you're probably right about the publishers of the paper, but hey, what can you do? The people with the most interest in th

      • ok I've heard people say this before like it's an issue. But I've never actually seen ANYONE use windows on the high end servers or datacenters. I've seen windows in small server roles, but I've never heard of someone actually USING clustering capabilities in windows or going beyond what a single system can do. A quad processor x86 box is hardly high end after all.
    • by haruchai ( 17472 ) on Saturday December 27, 2003 @02:28AM (#7816225)
      Well, their bias is clearly stated on their web page- this is the Linux Test project - which is dedicated to evaluating the capabilities and limits of Linux.

      They aren't making a comparison to other OSes or saying that Linux is more suitable than such-and-such operating system; just that it is suitable for particular tasks or environments.

      A comparison between different OSes should be carried out by an independent testing facility but, in this particular case, I don't see anything
      wrong with their modus operandi.
    • by antiMStroll ( 664213 ) on Saturday December 27, 2003 @02:36AM (#7816243)
      Second off, If this were M$ testing 2k3 and publishing the paper, everyone here would be crying foul. But because its, "Linux" it must be 100% unbais and true.

      An ironic assertion regarding bias. IBM isn't the author of Linux or any of its tools, add-ons, servers, etc. as Microsoft is of 2k3 and its support software. Microsoft also has a long and distinguished history of FUD. IBM doesn't have anywhere near the historical attachment to Linux that MS has to Windows, and IBM hasn't been caught lying about it yet. It would be irrational to treat the two equally at their word.

      • by Samrobb ( 12731 ) on Saturday December 27, 2003 @04:03AM (#7816401) Journal
        Microsoft also has a long and distinguished history of FUD.

        Pretty amusing that you would say that, considering the origin of the term: [catb.org]

        Defined by Gene Amdahl after he left IBM to found his own company: "FUD is the fear, uncertainty, and doubt that
        IBM sales people instill in the minds of potential customers who might be considering [Amdahl] products."

        Not that I disagree with your assertions - IBM doesn't have near the same ties to Linux that MS has to Windows. But it's amusing to see how much the technological landscape has changed, that a term coined to describe IBM can now be used to (in some sense) defend it.

  • WHAT is the failure? (Score:5, Interesting)

    by SharpFang ( 651121 ) on Saturday December 27, 2003 @02:02AM (#7816164) Homepage Journal
    95% success ratio... does that mean that 1 in 20 programs I run segfaults or what? What do they mean by "failure"? Not finishing given task in predefined time? Getting the results wrong? Hanging?

    Sorry but that means nothing. Even if there -was- a comparison to other systems, it would still mean nothing. 95% success ratio, 78% happiness factor and 93% user satisfaction.
    • by be-fan ( 61476 ) on Saturday December 27, 2003 @02:31AM (#7816235)
      Its the results from the Linux Text Project suite. 95% success rate, zero critical failures, means that 95% of the 2000 test cases completed successfully, and nothing crashed the kernel. To see what that means, just take a look at what test cases are in the LTP!
    • I suspect the 5% failure rate refers to things like dropped packets. This isn't considered a critical error because of the way TCP/IP works -- the client will just resend the dropped packet. If there were incorrect results or processes hanging/segfaulting, that surely would have counted as a critical error.
    • oh, yeah---and statistics. Statistics are the most manipulatable facts. When given with no context, they might as well be lies.
  • Failures needed (Score:4, Interesting)

    by Error27 ( 100234 ) <error27.gmail@com> on Saturday December 27, 2003 @02:21AM (#7816209) Homepage Journal
    This test would have been more interesting if there had been failures. Perhaps they could have tried the test on an older version of Linux, or a different operating system.

    I have been trying to write some tests of my own recently. So far I have found a filesystem OOPs, a ptrace BUG(), and my system locks up on low memory situations. Probably the lockup is because my ethernet driver allocates memory in the interrupt handler (GFP_ATOMIC) and can't handle the result when there is no memory available.

    I need to fix the lock up first of all so the other tests have time to run...

  • Here goes (Score:5, Insightful)

    by floydman ( 179924 ) <floydman@gmail.com> on Saturday December 27, 2003 @03:18AM (#7816309)
    FTA
    The tests demonstrate that the Linux system is reliable and stable over long durations and can provide a robust, enterprise-level environment.

    Ok, now i dont mean to troll here, so mod down if you wish, i really dont care.... BUT...
    I am a linux user/programmer/lover for the past few years now, and i wanna see a company that is not SO IN LOVE with linux say what have just been said by IBM above.
    In other words, i dont want to see companies who sell Linux, or who have benefit in selling Linux praise it. Does any one of you know of someone who fills in these criteria. Sun for one is not very fond of Linux, nor is MS ofcorse (despite the fact sometimes i doubt they have code in their stuff from Linux...)...to make a long story short
    It would be really nice if such a judgment came from someone else besides IBM/REDHAT/ORACLE...

    • Re:Here goes (Score:3, Informative)

      by Fnkmaster ( 89084 ) *
      You mean an "unbiased" industry analyst? The problem is that everybody needs their bills paid by somebody. And these days pretty much everybody in the computer industry has some interests tied up with either Microsoft or Linux (seeing as how most of the old Unix players are becoming Linux players as well - IBM, SGI, (sometimes) HP, Sun...).

      It takes a lot of time and money to do very thorough analyses of operating systems, hardware and enterprise apps. So that money has to come from somewhere. It would

    • Espically for thigns like this. If a company isn't the one performing the research, chances are they are bankrolling the company that is. This means that generally thigns will be stacked in the favour, and unfavourable results will often be suppressed.

      Even true independants are often not unbiased. For example some individual, with no teis to and OS developers or vendor, might decide to test OSes. Of course it might be that they are a huge Linux or Mac or Windows zealot so again stack things in their favour
    • any one of you know of someone who fills in these criteria.

      The closest you're likely to get is good testimonials from companies using Linux. IBM, SUN etc. all have a stake in Linux, and the 'independant' research outfits are probably funded by them, or by Microsoft (in case Linux needs a good bashing).

      My client is a big megacorp. Their strategy for the coming years is to migrate all Unix systems to Windows/.Net (client side), and to Linux or NT (server side, depending on which OS fits best). This is

  • by tonyr60 ( 32153 ) * on Saturday December 27, 2003 @03:42AM (#7816355)
    "The Linux kernel properly scaled to use hardware resources (CPU, memory, disk) on SMP systems."

    Sorry, but how can the scaleability of the CPU resource be proven on a 2 CPU system? Show incremental results on 1, 2, 4, 8, 16, 32 etc. etc. and then CPU scaleability may be proven.

    This is NOT an anti-Linux troll, rather the evaluation needs to justify it's outcomes or it starts to look like something from a company starting with M.
  • by Anonymous Coward on Saturday December 27, 2003 @03:45AM (#7816360)
    1) Linux is free if your time is worthless
    2) My time is not worthless
    3) Linux is not free
    4) Giving Linux as a Christmas present is not "cheap"
    5) Linux is a good Christmas present.
  • So what ? (Score:4, Insightful)

    by Alain Williams ( 2972 ) <addw@phcomp.co.uk> on Saturday December 27, 2003 @06:55AM (#7816650) Homepage
    All that this shows is that a Linux based system works in the way that it should. Would you expect anything else if you ran your: TV, central heating, ... for a long period ?

    The trouble is that, after a period of increased stability in the 1980's, in the last decade people have come to expect that computers fail, and they wonder with amasement if they don't.

    OK: 30 years ago I remember it being a good day if the mainframe stayed up 12 hours. But things have moved on, today you expect your: MVS, VMS, Unix, Linux machine to stay working. The only OS vendor who's products have not matured is the one in Redmond - largely because of rampant infestation with new features.

    The above is not intended to belittle the fantastic efforts of all those involved.
  • by idiotnot ( 302133 ) <sean@757.org> on Saturday December 27, 2003 @08:18AM (#7816758) Homepage Journal
    In fact, it's not much of a question for me anymore -- when there's a problem, it's normally hardware malfunction. I have several machines with 160+ day uptimes, which would be longer if not for an extended power outage at the office.

    IBM just confirmed what I already knew. Guess what, Win2k is pretty stable, too. Sorry, but it's true.

    But, jeeze, isn't anyone else drooling over those systems they tested on? Makes me hate my busted whiteboxes and horrible HP's a little more everyday.

    Repeat after me....."MMMM, dual Power4......MMMM, dual Power4...."
  • My experience (Score:2, Informative)

    by MadChicken ( 36468 )
    OK, so I don't have a paper, but I remember my old Linux/P166 running great for a day or so when the CPU fan had died. I only noticed when I rebooted into Windows!

    My notebook has a flaky RAM connection. 32 MB comes and goes depending on how the machine is squeezed. Win 9x products crash it hard, Linux and Win2k don't even notice.

    So in my experience, Linux doesn't mind a hostile platform.
  • IMHO (Score:4, Informative)

    by kmichels ( 161476 ) <konrad AT michels DOT co DOT uk> on Saturday December 27, 2003 @09:21AM (#7816861)
    Nice to see some number coming in on Linux stability, although, as someone once said: there are lies, damn lies, and statistics. Anyone who has done any stats at University will know that you can prove anything from any result set. And as has already been pointed out by someone, the fact that it has been done by a company who has a not insubstantial vested interest in getting Linux into as many big companies as they can, it carries about as much credability as a Windoze security evaluation paper done by anyone other than Linus.


    The very reason Linux has already made so many inroads into coporations in the first place is because of its reliability and stability, and not because some marketing campaign has churned out the words on header paper.


    Another point is that I personally expect the sytems I administer to run for a darn side longer than 30, 60 or 90 days unless I need to restart them because of a kernel upgrade. When my last bunch I worked for went tits-up, our SAMBA file server had a 790 day uptime, and had run the SAMBA daemons reliably throughout, as well as doing internal DNS and DHCP. That's what your average Linux sysadmin expects from a Linux server box.


    A Linux desktop being used for all manner of things though is completely another story: if I muck around with the Linux install on my laptop, as I do because that's what I do, then I expect to break it from time to time, and so "reliability" is not measured in the same way on a desktop/laptop system, IMHO.


    The ideal environment for Linux is as a networked server, where it can get on with doing what it was setup to do, and will continue doing so until someone pulls the power plug on it. In that context, there are few OS's playing on the same field that can rival it for reliability and stability.

  • by Lord Kano ( 13027 ) on Saturday December 27, 2003 @09:41AM (#7816889) Homepage Journal
    but with package and dependency madness.

    I couldn't tell you the number of times I tried to install something and it fails because I was missing "X-Widget-2.41.so.1", so I try to install that "X-Widget-2.41" package and the "X-Widget-2.41-devel" package and they fail because they are missing several other depends as well.

    Linux stability is fine. The GNU software stability is fine. We need a better way to install and maintain software.

    LK
  • by crow ( 16139 ) on Saturday December 27, 2003 @10:28AM (#7816976) Homepage Journal
    I've been using an old P120 laptop as a firewall/router for my house for the past several years running 2.2.something. I wondered why it rebooted after noticing an uptime of only a day or two, but found that instead I was experiencing the uptime rollover bug (at about 500 days; Windows used to crash on a similar bug after 48 days). About a month ago, it stopped giving out DHCP addresses. I went downstairs to investigate, as I couldn't log in remotely, and found that the hard drive was making that nasty clicking sound. I eventually managed to ssh in (sshd and sh were in ram; I just waited for the logging to time out). I was able to kill syslog and cron, and now dhcp is again giving out addresses.

    It's been running just fine for a month now with a dead hard drive.

    (Yes, I'm getting a replacement because it won't survive an extended power outage on that ancient battery.)

If all else fails, lower your standards.

Working...