Kernel Changes Draw Concern 685
Saeed al-Sahaf writes "Is the Linux kernel becoming fat and unstable? Computer Associates seems to think so. Sam Greenblatt, a senior vice president at Computer Associates, said the kernel is 'getting fatter. We are not interested in the game drivers and music drivers that are being added to the kernel. We are interested in a more stable kernel.' There continues to be a huge debate over what technology to fold into the Linux kernel, and Andrew Morton, the current maintainer of the Linux 2.6 kernel, expands on these subjects in this article at eWeek."
Just my $0.02 (Score:5, Interesting)
"We are not interested in the game drivers and music drivers that are being added to the kernel. We are interested in a more stable kernel."
If you don't want it, don't compile it in. Thats the best part about having the kernel opened and so easy to manipulate. With the GUI available for modifying the kernel as well as a detailed set of instructions built right in, anyone can sit there and remove support for the latest gaming joystick if they so choose to. No one is making you keep it. If the kernel didn't have the option of supporting it, or if they discontinue the building of, then Linux will never be ready for the desktop. Just because Morton or Linus decides to add/accept support for the desktop community doesn't mean that the kernel won't be any more stable. Who is to say that adding gaming support took time away from stabilizing the kernel? If I'm strictly a game hardware designer and send my contribution to support the latest device does not mean that I could have spent my time improving the kernel. I may not be comfortable doing that. In other words, maybe I can't stabilize the kernel but I can write new drivers for it. And if I spend my time doing that it doesn't mean that I take time away from those improving and stabilizing the kernel.
The part that really caught me off guard is the inclusion of the Xen virtualization technology. [crn.com] Big changes are coming to the kernel that are really going to improve Linux and its functionality in the buisness and home world.
Re:Just my $0.02 (Score:5, Insightful)
Which is exactly what Andrew Morton said. I think that the underlying issue is a human resources one. CA wants Linus and Andrew to spend all of their time working on "Enterprise" features and none of it on things like improving Linux's real-time performance and integrating drivers for non-server hardware. I think that they're being selfish and unreasonable, but that seems to be par for the course for CA.
Re:Just my $0.02 (Score:5, Insightful)
Good thing $0.02 is pretty worthless. (Score:5, Informative)
Re:Just my $0.02 (Score:4, Funny)
Re:Just my $0.02 (Score:4, Funny)
Because that's not how Microsoft does it. And the business world will never accept linux until it's changed to mimic MS Windows' design. Haven't you been listening to what people have been saying here for the past N years? It's routine to point out a good design feature of linux and claim that that's why linux Isn't Ready For The Desktop, and won't be until that design is changed. This is mentioned more often than the impending death of *BSD.
(Lessee, do I need a
Re:Just my $0.02 (Score:5, Interesting)
What's rediculous is having to recompile C code to remove unwanted bloaty functionality. What's that do for QA? No two compiled kernels are the same, depending on what got commented out, compiler, settings, etc. Thats the concern with stability.
Why worry about whether or not newly added stuff is going to break. If the scope of each layer is limited properly the kernel can be fundamentally stable.
Here's more info on micro kernels and why they rock [linuxfinances.info]
Re:Just my $0.02 (Score:5, Funny)
Re:Just my $0.02 (Score:5, Funny)
Re:Just my $0.02 (Score:5, Insightful)
And this points to the real evolution in linux that has Microsoft sweating: what CA wants is a kernel that works better for businesses. Why? Because businesses have come to rely on linux.
Business (in general, I'm not talking about CA specifically but about all the businesses that now use linux in their operations or, even more, in their firmware) to linux: "Linus, we didn't pay you to write the kernel, we didn't give you much help in writing it, we've often appropriated it and ignored our legal responsibility under the GPL while at the same time keeping out own drivers closed-source and binary only. But, ah, now that we use -- for free -- what started out as your hobby project, we expect you to give up your hobbyist ways and toe the line, because it's now our bottom line."
This really isn't all that much different from the RIAA's "buggy-whip manufacturers'" outlook on file-sharing: "we've always made buggy-whips, and we loved it when Linus and the rest of the OS community were producing free leather for us to make buggy-whips, but now that you're producing those infernal auto-mobiles, we'll, you better stop before you threaten our profits."
The one thing I've never liked about the GPL was that it gave the same rights to a for-profit business as to a fellow hobbyist. I'm more than glad to share my code with a fellow, who like me, is coding for the love of it. I'm a bit less happy to share with someone who just sees my uncompensated work as a way for him to parasite off it.
Linus should tell CA that businesses have gotten far far more -- just in dollars, I'm not talking intangibles -- from Linus than they can ever repay, and that he's going to go on doing what makes Linus happy. After all, that attitude worked out pretty well for the parasites last time around.
As for the rest of us, maybe those of us who can and do code should ask ourselves why we're so happy to give our work away for free to businesses that do their level best, day in and day out, not to give away anything for free.
Is the GPL really our best answer?
Re:Just my $0.02 (Score:5, Insightful)
But how about sharing it with a multinational company that is able to throw massive resources into helping you to develop your program? If you shut out all companies you shut out the freeloaders, but you also shut out companies that would otherwise be helping your project. The Linux kernel isn't mostly the work of hobbyists, and it hasn't been for a long time. For many years Linus worked for Transmeta, who hired him in part because they wanted to use Linux with their chips, and now he works for OSDL, which is funded by big corporate Linux users. Alan Cox works for Red Hat. Marcello Tostatti works for Conectiva (now Mandriva or whatever they're calling it). The list goes on and on.
And then there are the direct corporate code contributions. SGI has contributed XFS and a lot of work on NUMA. IBM has contributed a boatload of code including JFS, NUMA, and RCU, and they've tried to contribut more things that were eventually passed up because others came up with better solutions. Namesys developed ReiserFS. Many vendors have contributed drivers for their hardware. The Linux kernel wouldn't be nearly what it is today if those companies hadn't been contributing.
The key thing to understand is that freeloaders don't actually cost anything, except for the bandwidth they use for downloads, but contributors help to build the software. It's smart to let anyone use the software because then anyone can be a contributor. Help from the IBMs and Red Hats of the corporate world more than pays for all the freeloaders.
Re:Just my $0.02 (Score:4, Interesting)
Apparently not for you. The neat part about licenses, is that they're so damn easy to cook up. For example: Simple, no? You could say that: Hard, wasn't it?
See, as the licensor, you can put pretty much any term you want. There are *SOME* limitations, but they aren't what you might think.
Ever READ the GPL? It's written in plain English, not Lawyerspeak. (Oh, and IANAL, all that jazz) When dealing with anything legal, lawyerspeak is to English was code is to specifications. It's intentionally a little halting because it's precise.
If you figure your licensed product is worth millions, get an attorney. Otherwise, specify the terms you like, and enjoy!
Re:Just my $0.02 (Score:4, Informative)
Re:Just my $0.02 (Score:4, Insightful)
Where did you get the idea that hobbyists have customers? Corporations have customers and if they want changes in the kernel or another open source application then they can code it themselves. Hobbyists don't care if Linux "takes off" because they make no money off of it and don't care to. For most of us hobbyists Linux is good enough as it is and if we want something more then we'll code it for OURSELVES. It's nice when big corporations contribute code but we don't owe them a damn thing, they are using our free code after all.
Re:Just my $0.02 (Score:4, Insightful)
And major F/OSS projects like linux aren't artificially hampered by the commercial OS vendors that want to sell a "desktop" version and a "server" version, or worse yet charge per client licenses (WTF!) Linux is imminently tweekable, runs on everything from embedded ARM7 to supercomputer cluster IA64. Stable linux distributions like Slackware offer far more compatability from desktop to server than RedHat's offerings (okay, FC4 is a "committee" project, not unlike the proverbial horse that became a camel).
Perhaps CA just needs to hire some F/OSS consultants -- they could get on the cluetrain just by lurking on the forums like slashdot. So to CA, I say "Quit you're mewing!".
That makes no sense (Score:5, Insightful)
Re:Just my 5 bytes (Score:5, Informative)
There is no reason that these "experts" can't tune a 2.6 series kernel to around 1 MB (maybe less). Kernels with modest support for lots of hardware are still around only 1.5 MB at best. Anyone complaining about it is simply talking out of their asses.
You don't want "game drivers and music drivers", then exclude them. There is no science to it. But I *want them* in my kernel, and many other people do as well.
Additionally, if Greenblatt and co. want more "enterprise features", they're certainly welcome to add time and money into developing these components.
This e-week article is misleading. It's not drawing "concern" for anybody, especially not the "open-source community". Computer Associates is not the "open-source community".
Take a look at the linux-tiny patchset (Score:5, Interesting)
The linux-tiny [selenic.com] patchset is your friend here. Using it, I've gotten a relatively full-featured kernel booting on x86 weighing in at under 800K... and that's without doing any agressive trimming, and without module support. According to his OLS 2004 presentation [selenic.com], Mackall has achieved a linux 2.6 kernel weighing in at a mere 363K, and others have reportedly managed a kernel as small as 191K.
Some of the linux-tiny ideas have been making their way into the mainline kernel, so this isn't just a special-purpose patchset - it's really a proving ground for kernel size minimization techniques.
Re:Take a look at the linux-tiny patchset (Score:4, Funny)
Re:Just my 5 bytes (Score:4, Insightful)
Yes, and the US Pledge of Allegiance ends with the words "with liberty and justice for all". Just because you say something doesn't mean it's true.
Re:Just my $0.02 (Score:5, Insightful)
That line of thinking can be dangerous though (Score:5, Interesting)
This is espically true for companies who's core bussiness isn't IT or engineering or the like. If a company just uses computers as a means to an end and they don't really have a tech staff, it can be expensive, difficult and risky to contract someone to do the development they need. Better to just get a commercial solution.
I'm not saying this means OSS devs need to jump up and meet every request from every person that whines, that's clearly impossible. However I find that the OSS community in general is way to fast to say "It's open, if you want the feature, write it yourself!" Rather the merit of the request should be weighed, it may be worth your while to work on. If it's not, then you should give reasoning as to why not, and not just say "Do it yourself."
Re:That line of thinking can be dangerous though (Score:5, Insightful)
As for other situations. If you are going to get a certain level of support for a product (new features, custom installations), that is going to cost you a certain number of dollars, whether it be licensing costs (you need to be a large enough customer to have that level of influence with a vendor), or it be in hiring developer time to work on an OS project.
I would love to see some sort of feature wishlist where smaller companies could vote with their dollars on certain bugs or features. I've heard of bounty systems like this being tried, and I would love to hear more about why they haven't really worked yet.
You are right about the OS community being quick to jump on the "code it yourself" excuse. But that is reality of dealing with volunteers. Some are motivated by competing with commercial products, and will work on features to make that happen. Others are totally unconcerned with what corporations think about their work. At the end of the day, many developers are scratching their own itch and shouldn't be expected to care about what other people want their software to do.
At the same time as some people are quick to jump on this excuse, others are quick to assume that the goal of OS should be to beat proprietary software. This is simply not many peoples goal.
Re:That line of thinking can be dangerous though (Score:5, Interesting)
I cant believe any such a debate emerges over quotes from one of the worst managed, maligned companies in enterprise software. Slashdot is doing them a favor by advertising this dude's comments...
Re:Just my $0.02 (Score:4, Insightful)
If you don't want it, don't compile it in.
It gets better. If someone says "but I use a stock kernel," remind them that they don't have to load every module under the sun.
This guy would be better off going off to tell hardware manufacturers to quit making new hardware. Yeah right! Also, why does he not complain about bloat in the Windows kernel? IIRC, there is a much larger segment of hardware supported in Windows than in Linux. Mehtinks his statement should be modded -1 Flamebait.
Re:Just my $0.02 (Score:5, Interesting)
grr let me rephrase : an option to download only stuff i need, eg i could only get the sources that i actually selec in the kernel config gconfig,menuconfig,config, whatever feels good
on the other hand if you cannot download 40 megs buy a distribution on cd/dvd or use windoze
I am happy with the kernel, and however is monkey enough to compile everything IN and than complain about it being big well uhm
i love the kernel supporting more and more of the junk i can stuff into my machine to enhance my gamin...video... I mean work and productivity
Re:Just my $0.02 (Score:5, Informative)
Or just download the patch instead. That's what those patches are there for, you know
Re:Just my $0.02 (Score:5, Funny)
Yeah, and we know that Linux Will Never Be Ready For The Desktop until firefox and thunderbird are integrated into the kernel.
Re:Just my $0.02 (Score:5, Funny)
Thanks for this exposition of conventional wisdom.
When you have something specific to pin this on, I'm sure we'd all like to hear from you again.
Re:Just my $0.02 (Score:4, Insightful)
Re:Just my $0.02 (Score:4, Insightful)
So now instead of paying Microsoft to make your choices for you, you pay Red Hat or Novell to do it. You can even hire a consultant that will tailor the kernel to your specific needs if it's that big of an issue, and if it is, I doubt that Windows would suffice anyway.
Choice alone is a good thing, and when your choices are open it's even better. Find someone to do what you want well for as cheap as you can, or take one of the prepackaged solutions. It's not that big of a deal.
Re:Just my $0.02 (Score:5, Insightful)
Re:Just my $0.02 (Score:4, Insightful)
1. Modprobe/insmod/rmmod.
2. The OpenVMS kernel is written in VAX assembler (http://research.compaq.com/wrl/DECarchives/DTJ/DT J807/DTJ807SC.TXT [compaq.com]). It was not written in "languages like" Ada. Jesus christ.
No problem (Score:5, Funny)
Isn't that why, (Score:3, Interesting)
FP?
Re:Isn't that why, (Score:4, Informative)
"fatter" (Score:3, Interesting)
And about "fatter": I don't get it. You will probably use ONE sound driver, ONE (or perhaps two) network drivers, etc. Just the fact that the *amount* of drivers is gettling bigger, does not mean the kernel "is getting fatter".
Re:"fatter" (Score:5, Insightful)
The kernel is fine it's the setup that sucks.
Re:"fatter" (Score:4, Insightful)
Re:"fatter" (Score:4, Insightful)
So that loading the kernel on 100s of machines is as easy as distributing a single file rather than a distribution of files.
Personally? I never used modules when I could just compile it all in. It's easier to transport that way.
Re:"fatter" (Score:5, Interesting)
I already have a third part driver, from linux-wlan-ng, and every time I upgrade the kernel I have to remember to recompile it again.
The kernel should have everything. Obvious, for licensing reasons, only GPL stuff, but everything that's GPL, and is a kernel driver, and is up to minimal code standards.
In fact, having to track down third party drivers has been a much more valid complaint than 'Too many drivers', which is just idiotic.
Re:"fatter" (Score:3, Interesting)
How many people actually use I2O, HAM and all that exotic hardware Linux can support? Spinning off all the exotic sections into separate downloads would seriously reduce the average download size. For fairness to server people, I suppose even the sound system could be dumped into a separate archive.
If "make *config" conveniently
Re:"fatter" (Score:3, Insightful)
If you're going to run a typical "server" for a business then a 20-50mb download isn't that much. combine it with it's source so you can build a different kernel for each server (if needed).
Yes there are large sections if the kernel I've never touched (and I doubt I ever would), but I for one still want to see it in th
Two Sides (Score:3, Insightful)
I think they have this nifty thing called CONFIG (Score:5, Funny)
Re:I think they have this nifty thing called CONFI (Score:5, Funny)
Hypocritical (Score:5, Insightful)
Re:Hypocritical (Score:5, Funny)
Hey, if they're experts on it....
Compiled Kernel not necessarily getting fatter. (Score:3, Informative)
Drivers that are not compiled are not taking any additional space. Drivers that are not used all the time can be compiled as modules...
So I guess I do not understand the criticism here.
Re:Compiled Kernel not necessarily getting fatter. (Score:4, Interesting)
Re:Compiled Kernel not necessarily getting fatter. (Score:3, Informative)
Does it actually load them or does it just print a message which indicates it's going to do so, check for hardware, and exit when it can't find it? If it does actually load the modules, won't they be unloaded after a while if they aren't used at all?
For example, if I try to load a pcmcia module on a destop machine from the command line it indicates it cannot find the hardware, and exits. I suspect the only diff
Re:Compiled Kernel not necessarily getting fatter. (Score:5, Insightful)
Have you ever installed a late version of Windows?
Watch the installer load device drivers for every known weird form of RAID before it even begins to ask you how you want to install the OS?
And then how long does it take to do "hardware detection" - versus Knoppix that does it all in the three minutes or so it takes to boot from CD?
Yes, Windows is bloated - bloated with (so-called) "features", not drivers. If Linux makes THAT mistake, we can complain. Having a bunch of drivers and support for oddball subsystems loaded into the kernel is not serious and until somebody DEMONSTRATES a stability problem, it's bullshit.
So far I've heard nobody say the 2.6 kernel is in FACT unstable because of x, y, z drivers or subsystems.
Re:Compiled Kernel not necessarily getting fatter. (Score:4, Interesting)
I'll say it: the 2.6 kernel is unstable on x86_64 platforms with USB 2.0 mass storeage devices. There are bug reports everywhere. The response? "It's fixed." The reality? The system locks up like Fort Knox whenever it's booted with a USB 2.0 mass storeage device attached.
Re:Compiled Kernel not necessarily getting fatter. (Score:3, Informative)
There's also a GUI tool for this. For that matter, you could not select those services to start in the first place. There's a dialog for it in the installer.
WTF? (Score:3, Insightful)
Huh????
Re:WTF? (Score:3, Interesting)
Can someone explain it to me or is this just a badly written article that is referring to the license of other virtualization technologies.
Re:WTF? (Score:5, Informative)
I'd say the parent is a fair question, not a troll.
Morton's point appears to be this:
* x86 is notoriously unco-operative to full virtualisation
* trying to fully virtualise it (as VMWare and Virtual PC do) is a work around for the fact you can't modify the guest OS because it's closed source
* fully virtualising x86 in software results in rather painful performance hits for many workloads and a very complex hypervisor
* for open source OSs, it therefore makes sense to use paravirtualisation. This involves porting the OS to a special virtual machine-oriented "architecture", closely resembling the real hardware but without the costly-to-virtualise parts.
* paravirtualisation can be argued to be better than full virtualisation because (esp. on x86) the performance hit is much lower.
Porting of open source OSs is happening: Linux 2.4 and 2.6, NetBSD, FreeBSD 5.3 and Plan 9 can run on Xen (although currently only the Linuxes are supported as "host" or "Dom0" operating systems).
Re:WTF? (Score:4, Informative)
Something I think Sam missed is that Xen also supports VT which provides full-virtualization on the x86 (which makes Xen undeniably a true-hypervisor).
Compiler-driven para-virtualization is an interesting emerging area of research too that should make porting OSes to Xen much simplier.
All we need now is a really cool hypervisor-aware file system.. like a XenFS
Re:WTF? (Score:5, Interesting)
You need a license for the host OS, and a license for the hosted OS. It's also having to provide fake hardware.
With Xen, maybe it's not that extreme. With the same OS inside and out, and it knowing about itself, it might not be running two copies at all, acting like a really extreme version of chroot instead. Hence the licencing being better.
And it would seem to be a lot saner. I mean, think about disk files. With VMWare, VMWare takes the file, fakes a device from it, and Linux accesses the device, but that's rather goofy when you think about it, because Linux can already mount files. With kernel support, the host kernel could let the hosted OS have direct disk access to that file, and only that file.
In the Linux kernel, there are a lot of 'loopback' and 'fake devices' concepts like that. There's the loopback mounter, there's SCSI emulation, there's fake network devices, there's the fake PS/2 mouse in /dev/input/mice, there's all sorts of pretend hardware. With Linux-on-Linux support in the kernel, that fake hardware could trivially turn into 'real' hardware for a hosted machine, where the hosted kernel know it's accessing something fake, and the host kernel just needs to restrict access.
Hopefully this will be extendable enough that the 'devices' the hosted kernel use can be shared with Linux-on-other-platforms, like coLinux on Windows. And the devices exposed to the hosted machine could be exposed to other emulators.
Re:WTF? (Score:3, Informative)
Not what we want. (Score:5, Funny)
also we want a sandwich.
That is all.
Re:Not what we want. (Score:5, Funny)
What about older hardware! (Score:5, Insightful)
The problem, I think, is that developers tend to be people who love computers. And people who love computers tend to have nice rigs, just as people who enjoy cars tend to spend a disproportionatly large amount of their income on cars (ever see the parking lot at a lan party--complete with people pulling multi-thousand dollar machines out of the hatch of a Hyundai?).
Perhaps Linux needs more developers from third world nations; the kid from a rural village with intermitant electricity getting his hands on an old, but useful machine and learning that he, too, can tell it to do all sorts of things!
Re:What about older hardware! (Score:5, Informative)
This proud owner of an AMD K6 300 MHz has compiled and runs Linux 2.6.11.7 without a hitch, and continues to not see the problem.
Re:What about older hardware! (Score:5, Insightful)
My point with this is that it's not the kernel that's making GNU/Linux systems crawl on older hardware. It's the newer versions of GNOME and KDE. As long as you aren't running GNOME or KDE, older hardware works just fine. My servers chug along just fine, and my 233 MHz laptop with 64 MBs of RAM running Sawfish also suffices just fine to do virtually all my common tasks (except running any Mozilla product :-P ).
So, certainly, GNU/Linux may need more developers from third world nations, as you put it. Linux, however, does not.
Re:What about older hardware! (Score:5, Insightful)
It's ridiculous to suggest that the kernel layout should be restricted to the level of a 486.
First of all, you can already do that if you know what you're doing. People in the Third World either know what they're doing or get their machines from people who do - just like in the rest of the world.
Secondly, there are tons of stripped down distros. Pick one.
This is merely asking for your cake and eating it, too - you want the latest kernel and everything it can support to run on the oldest hardware.
Try it with Windows 20003 Server.
Then go back and read the specs for Longhorn: a GB of RAM, a terabyte of hard disk, and a minimum 3GHz CPU.
The Linux kernel is intended to push the boundaries of OS technology - not run on every Third World machine in existence.
Yet, at that, as I pointed out, Linux is incredibly flexible in what it will run on compared to virtually every other OS in existence.
All of this is just utterly pointless criticism.
Re:What about older hardware! (Score:4, Funny)
You're shitting me. Who the hell is going to use it with those kind of requirements?
Re:What about older hardware! (Score:5, Informative)
http://technovia.typepad.com/technovia/2004/05/lon ghorn_specs_.html
[typepad.com]
In a nutshell, it comes from a slide at a developer's conference, indicating the kind of machines that may be around for Longhorn's lifetime, and that the OS should be able to take advantage of such high specs, not that it will require such high specs.
this is nothing new (Score:5, Interesting)
What would be cool is if the linux distros had default kernel options, much the way some of the majors have Workstation, Server, etc... that would adjust the kernel based on how the machine was being used.
Yes, I know one can reconfigure the kernel by one's self, but it then requires personal care and feeding for patches, upgrades, etc... It becomes one more thing one has to do. Personally, unless I really need it, I'm not goign to bother... too much of a PITA
Re:this is nothing new (Score:3, Informative)
Slackware has this (or something rather like this) -- it comes with a whole set of kernels compiled for different kinds of hardware.
BS (Score:5, Funny)
can't please everyone all of the time (Score:3, Insightful)
I myself would like better multimedia drivers, good solid and easy to install and configure drivers for my PVR-250 and pcHDTV tuner cards in my MythTV box. CA may not give a darn about those at all, but this is my primary Linux goal and getting my particular MythTV rig running is the only application I myself presently give a darn about in all of Linux land.
I myself do not give a darn about gaming support either right now. That may change in the future if I decide to expand on MythTV and turn the thing into a high-end game console as well. But for the moment I'm not interested, just as many gamers may not be particularly interested in TV tuner drivers.
Though keeping stability and efficiency as primary goalsagreeably is a good idea. But I think high-quality (ie. NOT alpha or beta) drivers for more hardware should also be important.
Probably true (Score:3, Interesting)
On the other hand I was screwed so badly by CA that my automatic reation to anything they say or do is to discount it as coming from that Den of Thieves and Liars.
What is it with CA? (Score:3, Insightful)
Re:What is it with CA? (Score:5, Insightful)
Re:What is it with CA? (Score:5, Insightful)
Re:What is it with CA? (Score:3, Insightful)
This sucks because of something explained clearly = criticism.
This sucks because of something I cannot understand enough to clearly articulate or really know whether it sucks = bashing.
"Getting fatter" is an analogy in the first place, and since it is talking about the size of the download and not the executable, not paticularly relevent. It isn't clear either whether "stable" is used in the context of "more code just keeps coming
We must listen to CA ! (Score:5, Insightful)
CA have contributed so much to the Linux kernel, so they know what they're talking about. NOT.
What is CA's motive in saying this ? They have no real experience in developing operating systems, nor are they producing data and a testing methodology to backup their opinion.
It seems to me they might be talking through their hat. [cambridge.org]
Inevitable event (Score:3, Interesting)
Re:Inevitable event (Score:3, Insightful)
Re:Inevitable event (Score:5, Funny)
Actually, that's not the case at all according to this new NY Times Article [nytimes.com]
...the Purdue researchers say the real explosive secret lies in the hull, or pericarp ... In some varieties, the pericarp becomes more moistureproof as it is heated, sealing in the steam until the pressure gets so high that the hull fractures and the kernel goes pop.
In other varieties that don't undergo heat-induced change, the moisture escapes, the hull never breaks and then the kernel goes pfffft.
BAAA (Score:4, Insightful)
If anything the extra junk benefits them because the folks developing those drivers are likely to find bugs in the kernel proper.
They've Been Complaining about That Since 1.3 (Score:5, Informative)
Natural evolution of an OS (Score:5, Insightful)
The trick, for Linux, will be to do what Apple did in moving to OS X -- create a new, "from-scratch" (yes, I know Apple borrowed a lot from others), OS with some form of compatibility-creating layer or old-kernal box. Incrementalism only takes an OS so far before revolution is needed to build a new, better system from the ground up.
I'm torn (Score:5, Interesting)
I really miss the days when I could run on a P166 with 32 MB of RAM, and KDE ran not too badly (as long as you don't try to open Netscape or StarOffice). I don't think this kind of performance is attainable at all anymore.
But on the other hand, I'd be loth to run a kernel that didn't at least support USB! I love having ALSA instead of the old mishmash of sound drivers. Ext3 was a relief. I must say that for me at least ip[tables|filter|chains] was confusing, but I trust that the best choices were made... Going back to a kernel that didn't have those features would be simply unnaceptable.
Has the kernel reached a level of complexity where the ol'time stability isn't likely to happen anymore? We just need to react with patches, just like the other OSs out there?
Re:I'm torn (Score:5, Insightful)
It's my belief that the kernel won't really stabilize until they branch off to 2.7. They're too focused on adding new features for the code to ever really shake out and get stable. They're shoveling new stuff in there way way way faster than it can really be debugged.
And they just wave their hands in the air and say that it's up to the distros to make this mess usable.
Until they get over this phase, in which they're pushing the hard work of debugging onto everyone else in the world, the kernel is not going to stabilize. And we will be held hostage by particular vendor kernels, instead of being able to track the 'one true Linux'. If we start with Redhat, we're stuck with Redhat. In the past, we were able to fall back on the One True Kernel if Redhat or Mandrake made a mistake. But that's not really an option anymore... tracking the One True Linux is now dangerous, because the kernel devs don't really care if it works right.
I can't find the precise quote right now, because I can't see my old comments on Slashdot... apparently I now have to pay for the privilege of seeing my OWN old comments
Until that mindset changes, Linux is just not trustworthy. It needs to be made right BY THE PEOPLE WHO WRITE IT. You can't hack reliability in as an afterthought, it has to be a major focus all the way along. This is exactly the sort of crap we always derided Microsoft for... ship it buggy and then fix it later. I hated this behavior in Microsoft. I hate it just as much in Linux. I switched to Linux because it was, first and foremost, reliable. It no longer offers me that, and I am starting to switch machines over to the BSDs now.
Waving one's hand and expecting 'the distributions' to do the grunt work of actually making the kernel stable is just wishful thinking... it's expecting other people to do the job that should be the very first one on their list. Reliability is THE MOST IMPORTANT FEATURE. It's not fun, it's not glamorous, but it's what got Linux so popular that these guys actually get paid to do it. If it doesn't return to relatively bulletproof status, then people are going to use other solutions instead, and there won't be as many Linux jobs available.
It's the reliability that creates the jobs. I wonder if they really grok this?
Re:I'm torn (Score:4, Insightful)
I agreed with you up to this... This is just FUD.
Many commercial vendors are famous for leaving serious open bugs, and not fixing them for a LONG time.
Now, it's true that OSS/FS developers aren't compelled to fix the problems you are having, but that doesn't mean you're screwed. If you are having a problem, you can fix it yourself, you aren't stuck if the company decides they aren't interested in fixing it. With plenty of developers using it, small bugs like yours get unoffical patches pretty quickly.
As I said, I agreed with you up to that point. Linux does seem to be very poor at stability testing before releasing. I would suggest switching to on of the BSDs if you want a rock-solid system... I know comments like this get marked as trolls here on
I see a lot of clueless replies (Score:5, Interesting)
Problem is, that doesn't affect the main problem, which is that 3 million lines of options code is a LOT harder to keep bug free among all the different combinations than 1 million loc.
All bugs may be shallow given enough eyeballs, but the difficulty of debugging the linux codebase may well be increasing faster than the number of eyeballs.
Re:I see a lot of clueless replies (Score:5, Insightful)
But isn't most of that code base specific drivers for specific hardware, maintained by individuals who wrote that code? Are you saying that instead of including possibly buggy drivers, it would be better to leave them out and give no support at all to people who happen to have that hardware??
Remember, any potential bugs in drivers won't affect anyone who doesn't have that hardware - these drivers are compiled in default kernel distributions as modules and never get loaded unless they're needed. All it means is that the kernel modules take up a bit of disk space, which is trivial compared to the sizes of current hard disks. They don't impede performance and they don't do any other harm. I really can't work out what all the fuss is about
Re:I see a lot of clueless replies (Score:4, Insightful)
code in the tree, even if it's perfectly disconnected from the rest, still has to be modified when an API changes. With the 2.6 the de facto development codebase, that's not something to ignore.
Thanks, CA (Score:4, Informative)
How about trying out this GREAT utility called "menuconfig"...then you can unbloat your kernel. In the time it saves you from manually editing your
Re:Thanks, CA (Score:5, Insightful)
Menuconfig is just the window to the maze that is the kernel ifdefs. You have no idea of the size or speed impacts of the options you through if the help doesn't tell you. You have no idea of the component interactions.
Menuconfig is just a parking place for problems. The real problem is too many options, and not enough testing of the combinations. That is what CA is complaining about.
Re:Thanks, CA (Score:4, Insightful)
Is it? Because in the article Greenblatt snivels about "too many game drivers!@" and then breaks down completely and starts complaining that Xen "doesn't do enough." I'm not sure which side of the fence he's on. I do know that if I don't have an ATI Radeon in my system I'm not going to be totally baffled by the vast array of ATI driver options. But I don't work for CA.
Has Sam Greenblatt EVER compiled a linux kernel? (Score:3, Insightful)
Yeah, I personally find increased driver support a real problem
If he wants an OS for which you can't optimise the kernel in anyway try microsoft.com. I hear there are a couple there.
CA's kernel demands (Score:5, Informative)
No offense, but he sounds pretty clueless here - not to mention the fact that there is no "game driver" or "music driver", perhaps he is referring to device drivers and/or low-latency features, which allow for a better gaming/multimedia experience...
In any case, he completely misses the point that the kernel, as shipped by the distros, is modular. That means, if a device isn't present, or isn't used, the driver for that device never gets loaded into memory. So it doesn't really matter how many devices are supported, the only device drivers affecting the size of the kernel are the ones loaded into memory on the machine in question.
I find Greenblat's attitude ridiculous, since he seems to be saying that the kernel developers need to focus on what Sanm Greenblat is interested in, and to hell with people who want to do cool and interesting things with linux, which aren't part of CA's business plan.
I could go on, but that's enough for a first impression.
The Big Bloat (Score:5, Insightful)
linux-2.6.11 is forty four megabytes. Gzipped up. I don't want to waste my bandwidth downloading it to see what it is unzipped, but trust me, it's massive. Where does all this bloat come from? Drivers. Drivers are good, but the current kernel paradigm (and Linux isn't alone in this) is that every driver has to be included with the kernel. So we end up with huge packages and huger repositories where everything is required to reside.
Imagine the size of Linux when we finally get to the goal of having every past and current device with a dedicated driver in the source tree. You're talking possibly ten gigabytes uncompressed. Even if you're not using 99.9% of those drivers, they're still there. The day may come when you can actually build the kernel faster than you can make its dependencies.
Could you imagine a KDE or GNOME where every core, addon, auxiliary and experimental component was all part of one single tarball? Even if you only wanted GTK+ and GIMP, you still have to download and configure the entirety of the GNOME repository to get it. That's what it's like with the Linux kernel.
It's time non-core drivers got split off from the main Linux project. If you don't need to add anything into the kernel to get driver to work, then put it in the driver subproject and don't bug the big guys with this penny ante crap.
Re:Heading Down the Windows Path (Score:3, Informative)
you have options and one of those options is to not use stuff.
Re:Straight from a horses mouth. (Score:5, Interesting)
Re:Microkernels... (Score:3, Insightful)
of separate interfaces for every kind of object a single regular interface could be used at least as a starting point
There is; the C function interface. Abstract as much as useful, but n
Re:Microkernels... (Score:3, Interesting)
Ah, but they only require marshalling when you cross a protection boundary. Without that, they can be as efficient as the Amiga message primitive which was four instructions long.
The C function interface does not suffice. That's the same problem that RPC mechanisms have. The C function interface is synchronous, messages can be asynchronous, they can be buffered and queued, they make concurren