Slashdot is powered by your submissions, so send in your scoop

 



Forgot your password?
typodupeerror
×
Operating Systems Software Linux

Linux Kernel 2.6.30 Released 341

diegocgteleline.es writes "Linux kernel 2.6.30 has been released. The list of new features includes NILFS2 (a new, log-structured filesystem), a filesystem for object-based storage devices called exofs, local caching for NFS, the RDS protocol (which delivers high-performance reliable connections between the servers of a cluster), a new distributed networking filesystem (POHMELFS), automatic flushing of files on renames/truncates in ext3, ext4 and btrfs, preliminary support for the 802.11w drafts, support for the Microblaze architecture, the Tomoyo security MAC, DRM support for the Radeon R6xx/R7xx graphic cards, asynchronous scanning of devices and partitions for faster bootup, the preadv/pwritev syscalls, several new drivers and many other small improvements."
This discussion has been archived. No new comments can be posted.

Linux Kernel 2.6.30 Released

Comments Filter:
  • DRM? (Score:4, Informative)

    by corsec67 ( 627446 ) on Wednesday June 10, 2009 @09:44AM (#28278743) Homepage Journal

    Why would DRM be listed as a "feature"?

    Oh, wrong kind of DRM?

    • Re:DRM? (Score:5, Informative)

      by sakdoctor ( 1087155 ) on Wednesday June 10, 2009 @09:55AM (#28278893) Homepage

      The Direct Rendering Manager (DRM) is a component of the Direct Rendering Infrastructure, a system to provide efficient video acceleration (especially 3D rendering) on Unix-like operating systems, e.g. Linux, FreeBSD, NetBSD, and OpenBSD.
      It consists of two in-kernel drivers (realized as kernel modules on Linux), a generic drm driver, and another which has specific support for the video hardware. This pair of drivers allows a userspace client direct access to the video hardware.

      I assume it's this. Either that, or linux now has Direct response marketing in the kernel.

    • Re:DRM? (Score:4, Informative)

      by Stoian Ivanov ( 818158 ) on Wednesday June 10, 2009 @09:57AM (#28278915)
      Direct Rendering Managment - this DRM not the bad one
    • Nonono, the right kind of DRM!

  • by Anonymous Coward on Wednesday June 10, 2009 @09:49AM (#28278815)

    Meanwhile back at GNOME H.Q. the developers are still undecided whether to move the "Ok" button on the default help screen 10 pixels to the right. Most think it would be a good idea but a hard core few insist that such a momentous change requires further study as it may confuse new users.

    A new version of the dialogue is expected in 2037.

    • Re: (Score:3, Insightful)

      by moosesocks ( 264553 )

      Say what you want about the glacial speed with which GNOME progresses. Their developers don't rip out 2/3 of the features of their applications, and call it a " major upgrade."

      There's also a key difference between 'minimalism' and 'feature-deprived'. Apple understand this, and the GNOME team seem to be catching on. XFce's flexibility also makes it a surprisingly good environment to work in, despite being billed as a 'bare bones' environment. KDE almost certainly doesn't understand this distinction, and

      • Re: (Score:3, Insightful)

        by kigrwik ( 462930 )

        Say what you want about the glacial speed with which GNOME progresses. Their developers don't rip out 2/3 of the features of their applications, and call it a " major upgrade."

        You obviously don't remember gnome 2.0

  • POHMEL (Score:4, Funny)

    by dimethylxanthine ( 946092 ) <mr,fruit&gmail,com> on Wednesday June 10, 2009 @09:56AM (#28278901) Homepage
    Not sure about the story behind naming POHMELFS what it is, but "pohmel'e" in Russian means "hangover". You can only guess...
    • Re: (Score:3, Informative)

      by Anonymous Coward

      And Evgeniy Polyakov (the POHMELFS dev) sounds like a russian name. I guess he knows.

      in soviet russia file systems name hangovers after you

  • DRM for Trolls (Score:4, Informative)

    by chill ( 34294 ) on Wednesday June 10, 2009 @09:57AM (#28278919) Journal

    The Direct Rendering Manager (DRM) is a component of the Direct Rendering Infrastructure, a system to provide efficient video acceleration (especially 3D rendering) on Unix-like operating systems, e.g. Linux, FreeBSD, NetBSD, and OpenBSD.

    It consists of two in-kernel drivers (realized as kernel modules on Linux), a generic drm driver, and another which has specific support for the video hardware. This pair of drivers allows a userspace client direct access to the video hardware.

    From WikiPedia.

    Karma Whoring FTW!

  • by zevans ( 101778 ) <zacktesting.googlemail@com> on Wednesday June 10, 2009 @09:57AM (#28278935)

    If you're using 2.7.x Intel xorg drivers you NEED this kernel. Anyone struggling with weird freezes, font corruption, and various other troubles - turns out most of these problems weren't in the Intel drivers at all, but in the GEM and DRI code in the kernel. Mine's been rock solid since RC5 for stability, and RC8 finally fixed the problem with fonts under UXA.

    • I could never get XvMC working on anything, but using kernel modesetting/DRI2 crashed X every time I tried to play a video in Mplayer. I hope this will fix both issues.

      • Well, I noticed that it has to fit the driver. And some installations change the association. So you then end up with an XvMC of nVidia, with the main driver from Xorg, or something like that. It gets even worse, when you did not reinstall the external driver after a kernel update, so that the module can't get loaded anyway.

        In gentoo you would do
        emerge -atv nvidia-drivers
        and
        eselect xvmc set nvidia
        after a kernel update, when using the nvidia binary blob drivers.

    • Performance? (Score:3, Informative)

      Intel's integrated graphics performance has been pretty progressively worse ever since switching from XAA, and rather abysmal ever since Xorg 1.5. Since then every release of X/mesa/xf86-video-intel made it even worse. Hopefully this release brings the entire GEM/UXA/KMS/whatever stack to a usable state. All this on a 945GM.

      What's your experience with it so far? I'll try it out myself in a few days, but I'm eager to hear the results...

  • Thottle Capability (Score:5, Interesting)

    by kenp2002 ( 545495 ) on Wednesday June 10, 2009 @10:01AM (#28278997) Homepage Journal

    Still no support for SLA\95% throttling of processing power allocated to VMs.

    Case in Point:

    VM 1 : 80% Of processor utilization
    VM 2 : 20% of processor utilization
              : Can borrow up to 20% of VM1's allocation
              : if unused.

    The scheduler does great things don't get me wrong but when it comes to provisioning systems for various clients some want a garuntee on the level of processing power that is available at any time. This is true in test systems as well where yout Integration, Acceptance, and Performance virtual environments may share Bare Iron with some production VMs.

    Now this is old hat easy with mainframes (MIP allocation\weights between LPARS\SYSPLEX) but with more and more focus on VMs and hosted VMs SLAs on processing power is becoming more of an issue.

    Nice values are not enough when writing contracts... Great work Linux team but could we get some more granular control over VM provisioning with SLAs in mind? Yeah we can build user space systems to help manage VMs but kernel level provisioning and auditing is something we need with KVM. Gotta have the reports to show the customer you are meeting the agreeded upon SLAs.

    And for my own personal use, I'd love to be able to throttle a dos 6.22 VM to 486 speeds so some of those ancient programs can be ran for historical purposes. (Without bombing the processor with dummy NOP and other MOSLO crap so we keep our power consumption down.)

    Just some musings as Linux rolls along...

    • Nice, But... (Score:4, Insightful)

      by maz2331 ( 1104901 ) on Wednesday June 10, 2009 @10:06AM (#28279053)

      If you want a mainframe, maybe calling IBM and ordering one is a better way to go?

  • by Psiren ( 6145 ) on Wednesday June 10, 2009 @10:04AM (#28279015)

    Can anyone explain to me why Linux has so many filesystems? Windows has had NTFS for years (admittedly, several versions, but never any compatibility issues that I've come across), and Linux has, what, 73 or something?! Is it really that hard to get it right?

    • by fbjon ( 692006 ) on Wednesday June 10, 2009 @10:07AM (#28279069) Homepage Journal
      This is something quite different and exciting: a log-structured file system, for storing your files on dead trees.
    • by harryandthehenderson ( 1559721 ) on Wednesday June 10, 2009 @10:08AM (#28279089)

      Can anyone explain to me why Linux has so many filesystems?

      Because one filesystem isn't optimal for all cases? Because people want to experiment with new things? Why does it matter?

      Windows has had NTFS for years (admittedly, several versions, but never any compatibility issues that I've come across), and Linux has, what, 73 or something?! Is it really that hard to get it right?

      And Windows has had FAT12, FAT16, FAT32, NTFS, exFAT, VFAT, FFS2, DFS, EFS. Was it really that hard to get it right?

      • by peppepz ( 1311345 ) on Wednesday June 10, 2009 @10:33AM (#28279497)

        You forgot High Sierra, ISO9660, UDF.

        And WinFS. Oh, wait...

      • Re: (Score:2, Informative)

        Because one filesystem isn't optimal for all cases?

        Exactly. You wouldn't use a journaling filesystem (ext3, JFS, XFS) on an SD card. In networked environments, some filesystems are optimized for general use (CIFS, NFS) while others are optimized for a clustered environment (GFS, VMFS), while others are optmized for a distributed environment (Andrew Filesystem, CODA Filesystem). Log-structured filesystems are a new technology that maximizes write throughput, something that is key to optimizing speed in write-heavy environments: this is as opposed to conv

        • by TheRaven64 ( 641858 ) on Wednesday June 10, 2009 @10:54AM (#28279807) Journal

          Log-structured filesystems are a new technology

          Haha! This is the kind of wonderful comment I see a lot from Linux users. The first operating system to ship with a log-structured filesystem was the Sprite kernel in 1990. It was rewritten for 4.4BSD, which was released in 1995. Then, 15 years later, suddenly Linux developers hear about it and it's a brand new technology.

          Linux is not the whole world. Most of the 'new' technologies in Linux appeared in other UNIX-like systems first, and many of the implementations in Linux are inferior to the originals (although some are better).

      • by Anonymous Coward on Wednesday June 10, 2009 @10:42AM (#28279621)

        Because one filesystem isn't optimal for all cases?

        The thing is, Linux strives so hard for the "optimum" that, while doing so, they end up in mediocrity. That's because its programmers are so concerned with micro-optimizations and top speed that they lack the ability to design properly and make good abstractions.

        Would it really be that hard to have ONE good fs that you could tune to different use cases? Probably not. But the average Linux coder sees that something isn't fast in case X and goes ahead redoing the entire wheel. And why? Because the thing he just looked at wasn't designed very well either and can't be adapted easily to different use scenarios. And why? Because it was done by a half-assed coder like himself. And so the circle closes.

        Linux needs more people that can properly design software and make good abstractions - instead of narrow-minded code monkeys that can't see beyond their own crap that they are willing to completely rewrite in two revisions anyway because they lost the big picture.

        • by profplump ( 309017 ) <zach-slashjunk@kotlarek.com> on Wednesday June 10, 2009 @01:24PM (#28282023)

          Did you miss the abstraction layer linux already has for file systems -- VFS? The layer that lets all file-related system calls like be unified among all file systems, so that a file system is only responsible for actually talking to the disk? The same sort of system used by BSD and Windows? Doesn't that essentially make new file systems as minimal as possible while still allowing "tuning"?

      • Why does it matter?

        Complexity, attack vectors, having to spread out resources to commit to maintaining multiple file systems. I'm not saying I'm an advocate for anything specific here, and I'm not saying Windows is better, I'm just trying to answer your question on why it *can* be a bad idea to just keep pushing out file systems. Of course, Linux might need this direly, and then this fs might be a good idea.

    • by tick-tock-atona ( 1145909 ) on Wednesday June 10, 2009 @10:09AM (#28279099)
      Can anyone explain to me why Windows has so many viruses? Linux has had no viruses for years (admittedly, several attempts, but never any in the wild that I've come across), and Windows has, what, 73 billion or something?! Is it really that hard to get it right?
      • by LWATCDR ( 28044 )

        Funny but has Windows had many Viruses lately. I know that Virus has become a common term for all sorts of programs but this is Slashdot. I thought that a Virus was a program that would self propagate usually using a boot block method. Not many people boot from floppies any more.
        You do have some email viruses but the vast majority of those I would call Trojans and depend on somebody to actually run them.
        Then of course your have all sorts of exploits that then try to infect other systems but wouldn't those

      • Re: (Score:3, Insightful)

        by mcgrew ( 92797 )

        Must... not... feed... [kuro5hin.org] Ah, screw it.

        There are a lot of reasons why windows has so many viruses. The one touted by Windows fans is that 90% of PCs have Windows, making it a fat target. Of course, this discounts the fact that Apple sells millions of computers every year, which should make it a fat target, too, but I don't see any Apple viruses either.

        But the 90% seems to me to be the reason, but a different reason - Microsoft has no incentive to "get it right". As long as they can get their OS preinstalled on

        • The reason why you have so much Windows malware and so little for Mac (aside of the smaller target) is simply the same why you get more Windows software and fewer Mac software (at least in areas where core system knowledge is required, as is for malware): Fewer programmers who know the inner workings of the OS.

          It's actually that simple. How many people do you know that could write a driver for MacOS? And how many for Windows?

          Malware doesn't come to life automagically when someone wants it. Like any software

      • Because one virus isn't optimal for all cases? Because people want to experiment with new things? Why does it matter?

    • POHMELFS is a filesystem that can do what a lot of people have been wanting to do for a while: Use that extra 100GB (or TB/PB for our future readers) that no one ever uses on their workstations as redundant distributed network storage, or the same for clusters instead of buying dedicated storage machines. Of course, this requires Linux with kernel 2.6.30 running on all those workstations.
    • by Dystopian Rebel ( 714995 ) * on Wednesday June 10, 2009 @10:27AM (#28279409) Journal

      NILFS2 is the successor to MILFS2, which was based on the "Mother" specification.

      NILFS2 is based on the "Nanny" specification, which means it is younger, firmer, *and* keeps the child nodes quiet when you are not actively updating its data.

    • Re: (Score:3, Informative)

      by wolrahnaes ( 632574 )

      It's the combination of a bit of NIH plus the freedom that Linux brings to a programmer. If you know enough C to not break things horribly and can operate Google, you can create a filesystem. There are also hundreds of proprietary filesystems from older hardware running other OSes, and Linux supports a number of those thanks to users of those older systems developing drivers for them.

      I'd bet that the vast majority of filesystems supported by Linux are rarely if ever used, and when used they're operated in

    • Re: (Score:3, Informative)

      by Ephemeriis ( 315124 )

      Can anyone explain to me why Linux has so many filesystems? Windows has had NTFS for years (admittedly, several versions, but never any compatibility issues that I've come across), and Linux has, what, 73 or something?! Is it really that hard to get it right?

      First up, you've got some incorrect assumptions/information about Windows.

      Windows has not had just NTFS for years. Windows has gone through several different flavors of FAT (FAT12, FAT16, FAT32, exFAT, VFAT).

      As far as NTFS goes... You dismiss the various versions, but then you're counting revisions to the various filesystems in Linux. NTFS has gone through four or five major revisions. Microsoft doesn't really advertise those revisions... They just keep calling it NTFS. But those revisions have added

    • by LWATCDR ( 28044 )

      Well the idea that NTFS got it right is funny to start with. Defrag??? I still have to freaking Defrag?
      Take a look at the feature set of ZFS and NTFS and tell me that NTFS got it right.
      Also Windows does have few more file systems like Fat32 and VFat. You can also install EXT support in XP as well.
      The real answer is that one file system doesn't work for everything. You only think that Windows has a single filesystem because that is the the default option. If you install Ubuntu you will get EXT3 I think by

    • by ifrag ( 984323 )

      Windows has had NTFS for years (admittedly, several versions, but never any compatibility issues that I've come across)

      XP will kill Shadow Copy data from Vista on NTFS volumes. Granted, all the data should be there and read / writes should work fine so it's not really a serious "compatibility" issue, it's more just like feature incompatibility. Of course after going back to Vista if you needed a prior version it's gone. And there might be some problems with System Restore if it's using Shadow Copy featu

    • by leoc ( 4746 ) on Wednesday June 10, 2009 @11:55AM (#28280579) Homepage

      Can anyone explain to me why Ford has so many kinds of cars? Tesla has had a 2 seat roadster for years (admittedly, several versions, but never any compatibility issues that I've come across), and Ford has, what, 73 or something?! Is it really that hard to get it right?

  • by Bob9113 ( 14996 ) on Wednesday June 10, 2009 @10:05AM (#28279029) Homepage

    Integrity Management Architecture

    Contributor: IBM

    Recommended LWN article: http://lwn.net/Articles/227937/ [lwn.net]

    The Trusted Computing Group(TCG) runtime Integrity Measurement Architecture(IMA) maintains a list of hash values of executables and other sensitive system files, as they are read or executed. If an attacker manages to change the contents of an important system file being measured, we can tell. If your system has a TPM chip, then IMA also maintains an aggregate integrity value over this list inside the TPM hardware, so that the TPM can prove to a third party whether or not critical system files have been modified.

    From the recommended article, the key dilemma:

    There are clear advantages to a structure like this. A Linux-based teller machine, say, or a voting machine could ensure that it has not been compromised and prove its integrity to the network. Administrators in charge of web servers can use the integrity code in similar ways. In general, integrity management can be a powerful tool for people who want to be sure that the systems they own (or manage) have not be reconfigured into spam servers when they weren't looking.

    The other side of this coin is that integrity management can be a powerful tool for those who wish to maintain control over systems they do not own. Should it be merged, the kernel will come with the tools needed to create a locked-down system out of the box. As these modules get closer to mainline confusion, we may begin to see more people getting worried about them. Quite a few kernel developers may oppose license terms intended to prevent "tivoization," but that doesn't mean they want to actively support that sort of use of their software. Certainly it would be harder to argue against the shipping of locked-down, Linux-based gadgets when the kernel, itself, provides the lockdown tools.

    OK, maybe this is overdramatic, but trading freedom from third-party oversight through trusted computing for the security of first-party oversight through trusted computing seems a little like:

    "They that can give up essential liberty to obtain a little temporary safety deserve neither liberty nor safety." - Benjamin Franklin

    But I can see both sides. Pondering... what are your thoughts?

  • I found the kernel thread where the original author of the FS-Cache patches, David Howell, makes it clear that on a quiet network with a quite fast server metadata will take longer from the cache. However, at my work we have very busy large NFS servers connected over the building network which is very busy. When you try to read a large file repeatedly in the middle of the day the traditional NFS caching just doesn't work if the time between reads is more than about 5 minutes. I've resorted to manually c
    • Re: (Score:3, Interesting)

      by TooMuchToDo ( 882796 )
      Have you looked at pnfs for performance reasons? We use it with upwards of 300TB of spinning media and 17PB of tape, and it works like a champ.
  • by mrpacmanjel ( 38218 ) on Wednesday June 10, 2009 @10:10AM (#28279101)

    Have wireless "issues" been fixed with this release.

    I have a laptop with generic realtek rt2500 wifi hardware.
    For many kernel releases I have to compile seperate drivers (Legacy serialmonkey) because the "stock" drivers are woefully unstable.
    I either lose my connection, painfully slow( have tried the "rate 54" fix) or I cannot reconnect to my network at all.

    I don't mind compiling seperate drivers (a huge benefit of open source stuff & Linux) but I am concerned how long I will be able to do this (E.g. something changes in the kernel makes the "external" driver break - in fact actual development of the legacy drivers has ceased - http://rt2x00.serialmonkey.com/wiki/index.php/Main_Page [serialmonkey.com])?

    I know I should not be moaning about this but this issue has been around for ages and seems to affect a lot of hardware.

    This is my only niggle with Linux and I am grateful for everything. Computing become much more interesting and fun again.

    Huge thanks to Linus and the kernel developers.

    • by Nursie ( 632944 )

      Can't remember when I last had dealings with this. I have used ralink devices though....

      Looking at the rt2x00, kernels from 2.6.24 onwards should have the rt2x00 driver right there in the kernel, so it should Just Work(TM). You shouldn't need to build the older, legacy drivers any more.

      I agree though, it's a pain making sure to rebuild the driver modules every time you have a kernel update. I've had to do it with the atheros chipset in my laptop. Hopefully, as these device drivers become official, we get to

      • That's right - think it was that version of the kernel I started having problems.

        Maybe the rt2500 in some laptops may affect the driver - I'm not sure.

        I have to blacklist the following drivers; rt2x00pci, rt2500pci and rt2500lib. (I think thats right)

        They just don't seem to work properly.

        I then compile the legacy version of the rt2500 drivers.

        (Mind you I could not get the drivers to compile properly on SUSE 11.1 - it's the only thing that stopped me using SUSE)

    • by Bigby ( 659157 )

      I don't have any issues with rt2500 in 2.6.28. However, I do have issues with ath9k and also issues with using smbfs and autofs, like my card is half duplex.

    • Re: (Score:3, Insightful)

      by CAIMLAS ( 41445 )

      That's not the half of it. The kernel devs appear to break things - intentionally - and leave them that way.

      Case in point, PCMCIA was/is supposedly being rewritten. It broke around kernel 2.6.27 for me (I think) on several systems with ricoh integrated chipsets: I'm unable to use my cardbus or CF slot unless I boot with the device in the slot (and not remove it). Supposedly (according to mailing list info I found) this is due to a 'rewrite' of the pcmcia architecture code. I guess they didn't want to leave

  • by Vu1turEMaN ( 1270774 ) on Wednesday June 10, 2009 @10:19AM (#28279247)

    When they say "Support for rt3070 driver for recent RaLink Wi-Fi chipsets", they really mean support for RT2870, RT2770, RT307X, RT3572 chipsets (they're all the same, with just features enabled or disabled, or signal strength improved between them).

    This was the one last thing for me to fully switch over to linux. Netgear and alot of other Wireless-N USB adapters use these chipsets, and they are the best around.

    Previously, the method of installing this driver was the largest pain in the ass I've ever had to go through as a linux noob (http://ubuntuforums.org/showthread.php?t=960642) and I'm so very very glad to see that this chipset is now supported.

    The reason it was so hard is that the normal controlling app for the USB device has many advanced features you normally don't see on a wireless adapter (act as a router, full cisco network compatibility, etc etc).

  • by Anonymous Coward

    If Linux is ever going to make it on the desktop, developers are going to need to get their shit together and: make webcams work (they don't in the majority of cases at the moment); stop regressions in graphics drivers; get other hardware working, e.g. iPods; make dual-screen work without spending 20 minutes fucking around (see Lunduke's presentation); get GNOME on to QT and develop a decent HIG (sorry, the current GNOME HIG is an excuse to put off doing anything about bugs, see Apple's for how this should

  • by DrogMan ( 708650 ) on Wednesday June 10, 2009 @12:30PM (#28281161) Homepage
    Downloaded. Compiled. Installed and rebooted, and it's running on a little test "embedded" box I'm playing with. (Geode LX800) It's passed all my own tests, and that's that.

    Like the new compression stuff. Compressed kernel under 1MB again - First time I've seen that for a while.

    Now to try it on my Acer Aspire One...

  • NILFS2 (Score:3, Interesting)

    by Eil ( 82413 ) on Wednesday June 10, 2009 @05:00PM (#28285203) Homepage Journal

    So I've been reading that NILFS is the dog bollocks when it comes to solid-state disks in terms of speed and longevity of the disk. However, what I'd like to know is whether any of the advantages will hold for regular old mechanical disks as well. If so, I'd love to try NILFS. Having a real honest-to-goodness versioning filesystem with instant snapshots on my file servers would be so great, I can hardly find the words to describe it.

"What man has done, man can aspire to do." -- Jerry Pournelle, about space flight

Working...