1. 27 Sep, 2002 7 commits
  2. 26 Sep, 2002 5 commits
    • Tim Hockin's avatar
      Add dp83816 support to drivers/net/natsemi.c · 1766b717
      Tim Hockin authored
      1766b717
    • Jes Sorensen's avatar
    • Jeff Garzik's avatar
      Merge mandrakesoft.com:/home/jgarzik/repo/linus-2.5 · 3319d0af
      Jeff Garzik authored
      into mandrakesoft.com:/home/jgarzik/repo/net-drivers-2.5
      3319d0af
    • Jens Axboe's avatar
      [PATCH] deadline ioscheduler cleanups · 82c8a4cc
      Jens Axboe authored
      Some various small cleanups, optimizations, and fixes.
      
      o Make fifo_batch=32 as default, from testing this appears a good
        default value. We still get good throughput, and latency is good.
      
      o Reintroduce the merge_cleanup logic. We need it for deadline for
        rehashing requests when they have been merged.
      
      o Cleanup last_merge logic. Move it to the new elv_merged_request(),
        this is where it really belongs. Doing it inside the io scheduler core
        can causes false positives, when the queue merge functions reject an
        otherwise good merge
      
      o Have deadline_move_requests() account from last entry on the dispatch
        queue, if it is non-empty. It doesn't really matter what the last
        extracted sector was, if we are not right behind it.
      
      o Clean/optimize deadline_move_requests()
      
      o Account size of a request just a little bit. Streaming transfer isn't
        for free, it's just a lot cheaper than a seek.
      
      o Make deadline_check_fifo() more readable.
      82c8a4cc
    • Ingo Molnar's avatar
      [PATCH] exit-fix-2.5.38-F0 · 9b34eb70
      Ingo Molnar authored
      From Andrew Morton.
      
      There are a couple of places where we would enable interrupts while
      write-holding the tasklist_lock ...  nasty.
      9b34eb70
  3. 25 Sep, 2002 25 commits
    • Jeff Garzik's avatar
      Merge mandrakesoft.com:/home/jgarzik/repo/linus-2.5 · 664a55c9
      Jeff Garzik authored
      into mandrakesoft.com:/home/jgarzik/repo/net-drivers-2.5
      664a55c9
    • Linus Torvalds's avatar
      Merge bk://ldm.bkbits.net/linux-2.5 · 2ce067b0
      Linus Torvalds authored
      into home.transmeta.com:/home/torvalds/v2.5/linux
      2ce067b0
    • Andrew Morton's avatar
      [PATCH] tighter locking in pdflush · 4ab1a3e6
      Andrew Morton authored
      Had a weird oops from Bill Irwin - the pdflush_list was corrupt.
      
      The only thing I can think of is that something sprayed out a wakeup
      when it shouldn't.  So tighten things up against that, and add some
      printks to catch it if it happens again.
      4ab1a3e6
    • Andrew Morton's avatar
      [PATCH] speed up sys_sync() · 57eb0613
      Andrew Morton authored
      Well it's a one-liner.  sys_sync() only syncs one queue at a time, and
      can be slow if you have a lot of disks.  So poke pdflush, which knows
      how to write all the queues in parallel.
      57eb0613
    • Andrew Morton's avatar
      [PATCH] increase traffic on linux-kernel · 4f3e8109
      Andrew Morton authored
      [This has four scalps already.  Thomas Molina has agreed
       to track things as they are identified ]
      
      Infrastructure to detect sleep-inside-spinlock bugs.  Really only
      useful if compiled with CONFIG_PREEMPT=y.  It prints out a whiny
      message and a stack backtrace if someone calls a function which might
      sleep from within an atomic region.
      
      This patch generates a storm of output at boot, due to
      drivers/ide/ide-probe.c:init_irq() calling lots of things which it
      shouldn't under ide_lock.
      
      It'll find other bugs too.
      4f3e8109
    • Andrew Morton's avatar
      [PATCH] slab reclaim balancing · b65bbded
      Andrew Morton authored
      A patch from Ed Tomlinson which improves the way in which the kernel
      reclaims slab objects.
      
      The theory is: a cached object's usefulness is measured in terms of the
      number of disk seeks which it saves.  Furthermore, we assume that one
      dentry or inode saves as many seeks as one pagecache page.
      
      So we reap slab objects at the same rate as we reclaim pages.  For each
      1% of reclaimed pagecache we reclaim 1% of slab.  (Actually, we _scan_
      1% of slab for each 1% of scanned pages).
      
      Furthermore we assume that one swapout costs twice as many seeks as one
      pagecache page, and twice as many seeks as one slab object.  So we
      double the pressure on slab when anonymous pages are being considered
      for eviction.
      
      The code works nicely, and smoothly.  Possibly it does not shrink slab
      hard enough, but that is now very easy to tune up and down.  It is just:
      
      	ratio *= 3;
      
      in shrink_caches().
      
      Slab caches no longer hold onto completely empty pages.  Instead, pages
      are freed as soon as they have zero objects.  This is possibly a
      performance hit for slabs which have constructors, but it's doubtful.
      Most allocations after a batch of frees are satisfied from inside
      internally-fragmented pages and by the time slab gets back onto using
      the wholly-empty pages they'll be cache-cold.  slab would be better off
      going and requesting a new, cache-warm page and reconstructing the
      objects therein.  (Once we have the per-cpu hot-page allocator in
      place.  It's happening).
      
      As a consequence of the above, kmem_cache_shrink() is now unused.  No
      great loss there - the serialising effect of kmem_cache_shrink and its
      semaphore in front of page reclaim was measurably bad.
      
      Still todo:
      
      - batch up the shrinking so we don't call into prune_dcache and
        friends at high frequency asking for a tiny number of objects.
      
      - Maybe expose the shrink ratio via a tunable.
      
      - clean up slab.c
      
      - highmem page reclaim in prune_icache: highmem pages can pin
        inodes.
      b65bbded
    • Andrew Morton's avatar
      [PATCH] use prepare_to_wait in VM/VFS · dfdacf59
      Andrew Morton authored
      This uses the new wakeup machinery in some hot parts of the VFS and
      block layers.
      
      wait_on_buffer(), wait_on_page(), lock_page(), blk_congestion_wait().
      Also in get_request_wait(), although the benefit for exclusive wakeups
      will be lower.
      dfdacf59
    • Andrew Morton's avatar
      [PATCH] prepare_to_wait/finish_wait sleep/wakeup API · 3da08d6c
      Andrew Morton authored
      This is worth a whopping 2% on spwecweb on an 8-way.  Which is faintly
      surprising because __wake_up and other wait/wakeup functions are not
      apparent in the specweb profiles which I've seen.
      
      
      The main objective of this is to reduce the CPU cost of the wait/wakeup
      operation.  When a task is woken up, its waitqueue is removed from the
      waitqueue_head by the waker (ie: immediately), rather than by the woken
      process.
      
      This means that a subsequent wakeup does not need to revisit the
      just-woken task.  It also means that the just-woken task does not need
      to take the waitqueue_head's lock, which may well reside in another
      CPU's cache.
      
      I have no decent measurements on the effect of this change - possibly a
      20-30% drop in __wake_up cost in Badari's 40-dds-to-40-disks test (it
      was the most expensive function), but it's inconclusive.  And no
      quantitative testing of which I am aware has been performed by
      networking people.
      
      The API is very simple to use (Linus thought it up):
      
      my_func(waitqueue_head_t *wqh)
      {
      	DEFINE_WAIT(wait);
      
      	prepare_to_wait(wqh, &wait, TASK_UNINTERRUPTIBLE);
      	if (!some_test)
      		schedule();
      	finish_wait(wqh, &wait);
      }
      
      or:
      
      	DEFINE_WAIT(wait);
      
      	while (!some_test_1) {
      		prepare_to_wait(wqh, &wait, TASK_UNINTERRUPTIBLE);
      		if (!some_test_2)
      			schedule();
      		...
      	}
      	finish_wait(wqh, &wait);
      
      You need to bear in mind that once prepare_to_wait has been performed,
      your task could be removed from the waitqueue_head and placed into
      TASK_RUNNING at any time.  You don't know whether or not you're still
      on the waitqueue_head.
      
      Running prepare_to_wait() when you're already on the waitqueue_head is
      fine - it will do the right thing.
      
      Running finish_wait() when you're actually not on the waitqueue_head is
      fine.
      
      Running finish_wait() when you've _never_ been on the waitqueue_head is
      fine, as ling as the DEFINE_WAIT() macro was used to initialise the
      waitqueue.
      
      You don't need to fiddle with current->state.  prepare_to_wait() and
      finish_wait() will do that.  finish_wait() will always return in state
      TASK_RUNNING.
      
      There are plenty of usage examples in vm-wakeups.patch and
      tcp-wakeups.patch.
      3da08d6c
    • Andrew Morton's avatar
      [PATCH] mprotect_fixup fix · 02b1783c
      Andrew Morton authored
      From David M-T.
      
      When this function successfully merges the new range into an existing
      VMA, it forgets to extend the new protection mode into the just-merged
      pages.
      02b1783c
    • Andrew Morton's avatar
      [PATCH] hugetlb fix · 5538fdaa
      Andrew Morton authored
      Patch from Rohit Seth
      
      It fixes the problem which Andrea noted in his initial review of the
      hugetlb code:
      
      "In short doing "addr = vma->vm_end" and then checking if vm_end + len
       is below vm_next->vm_start is broken, because there's no guarantee
       that "addr" will be a largepage aligned address.  the LPAGE_ALIGN in
       found_addr should be dropped becaue moving the addr ahead without
       checking that addr+len doesn't then fall into a vma, will generate
       do_munmaps and in turn userspace mem corruption."
      5538fdaa
    • Martin J. Bligh's avatar
      [PATCH] NUMA-Q fixes · bce5aeb5
      Martin J. Bligh authored
       - Remove the const that someone incorrectly stuck in there, it type conflicts.
         Alan has a better plan for fixing this long term, but this fixes the compile
         warning for now.
      
       - Move the printk of the xquad_portio setup *after* we put something in the variable
         so it actually prints something useful, not 0 ;-)
      
       - To derive the size of the xquad_portio area, multiply the number of nodes by the
         size of each nodes, not the size of two nodes (and remove define). Doh!
      bce5aeb5
    • Linus Torvalds's avatar
      Remove busy-wait for short RT nanosleeps. It's a random special case · 98ae8e2b
      Linus Torvalds authored
      and does the wrong thing for higher HZ values anyway.
      98ae8e2b
    • Patrick Mochel's avatar
      Merge bk://ldm@bkbits.net/linux-2.5 · 56d8b39d
      Patrick Mochel authored
      into osdl.org:/home/mochel/src/kernel/devel/linux-2.5
      56d8b39d
    • Patrick Mochel's avatar
      add disk device class · abe2e064
      Patrick Mochel authored
      abe2e064
    • Patrick Mochel's avatar
      Merge bk://linus.bkbits.net/linux-2.5 · 4a99b33d
      Patrick Mochel authored
      into hostme.bitkeeper.com:/ua/repos/l/ldm/linux-2.5
      4a99b33d
    • Ingo Molnar's avatar
      [PATCH] exit-fix-2.5.38-E3 · 5dd6a6e5
      Ingo Molnar authored
      This fixes a number of bugs in the thread-release code:
      
       - notify parents only if the group leader is a zombie,
         and if it's not a detached thread.
      
       - do not reparent children to zombie tasks.
      
       - introduce the TASK_DEAD state for tasks, to serialize the task-release
         path. (to some it might be confusing that tasks are zombies first, then
         dead :-)
      
       - simplify tasklist_lock usage in release_task().
      
      the effect of the above bugs ranged from unkillable hung zombies to kernel
      crashes. None of those happens with the patch applied.
      5dd6a6e5
    • Jens Axboe's avatar
      [PATCH] remove elevator_linus · 2684cd69
      Jens Axboe authored
      Patch killing off elevator_linus for good. Sniffle.
      2684cd69
    • Jens Axboe's avatar
      [PATCH] deadline scheduler · 85b2148a
      Jens Axboe authored
      This introduces the deadline-ioscheduler, making it the default.  2nd
      patch coming that deletes elevator_linus in a minute.
      
      This one has read_expire at 500ms, and writes_starved at 2.
      85b2148a
    • Thomas Hood's avatar
      [PATCH] PnP BIOS ESCD sanity check · 650e56ee
      Thomas Hood authored
      Sanity checkthe ESCD size. From 2.4.
      650e56ee
    • Ivan Kokshaysky's avatar
      [PATCH] ALi and Cypress IDE fixes · 26b90050
      Ivan Kokshaysky authored
      These two chipsets are most common on alpha.
      - cy82c693: allow the generic IDE setup code to work correctly
        with broken PCI registers layout of this chip. This fixes
        quite a few problems with secondary channel, plus some hacks in
        arch code can go away.
      - ALi M5229: enable DMA.
      26b90050
    • Adam Radford's avatar
      [PATCH] 3ware driver update for 2.5.35 · 92f2c52c
      Adam Radford authored
      92f2c52c
    • Ingo Molnar's avatar
      [PATCH] pidhash-2.5.38-A0 · 5191a147
      Ingo Molnar authored
      This removes the cmpxchg from the PID allocator and replaces it with a
      spinlock.  This spinlock is hit only a couple of times per bootup, so
      it's not a performance issue.
      5191a147
    • Ingo Molnar's avatar
      [PATCH] thread-flock-2.5.38-A3 · a16435af
      Ingo Molnar authored
      Ulrich found another small detail wrt. POSIX requirements for threads -
      this time it's the recursion features (read-held lock being write-locked
      means an upgrade if the same 'process' is the owner, means a deadlock if a
      different 'process').
      
      this requirement even makes some sense - the group of threads who own a
      lock really own all rights to the lock as well.
      
      These changes fix this, all testcases pass now.  (inter-process
      testcases as well, which are not affected by this patch.)
      
      (SIGURG and SIGIO semantics should also continue to work - there's some
      more stuff we can optimize with the new pidhash in this area, but that's
      for later.)
      a16435af
    • Theodore Y. Ts'o's avatar
      [PATCH] loop device broken in 2.5.38 · 86b18ae3
      Theodore Y. Ts'o authored
      The loop device driver was broken in 2.5.38 when it was converted over
      to use gendisk.  I discovered this while doing final regression testing
      on the ext3 htree code.
      
      The problem is that figure_loop_size() is setting the capacity of the
      loop device in kilobytes (because that's what compute_loop_size()
      returns), but set_capacity() expects the size in 512 byte sectors.
      
      I've enclosed a patch which fixes the problem, as well as simplifying
      the code by eliminating compute_loop_size(), since it is a static
      function is only used once by figure_loop_size().
      86b18ae3
    • Dave Kleikamp's avatar
      Merge jfs@jfs.bkbits.net:linux-2.5 · 0de4d503
      Dave Kleikamp authored
      into kleikamp.austin.ibm.com:/home/shaggy/bk/jfs-2.5
      0de4d503
  4. 24 Sep, 2002 3 commits
    • Jeff Garzik's avatar
      Merge mandrakesoft.com:/home/jgarzik/repo/linus-2.5 · 40e59196
      Jeff Garzik authored
      into mandrakesoft.com:/home/jgarzik/repo/net-drivers-2.5
      40e59196
    • Paul Mackerras's avatar
      [PATCH] fix null dereference in sys_mprotect · 0cd9efe3
      Paul Mackerras authored
      As it is at the moment, sys_mprotect will dereference a null pointer
      if you use it on a region that is contained within the first vma.  I
      have a little program that demonstrates this (I'll post it if anyone
      is interested).  What happens then is that the process hangs in
      do_page_fault at the down_read on the mm->mmap_sem, since sys_mprotect
      has done a down_write on mm->mmap_sem.
      
      The problem is that mprotect_fixup isn't updating prev properly.  Thus
      we can finish the main loop in sys_mprotect with prev == NULL.  This
      has been the case since Christoph's cleanups went in.  Prior to that,
      mprotect_fixup always set prev to something non-NULL.  I suspect that
      not updating prev could also cause vmas to get dropped completely if
      the region being mprotected spans more than one vma.
      
      The patch below fixes the problem by making mprotect_fixup set prev to
      a reasonable value in all circumstances.
      0cd9efe3
    • Linus Torvalds's avatar
      efae82c0