1. 05 Oct, 2002 33 commits
  2. 06 Oct, 2002 2 commits
  3. 05 Oct, 2002 5 commits
    • Kai Germaschewski's avatar
      kbuild: Nicer warnings · 56a8f5d4
      Kai Germaschewski authored
      Improve the warning messages when using obsolete features, kill one
      remaining user of $(list-multi)
      
      (by Sam Ravnborg)
      
      I also made O_TARGET != built-in.o an error, since compatibility code for
      that case has already been dropped 
      56a8f5d4
    • Linus Torvalds's avatar
      Merge bk://linux-bt.bkbits.net/bt-2.5 · 6cab0e06
      Linus Torvalds authored
      into home.transmeta.com:/home/torvalds/v2.5/linux
      6cab0e06
    • Andrew Morton's avatar
      [PATCH] clean up ll_rw_block() · 61c4b8fb
      Andrew Morton authored
      Hardly anything uses this function, so the debug checks in there are
      not of much value.
      
      The check for bdev_readonly() should be done in submit_bio().
      
      Local variable `major' was altogether unused.
      61c4b8fb
    • Andrew Morton's avatar
      [PATCH] stricter dirty memory clamping · 3669e824
      Andrew Morton authored
      The ratelimiting logic in balance_dirty_pages_ratelimited() is designed
      to prevent excessive calls to the expensive get_page_state(): On a big
      machine we only check to see if we're over dirty memory limits once per
      1024 dirtyings per cpu.
      
      This works OK normally, but it has the effect of allowing each process
      to go 1024 pages over the dirty limit before it gets throttled.
      
      So if someone runs 16000 tiobench threads, they can go 16G over the
      dirty memory threshold and die the death of buffer_head consumption.
      Because page dirtiness pins the page's buffer_heads, defeating the
      special buffer_head reclaim logic.
      
      I'd left this overshoot artifact in place because it provides a degree
      of adaptivity - of someone if running hundreds of dirtying processes
      (dbench!) then they do want to overshoot the dirty memory limit.
      
      But it's hard to balance, and is really not worth the futzing around.
      So change the logic to only perform the get_page_state() call rate
      limiting if we're known to be under the dirty memory threshold.
      3669e824
    • Andrew Morton's avatar
      [PATCH] remove page->virtual · a27efcaf
      Andrew Morton authored
      The patch removes page->virtual for all architectures which do not
      define WANT_PAGE_VIRTUAL.  Hash for it instead.
      
      Possibly we could define WANT_PAGE_VIRTUAL for CONFIG_HIGHMEM4G, but it
      seems unlikely.
      
      A lot of the pressure went off kmap() and page_address() as a result of
      the move to kmap_atomic().  That should be the preferred way to address
      CPU load in the set_page_address() and page_address() hashing and
      locking.
      
      If kmap_atomic is not usable then the next best approach is for users
      to cache the result of kmap() in a local rather than calling
      page_address() repeatedly.
      
      One heavy user of kmap() and page_address() is the ext2 directory code.
      
      On a 7G Quad PIII, running four concurrent instances of
      
      	while true
      	do
      		find /usr/src/linux > /dev/null
      	done
      
      on ext2 with everything cached, profiling shows that the new hashed
      set_page_address() and page_address() implementations consume 0.4% and
      1.3% of CPU time respectively.   I think that's OK.
      a27efcaf