- 26 Nov, 2002 40 commits
-
-
Miles Bader authored
Update v850 to use kstat_cpu in irqs.c
-
Miles Bader authored
The v850 doesn't need sys_old_getrlimit
-
Miles Bader authored
Add TASK_UNMAPPED_BASE for v850
-
Miles Bader authored
Update v850 usage of do_fork to supply new args
-
Miles Bader authored
Add v850 support for initramfs
-
Miles Bader authored
Random minor fixes for v850 `anna' platform
-
Miles Bader authored
Update `nb85e_uart' driver for recent serial framework changes
-
Miles Bader authored
Make NB85E_UART_CKSR_MAX_FREQ (in nb85e_uart driver) overridable [The proper value is CPU-dependent]
-
Miles Bader authored
Add v850 nanosecond stat fields
-
Miles Bader authored
Add support for AS85EP1 platform to v850 arch
-
Miles Bader authored
Add v850 versions of dump_stack and show_stack Also move v850 show_regs to same source file (bug.c)
-
Miles Bader authored
Add v850 support for new module loader
-
Miles Bader authored
Add asm macros on v850 so the size of asm data objects can be recorded [This just makes it easier to see the sizes using readelf or whatever]
-
Miles Bader authored
Fix some minor type problems in v850 code [Mostly signed/unsigned mismatches]
-
Miles Bader authored
Consolidate various v850 platform banner printks
-
Andrew Morton authored
cpu_buffer.c needs errno.h for ENOMEM.
-
-
http://linux-acpi.bkbits.net/linux-acpiLinus Torvalds authored
into home.transmeta.com:/home/torvalds/v2.5/linux
-
bk://lsm.bkbits.net/linus-2.5Linus Torvalds authored
into penguin.transmeta.com:/home/penguin/torvalds/repositories/kernel/lsm-tree
-
Andries E. Brouwer authored
That annoying ldm message is just debugging output somebody forgot to remove.
-
Greg Kroah-Hartman authored
into kroah.com:/home/greg/linux/BK/lsm-2.5
-
Andy Grover authored
into groveronline.com:/root/bk/linux-acpi
-
-
Andrew Morton authored
Shrinking a huge number of dentries or inodes can hold dcache_lock or inode_lock for a long time. Not only does this hold off preemption - holding those locks basically shuts down the whole VFS. A neat fix for all such caches is to chunk the work up at the shrink_slab() level. I made the chunksize pretty small, for scalability reasons - avoid holding the lock for too long so another CPU can come in, acquire it and go off to do some work.
-
Andrew Morton authored
Running sync_blockdev() inside lock_kernel causes long scheduling stalls under some conditions because the lock_kernel() disables preemption. We don't need the BKL across sync_blockdev(), so move it down a bit. This is a minimal fix - we probably don't need BKL for much at all in there.
-
Andrew Morton authored
When the VM is under heavy load and a page allocator caller is holding the BKL there can be several-millisecond scheduling stalls. So add a cond_resched() at every 32nd page (approx).
-
Andrew Morton authored
From Hugh. Remove some strangeness in the swapoff path. "it dates from the days when that block set *swap_map to 0, but in 2.4.13 I changed it to set *swap_map to 1 and delete_from_swap_cache afterwards: it's been wrong ever since, hasn't it? swap_list_locking can go, it was there to guard nr_swap_pages for si_swapinfo; the swap_device_locking has to stay because swap_map is an array of unsigned _shorts_."
-
Andrew Morton authored
There are a couple of statistical functions which scan the entire swap map counting things up, to display in /proc. On my machine, these hold spinlocks for 19 milliseconds which is unacceptable from a scheduling latency point of view. And an application which sits in a loop reading /proc/swaps on a large machine is probably a decent denial-of-service attack - it will limit swap allocations to tens of pages per second. So add a counter to swap_info_struct and use it to track how many pages are currently in use, so those reporting functions don't need to add them all up.
-
Andrew Morton authored
This is the first in a little batch of patches which address long-held locks in the VFS/MM layer which are affecting our worst-case scheduling latency, and are making CONFIG_PREEMPT not very useful. We end up with a worst-case of 500 microseconds at 500MHz, which is very good. Unless you do an exit with lots of mmapped memory. unmap_page_range() needs work. Some of these patches also add rescheduling points for non-preemptible kernels - where I felt that the code path could be long enough to be perceptible. Three places in the generic pagecache functions need manual rescheduling points even for non-preemptible kernels: - generic_file_read() (Can hold the CPU for seconds) - generic_file_write() (ditto) - filemap_fdatawait(). This won't hold the CPU for so long, but it can walk many thousands of pages under the lock. It needs a lock drop and scheduling point for both preemptible and non-preemptible kernels. (This one's a bit ugly...)
-
Andrew Morton authored
pdflush writes back chunks of ~1000 pages. It currently takes a short nap if it writes back no pages at all. That could cause it to write back lots of small batches of pages, as it bounces against a congested queue. Change it to sleep if it failed to write back the entire batch against a congested queue. Ths reduces the context switch rate a little. The context switch rate is still fairly high (150/sec) - this appears to be due to add_disk_randomness() scheduling a work function.
-
Andrew Morton authored
Patch from Bill Irwin. It has the potential to break userspace monitoring tools a little bit, and I'm a rater uncertain about how useful the per-process per-cpu accounting is. Bill sent this out as an RFC on July 29: "These statistics severely bloat the task_struct and nothing in userspace can rely on them as they're conditional on CONFIG_SMP. If anyone is using them (or just wants them around), please speak up." And nobody spoke up. If we apply this, the contents of /proc/783/cpu will go from cpu 1 1 cpu0 0 0 cpu1 0 0 cpu2 1 1 cpu3 0 0 to cpu 1 1 And we shall save 256 bytes from the ia32 task_struct. On my SMP build with NR_CPUS=32: Without this patch, sizeof(task_struct) is 1824, slab uses a 1-order allocation and we are getting 2 task_structs per page. With this patch, sizeof(task_struct) is 1568, slab uses a 2-order allocation and we are getting 2.5 task_structs per page. So it seems worthwhile. (Maybe this highlights a shortcoming in slab. For the 1824-byte case it could have used a 0-order allocation)
-
Andrew Morton authored
Patch from Petr Baudis <pasky@ucw.cz> This patch (against 2.5.49) updates Documentation/kernel-parameters.txt to the current state of kernel. It was somehow abadonded lately, so I did my best, but it's possible that I still missed some of the options - thus, if you will notice your favourite boot option missing there, please speak up. Note also that I will probably send up another update after few further kernel releases.. Also, I attempted to introduce some uniform format to the entries, I added the format description where I was able to find it out and decypher it, and I also added gross amount of external links to the headers of the source files or to the README-like files, where the options are described into more degree. This way, hopefully this file has a chance to be actually usable for the users ;-). There are almost certainly some entries which I missed, it was really a huge number and the main reason is that some of the boot options don't use the __setup macro, which I grep'd for. I hope the patch is ok, there should be no problems with it. Please apply. Note that this is the fourth submission of the patch - I took the opportunity and updated the patch from 2.5.48 to 2.5.49. AFAIK mutt shouldn't mangle the patch in any way, so it should apply cleanly to your tree, Linus.
-
Andrew Morton authored
Patch from Dipankar Sarma <dipankar@gamebox.net> This is a trivial cleanup removing two old unused macros from kernel_stat.h that made no sense with the new per-CPU kstat. Also included a few finicky coding style changes. Please apply.
-
Andrew Morton authored
Patch from Ravikiran G Thirumalai <kiran@in.ibm.com> Allows the kernel profiler to use a shift value of zero for high-resolution profiling. So instead of overloading a shift of zero to mean "not profiling", use a separate flag for that. This means that anyone who is currently using "profile=0" to disable profiling has accidentally turned profiling on, so I added a printk.
-
Andrew Morton authored
- Missing parentheses around macro arg in __pte_offest() - struct page_state needs to be cacheline aligned (the first seven fields should be in the same cacheline)
-
Andrew Morton authored
blk_run_queues() places queues onto a local list without locking. But interrupt-time activity in scsi_lib.c will replug these queues, which involves accessing that list which blk_run_queues() is walking. Net effect: list corruption and an oops. Very hard to reproduce... So hold the lock while blk_run_queues() is walking the local list. The patch was acked by Jens. It also uninlines a few largeish functions and adds a couple of WARN_ON()s in code which worried me.
-
Linus Torvalds authored
Cset exclude: davej@codemonkey.org.uk|ChangeSet|20021126023731|33210
-
Linus Torvalds authored
-
Linus Torvalds authored
into home.transmeta.com:/home/torvalds/v2.5/linux
-
Linus Torvalds authored
Cset exclude: alan@lxorguk.ukuu.org.uk|ChangeSet|20021126021252|43411
-