On Sun, Dec 16, 2007 at 07:59:30PM +0800, Fengguang Wu wrote: > @@ -1321,78 +1401,69 @@ int filemap_fault(struct vm_area_struct > struct address_space *mapping = file->f_mapping; > struct file_ra_state *ra = &file->f_ra; > struct inode *inode = mapping->host; > + pgoff_t offset = vmf->pgoff; > struct page *page; > unsigned long size; > - int did_readaround = 0; > int ret = 0; > > size = (i_size_read(inode) + PAGE_CACHE_SIZE - 1) >> PAGE_CACHE_SHIFT; > if (vmf->pgoff >= size) > return VM_FAULT_SIGBUS; > > - /* If we don't want any read-ahead, don't bother */ > - if (VM_RandomReadHint(vma)) > - goto no_cached_page; > - > /* > - * Do we have something in the page cache already? > + * Do we have something in the page cache already that > + * is unlocked and already up-to-date? > */ > -retry_find: > - page = find_lock_page(mapping, vmf->pgoff); > - /* > - * For sequential accesses, we use the generic readahead logic. > - */ > - if (VM_SequentialReadHint(vma)) { > - if (!page) { > - page_cache_sync_readahead(mapping, ra, file, > - vmf->pgoff, 1); > - page = find_lock_page(mapping, vmf->pgoff); > - if (!page) > - goto no_cached_page; > - } > - if (PageReadahead(page)) { > - page_cache_async_readahead(mapping, ra, file, page, > - vmf->pgoff, 1); > - } > - } > + read_lock_irq(&mapping->tree_lock); > + page = radix_tree_lookup(&mapping->page_tree, offset); > + if (likely(page)) { > + int got_lock, uptodate; > + page_cache_get(page); > + > + got_lock = !TestSetPageLocked(page); > + uptodate = PageUptodate(page); > + read_unlock_irq(&mapping->tree_lock);
If we could avoid open coding tree_lock here (and expanding its coverage to PageUptodate), that would be nice. I don't think it gains us too much. -- To unsubscribe from this list: send the line "unsubscribe linux-kernel" in the body of a message to [EMAIL PROTECTED] More majordomo info at http://vger.kernel.org/majordomo-info.html Please read the FAQ at http://www.tux.org/lkml/