On Sat, 28 Jul 2007 23:52:30 +0100
Andy Whitcroft <[EMAIL PROTECTED]> wrote:

> 
> From: Mel Gorman <[EMAIL PROTECTED]>
> 
> Lumpy reclaim works by selecting a lead page from the LRU list and then
> selecting pages for reclaim from the order-aligned area of pages. In the
> situation were all pages in that region are inactive and not referenced by
> any process over time, it works well.
> 
> In the situation where there is even light load on the system, the pages may
> not free quickly. Out of a area of 1024 pages, maybe only 950 of them are
> freed when the allocation attempt occurs because lumpy reclaim returned early.
> This patch alters the behaviour of direct reclaim for large contiguous blocks.
> 
> The first attempt to call shrink_page_list() is asynchronous but if it
> fails, the pages are submitted a second time and the calling process waits
> for the IO to complete. It'll retry up to 5 times for the pages to be
> fully freed. This may stall allocators waiting for contiguous memory but
> that should be expected behaviour for high-order users. It is preferable
> behaviour to potentially queueing unnecessary areas for IO. Note that kswapd
> will not stall in this fashion.

I agree with the intent.

> +/* Request for sync pageout. */
> +typedef enum {
> +     PAGEOUT_IO_ASYNC,
> +     PAGEOUT_IO_SYNC,
> +} pageout_io_t;

no typedefs.

(checkpatch.pl knew that ;))

>  /* possible outcome of pageout() */
>  typedef enum {
>       /* failed to write page out, page is locked */
> @@ -287,7 +293,8 @@ typedef enum {
>   * pageout is called by shrink_page_list() for each dirty page.
>   * Calls ->writepage().
>   */
> -static pageout_t pageout(struct page *page, struct address_space *mapping)
> +static pageout_t pageout(struct page *page, struct address_space *mapping,
> +                                             pageout_io_t sync_writeback)
>  {
>       /*
>        * If the page is dirty, only perform writeback if that write
> @@ -346,6 +353,15 @@ static pageout_t pageout(struct page *page, struct 
> address_space *mapping)
>                       ClearPageReclaim(page);
>                       return PAGE_ACTIVATE;
>               }
> +
> +             /*
> +              * Wait on writeback if requested to. This happens when
> +              * direct reclaiming a large contiguous area and the
> +              * first attempt to free a ranage of pages fails

cnat tpye.

> +              */
> +             if (PageWriteback(page) && sync_writeback == PAGEOUT_IO_SYNC)
> +                     wait_on_page_writeback(page);
> +
>
>               if (!PageWriteback(page)) {
>                       /* synchronous write or broken a_ops? */
>                       ClearPageReclaim(page);
> @@ -423,7 +439,8 @@ cannot_free:
>   * shrink_page_list() returns the number of reclaimed pages
>   */
>  static unsigned long shrink_page_list(struct list_head *page_list,
> -                                     struct scan_control *sc)
> +                                     struct scan_control *sc,
> +                                     pageout_io_t sync_writeback)
>  {
>       LIST_HEAD(ret_pages);
>       struct pagevec freed_pvec;
> @@ -458,8 +475,12 @@ static unsigned long shrink_page_list(struct list_head 
> *page_list,
>               if (page_mapped(page) || PageSwapCache(page))
>                       sc->nr_scanned++;
>  
> -             if (PageWriteback(page))
> -                     goto keep_locked;
> +             if (PageWriteback(page)) {
> +                     if (sync_writeback == PAGEOUT_IO_SYNC)
> +                             wait_on_page_writeback(page);
> +                     else
> +                             goto keep_locked;
> +             }

This is unneeded and conceivably deadlocky for !__GFP_FS allocations. 
Probably we avoid doing all this if the test which may_enter_fs uses is
false.

It's unlikely that any very-high-order allocators are using GFP_NOIO or
whatever, but still...


-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Reply via email to