From: Wu Fengguang backing_dev_info.ra_expect_bytes is dynamicly updated to be the expected read pages on start-of-file. It allows the initial readahead to be more aggressive and hence efficient. Signed-off-by: Wu Fengguang Signed-off-by: Andrew Morton --- fs/file_table.c | 7 +++++ include/linux/mm.h | 1 mm/readahead.c | 56 +++++++++++++++++++++++++++++++++++++++++++ 3 files changed, 64 insertions(+) diff -puN fs/file_table.c~readahead-initial-method-expected-read-size fs/file_table.c --- a/fs/file_table.c~readahead-initial-method-expected-read-size +++ a/fs/file_table.c @@ -11,6 +11,7 @@ #include #include #include +#include #include #include #include @@ -159,6 +160,12 @@ void fastcall __fput(struct file *file) might_sleep(); fsnotify_close(file); + +#ifdef CONFIG_ADAPTIVE_READAHEAD + if (file->f_ra.flags & RA_FLAG_EOF) + readahead_close(file); +#endif + /* * The function eventpoll_release() should be the first called * in the file cleanup chain. diff -puN include/linux/mm.h~readahead-initial-method-expected-read-size include/linux/mm.h --- a/include/linux/mm.h~readahead-initial-method-expected-read-size +++ a/include/linux/mm.h @@ -1067,6 +1067,7 @@ unsigned long page_cache_readahead(struc void handle_ra_miss(struct address_space *mapping, struct file_ra_state *ra, pgoff_t offset); unsigned long max_sane_readahead(unsigned long nr); +void fastcall readahead_close(struct file *file); #ifdef CONFIG_ADAPTIVE_READAHEAD extern int readahead_ratio; diff -puN mm/readahead.c~readahead-initial-method-expected-read-size mm/readahead.c --- a/mm/readahead.c~readahead-initial-method-expected-read-size +++ a/mm/readahead.c @@ -1559,6 +1559,62 @@ static inline void get_readahead_bounds( (128*1024) / PAGE_CACHE_SIZE), *ra_max / 2); } +/* + * When closing a normal readonly file, + * - on cache hit: increase `backing_dev_info.ra_expect_bytes' slowly; + * - on cache miss: decrease it rapidly. + * + * The resulted `ra_expect_bytes' answers the question of: + * How many pages are expected to be read on start-of-file? + */ +void fastcall readahead_close(struct file *file) +{ + struct inode *inode = file->f_dentry->d_inode; + struct address_space *mapping = inode->i_mapping; + struct backing_dev_info *bdi = mapping->backing_dev_info; + unsigned long pos = file->f_pos; /* supposed to be small */ + unsigned long pgrahit = file->f_ra.hit0; + unsigned long pgcached = mapping->nrpages; + unsigned long pgaccess; + + if (!pos) /* pread */ + return; + + if (pgcached > bdi->ra_pages0) /* excessive reads */ + return; + + pgaccess = max(pgrahit, 1 + pos / PAGE_CACHE_SIZE); + if (pgaccess >= pgcached) { + if (bdi->ra_expect_bytes < bdi->ra_pages0 * PAGE_CACHE_SIZE) + bdi->ra_expect_bytes += pgcached * PAGE_CACHE_SIZE / 8; + + debug_inc(initial_ra_hit); + dprintk("initial_ra_hit on file %s size %lluK " + "pos %lu by %s(%d)\n", + file->f_dentry->d_name.name, + i_size_read(inode) / 1024, + pos, + current->comm, current->pid); + } else { + unsigned long missed; + + missed = (pgcached - pgaccess) * PAGE_CACHE_SIZE; + if (bdi->ra_expect_bytes >= missed / 2) + bdi->ra_expect_bytes -= missed / 2; + + debug_inc(initial_ra_miss); + dprintk("initial_ra_miss on file %s " + "size %lluK cached %luK hit %luK " + "pos %lu by %s(%d)\n", + file->f_dentry->d_name.name, + i_size_read(inode) / 1024, + pgcached << (PAGE_CACHE_SHIFT - 10), + pgrahit << (PAGE_CACHE_SHIFT - 10), + pos, + current->comm, current->pid); + } +} + #endif /* CONFIG_ADAPTIVE_READAHEAD */ /* _