mirror of
https://github.com/FEX-Emu/linux.git
synced 2024-12-01 03:22:08 +00:00
[PATCH] gfp_t: mm/* (easy parts)
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk> Signed-off-by: Linus Torvalds <torvalds@osdl.org>
This commit is contained in:
parent
af4ca457ea
commit
6daa0e2862
@ -747,7 +747,7 @@ extern unsigned long do_mremap(unsigned long addr,
|
||||
* The callback will be passed nr_to_scan == 0 when the VM is querying the
|
||||
* cache size, so a fastpath for that case is appropriate.
|
||||
*/
|
||||
typedef int (*shrinker_t)(int nr_to_scan, unsigned int gfp_mask);
|
||||
typedef int (*shrinker_t)(int nr_to_scan, gfp_t gfp_mask);
|
||||
|
||||
/*
|
||||
* Add an aging callback. The int is the number of 'seeks' it takes
|
||||
|
@ -69,7 +69,7 @@ extern struct page * find_lock_page(struct address_space *mapping,
|
||||
extern struct page * find_trylock_page(struct address_space *mapping,
|
||||
unsigned long index);
|
||||
extern struct page * find_or_create_page(struct address_space *mapping,
|
||||
unsigned long index, unsigned int gfp_mask);
|
||||
unsigned long index, gfp_t gfp_mask);
|
||||
unsigned find_get_pages(struct address_space *mapping, pgoff_t start,
|
||||
unsigned int nr_pages, struct page **pages);
|
||||
unsigned find_get_pages_tag(struct address_space *mapping, pgoff_t *index,
|
||||
@ -92,9 +92,9 @@ extern int read_cache_pages(struct address_space *mapping,
|
||||
struct list_head *pages, filler_t *filler, void *data);
|
||||
|
||||
int add_to_page_cache(struct page *page, struct address_space *mapping,
|
||||
unsigned long index, int gfp_mask);
|
||||
unsigned long index, gfp_t gfp_mask);
|
||||
int add_to_page_cache_lru(struct page *page, struct address_space *mapping,
|
||||
unsigned long index, int gfp_mask);
|
||||
unsigned long index, gfp_t gfp_mask);
|
||||
extern void remove_from_page_cache(struct page *page);
|
||||
extern void __remove_from_page_cache(struct page *page);
|
||||
|
||||
|
@ -121,7 +121,7 @@ extern unsigned int ksize(const void *);
|
||||
extern void *kmem_cache_alloc_node(kmem_cache_t *, gfp_t flags, int node);
|
||||
extern void *kmalloc_node(size_t size, gfp_t flags, int node);
|
||||
#else
|
||||
static inline void *kmem_cache_alloc_node(kmem_cache_t *cachep, int flags, int node)
|
||||
static inline void *kmem_cache_alloc_node(kmem_cache_t *cachep, gfp_t flags, int node)
|
||||
{
|
||||
return kmem_cache_alloc(cachep, flags);
|
||||
}
|
||||
|
@ -171,8 +171,8 @@ extern int rotate_reclaimable_page(struct page *page);
|
||||
extern void swap_setup(void);
|
||||
|
||||
/* linux/mm/vmscan.c */
|
||||
extern int try_to_free_pages(struct zone **, unsigned int);
|
||||
extern int zone_reclaim(struct zone *, unsigned int, unsigned int);
|
||||
extern int try_to_free_pages(struct zone **, gfp_t);
|
||||
extern int zone_reclaim(struct zone *, gfp_t, unsigned int);
|
||||
extern int shrink_all_memory(int);
|
||||
extern int vm_swappiness;
|
||||
|
||||
|
@ -377,7 +377,7 @@ int filemap_write_and_wait_range(struct address_space *mapping,
|
||||
* This function does not add the page to the LRU. The caller must do that.
|
||||
*/
|
||||
int add_to_page_cache(struct page *page, struct address_space *mapping,
|
||||
pgoff_t offset, int gfp_mask)
|
||||
pgoff_t offset, gfp_t gfp_mask)
|
||||
{
|
||||
int error = radix_tree_preload(gfp_mask & ~__GFP_HIGHMEM);
|
||||
|
||||
@ -401,7 +401,7 @@ int add_to_page_cache(struct page *page, struct address_space *mapping,
|
||||
EXPORT_SYMBOL(add_to_page_cache);
|
||||
|
||||
int add_to_page_cache_lru(struct page *page, struct address_space *mapping,
|
||||
pgoff_t offset, int gfp_mask)
|
||||
pgoff_t offset, gfp_t gfp_mask)
|
||||
{
|
||||
int ret = add_to_page_cache(page, mapping, offset, gfp_mask);
|
||||
if (ret == 0)
|
||||
@ -591,7 +591,7 @@ EXPORT_SYMBOL(find_lock_page);
|
||||
* memory exhaustion.
|
||||
*/
|
||||
struct page *find_or_create_page(struct address_space *mapping,
|
||||
unsigned long index, unsigned int gfp_mask)
|
||||
unsigned long index, gfp_t gfp_mask)
|
||||
{
|
||||
struct page *page, *cached_page = NULL;
|
||||
int err;
|
||||
@ -683,7 +683,7 @@ struct page *
|
||||
grab_cache_page_nowait(struct address_space *mapping, unsigned long index)
|
||||
{
|
||||
struct page *page = find_get_page(mapping, index);
|
||||
unsigned int gfp_mask;
|
||||
gfp_t gfp_mask;
|
||||
|
||||
if (page) {
|
||||
if (!TestSetPageLocked(page))
|
||||
|
@ -205,7 +205,7 @@ void * mempool_alloc(mempool_t *pool, gfp_t gfp_mask)
|
||||
void *element;
|
||||
unsigned long flags;
|
||||
wait_queue_t wait;
|
||||
unsigned int gfp_temp;
|
||||
gfp_t gfp_temp;
|
||||
|
||||
might_sleep_if(gfp_mask & __GFP_WAIT);
|
||||
|
||||
|
@ -85,7 +85,7 @@ enum sgp_type {
|
||||
static int shmem_getpage(struct inode *inode, unsigned long idx,
|
||||
struct page **pagep, enum sgp_type sgp, int *type);
|
||||
|
||||
static inline struct page *shmem_dir_alloc(unsigned int gfp_mask)
|
||||
static inline struct page *shmem_dir_alloc(gfp_t gfp_mask)
|
||||
{
|
||||
/*
|
||||
* The above definition of ENTRIES_PER_PAGE, and the use of
|
||||
@ -898,7 +898,7 @@ struct page *shmem_swapin(struct shmem_inode_info *info, swp_entry_t entry,
|
||||
}
|
||||
|
||||
static struct page *
|
||||
shmem_alloc_page(unsigned long gfp, struct shmem_inode_info *info,
|
||||
shmem_alloc_page(gfp_t gfp, struct shmem_inode_info *info,
|
||||
unsigned long idx)
|
||||
{
|
||||
struct vm_area_struct pvma;
|
||||
|
@ -386,7 +386,7 @@ struct kmem_cache_s {
|
||||
unsigned int gfporder;
|
||||
|
||||
/* force GFP flags, e.g. GFP_DMA */
|
||||
unsigned int gfpflags;
|
||||
gfp_t gfpflags;
|
||||
|
||||
size_t colour; /* cache colouring range */
|
||||
unsigned int colour_off; /* colour offset */
|
||||
@ -2117,7 +2117,7 @@ static void cache_init_objs(kmem_cache_t *cachep,
|
||||
slabp->free = 0;
|
||||
}
|
||||
|
||||
static void kmem_flagcheck(kmem_cache_t *cachep, unsigned int flags)
|
||||
static void kmem_flagcheck(kmem_cache_t *cachep, gfp_t flags)
|
||||
{
|
||||
if (flags & SLAB_DMA) {
|
||||
if (!(cachep->gfpflags & GFP_DMA))
|
||||
@ -2152,7 +2152,7 @@ static int cache_grow(kmem_cache_t *cachep, gfp_t flags, int nodeid)
|
||||
struct slab *slabp;
|
||||
void *objp;
|
||||
size_t offset;
|
||||
unsigned int local_flags;
|
||||
gfp_t local_flags;
|
||||
unsigned long ctor_flags;
|
||||
struct kmem_list3 *l3;
|
||||
|
||||
@ -2546,7 +2546,7 @@ static inline void *__cache_alloc(kmem_cache_t *cachep, gfp_t flags)
|
||||
/*
|
||||
* A interface to enable slab creation on nodeid
|
||||
*/
|
||||
static void *__cache_alloc_node(kmem_cache_t *cachep, int flags, int nodeid)
|
||||
static void *__cache_alloc_node(kmem_cache_t *cachep, gfp_t flags, int nodeid)
|
||||
{
|
||||
struct list_head *entry;
|
||||
struct slab *slabp;
|
||||
|
@ -70,7 +70,7 @@ struct scan_control {
|
||||
unsigned int priority;
|
||||
|
||||
/* This context's GFP mask */
|
||||
unsigned int gfp_mask;
|
||||
gfp_t gfp_mask;
|
||||
|
||||
int may_writepage;
|
||||
|
||||
@ -186,7 +186,7 @@ EXPORT_SYMBOL(remove_shrinker);
|
||||
*
|
||||
* Returns the number of slab objects which we shrunk.
|
||||
*/
|
||||
static int shrink_slab(unsigned long scanned, unsigned int gfp_mask,
|
||||
static int shrink_slab(unsigned long scanned, gfp_t gfp_mask,
|
||||
unsigned long lru_pages)
|
||||
{
|
||||
struct shrinker *shrinker;
|
||||
@ -926,7 +926,7 @@ shrink_caches(struct zone **zones, struct scan_control *sc)
|
||||
* holds filesystem locks which prevent writeout this might not work, and the
|
||||
* allocation attempt will fail.
|
||||
*/
|
||||
int try_to_free_pages(struct zone **zones, unsigned int gfp_mask)
|
||||
int try_to_free_pages(struct zone **zones, gfp_t gfp_mask)
|
||||
{
|
||||
int priority;
|
||||
int ret = 0;
|
||||
@ -1338,7 +1338,7 @@ module_init(kswapd_init)
|
||||
/*
|
||||
* Try to free up some pages from this zone through reclaim.
|
||||
*/
|
||||
int zone_reclaim(struct zone *zone, unsigned int gfp_mask, unsigned int order)
|
||||
int zone_reclaim(struct zone *zone, gfp_t gfp_mask, unsigned int order)
|
||||
{
|
||||
struct scan_control sc;
|
||||
int nr_pages = 1 << order;
|
||||
|
Loading…
Reference in New Issue
Block a user