mm/page_alloc.c: fix coding style and spelling

Fix all errors reported by checkpatch and some small spelling mistakes.

Signed-off-by: Pintu Kumar <pintu.k@samsung.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
Pintu Kumar
2013-09-11 14:20:34 -07:00
committed by Linus Torvalds
parent ebc2a1a691
commit b8af29418a

View File

@@ -721,7 +721,8 @@ static bool free_pages_prepare(struct page *page, unsigned int order)
return false; return false;
if (!PageHighMem(page)) { if (!PageHighMem(page)) {
debug_check_no_locks_freed(page_address(page),PAGE_SIZE<<order); debug_check_no_locks_freed(page_address(page),
PAGE_SIZE << order);
debug_check_no_obj_freed(page_address(page), debug_check_no_obj_freed(page_address(page),
PAGE_SIZE << order); PAGE_SIZE << order);
} }
@@ -5047,7 +5048,7 @@ restart:
/* /*
* Some kernelcore has been met, update counts and * Some kernelcore has been met, update counts and
* break if the kernelcore for this node has been * break if the kernelcore for this node has been
* satisified * satisfied
*/ */
required_kernelcore -= min(required_kernelcore, required_kernelcore -= min(required_kernelcore,
size_pages); size_pages);
@@ -5061,7 +5062,7 @@ restart:
* If there is still required_kernelcore, we do another pass with one * If there is still required_kernelcore, we do another pass with one
* less node in the count. This will push zone_movable_pfn[nid] further * less node in the count. This will push zone_movable_pfn[nid] further
* along on the nodes that still have memory until kernelcore is * along on the nodes that still have memory until kernelcore is
* satisified * satisfied
*/ */
usable_nodes--; usable_nodes--;
if (usable_nodes && required_kernelcore > usable_nodes) if (usable_nodes && required_kernelcore > usable_nodes)
@@ -5286,8 +5287,10 @@ void __init mem_init_print_info(const char *str)
* 3) .rodata.* may be embedded into .text or .data sections. * 3) .rodata.* may be embedded into .text or .data sections.
*/ */
#define adj_init_size(start, end, size, pos, adj) \ #define adj_init_size(start, end, size, pos, adj) \
do { \
if (start <= pos && pos < end && size > adj) \ if (start <= pos && pos < end && size > adj) \
size -= adj; size -= adj; \
} while (0)
adj_init_size(__init_begin, __init_end, init_data_size, adj_init_size(__init_begin, __init_end, init_data_size,
_sinittext, init_code_size); _sinittext, init_code_size);
@@ -5682,8 +5685,8 @@ int lowmem_reserve_ratio_sysctl_handler(ctl_table *table, int write,
/* /*
* percpu_pagelist_fraction - changes the pcp->high for each zone on each * percpu_pagelist_fraction - changes the pcp->high for each zone on each
* cpu. It is the fraction of total pages in each zone that a hot per cpu pagelist * cpu. It is the fraction of total pages in each zone that a hot per cpu
* can have before it gets flushed back to buddy allocator. * pagelist can have before it gets flushed back to buddy allocator.
*/ */
int percpu_pagelist_fraction_sysctl_handler(ctl_table *table, int write, int percpu_pagelist_fraction_sysctl_handler(ctl_table *table, int write,
void __user *buffer, size_t *length, loff_t *ppos) void __user *buffer, size_t *length, loff_t *ppos)
@@ -5901,7 +5904,7 @@ void set_pageblock_flags_group(struct page *page, unsigned long flags,
* This function checks whether pageblock includes unmovable pages or not. * This function checks whether pageblock includes unmovable pages or not.
* If @count is not zero, it is okay to include less @count unmovable pages * If @count is not zero, it is okay to include less @count unmovable pages
* *
* PageLRU check wihtout isolation or lru_lock could race so that * PageLRU check without isolation or lru_lock could race so that
* MIGRATE_MOVABLE block might include unmovable pages. It means you can't * MIGRATE_MOVABLE block might include unmovable pages. It means you can't
* expect this function should be exact. * expect this function should be exact.
*/ */