126b3fcdec
If CONFIG_VIRTUAL_MEM_MAP is enabled, ia64 defines macro VMALLOC_END as unsigned long variable vmalloc_end which is adjusted to prepare room for vmemmap. This becomes probnlematic if a local variables vmalloc_end is defined in some function (not very unlikely) and VMALLOC_END is used in the function - the function thinks its referencing the global VMALLOC_END value but would be referencing its own local vmalloc_end variable. There's no reason VMALLOC_END should be a macro. Just define it as an unsigned long variable if CONFIG_VIRTUAL_MEM_MAP is set to avoid nasty surprises. Signed-off-by: Tejun Heo <tj@kernel.org> Acked-by: Tony Luck <tony.luck@intel.com> Cc: Fenghua Yu <fenghua.yu@intel.com> Cc: linux-ia64 <linux-ia64@vger.kernel.org> Cc: Christoph Lameter <cl@linux-foundation.org>
76 lines
2.3 KiB
C
76 lines
2.3 KiB
C
#ifndef meminit_h
|
|
#define meminit_h
|
|
|
|
/*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*/
|
|
|
|
|
|
/*
|
|
* Entries defined so far:
|
|
* - boot param structure itself
|
|
* - memory map
|
|
* - initrd (optional)
|
|
* - command line string
|
|
* - kernel code & data
|
|
* - crash dumping code reserved region
|
|
* - Kernel memory map built from EFI memory map
|
|
* - ELF core header
|
|
* - xen start info if CONFIG_XEN
|
|
*
|
|
* More could be added if necessary
|
|
*/
|
|
#define IA64_MAX_RSVD_REGIONS 9
|
|
|
|
struct rsvd_region {
|
|
u64 start; /* virtual address of beginning of element */
|
|
u64 end; /* virtual address of end of element + 1 */
|
|
};
|
|
|
|
extern struct rsvd_region rsvd_region[IA64_MAX_RSVD_REGIONS + 1];
|
|
extern int num_rsvd_regions;
|
|
|
|
extern void find_memory (void);
|
|
extern void reserve_memory (void);
|
|
extern void find_initrd (void);
|
|
extern int filter_rsvd_memory (u64 start, u64 end, void *arg);
|
|
extern int filter_memory (u64 start, u64 end, void *arg);
|
|
extern unsigned long efi_memmap_init(u64 *s, u64 *e);
|
|
extern int find_max_min_low_pfn (u64, u64, void *);
|
|
|
|
extern unsigned long vmcore_find_descriptor_size(unsigned long address);
|
|
extern int reserve_elfcorehdr(u64 *start, u64 *end);
|
|
|
|
/*
|
|
* For rounding an address to the next IA64_GRANULE_SIZE or order
|
|
*/
|
|
#define GRANULEROUNDDOWN(n) ((n) & ~(IA64_GRANULE_SIZE-1))
|
|
#define GRANULEROUNDUP(n) (((n)+IA64_GRANULE_SIZE-1) & ~(IA64_GRANULE_SIZE-1))
|
|
|
|
#ifdef CONFIG_NUMA
|
|
extern void call_pernode_memory (unsigned long start, unsigned long len, void *func);
|
|
#else
|
|
# define call_pernode_memory(start, len, func) (*func)(start, len, 0)
|
|
#endif
|
|
|
|
#define IGNORE_PFN0 1 /* XXX fix me: ignore pfn 0 until TLB miss handler is updated... */
|
|
|
|
extern int register_active_ranges(u64 start, u64 len, int nid);
|
|
|
|
#ifdef CONFIG_VIRTUAL_MEM_MAP
|
|
# define LARGE_GAP 0x40000000 /* Use virtual mem map if hole is > than this */
|
|
extern unsigned long VMALLOC_END;
|
|
extern struct page *vmem_map;
|
|
extern int find_largest_hole(u64 start, u64 end, void *arg);
|
|
extern int create_mem_map_page_table(u64 start, u64 end, void *arg);
|
|
extern int vmemmap_find_next_valid_pfn(int, int);
|
|
#else
|
|
static inline int vmemmap_find_next_valid_pfn(int node, int i)
|
|
{
|
|
return i + 1;
|
|
}
|
|
#endif
|
|
#endif /* meminit_h */
|