| /* init.c: memory initialisation for FRV | 
 |  * | 
 |  * Copyright (C) 2004 Red Hat, Inc. All Rights Reserved. | 
 |  * Written by David Howells (dhowells@redhat.com) | 
 |  * | 
 |  * This program is free software; you can redistribute it and/or | 
 |  * modify it under the terms of the GNU General Public License | 
 |  * as published by the Free Software Foundation; either version | 
 |  * 2 of the License, or (at your option) any later version. | 
 |  * | 
 |  * Derived from: | 
 |  *  - linux/arch/m68knommu/mm/init.c | 
 |  *    - Copyright (C) 1998  D. Jeff Dionne <jeff@lineo.ca>, Kenneth Albanowski <kjahds@kjahds.com>, | 
 |  *    - Copyright (C) 2000  Lineo, Inc.  (www.lineo.com) | 
 |  *  - linux/arch/m68k/mm/init.c | 
 |  *    - Copyright (C) 1995  Hamish Macdonald | 
 |  */ | 
 |  | 
 | #include <linux/signal.h> | 
 | #include <linux/sched.h> | 
 | #include <linux/pagemap.h> | 
 | #include <linux/swap.h> | 
 | #include <linux/mm.h> | 
 | #include <linux/kernel.h> | 
 | #include <linux/string.h> | 
 | #include <linux/types.h> | 
 | #include <linux/bootmem.h> | 
 | #include <linux/highmem.h> | 
 |  | 
 | #include <asm/setup.h> | 
 | #include <asm/segment.h> | 
 | #include <asm/page.h> | 
 | #include <asm/pgtable.h> | 
 | #include <asm/system.h> | 
 | #include <asm/mmu_context.h> | 
 | #include <asm/virtconvert.h> | 
 | #include <asm/sections.h> | 
 | #include <asm/tlb.h> | 
 |  | 
 | #undef DEBUG | 
 |  | 
 | DEFINE_PER_CPU(struct mmu_gather, mmu_gathers); | 
 |  | 
 | /* | 
 |  * BAD_PAGE is the page that is used for page faults when linux | 
 |  * is out-of-memory. Older versions of linux just did a | 
 |  * do_exit(), but using this instead means there is less risk | 
 |  * for a process dying in kernel mode, possibly leaving a inode | 
 |  * unused etc.. | 
 |  * | 
 |  * BAD_PAGETABLE is the accompanying page-table: it is initialized | 
 |  * to point to BAD_PAGE entries. | 
 |  * | 
 |  * ZERO_PAGE is a special page that is used for zero-initialized | 
 |  * data and COW. | 
 |  */ | 
 | static unsigned long empty_bad_page_table; | 
 | static unsigned long empty_bad_page; | 
 | unsigned long empty_zero_page; | 
 |  | 
 | /*****************************************************************************/ | 
 | /* | 
 |  * | 
 |  */ | 
 | void show_mem(void) | 
 | { | 
 | 	unsigned long i; | 
 | 	int free = 0, total = 0, reserved = 0, shared = 0; | 
 |  | 
 | 	printk("\nMem-info:\n"); | 
 | 	show_free_areas(); | 
 | 	i = max_mapnr; | 
 | 	while (i-- > 0) { | 
 | 		struct page *page = &mem_map[i]; | 
 |  | 
 | 		total++; | 
 | 		if (PageReserved(page)) | 
 | 			reserved++; | 
 | 		else if (!page_count(page)) | 
 | 			free++; | 
 | 		else | 
 | 			shared += page_count(page) - 1; | 
 | 	} | 
 |  | 
 | 	printk("%d pages of RAM\n",total); | 
 | 	printk("%d free pages\n",free); | 
 | 	printk("%d reserved pages\n",reserved); | 
 | 	printk("%d pages shared\n",shared); | 
 |  | 
 | } /* end show_mem() */ | 
 |  | 
 | /*****************************************************************************/ | 
 | /* | 
 |  * paging_init() continues the virtual memory environment setup which | 
 |  * was begun by the code in arch/head.S. | 
 |  * The parameters are pointers to where to stick the starting and ending | 
 |  * addresses  of available kernel virtual memory. | 
 |  */ | 
 | void __init paging_init(void) | 
 | { | 
 | 	unsigned long zones_size[MAX_NR_ZONES] = {0, }; | 
 |  | 
 | 	/* allocate some pages for kernel housekeeping tasks */ | 
 | 	empty_bad_page_table	= (unsigned long) alloc_bootmem_pages(PAGE_SIZE); | 
 | 	empty_bad_page		= (unsigned long) alloc_bootmem_pages(PAGE_SIZE); | 
 | 	empty_zero_page		= (unsigned long) alloc_bootmem_pages(PAGE_SIZE); | 
 |  | 
 | 	memset((void *) empty_zero_page, 0, PAGE_SIZE); | 
 |  | 
 | #ifdef CONFIG_HIGHMEM | 
 | 	if (num_physpages - num_mappedpages) { | 
 | 		pgd_t *pge; | 
 | 		pud_t *pue; | 
 | 		pmd_t *pme; | 
 |  | 
 | 		pkmap_page_table = alloc_bootmem_pages(PAGE_SIZE); | 
 |  | 
 | 		memset(pkmap_page_table, 0, PAGE_SIZE); | 
 |  | 
 | 		pge = swapper_pg_dir + pgd_index_k(PKMAP_BASE); | 
 | 		pue = pud_offset(pge, PKMAP_BASE); | 
 | 		pme = pmd_offset(pue, PKMAP_BASE); | 
 | 		__set_pmd(pme, virt_to_phys(pkmap_page_table) | _PAGE_TABLE); | 
 | 	} | 
 | #endif | 
 |  | 
 | 	/* distribute the allocatable pages across the various zones and pass them to the allocator | 
 | 	 */ | 
 | 	zones_size[ZONE_DMA]     = max_low_pfn - min_low_pfn; | 
 | 	zones_size[ZONE_NORMAL]  = 0; | 
 | #ifdef CONFIG_HIGHMEM | 
 | 	zones_size[ZONE_HIGHMEM] = num_physpages - num_mappedpages; | 
 | #endif | 
 |  | 
 | 	free_area_init(zones_size); | 
 |  | 
 | #ifdef CONFIG_MMU | 
 | 	/* initialise init's MMU context */ | 
 | 	init_new_context(&init_task, &init_mm); | 
 | #endif | 
 |  | 
 | } /* end paging_init() */ | 
 |  | 
 | /*****************************************************************************/ | 
 | /* | 
 |  * | 
 |  */ | 
 | void __init mem_init(void) | 
 | { | 
 | 	unsigned long npages = (memory_end - memory_start) >> PAGE_SHIFT; | 
 | 	unsigned long tmp; | 
 | #ifdef CONFIG_MMU | 
 | 	unsigned long loop, pfn; | 
 | 	int datapages = 0; | 
 | #endif | 
 | 	int codek = 0, datak = 0; | 
 |  | 
 | 	/* this will put all memory onto the freelists */ | 
 | 	totalram_pages = free_all_bootmem(); | 
 |  | 
 | #ifdef CONFIG_MMU | 
 | 	for (loop = 0 ; loop < npages ; loop++) | 
 | 		if (PageReserved(&mem_map[loop])) | 
 | 			datapages++; | 
 |  | 
 | #ifdef CONFIG_HIGHMEM | 
 | 	for (pfn = num_physpages - 1; pfn >= num_mappedpages; pfn--) { | 
 | 		struct page *page = &mem_map[pfn]; | 
 |  | 
 | 		ClearPageReserved(page); | 
 | 		init_page_count(page); | 
 | 		__free_page(page); | 
 | 		totalram_pages++; | 
 | 	} | 
 | #endif | 
 |  | 
 | 	codek = ((unsigned long) &_etext - (unsigned long) &_stext) >> 10; | 
 | 	datak = datapages << (PAGE_SHIFT - 10); | 
 |  | 
 | #else | 
 | 	codek = (_etext - _stext) >> 10; | 
 | 	datak = 0; //(_ebss - _sdata) >> 10; | 
 | #endif | 
 |  | 
 | 	tmp = nr_free_pages() << PAGE_SHIFT; | 
 | 	printk("Memory available: %luKiB/%luKiB RAM, %luKiB/%luKiB ROM (%dKiB kernel code, %dKiB data)\n", | 
 | 	       tmp >> 10, | 
 | 	       npages << (PAGE_SHIFT - 10), | 
 | 	       (rom_length > 0) ? ((rom_length >> 10) - codek) : 0, | 
 | 	       rom_length >> 10, | 
 | 	       codek, | 
 | 	       datak | 
 | 	       ); | 
 |  | 
 | } /* end mem_init() */ | 
 |  | 
 | /*****************************************************************************/ | 
 | /* | 
 |  * free the memory that was only required for initialisation | 
 |  */ | 
 | void __init free_initmem(void) | 
 | { | 
 | #if defined(CONFIG_RAMKERNEL) && !defined(CONFIG_PROTECT_KERNEL) | 
 | 	unsigned long start, end, addr; | 
 |  | 
 | 	start = PAGE_ALIGN((unsigned long) &__init_begin);	/* round up */ | 
 | 	end   = ((unsigned long) &__init_end) & PAGE_MASK;	/* round down */ | 
 |  | 
 | 	/* next to check that the page we free is not a partial page */ | 
 | 	for (addr = start; addr < end; addr += PAGE_SIZE) { | 
 | 		ClearPageReserved(virt_to_page(addr)); | 
 | 		init_page_count(virt_to_page(addr)); | 
 | 		free_page(addr); | 
 | 		totalram_pages++; | 
 | 	} | 
 |  | 
 | 	printk("Freeing unused kernel memory: %ldKiB freed (0x%lx - 0x%lx)\n", | 
 | 	       (end - start) >> 10, start, end); | 
 | #endif | 
 | } /* end free_initmem() */ | 
 |  | 
 | /*****************************************************************************/ | 
 | /* | 
 |  * free the initial ramdisk memory | 
 |  */ | 
 | #ifdef CONFIG_BLK_DEV_INITRD | 
 | void __init free_initrd_mem(unsigned long start, unsigned long end) | 
 | { | 
 | 	int pages = 0; | 
 | 	for (; start < end; start += PAGE_SIZE) { | 
 | 		ClearPageReserved(virt_to_page(start)); | 
 | 		init_page_count(virt_to_page(start)); | 
 | 		free_page(start); | 
 | 		totalram_pages++; | 
 | 		pages++; | 
 | 	} | 
 | 	printk("Freeing initrd memory: %dKiB freed\n", (pages * PAGE_SIZE) >> 10); | 
 | } /* end free_initrd_mem() */ | 
 | #endif |