linux/arch/frv/include/asm/mem-layout.h
FUJITA Tomonori a6eb9fe105 dma-mapping: rename ARCH_KMALLOC_MINALIGN to ARCH_DMA_MINALIGN
Now each architecture has the own dma_get_cache_alignment implementation.

dma_get_cache_alignment returns the minimum DMA alignment.  Architectures
define it as ARCH_KMALLOC_MINALIGN (it's used to make sure that malloc'ed
buffer is DMA-safe; the buffer doesn't share a cache with the others).  So
we can unify dma_get_cache_alignment implementations.

This patch:

dma_get_cache_alignment() needs to know if an architecture defines
ARCH_KMALLOC_MINALIGN or not (needs to know if architecture has DMA
alignment restriction).  However, slab.h define ARCH_KMALLOC_MINALIGN if
architectures doesn't define it.

Let's rename ARCH_KMALLOC_MINALIGN to ARCH_DMA_MINALIGN.
ARCH_KMALLOC_MINALIGN is used only in the internals of slab/slob/slub
(except for crypto).

Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp>
Cc: <linux-arch@vger.kernel.org>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2010-08-11 08:59:21 -07:00

86 lines
2.2 KiB
C

/* mem-layout.h: memory layout
*
* Copyright (C) 2004 Red Hat, Inc. All Rights Reserved.
* Written by David Howells (dhowells@redhat.com)
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* as published by the Free Software Foundation; either version
* 2 of the License, or (at your option) any later version.
*/
#ifndef _ASM_MEM_LAYOUT_H
#define _ASM_MEM_LAYOUT_H
#ifndef __ASSEMBLY__
#define __UL(X) ((unsigned long) (X))
#else
#define __UL(X) (X)
#endif
/*
* PAGE_SHIFT determines the page size
*/
#define PAGE_SHIFT 14
#ifndef __ASSEMBLY__
#define PAGE_SIZE (1UL << PAGE_SHIFT)
#else
#define PAGE_SIZE (1 << PAGE_SHIFT)
#endif
#define PAGE_MASK (~(PAGE_SIZE-1))
/*
* the slab must be aligned such that load- and store-double instructions don't
* fault if used
*/
#define ARCH_DMA_MINALIGN L1_CACHE_BYTES
#define ARCH_SLAB_MINALIGN L1_CACHE_BYTES
/*****************************************************************************/
/*
* virtual memory layout from kernel's point of view
*/
#define PAGE_OFFSET ((unsigned long) &__page_offset)
#ifdef CONFIG_MMU
/* see Documentation/frv/mmu-layout.txt */
#define KERNEL_LOWMEM_START __UL(0xc0000000)
#define KERNEL_LOWMEM_END __UL(0xd0000000)
#define VMALLOC_START __UL(0xd0000000)
#define VMALLOC_END __UL(0xd8000000)
#define PKMAP_BASE __UL(0xd8000000)
#define PKMAP_END __UL(0xdc000000)
#define KMAP_ATOMIC_SECONDARY_FRAME __UL(0xdc000000)
#define KMAP_ATOMIC_PRIMARY_FRAME __UL(0xdd000000)
#endif
#define KERNEL_IO_START __UL(0xe0000000)
/*****************************************************************************/
/*
* memory layout from userspace's point of view
*/
#define BRK_BASE __UL(2 * 1024 * 1024 + PAGE_SIZE)
#define STACK_TOP __UL(2 * 1024 * 1024)
#define STACK_TOP_MAX __UL(0xc0000000)
/* userspace process size */
#ifdef CONFIG_MMU
#define TASK_SIZE (PAGE_OFFSET)
#else
#define TASK_SIZE __UL(0xFFFFFFFFUL)
#endif
/* base of area at which unspecified mmaps will start */
#ifdef CONFIG_BINFMT_ELF_FDPIC
#define TASK_UNMAPPED_BASE __UL(16 * 1024 * 1024)
#else
#define TASK_UNMAPPED_BASE __UL(TASK_SIZE / 3)
#endif
#endif /* _ASM_MEM_LAYOUT_H */