提交 9c93af1e 编写于 作者: D David Howells

NOMMU: Rename ARM's struct vm_region

Rename ARM's struct vm_region so that I can introduce my own global version
for NOMMU.  It's feasible that the ARM version may wish to use my global one
instead.

The NOMMU vm_region struct defines areas of the physical memory map that are
under mmap.  This may include chunks of RAM or regions of memory mapped
devices, such as flash.  It is also used to retain copies of file content so
that shareable private memory mappings of files can be made.  As such, it may
be compatible with what is described in the banner comment for ARM's vm_region
struct.
Signed-off-by: NDavid Howells <dhowells@redhat.com>
上级 0e8f989a
...@@ -71,7 +71,7 @@ static DEFINE_SPINLOCK(consistent_lock); ...@@ -71,7 +71,7 @@ static DEFINE_SPINLOCK(consistent_lock);
* the amount of RAM found at boot time.) I would imagine that get_vm_area() * the amount of RAM found at boot time.) I would imagine that get_vm_area()
* would have to initialise this each time prior to calling vm_region_alloc(). * would have to initialise this each time prior to calling vm_region_alloc().
*/ */
struct vm_region { struct arm_vm_region {
struct list_head vm_list; struct list_head vm_list;
unsigned long vm_start; unsigned long vm_start;
unsigned long vm_end; unsigned long vm_end;
...@@ -79,20 +79,20 @@ struct vm_region { ...@@ -79,20 +79,20 @@ struct vm_region {
int vm_active; int vm_active;
}; };
static struct vm_region consistent_head = { static struct arm_vm_region consistent_head = {
.vm_list = LIST_HEAD_INIT(consistent_head.vm_list), .vm_list = LIST_HEAD_INIT(consistent_head.vm_list),
.vm_start = CONSISTENT_BASE, .vm_start = CONSISTENT_BASE,
.vm_end = CONSISTENT_END, .vm_end = CONSISTENT_END,
}; };
static struct vm_region * static struct arm_vm_region *
vm_region_alloc(struct vm_region *head, size_t size, gfp_t gfp) arm_vm_region_alloc(struct arm_vm_region *head, size_t size, gfp_t gfp)
{ {
unsigned long addr = head->vm_start, end = head->vm_end - size; unsigned long addr = head->vm_start, end = head->vm_end - size;
unsigned long flags; unsigned long flags;
struct vm_region *c, *new; struct arm_vm_region *c, *new;
new = kmalloc(sizeof(struct vm_region), gfp); new = kmalloc(sizeof(struct arm_vm_region), gfp);
if (!new) if (!new)
goto out; goto out;
...@@ -127,9 +127,9 @@ vm_region_alloc(struct vm_region *head, size_t size, gfp_t gfp) ...@@ -127,9 +127,9 @@ vm_region_alloc(struct vm_region *head, size_t size, gfp_t gfp)
return NULL; return NULL;
} }
static struct vm_region *vm_region_find(struct vm_region *head, unsigned long addr) static struct arm_vm_region *arm_vm_region_find(struct arm_vm_region *head, unsigned long addr)
{ {
struct vm_region *c; struct arm_vm_region *c;
list_for_each_entry(c, &head->vm_list, vm_list) { list_for_each_entry(c, &head->vm_list, vm_list) {
if (c->vm_active && c->vm_start == addr) if (c->vm_active && c->vm_start == addr)
...@@ -149,7 +149,7 @@ __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp, ...@@ -149,7 +149,7 @@ __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp,
pgprot_t prot) pgprot_t prot)
{ {
struct page *page; struct page *page;
struct vm_region *c; struct arm_vm_region *c;
unsigned long order; unsigned long order;
u64 mask = ISA_DMA_THRESHOLD, limit; u64 mask = ISA_DMA_THRESHOLD, limit;
...@@ -214,7 +214,7 @@ __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp, ...@@ -214,7 +214,7 @@ __dma_alloc(struct device *dev, size_t size, dma_addr_t *handle, gfp_t gfp,
/* /*
* Allocate a virtual address in the consistent mapping region. * Allocate a virtual address in the consistent mapping region.
*/ */
c = vm_region_alloc(&consistent_head, size, c = arm_vm_region_alloc(&consistent_head, size,
gfp & ~(__GFP_DMA | __GFP_HIGHMEM)); gfp & ~(__GFP_DMA | __GFP_HIGHMEM));
if (c) { if (c) {
pte_t *pte; pte_t *pte;
...@@ -311,13 +311,13 @@ static int dma_mmap(struct device *dev, struct vm_area_struct *vma, ...@@ -311,13 +311,13 @@ static int dma_mmap(struct device *dev, struct vm_area_struct *vma,
void *cpu_addr, dma_addr_t dma_addr, size_t size) void *cpu_addr, dma_addr_t dma_addr, size_t size)
{ {
unsigned long flags, user_size, kern_size; unsigned long flags, user_size, kern_size;
struct vm_region *c; struct arm_vm_region *c;
int ret = -ENXIO; int ret = -ENXIO;
user_size = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT; user_size = (vma->vm_end - vma->vm_start) >> PAGE_SHIFT;
spin_lock_irqsave(&consistent_lock, flags); spin_lock_irqsave(&consistent_lock, flags);
c = vm_region_find(&consistent_head, (unsigned long)cpu_addr); c = arm_vm_region_find(&consistent_head, (unsigned long)cpu_addr);
spin_unlock_irqrestore(&consistent_lock, flags); spin_unlock_irqrestore(&consistent_lock, flags);
if (c) { if (c) {
...@@ -359,7 +359,7 @@ EXPORT_SYMBOL(dma_mmap_writecombine); ...@@ -359,7 +359,7 @@ EXPORT_SYMBOL(dma_mmap_writecombine);
*/ */
void dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr_t handle) void dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr_t handle)
{ {
struct vm_region *c; struct arm_vm_region *c;
unsigned long flags, addr; unsigned long flags, addr;
pte_t *ptep; pte_t *ptep;
int idx; int idx;
...@@ -378,7 +378,7 @@ void dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr ...@@ -378,7 +378,7 @@ void dma_free_coherent(struct device *dev, size_t size, void *cpu_addr, dma_addr
size = PAGE_ALIGN(size); size = PAGE_ALIGN(size);
spin_lock_irqsave(&consistent_lock, flags); spin_lock_irqsave(&consistent_lock, flags);
c = vm_region_find(&consistent_head, (unsigned long)cpu_addr); c = arm_vm_region_find(&consistent_head, (unsigned long)cpu_addr);
if (!c) if (!c)
goto no_area; goto no_area;
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册