提交 06d38148 编写于 作者: L Linus Torvalds

Merge branch 'stable/vmalloc-3.2' of git://git.kernel.org/pub/scm/linux/kernel/git/konrad/xen

* 'stable/vmalloc-3.2' of git://git.kernel.org/pub/scm/linux/kernel/git/konrad/xen:
  net: xen-netback: use API provided by xenbus module to map rings
  block: xen-blkback: use API provided by xenbus module to map rings
  xen: use generic functions instead of xen_{alloc, free}_vm_area()
/******************************************************************************
* arch/ia64/include/asm/xen/grant_table.h
*
* Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp>
* VA Linux Systems Japan K.K.
*
* This program is free software; you can redistribute it and/or modify
* it under the terms of the GNU General Public License as published by
* the Free Software Foundation; either version 2 of the License, or
* (at your option) any later version.
*
* This program is distributed in the hope that it will be useful,
* but WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
* GNU General Public License for more details.
*
* You should have received a copy of the GNU General Public License
* along with this program; if not, write to the Free Software
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
*
*/
#ifndef _ASM_IA64_XEN_GRANT_TABLE_H
#define _ASM_IA64_XEN_GRANT_TABLE_H
struct vm_struct *xen_alloc_vm_area(unsigned long size);
void xen_free_vm_area(struct vm_struct *area);
#endif /* _ASM_IA64_XEN_GRANT_TABLE_H */
...@@ -31,68 +31,6 @@ ...@@ -31,68 +31,6 @@
#include <asm/xen/hypervisor.h> #include <asm/xen/hypervisor.h>
struct vm_struct *xen_alloc_vm_area(unsigned long size)
{
int order;
unsigned long virt;
unsigned long nr_pages;
struct vm_struct *area;
order = get_order(size);
virt = __get_free_pages(GFP_KERNEL, order);
if (virt == 0)
goto err0;
nr_pages = 1 << order;
scrub_pages(virt, nr_pages);
area = kmalloc(sizeof(*area), GFP_KERNEL);
if (area == NULL)
goto err1;
area->flags = VM_IOREMAP;
area->addr = (void *)virt;
area->size = size;
area->pages = NULL;
area->nr_pages = nr_pages;
area->phys_addr = 0; /* xenbus_map_ring_valloc uses this field! */
return area;
err1:
free_pages(virt, order);
err0:
return NULL;
}
EXPORT_SYMBOL_GPL(xen_alloc_vm_area);
void xen_free_vm_area(struct vm_struct *area)
{
unsigned int order = get_order(area->size);
unsigned long i;
unsigned long phys_addr = __pa(area->addr);
/* This area is used for foreign page mappping.
* So underlying machine page may not be assigned. */
for (i = 0; i < (1 << order); i++) {
unsigned long ret;
unsigned long gpfn = (phys_addr >> PAGE_SHIFT) + i;
struct xen_memory_reservation reservation = {
.nr_extents = 1,
.address_bits = 0,
.extent_order = 0,
.domid = DOMID_SELF
};
set_xen_guest_handle(reservation.extent_start, &gpfn);
ret = HYPERVISOR_memory_op(XENMEM_populate_physmap,
&reservation);
BUG_ON(ret != 1);
}
free_pages((unsigned long)area->addr, order);
kfree(area);
}
EXPORT_SYMBOL_GPL(xen_free_vm_area);
/**************************************************************************** /****************************************************************************
* grant table hack * grant table hack
* cmd: GNTTABOP_xxx * cmd: GNTTABOP_xxx
......
#ifndef _ASM_X86_XEN_GRANT_TABLE_H
#define _ASM_X86_XEN_GRANT_TABLE_H
#define xen_alloc_vm_area(size) alloc_vm_area(size)
#define xen_free_vm_area(area) free_vm_area(area)
#endif /* _ASM_X86_XEN_GRANT_TABLE_H */
...@@ -71,7 +71,7 @@ int arch_gnttab_map_shared(unsigned long *frames, unsigned long nr_gframes, ...@@ -71,7 +71,7 @@ int arch_gnttab_map_shared(unsigned long *frames, unsigned long nr_gframes,
if (shared == NULL) { if (shared == NULL) {
struct vm_struct *area = struct vm_struct *area =
xen_alloc_vm_area(PAGE_SIZE * max_nr_gframes); alloc_vm_area(PAGE_SIZE * max_nr_gframes);
BUG_ON(area == NULL); BUG_ON(area == NULL);
shared = area->addr; shared = area->addr;
*__shared = shared; *__shared = shared;
......
...@@ -170,7 +170,7 @@ struct xen_blkif { ...@@ -170,7 +170,7 @@ struct xen_blkif {
enum blkif_protocol blk_protocol; enum blkif_protocol blk_protocol;
enum blkif_backend_type blk_backend_type; enum blkif_backend_type blk_backend_type;
union blkif_back_rings blk_rings; union blkif_back_rings blk_rings;
struct vm_struct *blk_ring_area; void *blk_ring;
/* The VBD attached to this interface. */ /* The VBD attached to this interface. */
struct xen_vbd vbd; struct xen_vbd vbd;
/* Back pointer to the backend_info. */ /* Back pointer to the backend_info. */
...@@ -198,9 +198,6 @@ struct xen_blkif { ...@@ -198,9 +198,6 @@ struct xen_blkif {
int st_wr_sect; int st_wr_sect;
wait_queue_head_t waiting_to_free; wait_queue_head_t waiting_to_free;
grant_handle_t shmem_handle;
grant_ref_t shmem_ref;
}; };
......
...@@ -122,38 +122,6 @@ static struct xen_blkif *xen_blkif_alloc(domid_t domid) ...@@ -122,38 +122,6 @@ static struct xen_blkif *xen_blkif_alloc(domid_t domid)
return blkif; return blkif;
} }
static int map_frontend_page(struct xen_blkif *blkif, unsigned long shared_page)
{
struct gnttab_map_grant_ref op;
gnttab_set_map_op(&op, (unsigned long)blkif->blk_ring_area->addr,
GNTMAP_host_map, shared_page, blkif->domid);
if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
BUG();
if (op.status) {
DPRINTK("Grant table operation failure !\n");
return op.status;
}
blkif->shmem_ref = shared_page;
blkif->shmem_handle = op.handle;
return 0;
}
static void unmap_frontend_page(struct xen_blkif *blkif)
{
struct gnttab_unmap_grant_ref op;
gnttab_set_unmap_op(&op, (unsigned long)blkif->blk_ring_area->addr,
GNTMAP_host_map, blkif->shmem_handle);
if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
BUG();
}
static int xen_blkif_map(struct xen_blkif *blkif, unsigned long shared_page, static int xen_blkif_map(struct xen_blkif *blkif, unsigned long shared_page,
unsigned int evtchn) unsigned int evtchn)
{ {
...@@ -163,35 +131,29 @@ static int xen_blkif_map(struct xen_blkif *blkif, unsigned long shared_page, ...@@ -163,35 +131,29 @@ static int xen_blkif_map(struct xen_blkif *blkif, unsigned long shared_page,
if (blkif->irq) if (blkif->irq)
return 0; return 0;
blkif->blk_ring_area = alloc_vm_area(PAGE_SIZE); err = xenbus_map_ring_valloc(blkif->be->dev, shared_page, &blkif->blk_ring);
if (!blkif->blk_ring_area) if (err < 0)
return -ENOMEM;
err = map_frontend_page(blkif, shared_page);
if (err) {
free_vm_area(blkif->blk_ring_area);
return err; return err;
}
switch (blkif->blk_protocol) { switch (blkif->blk_protocol) {
case BLKIF_PROTOCOL_NATIVE: case BLKIF_PROTOCOL_NATIVE:
{ {
struct blkif_sring *sring; struct blkif_sring *sring;
sring = (struct blkif_sring *)blkif->blk_ring_area->addr; sring = (struct blkif_sring *)blkif->blk_ring;
BACK_RING_INIT(&blkif->blk_rings.native, sring, PAGE_SIZE); BACK_RING_INIT(&blkif->blk_rings.native, sring, PAGE_SIZE);
break; break;
} }
case BLKIF_PROTOCOL_X86_32: case BLKIF_PROTOCOL_X86_32:
{ {
struct blkif_x86_32_sring *sring_x86_32; struct blkif_x86_32_sring *sring_x86_32;
sring_x86_32 = (struct blkif_x86_32_sring *)blkif->blk_ring_area->addr; sring_x86_32 = (struct blkif_x86_32_sring *)blkif->blk_ring;
BACK_RING_INIT(&blkif->blk_rings.x86_32, sring_x86_32, PAGE_SIZE); BACK_RING_INIT(&blkif->blk_rings.x86_32, sring_x86_32, PAGE_SIZE);
break; break;
} }
case BLKIF_PROTOCOL_X86_64: case BLKIF_PROTOCOL_X86_64:
{ {
struct blkif_x86_64_sring *sring_x86_64; struct blkif_x86_64_sring *sring_x86_64;
sring_x86_64 = (struct blkif_x86_64_sring *)blkif->blk_ring_area->addr; sring_x86_64 = (struct blkif_x86_64_sring *)blkif->blk_ring;
BACK_RING_INIT(&blkif->blk_rings.x86_64, sring_x86_64, PAGE_SIZE); BACK_RING_INIT(&blkif->blk_rings.x86_64, sring_x86_64, PAGE_SIZE);
break; break;
} }
...@@ -203,8 +165,7 @@ static int xen_blkif_map(struct xen_blkif *blkif, unsigned long shared_page, ...@@ -203,8 +165,7 @@ static int xen_blkif_map(struct xen_blkif *blkif, unsigned long shared_page,
xen_blkif_be_int, 0, xen_blkif_be_int, 0,
"blkif-backend", blkif); "blkif-backend", blkif);
if (err < 0) { if (err < 0) {
unmap_frontend_page(blkif); xenbus_unmap_ring_vfree(blkif->be->dev, blkif->blk_ring);
free_vm_area(blkif->blk_ring_area);
blkif->blk_rings.common.sring = NULL; blkif->blk_rings.common.sring = NULL;
return err; return err;
} }
...@@ -230,8 +191,7 @@ static void xen_blkif_disconnect(struct xen_blkif *blkif) ...@@ -230,8 +191,7 @@ static void xen_blkif_disconnect(struct xen_blkif *blkif)
} }
if (blkif->blk_rings.common.sring) { if (blkif->blk_rings.common.sring) {
unmap_frontend_page(blkif); xenbus_unmap_ring_vfree(blkif->be->dev, blkif->blk_ring);
free_vm_area(blkif->blk_ring_area);
blkif->blk_rings.common.sring = NULL; blkif->blk_rings.common.sring = NULL;
} }
} }
......
...@@ -58,10 +58,6 @@ struct xenvif { ...@@ -58,10 +58,6 @@ struct xenvif {
u8 fe_dev_addr[6]; u8 fe_dev_addr[6];
/* Physical parameters of the comms window. */ /* Physical parameters of the comms window. */
grant_handle_t tx_shmem_handle;
grant_ref_t tx_shmem_ref;
grant_handle_t rx_shmem_handle;
grant_ref_t rx_shmem_ref;
unsigned int irq; unsigned int irq;
/* List of frontends to notify after a batch of frames sent. */ /* List of frontends to notify after a batch of frames sent. */
...@@ -70,8 +66,6 @@ struct xenvif { ...@@ -70,8 +66,6 @@ struct xenvif {
/* The shared rings and indexes. */ /* The shared rings and indexes. */
struct xen_netif_tx_back_ring tx; struct xen_netif_tx_back_ring tx;
struct xen_netif_rx_back_ring rx; struct xen_netif_rx_back_ring rx;
struct vm_struct *tx_comms_area;
struct vm_struct *rx_comms_area;
/* Frontend feature information. */ /* Frontend feature information. */
u8 can_sg:1; u8 can_sg:1;
...@@ -106,6 +100,11 @@ struct xenvif { ...@@ -106,6 +100,11 @@ struct xenvif {
wait_queue_head_t waiting_to_free; wait_queue_head_t waiting_to_free;
}; };
static inline struct xenbus_device *xenvif_to_xenbus_device(struct xenvif *vif)
{
return to_xenbus_device(vif->dev->dev.parent);
}
#define XEN_NETIF_TX_RING_SIZE __CONST_RING_SIZE(xen_netif_tx, PAGE_SIZE) #define XEN_NETIF_TX_RING_SIZE __CONST_RING_SIZE(xen_netif_tx, PAGE_SIZE)
#define XEN_NETIF_RX_RING_SIZE __CONST_RING_SIZE(xen_netif_rx, PAGE_SIZE) #define XEN_NETIF_RX_RING_SIZE __CONST_RING_SIZE(xen_netif_rx, PAGE_SIZE)
......
...@@ -1589,88 +1589,42 @@ static int xen_netbk_kthread(void *data) ...@@ -1589,88 +1589,42 @@ static int xen_netbk_kthread(void *data)
void xen_netbk_unmap_frontend_rings(struct xenvif *vif) void xen_netbk_unmap_frontend_rings(struct xenvif *vif)
{ {
struct gnttab_unmap_grant_ref op; if (vif->tx.sring)
xenbus_unmap_ring_vfree(xenvif_to_xenbus_device(vif),
if (vif->tx.sring) { vif->tx.sring);
gnttab_set_unmap_op(&op, (unsigned long)vif->tx_comms_area->addr, if (vif->rx.sring)
GNTMAP_host_map, vif->tx_shmem_handle); xenbus_unmap_ring_vfree(xenvif_to_xenbus_device(vif),
vif->rx.sring);
if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
BUG();
}
if (vif->rx.sring) {
gnttab_set_unmap_op(&op, (unsigned long)vif->rx_comms_area->addr,
GNTMAP_host_map, vif->rx_shmem_handle);
if (HYPERVISOR_grant_table_op(GNTTABOP_unmap_grant_ref, &op, 1))
BUG();
}
if (vif->rx_comms_area)
free_vm_area(vif->rx_comms_area);
if (vif->tx_comms_area)
free_vm_area(vif->tx_comms_area);
} }
int xen_netbk_map_frontend_rings(struct xenvif *vif, int xen_netbk_map_frontend_rings(struct xenvif *vif,
grant_ref_t tx_ring_ref, grant_ref_t tx_ring_ref,
grant_ref_t rx_ring_ref) grant_ref_t rx_ring_ref)
{ {
struct gnttab_map_grant_ref op; void *addr;
struct xen_netif_tx_sring *txs; struct xen_netif_tx_sring *txs;
struct xen_netif_rx_sring *rxs; struct xen_netif_rx_sring *rxs;
int err = -ENOMEM; int err = -ENOMEM;
vif->tx_comms_area = alloc_vm_area(PAGE_SIZE); err = xenbus_map_ring_valloc(xenvif_to_xenbus_device(vif),
if (vif->tx_comms_area == NULL) tx_ring_ref, &addr);
if (err)
goto err; goto err;
vif->rx_comms_area = alloc_vm_area(PAGE_SIZE); txs = (struct xen_netif_tx_sring *)addr;
if (vif->rx_comms_area == NULL)
goto err;
gnttab_set_map_op(&op, (unsigned long)vif->tx_comms_area->addr,
GNTMAP_host_map, tx_ring_ref, vif->domid);
if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
BUG();
if (op.status) {
netdev_warn(vif->dev,
"failed to map tx ring. err=%d status=%d\n",
err, op.status);
err = op.status;
goto err;
}
vif->tx_shmem_ref = tx_ring_ref;
vif->tx_shmem_handle = op.handle;
txs = (struct xen_netif_tx_sring *)vif->tx_comms_area->addr;
BACK_RING_INIT(&vif->tx, txs, PAGE_SIZE); BACK_RING_INIT(&vif->tx, txs, PAGE_SIZE);
gnttab_set_map_op(&op, (unsigned long)vif->rx_comms_area->addr, err = xenbus_map_ring_valloc(xenvif_to_xenbus_device(vif),
GNTMAP_host_map, rx_ring_ref, vif->domid); rx_ring_ref, &addr);
if (err)
if (HYPERVISOR_grant_table_op(GNTTABOP_map_grant_ref, &op, 1))
BUG();
if (op.status) {
netdev_warn(vif->dev,
"failed to map rx ring. err=%d status=%d\n",
err, op.status);
err = op.status;
goto err; goto err;
}
vif->rx_shmem_ref = rx_ring_ref;
vif->rx_shmem_handle = op.handle;
vif->rx_req_cons_peek = 0;
rxs = (struct xen_netif_rx_sring *)vif->rx_comms_area->addr; rxs = (struct xen_netif_rx_sring *)addr;
BACK_RING_INIT(&vif->rx, rxs, PAGE_SIZE); BACK_RING_INIT(&vif->rx, rxs, PAGE_SIZE);
vif->rx_req_cons_peek = 0;
return 0; return 0;
err: err:
......
...@@ -443,7 +443,7 @@ int xenbus_map_ring_valloc(struct xenbus_device *dev, int gnt_ref, void **vaddr) ...@@ -443,7 +443,7 @@ int xenbus_map_ring_valloc(struct xenbus_device *dev, int gnt_ref, void **vaddr)
*vaddr = NULL; *vaddr = NULL;
area = xen_alloc_vm_area(PAGE_SIZE); area = alloc_vm_area(PAGE_SIZE);
if (!area) if (!area)
return -ENOMEM; return -ENOMEM;
...@@ -453,7 +453,7 @@ int xenbus_map_ring_valloc(struct xenbus_device *dev, int gnt_ref, void **vaddr) ...@@ -453,7 +453,7 @@ int xenbus_map_ring_valloc(struct xenbus_device *dev, int gnt_ref, void **vaddr)
BUG(); BUG();
if (op.status != GNTST_okay) { if (op.status != GNTST_okay) {
xen_free_vm_area(area); free_vm_area(area);
xenbus_dev_fatal(dev, op.status, xenbus_dev_fatal(dev, op.status,
"mapping in shared page %d from domain %d", "mapping in shared page %d from domain %d",
gnt_ref, dev->otherend_id); gnt_ref, dev->otherend_id);
...@@ -552,7 +552,7 @@ int xenbus_unmap_ring_vfree(struct xenbus_device *dev, void *vaddr) ...@@ -552,7 +552,7 @@ int xenbus_unmap_ring_vfree(struct xenbus_device *dev, void *vaddr)
BUG(); BUG();
if (op.status == GNTST_okay) if (op.status == GNTST_okay)
xen_free_vm_area(area); free_vm_area(area);
else else
xenbus_dev_error(dev, op.status, xenbus_dev_error(dev, op.status,
"unmapping page at handle %d error %d", "unmapping page at handle %d error %d",
......
...@@ -43,7 +43,6 @@ ...@@ -43,7 +43,6 @@
#include <xen/interface/grant_table.h> #include <xen/interface/grant_table.h>
#include <asm/xen/hypervisor.h> #include <asm/xen/hypervisor.h>
#include <asm/xen/grant_table.h>
#include <xen/features.h> #include <xen/features.h>
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册