提交 ee3f6ba8 编写于 作者: A Alexander Duyck 提交者: Konrad Rzeszutek Wilk

swiotlb: Make io_tlb_overflow_buffer a physical address

This change makes it so that we can avoid virt_to_phys overhead when using the
io_tlb_overflow_buffer.  My original plan was to completely remove the value
and replace it with a constant but I had seen that there were recent patches
that stated this couldn't be done until all device drivers that depended on
that functionality be updated.
Signed-off-by: NAlexander Duyck <alexander.h.duyck@intel.com>
Signed-off-by: NKonrad Rzeszutek Wilk <konrad.wilk@oracle.com>
上级 ff7204a7
...@@ -70,7 +70,7 @@ static unsigned long io_tlb_nslabs; ...@@ -70,7 +70,7 @@ static unsigned long io_tlb_nslabs;
*/ */
static unsigned long io_tlb_overflow = 32*1024; static unsigned long io_tlb_overflow = 32*1024;
static void *io_tlb_overflow_buffer; static phys_addr_t io_tlb_overflow_buffer;
/* /*
* This is a free list describing the number of free entries available from * This is a free list describing the number of free entries available from
...@@ -138,6 +138,7 @@ void swiotlb_print_info(void) ...@@ -138,6 +138,7 @@ void swiotlb_print_info(void)
void __init swiotlb_init_with_tbl(char *tlb, unsigned long nslabs, int verbose) void __init swiotlb_init_with_tbl(char *tlb, unsigned long nslabs, int verbose)
{ {
void *v_overflow_buffer;
unsigned long i, bytes; unsigned long i, bytes;
bytes = nslabs << IO_TLB_SHIFT; bytes = nslabs << IO_TLB_SHIFT;
...@@ -146,6 +147,15 @@ void __init swiotlb_init_with_tbl(char *tlb, unsigned long nslabs, int verbose) ...@@ -146,6 +147,15 @@ void __init swiotlb_init_with_tbl(char *tlb, unsigned long nslabs, int verbose)
io_tlb_start = __pa(tlb); io_tlb_start = __pa(tlb);
io_tlb_end = io_tlb_start + bytes; io_tlb_end = io_tlb_start + bytes;
/*
* Get the overflow emergency buffer
*/
v_overflow_buffer = alloc_bootmem_low_pages(PAGE_ALIGN(io_tlb_overflow));
if (!v_overflow_buffer)
panic("Cannot allocate SWIOTLB overflow buffer!\n");
io_tlb_overflow_buffer = __pa(v_overflow_buffer);
/* /*
* Allocate and initialize the free list array. This array is used * Allocate and initialize the free list array. This array is used
* to find contiguous free memory regions of size up to IO_TLB_SEGSIZE * to find contiguous free memory regions of size up to IO_TLB_SEGSIZE
...@@ -157,12 +167,6 @@ void __init swiotlb_init_with_tbl(char *tlb, unsigned long nslabs, int verbose) ...@@ -157,12 +167,6 @@ void __init swiotlb_init_with_tbl(char *tlb, unsigned long nslabs, int verbose)
io_tlb_index = 0; io_tlb_index = 0;
io_tlb_orig_addr = alloc_bootmem_pages(PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t))); io_tlb_orig_addr = alloc_bootmem_pages(PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t)));
/*
* Get the overflow emergency buffer
*/
io_tlb_overflow_buffer = alloc_bootmem_low_pages(PAGE_ALIGN(io_tlb_overflow));
if (!io_tlb_overflow_buffer)
panic("Cannot allocate SWIOTLB overflow buffer!\n");
if (verbose) if (verbose)
swiotlb_print_info(); swiotlb_print_info();
} }
...@@ -252,6 +256,7 @@ int ...@@ -252,6 +256,7 @@ int
swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs) swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs)
{ {
unsigned long i, bytes; unsigned long i, bytes;
unsigned char *v_overflow_buffer;
bytes = nslabs << IO_TLB_SHIFT; bytes = nslabs << IO_TLB_SHIFT;
...@@ -261,6 +266,16 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs) ...@@ -261,6 +266,16 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs)
memset(tlb, 0, bytes); memset(tlb, 0, bytes);
/*
* Get the overflow emergency buffer
*/
v_overflow_buffer = (void *)__get_free_pages(GFP_DMA,
get_order(io_tlb_overflow));
if (!v_overflow_buffer)
goto cleanup2;
io_tlb_overflow_buffer = virt_to_phys(v_overflow_buffer);
/* /*
* Allocate and initialize the free list array. This array is used * Allocate and initialize the free list array. This array is used
* to find contiguous free memory regions of size up to IO_TLB_SEGSIZE * to find contiguous free memory regions of size up to IO_TLB_SEGSIZE
...@@ -269,7 +284,7 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs) ...@@ -269,7 +284,7 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs)
io_tlb_list = (unsigned int *)__get_free_pages(GFP_KERNEL, io_tlb_list = (unsigned int *)__get_free_pages(GFP_KERNEL,
get_order(io_tlb_nslabs * sizeof(int))); get_order(io_tlb_nslabs * sizeof(int)));
if (!io_tlb_list) if (!io_tlb_list)
goto cleanup2; goto cleanup3;
for (i = 0; i < io_tlb_nslabs; i++) for (i = 0; i < io_tlb_nslabs; i++)
io_tlb_list[i] = IO_TLB_SEGSIZE - OFFSET(i, IO_TLB_SEGSIZE); io_tlb_list[i] = IO_TLB_SEGSIZE - OFFSET(i, IO_TLB_SEGSIZE);
...@@ -280,18 +295,10 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs) ...@@ -280,18 +295,10 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs)
get_order(io_tlb_nslabs * get_order(io_tlb_nslabs *
sizeof(phys_addr_t))); sizeof(phys_addr_t)));
if (!io_tlb_orig_addr) if (!io_tlb_orig_addr)
goto cleanup3; goto cleanup4;
memset(io_tlb_orig_addr, 0, io_tlb_nslabs * sizeof(phys_addr_t)); memset(io_tlb_orig_addr, 0, io_tlb_nslabs * sizeof(phys_addr_t));
/*
* Get the overflow emergency buffer
*/
io_tlb_overflow_buffer = (void *)__get_free_pages(GFP_DMA,
get_order(io_tlb_overflow));
if (!io_tlb_overflow_buffer)
goto cleanup4;
swiotlb_print_info(); swiotlb_print_info();
late_alloc = 1; late_alloc = 1;
...@@ -299,13 +306,13 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs) ...@@ -299,13 +306,13 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs)
return 0; return 0;
cleanup4: cleanup4:
free_pages((unsigned long)io_tlb_orig_addr,
get_order(io_tlb_nslabs * sizeof(phys_addr_t)));
io_tlb_orig_addr = NULL;
cleanup3:
free_pages((unsigned long)io_tlb_list, get_order(io_tlb_nslabs * free_pages((unsigned long)io_tlb_list, get_order(io_tlb_nslabs *
sizeof(int))); sizeof(int)));
io_tlb_list = NULL; io_tlb_list = NULL;
cleanup3:
free_pages((unsigned long)v_overflow_buffer,
get_order(io_tlb_overflow));
io_tlb_overflow_buffer = 0;
cleanup2: cleanup2:
io_tlb_end = 0; io_tlb_end = 0;
io_tlb_start = 0; io_tlb_start = 0;
...@@ -315,11 +322,11 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs) ...@@ -315,11 +322,11 @@ swiotlb_late_init_with_tbl(char *tlb, unsigned long nslabs)
void __init swiotlb_free(void) void __init swiotlb_free(void)
{ {
if (!io_tlb_overflow_buffer) if (!io_tlb_orig_addr)
return; return;
if (late_alloc) { if (late_alloc) {
free_pages((unsigned long)io_tlb_overflow_buffer, free_pages((unsigned long)phys_to_virt(io_tlb_overflow_buffer),
get_order(io_tlb_overflow)); get_order(io_tlb_overflow));
free_pages((unsigned long)io_tlb_orig_addr, free_pages((unsigned long)io_tlb_orig_addr,
get_order(io_tlb_nslabs * sizeof(phys_addr_t))); get_order(io_tlb_nslabs * sizeof(phys_addr_t)));
...@@ -328,7 +335,7 @@ void __init swiotlb_free(void) ...@@ -328,7 +335,7 @@ void __init swiotlb_free(void)
free_pages((unsigned long)phys_to_virt(io_tlb_start), free_pages((unsigned long)phys_to_virt(io_tlb_start),
get_order(io_tlb_nslabs << IO_TLB_SHIFT)); get_order(io_tlb_nslabs << IO_TLB_SHIFT));
} else { } else {
free_bootmem_late(__pa(io_tlb_overflow_buffer), free_bootmem_late(io_tlb_overflow_buffer,
PAGE_ALIGN(io_tlb_overflow)); PAGE_ALIGN(io_tlb_overflow));
free_bootmem_late(__pa(io_tlb_orig_addr), free_bootmem_late(__pa(io_tlb_orig_addr),
PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t))); PAGE_ALIGN(io_tlb_nslabs * sizeof(phys_addr_t)));
...@@ -698,7 +705,7 @@ dma_addr_t swiotlb_map_page(struct device *dev, struct page *page, ...@@ -698,7 +705,7 @@ dma_addr_t swiotlb_map_page(struct device *dev, struct page *page,
map = map_single(dev, phys, size, dir); map = map_single(dev, phys, size, dir);
if (!map) { if (!map) {
swiotlb_full(dev, size, dir, 1); swiotlb_full(dev, size, dir, 1);
map = io_tlb_overflow_buffer; return phys_to_dma(dev, io_tlb_overflow_buffer);
} }
dev_addr = swiotlb_virt_to_bus(dev, map); dev_addr = swiotlb_virt_to_bus(dev, map);
...@@ -708,7 +715,7 @@ dma_addr_t swiotlb_map_page(struct device *dev, struct page *page, ...@@ -708,7 +715,7 @@ dma_addr_t swiotlb_map_page(struct device *dev, struct page *page,
*/ */
if (!dma_capable(dev, dev_addr, size)) { if (!dma_capable(dev, dev_addr, size)) {
swiotlb_tbl_unmap_single(dev, map, size, dir); swiotlb_tbl_unmap_single(dev, map, size, dir);
dev_addr = swiotlb_virt_to_bus(dev, io_tlb_overflow_buffer); return phys_to_dma(dev, io_tlb_overflow_buffer);
} }
return dev_addr; return dev_addr;
...@@ -927,7 +934,7 @@ EXPORT_SYMBOL(swiotlb_sync_sg_for_device); ...@@ -927,7 +934,7 @@ EXPORT_SYMBOL(swiotlb_sync_sg_for_device);
int int
swiotlb_dma_mapping_error(struct device *hwdev, dma_addr_t dma_addr) swiotlb_dma_mapping_error(struct device *hwdev, dma_addr_t dma_addr)
{ {
return (dma_addr == swiotlb_virt_to_bus(hwdev, io_tlb_overflow_buffer)); return (dma_addr == phys_to_dma(hwdev, io_tlb_overflow_buffer));
} }
EXPORT_SYMBOL(swiotlb_dma_mapping_error); EXPORT_SYMBOL(swiotlb_dma_mapping_error);
......
Markdown is supported
0% .
You are about to add 0 people to the discussion. Proceed with caution.
先完成此消息的编辑!
想要评论请 注册