of_iommu.c 6.8 KB
Newer Older
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20
/*
 * OF helpers for IOMMU
 *
 * Copyright (c) 2012, NVIDIA CORPORATION.  All rights reserved.
 *
 * This program is free software; you can redistribute it and/or modify it
 * under the terms and conditions of the GNU General Public License,
 * version 2, as published by the Free Software Foundation.
 *
 * This program is distributed in the hope it will be useful, but WITHOUT
 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
 * FITNESS FOR A PARTICULAR PURPOSE.  See the GNU General Public License for
 * more details.
 *
 * You should have received a copy of the GNU General Public License along with
 * this program; if not, write to the Free Software Foundation, Inc.,
 * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
 */

#include <linux/export.h>
21
#include <linux/iommu.h>
22 23
#include <linux/limits.h>
#include <linux/of.h>
B
Brian Norris 已提交
24
#include <linux/of_iommu.h>
25
#include <linux/of_pci.h>
26
#include <linux/slab.h>
27

28 29 30
static const struct of_device_id __iommu_of_table_sentinel
	__used __section(__iommu_of_table_end);

31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97
/**
 * of_get_dma_window - Parse *dma-window property and returns 0 if found.
 *
 * @dn: device node
 * @prefix: prefix for property name if any
 * @index: index to start to parse
 * @busno: Returns busno if supported. Otherwise pass NULL
 * @addr: Returns address that DMA starts
 * @size: Returns the range that DMA can handle
 *
 * This supports different formats flexibly. "prefix" can be
 * configured if any. "busno" and "index" are optionally
 * specified. Set 0(or NULL) if not used.
 */
int of_get_dma_window(struct device_node *dn, const char *prefix, int index,
		      unsigned long *busno, dma_addr_t *addr, size_t *size)
{
	const __be32 *dma_window, *end;
	int bytes, cur_index = 0;
	char propname[NAME_MAX], addrname[NAME_MAX], sizename[NAME_MAX];

	if (!dn || !addr || !size)
		return -EINVAL;

	if (!prefix)
		prefix = "";

	snprintf(propname, sizeof(propname), "%sdma-window", prefix);
	snprintf(addrname, sizeof(addrname), "%s#dma-address-cells", prefix);
	snprintf(sizename, sizeof(sizename), "%s#dma-size-cells", prefix);

	dma_window = of_get_property(dn, propname, &bytes);
	if (!dma_window)
		return -ENODEV;
	end = dma_window + bytes / sizeof(*dma_window);

	while (dma_window < end) {
		u32 cells;
		const void *prop;

		/* busno is one cell if supported */
		if (busno)
			*busno = be32_to_cpup(dma_window++);

		prop = of_get_property(dn, addrname, NULL);
		if (!prop)
			prop = of_get_property(dn, "#address-cells", NULL);

		cells = prop ? be32_to_cpup(prop) : of_n_addr_cells(dn);
		if (!cells)
			return -EINVAL;
		*addr = of_read_number(dma_window, cells);
		dma_window += cells;

		prop = of_get_property(dn, sizename, NULL);
		cells = prop ? be32_to_cpup(prop) : of_n_size_cells(dn);
		if (!cells)
			return -EINVAL;
		*size = of_read_number(dma_window, cells);
		dma_window += cells;

		if (cur_index++ == index)
			break;
	}
	return 0;
}
EXPORT_SYMBOL_GPL(of_get_dma_window);
98

99 100 101 102 103 104 105
static bool of_iommu_driver_present(struct device_node *np)
{
	/*
	 * If the IOMMU still isn't ready by the time we reach init, assume
	 * it never will be. We don't want to defer indefinitely, nor attempt
	 * to dereference __iommu_of_table after it's been freed.
	 */
106
	if (system_state >= SYSTEM_RUNNING)
107 108 109 110 111
		return false;

	return of_match_node(&__iommu_of_table, np);
}

112 113 114 115 116 117 118 119
static const struct iommu_ops
*of_iommu_xlate(struct device *dev, struct of_phandle_args *iommu_spec)
{
	const struct iommu_ops *ops;
	struct fwnode_handle *fwnode = &iommu_spec->np->fwnode;
	int err;

	ops = iommu_ops_from_fwnode(fwnode);
120
	if ((ops && !ops->of_xlate) ||
121
	    !of_device_is_available(iommu_spec->np) ||
122
	    (!ops && !of_iommu_driver_present(iommu_spec->np)))
123 124 125 126 127
		return NULL;

	err = iommu_fwspec_init(dev, &iommu_spec->np->fwnode, ops);
	if (err)
		return ERR_PTR(err);
128 129 130 131 132 133 134
	/*
	 * The otherwise-empty fwspec handily serves to indicate the specific
	 * IOMMU device we're waiting for, which will be useful if we ever get
	 * a proper probe-ordering dependency mechanism in future.
	 */
	if (!ops)
		return ERR_PTR(-EPROBE_DEFER);
135 136 137 138 139 140 141 142

	err = ops->of_xlate(dev, iommu_spec);
	if (err)
		return ERR_PTR(err);

	return ops;
}

143 144 145 146
struct of_pci_iommu_alias_info {
	struct device *dev;
	struct device_node *np;
};
147

148
static int of_pci_iommu_init(struct pci_dev *pdev, u16 alias, void *data)
149
{
150
	struct of_pci_iommu_alias_info *info = data;
151
	const struct iommu_ops *ops;
152
	struct of_phandle_args iommu_spec = { .args_count = 1 };
153
	int err;
154

155
	err = of_pci_map_rid(info->np, alias, "iommu-map",
156 157 158
			     "iommu-map-mask", &iommu_spec.np,
			     iommu_spec.args);
	if (err)
159
		return err == -ENODEV ? 1 : err;
160

161
	ops = of_iommu_xlate(info->dev, &iommu_spec);
162
	of_node_put(iommu_spec.np);
163

164 165
	if (IS_ERR(ops))
		return PTR_ERR(ops);
166

167
	return info->np == pdev->bus->dev.of_node;
168 169 170 171 172
}

const struct iommu_ops *of_iommu_configure(struct device *dev,
					   struct device_node *master_np)
{
173
	const struct iommu_ops *ops = NULL;
174
	struct iommu_fwspec *fwspec = dev->iommu_fwspec;
175
	int err;
176 177 178 179

	if (!master_np)
		return NULL;

180 181 182 183 184 185 186 187
	if (fwspec) {
		if (fwspec->ops)
			return fwspec->ops;

		/* In the deferred case, start again from scratch */
		iommu_fwspec_free(dev);
	}

188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218
	/*
	 * We don't currently walk up the tree looking for a parent IOMMU.
	 * See the `Notes:' section of
	 * Documentation/devicetree/bindings/iommu/iommu.txt
	 */
	if (dev_is_pci(dev)) {
		struct of_pci_iommu_alias_info info = {
			.dev = dev,
			.np = master_np,
		};

		err = pci_for_each_dma_alias(to_pci_dev(dev),
					     of_pci_iommu_init, &info);
		if (err) /* err > 0 means the walk stopped, but non-fatally */
			ops = ERR_PTR(min(err, 0));
		else /* success implies both fwspec and ops are now valid */
			ops = dev->iommu_fwspec->ops;
	} else {
		struct of_phandle_args iommu_spec;
		int idx = 0;

		while (!of_parse_phandle_with_args(master_np, "iommus",
						   "#iommu-cells",
						   idx, &iommu_spec)) {
			ops = of_iommu_xlate(dev, &iommu_spec);
			of_node_put(iommu_spec.np);
			idx++;
			if (IS_ERR_OR_NULL(ops))
				break;
		}
	}
219 220 221 222 223 224
	/*
	 * If we have reason to believe the IOMMU driver missed the initial
	 * add_device callback for dev, replay it to get things in order.
	 */
	if (!IS_ERR_OR_NULL(ops) && ops->add_device &&
	    dev->bus && !dev->iommu_group) {
225
		err = ops->add_device(dev);
226 227 228
		if (err)
			ops = ERR_PTR(err);
	}
229

230 231 232 233 234 235
	/* Ignore all other errors apart from EPROBE_DEFER */
	if (IS_ERR(ops) && (PTR_ERR(ops) != -EPROBE_DEFER)) {
		dev_dbg(dev, "Adding to IOMMU failed: %ld\n", PTR_ERR(ops));
		ops = NULL;
	}

236
	return ops;
237 238
}

239
static int __init of_iommu_init(void)
240 241 242 243 244 245 246
{
	struct device_node *np;
	const struct of_device_id *match, *matches = &__iommu_of_table;

	for_each_matching_node_and_match(np, matches, &match) {
		const of_iommu_init_fn init_fn = match->data;

247
		if (init_fn && init_fn(np))
248 249 250
			pr_err("Failed to initialise IOMMU %s\n",
				of_node_full_name(np));
	}
251 252

	return 0;
253
}
254
postcore_initcall_sync(of_iommu_init);