mtk_iommu_v1.c 20.2 KB
Newer Older
1
// SPDX-License-Identifier: GPL-2.0-only
2
/*
3 4
 * IOMMU API for MTK architected m4u v1 implementations
 *
5 6 7 8 9 10 11 12 13
 * Copyright (c) 2015-2016 MediaTek Inc.
 * Author: Honghui Zhang <honghui.zhang@mediatek.com>
 *
 * Based on driver/iommu/mtk_iommu.c
 */
#include <linux/bug.h>
#include <linux/clk.h>
#include <linux/component.h>
#include <linux/device.h>
14
#include <linux/dma-mapping.h>
15 16 17 18 19 20
#include <linux/err.h>
#include <linux/interrupt.h>
#include <linux/io.h>
#include <linux/iommu.h>
#include <linux/iopoll.h>
#include <linux/list.h>
21
#include <linux/module.h>
22 23 24 25 26 27 28 29
#include <linux/of_address.h>
#include <linux/of_irq.h>
#include <linux/of_platform.h>
#include <linux/platform_device.h>
#include <linux/slab.h>
#include <linux/spinlock.h>
#include <asm/barrier.h>
#include <asm/dma-iommu.h>
30
#include <dt-bindings/memory/mtk-memory-port.h>
31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80
#include <dt-bindings/memory/mt2701-larb-port.h>
#include <soc/mediatek/smi.h>

#define REG_MMU_PT_BASE_ADDR			0x000

#define F_ALL_INVLD				0x2
#define F_MMU_INV_RANGE				0x1
#define F_INVLD_EN0				BIT(0)
#define F_INVLD_EN1				BIT(1)

#define F_MMU_FAULT_VA_MSK			0xfffff000
#define MTK_PROTECT_PA_ALIGN			128

#define REG_MMU_CTRL_REG			0x210
#define F_MMU_CTRL_COHERENT_EN			BIT(8)
#define REG_MMU_IVRP_PADDR			0x214
#define REG_MMU_INT_CONTROL			0x220
#define F_INT_TRANSLATION_FAULT			BIT(0)
#define F_INT_MAIN_MULTI_HIT_FAULT		BIT(1)
#define F_INT_INVALID_PA_FAULT			BIT(2)
#define F_INT_ENTRY_REPLACEMENT_FAULT		BIT(3)
#define F_INT_TABLE_WALK_FAULT			BIT(4)
#define F_INT_TLB_MISS_FAULT			BIT(5)
#define F_INT_PFH_DMA_FIFO_OVERFLOW		BIT(6)
#define F_INT_MISS_DMA_FIFO_OVERFLOW		BIT(7)

#define F_MMU_TF_PROTECT_SEL(prot)		(((prot) & 0x3) << 5)
#define F_INT_CLR_BIT				BIT(12)

#define REG_MMU_FAULT_ST			0x224
#define REG_MMU_FAULT_VA			0x228
#define REG_MMU_INVLD_PA			0x22C
#define REG_MMU_INT_ID				0x388
#define REG_MMU_INVALIDATE			0x5c0
#define REG_MMU_INVLD_START_A			0x5c4
#define REG_MMU_INVLD_END_A			0x5c8

#define REG_MMU_INV_SEL				0x5d8
#define REG_MMU_STANDARD_AXI_MODE		0x5e8

#define REG_MMU_DCM				0x5f0
#define F_MMU_DCM_ON				BIT(1)
#define REG_MMU_CPE_DONE			0x60c
#define F_DESC_VALID				0x2
#define F_DESC_NONSEC				BIT(3)
#define MT2701_M4U_TF_LARB(TF)			(6 - (((TF) >> 13) & 0x7))
#define MT2701_M4U_TF_PORT(TF)			(((TF) >> 8) & 0xF)
/* MTK generation one iommu HW only support 4K size mapping */
#define MT2701_IOMMU_PAGE_SHIFT			12
#define MT2701_IOMMU_PAGE_SIZE			(1UL << MT2701_IOMMU_PAGE_SHIFT)
81
#define MT2701_LARB_NR_MAX			3
82 83 84 85 86 87 88

/*
 * MTK m4u support 4GB iova address space, and only support 4K page
 * mapping. So the pagetable size should be exactly as 4M.
 */
#define M2701_IOMMU_PGT_SIZE			SZ_4M

89
struct mtk_iommu_v1_suspend_reg {
90 91 92 93 94 95
	u32			standard_axi_mode;
	u32			dcm_dis;
	u32			ctrl_reg;
	u32			int_control0;
};

96
struct mtk_iommu_v1_data {
97 98 99 100 101
	void __iomem			*base;
	int				irq;
	struct device			*dev;
	struct clk			*bclk;
	phys_addr_t			protect_base; /* protect memory base */
102
	struct mtk_iommu_v1_domain	*m4u_dom;
103 104 105 106 107

	struct iommu_device		iommu;
	struct dma_iommu_mapping	*mapping;
	struct mtk_smi_larb_iommu	larb_imu[MTK_LARB_NR_MAX];

108
	struct mtk_iommu_v1_suspend_reg	reg;
109 110
};

111
struct mtk_iommu_v1_domain {
112 113 114 115
	spinlock_t			pgtlock; /* lock for page table */
	struct iommu_domain		domain;
	u32				*pgt_va;
	dma_addr_t			pgt_pa;
116
	struct mtk_iommu_v1_data	*data;
117 118
};

119
static int mtk_iommu_v1_bind(struct device *dev)
120
{
121
	struct mtk_iommu_v1_data *data = dev_get_drvdata(dev);
122 123 124 125

	return component_bind_all(dev, &data->larb_imu);
}

126
static void mtk_iommu_v1_unbind(struct device *dev)
127
{
128
	struct mtk_iommu_v1_data *data = dev_get_drvdata(dev);
129 130 131 132

	component_unbind_all(dev, &data->larb_imu);
}

133
static struct mtk_iommu_v1_domain *to_mtk_domain(struct iommu_domain *dom)
134
{
135
	return container_of(dom, struct mtk_iommu_v1_domain, domain);
136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160
}

static const int mt2701_m4u_in_larb[] = {
	LARB0_PORT_OFFSET, LARB1_PORT_OFFSET,
	LARB2_PORT_OFFSET, LARB3_PORT_OFFSET
};

static inline int mt2701_m4u_to_larb(int id)
{
	int i;

	for (i = ARRAY_SIZE(mt2701_m4u_in_larb) - 1; i >= 0; i--)
		if ((id) >= mt2701_m4u_in_larb[i])
			return i;

	return 0;
}

static inline int mt2701_m4u_to_port(int id)
{
	int larb = mt2701_m4u_to_larb(id);

	return id - mt2701_m4u_in_larb[larb];
}

161
static void mtk_iommu_v1_tlb_flush_all(struct mtk_iommu_v1_data *data)
162 163 164 165 166 167 168
{
	writel_relaxed(F_INVLD_EN1 | F_INVLD_EN0,
			data->base + REG_MMU_INV_SEL);
	writel_relaxed(F_ALL_INVLD, data->base + REG_MMU_INVALIDATE);
	wmb(); /* Make sure the tlb flush all done */
}

169 170
static void mtk_iommu_v1_tlb_flush_range(struct mtk_iommu_v1_data *data,
					 unsigned long iova, size_t size)
171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187
{
	int ret;
	u32 tmp;

	writel_relaxed(F_INVLD_EN1 | F_INVLD_EN0,
		data->base + REG_MMU_INV_SEL);
	writel_relaxed(iova & F_MMU_FAULT_VA_MSK,
		data->base + REG_MMU_INVLD_START_A);
	writel_relaxed((iova + size - 1) & F_MMU_FAULT_VA_MSK,
		data->base + REG_MMU_INVLD_END_A);
	writel_relaxed(F_MMU_INV_RANGE, data->base + REG_MMU_INVALIDATE);

	ret = readl_poll_timeout_atomic(data->base + REG_MMU_CPE_DONE,
				tmp, tmp != 0, 10, 100000);
	if (ret) {
		dev_warn(data->dev,
			 "Partial TLB flush timed out, falling back to full flush\n");
188
		mtk_iommu_v1_tlb_flush_all(data);
189 190 191 192 193
	}
	/* Clear the CPE status */
	writel_relaxed(0, data->base + REG_MMU_CPE_DONE);
}

194
static irqreturn_t mtk_iommu_v1_isr(int irq, void *dev_id)
195
{
196 197
	struct mtk_iommu_v1_data *data = dev_id;
	struct mtk_iommu_v1_domain *dom = data->m4u_dom;
198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 222 223 224 225 226
	u32 int_state, regval, fault_iova, fault_pa;
	unsigned int fault_larb, fault_port;

	/* Read error information from registers */
	int_state = readl_relaxed(data->base + REG_MMU_FAULT_ST);
	fault_iova = readl_relaxed(data->base + REG_MMU_FAULT_VA);

	fault_iova &= F_MMU_FAULT_VA_MSK;
	fault_pa = readl_relaxed(data->base + REG_MMU_INVLD_PA);
	regval = readl_relaxed(data->base + REG_MMU_INT_ID);
	fault_larb = MT2701_M4U_TF_LARB(regval);
	fault_port = MT2701_M4U_TF_PORT(regval);

	/*
	 * MTK v1 iommu HW could not determine whether the fault is read or
	 * write fault, report as read fault.
	 */
	if (report_iommu_fault(&dom->domain, data->dev, fault_iova,
			IOMMU_FAULT_READ))
		dev_err_ratelimited(data->dev,
			"fault type=0x%x iova=0x%x pa=0x%x larb=%d port=%d\n",
			int_state, fault_iova, fault_pa,
			fault_larb, fault_port);

	/* Interrupt clear */
	regval = readl_relaxed(data->base + REG_MMU_INT_CONTROL);
	regval |= F_INT_CLR_BIT;
	writel_relaxed(regval, data->base + REG_MMU_INT_CONTROL);

227
	mtk_iommu_v1_tlb_flush_all(data);
228 229 230 231

	return IRQ_HANDLED;
}

232 233
static void mtk_iommu_v1_config(struct mtk_iommu_v1_data *data,
				struct device *dev, bool enable)
234 235 236
{
	struct mtk_smi_larb_iommu    *larb_mmu;
	unsigned int                 larbid, portid;
237
	struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
238
	int i;
239

240 241 242
	for (i = 0; i < fwspec->num_ids; ++i) {
		larbid = mt2701_m4u_to_larb(fwspec->ids[i]);
		portid = mt2701_m4u_to_port(fwspec->ids[i]);
243
		larb_mmu = &data->larb_imu[larbid];
244 245 246 247 248 249 250 251 252 253 254

		dev_dbg(dev, "%s iommu port: %d\n",
			enable ? "enable" : "disable", portid);

		if (enable)
			larb_mmu->mmu |= MTK_SMI_MMU_EN(portid);
		else
			larb_mmu->mmu &= ~MTK_SMI_MMU_EN(portid);
	}
}

255
static int mtk_iommu_v1_domain_finalise(struct mtk_iommu_v1_data *data)
256
{
257
	struct mtk_iommu_v1_domain *dom = data->m4u_dom;
258 259 260

	spin_lock_init(&dom->pgtlock);

261 262
	dom->pgt_va = dma_alloc_coherent(data->dev, M2701_IOMMU_PGT_SIZE,
					 &dom->pgt_pa, GFP_KERNEL);
263 264 265 266 267 268 269 270 271 272
	if (!dom->pgt_va)
		return -ENOMEM;

	writel(dom->pgt_pa, data->base + REG_MMU_PT_BASE_ADDR);

	dom->data = data;

	return 0;
}

273
static struct iommu_domain *mtk_iommu_v1_domain_alloc_paging(struct device *dev)
274
{
275
	struct mtk_iommu_v1_domain *dom;
276 277 278 279 280 281 282 283

	dom = kzalloc(sizeof(*dom), GFP_KERNEL);
	if (!dom)
		return NULL;

	return &dom->domain;
}

284
static void mtk_iommu_v1_domain_free(struct iommu_domain *domain)
285
{
286 287
	struct mtk_iommu_v1_domain *dom = to_mtk_domain(domain);
	struct mtk_iommu_v1_data *data = dom->data;
288 289 290 291 292 293

	dma_free_coherent(data->dev, M2701_IOMMU_PGT_SIZE,
			dom->pgt_va, dom->pgt_pa);
	kfree(to_mtk_domain(domain));
}

294
static int mtk_iommu_v1_attach_device(struct iommu_domain *domain, struct device *dev)
295
{
296 297
	struct mtk_iommu_v1_data *data = dev_iommu_priv_get(dev);
	struct mtk_iommu_v1_domain *dom = to_mtk_domain(domain);
298
	struct dma_iommu_mapping *mtk_mapping;
299 300
	int ret;

301
	/* Only allow the domain created internally. */
302
	mtk_mapping = data->mapping;
303 304
	if (mtk_mapping->domain != domain)
		return 0;
305 306 307

	if (!data->m4u_dom) {
		data->m4u_dom = dom;
308
		ret = mtk_iommu_v1_domain_finalise(data);
309 310 311 312 313 314
		if (ret) {
			data->m4u_dom = NULL;
			return ret;
		}
	}

315
	mtk_iommu_v1_config(data, dev, true);
316 317 318
	return 0;
}

319 320
static int mtk_iommu_v1_identity_attach(struct iommu_domain *identity_domain,
					struct device *dev)
321
{
322
	struct mtk_iommu_v1_data *data = dev_iommu_priv_get(dev);
323

324
	mtk_iommu_v1_config(data, dev, false);
325 326 327 328 329 330 331 332 333 334 335 336
	return 0;
}

static struct iommu_domain_ops mtk_iommu_v1_identity_ops = {
	.attach_dev = mtk_iommu_v1_identity_attach,
};

static struct iommu_domain mtk_iommu_v1_identity_domain = {
	.type = IOMMU_DOMAIN_IDENTITY,
	.ops = &mtk_iommu_v1_identity_ops,
};

337
static int mtk_iommu_v1_map(struct iommu_domain *domain, unsigned long iova,
338 339
			    phys_addr_t paddr, size_t pgsize, size_t pgcount,
			    int prot, gfp_t gfp, size_t *mapped)
340
{
341
	struct mtk_iommu_v1_domain *dom = to_mtk_domain(domain);
342 343 344 345 346 347
	unsigned long flags;
	unsigned int i;
	u32 *pgt_base_iova = dom->pgt_va + (iova  >> MT2701_IOMMU_PAGE_SHIFT);
	u32 pabase = (u32)paddr;

	spin_lock_irqsave(&dom->pgtlock, flags);
348 349
	for (i = 0; i < pgcount; i++) {
		if (pgt_base_iova[i])
350 351 352 353 354 355 356
			break;
		pgt_base_iova[i] = pabase | F_DESC_VALID | F_DESC_NONSEC;
		pabase += MT2701_IOMMU_PAGE_SIZE;
	}

	spin_unlock_irqrestore(&dom->pgtlock, flags);

357 358
	*mapped = i * MT2701_IOMMU_PAGE_SIZE;
	mtk_iommu_v1_tlb_flush_range(dom->data, iova, *mapped);
359

360
	return i == pgcount ? 0 : -EEXIST;
361 362
}

363
static size_t mtk_iommu_v1_unmap(struct iommu_domain *domain, unsigned long iova,
364 365
				 size_t pgsize, size_t pgcount,
				 struct iommu_iotlb_gather *gather)
366
{
367
	struct mtk_iommu_v1_domain *dom = to_mtk_domain(domain);
368 369
	unsigned long flags;
	u32 *pgt_base_iova = dom->pgt_va + (iova  >> MT2701_IOMMU_PAGE_SHIFT);
370
	size_t size = pgcount * MT2701_IOMMU_PAGE_SIZE;
371 372

	spin_lock_irqsave(&dom->pgtlock, flags);
373
	memset(pgt_base_iova, 0, pgcount * sizeof(u32));
374 375
	spin_unlock_irqrestore(&dom->pgtlock, flags);

376
	mtk_iommu_v1_tlb_flush_range(dom->data, iova, size);
377 378 379 380

	return size;
}

381
static phys_addr_t mtk_iommu_v1_iova_to_phys(struct iommu_domain *domain, dma_addr_t iova)
382
{
383
	struct mtk_iommu_v1_domain *dom = to_mtk_domain(domain);
384 385 386 387 388 389 390 391 392 393 394
	unsigned long flags;
	phys_addr_t pa;

	spin_lock_irqsave(&dom->pgtlock, flags);
	pa = *(dom->pgt_va + (iova >> MT2701_IOMMU_PAGE_SHIFT));
	pa = pa & (~(MT2701_IOMMU_PAGE_SIZE - 1));
	spin_unlock_irqrestore(&dom->pgtlock, flags);

	return pa;
}

395
static const struct iommu_ops mtk_iommu_v1_ops;
396

397 398 399 400
/*
 * MTK generation one iommu HW only support one iommu domain, and all the client
 * sharing the same iova address space.
 */
401 402
static int mtk_iommu_v1_create_mapping(struct device *dev,
				       const struct of_phandle_args *args)
403
{
404
	struct mtk_iommu_v1_data *data;
405 406 407 408 409 410 411 412 413 414
	struct platform_device *m4updev;
	struct dma_iommu_mapping *mtk_mapping;
	int ret;

	if (args->args_count != 1) {
		dev_err(dev, "invalid #iommu-cells(%d) property for IOMMU\n",
			args->args_count);
		return -EINVAL;
	}

415
	ret = iommu_fwspec_init(dev, of_fwnode_handle(args->np));
416 417
	if (ret)
		return ret;
418

419
	if (!dev_iommu_priv_get(dev)) {
420 421 422 423 424
		/* Get the m4u device */
		m4updev = of_find_device_by_node(args->np);
		if (WARN_ON(!m4updev))
			return -EINVAL;

425
		dev_iommu_priv_set(dev, platform_get_drvdata(m4updev));
426 427
	}

428 429 430
	ret = iommu_fwspec_add_ids(dev, args->args, 1);
	if (ret)
		return ret;
431

432
	data = dev_iommu_priv_get(dev);
433
	mtk_mapping = data->mapping;
434 435
	if (!mtk_mapping) {
		/* MTK iommu support 4GB iova address space. */
436
		mtk_mapping = arm_iommu_create_mapping(dev, 0, 1ULL << 32);
437 438 439
		if (IS_ERR(mtk_mapping))
			return PTR_ERR(mtk_mapping);

440
		data->mapping = mtk_mapping;
441 442 443 444 445
	}

	return 0;
}

446
static struct iommu_device *mtk_iommu_v1_probe_device(struct device *dev)
447
{
448
	struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
449
	struct of_phandle_args iommu_spec;
450
	struct mtk_iommu_v1_data *data;
451 452 453
	int err, idx = 0, larbid, larbidx;
	struct device_link *link;
	struct device *larbdev;
454

455 456 457 458 459 460 461 462 463
	/*
	 * In the deferred case, free the existed fwspec.
	 * Always initialize the fwspec internally.
	 */
	if (fwspec) {
		iommu_fwspec_free(dev);
		fwspec = dev_iommu_fwspec_get(dev);
	}

464 465 466
	while (!of_parse_phandle_with_args(dev->of_node, "iommus",
					   "#iommu-cells",
					   idx, &iommu_spec)) {
467

468
		err = mtk_iommu_v1_create_mapping(dev, &iommu_spec);
469 470 471
		of_node_put(iommu_spec.np);
		if (err)
			return ERR_PTR(err);
472 473 474

		/* dev->iommu_fwspec might have changed */
		fwspec = dev_iommu_fwspec_get(dev);
475
		idx++;
476 477
	}

478
	data = dev_iommu_priv_get(dev);
479

480 481
	/* Link the consumer device with the smi-larb device(supplier) */
	larbid = mt2701_m4u_to_larb(fwspec->ids[0]);
482 483 484
	if (larbid >= MT2701_LARB_NR_MAX)
		return ERR_PTR(-EINVAL);

485 486 487 488 489 490 491 492 493 494
	for (idx = 1; idx < fwspec->num_ids; idx++) {
		larbidx = mt2701_m4u_to_larb(fwspec->ids[idx]);
		if (larbid != larbidx) {
			dev_err(dev, "Can only use one larb. Fail@larb%d-%d.\n",
				larbid, larbidx);
			return ERR_PTR(-EINVAL);
		}
	}

	larbdev = data->larb_imu[larbid].dev;
495 496 497
	if (!larbdev)
		return ERR_PTR(-EINVAL);

498 499 500 501 502
	link = device_link_add(dev, larbdev,
			       DL_FLAG_PM_RUNTIME | DL_FLAG_STATELESS);
	if (!link)
		dev_err(dev, "Unable to link %s\n", dev_name(larbdev));

503 504
	return &data->iommu;
}
505

506
static void mtk_iommu_v1_probe_finalize(struct device *dev)
507 508
{
	struct dma_iommu_mapping *mtk_mapping;
509
	struct mtk_iommu_v1_data *data;
510 511 512
	int err;

	data        = dev_iommu_priv_get(dev);
513
	mtk_mapping = data->mapping;
514

515 516 517
	err = arm_iommu_attach_device(dev, mtk_mapping);
	if (err)
		dev_err(dev, "Can't create IOMMU mapping - DMA-OPS will not work\n");
518 519
}

520
static void mtk_iommu_v1_release_device(struct device *dev)
521
{
522
	struct iommu_fwspec *fwspec = dev_iommu_fwspec_get(dev);
523
	struct mtk_iommu_v1_data *data;
524 525
	struct device *larbdev;
	unsigned int larbid;
526

527 528 529 530
	data = dev_iommu_priv_get(dev);
	larbid = mt2701_m4u_to_larb(fwspec->ids[0]);
	larbdev = data->larb_imu[larbid].dev;
	device_link_remove(dev, larbdev);
531 532
}

533
static int mtk_iommu_v1_hw_init(const struct mtk_iommu_v1_data *data)
534 535 536 537 538 539 540 541 542 543 544 545 546 547 548 549 550 551 552 553 554 555 556 557 558 559 560 561 562
{
	u32 regval;
	int ret;

	ret = clk_prepare_enable(data->bclk);
	if (ret) {
		dev_err(data->dev, "Failed to enable iommu bclk(%d)\n", ret);
		return ret;
	}

	regval = F_MMU_CTRL_COHERENT_EN | F_MMU_TF_PROTECT_SEL(2);
	writel_relaxed(regval, data->base + REG_MMU_CTRL_REG);

	regval = F_INT_TRANSLATION_FAULT |
		F_INT_MAIN_MULTI_HIT_FAULT |
		F_INT_INVALID_PA_FAULT |
		F_INT_ENTRY_REPLACEMENT_FAULT |
		F_INT_TABLE_WALK_FAULT |
		F_INT_TLB_MISS_FAULT |
		F_INT_PFH_DMA_FIFO_OVERFLOW |
		F_INT_MISS_DMA_FIFO_OVERFLOW;
	writel_relaxed(regval, data->base + REG_MMU_INT_CONTROL);

	/* protect memory,hw will write here while translation fault */
	writel_relaxed(data->protect_base,
			data->base + REG_MMU_IVRP_PADDR);

	writel_relaxed(F_MMU_DCM_ON, data->base + REG_MMU_DCM);

563
	if (devm_request_irq(data->dev, data->irq, mtk_iommu_v1_isr, 0,
564 565 566 567 568 569 570 571 572 573
			     dev_name(data->dev), (void *)data)) {
		writel_relaxed(0, data->base + REG_MMU_PT_BASE_ADDR);
		clk_disable_unprepare(data->bclk);
		dev_err(data->dev, "Failed @ IRQ-%d Request\n", data->irq);
		return -ENODEV;
	}

	return 0;
}

574
static const struct iommu_ops mtk_iommu_v1_ops = {
575
	.identity_domain = &mtk_iommu_v1_identity_domain,
576
	.domain_alloc_paging = mtk_iommu_v1_domain_alloc_paging,
577 578 579
	.probe_device	= mtk_iommu_v1_probe_device,
	.probe_finalize = mtk_iommu_v1_probe_finalize,
	.release_device	= mtk_iommu_v1_release_device,
580
	.device_group	= generic_device_group,
581
	.pgsize_bitmap	= MT2701_IOMMU_PAGE_SIZE,
582
	.owner          = THIS_MODULE,
Lu Baolu's avatar
Lu Baolu committed
583
	.default_domain_ops = &(const struct iommu_domain_ops) {
584
		.attach_dev	= mtk_iommu_v1_attach_device,
585 586
		.map_pages	= mtk_iommu_v1_map,
		.unmap_pages	= mtk_iommu_v1_unmap,
587 588
		.iova_to_phys	= mtk_iommu_v1_iova_to_phys,
		.free		= mtk_iommu_v1_domain_free,
Lu Baolu's avatar
Lu Baolu committed
589
	}
590 591
};

592
static const struct of_device_id mtk_iommu_v1_of_ids[] = {
593 594 595
	{ .compatible = "mediatek,mt2701-m4u", },
	{}
};
596
MODULE_DEVICE_TABLE(of, mtk_iommu_v1_of_ids);
597

598 599 600
static const struct component_master_ops mtk_iommu_v1_com_ops = {
	.bind		= mtk_iommu_v1_bind,
	.unbind		= mtk_iommu_v1_unbind,
601 602
};

603
static int mtk_iommu_v1_probe(struct platform_device *pdev)
604 605
{
	struct device			*dev = &pdev->dev;
606
	struct mtk_iommu_v1_data	*data;
607 608 609
	struct resource			*res;
	struct component_match		*match = NULL;
	void				*protect;
610
	int				larb_nr, ret, i;
611 612 613 614 615 616 617 618

	data = devm_kzalloc(dev, sizeof(*data), GFP_KERNEL);
	if (!data)
		return -ENOMEM;

	data->dev = dev;

	/* Protect memory. HW will access here while translation fault.*/
619 620
	protect = devm_kcalloc(dev, 2, MTK_PROTECT_PA_ALIGN,
			       GFP_KERNEL | GFP_DMA);
621 622 623 624 625 626 627 628 629 630 631 632 633 634 635 636 637
	if (!protect)
		return -ENOMEM;
	data->protect_base = ALIGN(virt_to_phys(protect), MTK_PROTECT_PA_ALIGN);

	res = platform_get_resource(pdev, IORESOURCE_MEM, 0);
	data->base = devm_ioremap_resource(dev, res);
	if (IS_ERR(data->base))
		return PTR_ERR(data->base);

	data->irq = platform_get_irq(pdev, 0);
	if (data->irq < 0)
		return data->irq;

	data->bclk = devm_clk_get(dev, "bclk");
	if (IS_ERR(data->bclk))
		return PTR_ERR(data->bclk);

638 639 640 641 642 643 644
	larb_nr = of_count_phandle_with_args(dev->of_node,
					     "mediatek,larbs", NULL);
	if (larb_nr < 0)
		return larb_nr;

	for (i = 0; i < larb_nr; i++) {
		struct device_node *larbnode;
645 646
		struct platform_device *plarbdev;

647 648 649
		larbnode = of_parse_phandle(dev->of_node, "mediatek,larbs", i);
		if (!larbnode)
			return -EINVAL;
650

651 652
		if (!of_device_is_available(larbnode)) {
			of_node_put(larbnode);
653
			continue;
654
		}
655

656
		plarbdev = of_find_device_by_node(larbnode);
657
		if (!plarbdev) {
658
			of_node_put(larbnode);
659
			return -ENODEV;
660
		}
661 662 663 664
		if (!plarbdev->dev.driver) {
			of_node_put(larbnode);
			return -EPROBE_DEFER;
		}
665
		data->larb_imu[i].dev = &plarbdev->dev;
666

667 668
		component_match_add_release(dev, &match, component_release_of,
					    component_compare_of, larbnode);
669 670 671 672
	}

	platform_set_drvdata(pdev, data);

673
	ret = mtk_iommu_v1_hw_init(data);
674 675 676
	if (ret)
		return ret;

677 678 679
	ret = iommu_device_sysfs_add(&data->iommu, &pdev->dev, NULL,
				     dev_name(&pdev->dev));
	if (ret)
680
		goto out_clk_unprepare;
681

682
	ret = iommu_device_register(&data->iommu, &mtk_iommu_v1_ops, dev);
683
	if (ret)
684
		goto out_sysfs_remove;
685

686
	ret = component_master_add_with_match(dev, &mtk_iommu_v1_com_ops, match);
687
	if (ret)
688
		goto out_dev_unreg;
689 690 691 692 693 694
	return ret;

out_dev_unreg:
	iommu_device_unregister(&data->iommu);
out_sysfs_remove:
	iommu_device_sysfs_remove(&data->iommu);
695 696
out_clk_unprepare:
	clk_disable_unprepare(data->bclk);
697
	return ret;
698 699
}

700
static void mtk_iommu_v1_remove(struct platform_device *pdev)
701
{
702
	struct mtk_iommu_v1_data *data = platform_get_drvdata(pdev);
703

704 705 706
	iommu_device_sysfs_remove(&data->iommu);
	iommu_device_unregister(&data->iommu);

707 708
	clk_disable_unprepare(data->bclk);
	devm_free_irq(&pdev->dev, data->irq, data);
709
	component_master_del(&pdev->dev, &mtk_iommu_v1_com_ops);
710 711
}

712
static int __maybe_unused mtk_iommu_v1_suspend(struct device *dev)
713
{
714 715
	struct mtk_iommu_v1_data *data = dev_get_drvdata(dev);
	struct mtk_iommu_v1_suspend_reg *reg = &data->reg;
716 717 718 719 720 721 722 723 724 725
	void __iomem *base = data->base;

	reg->standard_axi_mode = readl_relaxed(base +
					       REG_MMU_STANDARD_AXI_MODE);
	reg->dcm_dis = readl_relaxed(base + REG_MMU_DCM);
	reg->ctrl_reg = readl_relaxed(base + REG_MMU_CTRL_REG);
	reg->int_control0 = readl_relaxed(base + REG_MMU_INT_CONTROL);
	return 0;
}

726
static int __maybe_unused mtk_iommu_v1_resume(struct device *dev)
727
{
728 729
	struct mtk_iommu_v1_data *data = dev_get_drvdata(dev);
	struct mtk_iommu_v1_suspend_reg *reg = &data->reg;
730 731 732 733 734 735 736 737 738 739 740 741
	void __iomem *base = data->base;

	writel_relaxed(data->m4u_dom->pgt_pa, base + REG_MMU_PT_BASE_ADDR);
	writel_relaxed(reg->standard_axi_mode,
		       base + REG_MMU_STANDARD_AXI_MODE);
	writel_relaxed(reg->dcm_dis, base + REG_MMU_DCM);
	writel_relaxed(reg->ctrl_reg, base + REG_MMU_CTRL_REG);
	writel_relaxed(reg->int_control0, base + REG_MMU_INT_CONTROL);
	writel_relaxed(data->protect_base, base + REG_MMU_IVRP_PADDR);
	return 0;
}

742 743
static const struct dev_pm_ops mtk_iommu_v1_pm_ops = {
	SET_SYSTEM_SLEEP_PM_OPS(mtk_iommu_v1_suspend, mtk_iommu_v1_resume)
744 745
};

746 747
static struct platform_driver mtk_iommu_v1_driver = {
	.probe	= mtk_iommu_v1_probe,
748
	.remove_new = mtk_iommu_v1_remove,
749
	.driver	= {
750
		.name = "mtk-iommu-v1",
751 752
		.of_match_table = mtk_iommu_v1_of_ids,
		.pm = &mtk_iommu_v1_pm_ops,
753 754
	}
};
755
module_platform_driver(mtk_iommu_v1_driver);
756

757 758
MODULE_DESCRIPTION("IOMMU API for MediaTek M4U v1 implementations");
MODULE_LICENSE("GPL v2");