xref: /linux/drivers/peci/controller/peci-aspeed.c (revision c532de5a67a70f8533d495f8f2aaa9a0491c3ad0)
1 // SPDX-License-Identifier: GPL-2.0-only
2 // Copyright (c) 2012-2017 ASPEED Technology Inc.
3 // Copyright (c) 2018-2021 Intel Corporation
4 
5 #include <linux/unaligned.h>
6 
7 #include <linux/bitfield.h>
8 #include <linux/clk.h>
9 #include <linux/clkdev.h>
10 #include <linux/clk-provider.h>
11 #include <linux/delay.h>
12 #include <linux/interrupt.h>
13 #include <linux/io.h>
14 #include <linux/iopoll.h>
15 #include <linux/jiffies.h>
16 #include <linux/math.h>
17 #include <linux/module.h>
18 #include <linux/of.h>
19 #include <linux/peci.h>
20 #include <linux/platform_device.h>
21 #include <linux/reset.h>
22 
23 /* ASPEED PECI Registers */
24 /* Control Register */
25 #define ASPEED_PECI_CTRL			0x00
26 #define   ASPEED_PECI_CTRL_SAMPLING_MASK	GENMASK(19, 16)
27 #define   ASPEED_PECI_CTRL_RD_MODE_MASK		GENMASK(13, 12)
28 #define     ASPEED_PECI_CTRL_RD_MODE_DBG	BIT(13)
29 #define     ASPEED_PECI_CTRL_RD_MODE_COUNT	BIT(12)
30 #define   ASPEED_PECI_CTRL_CLK_SRC_HCLK		BIT(11)
31 #define   ASPEED_PECI_CTRL_CLK_DIV_MASK		GENMASK(10, 8)
32 #define   ASPEED_PECI_CTRL_INVERT_OUT		BIT(7)
33 #define   ASPEED_PECI_CTRL_INVERT_IN		BIT(6)
34 #define   ASPEED_PECI_CTRL_BUS_CONTENTION_EN	BIT(5)
35 #define   ASPEED_PECI_CTRL_PECI_EN		BIT(4)
36 #define   ASPEED_PECI_CTRL_PECI_CLK_EN		BIT(0)
37 
38 /* Timing Negotiation Register */
39 #define ASPEED_PECI_TIMING_NEGOTIATION		0x04
40 #define   ASPEED_PECI_T_NEGO_MSG_MASK		GENMASK(15, 8)
41 #define   ASPEED_PECI_T_NEGO_ADDR_MASK		GENMASK(7, 0)
42 
43 /* Command Register */
44 #define ASPEED_PECI_CMD				0x08
45 #define   ASPEED_PECI_CMD_PIN_MONITORING	BIT(31)
46 #define   ASPEED_PECI_CMD_STS_MASK		GENMASK(27, 24)
47 #define     ASPEED_PECI_CMD_STS_ADDR_T_NEGO	0x3
48 #define   ASPEED_PECI_CMD_IDLE_MASK		\
49 	  (ASPEED_PECI_CMD_STS_MASK | ASPEED_PECI_CMD_PIN_MONITORING)
50 #define   ASPEED_PECI_CMD_FIRE			BIT(0)
51 
52 /* Read/Write Length Register */
53 #define ASPEED_PECI_RW_LENGTH			0x0c
54 #define   ASPEED_PECI_AW_FCS_EN			BIT(31)
55 #define   ASPEED_PECI_RD_LEN_MASK		GENMASK(23, 16)
56 #define   ASPEED_PECI_WR_LEN_MASK		GENMASK(15, 8)
57 #define   ASPEED_PECI_TARGET_ADDR_MASK		GENMASK(7, 0)
58 
59 /* Expected FCS Data Register */
60 #define ASPEED_PECI_EXPECTED_FCS		0x10
61 #define   ASPEED_PECI_EXPECTED_RD_FCS_MASK	GENMASK(23, 16)
62 #define   ASPEED_PECI_EXPECTED_AW_FCS_AUTO_MASK	GENMASK(15, 8)
63 #define   ASPEED_PECI_EXPECTED_WR_FCS_MASK	GENMASK(7, 0)
64 
65 /* Captured FCS Data Register */
66 #define ASPEED_PECI_CAPTURED_FCS		0x14
67 #define   ASPEED_PECI_CAPTURED_RD_FCS_MASK	GENMASK(23, 16)
68 #define   ASPEED_PECI_CAPTURED_WR_FCS_MASK	GENMASK(7, 0)
69 
70 /* Interrupt Register */
71 #define ASPEED_PECI_INT_CTRL			0x18
72 #define   ASPEED_PECI_TIMING_NEGO_SEL_MASK	GENMASK(31, 30)
73 #define     ASPEED_PECI_1ST_BIT_OF_ADDR_NEGO	0
74 #define     ASPEED_PECI_2ND_BIT_OF_ADDR_NEGO	1
75 #define     ASPEED_PECI_MESSAGE_NEGO		2
76 #define   ASPEED_PECI_INT_MASK			GENMASK(4, 0)
77 #define     ASPEED_PECI_INT_BUS_TIMEOUT		BIT(4)
78 #define     ASPEED_PECI_INT_BUS_CONTENTION	BIT(3)
79 #define     ASPEED_PECI_INT_WR_FCS_BAD		BIT(2)
80 #define     ASPEED_PECI_INT_WR_FCS_ABORT	BIT(1)
81 #define     ASPEED_PECI_INT_CMD_DONE		BIT(0)
82 
83 /* Interrupt Status Register */
84 #define ASPEED_PECI_INT_STS			0x1c
85 #define   ASPEED_PECI_INT_TIMING_RESULT_MASK	GENMASK(29, 16)
86 	  /* bits[4..0]: Same bit fields in the 'Interrupt Register' */
87 
88 /* Rx/Tx Data Buffer Registers */
89 #define ASPEED_PECI_WR_DATA0			0x20
90 #define ASPEED_PECI_WR_DATA1			0x24
91 #define ASPEED_PECI_WR_DATA2			0x28
92 #define ASPEED_PECI_WR_DATA3			0x2c
93 #define ASPEED_PECI_RD_DATA0			0x30
94 #define ASPEED_PECI_RD_DATA1			0x34
95 #define ASPEED_PECI_RD_DATA2			0x38
96 #define ASPEED_PECI_RD_DATA3			0x3c
97 #define ASPEED_PECI_WR_DATA4			0x40
98 #define ASPEED_PECI_WR_DATA5			0x44
99 #define ASPEED_PECI_WR_DATA6			0x48
100 #define ASPEED_PECI_WR_DATA7			0x4c
101 #define ASPEED_PECI_RD_DATA4			0x50
102 #define ASPEED_PECI_RD_DATA5			0x54
103 #define ASPEED_PECI_RD_DATA6			0x58
104 #define ASPEED_PECI_RD_DATA7			0x5c
105 #define   ASPEED_PECI_DATA_BUF_SIZE_MAX		32
106 
107 /* Timing Negotiation */
108 #define ASPEED_PECI_CLK_FREQUENCY_MIN		2000
109 #define ASPEED_PECI_CLK_FREQUENCY_DEFAULT	1000000
110 #define ASPEED_PECI_CLK_FREQUENCY_MAX		2000000
111 #define ASPEED_PECI_RD_SAMPLING_POINT_DEFAULT	8
112 /* Timeout */
113 #define ASPEED_PECI_IDLE_CHECK_TIMEOUT_US	(50 * USEC_PER_MSEC)
114 #define ASPEED_PECI_IDLE_CHECK_INTERVAL_US	(10 * USEC_PER_MSEC)
115 #define ASPEED_PECI_CMD_TIMEOUT_MS_DEFAULT	1000
116 #define ASPEED_PECI_CMD_TIMEOUT_MS_MAX		1000
117 
118 #define ASPEED_PECI_CLK_DIV1(msg_timing) (4 * (msg_timing) + 1)
119 #define ASPEED_PECI_CLK_DIV2(clk_div_exp) BIT(clk_div_exp)
120 #define ASPEED_PECI_CLK_DIV(msg_timing, clk_div_exp) \
121 	(4 * ASPEED_PECI_CLK_DIV1(msg_timing) * ASPEED_PECI_CLK_DIV2(clk_div_exp))
122 
123 struct aspeed_peci {
124 	struct peci_controller *controller;
125 	struct device *dev;
126 	void __iomem *base;
127 	struct reset_control *rst;
128 	int irq;
129 	spinlock_t lock; /* to sync completion status handling */
130 	struct completion xfer_complete;
131 	struct clk *clk;
132 	u32 clk_frequency;
133 	u32 status;
134 	u32 cmd_timeout_ms;
135 };
136 
137 struct clk_aspeed_peci {
138 	struct clk_hw hw;
139 	struct aspeed_peci *aspeed_peci;
140 };
141 
142 static void aspeed_peci_controller_enable(struct aspeed_peci *priv)
143 {
144 	u32 val = readl(priv->base + ASPEED_PECI_CTRL);
145 
146 	val |= ASPEED_PECI_CTRL_PECI_CLK_EN;
147 	val |= ASPEED_PECI_CTRL_PECI_EN;
148 
149 	writel(val, priv->base + ASPEED_PECI_CTRL);
150 }
151 
152 static void aspeed_peci_init_regs(struct aspeed_peci *priv)
153 {
154 	u32 val;
155 
156 	/* Clear interrupts */
157 	writel(ASPEED_PECI_INT_MASK, priv->base + ASPEED_PECI_INT_STS);
158 
159 	/* Set timing negotiation mode and enable interrupts */
160 	val = FIELD_PREP(ASPEED_PECI_TIMING_NEGO_SEL_MASK, ASPEED_PECI_1ST_BIT_OF_ADDR_NEGO);
161 	val |= ASPEED_PECI_INT_MASK;
162 	writel(val, priv->base + ASPEED_PECI_INT_CTRL);
163 
164 	val = FIELD_PREP(ASPEED_PECI_CTRL_SAMPLING_MASK, ASPEED_PECI_RD_SAMPLING_POINT_DEFAULT);
165 	writel(val, priv->base + ASPEED_PECI_CTRL);
166 }
167 
168 static int aspeed_peci_check_idle(struct aspeed_peci *priv)
169 {
170 	u32 cmd_sts = readl(priv->base + ASPEED_PECI_CMD);
171 	int ret;
172 
173 	/*
174 	 * Under normal circumstances, we expect to be idle here.
175 	 * In case there were any errors/timeouts that led to the situation
176 	 * where the hardware is not in idle state - we need to reset and
177 	 * reinitialize it to avoid potential controller hang.
178 	 */
179 	if (FIELD_GET(ASPEED_PECI_CMD_STS_MASK, cmd_sts)) {
180 		ret = reset_control_assert(priv->rst);
181 		if (ret) {
182 			dev_err(priv->dev, "cannot assert reset control\n");
183 			return ret;
184 		}
185 
186 		ret = reset_control_deassert(priv->rst);
187 		if (ret) {
188 			dev_err(priv->dev, "cannot deassert reset control\n");
189 			return ret;
190 		}
191 
192 		aspeed_peci_init_regs(priv);
193 
194 		ret = clk_set_rate(priv->clk, priv->clk_frequency);
195 		if (ret < 0) {
196 			dev_err(priv->dev, "cannot set clock frequency\n");
197 			return ret;
198 		}
199 
200 		aspeed_peci_controller_enable(priv);
201 	}
202 
203 	return readl_poll_timeout(priv->base + ASPEED_PECI_CMD,
204 				  cmd_sts,
205 				  !(cmd_sts & ASPEED_PECI_CMD_IDLE_MASK),
206 				  ASPEED_PECI_IDLE_CHECK_INTERVAL_US,
207 				  ASPEED_PECI_IDLE_CHECK_TIMEOUT_US);
208 }
209 
210 static int aspeed_peci_xfer(struct peci_controller *controller,
211 			    u8 addr, struct peci_request *req)
212 {
213 	struct aspeed_peci *priv = dev_get_drvdata(controller->dev.parent);
214 	unsigned long timeout = msecs_to_jiffies(priv->cmd_timeout_ms);
215 	u32 peci_head;
216 	int ret, i;
217 
218 	if (req->tx.len > ASPEED_PECI_DATA_BUF_SIZE_MAX ||
219 	    req->rx.len > ASPEED_PECI_DATA_BUF_SIZE_MAX)
220 		return -EINVAL;
221 
222 	/* Check command sts and bus idle state */
223 	ret = aspeed_peci_check_idle(priv);
224 	if (ret)
225 		return ret; /* -ETIMEDOUT */
226 
227 	spin_lock_irq(&priv->lock);
228 	reinit_completion(&priv->xfer_complete);
229 
230 	peci_head = FIELD_PREP(ASPEED_PECI_TARGET_ADDR_MASK, addr) |
231 		    FIELD_PREP(ASPEED_PECI_WR_LEN_MASK, req->tx.len) |
232 		    FIELD_PREP(ASPEED_PECI_RD_LEN_MASK, req->rx.len);
233 
234 	writel(peci_head, priv->base + ASPEED_PECI_RW_LENGTH);
235 
236 	for (i = 0; i < req->tx.len; i += 4) {
237 		u32 reg = (i < 16 ? ASPEED_PECI_WR_DATA0 : ASPEED_PECI_WR_DATA4) + i % 16;
238 
239 		writel(get_unaligned_le32(&req->tx.buf[i]), priv->base + reg);
240 	}
241 
242 #if IS_ENABLED(CONFIG_DYNAMIC_DEBUG)
243 	dev_dbg(priv->dev, "HEAD : %#08x\n", peci_head);
244 	print_hex_dump_bytes("TX : ", DUMP_PREFIX_NONE, req->tx.buf, req->tx.len);
245 #endif
246 
247 	priv->status = 0;
248 	writel(ASPEED_PECI_CMD_FIRE, priv->base + ASPEED_PECI_CMD);
249 	spin_unlock_irq(&priv->lock);
250 
251 	ret = wait_for_completion_interruptible_timeout(&priv->xfer_complete, timeout);
252 	if (ret < 0)
253 		return ret;
254 
255 	if (ret == 0) {
256 		dev_dbg(priv->dev, "timeout waiting for a response\n");
257 		return -ETIMEDOUT;
258 	}
259 
260 	spin_lock_irq(&priv->lock);
261 
262 	if (priv->status != ASPEED_PECI_INT_CMD_DONE) {
263 		spin_unlock_irq(&priv->lock);
264 		dev_dbg(priv->dev, "no valid response, status: %#02x\n", priv->status);
265 		return -EIO;
266 	}
267 
268 	spin_unlock_irq(&priv->lock);
269 
270 	/*
271 	 * We need to use dword reads for register access, make sure that the
272 	 * buffer size is multiple of 4-bytes.
273 	 */
274 	BUILD_BUG_ON(PECI_REQUEST_MAX_BUF_SIZE % 4);
275 
276 	for (i = 0; i < req->rx.len; i += 4) {
277 		u32 reg = (i < 16 ? ASPEED_PECI_RD_DATA0 : ASPEED_PECI_RD_DATA4) + i % 16;
278 		u32 rx_data = readl(priv->base + reg);
279 
280 		put_unaligned_le32(rx_data, &req->rx.buf[i]);
281 	}
282 
283 #if IS_ENABLED(CONFIG_DYNAMIC_DEBUG)
284 	print_hex_dump_bytes("RX : ", DUMP_PREFIX_NONE, req->rx.buf, req->rx.len);
285 #endif
286 	return 0;
287 }
288 
289 static irqreturn_t aspeed_peci_irq_handler(int irq, void *arg)
290 {
291 	struct aspeed_peci *priv = arg;
292 	u32 status;
293 
294 	spin_lock(&priv->lock);
295 	status = readl(priv->base + ASPEED_PECI_INT_STS);
296 	writel(status, priv->base + ASPEED_PECI_INT_STS);
297 	priv->status |= (status & ASPEED_PECI_INT_MASK);
298 
299 	/*
300 	 * All commands should be ended up with a ASPEED_PECI_INT_CMD_DONE bit
301 	 * set even in an error case.
302 	 */
303 	if (status & ASPEED_PECI_INT_CMD_DONE)
304 		complete(&priv->xfer_complete);
305 
306 	writel(0, priv->base + ASPEED_PECI_CMD);
307 
308 	spin_unlock(&priv->lock);
309 
310 	return IRQ_HANDLED;
311 }
312 
313 static void clk_aspeed_peci_find_div_values(unsigned long rate, int *msg_timing, int *clk_div_exp)
314 {
315 	unsigned long best_diff = ~0ul, diff;
316 	int msg_timing_temp, clk_div_exp_temp, i, j;
317 
318 	for (i = 1; i <= 255; i++)
319 		for (j = 0; j < 8; j++) {
320 			diff = abs(rate - ASPEED_PECI_CLK_DIV1(i) * ASPEED_PECI_CLK_DIV2(j));
321 			if (diff < best_diff) {
322 				msg_timing_temp = i;
323 				clk_div_exp_temp = j;
324 				best_diff = diff;
325 			}
326 		}
327 
328 	*msg_timing = msg_timing_temp;
329 	*clk_div_exp = clk_div_exp_temp;
330 }
331 
332 static int clk_aspeed_peci_get_div(unsigned long rate, const unsigned long *prate)
333 {
334 	unsigned long this_rate = *prate / (4 * rate);
335 	int msg_timing, clk_div_exp;
336 
337 	clk_aspeed_peci_find_div_values(this_rate, &msg_timing, &clk_div_exp);
338 
339 	return ASPEED_PECI_CLK_DIV(msg_timing, clk_div_exp);
340 }
341 
342 static int clk_aspeed_peci_set_rate(struct clk_hw *hw, unsigned long rate,
343 				    unsigned long prate)
344 {
345 	struct clk_aspeed_peci *peci_clk = container_of(hw, struct clk_aspeed_peci, hw);
346 	struct aspeed_peci *aspeed_peci = peci_clk->aspeed_peci;
347 	unsigned long this_rate = prate / (4 * rate);
348 	int clk_div_exp, msg_timing;
349 	u32 val;
350 
351 	clk_aspeed_peci_find_div_values(this_rate, &msg_timing, &clk_div_exp);
352 
353 	val = readl(aspeed_peci->base + ASPEED_PECI_CTRL);
354 	val &= ~ASPEED_PECI_CTRL_CLK_DIV_MASK;
355 	val |= FIELD_PREP(ASPEED_PECI_CTRL_CLK_DIV_MASK, clk_div_exp);
356 	writel(val, aspeed_peci->base + ASPEED_PECI_CTRL);
357 
358 	val = FIELD_PREP(ASPEED_PECI_T_NEGO_MSG_MASK, msg_timing);
359 	val |= FIELD_PREP(ASPEED_PECI_T_NEGO_ADDR_MASK, msg_timing);
360 	writel(val, aspeed_peci->base + ASPEED_PECI_TIMING_NEGOTIATION);
361 
362 	return 0;
363 }
364 
365 static long clk_aspeed_peci_round_rate(struct clk_hw *hw, unsigned long rate,
366 				       unsigned long *prate)
367 {
368 	int div = clk_aspeed_peci_get_div(rate, prate);
369 
370 	return DIV_ROUND_UP_ULL(*prate, div);
371 }
372 
373 static unsigned long clk_aspeed_peci_recalc_rate(struct clk_hw *hw, unsigned long prate)
374 {
375 	struct clk_aspeed_peci *peci_clk = container_of(hw, struct clk_aspeed_peci, hw);
376 	struct aspeed_peci *aspeed_peci = peci_clk->aspeed_peci;
377 	int div, msg_timing, addr_timing, clk_div_exp;
378 	u32 reg;
379 
380 	reg = readl(aspeed_peci->base + ASPEED_PECI_TIMING_NEGOTIATION);
381 	msg_timing = FIELD_GET(ASPEED_PECI_T_NEGO_MSG_MASK, reg);
382 	addr_timing = FIELD_GET(ASPEED_PECI_T_NEGO_ADDR_MASK, reg);
383 
384 	if (msg_timing != addr_timing)
385 		return 0;
386 
387 	reg = readl(aspeed_peci->base + ASPEED_PECI_CTRL);
388 	clk_div_exp = FIELD_GET(ASPEED_PECI_CTRL_CLK_DIV_MASK, reg);
389 
390 	div = ASPEED_PECI_CLK_DIV(msg_timing, clk_div_exp);
391 
392 	return DIV_ROUND_UP_ULL(prate, div);
393 }
394 
395 static const struct clk_ops clk_aspeed_peci_ops = {
396 	.set_rate = clk_aspeed_peci_set_rate,
397 	.round_rate = clk_aspeed_peci_round_rate,
398 	.recalc_rate = clk_aspeed_peci_recalc_rate,
399 };
400 
401 /*
402  * PECI HW contains a clock divider which is a combination of:
403  *  div0: 4 (fixed divider)
404  *  div1: x + 1
405  *  div2: 1 << y
406  * In other words, out_clk = in_clk / (div0 * div1 * div2)
407  * The resulting frequency is used by PECI Controller to drive the PECI bus to
408  * negotiate optimal transfer rate.
409  */
410 static struct clk *devm_aspeed_peci_register_clk_div(struct device *dev, struct clk *parent,
411 						     struct aspeed_peci *priv)
412 {
413 	struct clk_aspeed_peci *peci_clk;
414 	struct clk_init_data init;
415 	const char *parent_name;
416 	char name[32];
417 	int ret;
418 
419 	snprintf(name, sizeof(name), "%s_div", dev_name(dev));
420 
421 	parent_name = __clk_get_name(parent);
422 
423 	init.ops = &clk_aspeed_peci_ops;
424 	init.name = name;
425 	init.parent_names = (const char* []) { parent_name };
426 	init.num_parents = 1;
427 	init.flags = 0;
428 
429 	peci_clk = devm_kzalloc(dev, sizeof(struct clk_aspeed_peci), GFP_KERNEL);
430 	if (!peci_clk)
431 		return ERR_PTR(-ENOMEM);
432 
433 	peci_clk->hw.init = &init;
434 	peci_clk->aspeed_peci = priv;
435 
436 	ret = devm_clk_hw_register(dev, &peci_clk->hw);
437 	if (ret)
438 		return ERR_PTR(ret);
439 
440 	return peci_clk->hw.clk;
441 }
442 
443 static void aspeed_peci_property_sanitize(struct device *dev, const char *propname,
444 					  u32 min, u32 max, u32 default_val, u32 *propval)
445 {
446 	u32 val;
447 	int ret;
448 
449 	ret = device_property_read_u32(dev, propname, &val);
450 	if (ret) {
451 		val = default_val;
452 	} else if (val > max || val < min) {
453 		dev_warn(dev, "invalid %s: %u, falling back to: %u\n",
454 			 propname, val, default_val);
455 
456 		val = default_val;
457 	}
458 
459 	*propval = val;
460 }
461 
462 static void aspeed_peci_property_setup(struct aspeed_peci *priv)
463 {
464 	aspeed_peci_property_sanitize(priv->dev, "clock-frequency",
465 				      ASPEED_PECI_CLK_FREQUENCY_MIN, ASPEED_PECI_CLK_FREQUENCY_MAX,
466 				      ASPEED_PECI_CLK_FREQUENCY_DEFAULT, &priv->clk_frequency);
467 	aspeed_peci_property_sanitize(priv->dev, "cmd-timeout-ms",
468 				      1, ASPEED_PECI_CMD_TIMEOUT_MS_MAX,
469 				      ASPEED_PECI_CMD_TIMEOUT_MS_DEFAULT, &priv->cmd_timeout_ms);
470 }
471 
472 static const struct peci_controller_ops aspeed_ops = {
473 	.xfer = aspeed_peci_xfer,
474 };
475 
476 static void aspeed_peci_reset_control_release(void *data)
477 {
478 	reset_control_assert(data);
479 }
480 
481 static int devm_aspeed_peci_reset_control_deassert(struct device *dev, struct reset_control *rst)
482 {
483 	int ret;
484 
485 	ret = reset_control_deassert(rst);
486 	if (ret)
487 		return ret;
488 
489 	return devm_add_action_or_reset(dev, aspeed_peci_reset_control_release, rst);
490 }
491 
492 static void aspeed_peci_clk_release(void *data)
493 {
494 	clk_disable_unprepare(data);
495 }
496 
497 static int devm_aspeed_peci_clk_enable(struct device *dev, struct clk *clk)
498 {
499 	int ret;
500 
501 	ret = clk_prepare_enable(clk);
502 	if (ret)
503 		return ret;
504 
505 	return devm_add_action_or_reset(dev, aspeed_peci_clk_release, clk);
506 }
507 
508 static int aspeed_peci_probe(struct platform_device *pdev)
509 {
510 	struct peci_controller *controller;
511 	struct aspeed_peci *priv;
512 	struct clk *ref_clk;
513 	int ret;
514 
515 	priv = devm_kzalloc(&pdev->dev, sizeof(*priv), GFP_KERNEL);
516 	if (!priv)
517 		return -ENOMEM;
518 
519 	priv->dev = &pdev->dev;
520 	dev_set_drvdata(priv->dev, priv);
521 
522 	priv->base = devm_platform_ioremap_resource(pdev, 0);
523 	if (IS_ERR(priv->base))
524 		return PTR_ERR(priv->base);
525 
526 	priv->irq = platform_get_irq(pdev, 0);
527 	if (priv->irq < 0)
528 		return priv->irq;
529 
530 	ret = devm_request_irq(&pdev->dev, priv->irq, aspeed_peci_irq_handler,
531 			       0, "peci-aspeed", priv);
532 	if (ret)
533 		return ret;
534 
535 	init_completion(&priv->xfer_complete);
536 	spin_lock_init(&priv->lock);
537 
538 	priv->rst = devm_reset_control_get(&pdev->dev, NULL);
539 	if (IS_ERR(priv->rst))
540 		return dev_err_probe(priv->dev, PTR_ERR(priv->rst),
541 				     "failed to get reset control\n");
542 
543 	ret = devm_aspeed_peci_reset_control_deassert(priv->dev, priv->rst);
544 	if (ret)
545 		return dev_err_probe(priv->dev, ret, "cannot deassert reset control\n");
546 
547 	aspeed_peci_property_setup(priv);
548 
549 	aspeed_peci_init_regs(priv);
550 
551 	ref_clk = devm_clk_get(priv->dev, NULL);
552 	if (IS_ERR(ref_clk))
553 		return dev_err_probe(priv->dev, PTR_ERR(ref_clk), "failed to get ref clock\n");
554 
555 	priv->clk = devm_aspeed_peci_register_clk_div(priv->dev, ref_clk, priv);
556 	if (IS_ERR(priv->clk))
557 		return dev_err_probe(priv->dev, PTR_ERR(priv->clk), "cannot register clock\n");
558 
559 	ret = clk_set_rate(priv->clk, priv->clk_frequency);
560 	if (ret < 0)
561 		return dev_err_probe(priv->dev, ret, "cannot set clock frequency\n");
562 
563 	ret = devm_aspeed_peci_clk_enable(priv->dev, priv->clk);
564 	if (ret)
565 		return dev_err_probe(priv->dev, ret, "failed to enable clock\n");
566 
567 	aspeed_peci_controller_enable(priv);
568 
569 	controller = devm_peci_controller_add(priv->dev, &aspeed_ops);
570 	if (IS_ERR(controller))
571 		return dev_err_probe(priv->dev, PTR_ERR(controller),
572 				     "failed to add aspeed peci controller\n");
573 
574 	priv->controller = controller;
575 
576 	return 0;
577 }
578 
579 static const struct of_device_id aspeed_peci_of_table[] = {
580 	{ .compatible = "aspeed,ast2400-peci", },
581 	{ .compatible = "aspeed,ast2500-peci", },
582 	{ .compatible = "aspeed,ast2600-peci", },
583 	{ }
584 };
585 MODULE_DEVICE_TABLE(of, aspeed_peci_of_table);
586 
587 static struct platform_driver aspeed_peci_driver = {
588 	.probe  = aspeed_peci_probe,
589 	.driver = {
590 		.name           = "peci-aspeed",
591 		.of_match_table = aspeed_peci_of_table,
592 	},
593 };
594 module_platform_driver(aspeed_peci_driver);
595 
596 MODULE_AUTHOR("Ryan Chen <ryan_chen@aspeedtech.com>");
597 MODULE_AUTHOR("Jae Hyun Yoo <jae.hyun.yoo@linux.intel.com>");
598 MODULE_DESCRIPTION("ASPEED PECI driver");
599 MODULE_LICENSE("GPL");
600 MODULE_IMPORT_NS(PECI);
601