1 /* 2 * drivers/w1/masters/omap_hdq.c 3 * 4 * Copyright (C) 2007,2012 Texas Instruments, Inc. 5 * 6 * This file is licensed under the terms of the GNU General Public License 7 * version 2. This program is licensed "as is" without any warranty of any 8 * kind, whether express or implied. 9 * 10 */ 11 #include <linux/kernel.h> 12 #include <linux/module.h> 13 #include <linux/platform_device.h> 14 #include <linux/interrupt.h> 15 #include <linux/slab.h> 16 #include <linux/err.h> 17 #include <linux/io.h> 18 #include <linux/sched.h> 19 #include <linux/pm_runtime.h> 20 21 #include "../w1.h" 22 #include "../w1_int.h" 23 24 #define MOD_NAME "OMAP_HDQ:" 25 26 #define OMAP_HDQ_REVISION 0x00 27 #define OMAP_HDQ_TX_DATA 0x04 28 #define OMAP_HDQ_RX_DATA 0x08 29 #define OMAP_HDQ_CTRL_STATUS 0x0c 30 #define OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK (1<<6) 31 #define OMAP_HDQ_CTRL_STATUS_CLOCKENABLE (1<<5) 32 #define OMAP_HDQ_CTRL_STATUS_GO (1<<4) 33 #define OMAP_HDQ_CTRL_STATUS_INITIALIZATION (1<<2) 34 #define OMAP_HDQ_CTRL_STATUS_DIR (1<<1) 35 #define OMAP_HDQ_CTRL_STATUS_MODE (1<<0) 36 #define OMAP_HDQ_INT_STATUS 0x10 37 #define OMAP_HDQ_INT_STATUS_TXCOMPLETE (1<<2) 38 #define OMAP_HDQ_INT_STATUS_RXCOMPLETE (1<<1) 39 #define OMAP_HDQ_INT_STATUS_TIMEOUT (1<<0) 40 #define OMAP_HDQ_SYSCONFIG 0x14 41 #define OMAP_HDQ_SYSCONFIG_SOFTRESET (1<<1) 42 #define OMAP_HDQ_SYSCONFIG_AUTOIDLE (1<<0) 43 #define OMAP_HDQ_SYSSTATUS 0x18 44 #define OMAP_HDQ_SYSSTATUS_RESETDONE (1<<0) 45 46 #define OMAP_HDQ_FLAG_CLEAR 0 47 #define OMAP_HDQ_FLAG_SET 1 48 #define OMAP_HDQ_TIMEOUT (HZ/5) 49 50 #define OMAP_HDQ_MAX_USER 4 51 52 static DECLARE_WAIT_QUEUE_HEAD(hdq_wait_queue); 53 static int w1_id; 54 55 struct hdq_data { 56 struct device *dev; 57 void __iomem *hdq_base; 58 /* lock status update */ 59 struct mutex hdq_mutex; 60 int hdq_usecount; 61 u8 hdq_irqstatus; 62 /* device lock */ 63 spinlock_t hdq_spinlock; 64 /* 65 * Used to control the call to omap_hdq_get and omap_hdq_put. 66 * HDQ Protocol: Write the CMD|REG_address first, followed by 67 * the data wrire or read. 68 */ 69 int init_trans; 70 }; 71 72 static int omap_hdq_probe(struct platform_device *pdev); 73 static int omap_hdq_remove(struct platform_device *pdev); 74 75 static struct of_device_id omap_hdq_dt_ids[] = { 76 { .compatible = "ti,omap3-1w" }, 77 {} 78 }; 79 MODULE_DEVICE_TABLE(of, omap_hdq_dt_ids); 80 81 static struct platform_driver omap_hdq_driver = { 82 .probe = omap_hdq_probe, 83 .remove = omap_hdq_remove, 84 .driver = { 85 .name = "omap_hdq", 86 .of_match_table = omap_hdq_dt_ids, 87 }, 88 }; 89 90 static u8 omap_w1_read_byte(void *_hdq); 91 static void omap_w1_write_byte(void *_hdq, u8 byte); 92 static u8 omap_w1_reset_bus(void *_hdq); 93 static void omap_w1_search_bus(void *_hdq, struct w1_master *master_dev, 94 u8 search_type, w1_slave_found_callback slave_found); 95 96 97 static struct w1_bus_master omap_w1_master = { 98 .read_byte = omap_w1_read_byte, 99 .write_byte = omap_w1_write_byte, 100 .reset_bus = omap_w1_reset_bus, 101 .search = omap_w1_search_bus, 102 }; 103 104 /* HDQ register I/O routines */ 105 static inline u8 hdq_reg_in(struct hdq_data *hdq_data, u32 offset) 106 { 107 return __raw_readl(hdq_data->hdq_base + offset); 108 } 109 110 static inline void hdq_reg_out(struct hdq_data *hdq_data, u32 offset, u8 val) 111 { 112 __raw_writel(val, hdq_data->hdq_base + offset); 113 } 114 115 static inline u8 hdq_reg_merge(struct hdq_data *hdq_data, u32 offset, 116 u8 val, u8 mask) 117 { 118 u8 new_val = (__raw_readl(hdq_data->hdq_base + offset) & ~mask) 119 | (val & mask); 120 __raw_writel(new_val, hdq_data->hdq_base + offset); 121 122 return new_val; 123 } 124 125 /* 126 * Wait for one or more bits in flag change. 127 * HDQ_FLAG_SET: wait until any bit in the flag is set. 128 * HDQ_FLAG_CLEAR: wait until all bits in the flag are cleared. 129 * return 0 on success and -ETIMEDOUT in the case of timeout. 130 */ 131 static int hdq_wait_for_flag(struct hdq_data *hdq_data, u32 offset, 132 u8 flag, u8 flag_set, u8 *status) 133 { 134 int ret = 0; 135 unsigned long timeout = jiffies + OMAP_HDQ_TIMEOUT; 136 137 if (flag_set == OMAP_HDQ_FLAG_CLEAR) { 138 /* wait for the flag clear */ 139 while (((*status = hdq_reg_in(hdq_data, offset)) & flag) 140 && time_before(jiffies, timeout)) { 141 schedule_timeout_uninterruptible(1); 142 } 143 if (*status & flag) 144 ret = -ETIMEDOUT; 145 } else if (flag_set == OMAP_HDQ_FLAG_SET) { 146 /* wait for the flag set */ 147 while (!((*status = hdq_reg_in(hdq_data, offset)) & flag) 148 && time_before(jiffies, timeout)) { 149 schedule_timeout_uninterruptible(1); 150 } 151 if (!(*status & flag)) 152 ret = -ETIMEDOUT; 153 } else 154 return -EINVAL; 155 156 return ret; 157 } 158 159 /* write out a byte and fill *status with HDQ_INT_STATUS */ 160 static int hdq_write_byte(struct hdq_data *hdq_data, u8 val, u8 *status) 161 { 162 int ret; 163 u8 tmp_status; 164 unsigned long irqflags; 165 166 *status = 0; 167 168 spin_lock_irqsave(&hdq_data->hdq_spinlock, irqflags); 169 /* clear interrupt flags via a dummy read */ 170 hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS); 171 /* ISR loads it with new INT_STATUS */ 172 hdq_data->hdq_irqstatus = 0; 173 spin_unlock_irqrestore(&hdq_data->hdq_spinlock, irqflags); 174 175 hdq_reg_out(hdq_data, OMAP_HDQ_TX_DATA, val); 176 177 /* set the GO bit */ 178 hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS, OMAP_HDQ_CTRL_STATUS_GO, 179 OMAP_HDQ_CTRL_STATUS_DIR | OMAP_HDQ_CTRL_STATUS_GO); 180 /* wait for the TXCOMPLETE bit */ 181 ret = wait_event_timeout(hdq_wait_queue, 182 hdq_data->hdq_irqstatus, OMAP_HDQ_TIMEOUT); 183 if (ret == 0) { 184 dev_dbg(hdq_data->dev, "TX wait elapsed\n"); 185 ret = -ETIMEDOUT; 186 goto out; 187 } 188 189 *status = hdq_data->hdq_irqstatus; 190 /* check irqstatus */ 191 if (!(*status & OMAP_HDQ_INT_STATUS_TXCOMPLETE)) { 192 dev_dbg(hdq_data->dev, "timeout waiting for" 193 " TXCOMPLETE/RXCOMPLETE, %x", *status); 194 ret = -ETIMEDOUT; 195 goto out; 196 } 197 198 /* wait for the GO bit return to zero */ 199 ret = hdq_wait_for_flag(hdq_data, OMAP_HDQ_CTRL_STATUS, 200 OMAP_HDQ_CTRL_STATUS_GO, 201 OMAP_HDQ_FLAG_CLEAR, &tmp_status); 202 if (ret) { 203 dev_dbg(hdq_data->dev, "timeout waiting GO bit" 204 " return to zero, %x", tmp_status); 205 } 206 207 out: 208 return ret; 209 } 210 211 /* HDQ Interrupt service routine */ 212 static irqreturn_t hdq_isr(int irq, void *_hdq) 213 { 214 struct hdq_data *hdq_data = _hdq; 215 unsigned long irqflags; 216 217 spin_lock_irqsave(&hdq_data->hdq_spinlock, irqflags); 218 hdq_data->hdq_irqstatus = hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS); 219 spin_unlock_irqrestore(&hdq_data->hdq_spinlock, irqflags); 220 dev_dbg(hdq_data->dev, "hdq_isr: %x", hdq_data->hdq_irqstatus); 221 222 if (hdq_data->hdq_irqstatus & 223 (OMAP_HDQ_INT_STATUS_TXCOMPLETE | OMAP_HDQ_INT_STATUS_RXCOMPLETE 224 | OMAP_HDQ_INT_STATUS_TIMEOUT)) { 225 /* wake up sleeping process */ 226 wake_up(&hdq_wait_queue); 227 } 228 229 return IRQ_HANDLED; 230 } 231 232 /* HDQ Mode: always return success */ 233 static u8 omap_w1_reset_bus(void *_hdq) 234 { 235 return 0; 236 } 237 238 /* W1 search callback function */ 239 static void omap_w1_search_bus(void *_hdq, struct w1_master *master_dev, 240 u8 search_type, w1_slave_found_callback slave_found) 241 { 242 u64 module_id, rn_le, cs, id; 243 244 if (w1_id) 245 module_id = w1_id; 246 else 247 module_id = 0x1; 248 249 rn_le = cpu_to_le64(module_id); 250 /* 251 * HDQ might not obey truly the 1-wire spec. 252 * So calculate CRC based on module parameter. 253 */ 254 cs = w1_calc_crc8((u8 *)&rn_le, 7); 255 id = (cs << 56) | module_id; 256 257 slave_found(master_dev, id); 258 } 259 260 static int _omap_hdq_reset(struct hdq_data *hdq_data) 261 { 262 int ret; 263 u8 tmp_status; 264 265 hdq_reg_out(hdq_data, OMAP_HDQ_SYSCONFIG, OMAP_HDQ_SYSCONFIG_SOFTRESET); 266 /* 267 * Select HDQ mode & enable clocks. 268 * It is observed that INT flags can't be cleared via a read and GO/INIT 269 * won't return to zero if interrupt is disabled. So we always enable 270 * interrupt. 271 */ 272 hdq_reg_out(hdq_data, OMAP_HDQ_CTRL_STATUS, 273 OMAP_HDQ_CTRL_STATUS_CLOCKENABLE | 274 OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK); 275 276 /* wait for reset to complete */ 277 ret = hdq_wait_for_flag(hdq_data, OMAP_HDQ_SYSSTATUS, 278 OMAP_HDQ_SYSSTATUS_RESETDONE, OMAP_HDQ_FLAG_SET, &tmp_status); 279 if (ret) 280 dev_dbg(hdq_data->dev, "timeout waiting HDQ reset, %x", 281 tmp_status); 282 else { 283 hdq_reg_out(hdq_data, OMAP_HDQ_CTRL_STATUS, 284 OMAP_HDQ_CTRL_STATUS_CLOCKENABLE | 285 OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK); 286 hdq_reg_out(hdq_data, OMAP_HDQ_SYSCONFIG, 287 OMAP_HDQ_SYSCONFIG_AUTOIDLE); 288 } 289 290 return ret; 291 } 292 293 /* Issue break pulse to the device */ 294 static int omap_hdq_break(struct hdq_data *hdq_data) 295 { 296 int ret = 0; 297 u8 tmp_status; 298 unsigned long irqflags; 299 300 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex); 301 if (ret < 0) { 302 dev_dbg(hdq_data->dev, "Could not acquire mutex\n"); 303 ret = -EINTR; 304 goto rtn; 305 } 306 307 spin_lock_irqsave(&hdq_data->hdq_spinlock, irqflags); 308 /* clear interrupt flags via a dummy read */ 309 hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS); 310 /* ISR loads it with new INT_STATUS */ 311 hdq_data->hdq_irqstatus = 0; 312 spin_unlock_irqrestore(&hdq_data->hdq_spinlock, irqflags); 313 314 /* set the INIT and GO bit */ 315 hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS, 316 OMAP_HDQ_CTRL_STATUS_INITIALIZATION | OMAP_HDQ_CTRL_STATUS_GO, 317 OMAP_HDQ_CTRL_STATUS_DIR | OMAP_HDQ_CTRL_STATUS_INITIALIZATION | 318 OMAP_HDQ_CTRL_STATUS_GO); 319 320 /* wait for the TIMEOUT bit */ 321 ret = wait_event_timeout(hdq_wait_queue, 322 hdq_data->hdq_irqstatus, OMAP_HDQ_TIMEOUT); 323 if (ret == 0) { 324 dev_dbg(hdq_data->dev, "break wait elapsed\n"); 325 ret = -EINTR; 326 goto out; 327 } 328 329 tmp_status = hdq_data->hdq_irqstatus; 330 /* check irqstatus */ 331 if (!(tmp_status & OMAP_HDQ_INT_STATUS_TIMEOUT)) { 332 dev_dbg(hdq_data->dev, "timeout waiting for TIMEOUT, %x", 333 tmp_status); 334 ret = -ETIMEDOUT; 335 goto out; 336 } 337 /* 338 * wait for both INIT and GO bits rerurn to zero. 339 * zero wait time expected for interrupt mode. 340 */ 341 ret = hdq_wait_for_flag(hdq_data, OMAP_HDQ_CTRL_STATUS, 342 OMAP_HDQ_CTRL_STATUS_INITIALIZATION | 343 OMAP_HDQ_CTRL_STATUS_GO, OMAP_HDQ_FLAG_CLEAR, 344 &tmp_status); 345 if (ret) 346 dev_dbg(hdq_data->dev, "timeout waiting INIT&GO bits" 347 " return to zero, %x", tmp_status); 348 349 out: 350 mutex_unlock(&hdq_data->hdq_mutex); 351 rtn: 352 return ret; 353 } 354 355 static int hdq_read_byte(struct hdq_data *hdq_data, u8 *val) 356 { 357 int ret = 0; 358 u8 status; 359 360 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex); 361 if (ret < 0) { 362 ret = -EINTR; 363 goto rtn; 364 } 365 366 if (!hdq_data->hdq_usecount) { 367 ret = -EINVAL; 368 goto out; 369 } 370 371 if (!(hdq_data->hdq_irqstatus & OMAP_HDQ_INT_STATUS_RXCOMPLETE)) { 372 hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS, 373 OMAP_HDQ_CTRL_STATUS_DIR | OMAP_HDQ_CTRL_STATUS_GO, 374 OMAP_HDQ_CTRL_STATUS_DIR | OMAP_HDQ_CTRL_STATUS_GO); 375 /* 376 * The RX comes immediately after TX. 377 */ 378 wait_event_timeout(hdq_wait_queue, 379 (hdq_data->hdq_irqstatus 380 & OMAP_HDQ_INT_STATUS_RXCOMPLETE), 381 OMAP_HDQ_TIMEOUT); 382 383 hdq_reg_merge(hdq_data, OMAP_HDQ_CTRL_STATUS, 0, 384 OMAP_HDQ_CTRL_STATUS_DIR); 385 status = hdq_data->hdq_irqstatus; 386 /* check irqstatus */ 387 if (!(status & OMAP_HDQ_INT_STATUS_RXCOMPLETE)) { 388 dev_dbg(hdq_data->dev, "timeout waiting for" 389 " RXCOMPLETE, %x", status); 390 ret = -ETIMEDOUT; 391 goto out; 392 } 393 } 394 /* the data is ready. Read it in! */ 395 *val = hdq_reg_in(hdq_data, OMAP_HDQ_RX_DATA); 396 out: 397 mutex_unlock(&hdq_data->hdq_mutex); 398 rtn: 399 return ret; 400 401 } 402 403 /* Enable clocks and set the controller to HDQ mode */ 404 static int omap_hdq_get(struct hdq_data *hdq_data) 405 { 406 int ret = 0; 407 408 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex); 409 if (ret < 0) { 410 ret = -EINTR; 411 goto rtn; 412 } 413 414 if (OMAP_HDQ_MAX_USER == hdq_data->hdq_usecount) { 415 dev_dbg(hdq_data->dev, "attempt to exceed the max use count"); 416 ret = -EINVAL; 417 goto out; 418 } else { 419 hdq_data->hdq_usecount++; 420 try_module_get(THIS_MODULE); 421 if (1 == hdq_data->hdq_usecount) { 422 423 pm_runtime_get_sync(hdq_data->dev); 424 425 /* make sure HDQ is out of reset */ 426 if (!(hdq_reg_in(hdq_data, OMAP_HDQ_SYSSTATUS) & 427 OMAP_HDQ_SYSSTATUS_RESETDONE)) { 428 ret = _omap_hdq_reset(hdq_data); 429 if (ret) 430 /* back up the count */ 431 hdq_data->hdq_usecount--; 432 } else { 433 /* select HDQ mode & enable clocks */ 434 hdq_reg_out(hdq_data, OMAP_HDQ_CTRL_STATUS, 435 OMAP_HDQ_CTRL_STATUS_CLOCKENABLE | 436 OMAP_HDQ_CTRL_STATUS_INTERRUPTMASK); 437 hdq_reg_out(hdq_data, OMAP_HDQ_SYSCONFIG, 438 OMAP_HDQ_SYSCONFIG_AUTOIDLE); 439 hdq_reg_in(hdq_data, OMAP_HDQ_INT_STATUS); 440 } 441 } 442 } 443 444 out: 445 mutex_unlock(&hdq_data->hdq_mutex); 446 rtn: 447 return ret; 448 } 449 450 /* Disable clocks to the module */ 451 static int omap_hdq_put(struct hdq_data *hdq_data) 452 { 453 int ret = 0; 454 455 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex); 456 if (ret < 0) 457 return -EINTR; 458 459 if (0 == hdq_data->hdq_usecount) { 460 dev_dbg(hdq_data->dev, "attempt to decrement use count" 461 " when it is zero"); 462 ret = -EINVAL; 463 } else { 464 hdq_data->hdq_usecount--; 465 module_put(THIS_MODULE); 466 if (0 == hdq_data->hdq_usecount) 467 pm_runtime_put_sync(hdq_data->dev); 468 } 469 mutex_unlock(&hdq_data->hdq_mutex); 470 471 return ret; 472 } 473 474 /* Read a byte of data from the device */ 475 static u8 omap_w1_read_byte(void *_hdq) 476 { 477 struct hdq_data *hdq_data = _hdq; 478 u8 val = 0; 479 int ret; 480 481 ret = hdq_read_byte(hdq_data, &val); 482 if (ret) { 483 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex); 484 if (ret < 0) { 485 dev_dbg(hdq_data->dev, "Could not acquire mutex\n"); 486 return -EINTR; 487 } 488 hdq_data->init_trans = 0; 489 mutex_unlock(&hdq_data->hdq_mutex); 490 omap_hdq_put(hdq_data); 491 return -1; 492 } 493 494 /* Write followed by a read, release the module */ 495 if (hdq_data->init_trans) { 496 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex); 497 if (ret < 0) { 498 dev_dbg(hdq_data->dev, "Could not acquire mutex\n"); 499 return -EINTR; 500 } 501 hdq_data->init_trans = 0; 502 mutex_unlock(&hdq_data->hdq_mutex); 503 omap_hdq_put(hdq_data); 504 } 505 506 return val; 507 } 508 509 /* Write a byte of data to the device */ 510 static void omap_w1_write_byte(void *_hdq, u8 byte) 511 { 512 struct hdq_data *hdq_data = _hdq; 513 int ret; 514 u8 status; 515 516 /* First write to initialize the transfer */ 517 if (hdq_data->init_trans == 0) 518 omap_hdq_get(hdq_data); 519 520 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex); 521 if (ret < 0) { 522 dev_dbg(hdq_data->dev, "Could not acquire mutex\n"); 523 return; 524 } 525 hdq_data->init_trans++; 526 mutex_unlock(&hdq_data->hdq_mutex); 527 528 ret = hdq_write_byte(hdq_data, byte, &status); 529 if (ret < 0) { 530 dev_dbg(hdq_data->dev, "TX failure:Ctrl status %x\n", status); 531 return; 532 } 533 534 /* Second write, data transferred. Release the module */ 535 if (hdq_data->init_trans > 1) { 536 omap_hdq_put(hdq_data); 537 ret = mutex_lock_interruptible(&hdq_data->hdq_mutex); 538 if (ret < 0) { 539 dev_dbg(hdq_data->dev, "Could not acquire mutex\n"); 540 return; 541 } 542 hdq_data->init_trans = 0; 543 mutex_unlock(&hdq_data->hdq_mutex); 544 } 545 } 546 547 static int omap_hdq_probe(struct platform_device *pdev) 548 { 549 struct device *dev = &pdev->dev; 550 struct hdq_data *hdq_data; 551 struct resource *res; 552 int ret, irq; 553 u8 rev; 554 555 hdq_data = devm_kzalloc(dev, sizeof(*hdq_data), GFP_KERNEL); 556 if (!hdq_data) { 557 dev_dbg(&pdev->dev, "unable to allocate memory\n"); 558 return -ENOMEM; 559 } 560 561 hdq_data->dev = dev; 562 platform_set_drvdata(pdev, hdq_data); 563 564 res = platform_get_resource(pdev, IORESOURCE_MEM, 0); 565 hdq_data->hdq_base = devm_ioremap_resource(dev, res); 566 if (IS_ERR(hdq_data->hdq_base)) 567 return PTR_ERR(hdq_data->hdq_base); 568 569 hdq_data->hdq_usecount = 0; 570 mutex_init(&hdq_data->hdq_mutex); 571 572 pm_runtime_enable(&pdev->dev); 573 pm_runtime_get_sync(&pdev->dev); 574 575 rev = hdq_reg_in(hdq_data, OMAP_HDQ_REVISION); 576 dev_info(&pdev->dev, "OMAP HDQ Hardware Rev %c.%c. Driver in %s mode\n", 577 (rev >> 4) + '0', (rev & 0x0f) + '0', "Interrupt"); 578 579 spin_lock_init(&hdq_data->hdq_spinlock); 580 581 irq = platform_get_irq(pdev, 0); 582 if (irq < 0) { 583 ret = -ENXIO; 584 goto err_irq; 585 } 586 587 ret = devm_request_irq(dev, irq, hdq_isr, 0, "omap_hdq", hdq_data); 588 if (ret < 0) { 589 dev_dbg(&pdev->dev, "could not request irq\n"); 590 goto err_irq; 591 } 592 593 omap_hdq_break(hdq_data); 594 595 pm_runtime_put_sync(&pdev->dev); 596 597 omap_w1_master.data = hdq_data; 598 599 ret = w1_add_master_device(&omap_w1_master); 600 if (ret) { 601 dev_dbg(&pdev->dev, "Failure in registering w1 master\n"); 602 goto err_w1; 603 } 604 605 return 0; 606 607 err_irq: 608 pm_runtime_put_sync(&pdev->dev); 609 err_w1: 610 pm_runtime_disable(&pdev->dev); 611 612 return ret; 613 } 614 615 static int omap_hdq_remove(struct platform_device *pdev) 616 { 617 struct hdq_data *hdq_data = platform_get_drvdata(pdev); 618 619 mutex_lock(&hdq_data->hdq_mutex); 620 621 if (hdq_data->hdq_usecount) { 622 dev_dbg(&pdev->dev, "removed when use count is not zero\n"); 623 mutex_unlock(&hdq_data->hdq_mutex); 624 return -EBUSY; 625 } 626 627 mutex_unlock(&hdq_data->hdq_mutex); 628 629 /* remove module dependency */ 630 pm_runtime_disable(&pdev->dev); 631 632 return 0; 633 } 634 635 module_platform_driver(omap_hdq_driver); 636 637 module_param(w1_id, int, S_IRUSR); 638 MODULE_PARM_DESC(w1_id, "1-wire id for the slave detection"); 639 640 MODULE_AUTHOR("Texas Instruments"); 641 MODULE_DESCRIPTION("HDQ driver Library"); 642 MODULE_LICENSE("GPL"); 643