1 /* 2 * linux/drivers/mmc/core/mmc.c 3 * 4 * Copyright (C) 2003-2004 Russell King, All Rights Reserved. 5 * Copyright (C) 2005-2007 Pierre Ossman, All Rights Reserved. 6 * MMCv4 support Copyright (C) 2006 Philip Langdale, All Rights Reserved. 7 * 8 * This program is free software; you can redistribute it and/or modify 9 * it under the terms of the GNU General Public License version 2 as 10 * published by the Free Software Foundation. 11 */ 12 13 #include <linux/err.h> 14 #include <linux/of.h> 15 #include <linux/slab.h> 16 #include <linux/stat.h> 17 #include <linux/pm_runtime.h> 18 19 #include <linux/mmc/host.h> 20 #include <linux/mmc/card.h> 21 #include <linux/mmc/mmc.h> 22 23 #include "core.h" 24 #include "host.h" 25 #include "bus.h" 26 #include "mmc_ops.h" 27 #include "sd_ops.h" 28 29 static const unsigned int tran_exp[] = { 30 10000, 100000, 1000000, 10000000, 31 0, 0, 0, 0 32 }; 33 34 static const unsigned char tran_mant[] = { 35 0, 10, 12, 13, 15, 20, 25, 30, 36 35, 40, 45, 50, 55, 60, 70, 80, 37 }; 38 39 static const unsigned int tacc_exp[] = { 40 1, 10, 100, 1000, 10000, 100000, 1000000, 10000000, 41 }; 42 43 static const unsigned int tacc_mant[] = { 44 0, 10, 12, 13, 15, 20, 25, 30, 45 35, 40, 45, 50, 55, 60, 70, 80, 46 }; 47 48 #define UNSTUFF_BITS(resp,start,size) \ 49 ({ \ 50 const int __size = size; \ 51 const u32 __mask = (__size < 32 ? 1 << __size : 0) - 1; \ 52 const int __off = 3 - ((start) / 32); \ 53 const int __shft = (start) & 31; \ 54 u32 __res; \ 55 \ 56 __res = resp[__off] >> __shft; \ 57 if (__size + __shft > 32) \ 58 __res |= resp[__off-1] << ((32 - __shft) % 32); \ 59 __res & __mask; \ 60 }) 61 62 /* 63 * Given the decoded CSD structure, decode the raw CID to our CID structure. 64 */ 65 static int mmc_decode_cid(struct mmc_card *card) 66 { 67 u32 *resp = card->raw_cid; 68 69 /* 70 * The selection of the format here is based upon published 71 * specs from sandisk and from what people have reported. 72 */ 73 switch (card->csd.mmca_vsn) { 74 case 0: /* MMC v1.0 - v1.2 */ 75 case 1: /* MMC v1.4 */ 76 card->cid.manfid = UNSTUFF_BITS(resp, 104, 24); 77 card->cid.prod_name[0] = UNSTUFF_BITS(resp, 96, 8); 78 card->cid.prod_name[1] = UNSTUFF_BITS(resp, 88, 8); 79 card->cid.prod_name[2] = UNSTUFF_BITS(resp, 80, 8); 80 card->cid.prod_name[3] = UNSTUFF_BITS(resp, 72, 8); 81 card->cid.prod_name[4] = UNSTUFF_BITS(resp, 64, 8); 82 card->cid.prod_name[5] = UNSTUFF_BITS(resp, 56, 8); 83 card->cid.prod_name[6] = UNSTUFF_BITS(resp, 48, 8); 84 card->cid.hwrev = UNSTUFF_BITS(resp, 44, 4); 85 card->cid.fwrev = UNSTUFF_BITS(resp, 40, 4); 86 card->cid.serial = UNSTUFF_BITS(resp, 16, 24); 87 card->cid.month = UNSTUFF_BITS(resp, 12, 4); 88 card->cid.year = UNSTUFF_BITS(resp, 8, 4) + 1997; 89 break; 90 91 case 2: /* MMC v2.0 - v2.2 */ 92 case 3: /* MMC v3.1 - v3.3 */ 93 case 4: /* MMC v4 */ 94 card->cid.manfid = UNSTUFF_BITS(resp, 120, 8); 95 card->cid.oemid = UNSTUFF_BITS(resp, 104, 16); 96 card->cid.prod_name[0] = UNSTUFF_BITS(resp, 96, 8); 97 card->cid.prod_name[1] = UNSTUFF_BITS(resp, 88, 8); 98 card->cid.prod_name[2] = UNSTUFF_BITS(resp, 80, 8); 99 card->cid.prod_name[3] = UNSTUFF_BITS(resp, 72, 8); 100 card->cid.prod_name[4] = UNSTUFF_BITS(resp, 64, 8); 101 card->cid.prod_name[5] = UNSTUFF_BITS(resp, 56, 8); 102 card->cid.prv = UNSTUFF_BITS(resp, 48, 8); 103 card->cid.serial = UNSTUFF_BITS(resp, 16, 32); 104 card->cid.month = UNSTUFF_BITS(resp, 12, 4); 105 card->cid.year = UNSTUFF_BITS(resp, 8, 4) + 1997; 106 break; 107 108 default: 109 pr_err("%s: card has unknown MMCA version %d\n", 110 mmc_hostname(card->host), card->csd.mmca_vsn); 111 return -EINVAL; 112 } 113 114 return 0; 115 } 116 117 static void mmc_set_erase_size(struct mmc_card *card) 118 { 119 if (card->ext_csd.erase_group_def & 1) 120 card->erase_size = card->ext_csd.hc_erase_size; 121 else 122 card->erase_size = card->csd.erase_size; 123 124 mmc_init_erase(card); 125 } 126 127 /* 128 * Given a 128-bit response, decode to our card CSD structure. 129 */ 130 static int mmc_decode_csd(struct mmc_card *card) 131 { 132 struct mmc_csd *csd = &card->csd; 133 unsigned int e, m, a, b; 134 u32 *resp = card->raw_csd; 135 136 /* 137 * We only understand CSD structure v1.1 and v1.2. 138 * v1.2 has extra information in bits 15, 11 and 10. 139 * We also support eMMC v4.4 & v4.41. 140 */ 141 csd->structure = UNSTUFF_BITS(resp, 126, 2); 142 if (csd->structure == 0) { 143 pr_err("%s: unrecognised CSD structure version %d\n", 144 mmc_hostname(card->host), csd->structure); 145 return -EINVAL; 146 } 147 148 csd->mmca_vsn = UNSTUFF_BITS(resp, 122, 4); 149 m = UNSTUFF_BITS(resp, 115, 4); 150 e = UNSTUFF_BITS(resp, 112, 3); 151 csd->tacc_ns = (tacc_exp[e] * tacc_mant[m] + 9) / 10; 152 csd->tacc_clks = UNSTUFF_BITS(resp, 104, 8) * 100; 153 154 m = UNSTUFF_BITS(resp, 99, 4); 155 e = UNSTUFF_BITS(resp, 96, 3); 156 csd->max_dtr = tran_exp[e] * tran_mant[m]; 157 csd->cmdclass = UNSTUFF_BITS(resp, 84, 12); 158 159 e = UNSTUFF_BITS(resp, 47, 3); 160 m = UNSTUFF_BITS(resp, 62, 12); 161 csd->capacity = (1 + m) << (e + 2); 162 163 csd->read_blkbits = UNSTUFF_BITS(resp, 80, 4); 164 csd->read_partial = UNSTUFF_BITS(resp, 79, 1); 165 csd->write_misalign = UNSTUFF_BITS(resp, 78, 1); 166 csd->read_misalign = UNSTUFF_BITS(resp, 77, 1); 167 csd->dsr_imp = UNSTUFF_BITS(resp, 76, 1); 168 csd->r2w_factor = UNSTUFF_BITS(resp, 26, 3); 169 csd->write_blkbits = UNSTUFF_BITS(resp, 22, 4); 170 csd->write_partial = UNSTUFF_BITS(resp, 21, 1); 171 172 if (csd->write_blkbits >= 9) { 173 a = UNSTUFF_BITS(resp, 42, 5); 174 b = UNSTUFF_BITS(resp, 37, 5); 175 csd->erase_size = (a + 1) * (b + 1); 176 csd->erase_size <<= csd->write_blkbits - 9; 177 } 178 179 return 0; 180 } 181 182 static void mmc_select_card_type(struct mmc_card *card) 183 { 184 struct mmc_host *host = card->host; 185 u8 card_type = card->ext_csd.raw_card_type; 186 u32 caps = host->caps, caps2 = host->caps2; 187 unsigned int hs_max_dtr = 0, hs200_max_dtr = 0; 188 unsigned int avail_type = 0; 189 190 if (caps & MMC_CAP_MMC_HIGHSPEED && 191 card_type & EXT_CSD_CARD_TYPE_HS_26) { 192 hs_max_dtr = MMC_HIGH_26_MAX_DTR; 193 avail_type |= EXT_CSD_CARD_TYPE_HS_26; 194 } 195 196 if (caps & MMC_CAP_MMC_HIGHSPEED && 197 card_type & EXT_CSD_CARD_TYPE_HS_52) { 198 hs_max_dtr = MMC_HIGH_52_MAX_DTR; 199 avail_type |= EXT_CSD_CARD_TYPE_HS_52; 200 } 201 202 if (caps & MMC_CAP_1_8V_DDR && 203 card_type & EXT_CSD_CARD_TYPE_DDR_1_8V) { 204 hs_max_dtr = MMC_HIGH_DDR_MAX_DTR; 205 avail_type |= EXT_CSD_CARD_TYPE_DDR_1_8V; 206 } 207 208 if (caps & MMC_CAP_1_2V_DDR && 209 card_type & EXT_CSD_CARD_TYPE_DDR_1_2V) { 210 hs_max_dtr = MMC_HIGH_DDR_MAX_DTR; 211 avail_type |= EXT_CSD_CARD_TYPE_DDR_1_2V; 212 } 213 214 if (caps2 & MMC_CAP2_HS200_1_8V_SDR && 215 card_type & EXT_CSD_CARD_TYPE_HS200_1_8V) { 216 hs200_max_dtr = MMC_HS200_MAX_DTR; 217 avail_type |= EXT_CSD_CARD_TYPE_HS200_1_8V; 218 } 219 220 if (caps2 & MMC_CAP2_HS200_1_2V_SDR && 221 card_type & EXT_CSD_CARD_TYPE_HS200_1_2V) { 222 hs200_max_dtr = MMC_HS200_MAX_DTR; 223 avail_type |= EXT_CSD_CARD_TYPE_HS200_1_2V; 224 } 225 226 if (caps2 & MMC_CAP2_HS400_1_8V && 227 card_type & EXT_CSD_CARD_TYPE_HS400_1_8V) { 228 hs200_max_dtr = MMC_HS200_MAX_DTR; 229 avail_type |= EXT_CSD_CARD_TYPE_HS400_1_8V; 230 } 231 232 if (caps2 & MMC_CAP2_HS400_1_2V && 233 card_type & EXT_CSD_CARD_TYPE_HS400_1_2V) { 234 hs200_max_dtr = MMC_HS200_MAX_DTR; 235 avail_type |= EXT_CSD_CARD_TYPE_HS400_1_2V; 236 } 237 238 card->ext_csd.hs_max_dtr = hs_max_dtr; 239 card->ext_csd.hs200_max_dtr = hs200_max_dtr; 240 card->mmc_avail_type = avail_type; 241 } 242 243 static void mmc_manage_enhanced_area(struct mmc_card *card, u8 *ext_csd) 244 { 245 u8 hc_erase_grp_sz, hc_wp_grp_sz; 246 247 /* 248 * Disable these attributes by default 249 */ 250 card->ext_csd.enhanced_area_offset = -EINVAL; 251 card->ext_csd.enhanced_area_size = -EINVAL; 252 253 /* 254 * Enhanced area feature support -- check whether the eMMC 255 * card has the Enhanced area enabled. If so, export enhanced 256 * area offset and size to user by adding sysfs interface. 257 */ 258 if ((ext_csd[EXT_CSD_PARTITION_SUPPORT] & 0x2) && 259 (ext_csd[EXT_CSD_PARTITION_ATTRIBUTE] & 0x1)) { 260 if (card->ext_csd.partition_setting_completed) { 261 hc_erase_grp_sz = 262 ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE]; 263 hc_wp_grp_sz = 264 ext_csd[EXT_CSD_HC_WP_GRP_SIZE]; 265 266 /* 267 * calculate the enhanced data area offset, in bytes 268 */ 269 card->ext_csd.enhanced_area_offset = 270 (((unsigned long long)ext_csd[139]) << 24) + 271 (((unsigned long long)ext_csd[138]) << 16) + 272 (((unsigned long long)ext_csd[137]) << 8) + 273 (((unsigned long long)ext_csd[136])); 274 if (mmc_card_blockaddr(card)) 275 card->ext_csd.enhanced_area_offset <<= 9; 276 /* 277 * calculate the enhanced data area size, in kilobytes 278 */ 279 card->ext_csd.enhanced_area_size = 280 (ext_csd[142] << 16) + (ext_csd[141] << 8) + 281 ext_csd[140]; 282 card->ext_csd.enhanced_area_size *= 283 (size_t)(hc_erase_grp_sz * hc_wp_grp_sz); 284 card->ext_csd.enhanced_area_size <<= 9; 285 } else { 286 pr_warn("%s: defines enhanced area without partition setting complete\n", 287 mmc_hostname(card->host)); 288 } 289 } 290 } 291 292 static void mmc_manage_gp_partitions(struct mmc_card *card, u8 *ext_csd) 293 { 294 int idx; 295 u8 hc_erase_grp_sz, hc_wp_grp_sz; 296 unsigned int part_size; 297 298 /* 299 * General purpose partition feature support -- 300 * If ext_csd has the size of general purpose partitions, 301 * set size, part_cfg, partition name in mmc_part. 302 */ 303 if (ext_csd[EXT_CSD_PARTITION_SUPPORT] & 304 EXT_CSD_PART_SUPPORT_PART_EN) { 305 hc_erase_grp_sz = 306 ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE]; 307 hc_wp_grp_sz = 308 ext_csd[EXT_CSD_HC_WP_GRP_SIZE]; 309 310 for (idx = 0; idx < MMC_NUM_GP_PARTITION; idx++) { 311 if (!ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3] && 312 !ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 1] && 313 !ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 2]) 314 continue; 315 if (card->ext_csd.partition_setting_completed == 0) { 316 pr_warn("%s: has partition size defined without partition complete\n", 317 mmc_hostname(card->host)); 318 break; 319 } 320 part_size = 321 (ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 2] 322 << 16) + 323 (ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 1] 324 << 8) + 325 ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3]; 326 part_size *= (size_t)(hc_erase_grp_sz * 327 hc_wp_grp_sz); 328 mmc_part_add(card, part_size << 19, 329 EXT_CSD_PART_CONFIG_ACC_GP0 + idx, 330 "gp%d", idx, false, 331 MMC_BLK_DATA_AREA_GP); 332 } 333 } 334 } 335 336 /* 337 * Decode extended CSD. 338 */ 339 static int mmc_decode_ext_csd(struct mmc_card *card, u8 *ext_csd) 340 { 341 int err = 0, idx; 342 unsigned int part_size; 343 struct device_node *np; 344 bool broken_hpi = false; 345 346 /* Version is coded in the CSD_STRUCTURE byte in the EXT_CSD register */ 347 card->ext_csd.raw_ext_csd_structure = ext_csd[EXT_CSD_STRUCTURE]; 348 if (card->csd.structure == 3) { 349 if (card->ext_csd.raw_ext_csd_structure > 2) { 350 pr_err("%s: unrecognised EXT_CSD structure " 351 "version %d\n", mmc_hostname(card->host), 352 card->ext_csd.raw_ext_csd_structure); 353 err = -EINVAL; 354 goto out; 355 } 356 } 357 358 np = mmc_of_find_child_device(card->host, 0); 359 if (np && of_device_is_compatible(np, "mmc-card")) 360 broken_hpi = of_property_read_bool(np, "broken-hpi"); 361 of_node_put(np); 362 363 /* 364 * The EXT_CSD format is meant to be forward compatible. As long 365 * as CSD_STRUCTURE does not change, all values for EXT_CSD_REV 366 * are authorized, see JEDEC JESD84-B50 section B.8. 367 */ 368 card->ext_csd.rev = ext_csd[EXT_CSD_REV]; 369 370 card->ext_csd.raw_sectors[0] = ext_csd[EXT_CSD_SEC_CNT + 0]; 371 card->ext_csd.raw_sectors[1] = ext_csd[EXT_CSD_SEC_CNT + 1]; 372 card->ext_csd.raw_sectors[2] = ext_csd[EXT_CSD_SEC_CNT + 2]; 373 card->ext_csd.raw_sectors[3] = ext_csd[EXT_CSD_SEC_CNT + 3]; 374 if (card->ext_csd.rev >= 2) { 375 card->ext_csd.sectors = 376 ext_csd[EXT_CSD_SEC_CNT + 0] << 0 | 377 ext_csd[EXT_CSD_SEC_CNT + 1] << 8 | 378 ext_csd[EXT_CSD_SEC_CNT + 2] << 16 | 379 ext_csd[EXT_CSD_SEC_CNT + 3] << 24; 380 381 /* Cards with density > 2GiB are sector addressed */ 382 if (card->ext_csd.sectors > (2u * 1024 * 1024 * 1024) / 512) 383 mmc_card_set_blockaddr(card); 384 } 385 386 card->ext_csd.raw_card_type = ext_csd[EXT_CSD_CARD_TYPE]; 387 mmc_select_card_type(card); 388 389 card->ext_csd.raw_s_a_timeout = ext_csd[EXT_CSD_S_A_TIMEOUT]; 390 card->ext_csd.raw_erase_timeout_mult = 391 ext_csd[EXT_CSD_ERASE_TIMEOUT_MULT]; 392 card->ext_csd.raw_hc_erase_grp_size = 393 ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE]; 394 if (card->ext_csd.rev >= 3) { 395 u8 sa_shift = ext_csd[EXT_CSD_S_A_TIMEOUT]; 396 card->ext_csd.part_config = ext_csd[EXT_CSD_PART_CONFIG]; 397 398 /* EXT_CSD value is in units of 10ms, but we store in ms */ 399 card->ext_csd.part_time = 10 * ext_csd[EXT_CSD_PART_SWITCH_TIME]; 400 401 /* Sleep / awake timeout in 100ns units */ 402 if (sa_shift > 0 && sa_shift <= 0x17) 403 card->ext_csd.sa_timeout = 404 1 << ext_csd[EXT_CSD_S_A_TIMEOUT]; 405 card->ext_csd.erase_group_def = 406 ext_csd[EXT_CSD_ERASE_GROUP_DEF]; 407 card->ext_csd.hc_erase_timeout = 300 * 408 ext_csd[EXT_CSD_ERASE_TIMEOUT_MULT]; 409 card->ext_csd.hc_erase_size = 410 ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE] << 10; 411 412 card->ext_csd.rel_sectors = ext_csd[EXT_CSD_REL_WR_SEC_C]; 413 414 /* 415 * There are two boot regions of equal size, defined in 416 * multiples of 128K. 417 */ 418 if (ext_csd[EXT_CSD_BOOT_MULT] && mmc_boot_partition_access(card->host)) { 419 for (idx = 0; idx < MMC_NUM_BOOT_PARTITION; idx++) { 420 part_size = ext_csd[EXT_CSD_BOOT_MULT] << 17; 421 mmc_part_add(card, part_size, 422 EXT_CSD_PART_CONFIG_ACC_BOOT0 + idx, 423 "boot%d", idx, true, 424 MMC_BLK_DATA_AREA_BOOT); 425 } 426 } 427 } 428 429 card->ext_csd.raw_hc_erase_gap_size = 430 ext_csd[EXT_CSD_HC_WP_GRP_SIZE]; 431 card->ext_csd.raw_sec_trim_mult = 432 ext_csd[EXT_CSD_SEC_TRIM_MULT]; 433 card->ext_csd.raw_sec_erase_mult = 434 ext_csd[EXT_CSD_SEC_ERASE_MULT]; 435 card->ext_csd.raw_sec_feature_support = 436 ext_csd[EXT_CSD_SEC_FEATURE_SUPPORT]; 437 card->ext_csd.raw_trim_mult = 438 ext_csd[EXT_CSD_TRIM_MULT]; 439 card->ext_csd.raw_partition_support = ext_csd[EXT_CSD_PARTITION_SUPPORT]; 440 card->ext_csd.raw_driver_strength = ext_csd[EXT_CSD_DRIVER_STRENGTH]; 441 if (card->ext_csd.rev >= 4) { 442 if (ext_csd[EXT_CSD_PARTITION_SETTING_COMPLETED] & 443 EXT_CSD_PART_SETTING_COMPLETED) 444 card->ext_csd.partition_setting_completed = 1; 445 else 446 card->ext_csd.partition_setting_completed = 0; 447 448 mmc_manage_enhanced_area(card, ext_csd); 449 450 mmc_manage_gp_partitions(card, ext_csd); 451 452 card->ext_csd.sec_trim_mult = 453 ext_csd[EXT_CSD_SEC_TRIM_MULT]; 454 card->ext_csd.sec_erase_mult = 455 ext_csd[EXT_CSD_SEC_ERASE_MULT]; 456 card->ext_csd.sec_feature_support = 457 ext_csd[EXT_CSD_SEC_FEATURE_SUPPORT]; 458 card->ext_csd.trim_timeout = 300 * 459 ext_csd[EXT_CSD_TRIM_MULT]; 460 461 /* 462 * Note that the call to mmc_part_add above defaults to read 463 * only. If this default assumption is changed, the call must 464 * take into account the value of boot_locked below. 465 */ 466 card->ext_csd.boot_ro_lock = ext_csd[EXT_CSD_BOOT_WP]; 467 card->ext_csd.boot_ro_lockable = true; 468 469 /* Save power class values */ 470 card->ext_csd.raw_pwr_cl_52_195 = 471 ext_csd[EXT_CSD_PWR_CL_52_195]; 472 card->ext_csd.raw_pwr_cl_26_195 = 473 ext_csd[EXT_CSD_PWR_CL_26_195]; 474 card->ext_csd.raw_pwr_cl_52_360 = 475 ext_csd[EXT_CSD_PWR_CL_52_360]; 476 card->ext_csd.raw_pwr_cl_26_360 = 477 ext_csd[EXT_CSD_PWR_CL_26_360]; 478 card->ext_csd.raw_pwr_cl_200_195 = 479 ext_csd[EXT_CSD_PWR_CL_200_195]; 480 card->ext_csd.raw_pwr_cl_200_360 = 481 ext_csd[EXT_CSD_PWR_CL_200_360]; 482 card->ext_csd.raw_pwr_cl_ddr_52_195 = 483 ext_csd[EXT_CSD_PWR_CL_DDR_52_195]; 484 card->ext_csd.raw_pwr_cl_ddr_52_360 = 485 ext_csd[EXT_CSD_PWR_CL_DDR_52_360]; 486 card->ext_csd.raw_pwr_cl_ddr_200_360 = 487 ext_csd[EXT_CSD_PWR_CL_DDR_200_360]; 488 } 489 490 if (card->ext_csd.rev >= 5) { 491 /* Adjust production date as per JEDEC JESD84-B451 */ 492 if (card->cid.year < 2010) 493 card->cid.year += 16; 494 495 /* check whether the eMMC card supports BKOPS */ 496 if (ext_csd[EXT_CSD_BKOPS_SUPPORT] & 0x1) { 497 card->ext_csd.bkops = 1; 498 card->ext_csd.man_bkops_en = 499 (ext_csd[EXT_CSD_BKOPS_EN] & 500 EXT_CSD_MANUAL_BKOPS_MASK); 501 card->ext_csd.raw_bkops_status = 502 ext_csd[EXT_CSD_BKOPS_STATUS]; 503 if (!card->ext_csd.man_bkops_en) 504 pr_info("%s: MAN_BKOPS_EN bit is not set\n", 505 mmc_hostname(card->host)); 506 } 507 508 /* check whether the eMMC card supports HPI */ 509 if (!broken_hpi && (ext_csd[EXT_CSD_HPI_FEATURES] & 0x1)) { 510 card->ext_csd.hpi = 1; 511 if (ext_csd[EXT_CSD_HPI_FEATURES] & 0x2) 512 card->ext_csd.hpi_cmd = MMC_STOP_TRANSMISSION; 513 else 514 card->ext_csd.hpi_cmd = MMC_SEND_STATUS; 515 /* 516 * Indicate the maximum timeout to close 517 * a command interrupted by HPI 518 */ 519 card->ext_csd.out_of_int_time = 520 ext_csd[EXT_CSD_OUT_OF_INTERRUPT_TIME] * 10; 521 } 522 523 card->ext_csd.rel_param = ext_csd[EXT_CSD_WR_REL_PARAM]; 524 card->ext_csd.rst_n_function = ext_csd[EXT_CSD_RST_N_FUNCTION]; 525 526 /* 527 * RPMB regions are defined in multiples of 128K. 528 */ 529 card->ext_csd.raw_rpmb_size_mult = ext_csd[EXT_CSD_RPMB_MULT]; 530 if (ext_csd[EXT_CSD_RPMB_MULT] && mmc_host_cmd23(card->host)) { 531 mmc_part_add(card, ext_csd[EXT_CSD_RPMB_MULT] << 17, 532 EXT_CSD_PART_CONFIG_ACC_RPMB, 533 "rpmb", 0, false, 534 MMC_BLK_DATA_AREA_RPMB); 535 } 536 } 537 538 card->ext_csd.raw_erased_mem_count = ext_csd[EXT_CSD_ERASED_MEM_CONT]; 539 if (ext_csd[EXT_CSD_ERASED_MEM_CONT]) 540 card->erased_byte = 0xFF; 541 else 542 card->erased_byte = 0x0; 543 544 /* eMMC v4.5 or later */ 545 if (card->ext_csd.rev >= 6) { 546 card->ext_csd.feature_support |= MMC_DISCARD_FEATURE; 547 548 card->ext_csd.generic_cmd6_time = 10 * 549 ext_csd[EXT_CSD_GENERIC_CMD6_TIME]; 550 card->ext_csd.power_off_longtime = 10 * 551 ext_csd[EXT_CSD_POWER_OFF_LONG_TIME]; 552 553 card->ext_csd.cache_size = 554 ext_csd[EXT_CSD_CACHE_SIZE + 0] << 0 | 555 ext_csd[EXT_CSD_CACHE_SIZE + 1] << 8 | 556 ext_csd[EXT_CSD_CACHE_SIZE + 2] << 16 | 557 ext_csd[EXT_CSD_CACHE_SIZE + 3] << 24; 558 559 if (ext_csd[EXT_CSD_DATA_SECTOR_SIZE] == 1) 560 card->ext_csd.data_sector_size = 4096; 561 else 562 card->ext_csd.data_sector_size = 512; 563 564 if ((ext_csd[EXT_CSD_DATA_TAG_SUPPORT] & 1) && 565 (ext_csd[EXT_CSD_TAG_UNIT_SIZE] <= 8)) { 566 card->ext_csd.data_tag_unit_size = 567 ((unsigned int) 1 << ext_csd[EXT_CSD_TAG_UNIT_SIZE]) * 568 (card->ext_csd.data_sector_size); 569 } else { 570 card->ext_csd.data_tag_unit_size = 0; 571 } 572 573 card->ext_csd.max_packed_writes = 574 ext_csd[EXT_CSD_MAX_PACKED_WRITES]; 575 card->ext_csd.max_packed_reads = 576 ext_csd[EXT_CSD_MAX_PACKED_READS]; 577 } else { 578 card->ext_csd.data_sector_size = 512; 579 } 580 581 /* eMMC v5 or later */ 582 if (card->ext_csd.rev >= 7) { 583 memcpy(card->ext_csd.fwrev, &ext_csd[EXT_CSD_FIRMWARE_VERSION], 584 MMC_FIRMWARE_LEN); 585 card->ext_csd.ffu_capable = 586 (ext_csd[EXT_CSD_SUPPORTED_MODE] & 0x1) && 587 !(ext_csd[EXT_CSD_FW_CONFIG] & 0x1); 588 } 589 out: 590 return err; 591 } 592 593 static int mmc_read_ext_csd(struct mmc_card *card) 594 { 595 u8 *ext_csd; 596 int err; 597 598 if (!mmc_can_ext_csd(card)) 599 return 0; 600 601 err = mmc_get_ext_csd(card, &ext_csd); 602 if (err) { 603 /* If the host or the card can't do the switch, 604 * fail more gracefully. */ 605 if ((err != -EINVAL) 606 && (err != -ENOSYS) 607 && (err != -EFAULT)) 608 return err; 609 610 /* 611 * High capacity cards should have this "magic" size 612 * stored in their CSD. 613 */ 614 if (card->csd.capacity == (4096 * 512)) { 615 pr_err("%s: unable to read EXT_CSD on a possible high capacity card. Card will be ignored.\n", 616 mmc_hostname(card->host)); 617 } else { 618 pr_warn("%s: unable to read EXT_CSD, performance might suffer\n", 619 mmc_hostname(card->host)); 620 err = 0; 621 } 622 623 return err; 624 } 625 626 err = mmc_decode_ext_csd(card, ext_csd); 627 kfree(ext_csd); 628 return err; 629 } 630 631 static int mmc_compare_ext_csds(struct mmc_card *card, unsigned bus_width) 632 { 633 u8 *bw_ext_csd; 634 int err; 635 636 if (bus_width == MMC_BUS_WIDTH_1) 637 return 0; 638 639 err = mmc_get_ext_csd(card, &bw_ext_csd); 640 if (err) 641 return err; 642 643 /* only compare read only fields */ 644 err = !((card->ext_csd.raw_partition_support == 645 bw_ext_csd[EXT_CSD_PARTITION_SUPPORT]) && 646 (card->ext_csd.raw_erased_mem_count == 647 bw_ext_csd[EXT_CSD_ERASED_MEM_CONT]) && 648 (card->ext_csd.rev == 649 bw_ext_csd[EXT_CSD_REV]) && 650 (card->ext_csd.raw_ext_csd_structure == 651 bw_ext_csd[EXT_CSD_STRUCTURE]) && 652 (card->ext_csd.raw_card_type == 653 bw_ext_csd[EXT_CSD_CARD_TYPE]) && 654 (card->ext_csd.raw_s_a_timeout == 655 bw_ext_csd[EXT_CSD_S_A_TIMEOUT]) && 656 (card->ext_csd.raw_hc_erase_gap_size == 657 bw_ext_csd[EXT_CSD_HC_WP_GRP_SIZE]) && 658 (card->ext_csd.raw_erase_timeout_mult == 659 bw_ext_csd[EXT_CSD_ERASE_TIMEOUT_MULT]) && 660 (card->ext_csd.raw_hc_erase_grp_size == 661 bw_ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE]) && 662 (card->ext_csd.raw_sec_trim_mult == 663 bw_ext_csd[EXT_CSD_SEC_TRIM_MULT]) && 664 (card->ext_csd.raw_sec_erase_mult == 665 bw_ext_csd[EXT_CSD_SEC_ERASE_MULT]) && 666 (card->ext_csd.raw_sec_feature_support == 667 bw_ext_csd[EXT_CSD_SEC_FEATURE_SUPPORT]) && 668 (card->ext_csd.raw_trim_mult == 669 bw_ext_csd[EXT_CSD_TRIM_MULT]) && 670 (card->ext_csd.raw_sectors[0] == 671 bw_ext_csd[EXT_CSD_SEC_CNT + 0]) && 672 (card->ext_csd.raw_sectors[1] == 673 bw_ext_csd[EXT_CSD_SEC_CNT + 1]) && 674 (card->ext_csd.raw_sectors[2] == 675 bw_ext_csd[EXT_CSD_SEC_CNT + 2]) && 676 (card->ext_csd.raw_sectors[3] == 677 bw_ext_csd[EXT_CSD_SEC_CNT + 3]) && 678 (card->ext_csd.raw_pwr_cl_52_195 == 679 bw_ext_csd[EXT_CSD_PWR_CL_52_195]) && 680 (card->ext_csd.raw_pwr_cl_26_195 == 681 bw_ext_csd[EXT_CSD_PWR_CL_26_195]) && 682 (card->ext_csd.raw_pwr_cl_52_360 == 683 bw_ext_csd[EXT_CSD_PWR_CL_52_360]) && 684 (card->ext_csd.raw_pwr_cl_26_360 == 685 bw_ext_csd[EXT_CSD_PWR_CL_26_360]) && 686 (card->ext_csd.raw_pwr_cl_200_195 == 687 bw_ext_csd[EXT_CSD_PWR_CL_200_195]) && 688 (card->ext_csd.raw_pwr_cl_200_360 == 689 bw_ext_csd[EXT_CSD_PWR_CL_200_360]) && 690 (card->ext_csd.raw_pwr_cl_ddr_52_195 == 691 bw_ext_csd[EXT_CSD_PWR_CL_DDR_52_195]) && 692 (card->ext_csd.raw_pwr_cl_ddr_52_360 == 693 bw_ext_csd[EXT_CSD_PWR_CL_DDR_52_360]) && 694 (card->ext_csd.raw_pwr_cl_ddr_200_360 == 695 bw_ext_csd[EXT_CSD_PWR_CL_DDR_200_360])); 696 697 if (err) 698 err = -EINVAL; 699 700 kfree(bw_ext_csd); 701 return err; 702 } 703 704 MMC_DEV_ATTR(cid, "%08x%08x%08x%08x\n", card->raw_cid[0], card->raw_cid[1], 705 card->raw_cid[2], card->raw_cid[3]); 706 MMC_DEV_ATTR(csd, "%08x%08x%08x%08x\n", card->raw_csd[0], card->raw_csd[1], 707 card->raw_csd[2], card->raw_csd[3]); 708 MMC_DEV_ATTR(date, "%02d/%04d\n", card->cid.month, card->cid.year); 709 MMC_DEV_ATTR(erase_size, "%u\n", card->erase_size << 9); 710 MMC_DEV_ATTR(preferred_erase_size, "%u\n", card->pref_erase << 9); 711 MMC_DEV_ATTR(ffu_capable, "%d\n", card->ext_csd.ffu_capable); 712 MMC_DEV_ATTR(hwrev, "0x%x\n", card->cid.hwrev); 713 MMC_DEV_ATTR(manfid, "0x%06x\n", card->cid.manfid); 714 MMC_DEV_ATTR(name, "%s\n", card->cid.prod_name); 715 MMC_DEV_ATTR(oemid, "0x%04x\n", card->cid.oemid); 716 MMC_DEV_ATTR(prv, "0x%x\n", card->cid.prv); 717 MMC_DEV_ATTR(serial, "0x%08x\n", card->cid.serial); 718 MMC_DEV_ATTR(enhanced_area_offset, "%llu\n", 719 card->ext_csd.enhanced_area_offset); 720 MMC_DEV_ATTR(enhanced_area_size, "%u\n", card->ext_csd.enhanced_area_size); 721 MMC_DEV_ATTR(raw_rpmb_size_mult, "%#x\n", card->ext_csd.raw_rpmb_size_mult); 722 MMC_DEV_ATTR(rel_sectors, "%#x\n", card->ext_csd.rel_sectors); 723 724 static ssize_t mmc_fwrev_show(struct device *dev, 725 struct device_attribute *attr, 726 char *buf) 727 { 728 struct mmc_card *card = mmc_dev_to_card(dev); 729 730 if (card->ext_csd.rev < 7) { 731 return sprintf(buf, "0x%x\n", card->cid.fwrev); 732 } else { 733 return sprintf(buf, "0x%*phN\n", MMC_FIRMWARE_LEN, 734 card->ext_csd.fwrev); 735 } 736 } 737 738 static DEVICE_ATTR(fwrev, S_IRUGO, mmc_fwrev_show, NULL); 739 740 static struct attribute *mmc_std_attrs[] = { 741 &dev_attr_cid.attr, 742 &dev_attr_csd.attr, 743 &dev_attr_date.attr, 744 &dev_attr_erase_size.attr, 745 &dev_attr_preferred_erase_size.attr, 746 &dev_attr_fwrev.attr, 747 &dev_attr_ffu_capable.attr, 748 &dev_attr_hwrev.attr, 749 &dev_attr_manfid.attr, 750 &dev_attr_name.attr, 751 &dev_attr_oemid.attr, 752 &dev_attr_prv.attr, 753 &dev_attr_serial.attr, 754 &dev_attr_enhanced_area_offset.attr, 755 &dev_attr_enhanced_area_size.attr, 756 &dev_attr_raw_rpmb_size_mult.attr, 757 &dev_attr_rel_sectors.attr, 758 NULL, 759 }; 760 ATTRIBUTE_GROUPS(mmc_std); 761 762 static struct device_type mmc_type = { 763 .groups = mmc_std_groups, 764 }; 765 766 /* 767 * Select the PowerClass for the current bus width 768 * If power class is defined for 4/8 bit bus in the 769 * extended CSD register, select it by executing the 770 * mmc_switch command. 771 */ 772 static int __mmc_select_powerclass(struct mmc_card *card, 773 unsigned int bus_width) 774 { 775 struct mmc_host *host = card->host; 776 struct mmc_ext_csd *ext_csd = &card->ext_csd; 777 unsigned int pwrclass_val = 0; 778 int err = 0; 779 780 switch (1 << host->ios.vdd) { 781 case MMC_VDD_165_195: 782 if (host->ios.clock <= MMC_HIGH_26_MAX_DTR) 783 pwrclass_val = ext_csd->raw_pwr_cl_26_195; 784 else if (host->ios.clock <= MMC_HIGH_52_MAX_DTR) 785 pwrclass_val = (bus_width <= EXT_CSD_BUS_WIDTH_8) ? 786 ext_csd->raw_pwr_cl_52_195 : 787 ext_csd->raw_pwr_cl_ddr_52_195; 788 else if (host->ios.clock <= MMC_HS200_MAX_DTR) 789 pwrclass_val = ext_csd->raw_pwr_cl_200_195; 790 break; 791 case MMC_VDD_27_28: 792 case MMC_VDD_28_29: 793 case MMC_VDD_29_30: 794 case MMC_VDD_30_31: 795 case MMC_VDD_31_32: 796 case MMC_VDD_32_33: 797 case MMC_VDD_33_34: 798 case MMC_VDD_34_35: 799 case MMC_VDD_35_36: 800 if (host->ios.clock <= MMC_HIGH_26_MAX_DTR) 801 pwrclass_val = ext_csd->raw_pwr_cl_26_360; 802 else if (host->ios.clock <= MMC_HIGH_52_MAX_DTR) 803 pwrclass_val = (bus_width <= EXT_CSD_BUS_WIDTH_8) ? 804 ext_csd->raw_pwr_cl_52_360 : 805 ext_csd->raw_pwr_cl_ddr_52_360; 806 else if (host->ios.clock <= MMC_HS200_MAX_DTR) 807 pwrclass_val = (bus_width == EXT_CSD_DDR_BUS_WIDTH_8) ? 808 ext_csd->raw_pwr_cl_ddr_200_360 : 809 ext_csd->raw_pwr_cl_200_360; 810 break; 811 default: 812 pr_warn("%s: Voltage range not supported for power class\n", 813 mmc_hostname(host)); 814 return -EINVAL; 815 } 816 817 if (bus_width & (EXT_CSD_BUS_WIDTH_8 | EXT_CSD_DDR_BUS_WIDTH_8)) 818 pwrclass_val = (pwrclass_val & EXT_CSD_PWR_CL_8BIT_MASK) >> 819 EXT_CSD_PWR_CL_8BIT_SHIFT; 820 else 821 pwrclass_val = (pwrclass_val & EXT_CSD_PWR_CL_4BIT_MASK) >> 822 EXT_CSD_PWR_CL_4BIT_SHIFT; 823 824 /* If the power class is different from the default value */ 825 if (pwrclass_val > 0) { 826 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 827 EXT_CSD_POWER_CLASS, 828 pwrclass_val, 829 card->ext_csd.generic_cmd6_time); 830 } 831 832 return err; 833 } 834 835 static int mmc_select_powerclass(struct mmc_card *card) 836 { 837 struct mmc_host *host = card->host; 838 u32 bus_width, ext_csd_bits; 839 int err, ddr; 840 841 /* Power class selection is supported for versions >= 4.0 */ 842 if (!mmc_can_ext_csd(card)) 843 return 0; 844 845 bus_width = host->ios.bus_width; 846 /* Power class values are defined only for 4/8 bit bus */ 847 if (bus_width == MMC_BUS_WIDTH_1) 848 return 0; 849 850 ddr = card->mmc_avail_type & EXT_CSD_CARD_TYPE_DDR_52; 851 if (ddr) 852 ext_csd_bits = (bus_width == MMC_BUS_WIDTH_8) ? 853 EXT_CSD_DDR_BUS_WIDTH_8 : EXT_CSD_DDR_BUS_WIDTH_4; 854 else 855 ext_csd_bits = (bus_width == MMC_BUS_WIDTH_8) ? 856 EXT_CSD_BUS_WIDTH_8 : EXT_CSD_BUS_WIDTH_4; 857 858 err = __mmc_select_powerclass(card, ext_csd_bits); 859 if (err) 860 pr_warn("%s: power class selection to bus width %d ddr %d failed\n", 861 mmc_hostname(host), 1 << bus_width, ddr); 862 863 return err; 864 } 865 866 /* 867 * Set the bus speed for the selected speed mode. 868 */ 869 static void mmc_set_bus_speed(struct mmc_card *card) 870 { 871 unsigned int max_dtr = (unsigned int)-1; 872 873 if ((mmc_card_hs200(card) || mmc_card_hs400(card)) && 874 max_dtr > card->ext_csd.hs200_max_dtr) 875 max_dtr = card->ext_csd.hs200_max_dtr; 876 else if (mmc_card_hs(card) && max_dtr > card->ext_csd.hs_max_dtr) 877 max_dtr = card->ext_csd.hs_max_dtr; 878 else if (max_dtr > card->csd.max_dtr) 879 max_dtr = card->csd.max_dtr; 880 881 mmc_set_clock(card->host, max_dtr); 882 } 883 884 /* 885 * Select the bus width amoung 4-bit and 8-bit(SDR). 886 * If the bus width is changed successfully, return the selected width value. 887 * Zero is returned instead of error value if the wide width is not supported. 888 */ 889 static int mmc_select_bus_width(struct mmc_card *card) 890 { 891 static unsigned ext_csd_bits[] = { 892 EXT_CSD_BUS_WIDTH_8, 893 EXT_CSD_BUS_WIDTH_4, 894 }; 895 static unsigned bus_widths[] = { 896 MMC_BUS_WIDTH_8, 897 MMC_BUS_WIDTH_4, 898 }; 899 struct mmc_host *host = card->host; 900 unsigned idx, bus_width = 0; 901 int err = 0; 902 903 if (!mmc_can_ext_csd(card) || 904 !(host->caps & (MMC_CAP_4_BIT_DATA | MMC_CAP_8_BIT_DATA))) 905 return 0; 906 907 idx = (host->caps & MMC_CAP_8_BIT_DATA) ? 0 : 1; 908 909 /* 910 * Unlike SD, MMC cards dont have a configuration register to notify 911 * supported bus width. So bus test command should be run to identify 912 * the supported bus width or compare the ext csd values of current 913 * bus width and ext csd values of 1 bit mode read earlier. 914 */ 915 for (; idx < ARRAY_SIZE(bus_widths); idx++) { 916 /* 917 * Host is capable of 8bit transfer, then switch 918 * the device to work in 8bit transfer mode. If the 919 * mmc switch command returns error then switch to 920 * 4bit transfer mode. On success set the corresponding 921 * bus width on the host. 922 */ 923 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 924 EXT_CSD_BUS_WIDTH, 925 ext_csd_bits[idx], 926 card->ext_csd.generic_cmd6_time); 927 if (err) 928 continue; 929 930 bus_width = bus_widths[idx]; 931 mmc_set_bus_width(host, bus_width); 932 933 /* 934 * If controller can't handle bus width test, 935 * compare ext_csd previously read in 1 bit mode 936 * against ext_csd at new bus width 937 */ 938 if (!(host->caps & MMC_CAP_BUS_WIDTH_TEST)) 939 err = mmc_compare_ext_csds(card, bus_width); 940 else 941 err = mmc_bus_test(card, bus_width); 942 943 if (!err) { 944 err = bus_width; 945 break; 946 } else { 947 pr_warn("%s: switch to bus width %d failed\n", 948 mmc_hostname(host), ext_csd_bits[idx]); 949 } 950 } 951 952 return err; 953 } 954 955 /* 956 * Switch to the high-speed mode 957 */ 958 static int mmc_select_hs(struct mmc_card *card) 959 { 960 int err; 961 962 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 963 EXT_CSD_HS_TIMING, EXT_CSD_TIMING_HS, 964 card->ext_csd.generic_cmd6_time, 965 true, true, true); 966 if (!err) 967 mmc_set_timing(card->host, MMC_TIMING_MMC_HS); 968 969 return err; 970 } 971 972 /* 973 * Activate wide bus and DDR if supported. 974 */ 975 static int mmc_select_hs_ddr(struct mmc_card *card) 976 { 977 struct mmc_host *host = card->host; 978 u32 bus_width, ext_csd_bits; 979 int err = 0; 980 981 if (!(card->mmc_avail_type & EXT_CSD_CARD_TYPE_DDR_52)) 982 return 0; 983 984 bus_width = host->ios.bus_width; 985 if (bus_width == MMC_BUS_WIDTH_1) 986 return 0; 987 988 ext_csd_bits = (bus_width == MMC_BUS_WIDTH_8) ? 989 EXT_CSD_DDR_BUS_WIDTH_8 : EXT_CSD_DDR_BUS_WIDTH_4; 990 991 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 992 EXT_CSD_BUS_WIDTH, 993 ext_csd_bits, 994 card->ext_csd.generic_cmd6_time); 995 if (err) { 996 pr_err("%s: switch to bus width %d ddr failed\n", 997 mmc_hostname(host), 1 << bus_width); 998 return err; 999 } 1000 1001 /* 1002 * eMMC cards can support 3.3V to 1.2V i/o (vccq) 1003 * signaling. 1004 * 1005 * EXT_CSD_CARD_TYPE_DDR_1_8V means 3.3V or 1.8V vccq. 1006 * 1007 * 1.8V vccq at 3.3V core voltage (vcc) is not required 1008 * in the JEDEC spec for DDR. 1009 * 1010 * Even (e)MMC card can support 3.3v to 1.2v vccq, but not all 1011 * host controller can support this, like some of the SDHCI 1012 * controller which connect to an eMMC device. Some of these 1013 * host controller still needs to use 1.8v vccq for supporting 1014 * DDR mode. 1015 * 1016 * So the sequence will be: 1017 * if (host and device can both support 1.2v IO) 1018 * use 1.2v IO; 1019 * else if (host and device can both support 1.8v IO) 1020 * use 1.8v IO; 1021 * so if host and device can only support 3.3v IO, this is the 1022 * last choice. 1023 * 1024 * WARNING: eMMC rules are NOT the same as SD DDR 1025 */ 1026 err = -EINVAL; 1027 if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_DDR_1_2V) 1028 err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_120); 1029 1030 if (err && (card->mmc_avail_type & EXT_CSD_CARD_TYPE_DDR_1_8V)) 1031 err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_180); 1032 1033 /* make sure vccq is 3.3v after switching disaster */ 1034 if (err) 1035 err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_330); 1036 1037 if (!err) 1038 mmc_set_timing(host, MMC_TIMING_MMC_DDR52); 1039 1040 return err; 1041 } 1042 1043 /* Caller must hold re-tuning */ 1044 static int mmc_switch_status(struct mmc_card *card) 1045 { 1046 u32 status; 1047 int err; 1048 1049 err = mmc_send_status(card, &status); 1050 if (err) 1051 return err; 1052 1053 return mmc_switch_status_error(card->host, status); 1054 } 1055 1056 static int mmc_select_hs400(struct mmc_card *card) 1057 { 1058 struct mmc_host *host = card->host; 1059 bool send_status = true; 1060 unsigned int max_dtr; 1061 int err = 0; 1062 u8 val; 1063 1064 /* 1065 * HS400 mode requires 8-bit bus width 1066 */ 1067 if (!(card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS400 && 1068 host->ios.bus_width == MMC_BUS_WIDTH_8)) 1069 return 0; 1070 1071 if (host->caps & MMC_CAP_WAIT_WHILE_BUSY) 1072 send_status = false; 1073 1074 /* Reduce frequency to HS frequency */ 1075 max_dtr = card->ext_csd.hs_max_dtr; 1076 mmc_set_clock(host, max_dtr); 1077 1078 /* Switch card to HS mode */ 1079 val = EXT_CSD_TIMING_HS | 1080 card->drive_strength << EXT_CSD_DRV_STR_SHIFT; 1081 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1082 EXT_CSD_HS_TIMING, val, 1083 card->ext_csd.generic_cmd6_time, 1084 true, send_status, true); 1085 if (err) { 1086 pr_err("%s: switch to high-speed from hs200 failed, err:%d\n", 1087 mmc_hostname(host), err); 1088 return err; 1089 } 1090 1091 /* Set host controller to HS timing */ 1092 mmc_set_timing(card->host, MMC_TIMING_MMC_HS); 1093 1094 if (!send_status) { 1095 err = mmc_switch_status(card); 1096 if (err) 1097 goto out_err; 1098 } 1099 1100 /* Switch card to DDR */ 1101 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1102 EXT_CSD_BUS_WIDTH, 1103 EXT_CSD_DDR_BUS_WIDTH_8, 1104 card->ext_csd.generic_cmd6_time); 1105 if (err) { 1106 pr_err("%s: switch to bus width for hs400 failed, err:%d\n", 1107 mmc_hostname(host), err); 1108 return err; 1109 } 1110 1111 /* Switch card to HS400 */ 1112 val = EXT_CSD_TIMING_HS400 | 1113 card->drive_strength << EXT_CSD_DRV_STR_SHIFT; 1114 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1115 EXT_CSD_HS_TIMING, val, 1116 card->ext_csd.generic_cmd6_time, 1117 true, send_status, true); 1118 if (err) { 1119 pr_err("%s: switch to hs400 failed, err:%d\n", 1120 mmc_hostname(host), err); 1121 return err; 1122 } 1123 1124 /* Set host controller to HS400 timing and frequency */ 1125 mmc_set_timing(host, MMC_TIMING_MMC_HS400); 1126 mmc_set_bus_speed(card); 1127 1128 if (!send_status) { 1129 err = mmc_switch_status(card); 1130 if (err) 1131 goto out_err; 1132 } 1133 1134 return 0; 1135 1136 out_err: 1137 pr_err("%s: %s failed, error %d\n", mmc_hostname(card->host), 1138 __func__, err); 1139 return err; 1140 } 1141 1142 int mmc_hs200_to_hs400(struct mmc_card *card) 1143 { 1144 return mmc_select_hs400(card); 1145 } 1146 1147 int mmc_hs400_to_hs200(struct mmc_card *card) 1148 { 1149 struct mmc_host *host = card->host; 1150 bool send_status = true; 1151 unsigned int max_dtr; 1152 int err; 1153 u8 val; 1154 1155 if (host->caps & MMC_CAP_WAIT_WHILE_BUSY) 1156 send_status = false; 1157 1158 /* Reduce frequency to HS */ 1159 max_dtr = card->ext_csd.hs_max_dtr; 1160 mmc_set_clock(host, max_dtr); 1161 1162 /* Switch HS400 to HS DDR */ 1163 val = EXT_CSD_TIMING_HS | 1164 card->drive_strength << EXT_CSD_DRV_STR_SHIFT; 1165 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_HS_TIMING, 1166 val, card->ext_csd.generic_cmd6_time, 1167 true, send_status, true); 1168 if (err) 1169 goto out_err; 1170 1171 mmc_set_timing(host, MMC_TIMING_MMC_DDR52); 1172 1173 if (!send_status) { 1174 err = mmc_switch_status(card); 1175 if (err) 1176 goto out_err; 1177 } 1178 1179 /* Switch HS DDR to HS */ 1180 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_BUS_WIDTH, 1181 EXT_CSD_BUS_WIDTH_8, card->ext_csd.generic_cmd6_time, 1182 true, send_status, true); 1183 if (err) 1184 goto out_err; 1185 1186 mmc_set_timing(host, MMC_TIMING_MMC_HS); 1187 1188 if (!send_status) { 1189 err = mmc_switch_status(card); 1190 if (err) 1191 goto out_err; 1192 } 1193 1194 /* Switch HS to HS200 */ 1195 val = EXT_CSD_TIMING_HS200 | 1196 card->drive_strength << EXT_CSD_DRV_STR_SHIFT; 1197 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_HS_TIMING, 1198 val, card->ext_csd.generic_cmd6_time, true, 1199 send_status, true); 1200 if (err) 1201 goto out_err; 1202 1203 mmc_set_timing(host, MMC_TIMING_MMC_HS200); 1204 1205 if (!send_status) { 1206 err = mmc_switch_status(card); 1207 if (err) 1208 goto out_err; 1209 } 1210 1211 mmc_set_bus_speed(card); 1212 1213 return 0; 1214 1215 out_err: 1216 pr_err("%s: %s failed, error %d\n", mmc_hostname(card->host), 1217 __func__, err); 1218 return err; 1219 } 1220 1221 static void mmc_select_driver_type(struct mmc_card *card) 1222 { 1223 int card_drv_type, drive_strength, drv_type; 1224 1225 card_drv_type = card->ext_csd.raw_driver_strength | 1226 mmc_driver_type_mask(0); 1227 1228 drive_strength = mmc_select_drive_strength(card, 1229 card->ext_csd.hs200_max_dtr, 1230 card_drv_type, &drv_type); 1231 1232 card->drive_strength = drive_strength; 1233 1234 if (drv_type) 1235 mmc_set_driver_type(card->host, drv_type); 1236 } 1237 1238 /* 1239 * For device supporting HS200 mode, the following sequence 1240 * should be done before executing the tuning process. 1241 * 1. set the desired bus width(4-bit or 8-bit, 1-bit is not supported) 1242 * 2. switch to HS200 mode 1243 * 3. set the clock to > 52Mhz and <=200MHz 1244 */ 1245 static int mmc_select_hs200(struct mmc_card *card) 1246 { 1247 struct mmc_host *host = card->host; 1248 bool send_status = true; 1249 unsigned int old_timing; 1250 int err = -EINVAL; 1251 u8 val; 1252 1253 if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS200_1_2V) 1254 err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_120); 1255 1256 if (err && card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS200_1_8V) 1257 err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_180); 1258 1259 /* If fails try again during next card power cycle */ 1260 if (err) 1261 goto err; 1262 1263 mmc_select_driver_type(card); 1264 1265 if (host->caps & MMC_CAP_WAIT_WHILE_BUSY) 1266 send_status = false; 1267 1268 /* 1269 * Set the bus width(4 or 8) with host's support and 1270 * switch to HS200 mode if bus width is set successfully. 1271 */ 1272 err = mmc_select_bus_width(card); 1273 if (!IS_ERR_VALUE(err)) { 1274 val = EXT_CSD_TIMING_HS200 | 1275 card->drive_strength << EXT_CSD_DRV_STR_SHIFT; 1276 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1277 EXT_CSD_HS_TIMING, val, 1278 card->ext_csd.generic_cmd6_time, 1279 true, send_status, true); 1280 if (err) 1281 goto err; 1282 old_timing = host->ios.timing; 1283 mmc_set_timing(host, MMC_TIMING_MMC_HS200); 1284 if (!send_status) { 1285 err = mmc_switch_status(card); 1286 /* 1287 * mmc_select_timing() assumes timing has not changed if 1288 * it is a switch error. 1289 */ 1290 if (err == -EBADMSG) 1291 mmc_set_timing(host, old_timing); 1292 } 1293 } 1294 err: 1295 if (err) 1296 pr_err("%s: %s failed, error %d\n", mmc_hostname(card->host), 1297 __func__, err); 1298 return err; 1299 } 1300 1301 /* 1302 * Activate High Speed or HS200 mode if supported. 1303 */ 1304 static int mmc_select_timing(struct mmc_card *card) 1305 { 1306 int err = 0; 1307 1308 if (!mmc_can_ext_csd(card)) 1309 goto bus_speed; 1310 1311 if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS200) 1312 err = mmc_select_hs200(card); 1313 else if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS) 1314 err = mmc_select_hs(card); 1315 1316 if (err && err != -EBADMSG) 1317 return err; 1318 1319 if (err) { 1320 pr_warn("%s: switch to %s failed\n", 1321 mmc_card_hs(card) ? "high-speed" : 1322 (mmc_card_hs200(card) ? "hs200" : ""), 1323 mmc_hostname(card->host)); 1324 err = 0; 1325 } 1326 1327 bus_speed: 1328 /* 1329 * Set the bus speed to the selected bus timing. 1330 * If timing is not selected, backward compatible is the default. 1331 */ 1332 mmc_set_bus_speed(card); 1333 return err; 1334 } 1335 1336 /* 1337 * Execute tuning sequence to seek the proper bus operating 1338 * conditions for HS200 and HS400, which sends CMD21 to the device. 1339 */ 1340 static int mmc_hs200_tuning(struct mmc_card *card) 1341 { 1342 struct mmc_host *host = card->host; 1343 1344 /* 1345 * Timing should be adjusted to the HS400 target 1346 * operation frequency for tuning process 1347 */ 1348 if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS400 && 1349 host->ios.bus_width == MMC_BUS_WIDTH_8) 1350 if (host->ops->prepare_hs400_tuning) 1351 host->ops->prepare_hs400_tuning(host, &host->ios); 1352 1353 return mmc_execute_tuning(card); 1354 } 1355 1356 /* 1357 * Handle the detection and initialisation of a card. 1358 * 1359 * In the case of a resume, "oldcard" will contain the card 1360 * we're trying to reinitialise. 1361 */ 1362 static int mmc_init_card(struct mmc_host *host, u32 ocr, 1363 struct mmc_card *oldcard) 1364 { 1365 struct mmc_card *card; 1366 int err; 1367 u32 cid[4]; 1368 u32 rocr; 1369 1370 BUG_ON(!host); 1371 WARN_ON(!host->claimed); 1372 1373 /* Set correct bus mode for MMC before attempting init */ 1374 if (!mmc_host_is_spi(host)) 1375 mmc_set_bus_mode(host, MMC_BUSMODE_OPENDRAIN); 1376 1377 /* 1378 * Since we're changing the OCR value, we seem to 1379 * need to tell some cards to go back to the idle 1380 * state. We wait 1ms to give cards time to 1381 * respond. 1382 * mmc_go_idle is needed for eMMC that are asleep 1383 */ 1384 mmc_go_idle(host); 1385 1386 /* The extra bit indicates that we support high capacity */ 1387 err = mmc_send_op_cond(host, ocr | (1 << 30), &rocr); 1388 if (err) 1389 goto err; 1390 1391 /* 1392 * For SPI, enable CRC as appropriate. 1393 */ 1394 if (mmc_host_is_spi(host)) { 1395 err = mmc_spi_set_crc(host, use_spi_crc); 1396 if (err) 1397 goto err; 1398 } 1399 1400 /* 1401 * Fetch CID from card. 1402 */ 1403 if (mmc_host_is_spi(host)) 1404 err = mmc_send_cid(host, cid); 1405 else 1406 err = mmc_all_send_cid(host, cid); 1407 if (err) 1408 goto err; 1409 1410 if (oldcard) { 1411 if (memcmp(cid, oldcard->raw_cid, sizeof(cid)) != 0) { 1412 err = -ENOENT; 1413 goto err; 1414 } 1415 1416 card = oldcard; 1417 } else { 1418 /* 1419 * Allocate card structure. 1420 */ 1421 card = mmc_alloc_card(host, &mmc_type); 1422 if (IS_ERR(card)) { 1423 err = PTR_ERR(card); 1424 goto err; 1425 } 1426 1427 card->ocr = ocr; 1428 card->type = MMC_TYPE_MMC; 1429 card->rca = 1; 1430 memcpy(card->raw_cid, cid, sizeof(card->raw_cid)); 1431 } 1432 1433 /* 1434 * Call the optional HC's init_card function to handle quirks. 1435 */ 1436 if (host->ops->init_card) 1437 host->ops->init_card(host, card); 1438 1439 /* 1440 * For native busses: set card RCA and quit open drain mode. 1441 */ 1442 if (!mmc_host_is_spi(host)) { 1443 err = mmc_set_relative_addr(card); 1444 if (err) 1445 goto free_card; 1446 1447 mmc_set_bus_mode(host, MMC_BUSMODE_PUSHPULL); 1448 } 1449 1450 if (!oldcard) { 1451 /* 1452 * Fetch CSD from card. 1453 */ 1454 err = mmc_send_csd(card, card->raw_csd); 1455 if (err) 1456 goto free_card; 1457 1458 err = mmc_decode_csd(card); 1459 if (err) 1460 goto free_card; 1461 err = mmc_decode_cid(card); 1462 if (err) 1463 goto free_card; 1464 } 1465 1466 /* 1467 * handling only for cards supporting DSR and hosts requesting 1468 * DSR configuration 1469 */ 1470 if (card->csd.dsr_imp && host->dsr_req) 1471 mmc_set_dsr(host); 1472 1473 /* 1474 * Select card, as all following commands rely on that. 1475 */ 1476 if (!mmc_host_is_spi(host)) { 1477 err = mmc_select_card(card); 1478 if (err) 1479 goto free_card; 1480 } 1481 1482 if (!oldcard) { 1483 /* Read extended CSD. */ 1484 err = mmc_read_ext_csd(card); 1485 if (err) 1486 goto free_card; 1487 1488 /* If doing byte addressing, check if required to do sector 1489 * addressing. Handle the case of <2GB cards needing sector 1490 * addressing. See section 8.1 JEDEC Standard JED84-A441; 1491 * ocr register has bit 30 set for sector addressing. 1492 */ 1493 if (!(mmc_card_blockaddr(card)) && (rocr & (1<<30))) 1494 mmc_card_set_blockaddr(card); 1495 1496 /* Erase size depends on CSD and Extended CSD */ 1497 mmc_set_erase_size(card); 1498 } 1499 1500 /* 1501 * If enhanced_area_en is TRUE, host needs to enable ERASE_GRP_DEF 1502 * bit. This bit will be lost every time after a reset or power off. 1503 */ 1504 if (card->ext_csd.partition_setting_completed || 1505 (card->ext_csd.rev >= 3 && (host->caps2 & MMC_CAP2_HC_ERASE_SZ))) { 1506 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1507 EXT_CSD_ERASE_GROUP_DEF, 1, 1508 card->ext_csd.generic_cmd6_time); 1509 1510 if (err && err != -EBADMSG) 1511 goto free_card; 1512 1513 if (err) { 1514 err = 0; 1515 /* 1516 * Just disable enhanced area off & sz 1517 * will try to enable ERASE_GROUP_DEF 1518 * during next time reinit 1519 */ 1520 card->ext_csd.enhanced_area_offset = -EINVAL; 1521 card->ext_csd.enhanced_area_size = -EINVAL; 1522 } else { 1523 card->ext_csd.erase_group_def = 1; 1524 /* 1525 * enable ERASE_GRP_DEF successfully. 1526 * This will affect the erase size, so 1527 * here need to reset erase size 1528 */ 1529 mmc_set_erase_size(card); 1530 } 1531 } 1532 1533 /* 1534 * Ensure eMMC user default partition is enabled 1535 */ 1536 if (card->ext_csd.part_config & EXT_CSD_PART_CONFIG_ACC_MASK) { 1537 card->ext_csd.part_config &= ~EXT_CSD_PART_CONFIG_ACC_MASK; 1538 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_PART_CONFIG, 1539 card->ext_csd.part_config, 1540 card->ext_csd.part_time); 1541 if (err && err != -EBADMSG) 1542 goto free_card; 1543 } 1544 1545 /* 1546 * Enable power_off_notification byte in the ext_csd register 1547 */ 1548 if (card->ext_csd.rev >= 6) { 1549 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1550 EXT_CSD_POWER_OFF_NOTIFICATION, 1551 EXT_CSD_POWER_ON, 1552 card->ext_csd.generic_cmd6_time); 1553 if (err && err != -EBADMSG) 1554 goto free_card; 1555 1556 /* 1557 * The err can be -EBADMSG or 0, 1558 * so check for success and update the flag 1559 */ 1560 if (!err) 1561 card->ext_csd.power_off_notification = EXT_CSD_POWER_ON; 1562 } 1563 1564 /* 1565 * Select timing interface 1566 */ 1567 err = mmc_select_timing(card); 1568 if (err) 1569 goto free_card; 1570 1571 if (mmc_card_hs200(card)) { 1572 err = mmc_hs200_tuning(card); 1573 if (err) 1574 goto free_card; 1575 1576 err = mmc_select_hs400(card); 1577 if (err) 1578 goto free_card; 1579 } else if (mmc_card_hs(card)) { 1580 /* Select the desired bus width optionally */ 1581 err = mmc_select_bus_width(card); 1582 if (!IS_ERR_VALUE(err)) { 1583 err = mmc_select_hs_ddr(card); 1584 if (err) 1585 goto free_card; 1586 } 1587 } 1588 1589 /* 1590 * Choose the power class with selected bus interface 1591 */ 1592 mmc_select_powerclass(card); 1593 1594 /* 1595 * Enable HPI feature (if supported) 1596 */ 1597 if (card->ext_csd.hpi) { 1598 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1599 EXT_CSD_HPI_MGMT, 1, 1600 card->ext_csd.generic_cmd6_time); 1601 if (err && err != -EBADMSG) 1602 goto free_card; 1603 if (err) { 1604 pr_warn("%s: Enabling HPI failed\n", 1605 mmc_hostname(card->host)); 1606 err = 0; 1607 } else 1608 card->ext_csd.hpi_en = 1; 1609 } 1610 1611 /* 1612 * If cache size is higher than 0, this indicates 1613 * the existence of cache and it can be turned on. 1614 */ 1615 if (card->ext_csd.cache_size > 0) { 1616 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1617 EXT_CSD_CACHE_CTRL, 1, 1618 card->ext_csd.generic_cmd6_time); 1619 if (err && err != -EBADMSG) 1620 goto free_card; 1621 1622 /* 1623 * Only if no error, cache is turned on successfully. 1624 */ 1625 if (err) { 1626 pr_warn("%s: Cache is supported, but failed to turn on (%d)\n", 1627 mmc_hostname(card->host), err); 1628 card->ext_csd.cache_ctrl = 0; 1629 err = 0; 1630 } else { 1631 card->ext_csd.cache_ctrl = 1; 1632 } 1633 } 1634 1635 /* 1636 * The mandatory minimum values are defined for packed command. 1637 * read: 5, write: 3 1638 */ 1639 if (card->ext_csd.max_packed_writes >= 3 && 1640 card->ext_csd.max_packed_reads >= 5 && 1641 host->caps2 & MMC_CAP2_PACKED_CMD) { 1642 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1643 EXT_CSD_EXP_EVENTS_CTRL, 1644 EXT_CSD_PACKED_EVENT_EN, 1645 card->ext_csd.generic_cmd6_time); 1646 if (err && err != -EBADMSG) 1647 goto free_card; 1648 if (err) { 1649 pr_warn("%s: Enabling packed event failed\n", 1650 mmc_hostname(card->host)); 1651 card->ext_csd.packed_event_en = 0; 1652 err = 0; 1653 } else { 1654 card->ext_csd.packed_event_en = 1; 1655 } 1656 } 1657 1658 if (!oldcard) 1659 host->card = card; 1660 1661 return 0; 1662 1663 free_card: 1664 if (!oldcard) 1665 mmc_remove_card(card); 1666 err: 1667 return err; 1668 } 1669 1670 static int mmc_can_sleep(struct mmc_card *card) 1671 { 1672 return (card && card->ext_csd.rev >= 3); 1673 } 1674 1675 static int mmc_sleep(struct mmc_host *host) 1676 { 1677 struct mmc_command cmd = {0}; 1678 struct mmc_card *card = host->card; 1679 unsigned int timeout_ms = DIV_ROUND_UP(card->ext_csd.sa_timeout, 10000); 1680 int err; 1681 1682 /* Re-tuning can't be done once the card is deselected */ 1683 mmc_retune_hold(host); 1684 1685 err = mmc_deselect_cards(host); 1686 if (err) 1687 goto out_release; 1688 1689 cmd.opcode = MMC_SLEEP_AWAKE; 1690 cmd.arg = card->rca << 16; 1691 cmd.arg |= 1 << 15; 1692 1693 /* 1694 * If the max_busy_timeout of the host is specified, validate it against 1695 * the sleep cmd timeout. A failure means we need to prevent the host 1696 * from doing hw busy detection, which is done by converting to a R1 1697 * response instead of a R1B. 1698 */ 1699 if (host->max_busy_timeout && (timeout_ms > host->max_busy_timeout)) { 1700 cmd.flags = MMC_RSP_R1 | MMC_CMD_AC; 1701 } else { 1702 cmd.flags = MMC_RSP_R1B | MMC_CMD_AC; 1703 cmd.busy_timeout = timeout_ms; 1704 } 1705 1706 err = mmc_wait_for_cmd(host, &cmd, 0); 1707 if (err) 1708 goto out_release; 1709 1710 /* 1711 * If the host does not wait while the card signals busy, then we will 1712 * will have to wait the sleep/awake timeout. Note, we cannot use the 1713 * SEND_STATUS command to poll the status because that command (and most 1714 * others) is invalid while the card sleeps. 1715 */ 1716 if (!cmd.busy_timeout || !(host->caps & MMC_CAP_WAIT_WHILE_BUSY)) 1717 mmc_delay(timeout_ms); 1718 1719 out_release: 1720 mmc_retune_release(host); 1721 return err; 1722 } 1723 1724 static int mmc_can_poweroff_notify(const struct mmc_card *card) 1725 { 1726 return card && 1727 mmc_card_mmc(card) && 1728 (card->ext_csd.power_off_notification == EXT_CSD_POWER_ON); 1729 } 1730 1731 static int mmc_poweroff_notify(struct mmc_card *card, unsigned int notify_type) 1732 { 1733 unsigned int timeout = card->ext_csd.generic_cmd6_time; 1734 int err; 1735 1736 /* Use EXT_CSD_POWER_OFF_SHORT as default notification type. */ 1737 if (notify_type == EXT_CSD_POWER_OFF_LONG) 1738 timeout = card->ext_csd.power_off_longtime; 1739 1740 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1741 EXT_CSD_POWER_OFF_NOTIFICATION, 1742 notify_type, timeout, true, false, false); 1743 if (err) 1744 pr_err("%s: Power Off Notification timed out, %u\n", 1745 mmc_hostname(card->host), timeout); 1746 1747 /* Disable the power off notification after the switch operation. */ 1748 card->ext_csd.power_off_notification = EXT_CSD_NO_POWER_NOTIFICATION; 1749 1750 return err; 1751 } 1752 1753 /* 1754 * Host is being removed. Free up the current card. 1755 */ 1756 static void mmc_remove(struct mmc_host *host) 1757 { 1758 BUG_ON(!host); 1759 BUG_ON(!host->card); 1760 1761 mmc_remove_card(host->card); 1762 host->card = NULL; 1763 } 1764 1765 /* 1766 * Card detection - card is alive. 1767 */ 1768 static int mmc_alive(struct mmc_host *host) 1769 { 1770 return mmc_send_status(host->card, NULL); 1771 } 1772 1773 /* 1774 * Card detection callback from host. 1775 */ 1776 static void mmc_detect(struct mmc_host *host) 1777 { 1778 int err; 1779 1780 BUG_ON(!host); 1781 BUG_ON(!host->card); 1782 1783 mmc_get_card(host->card); 1784 1785 /* 1786 * Just check if our card has been removed. 1787 */ 1788 err = _mmc_detect_card_removed(host); 1789 1790 mmc_put_card(host->card); 1791 1792 if (err) { 1793 mmc_remove(host); 1794 1795 mmc_claim_host(host); 1796 mmc_detach_bus(host); 1797 mmc_power_off(host); 1798 mmc_release_host(host); 1799 } 1800 } 1801 1802 static int _mmc_suspend(struct mmc_host *host, bool is_suspend) 1803 { 1804 int err = 0; 1805 unsigned int notify_type = is_suspend ? EXT_CSD_POWER_OFF_SHORT : 1806 EXT_CSD_POWER_OFF_LONG; 1807 1808 BUG_ON(!host); 1809 BUG_ON(!host->card); 1810 1811 mmc_claim_host(host); 1812 1813 if (mmc_card_suspended(host->card)) 1814 goto out; 1815 1816 if (mmc_card_doing_bkops(host->card)) { 1817 err = mmc_stop_bkops(host->card); 1818 if (err) 1819 goto out; 1820 } 1821 1822 err = mmc_flush_cache(host->card); 1823 if (err) 1824 goto out; 1825 1826 if (mmc_can_poweroff_notify(host->card) && 1827 ((host->caps2 & MMC_CAP2_FULL_PWR_CYCLE) || !is_suspend)) 1828 err = mmc_poweroff_notify(host->card, notify_type); 1829 else if (mmc_can_sleep(host->card)) 1830 err = mmc_sleep(host); 1831 else if (!mmc_host_is_spi(host)) 1832 err = mmc_deselect_cards(host); 1833 1834 if (!err) { 1835 mmc_power_off(host); 1836 mmc_card_set_suspended(host->card); 1837 } 1838 out: 1839 mmc_release_host(host); 1840 return err; 1841 } 1842 1843 /* 1844 * Suspend callback 1845 */ 1846 static int mmc_suspend(struct mmc_host *host) 1847 { 1848 int err; 1849 1850 err = _mmc_suspend(host, true); 1851 if (!err) { 1852 pm_runtime_disable(&host->card->dev); 1853 pm_runtime_set_suspended(&host->card->dev); 1854 } 1855 1856 return err; 1857 } 1858 1859 /* 1860 * This function tries to determine if the same card is still present 1861 * and, if so, restore all state to it. 1862 */ 1863 static int _mmc_resume(struct mmc_host *host) 1864 { 1865 int err = 0; 1866 1867 BUG_ON(!host); 1868 BUG_ON(!host->card); 1869 1870 mmc_claim_host(host); 1871 1872 if (!mmc_card_suspended(host->card)) 1873 goto out; 1874 1875 mmc_power_up(host, host->card->ocr); 1876 err = mmc_init_card(host, host->card->ocr, host->card); 1877 mmc_card_clr_suspended(host->card); 1878 1879 out: 1880 mmc_release_host(host); 1881 return err; 1882 } 1883 1884 /* 1885 * Shutdown callback 1886 */ 1887 static int mmc_shutdown(struct mmc_host *host) 1888 { 1889 int err = 0; 1890 1891 /* 1892 * In a specific case for poweroff notify, we need to resume the card 1893 * before we can shutdown it properly. 1894 */ 1895 if (mmc_can_poweroff_notify(host->card) && 1896 !(host->caps2 & MMC_CAP2_FULL_PWR_CYCLE)) 1897 err = _mmc_resume(host); 1898 1899 if (!err) 1900 err = _mmc_suspend(host, false); 1901 1902 return err; 1903 } 1904 1905 /* 1906 * Callback for resume. 1907 */ 1908 static int mmc_resume(struct mmc_host *host) 1909 { 1910 int err = 0; 1911 1912 if (!(host->caps & MMC_CAP_RUNTIME_RESUME)) { 1913 err = _mmc_resume(host); 1914 pm_runtime_set_active(&host->card->dev); 1915 pm_runtime_mark_last_busy(&host->card->dev); 1916 } 1917 pm_runtime_enable(&host->card->dev); 1918 1919 return err; 1920 } 1921 1922 /* 1923 * Callback for runtime_suspend. 1924 */ 1925 static int mmc_runtime_suspend(struct mmc_host *host) 1926 { 1927 int err; 1928 1929 if (!(host->caps & MMC_CAP_AGGRESSIVE_PM)) 1930 return 0; 1931 1932 err = _mmc_suspend(host, true); 1933 if (err) 1934 pr_err("%s: error %d doing aggressive suspend\n", 1935 mmc_hostname(host), err); 1936 1937 return err; 1938 } 1939 1940 /* 1941 * Callback for runtime_resume. 1942 */ 1943 static int mmc_runtime_resume(struct mmc_host *host) 1944 { 1945 int err; 1946 1947 if (!(host->caps & (MMC_CAP_AGGRESSIVE_PM | MMC_CAP_RUNTIME_RESUME))) 1948 return 0; 1949 1950 err = _mmc_resume(host); 1951 if (err) 1952 pr_err("%s: error %d doing aggressive resume\n", 1953 mmc_hostname(host), err); 1954 1955 return 0; 1956 } 1957 1958 int mmc_can_reset(struct mmc_card *card) 1959 { 1960 u8 rst_n_function; 1961 1962 rst_n_function = card->ext_csd.rst_n_function; 1963 if ((rst_n_function & EXT_CSD_RST_N_EN_MASK) != EXT_CSD_RST_N_ENABLED) 1964 return 0; 1965 return 1; 1966 } 1967 EXPORT_SYMBOL(mmc_can_reset); 1968 1969 static int mmc_reset(struct mmc_host *host) 1970 { 1971 struct mmc_card *card = host->card; 1972 1973 if (!(host->caps & MMC_CAP_HW_RESET) || !host->ops->hw_reset) 1974 return -EOPNOTSUPP; 1975 1976 if (!mmc_can_reset(card)) 1977 return -EOPNOTSUPP; 1978 1979 mmc_set_clock(host, host->f_init); 1980 1981 host->ops->hw_reset(host); 1982 1983 /* Set initial state and call mmc_set_ios */ 1984 mmc_set_initial_state(host); 1985 1986 return mmc_init_card(host, card->ocr, card); 1987 } 1988 1989 static const struct mmc_bus_ops mmc_ops = { 1990 .remove = mmc_remove, 1991 .detect = mmc_detect, 1992 .suspend = mmc_suspend, 1993 .resume = mmc_resume, 1994 .runtime_suspend = mmc_runtime_suspend, 1995 .runtime_resume = mmc_runtime_resume, 1996 .alive = mmc_alive, 1997 .shutdown = mmc_shutdown, 1998 .reset = mmc_reset, 1999 }; 2000 2001 /* 2002 * Starting point for MMC card init. 2003 */ 2004 int mmc_attach_mmc(struct mmc_host *host) 2005 { 2006 int err; 2007 u32 ocr, rocr; 2008 2009 BUG_ON(!host); 2010 WARN_ON(!host->claimed); 2011 2012 /* Set correct bus mode for MMC before attempting attach */ 2013 if (!mmc_host_is_spi(host)) 2014 mmc_set_bus_mode(host, MMC_BUSMODE_OPENDRAIN); 2015 2016 err = mmc_send_op_cond(host, 0, &ocr); 2017 if (err) 2018 return err; 2019 2020 mmc_attach_bus(host, &mmc_ops); 2021 if (host->ocr_avail_mmc) 2022 host->ocr_avail = host->ocr_avail_mmc; 2023 2024 /* 2025 * We need to get OCR a different way for SPI. 2026 */ 2027 if (mmc_host_is_spi(host)) { 2028 err = mmc_spi_read_ocr(host, 1, &ocr); 2029 if (err) 2030 goto err; 2031 } 2032 2033 rocr = mmc_select_voltage(host, ocr); 2034 2035 /* 2036 * Can we support the voltage of the card? 2037 */ 2038 if (!rocr) { 2039 err = -EINVAL; 2040 goto err; 2041 } 2042 2043 /* 2044 * Detect and init the card. 2045 */ 2046 err = mmc_init_card(host, rocr, NULL); 2047 if (err) 2048 goto err; 2049 2050 mmc_release_host(host); 2051 err = mmc_add_card(host->card); 2052 if (err) 2053 goto remove_card; 2054 2055 mmc_claim_host(host); 2056 return 0; 2057 2058 remove_card: 2059 mmc_remove_card(host->card); 2060 mmc_claim_host(host); 2061 host->card = NULL; 2062 err: 2063 mmc_detach_bus(host); 2064 2065 pr_err("%s: error %d whilst initialising MMC card\n", 2066 mmc_hostname(host), err); 2067 2068 return err; 2069 } 2070