1 /* 2 * linux/drivers/mmc/core/mmc.c 3 * 4 * Copyright (C) 2003-2004 Russell King, All Rights Reserved. 5 * Copyright (C) 2005-2007 Pierre Ossman, All Rights Reserved. 6 * MMCv4 support Copyright (C) 2006 Philip Langdale, All Rights Reserved. 7 * 8 * This program is free software; you can redistribute it and/or modify 9 * it under the terms of the GNU General Public License version 2 as 10 * published by the Free Software Foundation. 11 */ 12 13 #include <linux/err.h> 14 #include <linux/of.h> 15 #include <linux/slab.h> 16 #include <linux/stat.h> 17 #include <linux/pm_runtime.h> 18 19 #include <linux/mmc/host.h> 20 #include <linux/mmc/card.h> 21 #include <linux/mmc/mmc.h> 22 23 #include "core.h" 24 #include "host.h" 25 #include "bus.h" 26 #include "mmc_ops.h" 27 #include "sd_ops.h" 28 29 static const unsigned int tran_exp[] = { 30 10000, 100000, 1000000, 10000000, 31 0, 0, 0, 0 32 }; 33 34 static const unsigned char tran_mant[] = { 35 0, 10, 12, 13, 15, 20, 25, 30, 36 35, 40, 45, 50, 55, 60, 70, 80, 37 }; 38 39 static const unsigned int tacc_exp[] = { 40 1, 10, 100, 1000, 10000, 100000, 1000000, 10000000, 41 }; 42 43 static const unsigned int tacc_mant[] = { 44 0, 10, 12, 13, 15, 20, 25, 30, 45 35, 40, 45, 50, 55, 60, 70, 80, 46 }; 47 48 #define UNSTUFF_BITS(resp,start,size) \ 49 ({ \ 50 const int __size = size; \ 51 const u32 __mask = (__size < 32 ? 1 << __size : 0) - 1; \ 52 const int __off = 3 - ((start) / 32); \ 53 const int __shft = (start) & 31; \ 54 u32 __res; \ 55 \ 56 __res = resp[__off] >> __shft; \ 57 if (__size + __shft > 32) \ 58 __res |= resp[__off-1] << ((32 - __shft) % 32); \ 59 __res & __mask; \ 60 }) 61 62 /* 63 * Given the decoded CSD structure, decode the raw CID to our CID structure. 64 */ 65 static int mmc_decode_cid(struct mmc_card *card) 66 { 67 u32 *resp = card->raw_cid; 68 69 /* 70 * The selection of the format here is based upon published 71 * specs from sandisk and from what people have reported. 72 */ 73 switch (card->csd.mmca_vsn) { 74 case 0: /* MMC v1.0 - v1.2 */ 75 case 1: /* MMC v1.4 */ 76 card->cid.manfid = UNSTUFF_BITS(resp, 104, 24); 77 card->cid.prod_name[0] = UNSTUFF_BITS(resp, 96, 8); 78 card->cid.prod_name[1] = UNSTUFF_BITS(resp, 88, 8); 79 card->cid.prod_name[2] = UNSTUFF_BITS(resp, 80, 8); 80 card->cid.prod_name[3] = UNSTUFF_BITS(resp, 72, 8); 81 card->cid.prod_name[4] = UNSTUFF_BITS(resp, 64, 8); 82 card->cid.prod_name[5] = UNSTUFF_BITS(resp, 56, 8); 83 card->cid.prod_name[6] = UNSTUFF_BITS(resp, 48, 8); 84 card->cid.hwrev = UNSTUFF_BITS(resp, 44, 4); 85 card->cid.fwrev = UNSTUFF_BITS(resp, 40, 4); 86 card->cid.serial = UNSTUFF_BITS(resp, 16, 24); 87 card->cid.month = UNSTUFF_BITS(resp, 12, 4); 88 card->cid.year = UNSTUFF_BITS(resp, 8, 4) + 1997; 89 break; 90 91 case 2: /* MMC v2.0 - v2.2 */ 92 case 3: /* MMC v3.1 - v3.3 */ 93 case 4: /* MMC v4 */ 94 card->cid.manfid = UNSTUFF_BITS(resp, 120, 8); 95 card->cid.oemid = UNSTUFF_BITS(resp, 104, 16); 96 card->cid.prod_name[0] = UNSTUFF_BITS(resp, 96, 8); 97 card->cid.prod_name[1] = UNSTUFF_BITS(resp, 88, 8); 98 card->cid.prod_name[2] = UNSTUFF_BITS(resp, 80, 8); 99 card->cid.prod_name[3] = UNSTUFF_BITS(resp, 72, 8); 100 card->cid.prod_name[4] = UNSTUFF_BITS(resp, 64, 8); 101 card->cid.prod_name[5] = UNSTUFF_BITS(resp, 56, 8); 102 card->cid.prv = UNSTUFF_BITS(resp, 48, 8); 103 card->cid.serial = UNSTUFF_BITS(resp, 16, 32); 104 card->cid.month = UNSTUFF_BITS(resp, 12, 4); 105 card->cid.year = UNSTUFF_BITS(resp, 8, 4) + 1997; 106 break; 107 108 default: 109 pr_err("%s: card has unknown MMCA version %d\n", 110 mmc_hostname(card->host), card->csd.mmca_vsn); 111 return -EINVAL; 112 } 113 114 return 0; 115 } 116 117 static void mmc_set_erase_size(struct mmc_card *card) 118 { 119 if (card->ext_csd.erase_group_def & 1) 120 card->erase_size = card->ext_csd.hc_erase_size; 121 else 122 card->erase_size = card->csd.erase_size; 123 124 mmc_init_erase(card); 125 } 126 127 /* 128 * Given a 128-bit response, decode to our card CSD structure. 129 */ 130 static int mmc_decode_csd(struct mmc_card *card) 131 { 132 struct mmc_csd *csd = &card->csd; 133 unsigned int e, m, a, b; 134 u32 *resp = card->raw_csd; 135 136 /* 137 * We only understand CSD structure v1.1 and v1.2. 138 * v1.2 has extra information in bits 15, 11 and 10. 139 * We also support eMMC v4.4 & v4.41. 140 */ 141 csd->structure = UNSTUFF_BITS(resp, 126, 2); 142 if (csd->structure == 0) { 143 pr_err("%s: unrecognised CSD structure version %d\n", 144 mmc_hostname(card->host), csd->structure); 145 return -EINVAL; 146 } 147 148 csd->mmca_vsn = UNSTUFF_BITS(resp, 122, 4); 149 m = UNSTUFF_BITS(resp, 115, 4); 150 e = UNSTUFF_BITS(resp, 112, 3); 151 csd->tacc_ns = (tacc_exp[e] * tacc_mant[m] + 9) / 10; 152 csd->tacc_clks = UNSTUFF_BITS(resp, 104, 8) * 100; 153 154 m = UNSTUFF_BITS(resp, 99, 4); 155 e = UNSTUFF_BITS(resp, 96, 3); 156 csd->max_dtr = tran_exp[e] * tran_mant[m]; 157 csd->cmdclass = UNSTUFF_BITS(resp, 84, 12); 158 159 e = UNSTUFF_BITS(resp, 47, 3); 160 m = UNSTUFF_BITS(resp, 62, 12); 161 csd->capacity = (1 + m) << (e + 2); 162 163 csd->read_blkbits = UNSTUFF_BITS(resp, 80, 4); 164 csd->read_partial = UNSTUFF_BITS(resp, 79, 1); 165 csd->write_misalign = UNSTUFF_BITS(resp, 78, 1); 166 csd->read_misalign = UNSTUFF_BITS(resp, 77, 1); 167 csd->dsr_imp = UNSTUFF_BITS(resp, 76, 1); 168 csd->r2w_factor = UNSTUFF_BITS(resp, 26, 3); 169 csd->write_blkbits = UNSTUFF_BITS(resp, 22, 4); 170 csd->write_partial = UNSTUFF_BITS(resp, 21, 1); 171 172 if (csd->write_blkbits >= 9) { 173 a = UNSTUFF_BITS(resp, 42, 5); 174 b = UNSTUFF_BITS(resp, 37, 5); 175 csd->erase_size = (a + 1) * (b + 1); 176 csd->erase_size <<= csd->write_blkbits - 9; 177 } 178 179 return 0; 180 } 181 182 static void mmc_select_card_type(struct mmc_card *card) 183 { 184 struct mmc_host *host = card->host; 185 u8 card_type = card->ext_csd.raw_card_type; 186 u32 caps = host->caps, caps2 = host->caps2; 187 unsigned int hs_max_dtr = 0, hs200_max_dtr = 0; 188 unsigned int avail_type = 0; 189 190 if (caps & MMC_CAP_MMC_HIGHSPEED && 191 card_type & EXT_CSD_CARD_TYPE_HS_26) { 192 hs_max_dtr = MMC_HIGH_26_MAX_DTR; 193 avail_type |= EXT_CSD_CARD_TYPE_HS_26; 194 } 195 196 if (caps & MMC_CAP_MMC_HIGHSPEED && 197 card_type & EXT_CSD_CARD_TYPE_HS_52) { 198 hs_max_dtr = MMC_HIGH_52_MAX_DTR; 199 avail_type |= EXT_CSD_CARD_TYPE_HS_52; 200 } 201 202 if (caps & MMC_CAP_1_8V_DDR && 203 card_type & EXT_CSD_CARD_TYPE_DDR_1_8V) { 204 hs_max_dtr = MMC_HIGH_DDR_MAX_DTR; 205 avail_type |= EXT_CSD_CARD_TYPE_DDR_1_8V; 206 } 207 208 if (caps & MMC_CAP_1_2V_DDR && 209 card_type & EXT_CSD_CARD_TYPE_DDR_1_2V) { 210 hs_max_dtr = MMC_HIGH_DDR_MAX_DTR; 211 avail_type |= EXT_CSD_CARD_TYPE_DDR_1_2V; 212 } 213 214 if (caps2 & MMC_CAP2_HS200_1_8V_SDR && 215 card_type & EXT_CSD_CARD_TYPE_HS200_1_8V) { 216 hs200_max_dtr = MMC_HS200_MAX_DTR; 217 avail_type |= EXT_CSD_CARD_TYPE_HS200_1_8V; 218 } 219 220 if (caps2 & MMC_CAP2_HS200_1_2V_SDR && 221 card_type & EXT_CSD_CARD_TYPE_HS200_1_2V) { 222 hs200_max_dtr = MMC_HS200_MAX_DTR; 223 avail_type |= EXT_CSD_CARD_TYPE_HS200_1_2V; 224 } 225 226 if (caps2 & MMC_CAP2_HS400_1_8V && 227 card_type & EXT_CSD_CARD_TYPE_HS400_1_8V) { 228 hs200_max_dtr = MMC_HS200_MAX_DTR; 229 avail_type |= EXT_CSD_CARD_TYPE_HS400_1_8V; 230 } 231 232 if (caps2 & MMC_CAP2_HS400_1_2V && 233 card_type & EXT_CSD_CARD_TYPE_HS400_1_2V) { 234 hs200_max_dtr = MMC_HS200_MAX_DTR; 235 avail_type |= EXT_CSD_CARD_TYPE_HS400_1_2V; 236 } 237 238 card->ext_csd.hs_max_dtr = hs_max_dtr; 239 card->ext_csd.hs200_max_dtr = hs200_max_dtr; 240 card->mmc_avail_type = avail_type; 241 } 242 243 static void mmc_manage_enhanced_area(struct mmc_card *card, u8 *ext_csd) 244 { 245 u8 hc_erase_grp_sz, hc_wp_grp_sz; 246 247 /* 248 * Disable these attributes by default 249 */ 250 card->ext_csd.enhanced_area_offset = -EINVAL; 251 card->ext_csd.enhanced_area_size = -EINVAL; 252 253 /* 254 * Enhanced area feature support -- check whether the eMMC 255 * card has the Enhanced area enabled. If so, export enhanced 256 * area offset and size to user by adding sysfs interface. 257 */ 258 if ((ext_csd[EXT_CSD_PARTITION_SUPPORT] & 0x2) && 259 (ext_csd[EXT_CSD_PARTITION_ATTRIBUTE] & 0x1)) { 260 if (card->ext_csd.partition_setting_completed) { 261 hc_erase_grp_sz = 262 ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE]; 263 hc_wp_grp_sz = 264 ext_csd[EXT_CSD_HC_WP_GRP_SIZE]; 265 266 /* 267 * calculate the enhanced data area offset, in bytes 268 */ 269 card->ext_csd.enhanced_area_offset = 270 (((unsigned long long)ext_csd[139]) << 24) + 271 (((unsigned long long)ext_csd[138]) << 16) + 272 (((unsigned long long)ext_csd[137]) << 8) + 273 (((unsigned long long)ext_csd[136])); 274 if (mmc_card_blockaddr(card)) 275 card->ext_csd.enhanced_area_offset <<= 9; 276 /* 277 * calculate the enhanced data area size, in kilobytes 278 */ 279 card->ext_csd.enhanced_area_size = 280 (ext_csd[142] << 16) + (ext_csd[141] << 8) + 281 ext_csd[140]; 282 card->ext_csd.enhanced_area_size *= 283 (size_t)(hc_erase_grp_sz * hc_wp_grp_sz); 284 card->ext_csd.enhanced_area_size <<= 9; 285 } else { 286 pr_warn("%s: defines enhanced area without partition setting complete\n", 287 mmc_hostname(card->host)); 288 } 289 } 290 } 291 292 static void mmc_manage_gp_partitions(struct mmc_card *card, u8 *ext_csd) 293 { 294 int idx; 295 u8 hc_erase_grp_sz, hc_wp_grp_sz; 296 unsigned int part_size; 297 298 /* 299 * General purpose partition feature support -- 300 * If ext_csd has the size of general purpose partitions, 301 * set size, part_cfg, partition name in mmc_part. 302 */ 303 if (ext_csd[EXT_CSD_PARTITION_SUPPORT] & 304 EXT_CSD_PART_SUPPORT_PART_EN) { 305 hc_erase_grp_sz = 306 ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE]; 307 hc_wp_grp_sz = 308 ext_csd[EXT_CSD_HC_WP_GRP_SIZE]; 309 310 for (idx = 0; idx < MMC_NUM_GP_PARTITION; idx++) { 311 if (!ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3] && 312 !ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 1] && 313 !ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 2]) 314 continue; 315 if (card->ext_csd.partition_setting_completed == 0) { 316 pr_warn("%s: has partition size defined without partition complete\n", 317 mmc_hostname(card->host)); 318 break; 319 } 320 part_size = 321 (ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 2] 322 << 16) + 323 (ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3 + 1] 324 << 8) + 325 ext_csd[EXT_CSD_GP_SIZE_MULT + idx * 3]; 326 part_size *= (size_t)(hc_erase_grp_sz * 327 hc_wp_grp_sz); 328 mmc_part_add(card, part_size << 19, 329 EXT_CSD_PART_CONFIG_ACC_GP0 + idx, 330 "gp%d", idx, false, 331 MMC_BLK_DATA_AREA_GP); 332 } 333 } 334 } 335 336 /* 337 * Decode extended CSD. 338 */ 339 static int mmc_decode_ext_csd(struct mmc_card *card, u8 *ext_csd) 340 { 341 int err = 0, idx; 342 unsigned int part_size; 343 struct device_node *np; 344 bool broken_hpi = false; 345 346 /* Version is coded in the CSD_STRUCTURE byte in the EXT_CSD register */ 347 card->ext_csd.raw_ext_csd_structure = ext_csd[EXT_CSD_STRUCTURE]; 348 if (card->csd.structure == 3) { 349 if (card->ext_csd.raw_ext_csd_structure > 2) { 350 pr_err("%s: unrecognised EXT_CSD structure " 351 "version %d\n", mmc_hostname(card->host), 352 card->ext_csd.raw_ext_csd_structure); 353 err = -EINVAL; 354 goto out; 355 } 356 } 357 358 np = mmc_of_find_child_device(card->host, 0); 359 if (np && of_device_is_compatible(np, "mmc-card")) 360 broken_hpi = of_property_read_bool(np, "broken-hpi"); 361 of_node_put(np); 362 363 /* 364 * The EXT_CSD format is meant to be forward compatible. As long 365 * as CSD_STRUCTURE does not change, all values for EXT_CSD_REV 366 * are authorized, see JEDEC JESD84-B50 section B.8. 367 */ 368 card->ext_csd.rev = ext_csd[EXT_CSD_REV]; 369 370 card->ext_csd.raw_sectors[0] = ext_csd[EXT_CSD_SEC_CNT + 0]; 371 card->ext_csd.raw_sectors[1] = ext_csd[EXT_CSD_SEC_CNT + 1]; 372 card->ext_csd.raw_sectors[2] = ext_csd[EXT_CSD_SEC_CNT + 2]; 373 card->ext_csd.raw_sectors[3] = ext_csd[EXT_CSD_SEC_CNT + 3]; 374 if (card->ext_csd.rev >= 2) { 375 card->ext_csd.sectors = 376 ext_csd[EXT_CSD_SEC_CNT + 0] << 0 | 377 ext_csd[EXT_CSD_SEC_CNT + 1] << 8 | 378 ext_csd[EXT_CSD_SEC_CNT + 2] << 16 | 379 ext_csd[EXT_CSD_SEC_CNT + 3] << 24; 380 381 /* Cards with density > 2GiB are sector addressed */ 382 if (card->ext_csd.sectors > (2u * 1024 * 1024 * 1024) / 512) 383 mmc_card_set_blockaddr(card); 384 } 385 386 card->ext_csd.raw_card_type = ext_csd[EXT_CSD_CARD_TYPE]; 387 mmc_select_card_type(card); 388 389 card->ext_csd.raw_s_a_timeout = ext_csd[EXT_CSD_S_A_TIMEOUT]; 390 card->ext_csd.raw_erase_timeout_mult = 391 ext_csd[EXT_CSD_ERASE_TIMEOUT_MULT]; 392 card->ext_csd.raw_hc_erase_grp_size = 393 ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE]; 394 if (card->ext_csd.rev >= 3) { 395 u8 sa_shift = ext_csd[EXT_CSD_S_A_TIMEOUT]; 396 card->ext_csd.part_config = ext_csd[EXT_CSD_PART_CONFIG]; 397 398 /* EXT_CSD value is in units of 10ms, but we store in ms */ 399 card->ext_csd.part_time = 10 * ext_csd[EXT_CSD_PART_SWITCH_TIME]; 400 401 /* Sleep / awake timeout in 100ns units */ 402 if (sa_shift > 0 && sa_shift <= 0x17) 403 card->ext_csd.sa_timeout = 404 1 << ext_csd[EXT_CSD_S_A_TIMEOUT]; 405 card->ext_csd.erase_group_def = 406 ext_csd[EXT_CSD_ERASE_GROUP_DEF]; 407 card->ext_csd.hc_erase_timeout = 300 * 408 ext_csd[EXT_CSD_ERASE_TIMEOUT_MULT]; 409 card->ext_csd.hc_erase_size = 410 ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE] << 10; 411 412 card->ext_csd.rel_sectors = ext_csd[EXT_CSD_REL_WR_SEC_C]; 413 414 /* 415 * There are two boot regions of equal size, defined in 416 * multiples of 128K. 417 */ 418 if (ext_csd[EXT_CSD_BOOT_MULT] && mmc_boot_partition_access(card->host)) { 419 for (idx = 0; idx < MMC_NUM_BOOT_PARTITION; idx++) { 420 part_size = ext_csd[EXT_CSD_BOOT_MULT] << 17; 421 mmc_part_add(card, part_size, 422 EXT_CSD_PART_CONFIG_ACC_BOOT0 + idx, 423 "boot%d", idx, true, 424 MMC_BLK_DATA_AREA_BOOT); 425 } 426 } 427 } 428 429 card->ext_csd.raw_hc_erase_gap_size = 430 ext_csd[EXT_CSD_HC_WP_GRP_SIZE]; 431 card->ext_csd.raw_sec_trim_mult = 432 ext_csd[EXT_CSD_SEC_TRIM_MULT]; 433 card->ext_csd.raw_sec_erase_mult = 434 ext_csd[EXT_CSD_SEC_ERASE_MULT]; 435 card->ext_csd.raw_sec_feature_support = 436 ext_csd[EXT_CSD_SEC_FEATURE_SUPPORT]; 437 card->ext_csd.raw_trim_mult = 438 ext_csd[EXT_CSD_TRIM_MULT]; 439 card->ext_csd.raw_partition_support = ext_csd[EXT_CSD_PARTITION_SUPPORT]; 440 card->ext_csd.raw_driver_strength = ext_csd[EXT_CSD_DRIVER_STRENGTH]; 441 if (card->ext_csd.rev >= 4) { 442 if (ext_csd[EXT_CSD_PARTITION_SETTING_COMPLETED] & 443 EXT_CSD_PART_SETTING_COMPLETED) 444 card->ext_csd.partition_setting_completed = 1; 445 else 446 card->ext_csd.partition_setting_completed = 0; 447 448 mmc_manage_enhanced_area(card, ext_csd); 449 450 mmc_manage_gp_partitions(card, ext_csd); 451 452 card->ext_csd.sec_trim_mult = 453 ext_csd[EXT_CSD_SEC_TRIM_MULT]; 454 card->ext_csd.sec_erase_mult = 455 ext_csd[EXT_CSD_SEC_ERASE_MULT]; 456 card->ext_csd.sec_feature_support = 457 ext_csd[EXT_CSD_SEC_FEATURE_SUPPORT]; 458 card->ext_csd.trim_timeout = 300 * 459 ext_csd[EXT_CSD_TRIM_MULT]; 460 461 /* 462 * Note that the call to mmc_part_add above defaults to read 463 * only. If this default assumption is changed, the call must 464 * take into account the value of boot_locked below. 465 */ 466 card->ext_csd.boot_ro_lock = ext_csd[EXT_CSD_BOOT_WP]; 467 card->ext_csd.boot_ro_lockable = true; 468 469 /* Save power class values */ 470 card->ext_csd.raw_pwr_cl_52_195 = 471 ext_csd[EXT_CSD_PWR_CL_52_195]; 472 card->ext_csd.raw_pwr_cl_26_195 = 473 ext_csd[EXT_CSD_PWR_CL_26_195]; 474 card->ext_csd.raw_pwr_cl_52_360 = 475 ext_csd[EXT_CSD_PWR_CL_52_360]; 476 card->ext_csd.raw_pwr_cl_26_360 = 477 ext_csd[EXT_CSD_PWR_CL_26_360]; 478 card->ext_csd.raw_pwr_cl_200_195 = 479 ext_csd[EXT_CSD_PWR_CL_200_195]; 480 card->ext_csd.raw_pwr_cl_200_360 = 481 ext_csd[EXT_CSD_PWR_CL_200_360]; 482 card->ext_csd.raw_pwr_cl_ddr_52_195 = 483 ext_csd[EXT_CSD_PWR_CL_DDR_52_195]; 484 card->ext_csd.raw_pwr_cl_ddr_52_360 = 485 ext_csd[EXT_CSD_PWR_CL_DDR_52_360]; 486 card->ext_csd.raw_pwr_cl_ddr_200_360 = 487 ext_csd[EXT_CSD_PWR_CL_DDR_200_360]; 488 } 489 490 if (card->ext_csd.rev >= 5) { 491 /* Adjust production date as per JEDEC JESD84-B451 */ 492 if (card->cid.year < 2010) 493 card->cid.year += 16; 494 495 /* check whether the eMMC card supports BKOPS */ 496 if (ext_csd[EXT_CSD_BKOPS_SUPPORT] & 0x1) { 497 card->ext_csd.bkops = 1; 498 card->ext_csd.man_bkops_en = 499 (ext_csd[EXT_CSD_BKOPS_EN] & 500 EXT_CSD_MANUAL_BKOPS_MASK); 501 card->ext_csd.raw_bkops_status = 502 ext_csd[EXT_CSD_BKOPS_STATUS]; 503 if (!card->ext_csd.man_bkops_en) 504 pr_info("%s: MAN_BKOPS_EN bit is not set\n", 505 mmc_hostname(card->host)); 506 } 507 508 /* check whether the eMMC card supports HPI */ 509 if (!broken_hpi && (ext_csd[EXT_CSD_HPI_FEATURES] & 0x1)) { 510 card->ext_csd.hpi = 1; 511 if (ext_csd[EXT_CSD_HPI_FEATURES] & 0x2) 512 card->ext_csd.hpi_cmd = MMC_STOP_TRANSMISSION; 513 else 514 card->ext_csd.hpi_cmd = MMC_SEND_STATUS; 515 /* 516 * Indicate the maximum timeout to close 517 * a command interrupted by HPI 518 */ 519 card->ext_csd.out_of_int_time = 520 ext_csd[EXT_CSD_OUT_OF_INTERRUPT_TIME] * 10; 521 } 522 523 card->ext_csd.rel_param = ext_csd[EXT_CSD_WR_REL_PARAM]; 524 card->ext_csd.rst_n_function = ext_csd[EXT_CSD_RST_N_FUNCTION]; 525 526 /* 527 * RPMB regions are defined in multiples of 128K. 528 */ 529 card->ext_csd.raw_rpmb_size_mult = ext_csd[EXT_CSD_RPMB_MULT]; 530 if (ext_csd[EXT_CSD_RPMB_MULT] && mmc_host_cmd23(card->host)) { 531 mmc_part_add(card, ext_csd[EXT_CSD_RPMB_MULT] << 17, 532 EXT_CSD_PART_CONFIG_ACC_RPMB, 533 "rpmb", 0, false, 534 MMC_BLK_DATA_AREA_RPMB); 535 } 536 } 537 538 card->ext_csd.raw_erased_mem_count = ext_csd[EXT_CSD_ERASED_MEM_CONT]; 539 if (ext_csd[EXT_CSD_ERASED_MEM_CONT]) 540 card->erased_byte = 0xFF; 541 else 542 card->erased_byte = 0x0; 543 544 /* eMMC v4.5 or later */ 545 if (card->ext_csd.rev >= 6) { 546 card->ext_csd.feature_support |= MMC_DISCARD_FEATURE; 547 548 card->ext_csd.generic_cmd6_time = 10 * 549 ext_csd[EXT_CSD_GENERIC_CMD6_TIME]; 550 card->ext_csd.power_off_longtime = 10 * 551 ext_csd[EXT_CSD_POWER_OFF_LONG_TIME]; 552 553 card->ext_csd.cache_size = 554 ext_csd[EXT_CSD_CACHE_SIZE + 0] << 0 | 555 ext_csd[EXT_CSD_CACHE_SIZE + 1] << 8 | 556 ext_csd[EXT_CSD_CACHE_SIZE + 2] << 16 | 557 ext_csd[EXT_CSD_CACHE_SIZE + 3] << 24; 558 559 if (ext_csd[EXT_CSD_DATA_SECTOR_SIZE] == 1) 560 card->ext_csd.data_sector_size = 4096; 561 else 562 card->ext_csd.data_sector_size = 512; 563 564 if ((ext_csd[EXT_CSD_DATA_TAG_SUPPORT] & 1) && 565 (ext_csd[EXT_CSD_TAG_UNIT_SIZE] <= 8)) { 566 card->ext_csd.data_tag_unit_size = 567 ((unsigned int) 1 << ext_csd[EXT_CSD_TAG_UNIT_SIZE]) * 568 (card->ext_csd.data_sector_size); 569 } else { 570 card->ext_csd.data_tag_unit_size = 0; 571 } 572 573 card->ext_csd.max_packed_writes = 574 ext_csd[EXT_CSD_MAX_PACKED_WRITES]; 575 card->ext_csd.max_packed_reads = 576 ext_csd[EXT_CSD_MAX_PACKED_READS]; 577 } else { 578 card->ext_csd.data_sector_size = 512; 579 } 580 581 /* eMMC v5 or later */ 582 if (card->ext_csd.rev >= 7) { 583 memcpy(card->ext_csd.fwrev, &ext_csd[EXT_CSD_FIRMWARE_VERSION], 584 MMC_FIRMWARE_LEN); 585 card->ext_csd.ffu_capable = 586 (ext_csd[EXT_CSD_SUPPORTED_MODE] & 0x1) && 587 !(ext_csd[EXT_CSD_FW_CONFIG] & 0x1); 588 } 589 out: 590 return err; 591 } 592 593 static int mmc_read_ext_csd(struct mmc_card *card) 594 { 595 u8 *ext_csd; 596 int err; 597 598 if (!mmc_can_ext_csd(card)) 599 return 0; 600 601 err = mmc_get_ext_csd(card, &ext_csd); 602 if (err) { 603 /* If the host or the card can't do the switch, 604 * fail more gracefully. */ 605 if ((err != -EINVAL) 606 && (err != -ENOSYS) 607 && (err != -EFAULT)) 608 return err; 609 610 /* 611 * High capacity cards should have this "magic" size 612 * stored in their CSD. 613 */ 614 if (card->csd.capacity == (4096 * 512)) { 615 pr_err("%s: unable to read EXT_CSD on a possible high capacity card. Card will be ignored.\n", 616 mmc_hostname(card->host)); 617 } else { 618 pr_warn("%s: unable to read EXT_CSD, performance might suffer\n", 619 mmc_hostname(card->host)); 620 err = 0; 621 } 622 623 return err; 624 } 625 626 err = mmc_decode_ext_csd(card, ext_csd); 627 kfree(ext_csd); 628 return err; 629 } 630 631 static int mmc_compare_ext_csds(struct mmc_card *card, unsigned bus_width) 632 { 633 u8 *bw_ext_csd; 634 int err; 635 636 if (bus_width == MMC_BUS_WIDTH_1) 637 return 0; 638 639 err = mmc_get_ext_csd(card, &bw_ext_csd); 640 if (err) 641 return err; 642 643 /* only compare read only fields */ 644 err = !((card->ext_csd.raw_partition_support == 645 bw_ext_csd[EXT_CSD_PARTITION_SUPPORT]) && 646 (card->ext_csd.raw_erased_mem_count == 647 bw_ext_csd[EXT_CSD_ERASED_MEM_CONT]) && 648 (card->ext_csd.rev == 649 bw_ext_csd[EXT_CSD_REV]) && 650 (card->ext_csd.raw_ext_csd_structure == 651 bw_ext_csd[EXT_CSD_STRUCTURE]) && 652 (card->ext_csd.raw_card_type == 653 bw_ext_csd[EXT_CSD_CARD_TYPE]) && 654 (card->ext_csd.raw_s_a_timeout == 655 bw_ext_csd[EXT_CSD_S_A_TIMEOUT]) && 656 (card->ext_csd.raw_hc_erase_gap_size == 657 bw_ext_csd[EXT_CSD_HC_WP_GRP_SIZE]) && 658 (card->ext_csd.raw_erase_timeout_mult == 659 bw_ext_csd[EXT_CSD_ERASE_TIMEOUT_MULT]) && 660 (card->ext_csd.raw_hc_erase_grp_size == 661 bw_ext_csd[EXT_CSD_HC_ERASE_GRP_SIZE]) && 662 (card->ext_csd.raw_sec_trim_mult == 663 bw_ext_csd[EXT_CSD_SEC_TRIM_MULT]) && 664 (card->ext_csd.raw_sec_erase_mult == 665 bw_ext_csd[EXT_CSD_SEC_ERASE_MULT]) && 666 (card->ext_csd.raw_sec_feature_support == 667 bw_ext_csd[EXT_CSD_SEC_FEATURE_SUPPORT]) && 668 (card->ext_csd.raw_trim_mult == 669 bw_ext_csd[EXT_CSD_TRIM_MULT]) && 670 (card->ext_csd.raw_sectors[0] == 671 bw_ext_csd[EXT_CSD_SEC_CNT + 0]) && 672 (card->ext_csd.raw_sectors[1] == 673 bw_ext_csd[EXT_CSD_SEC_CNT + 1]) && 674 (card->ext_csd.raw_sectors[2] == 675 bw_ext_csd[EXT_CSD_SEC_CNT + 2]) && 676 (card->ext_csd.raw_sectors[3] == 677 bw_ext_csd[EXT_CSD_SEC_CNT + 3]) && 678 (card->ext_csd.raw_pwr_cl_52_195 == 679 bw_ext_csd[EXT_CSD_PWR_CL_52_195]) && 680 (card->ext_csd.raw_pwr_cl_26_195 == 681 bw_ext_csd[EXT_CSD_PWR_CL_26_195]) && 682 (card->ext_csd.raw_pwr_cl_52_360 == 683 bw_ext_csd[EXT_CSD_PWR_CL_52_360]) && 684 (card->ext_csd.raw_pwr_cl_26_360 == 685 bw_ext_csd[EXT_CSD_PWR_CL_26_360]) && 686 (card->ext_csd.raw_pwr_cl_200_195 == 687 bw_ext_csd[EXT_CSD_PWR_CL_200_195]) && 688 (card->ext_csd.raw_pwr_cl_200_360 == 689 bw_ext_csd[EXT_CSD_PWR_CL_200_360]) && 690 (card->ext_csd.raw_pwr_cl_ddr_52_195 == 691 bw_ext_csd[EXT_CSD_PWR_CL_DDR_52_195]) && 692 (card->ext_csd.raw_pwr_cl_ddr_52_360 == 693 bw_ext_csd[EXT_CSD_PWR_CL_DDR_52_360]) && 694 (card->ext_csd.raw_pwr_cl_ddr_200_360 == 695 bw_ext_csd[EXT_CSD_PWR_CL_DDR_200_360])); 696 697 if (err) 698 err = -EINVAL; 699 700 kfree(bw_ext_csd); 701 return err; 702 } 703 704 MMC_DEV_ATTR(cid, "%08x%08x%08x%08x\n", card->raw_cid[0], card->raw_cid[1], 705 card->raw_cid[2], card->raw_cid[3]); 706 MMC_DEV_ATTR(csd, "%08x%08x%08x%08x\n", card->raw_csd[0], card->raw_csd[1], 707 card->raw_csd[2], card->raw_csd[3]); 708 MMC_DEV_ATTR(date, "%02d/%04d\n", card->cid.month, card->cid.year); 709 MMC_DEV_ATTR(erase_size, "%u\n", card->erase_size << 9); 710 MMC_DEV_ATTR(preferred_erase_size, "%u\n", card->pref_erase << 9); 711 MMC_DEV_ATTR(ffu_capable, "%d\n", card->ext_csd.ffu_capable); 712 MMC_DEV_ATTR(hwrev, "0x%x\n", card->cid.hwrev); 713 MMC_DEV_ATTR(manfid, "0x%06x\n", card->cid.manfid); 714 MMC_DEV_ATTR(name, "%s\n", card->cid.prod_name); 715 MMC_DEV_ATTR(oemid, "0x%04x\n", card->cid.oemid); 716 MMC_DEV_ATTR(prv, "0x%x\n", card->cid.prv); 717 MMC_DEV_ATTR(serial, "0x%08x\n", card->cid.serial); 718 MMC_DEV_ATTR(enhanced_area_offset, "%llu\n", 719 card->ext_csd.enhanced_area_offset); 720 MMC_DEV_ATTR(enhanced_area_size, "%u\n", card->ext_csd.enhanced_area_size); 721 MMC_DEV_ATTR(raw_rpmb_size_mult, "%#x\n", card->ext_csd.raw_rpmb_size_mult); 722 MMC_DEV_ATTR(rel_sectors, "%#x\n", card->ext_csd.rel_sectors); 723 724 static ssize_t mmc_fwrev_show(struct device *dev, 725 struct device_attribute *attr, 726 char *buf) 727 { 728 struct mmc_card *card = mmc_dev_to_card(dev); 729 730 if (card->ext_csd.rev < 7) { 731 return sprintf(buf, "0x%x\n", card->cid.fwrev); 732 } else { 733 return sprintf(buf, "0x%*phN\n", MMC_FIRMWARE_LEN, 734 card->ext_csd.fwrev); 735 } 736 } 737 738 static DEVICE_ATTR(fwrev, S_IRUGO, mmc_fwrev_show, NULL); 739 740 static struct attribute *mmc_std_attrs[] = { 741 &dev_attr_cid.attr, 742 &dev_attr_csd.attr, 743 &dev_attr_date.attr, 744 &dev_attr_erase_size.attr, 745 &dev_attr_preferred_erase_size.attr, 746 &dev_attr_fwrev.attr, 747 &dev_attr_ffu_capable.attr, 748 &dev_attr_hwrev.attr, 749 &dev_attr_manfid.attr, 750 &dev_attr_name.attr, 751 &dev_attr_oemid.attr, 752 &dev_attr_prv.attr, 753 &dev_attr_serial.attr, 754 &dev_attr_enhanced_area_offset.attr, 755 &dev_attr_enhanced_area_size.attr, 756 &dev_attr_raw_rpmb_size_mult.attr, 757 &dev_attr_rel_sectors.attr, 758 NULL, 759 }; 760 ATTRIBUTE_GROUPS(mmc_std); 761 762 static struct device_type mmc_type = { 763 .groups = mmc_std_groups, 764 }; 765 766 /* 767 * Select the PowerClass for the current bus width 768 * If power class is defined for 4/8 bit bus in the 769 * extended CSD register, select it by executing the 770 * mmc_switch command. 771 */ 772 static int __mmc_select_powerclass(struct mmc_card *card, 773 unsigned int bus_width) 774 { 775 struct mmc_host *host = card->host; 776 struct mmc_ext_csd *ext_csd = &card->ext_csd; 777 unsigned int pwrclass_val = 0; 778 int err = 0; 779 780 switch (1 << host->ios.vdd) { 781 case MMC_VDD_165_195: 782 if (host->ios.clock <= MMC_HIGH_26_MAX_DTR) 783 pwrclass_val = ext_csd->raw_pwr_cl_26_195; 784 else if (host->ios.clock <= MMC_HIGH_52_MAX_DTR) 785 pwrclass_val = (bus_width <= EXT_CSD_BUS_WIDTH_8) ? 786 ext_csd->raw_pwr_cl_52_195 : 787 ext_csd->raw_pwr_cl_ddr_52_195; 788 else if (host->ios.clock <= MMC_HS200_MAX_DTR) 789 pwrclass_val = ext_csd->raw_pwr_cl_200_195; 790 break; 791 case MMC_VDD_27_28: 792 case MMC_VDD_28_29: 793 case MMC_VDD_29_30: 794 case MMC_VDD_30_31: 795 case MMC_VDD_31_32: 796 case MMC_VDD_32_33: 797 case MMC_VDD_33_34: 798 case MMC_VDD_34_35: 799 case MMC_VDD_35_36: 800 if (host->ios.clock <= MMC_HIGH_26_MAX_DTR) 801 pwrclass_val = ext_csd->raw_pwr_cl_26_360; 802 else if (host->ios.clock <= MMC_HIGH_52_MAX_DTR) 803 pwrclass_val = (bus_width <= EXT_CSD_BUS_WIDTH_8) ? 804 ext_csd->raw_pwr_cl_52_360 : 805 ext_csd->raw_pwr_cl_ddr_52_360; 806 else if (host->ios.clock <= MMC_HS200_MAX_DTR) 807 pwrclass_val = (bus_width == EXT_CSD_DDR_BUS_WIDTH_8) ? 808 ext_csd->raw_pwr_cl_ddr_200_360 : 809 ext_csd->raw_pwr_cl_200_360; 810 break; 811 default: 812 pr_warn("%s: Voltage range not supported for power class\n", 813 mmc_hostname(host)); 814 return -EINVAL; 815 } 816 817 if (bus_width & (EXT_CSD_BUS_WIDTH_8 | EXT_CSD_DDR_BUS_WIDTH_8)) 818 pwrclass_val = (pwrclass_val & EXT_CSD_PWR_CL_8BIT_MASK) >> 819 EXT_CSD_PWR_CL_8BIT_SHIFT; 820 else 821 pwrclass_val = (pwrclass_val & EXT_CSD_PWR_CL_4BIT_MASK) >> 822 EXT_CSD_PWR_CL_4BIT_SHIFT; 823 824 /* If the power class is different from the default value */ 825 if (pwrclass_val > 0) { 826 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 827 EXT_CSD_POWER_CLASS, 828 pwrclass_val, 829 card->ext_csd.generic_cmd6_time); 830 } 831 832 return err; 833 } 834 835 static int mmc_select_powerclass(struct mmc_card *card) 836 { 837 struct mmc_host *host = card->host; 838 u32 bus_width, ext_csd_bits; 839 int err, ddr; 840 841 /* Power class selection is supported for versions >= 4.0 */ 842 if (!mmc_can_ext_csd(card)) 843 return 0; 844 845 bus_width = host->ios.bus_width; 846 /* Power class values are defined only for 4/8 bit bus */ 847 if (bus_width == MMC_BUS_WIDTH_1) 848 return 0; 849 850 ddr = card->mmc_avail_type & EXT_CSD_CARD_TYPE_DDR_52; 851 if (ddr) 852 ext_csd_bits = (bus_width == MMC_BUS_WIDTH_8) ? 853 EXT_CSD_DDR_BUS_WIDTH_8 : EXT_CSD_DDR_BUS_WIDTH_4; 854 else 855 ext_csd_bits = (bus_width == MMC_BUS_WIDTH_8) ? 856 EXT_CSD_BUS_WIDTH_8 : EXT_CSD_BUS_WIDTH_4; 857 858 err = __mmc_select_powerclass(card, ext_csd_bits); 859 if (err) 860 pr_warn("%s: power class selection to bus width %d ddr %d failed\n", 861 mmc_hostname(host), 1 << bus_width, ddr); 862 863 return err; 864 } 865 866 /* 867 * Set the bus speed for the selected speed mode. 868 */ 869 static void mmc_set_bus_speed(struct mmc_card *card) 870 { 871 unsigned int max_dtr = (unsigned int)-1; 872 873 if ((mmc_card_hs200(card) || mmc_card_hs400(card)) && 874 max_dtr > card->ext_csd.hs200_max_dtr) 875 max_dtr = card->ext_csd.hs200_max_dtr; 876 else if (mmc_card_hs(card) && max_dtr > card->ext_csd.hs_max_dtr) 877 max_dtr = card->ext_csd.hs_max_dtr; 878 else if (max_dtr > card->csd.max_dtr) 879 max_dtr = card->csd.max_dtr; 880 881 mmc_set_clock(card->host, max_dtr); 882 } 883 884 /* 885 * Select the bus width amoung 4-bit and 8-bit(SDR). 886 * If the bus width is changed successfully, return the selected width value. 887 * Zero is returned instead of error value if the wide width is not supported. 888 */ 889 static int mmc_select_bus_width(struct mmc_card *card) 890 { 891 static unsigned ext_csd_bits[] = { 892 EXT_CSD_BUS_WIDTH_8, 893 EXT_CSD_BUS_WIDTH_4, 894 }; 895 static unsigned bus_widths[] = { 896 MMC_BUS_WIDTH_8, 897 MMC_BUS_WIDTH_4, 898 }; 899 struct mmc_host *host = card->host; 900 unsigned idx, bus_width = 0; 901 int err = 0; 902 903 if (!mmc_can_ext_csd(card) || 904 !(host->caps & (MMC_CAP_4_BIT_DATA | MMC_CAP_8_BIT_DATA))) 905 return 0; 906 907 idx = (host->caps & MMC_CAP_8_BIT_DATA) ? 0 : 1; 908 909 /* 910 * Unlike SD, MMC cards dont have a configuration register to notify 911 * supported bus width. So bus test command should be run to identify 912 * the supported bus width or compare the ext csd values of current 913 * bus width and ext csd values of 1 bit mode read earlier. 914 */ 915 for (; idx < ARRAY_SIZE(bus_widths); idx++) { 916 /* 917 * Host is capable of 8bit transfer, then switch 918 * the device to work in 8bit transfer mode. If the 919 * mmc switch command returns error then switch to 920 * 4bit transfer mode. On success set the corresponding 921 * bus width on the host. 922 */ 923 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 924 EXT_CSD_BUS_WIDTH, 925 ext_csd_bits[idx], 926 card->ext_csd.generic_cmd6_time); 927 if (err) 928 continue; 929 930 bus_width = bus_widths[idx]; 931 mmc_set_bus_width(host, bus_width); 932 933 /* 934 * If controller can't handle bus width test, 935 * compare ext_csd previously read in 1 bit mode 936 * against ext_csd at new bus width 937 */ 938 if (!(host->caps & MMC_CAP_BUS_WIDTH_TEST)) 939 err = mmc_compare_ext_csds(card, bus_width); 940 else 941 err = mmc_bus_test(card, bus_width); 942 943 if (!err) { 944 err = bus_width; 945 break; 946 } else { 947 pr_warn("%s: switch to bus width %d failed\n", 948 mmc_hostname(host), ext_csd_bits[idx]); 949 } 950 } 951 952 return err; 953 } 954 955 /* 956 * Switch to the high-speed mode 957 */ 958 static int mmc_select_hs(struct mmc_card *card) 959 { 960 int err; 961 962 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 963 EXT_CSD_HS_TIMING, EXT_CSD_TIMING_HS, 964 card->ext_csd.generic_cmd6_time, 965 true, true, true); 966 if (!err) 967 mmc_set_timing(card->host, MMC_TIMING_MMC_HS); 968 969 return err; 970 } 971 972 /* 973 * Activate wide bus and DDR if supported. 974 */ 975 static int mmc_select_hs_ddr(struct mmc_card *card) 976 { 977 struct mmc_host *host = card->host; 978 u32 bus_width, ext_csd_bits; 979 int err = 0; 980 981 if (!(card->mmc_avail_type & EXT_CSD_CARD_TYPE_DDR_52)) 982 return 0; 983 984 bus_width = host->ios.bus_width; 985 if (bus_width == MMC_BUS_WIDTH_1) 986 return 0; 987 988 ext_csd_bits = (bus_width == MMC_BUS_WIDTH_8) ? 989 EXT_CSD_DDR_BUS_WIDTH_8 : EXT_CSD_DDR_BUS_WIDTH_4; 990 991 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 992 EXT_CSD_BUS_WIDTH, 993 ext_csd_bits, 994 card->ext_csd.generic_cmd6_time); 995 if (err) { 996 pr_err("%s: switch to bus width %d ddr failed\n", 997 mmc_hostname(host), 1 << bus_width); 998 return err; 999 } 1000 1001 /* 1002 * eMMC cards can support 3.3V to 1.2V i/o (vccq) 1003 * signaling. 1004 * 1005 * EXT_CSD_CARD_TYPE_DDR_1_8V means 3.3V or 1.8V vccq. 1006 * 1007 * 1.8V vccq at 3.3V core voltage (vcc) is not required 1008 * in the JEDEC spec for DDR. 1009 * 1010 * Even (e)MMC card can support 3.3v to 1.2v vccq, but not all 1011 * host controller can support this, like some of the SDHCI 1012 * controller which connect to an eMMC device. Some of these 1013 * host controller still needs to use 1.8v vccq for supporting 1014 * DDR mode. 1015 * 1016 * So the sequence will be: 1017 * if (host and device can both support 1.2v IO) 1018 * use 1.2v IO; 1019 * else if (host and device can both support 1.8v IO) 1020 * use 1.8v IO; 1021 * so if host and device can only support 3.3v IO, this is the 1022 * last choice. 1023 * 1024 * WARNING: eMMC rules are NOT the same as SD DDR 1025 */ 1026 err = -EINVAL; 1027 if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_DDR_1_2V) 1028 err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_120); 1029 1030 if (err && (card->mmc_avail_type & EXT_CSD_CARD_TYPE_DDR_1_8V)) 1031 err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_180); 1032 1033 /* make sure vccq is 3.3v after switching disaster */ 1034 if (err) 1035 err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_330); 1036 1037 if (!err) 1038 mmc_set_timing(host, MMC_TIMING_MMC_DDR52); 1039 1040 return err; 1041 } 1042 1043 static int mmc_select_hs400(struct mmc_card *card) 1044 { 1045 struct mmc_host *host = card->host; 1046 int err = 0; 1047 u8 val; 1048 1049 /* 1050 * HS400 mode requires 8-bit bus width 1051 */ 1052 if (!(card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS400 && 1053 host->ios.bus_width == MMC_BUS_WIDTH_8)) 1054 return 0; 1055 1056 /* 1057 * Before switching to dual data rate operation for HS400, 1058 * it is required to convert from HS200 mode to HS mode. 1059 */ 1060 mmc_set_timing(card->host, MMC_TIMING_MMC_HS); 1061 mmc_set_bus_speed(card); 1062 1063 val = EXT_CSD_TIMING_HS | 1064 card->drive_strength << EXT_CSD_DRV_STR_SHIFT; 1065 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1066 EXT_CSD_HS_TIMING, val, 1067 card->ext_csd.generic_cmd6_time, 1068 true, true, true); 1069 if (err) { 1070 pr_err("%s: switch to high-speed from hs200 failed, err:%d\n", 1071 mmc_hostname(host), err); 1072 return err; 1073 } 1074 1075 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1076 EXT_CSD_BUS_WIDTH, 1077 EXT_CSD_DDR_BUS_WIDTH_8, 1078 card->ext_csd.generic_cmd6_time); 1079 if (err) { 1080 pr_err("%s: switch to bus width for hs400 failed, err:%d\n", 1081 mmc_hostname(host), err); 1082 return err; 1083 } 1084 1085 val = EXT_CSD_TIMING_HS400 | 1086 card->drive_strength << EXT_CSD_DRV_STR_SHIFT; 1087 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1088 EXT_CSD_HS_TIMING, val, 1089 card->ext_csd.generic_cmd6_time, 1090 true, true, true); 1091 if (err) { 1092 pr_err("%s: switch to hs400 failed, err:%d\n", 1093 mmc_hostname(host), err); 1094 return err; 1095 } 1096 1097 mmc_set_timing(host, MMC_TIMING_MMC_HS400); 1098 mmc_set_bus_speed(card); 1099 1100 return 0; 1101 } 1102 1103 int mmc_hs200_to_hs400(struct mmc_card *card) 1104 { 1105 return mmc_select_hs400(card); 1106 } 1107 1108 /* Caller must hold re-tuning */ 1109 static int mmc_switch_status(struct mmc_card *card) 1110 { 1111 u32 status; 1112 int err; 1113 1114 err = mmc_send_status(card, &status); 1115 if (err) 1116 return err; 1117 1118 return mmc_switch_status_error(card->host, status); 1119 } 1120 1121 int mmc_hs400_to_hs200(struct mmc_card *card) 1122 { 1123 struct mmc_host *host = card->host; 1124 bool send_status = true; 1125 unsigned int max_dtr; 1126 int err; 1127 u8 val; 1128 1129 if (host->caps & MMC_CAP_WAIT_WHILE_BUSY) 1130 send_status = false; 1131 1132 /* Reduce frequency to HS */ 1133 max_dtr = card->ext_csd.hs_max_dtr; 1134 mmc_set_clock(host, max_dtr); 1135 1136 /* Switch HS400 to HS DDR */ 1137 val = EXT_CSD_TIMING_HS | 1138 card->drive_strength << EXT_CSD_DRV_STR_SHIFT; 1139 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_HS_TIMING, 1140 val, card->ext_csd.generic_cmd6_time, 1141 true, send_status, true); 1142 if (err) 1143 goto out_err; 1144 1145 mmc_set_timing(host, MMC_TIMING_MMC_DDR52); 1146 1147 if (!send_status) { 1148 err = mmc_switch_status(card); 1149 if (err) 1150 goto out_err; 1151 } 1152 1153 /* Switch HS DDR to HS */ 1154 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_BUS_WIDTH, 1155 EXT_CSD_BUS_WIDTH_8, card->ext_csd.generic_cmd6_time, 1156 true, send_status, true); 1157 if (err) 1158 goto out_err; 1159 1160 mmc_set_timing(host, MMC_TIMING_MMC_HS); 1161 1162 if (!send_status) { 1163 err = mmc_switch_status(card); 1164 if (err) 1165 goto out_err; 1166 } 1167 1168 /* Switch HS to HS200 */ 1169 val = EXT_CSD_TIMING_HS200 | 1170 card->drive_strength << EXT_CSD_DRV_STR_SHIFT; 1171 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_HS_TIMING, 1172 val, card->ext_csd.generic_cmd6_time, true, 1173 send_status, true); 1174 if (err) 1175 goto out_err; 1176 1177 mmc_set_timing(host, MMC_TIMING_MMC_HS200); 1178 1179 if (!send_status) { 1180 err = mmc_switch_status(card); 1181 if (err) 1182 goto out_err; 1183 } 1184 1185 mmc_set_bus_speed(card); 1186 1187 return 0; 1188 1189 out_err: 1190 pr_err("%s: %s failed, error %d\n", mmc_hostname(card->host), 1191 __func__, err); 1192 return err; 1193 } 1194 1195 static void mmc_select_driver_type(struct mmc_card *card) 1196 { 1197 int card_drv_type, drive_strength, drv_type; 1198 1199 card_drv_type = card->ext_csd.raw_driver_strength | 1200 mmc_driver_type_mask(0); 1201 1202 drive_strength = mmc_select_drive_strength(card, 1203 card->ext_csd.hs200_max_dtr, 1204 card_drv_type, &drv_type); 1205 1206 card->drive_strength = drive_strength; 1207 1208 if (drv_type) 1209 mmc_set_driver_type(card->host, drv_type); 1210 } 1211 1212 /* 1213 * For device supporting HS200 mode, the following sequence 1214 * should be done before executing the tuning process. 1215 * 1. set the desired bus width(4-bit or 8-bit, 1-bit is not supported) 1216 * 2. switch to HS200 mode 1217 * 3. set the clock to > 52Mhz and <=200MHz 1218 */ 1219 static int mmc_select_hs200(struct mmc_card *card) 1220 { 1221 struct mmc_host *host = card->host; 1222 int err = -EINVAL; 1223 u8 val; 1224 1225 if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS200_1_2V) 1226 err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_120); 1227 1228 if (err && card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS200_1_8V) 1229 err = __mmc_set_signal_voltage(host, MMC_SIGNAL_VOLTAGE_180); 1230 1231 /* If fails try again during next card power cycle */ 1232 if (err) 1233 goto err; 1234 1235 mmc_select_driver_type(card); 1236 1237 /* 1238 * Set the bus width(4 or 8) with host's support and 1239 * switch to HS200 mode if bus width is set successfully. 1240 */ 1241 err = mmc_select_bus_width(card); 1242 if (!IS_ERR_VALUE(err)) { 1243 val = EXT_CSD_TIMING_HS200 | 1244 card->drive_strength << EXT_CSD_DRV_STR_SHIFT; 1245 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1246 EXT_CSD_HS_TIMING, val, 1247 card->ext_csd.generic_cmd6_time, 1248 true, true, true); 1249 if (!err) 1250 mmc_set_timing(host, MMC_TIMING_MMC_HS200); 1251 } 1252 err: 1253 return err; 1254 } 1255 1256 /* 1257 * Activate High Speed or HS200 mode if supported. 1258 */ 1259 static int mmc_select_timing(struct mmc_card *card) 1260 { 1261 int err = 0; 1262 1263 if (!mmc_can_ext_csd(card)) 1264 goto bus_speed; 1265 1266 if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS200) 1267 err = mmc_select_hs200(card); 1268 else if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS) 1269 err = mmc_select_hs(card); 1270 1271 if (err && err != -EBADMSG) 1272 return err; 1273 1274 if (err) { 1275 pr_warn("%s: switch to %s failed\n", 1276 mmc_card_hs(card) ? "high-speed" : 1277 (mmc_card_hs200(card) ? "hs200" : ""), 1278 mmc_hostname(card->host)); 1279 err = 0; 1280 } 1281 1282 bus_speed: 1283 /* 1284 * Set the bus speed to the selected bus timing. 1285 * If timing is not selected, backward compatible is the default. 1286 */ 1287 mmc_set_bus_speed(card); 1288 return err; 1289 } 1290 1291 /* 1292 * Execute tuning sequence to seek the proper bus operating 1293 * conditions for HS200 and HS400, which sends CMD21 to the device. 1294 */ 1295 static int mmc_hs200_tuning(struct mmc_card *card) 1296 { 1297 struct mmc_host *host = card->host; 1298 1299 /* 1300 * Timing should be adjusted to the HS400 target 1301 * operation frequency for tuning process 1302 */ 1303 if (card->mmc_avail_type & EXT_CSD_CARD_TYPE_HS400 && 1304 host->ios.bus_width == MMC_BUS_WIDTH_8) 1305 if (host->ops->prepare_hs400_tuning) 1306 host->ops->prepare_hs400_tuning(host, &host->ios); 1307 1308 return mmc_execute_tuning(card); 1309 } 1310 1311 /* 1312 * Handle the detection and initialisation of a card. 1313 * 1314 * In the case of a resume, "oldcard" will contain the card 1315 * we're trying to reinitialise. 1316 */ 1317 static int mmc_init_card(struct mmc_host *host, u32 ocr, 1318 struct mmc_card *oldcard) 1319 { 1320 struct mmc_card *card; 1321 int err; 1322 u32 cid[4]; 1323 u32 rocr; 1324 1325 BUG_ON(!host); 1326 WARN_ON(!host->claimed); 1327 1328 /* Set correct bus mode for MMC before attempting init */ 1329 if (!mmc_host_is_spi(host)) 1330 mmc_set_bus_mode(host, MMC_BUSMODE_OPENDRAIN); 1331 1332 /* 1333 * Since we're changing the OCR value, we seem to 1334 * need to tell some cards to go back to the idle 1335 * state. We wait 1ms to give cards time to 1336 * respond. 1337 * mmc_go_idle is needed for eMMC that are asleep 1338 */ 1339 mmc_go_idle(host); 1340 1341 /* The extra bit indicates that we support high capacity */ 1342 err = mmc_send_op_cond(host, ocr | (1 << 30), &rocr); 1343 if (err) 1344 goto err; 1345 1346 /* 1347 * For SPI, enable CRC as appropriate. 1348 */ 1349 if (mmc_host_is_spi(host)) { 1350 err = mmc_spi_set_crc(host, use_spi_crc); 1351 if (err) 1352 goto err; 1353 } 1354 1355 /* 1356 * Fetch CID from card. 1357 */ 1358 if (mmc_host_is_spi(host)) 1359 err = mmc_send_cid(host, cid); 1360 else 1361 err = mmc_all_send_cid(host, cid); 1362 if (err) 1363 goto err; 1364 1365 if (oldcard) { 1366 if (memcmp(cid, oldcard->raw_cid, sizeof(cid)) != 0) { 1367 err = -ENOENT; 1368 goto err; 1369 } 1370 1371 card = oldcard; 1372 } else { 1373 /* 1374 * Allocate card structure. 1375 */ 1376 card = mmc_alloc_card(host, &mmc_type); 1377 if (IS_ERR(card)) { 1378 err = PTR_ERR(card); 1379 goto err; 1380 } 1381 1382 card->ocr = ocr; 1383 card->type = MMC_TYPE_MMC; 1384 card->rca = 1; 1385 memcpy(card->raw_cid, cid, sizeof(card->raw_cid)); 1386 } 1387 1388 /* 1389 * Call the optional HC's init_card function to handle quirks. 1390 */ 1391 if (host->ops->init_card) 1392 host->ops->init_card(host, card); 1393 1394 /* 1395 * For native busses: set card RCA and quit open drain mode. 1396 */ 1397 if (!mmc_host_is_spi(host)) { 1398 err = mmc_set_relative_addr(card); 1399 if (err) 1400 goto free_card; 1401 1402 mmc_set_bus_mode(host, MMC_BUSMODE_PUSHPULL); 1403 } 1404 1405 if (!oldcard) { 1406 /* 1407 * Fetch CSD from card. 1408 */ 1409 err = mmc_send_csd(card, card->raw_csd); 1410 if (err) 1411 goto free_card; 1412 1413 err = mmc_decode_csd(card); 1414 if (err) 1415 goto free_card; 1416 err = mmc_decode_cid(card); 1417 if (err) 1418 goto free_card; 1419 } 1420 1421 /* 1422 * handling only for cards supporting DSR and hosts requesting 1423 * DSR configuration 1424 */ 1425 if (card->csd.dsr_imp && host->dsr_req) 1426 mmc_set_dsr(host); 1427 1428 /* 1429 * Select card, as all following commands rely on that. 1430 */ 1431 if (!mmc_host_is_spi(host)) { 1432 err = mmc_select_card(card); 1433 if (err) 1434 goto free_card; 1435 } 1436 1437 if (!oldcard) { 1438 /* Read extended CSD. */ 1439 err = mmc_read_ext_csd(card); 1440 if (err) 1441 goto free_card; 1442 1443 /* If doing byte addressing, check if required to do sector 1444 * addressing. Handle the case of <2GB cards needing sector 1445 * addressing. See section 8.1 JEDEC Standard JED84-A441; 1446 * ocr register has bit 30 set for sector addressing. 1447 */ 1448 if (!(mmc_card_blockaddr(card)) && (rocr & (1<<30))) 1449 mmc_card_set_blockaddr(card); 1450 1451 /* Erase size depends on CSD and Extended CSD */ 1452 mmc_set_erase_size(card); 1453 } 1454 1455 /* 1456 * If enhanced_area_en is TRUE, host needs to enable ERASE_GRP_DEF 1457 * bit. This bit will be lost every time after a reset or power off. 1458 */ 1459 if (card->ext_csd.partition_setting_completed || 1460 (card->ext_csd.rev >= 3 && (host->caps2 & MMC_CAP2_HC_ERASE_SZ))) { 1461 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1462 EXT_CSD_ERASE_GROUP_DEF, 1, 1463 card->ext_csd.generic_cmd6_time); 1464 1465 if (err && err != -EBADMSG) 1466 goto free_card; 1467 1468 if (err) { 1469 err = 0; 1470 /* 1471 * Just disable enhanced area off & sz 1472 * will try to enable ERASE_GROUP_DEF 1473 * during next time reinit 1474 */ 1475 card->ext_csd.enhanced_area_offset = -EINVAL; 1476 card->ext_csd.enhanced_area_size = -EINVAL; 1477 } else { 1478 card->ext_csd.erase_group_def = 1; 1479 /* 1480 * enable ERASE_GRP_DEF successfully. 1481 * This will affect the erase size, so 1482 * here need to reset erase size 1483 */ 1484 mmc_set_erase_size(card); 1485 } 1486 } 1487 1488 /* 1489 * Ensure eMMC user default partition is enabled 1490 */ 1491 if (card->ext_csd.part_config & EXT_CSD_PART_CONFIG_ACC_MASK) { 1492 card->ext_csd.part_config &= ~EXT_CSD_PART_CONFIG_ACC_MASK; 1493 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, EXT_CSD_PART_CONFIG, 1494 card->ext_csd.part_config, 1495 card->ext_csd.part_time); 1496 if (err && err != -EBADMSG) 1497 goto free_card; 1498 } 1499 1500 /* 1501 * Enable power_off_notification byte in the ext_csd register 1502 */ 1503 if (card->ext_csd.rev >= 6) { 1504 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1505 EXT_CSD_POWER_OFF_NOTIFICATION, 1506 EXT_CSD_POWER_ON, 1507 card->ext_csd.generic_cmd6_time); 1508 if (err && err != -EBADMSG) 1509 goto free_card; 1510 1511 /* 1512 * The err can be -EBADMSG or 0, 1513 * so check for success and update the flag 1514 */ 1515 if (!err) 1516 card->ext_csd.power_off_notification = EXT_CSD_POWER_ON; 1517 } 1518 1519 /* 1520 * Select timing interface 1521 */ 1522 err = mmc_select_timing(card); 1523 if (err) 1524 goto free_card; 1525 1526 if (mmc_card_hs200(card)) { 1527 err = mmc_hs200_tuning(card); 1528 if (err) 1529 goto free_card; 1530 1531 err = mmc_select_hs400(card); 1532 if (err) 1533 goto free_card; 1534 } else if (mmc_card_hs(card)) { 1535 /* Select the desired bus width optionally */ 1536 err = mmc_select_bus_width(card); 1537 if (!IS_ERR_VALUE(err)) { 1538 err = mmc_select_hs_ddr(card); 1539 if (err) 1540 goto free_card; 1541 } 1542 } 1543 1544 /* 1545 * Choose the power class with selected bus interface 1546 */ 1547 mmc_select_powerclass(card); 1548 1549 /* 1550 * Enable HPI feature (if supported) 1551 */ 1552 if (card->ext_csd.hpi) { 1553 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1554 EXT_CSD_HPI_MGMT, 1, 1555 card->ext_csd.generic_cmd6_time); 1556 if (err && err != -EBADMSG) 1557 goto free_card; 1558 if (err) { 1559 pr_warn("%s: Enabling HPI failed\n", 1560 mmc_hostname(card->host)); 1561 err = 0; 1562 } else 1563 card->ext_csd.hpi_en = 1; 1564 } 1565 1566 /* 1567 * If cache size is higher than 0, this indicates 1568 * the existence of cache and it can be turned on. 1569 */ 1570 if (card->ext_csd.cache_size > 0) { 1571 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1572 EXT_CSD_CACHE_CTRL, 1, 1573 card->ext_csd.generic_cmd6_time); 1574 if (err && err != -EBADMSG) 1575 goto free_card; 1576 1577 /* 1578 * Only if no error, cache is turned on successfully. 1579 */ 1580 if (err) { 1581 pr_warn("%s: Cache is supported, but failed to turn on (%d)\n", 1582 mmc_hostname(card->host), err); 1583 card->ext_csd.cache_ctrl = 0; 1584 err = 0; 1585 } else { 1586 card->ext_csd.cache_ctrl = 1; 1587 } 1588 } 1589 1590 /* 1591 * The mandatory minimum values are defined for packed command. 1592 * read: 5, write: 3 1593 */ 1594 if (card->ext_csd.max_packed_writes >= 3 && 1595 card->ext_csd.max_packed_reads >= 5 && 1596 host->caps2 & MMC_CAP2_PACKED_CMD) { 1597 err = mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1598 EXT_CSD_EXP_EVENTS_CTRL, 1599 EXT_CSD_PACKED_EVENT_EN, 1600 card->ext_csd.generic_cmd6_time); 1601 if (err && err != -EBADMSG) 1602 goto free_card; 1603 if (err) { 1604 pr_warn("%s: Enabling packed event failed\n", 1605 mmc_hostname(card->host)); 1606 card->ext_csd.packed_event_en = 0; 1607 err = 0; 1608 } else { 1609 card->ext_csd.packed_event_en = 1; 1610 } 1611 } 1612 1613 if (!oldcard) 1614 host->card = card; 1615 1616 return 0; 1617 1618 free_card: 1619 if (!oldcard) 1620 mmc_remove_card(card); 1621 err: 1622 return err; 1623 } 1624 1625 static int mmc_can_sleep(struct mmc_card *card) 1626 { 1627 return (card && card->ext_csd.rev >= 3); 1628 } 1629 1630 static int mmc_sleep(struct mmc_host *host) 1631 { 1632 struct mmc_command cmd = {0}; 1633 struct mmc_card *card = host->card; 1634 unsigned int timeout_ms = DIV_ROUND_UP(card->ext_csd.sa_timeout, 10000); 1635 int err; 1636 1637 /* Re-tuning can't be done once the card is deselected */ 1638 mmc_retune_hold(host); 1639 1640 err = mmc_deselect_cards(host); 1641 if (err) 1642 goto out_release; 1643 1644 cmd.opcode = MMC_SLEEP_AWAKE; 1645 cmd.arg = card->rca << 16; 1646 cmd.arg |= 1 << 15; 1647 1648 /* 1649 * If the max_busy_timeout of the host is specified, validate it against 1650 * the sleep cmd timeout. A failure means we need to prevent the host 1651 * from doing hw busy detection, which is done by converting to a R1 1652 * response instead of a R1B. 1653 */ 1654 if (host->max_busy_timeout && (timeout_ms > host->max_busy_timeout)) { 1655 cmd.flags = MMC_RSP_R1 | MMC_CMD_AC; 1656 } else { 1657 cmd.flags = MMC_RSP_R1B | MMC_CMD_AC; 1658 cmd.busy_timeout = timeout_ms; 1659 } 1660 1661 err = mmc_wait_for_cmd(host, &cmd, 0); 1662 if (err) 1663 goto out_release; 1664 1665 /* 1666 * If the host does not wait while the card signals busy, then we will 1667 * will have to wait the sleep/awake timeout. Note, we cannot use the 1668 * SEND_STATUS command to poll the status because that command (and most 1669 * others) is invalid while the card sleeps. 1670 */ 1671 if (!cmd.busy_timeout || !(host->caps & MMC_CAP_WAIT_WHILE_BUSY)) 1672 mmc_delay(timeout_ms); 1673 1674 out_release: 1675 mmc_retune_release(host); 1676 return err; 1677 } 1678 1679 static int mmc_can_poweroff_notify(const struct mmc_card *card) 1680 { 1681 return card && 1682 mmc_card_mmc(card) && 1683 (card->ext_csd.power_off_notification == EXT_CSD_POWER_ON); 1684 } 1685 1686 static int mmc_poweroff_notify(struct mmc_card *card, unsigned int notify_type) 1687 { 1688 unsigned int timeout = card->ext_csd.generic_cmd6_time; 1689 int err; 1690 1691 /* Use EXT_CSD_POWER_OFF_SHORT as default notification type. */ 1692 if (notify_type == EXT_CSD_POWER_OFF_LONG) 1693 timeout = card->ext_csd.power_off_longtime; 1694 1695 err = __mmc_switch(card, EXT_CSD_CMD_SET_NORMAL, 1696 EXT_CSD_POWER_OFF_NOTIFICATION, 1697 notify_type, timeout, true, false, false); 1698 if (err) 1699 pr_err("%s: Power Off Notification timed out, %u\n", 1700 mmc_hostname(card->host), timeout); 1701 1702 /* Disable the power off notification after the switch operation. */ 1703 card->ext_csd.power_off_notification = EXT_CSD_NO_POWER_NOTIFICATION; 1704 1705 return err; 1706 } 1707 1708 /* 1709 * Host is being removed. Free up the current card. 1710 */ 1711 static void mmc_remove(struct mmc_host *host) 1712 { 1713 BUG_ON(!host); 1714 BUG_ON(!host->card); 1715 1716 mmc_remove_card(host->card); 1717 host->card = NULL; 1718 } 1719 1720 /* 1721 * Card detection - card is alive. 1722 */ 1723 static int mmc_alive(struct mmc_host *host) 1724 { 1725 return mmc_send_status(host->card, NULL); 1726 } 1727 1728 /* 1729 * Card detection callback from host. 1730 */ 1731 static void mmc_detect(struct mmc_host *host) 1732 { 1733 int err; 1734 1735 BUG_ON(!host); 1736 BUG_ON(!host->card); 1737 1738 mmc_get_card(host->card); 1739 1740 /* 1741 * Just check if our card has been removed. 1742 */ 1743 err = _mmc_detect_card_removed(host); 1744 1745 mmc_put_card(host->card); 1746 1747 if (err) { 1748 mmc_remove(host); 1749 1750 mmc_claim_host(host); 1751 mmc_detach_bus(host); 1752 mmc_power_off(host); 1753 mmc_release_host(host); 1754 } 1755 } 1756 1757 static int _mmc_suspend(struct mmc_host *host, bool is_suspend) 1758 { 1759 int err = 0; 1760 unsigned int notify_type = is_suspend ? EXT_CSD_POWER_OFF_SHORT : 1761 EXT_CSD_POWER_OFF_LONG; 1762 1763 BUG_ON(!host); 1764 BUG_ON(!host->card); 1765 1766 mmc_claim_host(host); 1767 1768 if (mmc_card_suspended(host->card)) 1769 goto out; 1770 1771 if (mmc_card_doing_bkops(host->card)) { 1772 err = mmc_stop_bkops(host->card); 1773 if (err) 1774 goto out; 1775 } 1776 1777 err = mmc_flush_cache(host->card); 1778 if (err) 1779 goto out; 1780 1781 if (mmc_can_poweroff_notify(host->card) && 1782 ((host->caps2 & MMC_CAP2_FULL_PWR_CYCLE) || !is_suspend)) 1783 err = mmc_poweroff_notify(host->card, notify_type); 1784 else if (mmc_can_sleep(host->card)) 1785 err = mmc_sleep(host); 1786 else if (!mmc_host_is_spi(host)) 1787 err = mmc_deselect_cards(host); 1788 1789 if (!err) { 1790 mmc_power_off(host); 1791 mmc_card_set_suspended(host->card); 1792 } 1793 out: 1794 mmc_release_host(host); 1795 return err; 1796 } 1797 1798 /* 1799 * Suspend callback 1800 */ 1801 static int mmc_suspend(struct mmc_host *host) 1802 { 1803 int err; 1804 1805 err = _mmc_suspend(host, true); 1806 if (!err) { 1807 pm_runtime_disable(&host->card->dev); 1808 pm_runtime_set_suspended(&host->card->dev); 1809 } 1810 1811 return err; 1812 } 1813 1814 /* 1815 * This function tries to determine if the same card is still present 1816 * and, if so, restore all state to it. 1817 */ 1818 static int _mmc_resume(struct mmc_host *host) 1819 { 1820 int err = 0; 1821 1822 BUG_ON(!host); 1823 BUG_ON(!host->card); 1824 1825 mmc_claim_host(host); 1826 1827 if (!mmc_card_suspended(host->card)) 1828 goto out; 1829 1830 mmc_power_up(host, host->card->ocr); 1831 err = mmc_init_card(host, host->card->ocr, host->card); 1832 mmc_card_clr_suspended(host->card); 1833 1834 out: 1835 mmc_release_host(host); 1836 return err; 1837 } 1838 1839 /* 1840 * Shutdown callback 1841 */ 1842 static int mmc_shutdown(struct mmc_host *host) 1843 { 1844 int err = 0; 1845 1846 /* 1847 * In a specific case for poweroff notify, we need to resume the card 1848 * before we can shutdown it properly. 1849 */ 1850 if (mmc_can_poweroff_notify(host->card) && 1851 !(host->caps2 & MMC_CAP2_FULL_PWR_CYCLE)) 1852 err = _mmc_resume(host); 1853 1854 if (!err) 1855 err = _mmc_suspend(host, false); 1856 1857 return err; 1858 } 1859 1860 /* 1861 * Callback for resume. 1862 */ 1863 static int mmc_resume(struct mmc_host *host) 1864 { 1865 int err = 0; 1866 1867 if (!(host->caps & MMC_CAP_RUNTIME_RESUME)) { 1868 err = _mmc_resume(host); 1869 pm_runtime_set_active(&host->card->dev); 1870 pm_runtime_mark_last_busy(&host->card->dev); 1871 } 1872 pm_runtime_enable(&host->card->dev); 1873 1874 return err; 1875 } 1876 1877 /* 1878 * Callback for runtime_suspend. 1879 */ 1880 static int mmc_runtime_suspend(struct mmc_host *host) 1881 { 1882 int err; 1883 1884 if (!(host->caps & MMC_CAP_AGGRESSIVE_PM)) 1885 return 0; 1886 1887 err = _mmc_suspend(host, true); 1888 if (err) 1889 pr_err("%s: error %d doing aggressive suspend\n", 1890 mmc_hostname(host), err); 1891 1892 return err; 1893 } 1894 1895 /* 1896 * Callback for runtime_resume. 1897 */ 1898 static int mmc_runtime_resume(struct mmc_host *host) 1899 { 1900 int err; 1901 1902 if (!(host->caps & (MMC_CAP_AGGRESSIVE_PM | MMC_CAP_RUNTIME_RESUME))) 1903 return 0; 1904 1905 err = _mmc_resume(host); 1906 if (err) 1907 pr_err("%s: error %d doing aggressive resume\n", 1908 mmc_hostname(host), err); 1909 1910 return 0; 1911 } 1912 1913 int mmc_can_reset(struct mmc_card *card) 1914 { 1915 u8 rst_n_function; 1916 1917 rst_n_function = card->ext_csd.rst_n_function; 1918 if ((rst_n_function & EXT_CSD_RST_N_EN_MASK) != EXT_CSD_RST_N_ENABLED) 1919 return 0; 1920 return 1; 1921 } 1922 EXPORT_SYMBOL(mmc_can_reset); 1923 1924 static int mmc_reset(struct mmc_host *host) 1925 { 1926 struct mmc_card *card = host->card; 1927 1928 if (!(host->caps & MMC_CAP_HW_RESET) || !host->ops->hw_reset) 1929 return -EOPNOTSUPP; 1930 1931 if (!mmc_can_reset(card)) 1932 return -EOPNOTSUPP; 1933 1934 mmc_host_clk_hold(host); 1935 mmc_set_clock(host, host->f_init); 1936 1937 host->ops->hw_reset(host); 1938 1939 /* Set initial state and call mmc_set_ios */ 1940 mmc_set_initial_state(host); 1941 mmc_host_clk_release(host); 1942 1943 return mmc_init_card(host, card->ocr, card); 1944 } 1945 1946 static const struct mmc_bus_ops mmc_ops = { 1947 .remove = mmc_remove, 1948 .detect = mmc_detect, 1949 .suspend = mmc_suspend, 1950 .resume = mmc_resume, 1951 .runtime_suspend = mmc_runtime_suspend, 1952 .runtime_resume = mmc_runtime_resume, 1953 .alive = mmc_alive, 1954 .shutdown = mmc_shutdown, 1955 .reset = mmc_reset, 1956 }; 1957 1958 /* 1959 * Starting point for MMC card init. 1960 */ 1961 int mmc_attach_mmc(struct mmc_host *host) 1962 { 1963 int err; 1964 u32 ocr, rocr; 1965 1966 BUG_ON(!host); 1967 WARN_ON(!host->claimed); 1968 1969 /* Set correct bus mode for MMC before attempting attach */ 1970 if (!mmc_host_is_spi(host)) 1971 mmc_set_bus_mode(host, MMC_BUSMODE_OPENDRAIN); 1972 1973 err = mmc_send_op_cond(host, 0, &ocr); 1974 if (err) 1975 return err; 1976 1977 mmc_attach_bus(host, &mmc_ops); 1978 if (host->ocr_avail_mmc) 1979 host->ocr_avail = host->ocr_avail_mmc; 1980 1981 /* 1982 * We need to get OCR a different way for SPI. 1983 */ 1984 if (mmc_host_is_spi(host)) { 1985 err = mmc_spi_read_ocr(host, 1, &ocr); 1986 if (err) 1987 goto err; 1988 } 1989 1990 rocr = mmc_select_voltage(host, ocr); 1991 1992 /* 1993 * Can we support the voltage of the card? 1994 */ 1995 if (!rocr) { 1996 err = -EINVAL; 1997 goto err; 1998 } 1999 2000 /* 2001 * Detect and init the card. 2002 */ 2003 err = mmc_init_card(host, rocr, NULL); 2004 if (err) 2005 goto err; 2006 2007 mmc_release_host(host); 2008 err = mmc_add_card(host->card); 2009 mmc_claim_host(host); 2010 if (err) 2011 goto remove_card; 2012 2013 return 0; 2014 2015 remove_card: 2016 mmc_release_host(host); 2017 mmc_remove_card(host->card); 2018 mmc_claim_host(host); 2019 host->card = NULL; 2020 err: 2021 mmc_detach_bus(host); 2022 2023 pr_err("%s: error %d whilst initialising MMC card\n", 2024 mmc_hostname(host), err); 2025 2026 return err; 2027 } 2028