1 /* 2 * PMC-Sierra SPC 8001 SAS/SATA based host adapters driver 3 * 4 * Copyright (c) 2008-2009 USI Co., Ltd. 5 * All rights reserved. 6 * 7 * Redistribution and use in source and binary forms, with or without 8 * modification, are permitted provided that the following conditions 9 * are met: 10 * 1. Redistributions of source code must retain the above copyright 11 * notice, this list of conditions, and the following disclaimer, 12 * without modification. 13 * 2. Redistributions in binary form must reproduce at minimum a disclaimer 14 * substantially similar to the "NO WARRANTY" disclaimer below 15 * ("Disclaimer") and any redistribution must be conditioned upon 16 * including a substantially similar Disclaimer requirement for further 17 * binary redistribution. 18 * 3. Neither the names of the above-listed copyright holders nor the names 19 * of any contributors may be used to endorse or promote products derived 20 * from this software without specific prior written permission. 21 * 22 * Alternatively, this software may be distributed under the terms of the 23 * GNU General Public License ("GPL") version 2 as published by the Free 24 * Software Foundation. 25 * 26 * NO WARRANTY 27 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS 28 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT 29 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTIBILITY AND FITNESS FOR 30 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT 31 * HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY, OR CONSEQUENTIAL 32 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS 33 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) 34 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, 35 * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING 36 * IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE 37 * POSSIBILITY OF SUCH DAMAGES. 38 * 39 */ 40 #include <linux/slab.h> 41 #include "pm8001_sas.h" 42 #include "pm8001_hwi.h" 43 #include "pm8001_chips.h" 44 #include "pm8001_ctl.h" 45 46 /** 47 * read_main_config_table - read the configure table and save it. 48 * @pm8001_ha: our hba card information 49 */ 50 static void read_main_config_table(struct pm8001_hba_info *pm8001_ha) 51 { 52 void __iomem *address = pm8001_ha->main_cfg_tbl_addr; 53 pm8001_ha->main_cfg_tbl.pm8001_tbl.signature = 54 pm8001_mr32(address, 0x00); 55 pm8001_ha->main_cfg_tbl.pm8001_tbl.interface_rev = 56 pm8001_mr32(address, 0x04); 57 pm8001_ha->main_cfg_tbl.pm8001_tbl.firmware_rev = 58 pm8001_mr32(address, 0x08); 59 pm8001_ha->main_cfg_tbl.pm8001_tbl.max_out_io = 60 pm8001_mr32(address, 0x0C); 61 pm8001_ha->main_cfg_tbl.pm8001_tbl.max_sgl = 62 pm8001_mr32(address, 0x10); 63 pm8001_ha->main_cfg_tbl.pm8001_tbl.ctrl_cap_flag = 64 pm8001_mr32(address, 0x14); 65 pm8001_ha->main_cfg_tbl.pm8001_tbl.gst_offset = 66 pm8001_mr32(address, 0x18); 67 pm8001_ha->main_cfg_tbl.pm8001_tbl.inbound_queue_offset = 68 pm8001_mr32(address, MAIN_IBQ_OFFSET); 69 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_queue_offset = 70 pm8001_mr32(address, MAIN_OBQ_OFFSET); 71 pm8001_ha->main_cfg_tbl.pm8001_tbl.hda_mode_flag = 72 pm8001_mr32(address, MAIN_HDA_FLAGS_OFFSET); 73 74 /* read analog Setting offset from the configuration table */ 75 pm8001_ha->main_cfg_tbl.pm8001_tbl.anolog_setup_table_offset = 76 pm8001_mr32(address, MAIN_ANALOG_SETUP_OFFSET); 77 78 /* read Error Dump Offset and Length */ 79 pm8001_ha->main_cfg_tbl.pm8001_tbl.fatal_err_dump_offset0 = 80 pm8001_mr32(address, MAIN_FATAL_ERROR_RDUMP0_OFFSET); 81 pm8001_ha->main_cfg_tbl.pm8001_tbl.fatal_err_dump_length0 = 82 pm8001_mr32(address, MAIN_FATAL_ERROR_RDUMP0_LENGTH); 83 pm8001_ha->main_cfg_tbl.pm8001_tbl.fatal_err_dump_offset1 = 84 pm8001_mr32(address, MAIN_FATAL_ERROR_RDUMP1_OFFSET); 85 pm8001_ha->main_cfg_tbl.pm8001_tbl.fatal_err_dump_length1 = 86 pm8001_mr32(address, MAIN_FATAL_ERROR_RDUMP1_LENGTH); 87 } 88 89 /** 90 * read_general_status_table - read the general status table and save it. 91 * @pm8001_ha: our hba card information 92 */ 93 static void read_general_status_table(struct pm8001_hba_info *pm8001_ha) 94 { 95 void __iomem *address = pm8001_ha->general_stat_tbl_addr; 96 pm8001_ha->gs_tbl.pm8001_tbl.gst_len_mpistate = 97 pm8001_mr32(address, 0x00); 98 pm8001_ha->gs_tbl.pm8001_tbl.iq_freeze_state0 = 99 pm8001_mr32(address, 0x04); 100 pm8001_ha->gs_tbl.pm8001_tbl.iq_freeze_state1 = 101 pm8001_mr32(address, 0x08); 102 pm8001_ha->gs_tbl.pm8001_tbl.msgu_tcnt = 103 pm8001_mr32(address, 0x0C); 104 pm8001_ha->gs_tbl.pm8001_tbl.iop_tcnt = 105 pm8001_mr32(address, 0x10); 106 pm8001_ha->gs_tbl.pm8001_tbl.rsvd = 107 pm8001_mr32(address, 0x14); 108 pm8001_ha->gs_tbl.pm8001_tbl.phy_state[0] = 109 pm8001_mr32(address, 0x18); 110 pm8001_ha->gs_tbl.pm8001_tbl.phy_state[1] = 111 pm8001_mr32(address, 0x1C); 112 pm8001_ha->gs_tbl.pm8001_tbl.phy_state[2] = 113 pm8001_mr32(address, 0x20); 114 pm8001_ha->gs_tbl.pm8001_tbl.phy_state[3] = 115 pm8001_mr32(address, 0x24); 116 pm8001_ha->gs_tbl.pm8001_tbl.phy_state[4] = 117 pm8001_mr32(address, 0x28); 118 pm8001_ha->gs_tbl.pm8001_tbl.phy_state[5] = 119 pm8001_mr32(address, 0x2C); 120 pm8001_ha->gs_tbl.pm8001_tbl.phy_state[6] = 121 pm8001_mr32(address, 0x30); 122 pm8001_ha->gs_tbl.pm8001_tbl.phy_state[7] = 123 pm8001_mr32(address, 0x34); 124 pm8001_ha->gs_tbl.pm8001_tbl.gpio_input_val = 125 pm8001_mr32(address, 0x38); 126 pm8001_ha->gs_tbl.pm8001_tbl.rsvd1[0] = 127 pm8001_mr32(address, 0x3C); 128 pm8001_ha->gs_tbl.pm8001_tbl.rsvd1[1] = 129 pm8001_mr32(address, 0x40); 130 pm8001_ha->gs_tbl.pm8001_tbl.recover_err_info[0] = 131 pm8001_mr32(address, 0x44); 132 pm8001_ha->gs_tbl.pm8001_tbl.recover_err_info[1] = 133 pm8001_mr32(address, 0x48); 134 pm8001_ha->gs_tbl.pm8001_tbl.recover_err_info[2] = 135 pm8001_mr32(address, 0x4C); 136 pm8001_ha->gs_tbl.pm8001_tbl.recover_err_info[3] = 137 pm8001_mr32(address, 0x50); 138 pm8001_ha->gs_tbl.pm8001_tbl.recover_err_info[4] = 139 pm8001_mr32(address, 0x54); 140 pm8001_ha->gs_tbl.pm8001_tbl.recover_err_info[5] = 141 pm8001_mr32(address, 0x58); 142 pm8001_ha->gs_tbl.pm8001_tbl.recover_err_info[6] = 143 pm8001_mr32(address, 0x5C); 144 pm8001_ha->gs_tbl.pm8001_tbl.recover_err_info[7] = 145 pm8001_mr32(address, 0x60); 146 } 147 148 /** 149 * read_inbnd_queue_table - read the inbound queue table and save it. 150 * @pm8001_ha: our hba card information 151 */ 152 static void read_inbnd_queue_table(struct pm8001_hba_info *pm8001_ha) 153 { 154 int i; 155 void __iomem *address = pm8001_ha->inbnd_q_tbl_addr; 156 for (i = 0; i < PM8001_MAX_INB_NUM; i++) { 157 u32 offset = i * 0x20; 158 pm8001_ha->inbnd_q_tbl[i].pi_pci_bar = 159 get_pci_bar_index(pm8001_mr32(address, (offset + 0x14))); 160 pm8001_ha->inbnd_q_tbl[i].pi_offset = 161 pm8001_mr32(address, (offset + 0x18)); 162 } 163 } 164 165 /** 166 * read_outbnd_queue_table - read the outbound queue table and save it. 167 * @pm8001_ha: our hba card information 168 */ 169 static void read_outbnd_queue_table(struct pm8001_hba_info *pm8001_ha) 170 { 171 int i; 172 void __iomem *address = pm8001_ha->outbnd_q_tbl_addr; 173 for (i = 0; i < PM8001_MAX_OUTB_NUM; i++) { 174 u32 offset = i * 0x24; 175 pm8001_ha->outbnd_q_tbl[i].ci_pci_bar = 176 get_pci_bar_index(pm8001_mr32(address, (offset + 0x14))); 177 pm8001_ha->outbnd_q_tbl[i].ci_offset = 178 pm8001_mr32(address, (offset + 0x18)); 179 } 180 } 181 182 /** 183 * init_default_table_values - init the default table. 184 * @pm8001_ha: our hba card information 185 */ 186 static void init_default_table_values(struct pm8001_hba_info *pm8001_ha) 187 { 188 int i; 189 u32 offsetib, offsetob; 190 void __iomem *addressib = pm8001_ha->inbnd_q_tbl_addr; 191 void __iomem *addressob = pm8001_ha->outbnd_q_tbl_addr; 192 u32 ib_offset = pm8001_ha->ib_offset; 193 u32 ob_offset = pm8001_ha->ob_offset; 194 u32 ci_offset = pm8001_ha->ci_offset; 195 u32 pi_offset = pm8001_ha->pi_offset; 196 197 pm8001_ha->main_cfg_tbl.pm8001_tbl.inbound_q_nppd_hppd = 0; 198 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_hw_event_pid0_3 = 0; 199 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_hw_event_pid4_7 = 0; 200 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_ncq_event_pid0_3 = 0; 201 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_ncq_event_pid4_7 = 0; 202 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_tgt_ITNexus_event_pid0_3 = 203 0; 204 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_tgt_ITNexus_event_pid4_7 = 205 0; 206 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_tgt_ssp_event_pid0_3 = 0; 207 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_tgt_ssp_event_pid4_7 = 0; 208 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_tgt_smp_event_pid0_3 = 0; 209 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_tgt_smp_event_pid4_7 = 0; 210 211 pm8001_ha->main_cfg_tbl.pm8001_tbl.upper_event_log_addr = 212 pm8001_ha->memoryMap.region[AAP1].phys_addr_hi; 213 pm8001_ha->main_cfg_tbl.pm8001_tbl.lower_event_log_addr = 214 pm8001_ha->memoryMap.region[AAP1].phys_addr_lo; 215 pm8001_ha->main_cfg_tbl.pm8001_tbl.event_log_size = 216 PM8001_EVENT_LOG_SIZE; 217 pm8001_ha->main_cfg_tbl.pm8001_tbl.event_log_option = 0x01; 218 pm8001_ha->main_cfg_tbl.pm8001_tbl.upper_iop_event_log_addr = 219 pm8001_ha->memoryMap.region[IOP].phys_addr_hi; 220 pm8001_ha->main_cfg_tbl.pm8001_tbl.lower_iop_event_log_addr = 221 pm8001_ha->memoryMap.region[IOP].phys_addr_lo; 222 pm8001_ha->main_cfg_tbl.pm8001_tbl.iop_event_log_size = 223 PM8001_EVENT_LOG_SIZE; 224 pm8001_ha->main_cfg_tbl.pm8001_tbl.iop_event_log_option = 0x01; 225 pm8001_ha->main_cfg_tbl.pm8001_tbl.fatal_err_interrupt = 0x01; 226 for (i = 0; i < PM8001_MAX_INB_NUM; i++) { 227 pm8001_ha->inbnd_q_tbl[i].element_pri_size_cnt = 228 PM8001_MPI_QUEUE | (pm8001_ha->iomb_size << 16) | (0x00<<30); 229 pm8001_ha->inbnd_q_tbl[i].upper_base_addr = 230 pm8001_ha->memoryMap.region[ib_offset + i].phys_addr_hi; 231 pm8001_ha->inbnd_q_tbl[i].lower_base_addr = 232 pm8001_ha->memoryMap.region[ib_offset + i].phys_addr_lo; 233 pm8001_ha->inbnd_q_tbl[i].base_virt = 234 (u8 *)pm8001_ha->memoryMap.region[ib_offset + i].virt_ptr; 235 pm8001_ha->inbnd_q_tbl[i].total_length = 236 pm8001_ha->memoryMap.region[ib_offset + i].total_len; 237 pm8001_ha->inbnd_q_tbl[i].ci_upper_base_addr = 238 pm8001_ha->memoryMap.region[ci_offset + i].phys_addr_hi; 239 pm8001_ha->inbnd_q_tbl[i].ci_lower_base_addr = 240 pm8001_ha->memoryMap.region[ci_offset + i].phys_addr_lo; 241 pm8001_ha->inbnd_q_tbl[i].ci_virt = 242 pm8001_ha->memoryMap.region[ci_offset + i].virt_ptr; 243 offsetib = i * 0x20; 244 pm8001_ha->inbnd_q_tbl[i].pi_pci_bar = 245 get_pci_bar_index(pm8001_mr32(addressib, 246 (offsetib + 0x14))); 247 pm8001_ha->inbnd_q_tbl[i].pi_offset = 248 pm8001_mr32(addressib, (offsetib + 0x18)); 249 pm8001_ha->inbnd_q_tbl[i].producer_idx = 0; 250 pm8001_ha->inbnd_q_tbl[i].consumer_index = 0; 251 } 252 for (i = 0; i < PM8001_MAX_OUTB_NUM; i++) { 253 pm8001_ha->outbnd_q_tbl[i].element_size_cnt = 254 PM8001_MPI_QUEUE | (pm8001_ha->iomb_size << 16) | (0x01<<30); 255 pm8001_ha->outbnd_q_tbl[i].upper_base_addr = 256 pm8001_ha->memoryMap.region[ob_offset + i].phys_addr_hi; 257 pm8001_ha->outbnd_q_tbl[i].lower_base_addr = 258 pm8001_ha->memoryMap.region[ob_offset + i].phys_addr_lo; 259 pm8001_ha->outbnd_q_tbl[i].base_virt = 260 (u8 *)pm8001_ha->memoryMap.region[ob_offset + i].virt_ptr; 261 pm8001_ha->outbnd_q_tbl[i].total_length = 262 pm8001_ha->memoryMap.region[ob_offset + i].total_len; 263 pm8001_ha->outbnd_q_tbl[i].pi_upper_base_addr = 264 pm8001_ha->memoryMap.region[pi_offset + i].phys_addr_hi; 265 pm8001_ha->outbnd_q_tbl[i].pi_lower_base_addr = 266 pm8001_ha->memoryMap.region[pi_offset + i].phys_addr_lo; 267 pm8001_ha->outbnd_q_tbl[i].interrup_vec_cnt_delay = 268 0 | (10 << 16) | (i << 24); 269 pm8001_ha->outbnd_q_tbl[i].pi_virt = 270 pm8001_ha->memoryMap.region[pi_offset + i].virt_ptr; 271 offsetob = i * 0x24; 272 pm8001_ha->outbnd_q_tbl[i].ci_pci_bar = 273 get_pci_bar_index(pm8001_mr32(addressob, 274 offsetob + 0x14)); 275 pm8001_ha->outbnd_q_tbl[i].ci_offset = 276 pm8001_mr32(addressob, (offsetob + 0x18)); 277 pm8001_ha->outbnd_q_tbl[i].consumer_idx = 0; 278 pm8001_ha->outbnd_q_tbl[i].producer_index = 0; 279 } 280 } 281 282 /** 283 * update_main_config_table - update the main default table to the HBA. 284 * @pm8001_ha: our hba card information 285 */ 286 static void update_main_config_table(struct pm8001_hba_info *pm8001_ha) 287 { 288 void __iomem *address = pm8001_ha->main_cfg_tbl_addr; 289 pm8001_mw32(address, 0x24, 290 pm8001_ha->main_cfg_tbl.pm8001_tbl.inbound_q_nppd_hppd); 291 pm8001_mw32(address, 0x28, 292 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_hw_event_pid0_3); 293 pm8001_mw32(address, 0x2C, 294 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_hw_event_pid4_7); 295 pm8001_mw32(address, 0x30, 296 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_ncq_event_pid0_3); 297 pm8001_mw32(address, 0x34, 298 pm8001_ha->main_cfg_tbl.pm8001_tbl.outbound_ncq_event_pid4_7); 299 pm8001_mw32(address, 0x38, 300 pm8001_ha->main_cfg_tbl.pm8001_tbl. 301 outbound_tgt_ITNexus_event_pid0_3); 302 pm8001_mw32(address, 0x3C, 303 pm8001_ha->main_cfg_tbl.pm8001_tbl. 304 outbound_tgt_ITNexus_event_pid4_7); 305 pm8001_mw32(address, 0x40, 306 pm8001_ha->main_cfg_tbl.pm8001_tbl. 307 outbound_tgt_ssp_event_pid0_3); 308 pm8001_mw32(address, 0x44, 309 pm8001_ha->main_cfg_tbl.pm8001_tbl. 310 outbound_tgt_ssp_event_pid4_7); 311 pm8001_mw32(address, 0x48, 312 pm8001_ha->main_cfg_tbl.pm8001_tbl. 313 outbound_tgt_smp_event_pid0_3); 314 pm8001_mw32(address, 0x4C, 315 pm8001_ha->main_cfg_tbl.pm8001_tbl. 316 outbound_tgt_smp_event_pid4_7); 317 pm8001_mw32(address, 0x50, 318 pm8001_ha->main_cfg_tbl.pm8001_tbl.upper_event_log_addr); 319 pm8001_mw32(address, 0x54, 320 pm8001_ha->main_cfg_tbl.pm8001_tbl.lower_event_log_addr); 321 pm8001_mw32(address, 0x58, 322 pm8001_ha->main_cfg_tbl.pm8001_tbl.event_log_size); 323 pm8001_mw32(address, 0x5C, 324 pm8001_ha->main_cfg_tbl.pm8001_tbl.event_log_option); 325 pm8001_mw32(address, 0x60, 326 pm8001_ha->main_cfg_tbl.pm8001_tbl.upper_iop_event_log_addr); 327 pm8001_mw32(address, 0x64, 328 pm8001_ha->main_cfg_tbl.pm8001_tbl.lower_iop_event_log_addr); 329 pm8001_mw32(address, 0x68, 330 pm8001_ha->main_cfg_tbl.pm8001_tbl.iop_event_log_size); 331 pm8001_mw32(address, 0x6C, 332 pm8001_ha->main_cfg_tbl.pm8001_tbl.iop_event_log_option); 333 pm8001_mw32(address, 0x70, 334 pm8001_ha->main_cfg_tbl.pm8001_tbl.fatal_err_interrupt); 335 } 336 337 /** 338 * update_inbnd_queue_table - update the inbound queue table to the HBA. 339 * @pm8001_ha: our hba card information 340 * @number: entry in the queue 341 */ 342 static void update_inbnd_queue_table(struct pm8001_hba_info *pm8001_ha, 343 int number) 344 { 345 void __iomem *address = pm8001_ha->inbnd_q_tbl_addr; 346 u16 offset = number * 0x20; 347 pm8001_mw32(address, offset + 0x00, 348 pm8001_ha->inbnd_q_tbl[number].element_pri_size_cnt); 349 pm8001_mw32(address, offset + 0x04, 350 pm8001_ha->inbnd_q_tbl[number].upper_base_addr); 351 pm8001_mw32(address, offset + 0x08, 352 pm8001_ha->inbnd_q_tbl[number].lower_base_addr); 353 pm8001_mw32(address, offset + 0x0C, 354 pm8001_ha->inbnd_q_tbl[number].ci_upper_base_addr); 355 pm8001_mw32(address, offset + 0x10, 356 pm8001_ha->inbnd_q_tbl[number].ci_lower_base_addr); 357 } 358 359 /** 360 * update_outbnd_queue_table - update the outbound queue table to the HBA. 361 * @pm8001_ha: our hba card information 362 * @number: entry in the queue 363 */ 364 static void update_outbnd_queue_table(struct pm8001_hba_info *pm8001_ha, 365 int number) 366 { 367 void __iomem *address = pm8001_ha->outbnd_q_tbl_addr; 368 u16 offset = number * 0x24; 369 pm8001_mw32(address, offset + 0x00, 370 pm8001_ha->outbnd_q_tbl[number].element_size_cnt); 371 pm8001_mw32(address, offset + 0x04, 372 pm8001_ha->outbnd_q_tbl[number].upper_base_addr); 373 pm8001_mw32(address, offset + 0x08, 374 pm8001_ha->outbnd_q_tbl[number].lower_base_addr); 375 pm8001_mw32(address, offset + 0x0C, 376 pm8001_ha->outbnd_q_tbl[number].pi_upper_base_addr); 377 pm8001_mw32(address, offset + 0x10, 378 pm8001_ha->outbnd_q_tbl[number].pi_lower_base_addr); 379 pm8001_mw32(address, offset + 0x1C, 380 pm8001_ha->outbnd_q_tbl[number].interrup_vec_cnt_delay); 381 } 382 383 /** 384 * pm8001_bar4_shift - function is called to shift BAR base address 385 * @pm8001_ha : our hba card infomation 386 * @shiftValue : shifting value in memory bar. 387 */ 388 int pm8001_bar4_shift(struct pm8001_hba_info *pm8001_ha, u32 shiftValue) 389 { 390 u32 regVal; 391 unsigned long start; 392 393 /* program the inbound AXI translation Lower Address */ 394 pm8001_cw32(pm8001_ha, 1, SPC_IBW_AXI_TRANSLATION_LOW, shiftValue); 395 396 /* confirm the setting is written */ 397 start = jiffies + HZ; /* 1 sec */ 398 do { 399 regVal = pm8001_cr32(pm8001_ha, 1, SPC_IBW_AXI_TRANSLATION_LOW); 400 } while ((regVal != shiftValue) && time_before(jiffies, start)); 401 402 if (regVal != shiftValue) { 403 pm8001_dbg(pm8001_ha, INIT, 404 "TIMEOUT:SPC_IBW_AXI_TRANSLATION_LOW = 0x%x\n", 405 regVal); 406 return -1; 407 } 408 return 0; 409 } 410 411 /** 412 * mpi_set_phys_g3_with_ssc 413 * @pm8001_ha: our hba card information 414 * @SSCbit: set SSCbit to 0 to disable all phys ssc; 1 to enable all phys ssc. 415 */ 416 static void mpi_set_phys_g3_with_ssc(struct pm8001_hba_info *pm8001_ha, 417 u32 SSCbit) 418 { 419 u32 offset, i; 420 unsigned long flags; 421 422 #define SAS2_SETTINGS_LOCAL_PHY_0_3_SHIFT_ADDR 0x00030000 423 #define SAS2_SETTINGS_LOCAL_PHY_4_7_SHIFT_ADDR 0x00040000 424 #define SAS2_SETTINGS_LOCAL_PHY_0_3_OFFSET 0x1074 425 #define SAS2_SETTINGS_LOCAL_PHY_4_7_OFFSET 0x1074 426 #define PHY_G3_WITHOUT_SSC_BIT_SHIFT 12 427 #define PHY_G3_WITH_SSC_BIT_SHIFT 13 428 #define SNW3_PHY_CAPABILITIES_PARITY 31 429 430 /* 431 * Using shifted destination address 0x3_0000:0x1074 + 0x4000*N (N=0:3) 432 * Using shifted destination address 0x4_0000:0x1074 + 0x4000*(N-4) (N=4:7) 433 */ 434 spin_lock_irqsave(&pm8001_ha->lock, flags); 435 if (-1 == pm8001_bar4_shift(pm8001_ha, 436 SAS2_SETTINGS_LOCAL_PHY_0_3_SHIFT_ADDR)) { 437 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 438 return; 439 } 440 441 for (i = 0; i < 4; i++) { 442 offset = SAS2_SETTINGS_LOCAL_PHY_0_3_OFFSET + 0x4000 * i; 443 pm8001_cw32(pm8001_ha, 2, offset, 0x80001501); 444 } 445 /* shift membase 3 for SAS2_SETTINGS_LOCAL_PHY 4 - 7 */ 446 if (-1 == pm8001_bar4_shift(pm8001_ha, 447 SAS2_SETTINGS_LOCAL_PHY_4_7_SHIFT_ADDR)) { 448 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 449 return; 450 } 451 for (i = 4; i < 8; i++) { 452 offset = SAS2_SETTINGS_LOCAL_PHY_4_7_OFFSET + 0x4000 * (i-4); 453 pm8001_cw32(pm8001_ha, 2, offset, 0x80001501); 454 } 455 /************************************************************* 456 Change the SSC upspreading value to 0x0 so that upspreading is disabled. 457 Device MABC SMOD0 Controls 458 Address: (via MEMBASE-III): 459 Using shifted destination address 0x0_0000: with Offset 0xD8 460 461 31:28 R/W Reserved Do not change 462 27:24 R/W SAS_SMOD_SPRDUP 0000 463 23:20 R/W SAS_SMOD_SPRDDN 0000 464 19:0 R/W Reserved Do not change 465 Upon power-up this register will read as 0x8990c016, 466 and I would like you to change the SAS_SMOD_SPRDUP bits to 0b0000 467 so that the written value will be 0x8090c016. 468 This will ensure only down-spreading SSC is enabled on the SPC. 469 *************************************************************/ 470 pm8001_cr32(pm8001_ha, 2, 0xd8); 471 pm8001_cw32(pm8001_ha, 2, 0xd8, 0x8000C016); 472 473 /*set the shifted destination address to 0x0 to avoid error operation */ 474 pm8001_bar4_shift(pm8001_ha, 0x0); 475 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 476 return; 477 } 478 479 /** 480 * mpi_set_open_retry_interval_reg 481 * @pm8001_ha: our hba card information 482 * @interval: interval time for each OPEN_REJECT (RETRY). The units are in 1us. 483 */ 484 static void mpi_set_open_retry_interval_reg(struct pm8001_hba_info *pm8001_ha, 485 u32 interval) 486 { 487 u32 offset; 488 u32 value; 489 u32 i; 490 unsigned long flags; 491 492 #define OPEN_RETRY_INTERVAL_PHY_0_3_SHIFT_ADDR 0x00030000 493 #define OPEN_RETRY_INTERVAL_PHY_4_7_SHIFT_ADDR 0x00040000 494 #define OPEN_RETRY_INTERVAL_PHY_0_3_OFFSET 0x30B4 495 #define OPEN_RETRY_INTERVAL_PHY_4_7_OFFSET 0x30B4 496 #define OPEN_RETRY_INTERVAL_REG_MASK 0x0000FFFF 497 498 value = interval & OPEN_RETRY_INTERVAL_REG_MASK; 499 spin_lock_irqsave(&pm8001_ha->lock, flags); 500 /* shift bar and set the OPEN_REJECT(RETRY) interval time of PHY 0 -3.*/ 501 if (-1 == pm8001_bar4_shift(pm8001_ha, 502 OPEN_RETRY_INTERVAL_PHY_0_3_SHIFT_ADDR)) { 503 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 504 return; 505 } 506 for (i = 0; i < 4; i++) { 507 offset = OPEN_RETRY_INTERVAL_PHY_0_3_OFFSET + 0x4000 * i; 508 pm8001_cw32(pm8001_ha, 2, offset, value); 509 } 510 511 if (-1 == pm8001_bar4_shift(pm8001_ha, 512 OPEN_RETRY_INTERVAL_PHY_4_7_SHIFT_ADDR)) { 513 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 514 return; 515 } 516 for (i = 4; i < 8; i++) { 517 offset = OPEN_RETRY_INTERVAL_PHY_4_7_OFFSET + 0x4000 * (i-4); 518 pm8001_cw32(pm8001_ha, 2, offset, value); 519 } 520 /*set the shifted destination address to 0x0 to avoid error operation */ 521 pm8001_bar4_shift(pm8001_ha, 0x0); 522 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 523 return; 524 } 525 526 /** 527 * mpi_init_check - check firmware initialization status. 528 * @pm8001_ha: our hba card information 529 */ 530 static int mpi_init_check(struct pm8001_hba_info *pm8001_ha) 531 { 532 u32 max_wait_count; 533 u32 value; 534 u32 gst_len_mpistate; 535 /* Write bit0=1 to Inbound DoorBell Register to tell the SPC FW the 536 table is updated */ 537 pm8001_cw32(pm8001_ha, 0, MSGU_IBDB_SET, SPC_MSGU_CFG_TABLE_UPDATE); 538 /* wait until Inbound DoorBell Clear Register toggled */ 539 max_wait_count = 1 * 1000 * 1000;/* 1 sec */ 540 do { 541 udelay(1); 542 value = pm8001_cr32(pm8001_ha, 0, MSGU_IBDB_SET); 543 value &= SPC_MSGU_CFG_TABLE_UPDATE; 544 } while ((value != 0) && (--max_wait_count)); 545 546 if (!max_wait_count) 547 return -1; 548 /* check the MPI-State for initialization */ 549 gst_len_mpistate = 550 pm8001_mr32(pm8001_ha->general_stat_tbl_addr, 551 GST_GSTLEN_MPIS_OFFSET); 552 if (GST_MPI_STATE_INIT != (gst_len_mpistate & GST_MPI_STATE_MASK)) 553 return -1; 554 /* check MPI Initialization error */ 555 gst_len_mpistate = gst_len_mpistate >> 16; 556 if (0x0000 != gst_len_mpistate) 557 return -1; 558 return 0; 559 } 560 561 /** 562 * check_fw_ready - The LLDD check if the FW is ready, if not, return error. 563 * @pm8001_ha: our hba card information 564 */ 565 static int check_fw_ready(struct pm8001_hba_info *pm8001_ha) 566 { 567 u32 value, value1; 568 u32 max_wait_count; 569 /* check error state */ 570 value = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_1); 571 value1 = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_2); 572 /* check AAP error */ 573 if (SCRATCH_PAD1_ERR == (value & SCRATCH_PAD_STATE_MASK)) { 574 /* error state */ 575 value = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_0); 576 return -1; 577 } 578 579 /* check IOP error */ 580 if (SCRATCH_PAD2_ERR == (value1 & SCRATCH_PAD_STATE_MASK)) { 581 /* error state */ 582 value1 = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_3); 583 return -1; 584 } 585 586 /* bit 4-31 of scratch pad1 should be zeros if it is not 587 in error state*/ 588 if (value & SCRATCH_PAD1_STATE_MASK) { 589 /* error case */ 590 pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_0); 591 return -1; 592 } 593 594 /* bit 2, 4-31 of scratch pad2 should be zeros if it is not 595 in error state */ 596 if (value1 & SCRATCH_PAD2_STATE_MASK) { 597 /* error case */ 598 return -1; 599 } 600 601 max_wait_count = 1 * 1000 * 1000;/* 1 sec timeout */ 602 603 /* wait until scratch pad 1 and 2 registers in ready state */ 604 do { 605 udelay(1); 606 value = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_1) 607 & SCRATCH_PAD1_RDY; 608 value1 = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_2) 609 & SCRATCH_PAD2_RDY; 610 if ((--max_wait_count) == 0) 611 return -1; 612 } while ((value != SCRATCH_PAD1_RDY) || (value1 != SCRATCH_PAD2_RDY)); 613 return 0; 614 } 615 616 static void init_pci_device_addresses(struct pm8001_hba_info *pm8001_ha) 617 { 618 void __iomem *base_addr; 619 u32 value; 620 u32 offset; 621 u32 pcibar; 622 u32 pcilogic; 623 624 value = pm8001_cr32(pm8001_ha, 0, 0x44); 625 offset = value & 0x03FFFFFF; 626 pm8001_dbg(pm8001_ha, INIT, "Scratchpad 0 Offset: %x\n", offset); 627 pcilogic = (value & 0xFC000000) >> 26; 628 pcibar = get_pci_bar_index(pcilogic); 629 pm8001_dbg(pm8001_ha, INIT, "Scratchpad 0 PCI BAR: %d\n", pcibar); 630 pm8001_ha->main_cfg_tbl_addr = base_addr = 631 pm8001_ha->io_mem[pcibar].memvirtaddr + offset; 632 pm8001_ha->general_stat_tbl_addr = 633 base_addr + pm8001_cr32(pm8001_ha, pcibar, offset + 0x18); 634 pm8001_ha->inbnd_q_tbl_addr = 635 base_addr + pm8001_cr32(pm8001_ha, pcibar, offset + 0x1C); 636 pm8001_ha->outbnd_q_tbl_addr = 637 base_addr + pm8001_cr32(pm8001_ha, pcibar, offset + 0x20); 638 } 639 640 /** 641 * pm8001_chip_init - the main init function that initialize whole PM8001 chip. 642 * @pm8001_ha: our hba card information 643 */ 644 static int pm8001_chip_init(struct pm8001_hba_info *pm8001_ha) 645 { 646 u8 i = 0; 647 u16 deviceid; 648 pci_read_config_word(pm8001_ha->pdev, PCI_DEVICE_ID, &deviceid); 649 /* 8081 controllers need BAR shift to access MPI space 650 * as this is shared with BIOS data */ 651 if (deviceid == 0x8081 || deviceid == 0x0042) { 652 if (-1 == pm8001_bar4_shift(pm8001_ha, GSM_SM_BASE)) { 653 pm8001_dbg(pm8001_ha, FAIL, 654 "Shift Bar4 to 0x%x failed\n", 655 GSM_SM_BASE); 656 return -1; 657 } 658 } 659 /* check the firmware status */ 660 if (-1 == check_fw_ready(pm8001_ha)) { 661 pm8001_dbg(pm8001_ha, FAIL, "Firmware is not ready!\n"); 662 return -EBUSY; 663 } 664 665 /* Initialize pci space address eg: mpi offset */ 666 init_pci_device_addresses(pm8001_ha); 667 init_default_table_values(pm8001_ha); 668 read_main_config_table(pm8001_ha); 669 read_general_status_table(pm8001_ha); 670 read_inbnd_queue_table(pm8001_ha); 671 read_outbnd_queue_table(pm8001_ha); 672 /* update main config table ,inbound table and outbound table */ 673 update_main_config_table(pm8001_ha); 674 for (i = 0; i < PM8001_MAX_INB_NUM; i++) 675 update_inbnd_queue_table(pm8001_ha, i); 676 for (i = 0; i < PM8001_MAX_OUTB_NUM; i++) 677 update_outbnd_queue_table(pm8001_ha, i); 678 /* 8081 controller donot require these operations */ 679 if (deviceid != 0x8081 && deviceid != 0x0042) { 680 mpi_set_phys_g3_with_ssc(pm8001_ha, 0); 681 /* 7->130ms, 34->500ms, 119->1.5s */ 682 mpi_set_open_retry_interval_reg(pm8001_ha, 119); 683 } 684 /* notify firmware update finished and check initialization status */ 685 if (0 == mpi_init_check(pm8001_ha)) { 686 pm8001_dbg(pm8001_ha, INIT, "MPI initialize successful!\n"); 687 } else 688 return -EBUSY; 689 /*This register is a 16-bit timer with a resolution of 1us. This is the 690 timer used for interrupt delay/coalescing in the PCIe Application Layer. 691 Zero is not a valid value. A value of 1 in the register will cause the 692 interrupts to be normal. A value greater than 1 will cause coalescing 693 delays.*/ 694 pm8001_cw32(pm8001_ha, 1, 0x0033c0, 0x1); 695 pm8001_cw32(pm8001_ha, 1, 0x0033c4, 0x0); 696 return 0; 697 } 698 699 static int mpi_uninit_check(struct pm8001_hba_info *pm8001_ha) 700 { 701 u32 max_wait_count; 702 u32 value; 703 u32 gst_len_mpistate; 704 u16 deviceid; 705 pci_read_config_word(pm8001_ha->pdev, PCI_DEVICE_ID, &deviceid); 706 if (deviceid == 0x8081 || deviceid == 0x0042) { 707 if (-1 == pm8001_bar4_shift(pm8001_ha, GSM_SM_BASE)) { 708 pm8001_dbg(pm8001_ha, FAIL, 709 "Shift Bar4 to 0x%x failed\n", 710 GSM_SM_BASE); 711 return -1; 712 } 713 } 714 init_pci_device_addresses(pm8001_ha); 715 /* Write bit1=1 to Inbound DoorBell Register to tell the SPC FW the 716 table is stop */ 717 pm8001_cw32(pm8001_ha, 0, MSGU_IBDB_SET, SPC_MSGU_CFG_TABLE_RESET); 718 719 /* wait until Inbound DoorBell Clear Register toggled */ 720 max_wait_count = 1 * 1000 * 1000;/* 1 sec */ 721 do { 722 udelay(1); 723 value = pm8001_cr32(pm8001_ha, 0, MSGU_IBDB_SET); 724 value &= SPC_MSGU_CFG_TABLE_RESET; 725 } while ((value != 0) && (--max_wait_count)); 726 727 if (!max_wait_count) { 728 pm8001_dbg(pm8001_ha, FAIL, "TIMEOUT:IBDB value/=0x%x\n", 729 value); 730 return -1; 731 } 732 733 /* check the MPI-State for termination in progress */ 734 /* wait until Inbound DoorBell Clear Register toggled */ 735 max_wait_count = 1 * 1000 * 1000; /* 1 sec */ 736 do { 737 udelay(1); 738 gst_len_mpistate = 739 pm8001_mr32(pm8001_ha->general_stat_tbl_addr, 740 GST_GSTLEN_MPIS_OFFSET); 741 if (GST_MPI_STATE_UNINIT == 742 (gst_len_mpistate & GST_MPI_STATE_MASK)) 743 break; 744 } while (--max_wait_count); 745 if (!max_wait_count) { 746 pm8001_dbg(pm8001_ha, FAIL, " TIME OUT MPI State = 0x%x\n", 747 gst_len_mpistate & GST_MPI_STATE_MASK); 748 return -1; 749 } 750 return 0; 751 } 752 753 /** 754 * soft_reset_ready_check - Function to check FW is ready for soft reset. 755 * @pm8001_ha: our hba card information 756 */ 757 static u32 soft_reset_ready_check(struct pm8001_hba_info *pm8001_ha) 758 { 759 u32 regVal, regVal1, regVal2; 760 if (mpi_uninit_check(pm8001_ha) != 0) { 761 pm8001_dbg(pm8001_ha, FAIL, "MPI state is not ready\n"); 762 return -1; 763 } 764 /* read the scratch pad 2 register bit 2 */ 765 regVal = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_2) 766 & SCRATCH_PAD2_FWRDY_RST; 767 if (regVal == SCRATCH_PAD2_FWRDY_RST) { 768 pm8001_dbg(pm8001_ha, INIT, "Firmware is ready for reset.\n"); 769 } else { 770 unsigned long flags; 771 /* Trigger NMI twice via RB6 */ 772 spin_lock_irqsave(&pm8001_ha->lock, flags); 773 if (-1 == pm8001_bar4_shift(pm8001_ha, RB6_ACCESS_REG)) { 774 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 775 pm8001_dbg(pm8001_ha, FAIL, 776 "Shift Bar4 to 0x%x failed\n", 777 RB6_ACCESS_REG); 778 return -1; 779 } 780 pm8001_cw32(pm8001_ha, 2, SPC_RB6_OFFSET, 781 RB6_MAGIC_NUMBER_RST); 782 pm8001_cw32(pm8001_ha, 2, SPC_RB6_OFFSET, RB6_MAGIC_NUMBER_RST); 783 /* wait for 100 ms */ 784 mdelay(100); 785 regVal = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_2) & 786 SCRATCH_PAD2_FWRDY_RST; 787 if (regVal != SCRATCH_PAD2_FWRDY_RST) { 788 regVal1 = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_1); 789 regVal2 = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_2); 790 pm8001_dbg(pm8001_ha, FAIL, "TIMEOUT:MSGU_SCRATCH_PAD1=0x%x, MSGU_SCRATCH_PAD2=0x%x\n", 791 regVal1, regVal2); 792 pm8001_dbg(pm8001_ha, FAIL, 793 "SCRATCH_PAD0 value = 0x%x\n", 794 pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_0)); 795 pm8001_dbg(pm8001_ha, FAIL, 796 "SCRATCH_PAD3 value = 0x%x\n", 797 pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_3)); 798 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 799 return -1; 800 } 801 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 802 } 803 return 0; 804 } 805 806 /** 807 * pm8001_chip_soft_rst - soft reset the PM8001 chip, so that the clear all 808 * the FW register status to the originated status. 809 * @pm8001_ha: our hba card information 810 */ 811 static int 812 pm8001_chip_soft_rst(struct pm8001_hba_info *pm8001_ha) 813 { 814 u32 regVal, toggleVal; 815 u32 max_wait_count; 816 u32 regVal1, regVal2, regVal3; 817 u32 signature = 0x252acbcd; /* for host scratch pad0 */ 818 unsigned long flags; 819 820 /* step1: Check FW is ready for soft reset */ 821 if (soft_reset_ready_check(pm8001_ha) != 0) { 822 pm8001_dbg(pm8001_ha, FAIL, "FW is not ready\n"); 823 return -1; 824 } 825 826 /* step 2: clear NMI status register on AAP1 and IOP, write the same 827 value to clear */ 828 /* map 0x60000 to BAR4(0x20), BAR2(win) */ 829 spin_lock_irqsave(&pm8001_ha->lock, flags); 830 if (-1 == pm8001_bar4_shift(pm8001_ha, MBIC_AAP1_ADDR_BASE)) { 831 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 832 pm8001_dbg(pm8001_ha, FAIL, "Shift Bar4 to 0x%x failed\n", 833 MBIC_AAP1_ADDR_BASE); 834 return -1; 835 } 836 regVal = pm8001_cr32(pm8001_ha, 2, MBIC_NMI_ENABLE_VPE0_IOP); 837 pm8001_dbg(pm8001_ha, INIT, "MBIC - NMI Enable VPE0 (IOP)= 0x%x\n", 838 regVal); 839 pm8001_cw32(pm8001_ha, 2, MBIC_NMI_ENABLE_VPE0_IOP, 0x0); 840 /* map 0x70000 to BAR4(0x20), BAR2(win) */ 841 if (-1 == pm8001_bar4_shift(pm8001_ha, MBIC_IOP_ADDR_BASE)) { 842 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 843 pm8001_dbg(pm8001_ha, FAIL, "Shift Bar4 to 0x%x failed\n", 844 MBIC_IOP_ADDR_BASE); 845 return -1; 846 } 847 regVal = pm8001_cr32(pm8001_ha, 2, MBIC_NMI_ENABLE_VPE0_AAP1); 848 pm8001_dbg(pm8001_ha, INIT, "MBIC - NMI Enable VPE0 (AAP1)= 0x%x\n", 849 regVal); 850 pm8001_cw32(pm8001_ha, 2, MBIC_NMI_ENABLE_VPE0_AAP1, 0x0); 851 852 regVal = pm8001_cr32(pm8001_ha, 1, PCIE_EVENT_INTERRUPT_ENABLE); 853 pm8001_dbg(pm8001_ha, INIT, "PCIE -Event Interrupt Enable = 0x%x\n", 854 regVal); 855 pm8001_cw32(pm8001_ha, 1, PCIE_EVENT_INTERRUPT_ENABLE, 0x0); 856 857 regVal = pm8001_cr32(pm8001_ha, 1, PCIE_EVENT_INTERRUPT); 858 pm8001_dbg(pm8001_ha, INIT, "PCIE - Event Interrupt = 0x%x\n", 859 regVal); 860 pm8001_cw32(pm8001_ha, 1, PCIE_EVENT_INTERRUPT, regVal); 861 862 regVal = pm8001_cr32(pm8001_ha, 1, PCIE_ERROR_INTERRUPT_ENABLE); 863 pm8001_dbg(pm8001_ha, INIT, "PCIE -Error Interrupt Enable = 0x%x\n", 864 regVal); 865 pm8001_cw32(pm8001_ha, 1, PCIE_ERROR_INTERRUPT_ENABLE, 0x0); 866 867 regVal = pm8001_cr32(pm8001_ha, 1, PCIE_ERROR_INTERRUPT); 868 pm8001_dbg(pm8001_ha, INIT, "PCIE - Error Interrupt = 0x%x\n", regVal); 869 pm8001_cw32(pm8001_ha, 1, PCIE_ERROR_INTERRUPT, regVal); 870 871 /* read the scratch pad 1 register bit 2 */ 872 regVal = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_1) 873 & SCRATCH_PAD1_RST; 874 toggleVal = regVal ^ SCRATCH_PAD1_RST; 875 876 /* set signature in host scratch pad0 register to tell SPC that the 877 host performs the soft reset */ 878 pm8001_cw32(pm8001_ha, 0, MSGU_HOST_SCRATCH_PAD_0, signature); 879 880 /* read required registers for confirmming */ 881 /* map 0x0700000 to BAR4(0x20), BAR2(win) */ 882 if (-1 == pm8001_bar4_shift(pm8001_ha, GSM_ADDR_BASE)) { 883 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 884 pm8001_dbg(pm8001_ha, FAIL, "Shift Bar4 to 0x%x failed\n", 885 GSM_ADDR_BASE); 886 return -1; 887 } 888 pm8001_dbg(pm8001_ha, INIT, 889 "GSM 0x0(0x00007b88)-GSM Configuration and Reset = 0x%x\n", 890 pm8001_cr32(pm8001_ha, 2, GSM_CONFIG_RESET)); 891 892 /* step 3: host read GSM Configuration and Reset register */ 893 regVal = pm8001_cr32(pm8001_ha, 2, GSM_CONFIG_RESET); 894 /* Put those bits to low */ 895 /* GSM XCBI offset = 0x70 0000 896 0x00 Bit 13 COM_SLV_SW_RSTB 1 897 0x00 Bit 12 QSSP_SW_RSTB 1 898 0x00 Bit 11 RAAE_SW_RSTB 1 899 0x00 Bit 9 RB_1_SW_RSTB 1 900 0x00 Bit 8 SM_SW_RSTB 1 901 */ 902 regVal &= ~(0x00003b00); 903 /* host write GSM Configuration and Reset register */ 904 pm8001_cw32(pm8001_ha, 2, GSM_CONFIG_RESET, regVal); 905 pm8001_dbg(pm8001_ha, INIT, 906 "GSM 0x0 (0x00007b88 ==> 0x00004088) - GSM Configuration and Reset is set to = 0x%x\n", 907 pm8001_cr32(pm8001_ha, 2, GSM_CONFIG_RESET)); 908 909 /* step 4: */ 910 /* disable GSM - Read Address Parity Check */ 911 regVal1 = pm8001_cr32(pm8001_ha, 2, GSM_READ_ADDR_PARITY_CHECK); 912 pm8001_dbg(pm8001_ha, INIT, 913 "GSM 0x700038 - Read Address Parity Check Enable = 0x%x\n", 914 regVal1); 915 pm8001_cw32(pm8001_ha, 2, GSM_READ_ADDR_PARITY_CHECK, 0x0); 916 pm8001_dbg(pm8001_ha, INIT, 917 "GSM 0x700038 - Read Address Parity Check Enable is set to = 0x%x\n", 918 pm8001_cr32(pm8001_ha, 2, GSM_READ_ADDR_PARITY_CHECK)); 919 920 /* disable GSM - Write Address Parity Check */ 921 regVal2 = pm8001_cr32(pm8001_ha, 2, GSM_WRITE_ADDR_PARITY_CHECK); 922 pm8001_dbg(pm8001_ha, INIT, 923 "GSM 0x700040 - Write Address Parity Check Enable = 0x%x\n", 924 regVal2); 925 pm8001_cw32(pm8001_ha, 2, GSM_WRITE_ADDR_PARITY_CHECK, 0x0); 926 pm8001_dbg(pm8001_ha, INIT, 927 "GSM 0x700040 - Write Address Parity Check Enable is set to = 0x%x\n", 928 pm8001_cr32(pm8001_ha, 2, GSM_WRITE_ADDR_PARITY_CHECK)); 929 930 /* disable GSM - Write Data Parity Check */ 931 regVal3 = pm8001_cr32(pm8001_ha, 2, GSM_WRITE_DATA_PARITY_CHECK); 932 pm8001_dbg(pm8001_ha, INIT, "GSM 0x300048 - Write Data Parity Check Enable = 0x%x\n", 933 regVal3); 934 pm8001_cw32(pm8001_ha, 2, GSM_WRITE_DATA_PARITY_CHECK, 0x0); 935 pm8001_dbg(pm8001_ha, INIT, 936 "GSM 0x300048 - Write Data Parity Check Enable is set to = 0x%x\n", 937 pm8001_cr32(pm8001_ha, 2, GSM_WRITE_DATA_PARITY_CHECK)); 938 939 /* step 5: delay 10 usec */ 940 udelay(10); 941 /* step 5-b: set GPIO-0 output control to tristate anyway */ 942 if (-1 == pm8001_bar4_shift(pm8001_ha, GPIO_ADDR_BASE)) { 943 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 944 pm8001_dbg(pm8001_ha, INIT, "Shift Bar4 to 0x%x failed\n", 945 GPIO_ADDR_BASE); 946 return -1; 947 } 948 regVal = pm8001_cr32(pm8001_ha, 2, GPIO_GPIO_0_0UTPUT_CTL_OFFSET); 949 pm8001_dbg(pm8001_ha, INIT, "GPIO Output Control Register: = 0x%x\n", 950 regVal); 951 /* set GPIO-0 output control to tri-state */ 952 regVal &= 0xFFFFFFFC; 953 pm8001_cw32(pm8001_ha, 2, GPIO_GPIO_0_0UTPUT_CTL_OFFSET, regVal); 954 955 /* Step 6: Reset the IOP and AAP1 */ 956 /* map 0x00000 to BAR4(0x20), BAR2(win) */ 957 if (-1 == pm8001_bar4_shift(pm8001_ha, SPC_TOP_LEVEL_ADDR_BASE)) { 958 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 959 pm8001_dbg(pm8001_ha, FAIL, "SPC Shift Bar4 to 0x%x failed\n", 960 SPC_TOP_LEVEL_ADDR_BASE); 961 return -1; 962 } 963 regVal = pm8001_cr32(pm8001_ha, 2, SPC_REG_RESET); 964 pm8001_dbg(pm8001_ha, INIT, "Top Register before resetting IOP/AAP1:= 0x%x\n", 965 regVal); 966 regVal &= ~(SPC_REG_RESET_PCS_IOP_SS | SPC_REG_RESET_PCS_AAP1_SS); 967 pm8001_cw32(pm8001_ha, 2, SPC_REG_RESET, regVal); 968 969 /* step 7: Reset the BDMA/OSSP */ 970 regVal = pm8001_cr32(pm8001_ha, 2, SPC_REG_RESET); 971 pm8001_dbg(pm8001_ha, INIT, "Top Register before resetting BDMA/OSSP: = 0x%x\n", 972 regVal); 973 regVal &= ~(SPC_REG_RESET_BDMA_CORE | SPC_REG_RESET_OSSP); 974 pm8001_cw32(pm8001_ha, 2, SPC_REG_RESET, regVal); 975 976 /* step 8: delay 10 usec */ 977 udelay(10); 978 979 /* step 9: bring the BDMA and OSSP out of reset */ 980 regVal = pm8001_cr32(pm8001_ha, 2, SPC_REG_RESET); 981 pm8001_dbg(pm8001_ha, INIT, 982 "Top Register before bringing up BDMA/OSSP:= 0x%x\n", 983 regVal); 984 regVal |= (SPC_REG_RESET_BDMA_CORE | SPC_REG_RESET_OSSP); 985 pm8001_cw32(pm8001_ha, 2, SPC_REG_RESET, regVal); 986 987 /* step 10: delay 10 usec */ 988 udelay(10); 989 990 /* step 11: reads and sets the GSM Configuration and Reset Register */ 991 /* map 0x0700000 to BAR4(0x20), BAR2(win) */ 992 if (-1 == pm8001_bar4_shift(pm8001_ha, GSM_ADDR_BASE)) { 993 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 994 pm8001_dbg(pm8001_ha, FAIL, "SPC Shift Bar4 to 0x%x failed\n", 995 GSM_ADDR_BASE); 996 return -1; 997 } 998 pm8001_dbg(pm8001_ha, INIT, 999 "GSM 0x0 (0x00007b88)-GSM Configuration and Reset = 0x%x\n", 1000 pm8001_cr32(pm8001_ha, 2, GSM_CONFIG_RESET)); 1001 regVal = pm8001_cr32(pm8001_ha, 2, GSM_CONFIG_RESET); 1002 /* Put those bits to high */ 1003 /* GSM XCBI offset = 0x70 0000 1004 0x00 Bit 13 COM_SLV_SW_RSTB 1 1005 0x00 Bit 12 QSSP_SW_RSTB 1 1006 0x00 Bit 11 RAAE_SW_RSTB 1 1007 0x00 Bit 9 RB_1_SW_RSTB 1 1008 0x00 Bit 8 SM_SW_RSTB 1 1009 */ 1010 regVal |= (GSM_CONFIG_RESET_VALUE); 1011 pm8001_cw32(pm8001_ha, 2, GSM_CONFIG_RESET, regVal); 1012 pm8001_dbg(pm8001_ha, INIT, "GSM (0x00004088 ==> 0x00007b88) - GSM Configuration and Reset is set to = 0x%x\n", 1013 pm8001_cr32(pm8001_ha, 2, GSM_CONFIG_RESET)); 1014 1015 /* step 12: Restore GSM - Read Address Parity Check */ 1016 regVal = pm8001_cr32(pm8001_ha, 2, GSM_READ_ADDR_PARITY_CHECK); 1017 /* just for debugging */ 1018 pm8001_dbg(pm8001_ha, INIT, 1019 "GSM 0x700038 - Read Address Parity Check Enable = 0x%x\n", 1020 regVal); 1021 pm8001_cw32(pm8001_ha, 2, GSM_READ_ADDR_PARITY_CHECK, regVal1); 1022 pm8001_dbg(pm8001_ha, INIT, "GSM 0x700038 - Read Address Parity Check Enable is set to = 0x%x\n", 1023 pm8001_cr32(pm8001_ha, 2, GSM_READ_ADDR_PARITY_CHECK)); 1024 /* Restore GSM - Write Address Parity Check */ 1025 regVal = pm8001_cr32(pm8001_ha, 2, GSM_WRITE_ADDR_PARITY_CHECK); 1026 pm8001_cw32(pm8001_ha, 2, GSM_WRITE_ADDR_PARITY_CHECK, regVal2); 1027 pm8001_dbg(pm8001_ha, INIT, 1028 "GSM 0x700040 - Write Address Parity Check Enable is set to = 0x%x\n", 1029 pm8001_cr32(pm8001_ha, 2, GSM_WRITE_ADDR_PARITY_CHECK)); 1030 /* Restore GSM - Write Data Parity Check */ 1031 regVal = pm8001_cr32(pm8001_ha, 2, GSM_WRITE_DATA_PARITY_CHECK); 1032 pm8001_cw32(pm8001_ha, 2, GSM_WRITE_DATA_PARITY_CHECK, regVal3); 1033 pm8001_dbg(pm8001_ha, INIT, 1034 "GSM 0x700048 - Write Data Parity Check Enable is set to = 0x%x\n", 1035 pm8001_cr32(pm8001_ha, 2, GSM_WRITE_DATA_PARITY_CHECK)); 1036 1037 /* step 13: bring the IOP and AAP1 out of reset */ 1038 /* map 0x00000 to BAR4(0x20), BAR2(win) */ 1039 if (-1 == pm8001_bar4_shift(pm8001_ha, SPC_TOP_LEVEL_ADDR_BASE)) { 1040 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1041 pm8001_dbg(pm8001_ha, FAIL, "Shift Bar4 to 0x%x failed\n", 1042 SPC_TOP_LEVEL_ADDR_BASE); 1043 return -1; 1044 } 1045 regVal = pm8001_cr32(pm8001_ha, 2, SPC_REG_RESET); 1046 regVal |= (SPC_REG_RESET_PCS_IOP_SS | SPC_REG_RESET_PCS_AAP1_SS); 1047 pm8001_cw32(pm8001_ha, 2, SPC_REG_RESET, regVal); 1048 1049 /* step 14: delay 10 usec - Normal Mode */ 1050 udelay(10); 1051 /* check Soft Reset Normal mode or Soft Reset HDA mode */ 1052 if (signature == SPC_SOFT_RESET_SIGNATURE) { 1053 /* step 15 (Normal Mode): wait until scratch pad1 register 1054 bit 2 toggled */ 1055 max_wait_count = 2 * 1000 * 1000;/* 2 sec */ 1056 do { 1057 udelay(1); 1058 regVal = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_1) & 1059 SCRATCH_PAD1_RST; 1060 } while ((regVal != toggleVal) && (--max_wait_count)); 1061 1062 if (!max_wait_count) { 1063 regVal = pm8001_cr32(pm8001_ha, 0, 1064 MSGU_SCRATCH_PAD_1); 1065 pm8001_dbg(pm8001_ha, FAIL, "TIMEOUT : ToggleVal 0x%x,MSGU_SCRATCH_PAD1 = 0x%x\n", 1066 toggleVal, regVal); 1067 pm8001_dbg(pm8001_ha, FAIL, 1068 "SCRATCH_PAD0 value = 0x%x\n", 1069 pm8001_cr32(pm8001_ha, 0, 1070 MSGU_SCRATCH_PAD_0)); 1071 pm8001_dbg(pm8001_ha, FAIL, 1072 "SCRATCH_PAD2 value = 0x%x\n", 1073 pm8001_cr32(pm8001_ha, 0, 1074 MSGU_SCRATCH_PAD_2)); 1075 pm8001_dbg(pm8001_ha, FAIL, 1076 "SCRATCH_PAD3 value = 0x%x\n", 1077 pm8001_cr32(pm8001_ha, 0, 1078 MSGU_SCRATCH_PAD_3)); 1079 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1080 return -1; 1081 } 1082 1083 /* step 16 (Normal) - Clear ODMR and ODCR */ 1084 pm8001_cw32(pm8001_ha, 0, MSGU_ODCR, ODCR_CLEAR_ALL); 1085 pm8001_cw32(pm8001_ha, 0, MSGU_ODMR, ODMR_CLEAR_ALL); 1086 1087 /* step 17 (Normal Mode): wait for the FW and IOP to get 1088 ready - 1 sec timeout */ 1089 /* Wait for the SPC Configuration Table to be ready */ 1090 if (check_fw_ready(pm8001_ha) == -1) { 1091 regVal = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_1); 1092 /* return error if MPI Configuration Table not ready */ 1093 pm8001_dbg(pm8001_ha, INIT, 1094 "FW not ready SCRATCH_PAD1 = 0x%x\n", 1095 regVal); 1096 regVal = pm8001_cr32(pm8001_ha, 0, MSGU_SCRATCH_PAD_2); 1097 /* return error if MPI Configuration Table not ready */ 1098 pm8001_dbg(pm8001_ha, INIT, 1099 "FW not ready SCRATCH_PAD2 = 0x%x\n", 1100 regVal); 1101 pm8001_dbg(pm8001_ha, INIT, 1102 "SCRATCH_PAD0 value = 0x%x\n", 1103 pm8001_cr32(pm8001_ha, 0, 1104 MSGU_SCRATCH_PAD_0)); 1105 pm8001_dbg(pm8001_ha, INIT, 1106 "SCRATCH_PAD3 value = 0x%x\n", 1107 pm8001_cr32(pm8001_ha, 0, 1108 MSGU_SCRATCH_PAD_3)); 1109 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1110 return -1; 1111 } 1112 } 1113 pm8001_bar4_shift(pm8001_ha, 0); 1114 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1115 1116 pm8001_dbg(pm8001_ha, INIT, "SPC soft reset Complete\n"); 1117 return 0; 1118 } 1119 1120 static void pm8001_hw_chip_rst(struct pm8001_hba_info *pm8001_ha) 1121 { 1122 u32 i; 1123 u32 regVal; 1124 pm8001_dbg(pm8001_ha, INIT, "chip reset start\n"); 1125 1126 /* do SPC chip reset. */ 1127 regVal = pm8001_cr32(pm8001_ha, 1, SPC_REG_RESET); 1128 regVal &= ~(SPC_REG_RESET_DEVICE); 1129 pm8001_cw32(pm8001_ha, 1, SPC_REG_RESET, regVal); 1130 1131 /* delay 10 usec */ 1132 udelay(10); 1133 1134 /* bring chip reset out of reset */ 1135 regVal = pm8001_cr32(pm8001_ha, 1, SPC_REG_RESET); 1136 regVal |= SPC_REG_RESET_DEVICE; 1137 pm8001_cw32(pm8001_ha, 1, SPC_REG_RESET, regVal); 1138 1139 /* delay 10 usec */ 1140 udelay(10); 1141 1142 /* wait for 20 msec until the firmware gets reloaded */ 1143 i = 20; 1144 do { 1145 mdelay(1); 1146 } while ((--i) != 0); 1147 1148 pm8001_dbg(pm8001_ha, INIT, "chip reset finished\n"); 1149 } 1150 1151 /** 1152 * pm8001_chip_iounmap - which maped when initialized. 1153 * @pm8001_ha: our hba card information 1154 */ 1155 void pm8001_chip_iounmap(struct pm8001_hba_info *pm8001_ha) 1156 { 1157 s8 bar, logical = 0; 1158 for (bar = 0; bar < PCI_STD_NUM_BARS; bar++) { 1159 /* 1160 ** logical BARs for SPC: 1161 ** bar 0 and 1 - logical BAR0 1162 ** bar 2 and 3 - logical BAR1 1163 ** bar4 - logical BAR2 1164 ** bar5 - logical BAR3 1165 ** Skip the appropriate assignments: 1166 */ 1167 if ((bar == 1) || (bar == 3)) 1168 continue; 1169 if (pm8001_ha->io_mem[logical].memvirtaddr) { 1170 iounmap(pm8001_ha->io_mem[logical].memvirtaddr); 1171 logical++; 1172 } 1173 } 1174 } 1175 1176 #ifndef PM8001_USE_MSIX 1177 /** 1178 * pm8001_chip_interrupt_enable - enable PM8001 chip interrupt 1179 * @pm8001_ha: our hba card information 1180 */ 1181 static void 1182 pm8001_chip_intx_interrupt_enable(struct pm8001_hba_info *pm8001_ha) 1183 { 1184 pm8001_cw32(pm8001_ha, 0, MSGU_ODMR, ODMR_CLEAR_ALL); 1185 pm8001_cw32(pm8001_ha, 0, MSGU_ODCR, ODCR_CLEAR_ALL); 1186 } 1187 1188 /** 1189 * pm8001_chip_intx_interrupt_disable- disable PM8001 chip interrupt 1190 * @pm8001_ha: our hba card information 1191 */ 1192 static void 1193 pm8001_chip_intx_interrupt_disable(struct pm8001_hba_info *pm8001_ha) 1194 { 1195 pm8001_cw32(pm8001_ha, 0, MSGU_ODMR, ODMR_MASK_ALL); 1196 } 1197 1198 #else 1199 1200 /** 1201 * pm8001_chip_msix_interrupt_enable - enable PM8001 chip interrupt 1202 * @pm8001_ha: our hba card information 1203 * @int_vec_idx: interrupt number to enable 1204 */ 1205 static void 1206 pm8001_chip_msix_interrupt_enable(struct pm8001_hba_info *pm8001_ha, 1207 u32 int_vec_idx) 1208 { 1209 u32 msi_index; 1210 u32 value; 1211 msi_index = int_vec_idx * MSIX_TABLE_ELEMENT_SIZE; 1212 msi_index += MSIX_TABLE_BASE; 1213 pm8001_cw32(pm8001_ha, 0, msi_index, MSIX_INTERRUPT_ENABLE); 1214 value = (1 << int_vec_idx); 1215 pm8001_cw32(pm8001_ha, 0, MSGU_ODCR, value); 1216 1217 } 1218 1219 /** 1220 * pm8001_chip_msix_interrupt_disable - disable PM8001 chip interrupt 1221 * @pm8001_ha: our hba card information 1222 * @int_vec_idx: interrupt number to disable 1223 */ 1224 static void 1225 pm8001_chip_msix_interrupt_disable(struct pm8001_hba_info *pm8001_ha, 1226 u32 int_vec_idx) 1227 { 1228 u32 msi_index; 1229 msi_index = int_vec_idx * MSIX_TABLE_ELEMENT_SIZE; 1230 msi_index += MSIX_TABLE_BASE; 1231 pm8001_cw32(pm8001_ha, 0, msi_index, MSIX_INTERRUPT_DISABLE); 1232 } 1233 #endif 1234 1235 /** 1236 * pm8001_chip_interrupt_enable - enable PM8001 chip interrupt 1237 * @pm8001_ha: our hba card information 1238 * @vec: unused 1239 */ 1240 static void 1241 pm8001_chip_interrupt_enable(struct pm8001_hba_info *pm8001_ha, u8 vec) 1242 { 1243 #ifdef PM8001_USE_MSIX 1244 pm8001_chip_msix_interrupt_enable(pm8001_ha, 0); 1245 #else 1246 pm8001_chip_intx_interrupt_enable(pm8001_ha); 1247 #endif 1248 } 1249 1250 /** 1251 * pm8001_chip_intx_interrupt_disable- disable PM8001 chip interrupt 1252 * @pm8001_ha: our hba card information 1253 * @vec: unused 1254 */ 1255 static void 1256 pm8001_chip_interrupt_disable(struct pm8001_hba_info *pm8001_ha, u8 vec) 1257 { 1258 #ifdef PM8001_USE_MSIX 1259 pm8001_chip_msix_interrupt_disable(pm8001_ha, 0); 1260 #else 1261 pm8001_chip_intx_interrupt_disable(pm8001_ha); 1262 #endif 1263 } 1264 1265 /** 1266 * pm8001_mpi_msg_free_get - get the free message buffer for transfer 1267 * inbound queue. 1268 * @circularQ: the inbound queue we want to transfer to HBA. 1269 * @messageSize: the message size of this transfer, normally it is 64 bytes 1270 * @messagePtr: the pointer to message. 1271 */ 1272 int pm8001_mpi_msg_free_get(struct inbound_queue_table *circularQ, 1273 u16 messageSize, void **messagePtr) 1274 { 1275 u32 offset, consumer_index; 1276 struct mpi_msg_hdr *msgHeader; 1277 u8 bcCount = 1; /* only support single buffer */ 1278 1279 /* Checks is the requested message size can be allocated in this queue*/ 1280 if (messageSize > IOMB_SIZE_SPCV) { 1281 *messagePtr = NULL; 1282 return -1; 1283 } 1284 1285 /* Stores the new consumer index */ 1286 consumer_index = pm8001_read_32(circularQ->ci_virt); 1287 circularQ->consumer_index = cpu_to_le32(consumer_index); 1288 if (((circularQ->producer_idx + bcCount) % PM8001_MPI_QUEUE) == 1289 le32_to_cpu(circularQ->consumer_index)) { 1290 *messagePtr = NULL; 1291 return -1; 1292 } 1293 /* get memory IOMB buffer address */ 1294 offset = circularQ->producer_idx * messageSize; 1295 /* increment to next bcCount element */ 1296 circularQ->producer_idx = (circularQ->producer_idx + bcCount) 1297 % PM8001_MPI_QUEUE; 1298 /* Adds that distance to the base of the region virtual address plus 1299 the message header size*/ 1300 msgHeader = (struct mpi_msg_hdr *)(circularQ->base_virt + offset); 1301 *messagePtr = ((void *)msgHeader) + sizeof(struct mpi_msg_hdr); 1302 return 0; 1303 } 1304 1305 /** 1306 * pm8001_mpi_build_cmd- build the message queue for transfer, update the PI to 1307 * FW to tell the fw to get this message from IOMB. 1308 * @pm8001_ha: our hba card information 1309 * @circularQ: the inbound queue we want to transfer to HBA. 1310 * @opCode: the operation code represents commands which LLDD and fw recognized. 1311 * @payload: the command payload of each operation command. 1312 * @nb: size in bytes of the command payload 1313 * @responseQueue: queue to interrupt on w/ command response (if any) 1314 */ 1315 int pm8001_mpi_build_cmd(struct pm8001_hba_info *pm8001_ha, 1316 struct inbound_queue_table *circularQ, 1317 u32 opCode, void *payload, size_t nb, 1318 u32 responseQueue) 1319 { 1320 u32 Header = 0, hpriority = 0, bc = 1, category = 0x02; 1321 void *pMessage; 1322 unsigned long flags; 1323 int q_index = circularQ - pm8001_ha->inbnd_q_tbl; 1324 int rv = -1; 1325 1326 WARN_ON(q_index >= PM8001_MAX_INB_NUM); 1327 spin_lock_irqsave(&circularQ->iq_lock, flags); 1328 rv = pm8001_mpi_msg_free_get(circularQ, pm8001_ha->iomb_size, 1329 &pMessage); 1330 if (rv < 0) { 1331 pm8001_dbg(pm8001_ha, IO, "No free mpi buffer\n"); 1332 rv = -ENOMEM; 1333 goto done; 1334 } 1335 1336 if (nb > (pm8001_ha->iomb_size - sizeof(struct mpi_msg_hdr))) 1337 nb = pm8001_ha->iomb_size - sizeof(struct mpi_msg_hdr); 1338 memcpy(pMessage, payload, nb); 1339 if (nb + sizeof(struct mpi_msg_hdr) < pm8001_ha->iomb_size) 1340 memset(pMessage + nb, 0, pm8001_ha->iomb_size - 1341 (nb + sizeof(struct mpi_msg_hdr))); 1342 1343 /*Build the header*/ 1344 Header = ((1 << 31) | (hpriority << 30) | ((bc & 0x1f) << 24) 1345 | ((responseQueue & 0x3F) << 16) 1346 | ((category & 0xF) << 12) | (opCode & 0xFFF)); 1347 1348 pm8001_write_32((pMessage - 4), 0, cpu_to_le32(Header)); 1349 /*Update the PI to the firmware*/ 1350 pm8001_cw32(pm8001_ha, circularQ->pi_pci_bar, 1351 circularQ->pi_offset, circularQ->producer_idx); 1352 pm8001_dbg(pm8001_ha, DEVIO, 1353 "INB Q %x OPCODE:%x , UPDATED PI=%d CI=%d\n", 1354 responseQueue, opCode, circularQ->producer_idx, 1355 circularQ->consumer_index); 1356 done: 1357 spin_unlock_irqrestore(&circularQ->iq_lock, flags); 1358 return rv; 1359 } 1360 1361 u32 pm8001_mpi_msg_free_set(struct pm8001_hba_info *pm8001_ha, void *pMsg, 1362 struct outbound_queue_table *circularQ, u8 bc) 1363 { 1364 u32 producer_index; 1365 struct mpi_msg_hdr *msgHeader; 1366 struct mpi_msg_hdr *pOutBoundMsgHeader; 1367 1368 msgHeader = (struct mpi_msg_hdr *)(pMsg - sizeof(struct mpi_msg_hdr)); 1369 pOutBoundMsgHeader = (struct mpi_msg_hdr *)(circularQ->base_virt + 1370 circularQ->consumer_idx * pm8001_ha->iomb_size); 1371 if (pOutBoundMsgHeader != msgHeader) { 1372 pm8001_dbg(pm8001_ha, FAIL, 1373 "consumer_idx = %d msgHeader = %p\n", 1374 circularQ->consumer_idx, msgHeader); 1375 1376 /* Update the producer index from SPC */ 1377 producer_index = pm8001_read_32(circularQ->pi_virt); 1378 circularQ->producer_index = cpu_to_le32(producer_index); 1379 pm8001_dbg(pm8001_ha, FAIL, 1380 "consumer_idx = %d producer_index = %dmsgHeader = %p\n", 1381 circularQ->consumer_idx, 1382 circularQ->producer_index, msgHeader); 1383 return 0; 1384 } 1385 /* free the circular queue buffer elements associated with the message*/ 1386 circularQ->consumer_idx = (circularQ->consumer_idx + bc) 1387 % PM8001_MPI_QUEUE; 1388 /* update the CI of outbound queue */ 1389 pm8001_cw32(pm8001_ha, circularQ->ci_pci_bar, circularQ->ci_offset, 1390 circularQ->consumer_idx); 1391 /* Update the producer index from SPC*/ 1392 producer_index = pm8001_read_32(circularQ->pi_virt); 1393 circularQ->producer_index = cpu_to_le32(producer_index); 1394 pm8001_dbg(pm8001_ha, IO, " CI=%d PI=%d\n", 1395 circularQ->consumer_idx, circularQ->producer_index); 1396 return 0; 1397 } 1398 1399 /** 1400 * pm8001_mpi_msg_consume- get the MPI message from outbound queue 1401 * message table. 1402 * @pm8001_ha: our hba card information 1403 * @circularQ: the outbound queue table. 1404 * @messagePtr1: the message contents of this outbound message. 1405 * @pBC: the message size. 1406 */ 1407 u32 pm8001_mpi_msg_consume(struct pm8001_hba_info *pm8001_ha, 1408 struct outbound_queue_table *circularQ, 1409 void **messagePtr1, u8 *pBC) 1410 { 1411 struct mpi_msg_hdr *msgHeader; 1412 __le32 msgHeader_tmp; 1413 u32 header_tmp; 1414 do { 1415 /* If there are not-yet-delivered messages ... */ 1416 if (le32_to_cpu(circularQ->producer_index) 1417 != circularQ->consumer_idx) { 1418 /*Get the pointer to the circular queue buffer element*/ 1419 msgHeader = (struct mpi_msg_hdr *) 1420 (circularQ->base_virt + 1421 circularQ->consumer_idx * pm8001_ha->iomb_size); 1422 /* read header */ 1423 header_tmp = pm8001_read_32(msgHeader); 1424 msgHeader_tmp = cpu_to_le32(header_tmp); 1425 pm8001_dbg(pm8001_ha, DEVIO, 1426 "outbound opcode msgheader:%x ci=%d pi=%d\n", 1427 msgHeader_tmp, circularQ->consumer_idx, 1428 circularQ->producer_index); 1429 if (0 != (le32_to_cpu(msgHeader_tmp) & 0x80000000)) { 1430 if (OPC_OUB_SKIP_ENTRY != 1431 (le32_to_cpu(msgHeader_tmp) & 0xfff)) { 1432 *messagePtr1 = 1433 ((u8 *)msgHeader) + 1434 sizeof(struct mpi_msg_hdr); 1435 *pBC = (u8)((le32_to_cpu(msgHeader_tmp) 1436 >> 24) & 0x1f); 1437 pm8001_dbg(pm8001_ha, IO, 1438 ": CI=%d PI=%d msgHeader=%x\n", 1439 circularQ->consumer_idx, 1440 circularQ->producer_index, 1441 msgHeader_tmp); 1442 return MPI_IO_STATUS_SUCCESS; 1443 } else { 1444 circularQ->consumer_idx = 1445 (circularQ->consumer_idx + 1446 ((le32_to_cpu(msgHeader_tmp) 1447 >> 24) & 0x1f)) 1448 % PM8001_MPI_QUEUE; 1449 msgHeader_tmp = 0; 1450 pm8001_write_32(msgHeader, 0, 0); 1451 /* update the CI of outbound queue */ 1452 pm8001_cw32(pm8001_ha, 1453 circularQ->ci_pci_bar, 1454 circularQ->ci_offset, 1455 circularQ->consumer_idx); 1456 } 1457 } else { 1458 circularQ->consumer_idx = 1459 (circularQ->consumer_idx + 1460 ((le32_to_cpu(msgHeader_tmp) >> 24) & 1461 0x1f)) % PM8001_MPI_QUEUE; 1462 msgHeader_tmp = 0; 1463 pm8001_write_32(msgHeader, 0, 0); 1464 /* update the CI of outbound queue */ 1465 pm8001_cw32(pm8001_ha, circularQ->ci_pci_bar, 1466 circularQ->ci_offset, 1467 circularQ->consumer_idx); 1468 return MPI_IO_STATUS_FAIL; 1469 } 1470 } else { 1471 u32 producer_index; 1472 void *pi_virt = circularQ->pi_virt; 1473 /* spurious interrupt during setup if 1474 * kexec-ing and driver doing a doorbell access 1475 * with the pre-kexec oq interrupt setup 1476 */ 1477 if (!pi_virt) 1478 break; 1479 /* Update the producer index from SPC */ 1480 producer_index = pm8001_read_32(pi_virt); 1481 circularQ->producer_index = cpu_to_le32(producer_index); 1482 } 1483 } while (le32_to_cpu(circularQ->producer_index) != 1484 circularQ->consumer_idx); 1485 /* while we don't have any more not-yet-delivered message */ 1486 /* report empty */ 1487 return MPI_IO_STATUS_BUSY; 1488 } 1489 1490 void pm8001_work_fn(struct work_struct *work) 1491 { 1492 struct pm8001_work *pw = container_of(work, struct pm8001_work, work); 1493 struct pm8001_device *pm8001_dev; 1494 struct domain_device *dev; 1495 1496 /* 1497 * So far, all users of this stash an associated structure here. 1498 * If we get here, and this pointer is null, then the action 1499 * was cancelled. This nullification happens when the device 1500 * goes away. 1501 */ 1502 pm8001_dev = pw->data; /* Most stash device structure */ 1503 if ((pm8001_dev == NULL) 1504 || ((pw->handler != IO_XFER_ERROR_BREAK) 1505 && (pm8001_dev->dev_type == SAS_PHY_UNUSED))) { 1506 kfree(pw); 1507 return; 1508 } 1509 1510 switch (pw->handler) { 1511 case IO_XFER_ERROR_BREAK: 1512 { /* This one stashes the sas_task instead */ 1513 struct sas_task *t = (struct sas_task *)pm8001_dev; 1514 u32 tag; 1515 struct pm8001_ccb_info *ccb; 1516 struct pm8001_hba_info *pm8001_ha = pw->pm8001_ha; 1517 unsigned long flags, flags1; 1518 struct task_status_struct *ts; 1519 int i; 1520 1521 if (pm8001_query_task(t) == TMF_RESP_FUNC_SUCC) 1522 break; /* Task still on lu */ 1523 spin_lock_irqsave(&pm8001_ha->lock, flags); 1524 1525 spin_lock_irqsave(&t->task_state_lock, flags1); 1526 if (unlikely((t->task_state_flags & SAS_TASK_STATE_DONE))) { 1527 spin_unlock_irqrestore(&t->task_state_lock, flags1); 1528 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1529 break; /* Task got completed by another */ 1530 } 1531 spin_unlock_irqrestore(&t->task_state_lock, flags1); 1532 1533 /* Search for a possible ccb that matches the task */ 1534 for (i = 0; ccb = NULL, i < PM8001_MAX_CCB; i++) { 1535 ccb = &pm8001_ha->ccb_info[i]; 1536 tag = ccb->ccb_tag; 1537 if ((tag != 0xFFFFFFFF) && (ccb->task == t)) 1538 break; 1539 } 1540 if (!ccb) { 1541 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1542 break; /* Task got freed by another */ 1543 } 1544 ts = &t->task_status; 1545 ts->resp = SAS_TASK_COMPLETE; 1546 /* Force the midlayer to retry */ 1547 ts->stat = SAS_QUEUE_FULL; 1548 pm8001_dev = ccb->device; 1549 if (pm8001_dev) 1550 atomic_dec(&pm8001_dev->running_req); 1551 spin_lock_irqsave(&t->task_state_lock, flags1); 1552 t->task_state_flags &= ~SAS_TASK_STATE_PENDING; 1553 t->task_state_flags &= ~SAS_TASK_AT_INITIATOR; 1554 t->task_state_flags |= SAS_TASK_STATE_DONE; 1555 if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) { 1556 spin_unlock_irqrestore(&t->task_state_lock, flags1); 1557 pm8001_dbg(pm8001_ha, FAIL, "task 0x%p done with event 0x%x resp 0x%x stat 0x%x but aborted by upper layer!\n", 1558 t, pw->handler, ts->resp, ts->stat); 1559 pm8001_ccb_task_free(pm8001_ha, t, ccb, tag); 1560 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1561 } else { 1562 spin_unlock_irqrestore(&t->task_state_lock, flags1); 1563 pm8001_ccb_task_free(pm8001_ha, t, ccb, tag); 1564 mb();/* in order to force CPU ordering */ 1565 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1566 t->task_done(t); 1567 } 1568 } break; 1569 case IO_XFER_OPEN_RETRY_TIMEOUT: 1570 { /* This one stashes the sas_task instead */ 1571 struct sas_task *t = (struct sas_task *)pm8001_dev; 1572 u32 tag; 1573 struct pm8001_ccb_info *ccb; 1574 struct pm8001_hba_info *pm8001_ha = pw->pm8001_ha; 1575 unsigned long flags, flags1; 1576 int i, ret = 0; 1577 1578 pm8001_dbg(pm8001_ha, IO, "IO_XFER_OPEN_RETRY_TIMEOUT\n"); 1579 1580 ret = pm8001_query_task(t); 1581 1582 if (ret == TMF_RESP_FUNC_SUCC) 1583 pm8001_dbg(pm8001_ha, IO, "...Task on lu\n"); 1584 else if (ret == TMF_RESP_FUNC_COMPLETE) 1585 pm8001_dbg(pm8001_ha, IO, "...Task NOT on lu\n"); 1586 else 1587 pm8001_dbg(pm8001_ha, DEVIO, "...query task failed!!!\n"); 1588 1589 spin_lock_irqsave(&pm8001_ha->lock, flags); 1590 1591 spin_lock_irqsave(&t->task_state_lock, flags1); 1592 1593 if (unlikely((t->task_state_flags & SAS_TASK_STATE_DONE))) { 1594 spin_unlock_irqrestore(&t->task_state_lock, flags1); 1595 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1596 if (ret == TMF_RESP_FUNC_SUCC) /* task on lu */ 1597 (void)pm8001_abort_task(t); 1598 break; /* Task got completed by another */ 1599 } 1600 1601 spin_unlock_irqrestore(&t->task_state_lock, flags1); 1602 1603 /* Search for a possible ccb that matches the task */ 1604 for (i = 0; ccb = NULL, i < PM8001_MAX_CCB; i++) { 1605 ccb = &pm8001_ha->ccb_info[i]; 1606 tag = ccb->ccb_tag; 1607 if ((tag != 0xFFFFFFFF) && (ccb->task == t)) 1608 break; 1609 } 1610 if (!ccb) { 1611 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1612 if (ret == TMF_RESP_FUNC_SUCC) /* task on lu */ 1613 (void)pm8001_abort_task(t); 1614 break; /* Task got freed by another */ 1615 } 1616 1617 pm8001_dev = ccb->device; 1618 dev = pm8001_dev->sas_device; 1619 1620 switch (ret) { 1621 case TMF_RESP_FUNC_SUCC: /* task on lu */ 1622 ccb->open_retry = 1; /* Snub completion */ 1623 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1624 ret = pm8001_abort_task(t); 1625 ccb->open_retry = 0; 1626 switch (ret) { 1627 case TMF_RESP_FUNC_SUCC: 1628 case TMF_RESP_FUNC_COMPLETE: 1629 break; 1630 default: /* device misbehavior */ 1631 ret = TMF_RESP_FUNC_FAILED; 1632 pm8001_dbg(pm8001_ha, IO, "...Reset phy\n"); 1633 pm8001_I_T_nexus_reset(dev); 1634 break; 1635 } 1636 break; 1637 1638 case TMF_RESP_FUNC_COMPLETE: /* task not on lu */ 1639 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1640 /* Do we need to abort the task locally? */ 1641 break; 1642 1643 default: /* device misbehavior */ 1644 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 1645 ret = TMF_RESP_FUNC_FAILED; 1646 pm8001_dbg(pm8001_ha, IO, "...Reset phy\n"); 1647 pm8001_I_T_nexus_reset(dev); 1648 } 1649 1650 if (ret == TMF_RESP_FUNC_FAILED) 1651 t = NULL; 1652 pm8001_open_reject_retry(pm8001_ha, t, pm8001_dev); 1653 pm8001_dbg(pm8001_ha, IO, "...Complete\n"); 1654 } break; 1655 case IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS: 1656 dev = pm8001_dev->sas_device; 1657 pm8001_I_T_nexus_event_handler(dev); 1658 break; 1659 case IO_OPEN_CNX_ERROR_STP_RESOURCES_BUSY: 1660 dev = pm8001_dev->sas_device; 1661 pm8001_I_T_nexus_reset(dev); 1662 break; 1663 case IO_DS_IN_ERROR: 1664 dev = pm8001_dev->sas_device; 1665 pm8001_I_T_nexus_reset(dev); 1666 break; 1667 case IO_DS_NON_OPERATIONAL: 1668 dev = pm8001_dev->sas_device; 1669 pm8001_I_T_nexus_reset(dev); 1670 break; 1671 } 1672 kfree(pw); 1673 } 1674 1675 int pm8001_handle_event(struct pm8001_hba_info *pm8001_ha, void *data, 1676 int handler) 1677 { 1678 struct pm8001_work *pw; 1679 int ret = 0; 1680 1681 pw = kmalloc(sizeof(struct pm8001_work), GFP_ATOMIC); 1682 if (pw) { 1683 pw->pm8001_ha = pm8001_ha; 1684 pw->data = data; 1685 pw->handler = handler; 1686 INIT_WORK(&pw->work, pm8001_work_fn); 1687 queue_work(pm8001_wq, &pw->work); 1688 } else 1689 ret = -ENOMEM; 1690 1691 return ret; 1692 } 1693 1694 static void pm8001_send_abort_all(struct pm8001_hba_info *pm8001_ha, 1695 struct pm8001_device *pm8001_ha_dev) 1696 { 1697 int res; 1698 u32 ccb_tag; 1699 struct pm8001_ccb_info *ccb; 1700 struct sas_task *task = NULL; 1701 struct task_abort_req task_abort; 1702 struct inbound_queue_table *circularQ; 1703 u32 opc = OPC_INB_SATA_ABORT; 1704 int ret; 1705 1706 if (!pm8001_ha_dev) { 1707 pm8001_dbg(pm8001_ha, FAIL, "dev is null\n"); 1708 return; 1709 } 1710 1711 task = sas_alloc_slow_task(GFP_ATOMIC); 1712 1713 if (!task) { 1714 pm8001_dbg(pm8001_ha, FAIL, "cannot allocate task\n"); 1715 return; 1716 } 1717 1718 task->task_done = pm8001_task_done; 1719 1720 res = pm8001_tag_alloc(pm8001_ha, &ccb_tag); 1721 if (res) 1722 return; 1723 1724 ccb = &pm8001_ha->ccb_info[ccb_tag]; 1725 ccb->device = pm8001_ha_dev; 1726 ccb->ccb_tag = ccb_tag; 1727 ccb->task = task; 1728 1729 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 1730 1731 memset(&task_abort, 0, sizeof(task_abort)); 1732 task_abort.abort_all = cpu_to_le32(1); 1733 task_abort.device_id = cpu_to_le32(pm8001_ha_dev->device_id); 1734 task_abort.tag = cpu_to_le32(ccb_tag); 1735 1736 ret = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &task_abort, 1737 sizeof(task_abort), 0); 1738 if (ret) 1739 pm8001_tag_free(pm8001_ha, ccb_tag); 1740 1741 } 1742 1743 static void pm8001_send_read_log(struct pm8001_hba_info *pm8001_ha, 1744 struct pm8001_device *pm8001_ha_dev) 1745 { 1746 struct sata_start_req sata_cmd; 1747 int res; 1748 u32 ccb_tag; 1749 struct pm8001_ccb_info *ccb; 1750 struct sas_task *task = NULL; 1751 struct host_to_dev_fis fis; 1752 struct domain_device *dev; 1753 struct inbound_queue_table *circularQ; 1754 u32 opc = OPC_INB_SATA_HOST_OPSTART; 1755 1756 task = sas_alloc_slow_task(GFP_ATOMIC); 1757 1758 if (!task) { 1759 pm8001_dbg(pm8001_ha, FAIL, "cannot allocate task !!!\n"); 1760 return; 1761 } 1762 task->task_done = pm8001_task_done; 1763 1764 res = pm8001_tag_alloc(pm8001_ha, &ccb_tag); 1765 if (res) { 1766 sas_free_task(task); 1767 pm8001_dbg(pm8001_ha, FAIL, "cannot allocate tag !!!\n"); 1768 return; 1769 } 1770 1771 /* allocate domain device by ourselves as libsas 1772 * is not going to provide any 1773 */ 1774 dev = kzalloc(sizeof(struct domain_device), GFP_ATOMIC); 1775 if (!dev) { 1776 sas_free_task(task); 1777 pm8001_tag_free(pm8001_ha, ccb_tag); 1778 pm8001_dbg(pm8001_ha, FAIL, 1779 "Domain device cannot be allocated\n"); 1780 return; 1781 } 1782 task->dev = dev; 1783 task->dev->lldd_dev = pm8001_ha_dev; 1784 1785 ccb = &pm8001_ha->ccb_info[ccb_tag]; 1786 ccb->device = pm8001_ha_dev; 1787 ccb->ccb_tag = ccb_tag; 1788 ccb->task = task; 1789 pm8001_ha_dev->id |= NCQ_READ_LOG_FLAG; 1790 pm8001_ha_dev->id |= NCQ_2ND_RLE_FLAG; 1791 1792 memset(&sata_cmd, 0, sizeof(sata_cmd)); 1793 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 1794 1795 /* construct read log FIS */ 1796 memset(&fis, 0, sizeof(struct host_to_dev_fis)); 1797 fis.fis_type = 0x27; 1798 fis.flags = 0x80; 1799 fis.command = ATA_CMD_READ_LOG_EXT; 1800 fis.lbal = 0x10; 1801 fis.sector_count = 0x1; 1802 1803 sata_cmd.tag = cpu_to_le32(ccb_tag); 1804 sata_cmd.device_id = cpu_to_le32(pm8001_ha_dev->device_id); 1805 sata_cmd.ncqtag_atap_dir_m |= ((0x1 << 7) | (0x5 << 9)); 1806 memcpy(&sata_cmd.sata_fis, &fis, sizeof(struct host_to_dev_fis)); 1807 1808 res = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &sata_cmd, 1809 sizeof(sata_cmd), 0); 1810 if (res) { 1811 sas_free_task(task); 1812 pm8001_tag_free(pm8001_ha, ccb_tag); 1813 kfree(dev); 1814 } 1815 } 1816 1817 /** 1818 * mpi_ssp_completion- process the event that FW response to the SSP request. 1819 * @pm8001_ha: our hba card information 1820 * @piomb: the message contents of this outbound message. 1821 * 1822 * When FW has completed a ssp request for example a IO request, after it has 1823 * filled the SG data with the data, it will trigger this event represent 1824 * that he has finished the job,please check the coresponding buffer. 1825 * So we will tell the caller who maybe waiting the result to tell upper layer 1826 * that the task has been finished. 1827 */ 1828 static void 1829 mpi_ssp_completion(struct pm8001_hba_info *pm8001_ha , void *piomb) 1830 { 1831 struct sas_task *t; 1832 struct pm8001_ccb_info *ccb; 1833 unsigned long flags; 1834 u32 status; 1835 u32 param; 1836 u32 tag; 1837 struct ssp_completion_resp *psspPayload; 1838 struct task_status_struct *ts; 1839 struct ssp_response_iu *iu; 1840 struct pm8001_device *pm8001_dev; 1841 psspPayload = (struct ssp_completion_resp *)(piomb + 4); 1842 status = le32_to_cpu(psspPayload->status); 1843 tag = le32_to_cpu(psspPayload->tag); 1844 ccb = &pm8001_ha->ccb_info[tag]; 1845 if ((status == IO_ABORTED) && ccb->open_retry) { 1846 /* Being completed by another */ 1847 ccb->open_retry = 0; 1848 return; 1849 } 1850 pm8001_dev = ccb->device; 1851 param = le32_to_cpu(psspPayload->param); 1852 1853 t = ccb->task; 1854 1855 if (status && status != IO_UNDERFLOW) 1856 pm8001_dbg(pm8001_ha, FAIL, "sas IO status 0x%x\n", status); 1857 if (unlikely(!t || !t->lldd_task || !t->dev)) 1858 return; 1859 ts = &t->task_status; 1860 /* Print sas address of IO failed device */ 1861 if ((status != IO_SUCCESS) && (status != IO_OVERFLOW) && 1862 (status != IO_UNDERFLOW)) 1863 pm8001_dbg(pm8001_ha, FAIL, "SAS Address of IO Failure Drive:%016llx\n", 1864 SAS_ADDR(t->dev->sas_addr)); 1865 1866 if (status) 1867 pm8001_dbg(pm8001_ha, IOERR, 1868 "status:0x%x, tag:0x%x, task:0x%p\n", 1869 status, tag, t); 1870 1871 switch (status) { 1872 case IO_SUCCESS: 1873 pm8001_dbg(pm8001_ha, IO, "IO_SUCCESS,param = %d\n", 1874 param); 1875 if (param == 0) { 1876 ts->resp = SAS_TASK_COMPLETE; 1877 ts->stat = SAM_STAT_GOOD; 1878 } else { 1879 ts->resp = SAS_TASK_COMPLETE; 1880 ts->stat = SAS_PROTO_RESPONSE; 1881 ts->residual = param; 1882 iu = &psspPayload->ssp_resp_iu; 1883 sas_ssp_task_response(pm8001_ha->dev, t, iu); 1884 } 1885 if (pm8001_dev) 1886 atomic_dec(&pm8001_dev->running_req); 1887 break; 1888 case IO_ABORTED: 1889 pm8001_dbg(pm8001_ha, IO, "IO_ABORTED IOMB Tag\n"); 1890 ts->resp = SAS_TASK_COMPLETE; 1891 ts->stat = SAS_ABORTED_TASK; 1892 break; 1893 case IO_UNDERFLOW: 1894 /* SSP Completion with error */ 1895 pm8001_dbg(pm8001_ha, IO, "IO_UNDERFLOW,param = %d\n", 1896 param); 1897 ts->resp = SAS_TASK_COMPLETE; 1898 ts->stat = SAS_DATA_UNDERRUN; 1899 ts->residual = param; 1900 if (pm8001_dev) 1901 atomic_dec(&pm8001_dev->running_req); 1902 break; 1903 case IO_NO_DEVICE: 1904 pm8001_dbg(pm8001_ha, IO, "IO_NO_DEVICE\n"); 1905 ts->resp = SAS_TASK_UNDELIVERED; 1906 ts->stat = SAS_PHY_DOWN; 1907 break; 1908 case IO_XFER_ERROR_BREAK: 1909 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_BREAK\n"); 1910 ts->resp = SAS_TASK_COMPLETE; 1911 ts->stat = SAS_OPEN_REJECT; 1912 /* Force the midlayer to retry */ 1913 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 1914 break; 1915 case IO_XFER_ERROR_PHY_NOT_READY: 1916 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_PHY_NOT_READY\n"); 1917 ts->resp = SAS_TASK_COMPLETE; 1918 ts->stat = SAS_OPEN_REJECT; 1919 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 1920 break; 1921 case IO_OPEN_CNX_ERROR_PROTOCOL_NOT_SUPPORTED: 1922 pm8001_dbg(pm8001_ha, IO, 1923 "IO_OPEN_CNX_ERROR_PROTOCOL_NOT_SUPPORTED\n"); 1924 ts->resp = SAS_TASK_COMPLETE; 1925 ts->stat = SAS_OPEN_REJECT; 1926 ts->open_rej_reason = SAS_OREJ_EPROTO; 1927 break; 1928 case IO_OPEN_CNX_ERROR_ZONE_VIOLATION: 1929 pm8001_dbg(pm8001_ha, IO, 1930 "IO_OPEN_CNX_ERROR_ZONE_VIOLATION\n"); 1931 ts->resp = SAS_TASK_COMPLETE; 1932 ts->stat = SAS_OPEN_REJECT; 1933 ts->open_rej_reason = SAS_OREJ_UNKNOWN; 1934 break; 1935 case IO_OPEN_CNX_ERROR_BREAK: 1936 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_BREAK\n"); 1937 ts->resp = SAS_TASK_COMPLETE; 1938 ts->stat = SAS_OPEN_REJECT; 1939 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 1940 break; 1941 case IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS: 1942 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS\n"); 1943 ts->resp = SAS_TASK_COMPLETE; 1944 ts->stat = SAS_OPEN_REJECT; 1945 ts->open_rej_reason = SAS_OREJ_UNKNOWN; 1946 if (!t->uldd_task) 1947 pm8001_handle_event(pm8001_ha, 1948 pm8001_dev, 1949 IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS); 1950 break; 1951 case IO_OPEN_CNX_ERROR_BAD_DESTINATION: 1952 pm8001_dbg(pm8001_ha, IO, 1953 "IO_OPEN_CNX_ERROR_BAD_DESTINATION\n"); 1954 ts->resp = SAS_TASK_COMPLETE; 1955 ts->stat = SAS_OPEN_REJECT; 1956 ts->open_rej_reason = SAS_OREJ_BAD_DEST; 1957 break; 1958 case IO_OPEN_CNX_ERROR_CONNECTION_RATE_NOT_SUPPORTED: 1959 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_CONNECTION_RATE_NOT_SUPPORTED\n"); 1960 ts->resp = SAS_TASK_COMPLETE; 1961 ts->stat = SAS_OPEN_REJECT; 1962 ts->open_rej_reason = SAS_OREJ_CONN_RATE; 1963 break; 1964 case IO_OPEN_CNX_ERROR_WRONG_DESTINATION: 1965 pm8001_dbg(pm8001_ha, IO, 1966 "IO_OPEN_CNX_ERROR_WRONG_DESTINATION\n"); 1967 ts->resp = SAS_TASK_UNDELIVERED; 1968 ts->stat = SAS_OPEN_REJECT; 1969 ts->open_rej_reason = SAS_OREJ_WRONG_DEST; 1970 break; 1971 case IO_XFER_ERROR_NAK_RECEIVED: 1972 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_NAK_RECEIVED\n"); 1973 ts->resp = SAS_TASK_COMPLETE; 1974 ts->stat = SAS_OPEN_REJECT; 1975 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 1976 break; 1977 case IO_XFER_ERROR_ACK_NAK_TIMEOUT: 1978 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_ACK_NAK_TIMEOUT\n"); 1979 ts->resp = SAS_TASK_COMPLETE; 1980 ts->stat = SAS_NAK_R_ERR; 1981 break; 1982 case IO_XFER_ERROR_DMA: 1983 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_DMA\n"); 1984 ts->resp = SAS_TASK_COMPLETE; 1985 ts->stat = SAS_OPEN_REJECT; 1986 break; 1987 case IO_XFER_OPEN_RETRY_TIMEOUT: 1988 pm8001_dbg(pm8001_ha, IO, "IO_XFER_OPEN_RETRY_TIMEOUT\n"); 1989 ts->resp = SAS_TASK_COMPLETE; 1990 ts->stat = SAS_OPEN_REJECT; 1991 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 1992 break; 1993 case IO_XFER_ERROR_OFFSET_MISMATCH: 1994 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_OFFSET_MISMATCH\n"); 1995 ts->resp = SAS_TASK_COMPLETE; 1996 ts->stat = SAS_OPEN_REJECT; 1997 break; 1998 case IO_PORT_IN_RESET: 1999 pm8001_dbg(pm8001_ha, IO, "IO_PORT_IN_RESET\n"); 2000 ts->resp = SAS_TASK_COMPLETE; 2001 ts->stat = SAS_OPEN_REJECT; 2002 break; 2003 case IO_DS_NON_OPERATIONAL: 2004 pm8001_dbg(pm8001_ha, IO, "IO_DS_NON_OPERATIONAL\n"); 2005 ts->resp = SAS_TASK_COMPLETE; 2006 ts->stat = SAS_OPEN_REJECT; 2007 if (!t->uldd_task) 2008 pm8001_handle_event(pm8001_ha, 2009 pm8001_dev, 2010 IO_DS_NON_OPERATIONAL); 2011 break; 2012 case IO_DS_IN_RECOVERY: 2013 pm8001_dbg(pm8001_ha, IO, "IO_DS_IN_RECOVERY\n"); 2014 ts->resp = SAS_TASK_COMPLETE; 2015 ts->stat = SAS_OPEN_REJECT; 2016 break; 2017 case IO_TM_TAG_NOT_FOUND: 2018 pm8001_dbg(pm8001_ha, IO, "IO_TM_TAG_NOT_FOUND\n"); 2019 ts->resp = SAS_TASK_COMPLETE; 2020 ts->stat = SAS_OPEN_REJECT; 2021 break; 2022 case IO_SSP_EXT_IU_ZERO_LEN_ERROR: 2023 pm8001_dbg(pm8001_ha, IO, "IO_SSP_EXT_IU_ZERO_LEN_ERROR\n"); 2024 ts->resp = SAS_TASK_COMPLETE; 2025 ts->stat = SAS_OPEN_REJECT; 2026 break; 2027 case IO_OPEN_CNX_ERROR_HW_RESOURCE_BUSY: 2028 pm8001_dbg(pm8001_ha, IO, 2029 "IO_OPEN_CNX_ERROR_HW_RESOURCE_BUSY\n"); 2030 ts->resp = SAS_TASK_COMPLETE; 2031 ts->stat = SAS_OPEN_REJECT; 2032 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 2033 break; 2034 default: 2035 pm8001_dbg(pm8001_ha, DEVIO, "Unknown status 0x%x\n", status); 2036 /* not allowed case. Therefore, return failed status */ 2037 ts->resp = SAS_TASK_COMPLETE; 2038 ts->stat = SAS_OPEN_REJECT; 2039 break; 2040 } 2041 pm8001_dbg(pm8001_ha, IO, "scsi_status = %x\n", 2042 psspPayload->ssp_resp_iu.status); 2043 spin_lock_irqsave(&t->task_state_lock, flags); 2044 t->task_state_flags &= ~SAS_TASK_STATE_PENDING; 2045 t->task_state_flags &= ~SAS_TASK_AT_INITIATOR; 2046 t->task_state_flags |= SAS_TASK_STATE_DONE; 2047 if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) { 2048 spin_unlock_irqrestore(&t->task_state_lock, flags); 2049 pm8001_dbg(pm8001_ha, FAIL, "task 0x%p done with io_status 0x%x resp 0x%x stat 0x%x but aborted by upper layer!\n", 2050 t, status, ts->resp, ts->stat); 2051 pm8001_ccb_task_free(pm8001_ha, t, ccb, tag); 2052 } else { 2053 spin_unlock_irqrestore(&t->task_state_lock, flags); 2054 pm8001_ccb_task_free(pm8001_ha, t, ccb, tag); 2055 mb();/* in order to force CPU ordering */ 2056 t->task_done(t); 2057 } 2058 } 2059 2060 /*See the comments for mpi_ssp_completion */ 2061 static void mpi_ssp_event(struct pm8001_hba_info *pm8001_ha , void *piomb) 2062 { 2063 struct sas_task *t; 2064 unsigned long flags; 2065 struct task_status_struct *ts; 2066 struct pm8001_ccb_info *ccb; 2067 struct pm8001_device *pm8001_dev; 2068 struct ssp_event_resp *psspPayload = 2069 (struct ssp_event_resp *)(piomb + 4); 2070 u32 event = le32_to_cpu(psspPayload->event); 2071 u32 tag = le32_to_cpu(psspPayload->tag); 2072 u32 port_id = le32_to_cpu(psspPayload->port_id); 2073 u32 dev_id = le32_to_cpu(psspPayload->device_id); 2074 2075 ccb = &pm8001_ha->ccb_info[tag]; 2076 t = ccb->task; 2077 pm8001_dev = ccb->device; 2078 if (event) 2079 pm8001_dbg(pm8001_ha, FAIL, "sas IO status 0x%x\n", event); 2080 if (unlikely(!t || !t->lldd_task || !t->dev)) 2081 return; 2082 ts = &t->task_status; 2083 pm8001_dbg(pm8001_ha, DEVIO, "port_id = %x,device_id = %x\n", 2084 port_id, dev_id); 2085 switch (event) { 2086 case IO_OVERFLOW: 2087 pm8001_dbg(pm8001_ha, IO, "IO_UNDERFLOW\n"); 2088 ts->resp = SAS_TASK_COMPLETE; 2089 ts->stat = SAS_DATA_OVERRUN; 2090 ts->residual = 0; 2091 if (pm8001_dev) 2092 atomic_dec(&pm8001_dev->running_req); 2093 break; 2094 case IO_XFER_ERROR_BREAK: 2095 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_BREAK\n"); 2096 pm8001_handle_event(pm8001_ha, t, IO_XFER_ERROR_BREAK); 2097 return; 2098 case IO_XFER_ERROR_PHY_NOT_READY: 2099 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_PHY_NOT_READY\n"); 2100 ts->resp = SAS_TASK_COMPLETE; 2101 ts->stat = SAS_OPEN_REJECT; 2102 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 2103 break; 2104 case IO_OPEN_CNX_ERROR_PROTOCOL_NOT_SUPPORTED: 2105 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_PROTOCOL_NOT_SUPPORTED\n"); 2106 ts->resp = SAS_TASK_COMPLETE; 2107 ts->stat = SAS_OPEN_REJECT; 2108 ts->open_rej_reason = SAS_OREJ_EPROTO; 2109 break; 2110 case IO_OPEN_CNX_ERROR_ZONE_VIOLATION: 2111 pm8001_dbg(pm8001_ha, IO, 2112 "IO_OPEN_CNX_ERROR_ZONE_VIOLATION\n"); 2113 ts->resp = SAS_TASK_COMPLETE; 2114 ts->stat = SAS_OPEN_REJECT; 2115 ts->open_rej_reason = SAS_OREJ_UNKNOWN; 2116 break; 2117 case IO_OPEN_CNX_ERROR_BREAK: 2118 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_BREAK\n"); 2119 ts->resp = SAS_TASK_COMPLETE; 2120 ts->stat = SAS_OPEN_REJECT; 2121 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 2122 break; 2123 case IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS: 2124 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS\n"); 2125 ts->resp = SAS_TASK_COMPLETE; 2126 ts->stat = SAS_OPEN_REJECT; 2127 ts->open_rej_reason = SAS_OREJ_UNKNOWN; 2128 if (!t->uldd_task) 2129 pm8001_handle_event(pm8001_ha, 2130 pm8001_dev, 2131 IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS); 2132 break; 2133 case IO_OPEN_CNX_ERROR_BAD_DESTINATION: 2134 pm8001_dbg(pm8001_ha, IO, 2135 "IO_OPEN_CNX_ERROR_BAD_DESTINATION\n"); 2136 ts->resp = SAS_TASK_COMPLETE; 2137 ts->stat = SAS_OPEN_REJECT; 2138 ts->open_rej_reason = SAS_OREJ_BAD_DEST; 2139 break; 2140 case IO_OPEN_CNX_ERROR_CONNECTION_RATE_NOT_SUPPORTED: 2141 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_CONNECTION_RATE_NOT_SUPPORTED\n"); 2142 ts->resp = SAS_TASK_COMPLETE; 2143 ts->stat = SAS_OPEN_REJECT; 2144 ts->open_rej_reason = SAS_OREJ_CONN_RATE; 2145 break; 2146 case IO_OPEN_CNX_ERROR_WRONG_DESTINATION: 2147 pm8001_dbg(pm8001_ha, IO, 2148 "IO_OPEN_CNX_ERROR_WRONG_DESTINATION\n"); 2149 ts->resp = SAS_TASK_COMPLETE; 2150 ts->stat = SAS_OPEN_REJECT; 2151 ts->open_rej_reason = SAS_OREJ_WRONG_DEST; 2152 break; 2153 case IO_XFER_ERROR_NAK_RECEIVED: 2154 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_NAK_RECEIVED\n"); 2155 ts->resp = SAS_TASK_COMPLETE; 2156 ts->stat = SAS_OPEN_REJECT; 2157 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 2158 break; 2159 case IO_XFER_ERROR_ACK_NAK_TIMEOUT: 2160 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_ACK_NAK_TIMEOUT\n"); 2161 ts->resp = SAS_TASK_COMPLETE; 2162 ts->stat = SAS_NAK_R_ERR; 2163 break; 2164 case IO_XFER_OPEN_RETRY_TIMEOUT: 2165 pm8001_dbg(pm8001_ha, IO, "IO_XFER_OPEN_RETRY_TIMEOUT\n"); 2166 pm8001_handle_event(pm8001_ha, t, IO_XFER_OPEN_RETRY_TIMEOUT); 2167 return; 2168 case IO_XFER_ERROR_UNEXPECTED_PHASE: 2169 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_UNEXPECTED_PHASE\n"); 2170 ts->resp = SAS_TASK_COMPLETE; 2171 ts->stat = SAS_DATA_OVERRUN; 2172 break; 2173 case IO_XFER_ERROR_XFER_RDY_OVERRUN: 2174 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_XFER_RDY_OVERRUN\n"); 2175 ts->resp = SAS_TASK_COMPLETE; 2176 ts->stat = SAS_DATA_OVERRUN; 2177 break; 2178 case IO_XFER_ERROR_XFER_RDY_NOT_EXPECTED: 2179 pm8001_dbg(pm8001_ha, IO, 2180 "IO_XFER_ERROR_XFER_RDY_NOT_EXPECTED\n"); 2181 ts->resp = SAS_TASK_COMPLETE; 2182 ts->stat = SAS_DATA_OVERRUN; 2183 break; 2184 case IO_XFER_ERROR_CMD_ISSUE_ACK_NAK_TIMEOUT: 2185 pm8001_dbg(pm8001_ha, IO, 2186 "IO_XFER_ERROR_CMD_ISSUE_ACK_NAK_TIMEOUT\n"); 2187 ts->resp = SAS_TASK_COMPLETE; 2188 ts->stat = SAS_DATA_OVERRUN; 2189 break; 2190 case IO_XFER_ERROR_OFFSET_MISMATCH: 2191 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_OFFSET_MISMATCH\n"); 2192 ts->resp = SAS_TASK_COMPLETE; 2193 ts->stat = SAS_DATA_OVERRUN; 2194 break; 2195 case IO_XFER_ERROR_XFER_ZERO_DATA_LEN: 2196 pm8001_dbg(pm8001_ha, IO, 2197 "IO_XFER_ERROR_XFER_ZERO_DATA_LEN\n"); 2198 ts->resp = SAS_TASK_COMPLETE; 2199 ts->stat = SAS_DATA_OVERRUN; 2200 break; 2201 case IO_XFER_CMD_FRAME_ISSUED: 2202 pm8001_dbg(pm8001_ha, IO, "IO_XFER_CMD_FRAME_ISSUED\n"); 2203 return; 2204 default: 2205 pm8001_dbg(pm8001_ha, DEVIO, "Unknown status 0x%x\n", event); 2206 /* not allowed case. Therefore, return failed status */ 2207 ts->resp = SAS_TASK_COMPLETE; 2208 ts->stat = SAS_DATA_OVERRUN; 2209 break; 2210 } 2211 spin_lock_irqsave(&t->task_state_lock, flags); 2212 t->task_state_flags &= ~SAS_TASK_STATE_PENDING; 2213 t->task_state_flags &= ~SAS_TASK_AT_INITIATOR; 2214 t->task_state_flags |= SAS_TASK_STATE_DONE; 2215 if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) { 2216 spin_unlock_irqrestore(&t->task_state_lock, flags); 2217 pm8001_dbg(pm8001_ha, FAIL, "task 0x%p done with event 0x%x resp 0x%x stat 0x%x but aborted by upper layer!\n", 2218 t, event, ts->resp, ts->stat); 2219 pm8001_ccb_task_free(pm8001_ha, t, ccb, tag); 2220 } else { 2221 spin_unlock_irqrestore(&t->task_state_lock, flags); 2222 pm8001_ccb_task_free(pm8001_ha, t, ccb, tag); 2223 mb();/* in order to force CPU ordering */ 2224 t->task_done(t); 2225 } 2226 } 2227 2228 /*See the comments for mpi_ssp_completion */ 2229 static void 2230 mpi_sata_completion(struct pm8001_hba_info *pm8001_ha, void *piomb) 2231 { 2232 struct sas_task *t; 2233 struct pm8001_ccb_info *ccb; 2234 u32 param; 2235 u32 status; 2236 u32 tag; 2237 int i, j; 2238 u8 sata_addr_low[4]; 2239 u32 temp_sata_addr_low; 2240 u8 sata_addr_hi[4]; 2241 u32 temp_sata_addr_hi; 2242 struct sata_completion_resp *psataPayload; 2243 struct task_status_struct *ts; 2244 struct ata_task_resp *resp ; 2245 u32 *sata_resp; 2246 struct pm8001_device *pm8001_dev; 2247 unsigned long flags; 2248 2249 psataPayload = (struct sata_completion_resp *)(piomb + 4); 2250 status = le32_to_cpu(psataPayload->status); 2251 tag = le32_to_cpu(psataPayload->tag); 2252 2253 if (!tag) { 2254 pm8001_dbg(pm8001_ha, FAIL, "tag null\n"); 2255 return; 2256 } 2257 ccb = &pm8001_ha->ccb_info[tag]; 2258 param = le32_to_cpu(psataPayload->param); 2259 if (ccb) { 2260 t = ccb->task; 2261 pm8001_dev = ccb->device; 2262 } else { 2263 pm8001_dbg(pm8001_ha, FAIL, "ccb null\n"); 2264 return; 2265 } 2266 2267 if (t) { 2268 if (t->dev && (t->dev->lldd_dev)) 2269 pm8001_dev = t->dev->lldd_dev; 2270 } else { 2271 pm8001_dbg(pm8001_ha, FAIL, "task null\n"); 2272 return; 2273 } 2274 2275 if ((pm8001_dev && !(pm8001_dev->id & NCQ_READ_LOG_FLAG)) 2276 && unlikely(!t || !t->lldd_task || !t->dev)) { 2277 pm8001_dbg(pm8001_ha, FAIL, "task or dev null\n"); 2278 return; 2279 } 2280 2281 ts = &t->task_status; 2282 if (!ts) { 2283 pm8001_dbg(pm8001_ha, FAIL, "ts null\n"); 2284 return; 2285 } 2286 2287 if (status) 2288 pm8001_dbg(pm8001_ha, IOERR, 2289 "status:0x%x, tag:0x%x, task::0x%p\n", 2290 status, tag, t); 2291 2292 /* Print sas address of IO failed device */ 2293 if ((status != IO_SUCCESS) && (status != IO_OVERFLOW) && 2294 (status != IO_UNDERFLOW)) { 2295 if (!((t->dev->parent) && 2296 (dev_is_expander(t->dev->parent->dev_type)))) { 2297 for (i = 0 , j = 4; j <= 7 && i <= 3; i++ , j++) 2298 sata_addr_low[i] = pm8001_ha->sas_addr[j]; 2299 for (i = 0 , j = 0; j <= 3 && i <= 3; i++ , j++) 2300 sata_addr_hi[i] = pm8001_ha->sas_addr[j]; 2301 memcpy(&temp_sata_addr_low, sata_addr_low, 2302 sizeof(sata_addr_low)); 2303 memcpy(&temp_sata_addr_hi, sata_addr_hi, 2304 sizeof(sata_addr_hi)); 2305 temp_sata_addr_hi = (((temp_sata_addr_hi >> 24) & 0xff) 2306 |((temp_sata_addr_hi << 8) & 2307 0xff0000) | 2308 ((temp_sata_addr_hi >> 8) 2309 & 0xff00) | 2310 ((temp_sata_addr_hi << 24) & 2311 0xff000000)); 2312 temp_sata_addr_low = ((((temp_sata_addr_low >> 24) 2313 & 0xff) | 2314 ((temp_sata_addr_low << 8) 2315 & 0xff0000) | 2316 ((temp_sata_addr_low >> 8) 2317 & 0xff00) | 2318 ((temp_sata_addr_low << 24) 2319 & 0xff000000)) + 2320 pm8001_dev->attached_phy + 2321 0x10); 2322 pm8001_dbg(pm8001_ha, FAIL, 2323 "SAS Address of IO Failure Drive:%08x%08x\n", 2324 temp_sata_addr_hi, 2325 temp_sata_addr_low); 2326 } else { 2327 pm8001_dbg(pm8001_ha, FAIL, 2328 "SAS Address of IO Failure Drive:%016llx\n", 2329 SAS_ADDR(t->dev->sas_addr)); 2330 } 2331 } 2332 switch (status) { 2333 case IO_SUCCESS: 2334 pm8001_dbg(pm8001_ha, IO, "IO_SUCCESS\n"); 2335 if (param == 0) { 2336 ts->resp = SAS_TASK_COMPLETE; 2337 ts->stat = SAM_STAT_GOOD; 2338 /* check if response is for SEND READ LOG */ 2339 if (pm8001_dev && 2340 (pm8001_dev->id & NCQ_READ_LOG_FLAG)) { 2341 /* set new bit for abort_all */ 2342 pm8001_dev->id |= NCQ_ABORT_ALL_FLAG; 2343 /* clear bit for read log */ 2344 pm8001_dev->id = pm8001_dev->id & 0x7FFFFFFF; 2345 pm8001_send_abort_all(pm8001_ha, pm8001_dev); 2346 /* Free the tag */ 2347 pm8001_tag_free(pm8001_ha, tag); 2348 sas_free_task(t); 2349 return; 2350 } 2351 } else { 2352 u8 len; 2353 ts->resp = SAS_TASK_COMPLETE; 2354 ts->stat = SAS_PROTO_RESPONSE; 2355 ts->residual = param; 2356 pm8001_dbg(pm8001_ha, IO, 2357 "SAS_PROTO_RESPONSE len = %d\n", 2358 param); 2359 sata_resp = &psataPayload->sata_resp[0]; 2360 resp = (struct ata_task_resp *)ts->buf; 2361 if (t->ata_task.dma_xfer == 0 && 2362 t->data_dir == DMA_FROM_DEVICE) { 2363 len = sizeof(struct pio_setup_fis); 2364 pm8001_dbg(pm8001_ha, IO, 2365 "PIO read len = %d\n", len); 2366 } else if (t->ata_task.use_ncq) { 2367 len = sizeof(struct set_dev_bits_fis); 2368 pm8001_dbg(pm8001_ha, IO, "FPDMA len = %d\n", 2369 len); 2370 } else { 2371 len = sizeof(struct dev_to_host_fis); 2372 pm8001_dbg(pm8001_ha, IO, "other len = %d\n", 2373 len); 2374 } 2375 if (SAS_STATUS_BUF_SIZE >= sizeof(*resp)) { 2376 resp->frame_len = len; 2377 memcpy(&resp->ending_fis[0], sata_resp, len); 2378 ts->buf_valid_size = sizeof(*resp); 2379 } else 2380 pm8001_dbg(pm8001_ha, IO, 2381 "response too large\n"); 2382 } 2383 if (pm8001_dev) 2384 atomic_dec(&pm8001_dev->running_req); 2385 break; 2386 case IO_ABORTED: 2387 pm8001_dbg(pm8001_ha, IO, "IO_ABORTED IOMB Tag\n"); 2388 ts->resp = SAS_TASK_COMPLETE; 2389 ts->stat = SAS_ABORTED_TASK; 2390 if (pm8001_dev) 2391 atomic_dec(&pm8001_dev->running_req); 2392 break; 2393 /* following cases are to do cases */ 2394 case IO_UNDERFLOW: 2395 /* SATA Completion with error */ 2396 pm8001_dbg(pm8001_ha, IO, "IO_UNDERFLOW param = %d\n", param); 2397 ts->resp = SAS_TASK_COMPLETE; 2398 ts->stat = SAS_DATA_UNDERRUN; 2399 ts->residual = param; 2400 if (pm8001_dev) 2401 atomic_dec(&pm8001_dev->running_req); 2402 break; 2403 case IO_NO_DEVICE: 2404 pm8001_dbg(pm8001_ha, IO, "IO_NO_DEVICE\n"); 2405 ts->resp = SAS_TASK_UNDELIVERED; 2406 ts->stat = SAS_PHY_DOWN; 2407 if (pm8001_dev) 2408 atomic_dec(&pm8001_dev->running_req); 2409 break; 2410 case IO_XFER_ERROR_BREAK: 2411 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_BREAK\n"); 2412 ts->resp = SAS_TASK_COMPLETE; 2413 ts->stat = SAS_INTERRUPTED; 2414 if (pm8001_dev) 2415 atomic_dec(&pm8001_dev->running_req); 2416 break; 2417 case IO_XFER_ERROR_PHY_NOT_READY: 2418 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_PHY_NOT_READY\n"); 2419 ts->resp = SAS_TASK_COMPLETE; 2420 ts->stat = SAS_OPEN_REJECT; 2421 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 2422 if (pm8001_dev) 2423 atomic_dec(&pm8001_dev->running_req); 2424 break; 2425 case IO_OPEN_CNX_ERROR_PROTOCOL_NOT_SUPPORTED: 2426 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_PROTOCOL_NOT_SUPPORTED\n"); 2427 ts->resp = SAS_TASK_COMPLETE; 2428 ts->stat = SAS_OPEN_REJECT; 2429 ts->open_rej_reason = SAS_OREJ_EPROTO; 2430 if (pm8001_dev) 2431 atomic_dec(&pm8001_dev->running_req); 2432 break; 2433 case IO_OPEN_CNX_ERROR_ZONE_VIOLATION: 2434 pm8001_dbg(pm8001_ha, IO, 2435 "IO_OPEN_CNX_ERROR_ZONE_VIOLATION\n"); 2436 ts->resp = SAS_TASK_COMPLETE; 2437 ts->stat = SAS_OPEN_REJECT; 2438 ts->open_rej_reason = SAS_OREJ_UNKNOWN; 2439 if (pm8001_dev) 2440 atomic_dec(&pm8001_dev->running_req); 2441 break; 2442 case IO_OPEN_CNX_ERROR_BREAK: 2443 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_BREAK\n"); 2444 ts->resp = SAS_TASK_COMPLETE; 2445 ts->stat = SAS_OPEN_REJECT; 2446 ts->open_rej_reason = SAS_OREJ_RSVD_CONT0; 2447 if (pm8001_dev) 2448 atomic_dec(&pm8001_dev->running_req); 2449 break; 2450 case IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS: 2451 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS\n"); 2452 ts->resp = SAS_TASK_COMPLETE; 2453 ts->stat = SAS_DEV_NO_RESPONSE; 2454 if (!t->uldd_task) { 2455 pm8001_handle_event(pm8001_ha, 2456 pm8001_dev, 2457 IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS); 2458 ts->resp = SAS_TASK_UNDELIVERED; 2459 ts->stat = SAS_QUEUE_FULL; 2460 pm8001_ccb_task_free_done(pm8001_ha, t, ccb, tag); 2461 return; 2462 } 2463 break; 2464 case IO_OPEN_CNX_ERROR_BAD_DESTINATION: 2465 pm8001_dbg(pm8001_ha, IO, 2466 "IO_OPEN_CNX_ERROR_BAD_DESTINATION\n"); 2467 ts->resp = SAS_TASK_UNDELIVERED; 2468 ts->stat = SAS_OPEN_REJECT; 2469 ts->open_rej_reason = SAS_OREJ_BAD_DEST; 2470 if (!t->uldd_task) { 2471 pm8001_handle_event(pm8001_ha, 2472 pm8001_dev, 2473 IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS); 2474 ts->resp = SAS_TASK_UNDELIVERED; 2475 ts->stat = SAS_QUEUE_FULL; 2476 pm8001_ccb_task_free_done(pm8001_ha, t, ccb, tag); 2477 return; 2478 } 2479 break; 2480 case IO_OPEN_CNX_ERROR_CONNECTION_RATE_NOT_SUPPORTED: 2481 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_CONNECTION_RATE_NOT_SUPPORTED\n"); 2482 ts->resp = SAS_TASK_COMPLETE; 2483 ts->stat = SAS_OPEN_REJECT; 2484 ts->open_rej_reason = SAS_OREJ_CONN_RATE; 2485 if (pm8001_dev) 2486 atomic_dec(&pm8001_dev->running_req); 2487 break; 2488 case IO_OPEN_CNX_ERROR_STP_RESOURCES_BUSY: 2489 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_STP_RESOURCES_BUSY\n"); 2490 ts->resp = SAS_TASK_COMPLETE; 2491 ts->stat = SAS_DEV_NO_RESPONSE; 2492 if (!t->uldd_task) { 2493 pm8001_handle_event(pm8001_ha, 2494 pm8001_dev, 2495 IO_OPEN_CNX_ERROR_STP_RESOURCES_BUSY); 2496 ts->resp = SAS_TASK_UNDELIVERED; 2497 ts->stat = SAS_QUEUE_FULL; 2498 pm8001_ccb_task_free_done(pm8001_ha, t, ccb, tag); 2499 return; 2500 } 2501 break; 2502 case IO_OPEN_CNX_ERROR_WRONG_DESTINATION: 2503 pm8001_dbg(pm8001_ha, IO, 2504 "IO_OPEN_CNX_ERROR_WRONG_DESTINATION\n"); 2505 ts->resp = SAS_TASK_COMPLETE; 2506 ts->stat = SAS_OPEN_REJECT; 2507 ts->open_rej_reason = SAS_OREJ_WRONG_DEST; 2508 if (pm8001_dev) 2509 atomic_dec(&pm8001_dev->running_req); 2510 break; 2511 case IO_XFER_ERROR_NAK_RECEIVED: 2512 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_NAK_RECEIVED\n"); 2513 ts->resp = SAS_TASK_COMPLETE; 2514 ts->stat = SAS_NAK_R_ERR; 2515 if (pm8001_dev) 2516 atomic_dec(&pm8001_dev->running_req); 2517 break; 2518 case IO_XFER_ERROR_ACK_NAK_TIMEOUT: 2519 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_ACK_NAK_TIMEOUT\n"); 2520 ts->resp = SAS_TASK_COMPLETE; 2521 ts->stat = SAS_NAK_R_ERR; 2522 if (pm8001_dev) 2523 atomic_dec(&pm8001_dev->running_req); 2524 break; 2525 case IO_XFER_ERROR_DMA: 2526 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_DMA\n"); 2527 ts->resp = SAS_TASK_COMPLETE; 2528 ts->stat = SAS_ABORTED_TASK; 2529 if (pm8001_dev) 2530 atomic_dec(&pm8001_dev->running_req); 2531 break; 2532 case IO_XFER_ERROR_SATA_LINK_TIMEOUT: 2533 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_SATA_LINK_TIMEOUT\n"); 2534 ts->resp = SAS_TASK_UNDELIVERED; 2535 ts->stat = SAS_DEV_NO_RESPONSE; 2536 if (pm8001_dev) 2537 atomic_dec(&pm8001_dev->running_req); 2538 break; 2539 case IO_XFER_ERROR_REJECTED_NCQ_MODE: 2540 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_REJECTED_NCQ_MODE\n"); 2541 ts->resp = SAS_TASK_COMPLETE; 2542 ts->stat = SAS_DATA_UNDERRUN; 2543 if (pm8001_dev) 2544 atomic_dec(&pm8001_dev->running_req); 2545 break; 2546 case IO_XFER_OPEN_RETRY_TIMEOUT: 2547 pm8001_dbg(pm8001_ha, IO, "IO_XFER_OPEN_RETRY_TIMEOUT\n"); 2548 ts->resp = SAS_TASK_COMPLETE; 2549 ts->stat = SAS_OPEN_TO; 2550 if (pm8001_dev) 2551 atomic_dec(&pm8001_dev->running_req); 2552 break; 2553 case IO_PORT_IN_RESET: 2554 pm8001_dbg(pm8001_ha, IO, "IO_PORT_IN_RESET\n"); 2555 ts->resp = SAS_TASK_COMPLETE; 2556 ts->stat = SAS_DEV_NO_RESPONSE; 2557 if (pm8001_dev) 2558 atomic_dec(&pm8001_dev->running_req); 2559 break; 2560 case IO_DS_NON_OPERATIONAL: 2561 pm8001_dbg(pm8001_ha, IO, "IO_DS_NON_OPERATIONAL\n"); 2562 ts->resp = SAS_TASK_COMPLETE; 2563 ts->stat = SAS_DEV_NO_RESPONSE; 2564 if (!t->uldd_task) { 2565 pm8001_handle_event(pm8001_ha, pm8001_dev, 2566 IO_DS_NON_OPERATIONAL); 2567 ts->resp = SAS_TASK_UNDELIVERED; 2568 ts->stat = SAS_QUEUE_FULL; 2569 pm8001_ccb_task_free_done(pm8001_ha, t, ccb, tag); 2570 return; 2571 } 2572 break; 2573 case IO_DS_IN_RECOVERY: 2574 pm8001_dbg(pm8001_ha, IO, " IO_DS_IN_RECOVERY\n"); 2575 ts->resp = SAS_TASK_COMPLETE; 2576 ts->stat = SAS_DEV_NO_RESPONSE; 2577 if (pm8001_dev) 2578 atomic_dec(&pm8001_dev->running_req); 2579 break; 2580 case IO_DS_IN_ERROR: 2581 pm8001_dbg(pm8001_ha, IO, "IO_DS_IN_ERROR\n"); 2582 ts->resp = SAS_TASK_COMPLETE; 2583 ts->stat = SAS_DEV_NO_RESPONSE; 2584 if (!t->uldd_task) { 2585 pm8001_handle_event(pm8001_ha, pm8001_dev, 2586 IO_DS_IN_ERROR); 2587 ts->resp = SAS_TASK_UNDELIVERED; 2588 ts->stat = SAS_QUEUE_FULL; 2589 pm8001_ccb_task_free_done(pm8001_ha, t, ccb, tag); 2590 return; 2591 } 2592 break; 2593 case IO_OPEN_CNX_ERROR_HW_RESOURCE_BUSY: 2594 pm8001_dbg(pm8001_ha, IO, 2595 "IO_OPEN_CNX_ERROR_HW_RESOURCE_BUSY\n"); 2596 ts->resp = SAS_TASK_COMPLETE; 2597 ts->stat = SAS_OPEN_REJECT; 2598 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 2599 if (pm8001_dev) 2600 atomic_dec(&pm8001_dev->running_req); 2601 break; 2602 default: 2603 pm8001_dbg(pm8001_ha, DEVIO, "Unknown status 0x%x\n", status); 2604 /* not allowed case. Therefore, return failed status */ 2605 ts->resp = SAS_TASK_COMPLETE; 2606 ts->stat = SAS_DEV_NO_RESPONSE; 2607 if (pm8001_dev) 2608 atomic_dec(&pm8001_dev->running_req); 2609 break; 2610 } 2611 spin_lock_irqsave(&t->task_state_lock, flags); 2612 t->task_state_flags &= ~SAS_TASK_STATE_PENDING; 2613 t->task_state_flags &= ~SAS_TASK_AT_INITIATOR; 2614 t->task_state_flags |= SAS_TASK_STATE_DONE; 2615 if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) { 2616 spin_unlock_irqrestore(&t->task_state_lock, flags); 2617 pm8001_dbg(pm8001_ha, FAIL, 2618 "task 0x%p done with io_status 0x%x resp 0x%x stat 0x%x but aborted by upper layer!\n", 2619 t, status, ts->resp, ts->stat); 2620 pm8001_ccb_task_free(pm8001_ha, t, ccb, tag); 2621 } else { 2622 spin_unlock_irqrestore(&t->task_state_lock, flags); 2623 pm8001_ccb_task_free_done(pm8001_ha, t, ccb, tag); 2624 } 2625 } 2626 2627 /*See the comments for mpi_ssp_completion */ 2628 static void mpi_sata_event(struct pm8001_hba_info *pm8001_ha , void *piomb) 2629 { 2630 struct sas_task *t; 2631 struct task_status_struct *ts; 2632 struct pm8001_ccb_info *ccb; 2633 struct pm8001_device *pm8001_dev; 2634 struct sata_event_resp *psataPayload = 2635 (struct sata_event_resp *)(piomb + 4); 2636 u32 event = le32_to_cpu(psataPayload->event); 2637 u32 tag = le32_to_cpu(psataPayload->tag); 2638 u32 port_id = le32_to_cpu(psataPayload->port_id); 2639 u32 dev_id = le32_to_cpu(psataPayload->device_id); 2640 unsigned long flags; 2641 2642 ccb = &pm8001_ha->ccb_info[tag]; 2643 2644 if (ccb) { 2645 t = ccb->task; 2646 pm8001_dev = ccb->device; 2647 } else { 2648 pm8001_dbg(pm8001_ha, FAIL, "No CCB !!!. returning\n"); 2649 } 2650 if (event) 2651 pm8001_dbg(pm8001_ha, FAIL, "SATA EVENT 0x%x\n", event); 2652 2653 /* Check if this is NCQ error */ 2654 if (event == IO_XFER_ERROR_ABORTED_NCQ_MODE) { 2655 /* find device using device id */ 2656 pm8001_dev = pm8001_find_dev(pm8001_ha, dev_id); 2657 /* send read log extension */ 2658 if (pm8001_dev) 2659 pm8001_send_read_log(pm8001_ha, pm8001_dev); 2660 return; 2661 } 2662 2663 ccb = &pm8001_ha->ccb_info[tag]; 2664 t = ccb->task; 2665 pm8001_dev = ccb->device; 2666 if (event) 2667 pm8001_dbg(pm8001_ha, FAIL, "sata IO status 0x%x\n", event); 2668 if (unlikely(!t || !t->lldd_task || !t->dev)) 2669 return; 2670 ts = &t->task_status; 2671 pm8001_dbg(pm8001_ha, DEVIO, 2672 "port_id:0x%x, device_id:0x%x, tag:0x%x, event:0x%x\n", 2673 port_id, dev_id, tag, event); 2674 switch (event) { 2675 case IO_OVERFLOW: 2676 pm8001_dbg(pm8001_ha, IO, "IO_UNDERFLOW\n"); 2677 ts->resp = SAS_TASK_COMPLETE; 2678 ts->stat = SAS_DATA_OVERRUN; 2679 ts->residual = 0; 2680 if (pm8001_dev) 2681 atomic_dec(&pm8001_dev->running_req); 2682 break; 2683 case IO_XFER_ERROR_BREAK: 2684 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_BREAK\n"); 2685 ts->resp = SAS_TASK_COMPLETE; 2686 ts->stat = SAS_INTERRUPTED; 2687 break; 2688 case IO_XFER_ERROR_PHY_NOT_READY: 2689 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_PHY_NOT_READY\n"); 2690 ts->resp = SAS_TASK_COMPLETE; 2691 ts->stat = SAS_OPEN_REJECT; 2692 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 2693 break; 2694 case IO_OPEN_CNX_ERROR_PROTOCOL_NOT_SUPPORTED: 2695 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_PROTOCOL_NOT_SUPPORTED\n"); 2696 ts->resp = SAS_TASK_COMPLETE; 2697 ts->stat = SAS_OPEN_REJECT; 2698 ts->open_rej_reason = SAS_OREJ_EPROTO; 2699 break; 2700 case IO_OPEN_CNX_ERROR_ZONE_VIOLATION: 2701 pm8001_dbg(pm8001_ha, IO, 2702 "IO_OPEN_CNX_ERROR_ZONE_VIOLATION\n"); 2703 ts->resp = SAS_TASK_COMPLETE; 2704 ts->stat = SAS_OPEN_REJECT; 2705 ts->open_rej_reason = SAS_OREJ_UNKNOWN; 2706 break; 2707 case IO_OPEN_CNX_ERROR_BREAK: 2708 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_BREAK\n"); 2709 ts->resp = SAS_TASK_COMPLETE; 2710 ts->stat = SAS_OPEN_REJECT; 2711 ts->open_rej_reason = SAS_OREJ_RSVD_CONT0; 2712 break; 2713 case IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS: 2714 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS\n"); 2715 ts->resp = SAS_TASK_UNDELIVERED; 2716 ts->stat = SAS_DEV_NO_RESPONSE; 2717 if (!t->uldd_task) { 2718 pm8001_handle_event(pm8001_ha, 2719 pm8001_dev, 2720 IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS); 2721 ts->resp = SAS_TASK_COMPLETE; 2722 ts->stat = SAS_QUEUE_FULL; 2723 pm8001_ccb_task_free_done(pm8001_ha, t, ccb, tag); 2724 return; 2725 } 2726 break; 2727 case IO_OPEN_CNX_ERROR_BAD_DESTINATION: 2728 pm8001_dbg(pm8001_ha, IO, 2729 "IO_OPEN_CNX_ERROR_BAD_DESTINATION\n"); 2730 ts->resp = SAS_TASK_UNDELIVERED; 2731 ts->stat = SAS_OPEN_REJECT; 2732 ts->open_rej_reason = SAS_OREJ_BAD_DEST; 2733 break; 2734 case IO_OPEN_CNX_ERROR_CONNECTION_RATE_NOT_SUPPORTED: 2735 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_CONNECTION_RATE_NOT_SUPPORTED\n"); 2736 ts->resp = SAS_TASK_COMPLETE; 2737 ts->stat = SAS_OPEN_REJECT; 2738 ts->open_rej_reason = SAS_OREJ_CONN_RATE; 2739 break; 2740 case IO_OPEN_CNX_ERROR_WRONG_DESTINATION: 2741 pm8001_dbg(pm8001_ha, IO, 2742 "IO_OPEN_CNX_ERROR_WRONG_DESTINATION\n"); 2743 ts->resp = SAS_TASK_COMPLETE; 2744 ts->stat = SAS_OPEN_REJECT; 2745 ts->open_rej_reason = SAS_OREJ_WRONG_DEST; 2746 break; 2747 case IO_XFER_ERROR_NAK_RECEIVED: 2748 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_NAK_RECEIVED\n"); 2749 ts->resp = SAS_TASK_COMPLETE; 2750 ts->stat = SAS_NAK_R_ERR; 2751 break; 2752 case IO_XFER_ERROR_PEER_ABORTED: 2753 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_PEER_ABORTED\n"); 2754 ts->resp = SAS_TASK_COMPLETE; 2755 ts->stat = SAS_NAK_R_ERR; 2756 break; 2757 case IO_XFER_ERROR_REJECTED_NCQ_MODE: 2758 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_REJECTED_NCQ_MODE\n"); 2759 ts->resp = SAS_TASK_COMPLETE; 2760 ts->stat = SAS_DATA_UNDERRUN; 2761 break; 2762 case IO_XFER_OPEN_RETRY_TIMEOUT: 2763 pm8001_dbg(pm8001_ha, IO, "IO_XFER_OPEN_RETRY_TIMEOUT\n"); 2764 ts->resp = SAS_TASK_COMPLETE; 2765 ts->stat = SAS_OPEN_TO; 2766 break; 2767 case IO_XFER_ERROR_UNEXPECTED_PHASE: 2768 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_UNEXPECTED_PHASE\n"); 2769 ts->resp = SAS_TASK_COMPLETE; 2770 ts->stat = SAS_OPEN_TO; 2771 break; 2772 case IO_XFER_ERROR_XFER_RDY_OVERRUN: 2773 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_XFER_RDY_OVERRUN\n"); 2774 ts->resp = SAS_TASK_COMPLETE; 2775 ts->stat = SAS_OPEN_TO; 2776 break; 2777 case IO_XFER_ERROR_XFER_RDY_NOT_EXPECTED: 2778 pm8001_dbg(pm8001_ha, IO, 2779 "IO_XFER_ERROR_XFER_RDY_NOT_EXPECTED\n"); 2780 ts->resp = SAS_TASK_COMPLETE; 2781 ts->stat = SAS_OPEN_TO; 2782 break; 2783 case IO_XFER_ERROR_OFFSET_MISMATCH: 2784 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_OFFSET_MISMATCH\n"); 2785 ts->resp = SAS_TASK_COMPLETE; 2786 ts->stat = SAS_OPEN_TO; 2787 break; 2788 case IO_XFER_ERROR_XFER_ZERO_DATA_LEN: 2789 pm8001_dbg(pm8001_ha, IO, 2790 "IO_XFER_ERROR_XFER_ZERO_DATA_LEN\n"); 2791 ts->resp = SAS_TASK_COMPLETE; 2792 ts->stat = SAS_OPEN_TO; 2793 break; 2794 case IO_XFER_CMD_FRAME_ISSUED: 2795 pm8001_dbg(pm8001_ha, IO, "IO_XFER_CMD_FRAME_ISSUED\n"); 2796 break; 2797 case IO_XFER_PIO_SETUP_ERROR: 2798 pm8001_dbg(pm8001_ha, IO, "IO_XFER_PIO_SETUP_ERROR\n"); 2799 ts->resp = SAS_TASK_COMPLETE; 2800 ts->stat = SAS_OPEN_TO; 2801 break; 2802 default: 2803 pm8001_dbg(pm8001_ha, DEVIO, "Unknown status 0x%x\n", event); 2804 /* not allowed case. Therefore, return failed status */ 2805 ts->resp = SAS_TASK_COMPLETE; 2806 ts->stat = SAS_OPEN_TO; 2807 break; 2808 } 2809 spin_lock_irqsave(&t->task_state_lock, flags); 2810 t->task_state_flags &= ~SAS_TASK_STATE_PENDING; 2811 t->task_state_flags &= ~SAS_TASK_AT_INITIATOR; 2812 t->task_state_flags |= SAS_TASK_STATE_DONE; 2813 if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) { 2814 spin_unlock_irqrestore(&t->task_state_lock, flags); 2815 pm8001_dbg(pm8001_ha, FAIL, 2816 "task 0x%p done with io_status 0x%x resp 0x%x stat 0x%x but aborted by upper layer!\n", 2817 t, event, ts->resp, ts->stat); 2818 pm8001_ccb_task_free(pm8001_ha, t, ccb, tag); 2819 } else { 2820 spin_unlock_irqrestore(&t->task_state_lock, flags); 2821 pm8001_ccb_task_free_done(pm8001_ha, t, ccb, tag); 2822 } 2823 } 2824 2825 /*See the comments for mpi_ssp_completion */ 2826 static void 2827 mpi_smp_completion(struct pm8001_hba_info *pm8001_ha, void *piomb) 2828 { 2829 struct sas_task *t; 2830 struct pm8001_ccb_info *ccb; 2831 unsigned long flags; 2832 u32 status; 2833 u32 tag; 2834 struct smp_completion_resp *psmpPayload; 2835 struct task_status_struct *ts; 2836 struct pm8001_device *pm8001_dev; 2837 2838 psmpPayload = (struct smp_completion_resp *)(piomb + 4); 2839 status = le32_to_cpu(psmpPayload->status); 2840 tag = le32_to_cpu(psmpPayload->tag); 2841 2842 ccb = &pm8001_ha->ccb_info[tag]; 2843 t = ccb->task; 2844 ts = &t->task_status; 2845 pm8001_dev = ccb->device; 2846 if (status) { 2847 pm8001_dbg(pm8001_ha, FAIL, "smp IO status 0x%x\n", status); 2848 pm8001_dbg(pm8001_ha, IOERR, 2849 "status:0x%x, tag:0x%x, task:0x%p\n", 2850 status, tag, t); 2851 } 2852 if (unlikely(!t || !t->lldd_task || !t->dev)) 2853 return; 2854 2855 switch (status) { 2856 case IO_SUCCESS: 2857 pm8001_dbg(pm8001_ha, IO, "IO_SUCCESS\n"); 2858 ts->resp = SAS_TASK_COMPLETE; 2859 ts->stat = SAM_STAT_GOOD; 2860 if (pm8001_dev) 2861 atomic_dec(&pm8001_dev->running_req); 2862 break; 2863 case IO_ABORTED: 2864 pm8001_dbg(pm8001_ha, IO, "IO_ABORTED IOMB\n"); 2865 ts->resp = SAS_TASK_COMPLETE; 2866 ts->stat = SAS_ABORTED_TASK; 2867 if (pm8001_dev) 2868 atomic_dec(&pm8001_dev->running_req); 2869 break; 2870 case IO_OVERFLOW: 2871 pm8001_dbg(pm8001_ha, IO, "IO_UNDERFLOW\n"); 2872 ts->resp = SAS_TASK_COMPLETE; 2873 ts->stat = SAS_DATA_OVERRUN; 2874 ts->residual = 0; 2875 if (pm8001_dev) 2876 atomic_dec(&pm8001_dev->running_req); 2877 break; 2878 case IO_NO_DEVICE: 2879 pm8001_dbg(pm8001_ha, IO, "IO_NO_DEVICE\n"); 2880 ts->resp = SAS_TASK_COMPLETE; 2881 ts->stat = SAS_PHY_DOWN; 2882 break; 2883 case IO_ERROR_HW_TIMEOUT: 2884 pm8001_dbg(pm8001_ha, IO, "IO_ERROR_HW_TIMEOUT\n"); 2885 ts->resp = SAS_TASK_COMPLETE; 2886 ts->stat = SAM_STAT_BUSY; 2887 break; 2888 case IO_XFER_ERROR_BREAK: 2889 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_BREAK\n"); 2890 ts->resp = SAS_TASK_COMPLETE; 2891 ts->stat = SAM_STAT_BUSY; 2892 break; 2893 case IO_XFER_ERROR_PHY_NOT_READY: 2894 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_PHY_NOT_READY\n"); 2895 ts->resp = SAS_TASK_COMPLETE; 2896 ts->stat = SAM_STAT_BUSY; 2897 break; 2898 case IO_OPEN_CNX_ERROR_PROTOCOL_NOT_SUPPORTED: 2899 pm8001_dbg(pm8001_ha, IO, 2900 "IO_OPEN_CNX_ERROR_PROTOCOL_NOT_SUPPORTED\n"); 2901 ts->resp = SAS_TASK_COMPLETE; 2902 ts->stat = SAS_OPEN_REJECT; 2903 ts->open_rej_reason = SAS_OREJ_UNKNOWN; 2904 break; 2905 case IO_OPEN_CNX_ERROR_ZONE_VIOLATION: 2906 pm8001_dbg(pm8001_ha, IO, 2907 "IO_OPEN_CNX_ERROR_ZONE_VIOLATION\n"); 2908 ts->resp = SAS_TASK_COMPLETE; 2909 ts->stat = SAS_OPEN_REJECT; 2910 ts->open_rej_reason = SAS_OREJ_UNKNOWN; 2911 break; 2912 case IO_OPEN_CNX_ERROR_BREAK: 2913 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_BREAK\n"); 2914 ts->resp = SAS_TASK_COMPLETE; 2915 ts->stat = SAS_OPEN_REJECT; 2916 ts->open_rej_reason = SAS_OREJ_RSVD_CONT0; 2917 break; 2918 case IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS: 2919 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS\n"); 2920 ts->resp = SAS_TASK_COMPLETE; 2921 ts->stat = SAS_OPEN_REJECT; 2922 ts->open_rej_reason = SAS_OREJ_UNKNOWN; 2923 pm8001_handle_event(pm8001_ha, 2924 pm8001_dev, 2925 IO_OPEN_CNX_ERROR_IT_NEXUS_LOSS); 2926 break; 2927 case IO_OPEN_CNX_ERROR_BAD_DESTINATION: 2928 pm8001_dbg(pm8001_ha, IO, 2929 "IO_OPEN_CNX_ERROR_BAD_DESTINATION\n"); 2930 ts->resp = SAS_TASK_COMPLETE; 2931 ts->stat = SAS_OPEN_REJECT; 2932 ts->open_rej_reason = SAS_OREJ_BAD_DEST; 2933 break; 2934 case IO_OPEN_CNX_ERROR_CONNECTION_RATE_NOT_SUPPORTED: 2935 pm8001_dbg(pm8001_ha, IO, "IO_OPEN_CNX_ERROR_CONNECTION_RATE_NOT_SUPPORTED\n"); 2936 ts->resp = SAS_TASK_COMPLETE; 2937 ts->stat = SAS_OPEN_REJECT; 2938 ts->open_rej_reason = SAS_OREJ_CONN_RATE; 2939 break; 2940 case IO_OPEN_CNX_ERROR_WRONG_DESTINATION: 2941 pm8001_dbg(pm8001_ha, IO, 2942 "IO_OPEN_CNX_ERROR_WRONG_DESTINATION\n"); 2943 ts->resp = SAS_TASK_COMPLETE; 2944 ts->stat = SAS_OPEN_REJECT; 2945 ts->open_rej_reason = SAS_OREJ_WRONG_DEST; 2946 break; 2947 case IO_XFER_ERROR_RX_FRAME: 2948 pm8001_dbg(pm8001_ha, IO, "IO_XFER_ERROR_RX_FRAME\n"); 2949 ts->resp = SAS_TASK_COMPLETE; 2950 ts->stat = SAS_DEV_NO_RESPONSE; 2951 break; 2952 case IO_XFER_OPEN_RETRY_TIMEOUT: 2953 pm8001_dbg(pm8001_ha, IO, "IO_XFER_OPEN_RETRY_TIMEOUT\n"); 2954 ts->resp = SAS_TASK_COMPLETE; 2955 ts->stat = SAS_OPEN_REJECT; 2956 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 2957 break; 2958 case IO_ERROR_INTERNAL_SMP_RESOURCE: 2959 pm8001_dbg(pm8001_ha, IO, "IO_ERROR_INTERNAL_SMP_RESOURCE\n"); 2960 ts->resp = SAS_TASK_COMPLETE; 2961 ts->stat = SAS_QUEUE_FULL; 2962 break; 2963 case IO_PORT_IN_RESET: 2964 pm8001_dbg(pm8001_ha, IO, "IO_PORT_IN_RESET\n"); 2965 ts->resp = SAS_TASK_COMPLETE; 2966 ts->stat = SAS_OPEN_REJECT; 2967 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 2968 break; 2969 case IO_DS_NON_OPERATIONAL: 2970 pm8001_dbg(pm8001_ha, IO, "IO_DS_NON_OPERATIONAL\n"); 2971 ts->resp = SAS_TASK_COMPLETE; 2972 ts->stat = SAS_DEV_NO_RESPONSE; 2973 break; 2974 case IO_DS_IN_RECOVERY: 2975 pm8001_dbg(pm8001_ha, IO, "IO_DS_IN_RECOVERY\n"); 2976 ts->resp = SAS_TASK_COMPLETE; 2977 ts->stat = SAS_OPEN_REJECT; 2978 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 2979 break; 2980 case IO_OPEN_CNX_ERROR_HW_RESOURCE_BUSY: 2981 pm8001_dbg(pm8001_ha, IO, 2982 "IO_OPEN_CNX_ERROR_HW_RESOURCE_BUSY\n"); 2983 ts->resp = SAS_TASK_COMPLETE; 2984 ts->stat = SAS_OPEN_REJECT; 2985 ts->open_rej_reason = SAS_OREJ_RSVD_RETRY; 2986 break; 2987 default: 2988 pm8001_dbg(pm8001_ha, DEVIO, "Unknown status 0x%x\n", status); 2989 ts->resp = SAS_TASK_COMPLETE; 2990 ts->stat = SAS_DEV_NO_RESPONSE; 2991 /* not allowed case. Therefore, return failed status */ 2992 break; 2993 } 2994 spin_lock_irqsave(&t->task_state_lock, flags); 2995 t->task_state_flags &= ~SAS_TASK_STATE_PENDING; 2996 t->task_state_flags &= ~SAS_TASK_AT_INITIATOR; 2997 t->task_state_flags |= SAS_TASK_STATE_DONE; 2998 if (unlikely((t->task_state_flags & SAS_TASK_STATE_ABORTED))) { 2999 spin_unlock_irqrestore(&t->task_state_lock, flags); 3000 pm8001_dbg(pm8001_ha, FAIL, "task 0x%p done with io_status 0x%x resp 0x%x stat 0x%x but aborted by upper layer!\n", 3001 t, status, ts->resp, ts->stat); 3002 pm8001_ccb_task_free(pm8001_ha, t, ccb, tag); 3003 } else { 3004 spin_unlock_irqrestore(&t->task_state_lock, flags); 3005 pm8001_ccb_task_free(pm8001_ha, t, ccb, tag); 3006 mb();/* in order to force CPU ordering */ 3007 t->task_done(t); 3008 } 3009 } 3010 3011 void pm8001_mpi_set_dev_state_resp(struct pm8001_hba_info *pm8001_ha, 3012 void *piomb) 3013 { 3014 struct set_dev_state_resp *pPayload = 3015 (struct set_dev_state_resp *)(piomb + 4); 3016 u32 tag = le32_to_cpu(pPayload->tag); 3017 struct pm8001_ccb_info *ccb = &pm8001_ha->ccb_info[tag]; 3018 struct pm8001_device *pm8001_dev = ccb->device; 3019 u32 status = le32_to_cpu(pPayload->status); 3020 u32 device_id = le32_to_cpu(pPayload->device_id); 3021 u8 pds = le32_to_cpu(pPayload->pds_nds) & PDS_BITS; 3022 u8 nds = le32_to_cpu(pPayload->pds_nds) & NDS_BITS; 3023 pm8001_dbg(pm8001_ha, MSG, "Set device id = 0x%x state from 0x%x to 0x%x status = 0x%x!\n", 3024 device_id, pds, nds, status); 3025 complete(pm8001_dev->setds_completion); 3026 ccb->task = NULL; 3027 ccb->ccb_tag = 0xFFFFFFFF; 3028 pm8001_tag_free(pm8001_ha, tag); 3029 } 3030 3031 void pm8001_mpi_set_nvmd_resp(struct pm8001_hba_info *pm8001_ha, void *piomb) 3032 { 3033 struct get_nvm_data_resp *pPayload = 3034 (struct get_nvm_data_resp *)(piomb + 4); 3035 u32 tag = le32_to_cpu(pPayload->tag); 3036 struct pm8001_ccb_info *ccb = &pm8001_ha->ccb_info[tag]; 3037 u32 dlen_status = le32_to_cpu(pPayload->dlen_status); 3038 complete(pm8001_ha->nvmd_completion); 3039 pm8001_dbg(pm8001_ha, MSG, "Set nvm data complete!\n"); 3040 if ((dlen_status & NVMD_STAT) != 0) { 3041 pm8001_dbg(pm8001_ha, FAIL, "Set nvm data error %x\n", 3042 dlen_status); 3043 } 3044 ccb->task = NULL; 3045 ccb->ccb_tag = 0xFFFFFFFF; 3046 pm8001_tag_free(pm8001_ha, tag); 3047 } 3048 3049 void 3050 pm8001_mpi_get_nvmd_resp(struct pm8001_hba_info *pm8001_ha, void *piomb) 3051 { 3052 struct fw_control_ex *fw_control_context; 3053 struct get_nvm_data_resp *pPayload = 3054 (struct get_nvm_data_resp *)(piomb + 4); 3055 u32 tag = le32_to_cpu(pPayload->tag); 3056 struct pm8001_ccb_info *ccb = &pm8001_ha->ccb_info[tag]; 3057 u32 dlen_status = le32_to_cpu(pPayload->dlen_status); 3058 u32 ir_tds_bn_dps_das_nvm = 3059 le32_to_cpu(pPayload->ir_tda_bn_dps_das_nvm); 3060 void *virt_addr = pm8001_ha->memoryMap.region[NVMD].virt_ptr; 3061 fw_control_context = ccb->fw_control_context; 3062 3063 pm8001_dbg(pm8001_ha, MSG, "Get nvm data complete!\n"); 3064 if ((dlen_status & NVMD_STAT) != 0) { 3065 pm8001_dbg(pm8001_ha, FAIL, "Get nvm data error %x\n", 3066 dlen_status); 3067 complete(pm8001_ha->nvmd_completion); 3068 /* We should free tag during failure also, the tag is not being 3069 * freed by requesting path anywhere. 3070 */ 3071 ccb->task = NULL; 3072 ccb->ccb_tag = 0xFFFFFFFF; 3073 pm8001_tag_free(pm8001_ha, tag); 3074 return; 3075 } 3076 if (ir_tds_bn_dps_das_nvm & IPMode) { 3077 /* indirect mode - IR bit set */ 3078 pm8001_dbg(pm8001_ha, MSG, "Get NVMD success, IR=1\n"); 3079 if ((ir_tds_bn_dps_das_nvm & NVMD_TYPE) == TWI_DEVICE) { 3080 if (ir_tds_bn_dps_das_nvm == 0x80a80200) { 3081 memcpy(pm8001_ha->sas_addr, 3082 ((u8 *)virt_addr + 4), 3083 SAS_ADDR_SIZE); 3084 pm8001_dbg(pm8001_ha, MSG, "Get SAS address from VPD successfully!\n"); 3085 } 3086 } else if (((ir_tds_bn_dps_das_nvm & NVMD_TYPE) == C_SEEPROM) 3087 || ((ir_tds_bn_dps_das_nvm & NVMD_TYPE) == VPD_FLASH) || 3088 ((ir_tds_bn_dps_das_nvm & NVMD_TYPE) == EXPAN_ROM)) { 3089 ; 3090 } else if (((ir_tds_bn_dps_das_nvm & NVMD_TYPE) == AAP1_RDUMP) 3091 || ((ir_tds_bn_dps_das_nvm & NVMD_TYPE) == IOP_RDUMP)) { 3092 ; 3093 } else { 3094 /* Should not be happened*/ 3095 pm8001_dbg(pm8001_ha, MSG, 3096 "(IR=1)Wrong Device type 0x%x\n", 3097 ir_tds_bn_dps_das_nvm); 3098 } 3099 } else /* direct mode */{ 3100 pm8001_dbg(pm8001_ha, MSG, 3101 "Get NVMD success, IR=0, dataLen=%d\n", 3102 (dlen_status & NVMD_LEN) >> 24); 3103 } 3104 /* Though fw_control_context is freed below, usrAddr still needs 3105 * to be updated as this holds the response to the request function 3106 */ 3107 memcpy(fw_control_context->usrAddr, 3108 pm8001_ha->memoryMap.region[NVMD].virt_ptr, 3109 fw_control_context->len); 3110 kfree(ccb->fw_control_context); 3111 /* To avoid race condition, complete should be 3112 * called after the message is copied to 3113 * fw_control_context->usrAddr 3114 */ 3115 complete(pm8001_ha->nvmd_completion); 3116 pm8001_dbg(pm8001_ha, MSG, "Set nvm data complete!\n"); 3117 ccb->task = NULL; 3118 ccb->ccb_tag = 0xFFFFFFFF; 3119 pm8001_tag_free(pm8001_ha, tag); 3120 } 3121 3122 int pm8001_mpi_local_phy_ctl(struct pm8001_hba_info *pm8001_ha, void *piomb) 3123 { 3124 u32 tag; 3125 struct local_phy_ctl_resp *pPayload = 3126 (struct local_phy_ctl_resp *)(piomb + 4); 3127 u32 status = le32_to_cpu(pPayload->status); 3128 u32 phy_id = le32_to_cpu(pPayload->phyop_phyid) & ID_BITS; 3129 u32 phy_op = le32_to_cpu(pPayload->phyop_phyid) & OP_BITS; 3130 tag = le32_to_cpu(pPayload->tag); 3131 if (status != 0) { 3132 pm8001_dbg(pm8001_ha, MSG, 3133 "%x phy execute %x phy op failed!\n", 3134 phy_id, phy_op); 3135 } else { 3136 pm8001_dbg(pm8001_ha, MSG, 3137 "%x phy execute %x phy op success!\n", 3138 phy_id, phy_op); 3139 pm8001_ha->phy[phy_id].reset_success = true; 3140 } 3141 if (pm8001_ha->phy[phy_id].enable_completion) { 3142 complete(pm8001_ha->phy[phy_id].enable_completion); 3143 pm8001_ha->phy[phy_id].enable_completion = NULL; 3144 } 3145 pm8001_tag_free(pm8001_ha, tag); 3146 return 0; 3147 } 3148 3149 /** 3150 * pm8001_bytes_dmaed - one of the interface function communication with libsas 3151 * @pm8001_ha: our hba card information 3152 * @i: which phy that received the event. 3153 * 3154 * when HBA driver received the identify done event or initiate FIS received 3155 * event(for SATA), it will invoke this function to notify the sas layer that 3156 * the sas toplogy has formed, please discover the the whole sas domain, 3157 * while receive a broadcast(change) primitive just tell the sas 3158 * layer to discover the changed domain rather than the whole domain. 3159 */ 3160 void pm8001_bytes_dmaed(struct pm8001_hba_info *pm8001_ha, int i) 3161 { 3162 struct pm8001_phy *phy = &pm8001_ha->phy[i]; 3163 struct asd_sas_phy *sas_phy = &phy->sas_phy; 3164 if (!phy->phy_attached) 3165 return; 3166 3167 if (sas_phy->phy) { 3168 struct sas_phy *sphy = sas_phy->phy; 3169 sphy->negotiated_linkrate = sas_phy->linkrate; 3170 sphy->minimum_linkrate = phy->minimum_linkrate; 3171 sphy->minimum_linkrate_hw = SAS_LINK_RATE_1_5_GBPS; 3172 sphy->maximum_linkrate = phy->maximum_linkrate; 3173 sphy->maximum_linkrate_hw = phy->maximum_linkrate; 3174 } 3175 3176 if (phy->phy_type & PORT_TYPE_SAS) { 3177 struct sas_identify_frame *id; 3178 id = (struct sas_identify_frame *)phy->frame_rcvd; 3179 id->dev_type = phy->identify.device_type; 3180 id->initiator_bits = SAS_PROTOCOL_ALL; 3181 id->target_bits = phy->identify.target_port_protocols; 3182 } else if (phy->phy_type & PORT_TYPE_SATA) { 3183 /*Nothing*/ 3184 } 3185 pm8001_dbg(pm8001_ha, MSG, "phy %d byte dmaded.\n", i); 3186 3187 sas_phy->frame_rcvd_size = phy->frame_rcvd_size; 3188 sas_notify_port_event(sas_phy, PORTE_BYTES_DMAED, GFP_ATOMIC); 3189 } 3190 3191 /* Get the link rate speed */ 3192 void pm8001_get_lrate_mode(struct pm8001_phy *phy, u8 link_rate) 3193 { 3194 struct sas_phy *sas_phy = phy->sas_phy.phy; 3195 3196 switch (link_rate) { 3197 case PHY_SPEED_120: 3198 phy->sas_phy.linkrate = SAS_LINK_RATE_12_0_GBPS; 3199 phy->sas_phy.phy->negotiated_linkrate = SAS_LINK_RATE_12_0_GBPS; 3200 break; 3201 case PHY_SPEED_60: 3202 phy->sas_phy.linkrate = SAS_LINK_RATE_6_0_GBPS; 3203 phy->sas_phy.phy->negotiated_linkrate = SAS_LINK_RATE_6_0_GBPS; 3204 break; 3205 case PHY_SPEED_30: 3206 phy->sas_phy.linkrate = SAS_LINK_RATE_3_0_GBPS; 3207 phy->sas_phy.phy->negotiated_linkrate = SAS_LINK_RATE_3_0_GBPS; 3208 break; 3209 case PHY_SPEED_15: 3210 phy->sas_phy.linkrate = SAS_LINK_RATE_1_5_GBPS; 3211 phy->sas_phy.phy->negotiated_linkrate = SAS_LINK_RATE_1_5_GBPS; 3212 break; 3213 } 3214 sas_phy->negotiated_linkrate = phy->sas_phy.linkrate; 3215 sas_phy->maximum_linkrate_hw = SAS_LINK_RATE_6_0_GBPS; 3216 sas_phy->minimum_linkrate_hw = SAS_LINK_RATE_1_5_GBPS; 3217 sas_phy->maximum_linkrate = SAS_LINK_RATE_6_0_GBPS; 3218 sas_phy->minimum_linkrate = SAS_LINK_RATE_1_5_GBPS; 3219 } 3220 3221 /** 3222 * asd_get_attached_sas_addr -- extract/generate attached SAS address 3223 * @phy: pointer to asd_phy 3224 * @sas_addr: pointer to buffer where the SAS address is to be written 3225 * 3226 * This function extracts the SAS address from an IDENTIFY frame 3227 * received. If OOB is SATA, then a SAS address is generated from the 3228 * HA tables. 3229 * 3230 * LOCKING: the frame_rcvd_lock needs to be held since this parses the frame 3231 * buffer. 3232 */ 3233 void pm8001_get_attached_sas_addr(struct pm8001_phy *phy, 3234 u8 *sas_addr) 3235 { 3236 if (phy->sas_phy.frame_rcvd[0] == 0x34 3237 && phy->sas_phy.oob_mode == SATA_OOB_MODE) { 3238 struct pm8001_hba_info *pm8001_ha = phy->sas_phy.ha->lldd_ha; 3239 /* FIS device-to-host */ 3240 u64 addr = be64_to_cpu(*(__be64 *)pm8001_ha->sas_addr); 3241 addr += phy->sas_phy.id; 3242 *(__be64 *)sas_addr = cpu_to_be64(addr); 3243 } else { 3244 struct sas_identify_frame *idframe = 3245 (void *) phy->sas_phy.frame_rcvd; 3246 memcpy(sas_addr, idframe->sas_addr, SAS_ADDR_SIZE); 3247 } 3248 } 3249 3250 /** 3251 * pm8001_hw_event_ack_req- For PM8001,some events need to acknowage to FW. 3252 * @pm8001_ha: our hba card information 3253 * @Qnum: the outbound queue message number. 3254 * @SEA: source of event to ack 3255 * @port_id: port id. 3256 * @phyId: phy id. 3257 * @param0: parameter 0. 3258 * @param1: parameter 1. 3259 */ 3260 static void pm8001_hw_event_ack_req(struct pm8001_hba_info *pm8001_ha, 3261 u32 Qnum, u32 SEA, u32 port_id, u32 phyId, u32 param0, u32 param1) 3262 { 3263 struct hw_event_ack_req payload; 3264 u32 opc = OPC_INB_SAS_HW_EVENT_ACK; 3265 3266 struct inbound_queue_table *circularQ; 3267 3268 memset((u8 *)&payload, 0, sizeof(payload)); 3269 circularQ = &pm8001_ha->inbnd_q_tbl[Qnum]; 3270 payload.tag = cpu_to_le32(1); 3271 payload.sea_phyid_portid = cpu_to_le32(((SEA & 0xFFFF) << 8) | 3272 ((phyId & 0x0F) << 4) | (port_id & 0x0F)); 3273 payload.param0 = cpu_to_le32(param0); 3274 payload.param1 = cpu_to_le32(param1); 3275 pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &payload, 3276 sizeof(payload), 0); 3277 } 3278 3279 static int pm8001_chip_phy_ctl_req(struct pm8001_hba_info *pm8001_ha, 3280 u32 phyId, u32 phy_op); 3281 3282 /** 3283 * hw_event_sas_phy_up -FW tells me a SAS phy up event. 3284 * @pm8001_ha: our hba card information 3285 * @piomb: IO message buffer 3286 */ 3287 static void 3288 hw_event_sas_phy_up(struct pm8001_hba_info *pm8001_ha, void *piomb) 3289 { 3290 struct hw_event_resp *pPayload = 3291 (struct hw_event_resp *)(piomb + 4); 3292 u32 lr_evt_status_phyid_portid = 3293 le32_to_cpu(pPayload->lr_evt_status_phyid_portid); 3294 u8 link_rate = 3295 (u8)((lr_evt_status_phyid_portid & 0xF0000000) >> 28); 3296 u8 port_id = (u8)(lr_evt_status_phyid_portid & 0x0000000F); 3297 u8 phy_id = 3298 (u8)((lr_evt_status_phyid_portid & 0x000000F0) >> 4); 3299 u32 npip_portstate = le32_to_cpu(pPayload->npip_portstate); 3300 u8 portstate = (u8)(npip_portstate & 0x0000000F); 3301 struct pm8001_port *port = &pm8001_ha->port[port_id]; 3302 struct pm8001_phy *phy = &pm8001_ha->phy[phy_id]; 3303 unsigned long flags; 3304 u8 deviceType = pPayload->sas_identify.dev_type; 3305 port->port_state = portstate; 3306 phy->phy_state = PHY_STATE_LINK_UP_SPC; 3307 pm8001_dbg(pm8001_ha, MSG, 3308 "HW_EVENT_SAS_PHY_UP port id = %d, phy id = %d\n", 3309 port_id, phy_id); 3310 3311 switch (deviceType) { 3312 case SAS_PHY_UNUSED: 3313 pm8001_dbg(pm8001_ha, MSG, "device type no device.\n"); 3314 break; 3315 case SAS_END_DEVICE: 3316 pm8001_dbg(pm8001_ha, MSG, "end device.\n"); 3317 pm8001_chip_phy_ctl_req(pm8001_ha, phy_id, 3318 PHY_NOTIFY_ENABLE_SPINUP); 3319 port->port_attached = 1; 3320 pm8001_get_lrate_mode(phy, link_rate); 3321 break; 3322 case SAS_EDGE_EXPANDER_DEVICE: 3323 pm8001_dbg(pm8001_ha, MSG, "expander device.\n"); 3324 port->port_attached = 1; 3325 pm8001_get_lrate_mode(phy, link_rate); 3326 break; 3327 case SAS_FANOUT_EXPANDER_DEVICE: 3328 pm8001_dbg(pm8001_ha, MSG, "fanout expander device.\n"); 3329 port->port_attached = 1; 3330 pm8001_get_lrate_mode(phy, link_rate); 3331 break; 3332 default: 3333 pm8001_dbg(pm8001_ha, DEVIO, "unknown device type(%x)\n", 3334 deviceType); 3335 break; 3336 } 3337 phy->phy_type |= PORT_TYPE_SAS; 3338 phy->identify.device_type = deviceType; 3339 phy->phy_attached = 1; 3340 if (phy->identify.device_type == SAS_END_DEVICE) 3341 phy->identify.target_port_protocols = SAS_PROTOCOL_SSP; 3342 else if (phy->identify.device_type != SAS_PHY_UNUSED) 3343 phy->identify.target_port_protocols = SAS_PROTOCOL_SMP; 3344 phy->sas_phy.oob_mode = SAS_OOB_MODE; 3345 sas_notify_phy_event(&phy->sas_phy, PHYE_OOB_DONE, GFP_ATOMIC); 3346 spin_lock_irqsave(&phy->sas_phy.frame_rcvd_lock, flags); 3347 memcpy(phy->frame_rcvd, &pPayload->sas_identify, 3348 sizeof(struct sas_identify_frame)-4); 3349 phy->frame_rcvd_size = sizeof(struct sas_identify_frame) - 4; 3350 pm8001_get_attached_sas_addr(phy, phy->sas_phy.attached_sas_addr); 3351 spin_unlock_irqrestore(&phy->sas_phy.frame_rcvd_lock, flags); 3352 if (pm8001_ha->flags == PM8001F_RUN_TIME) 3353 mdelay(200);/*delay a moment to wait disk to spinup*/ 3354 pm8001_bytes_dmaed(pm8001_ha, phy_id); 3355 } 3356 3357 /** 3358 * hw_event_sata_phy_up -FW tells me a SATA phy up event. 3359 * @pm8001_ha: our hba card information 3360 * @piomb: IO message buffer 3361 */ 3362 static void 3363 hw_event_sata_phy_up(struct pm8001_hba_info *pm8001_ha, void *piomb) 3364 { 3365 struct hw_event_resp *pPayload = 3366 (struct hw_event_resp *)(piomb + 4); 3367 u32 lr_evt_status_phyid_portid = 3368 le32_to_cpu(pPayload->lr_evt_status_phyid_portid); 3369 u8 link_rate = 3370 (u8)((lr_evt_status_phyid_portid & 0xF0000000) >> 28); 3371 u8 port_id = (u8)(lr_evt_status_phyid_portid & 0x0000000F); 3372 u8 phy_id = 3373 (u8)((lr_evt_status_phyid_portid & 0x000000F0) >> 4); 3374 u32 npip_portstate = le32_to_cpu(pPayload->npip_portstate); 3375 u8 portstate = (u8)(npip_portstate & 0x0000000F); 3376 struct pm8001_port *port = &pm8001_ha->port[port_id]; 3377 struct pm8001_phy *phy = &pm8001_ha->phy[phy_id]; 3378 unsigned long flags; 3379 pm8001_dbg(pm8001_ha, DEVIO, "HW_EVENT_SATA_PHY_UP port id = %d, phy id = %d\n", 3380 port_id, phy_id); 3381 port->port_state = portstate; 3382 phy->phy_state = PHY_STATE_LINK_UP_SPC; 3383 port->port_attached = 1; 3384 pm8001_get_lrate_mode(phy, link_rate); 3385 phy->phy_type |= PORT_TYPE_SATA; 3386 phy->phy_attached = 1; 3387 phy->sas_phy.oob_mode = SATA_OOB_MODE; 3388 sas_notify_phy_event(&phy->sas_phy, PHYE_OOB_DONE, GFP_ATOMIC); 3389 spin_lock_irqsave(&phy->sas_phy.frame_rcvd_lock, flags); 3390 memcpy(phy->frame_rcvd, ((u8 *)&pPayload->sata_fis - 4), 3391 sizeof(struct dev_to_host_fis)); 3392 phy->frame_rcvd_size = sizeof(struct dev_to_host_fis); 3393 phy->identify.target_port_protocols = SAS_PROTOCOL_SATA; 3394 phy->identify.device_type = SAS_SATA_DEV; 3395 pm8001_get_attached_sas_addr(phy, phy->sas_phy.attached_sas_addr); 3396 spin_unlock_irqrestore(&phy->sas_phy.frame_rcvd_lock, flags); 3397 pm8001_bytes_dmaed(pm8001_ha, phy_id); 3398 } 3399 3400 /** 3401 * hw_event_phy_down -we should notify the libsas the phy is down. 3402 * @pm8001_ha: our hba card information 3403 * @piomb: IO message buffer 3404 */ 3405 static void 3406 hw_event_phy_down(struct pm8001_hba_info *pm8001_ha, void *piomb) 3407 { 3408 struct hw_event_resp *pPayload = 3409 (struct hw_event_resp *)(piomb + 4); 3410 u32 lr_evt_status_phyid_portid = 3411 le32_to_cpu(pPayload->lr_evt_status_phyid_portid); 3412 u8 port_id = (u8)(lr_evt_status_phyid_portid & 0x0000000F); 3413 u8 phy_id = 3414 (u8)((lr_evt_status_phyid_portid & 0x000000F0) >> 4); 3415 u32 npip_portstate = le32_to_cpu(pPayload->npip_portstate); 3416 u8 portstate = (u8)(npip_portstate & 0x0000000F); 3417 struct pm8001_port *port = &pm8001_ha->port[port_id]; 3418 struct pm8001_phy *phy = &pm8001_ha->phy[phy_id]; 3419 port->port_state = portstate; 3420 phy->phy_type = 0; 3421 phy->identify.device_type = 0; 3422 phy->phy_attached = 0; 3423 memset(&phy->dev_sas_addr, 0, SAS_ADDR_SIZE); 3424 switch (portstate) { 3425 case PORT_VALID: 3426 break; 3427 case PORT_INVALID: 3428 pm8001_dbg(pm8001_ha, MSG, " PortInvalid portID %d\n", 3429 port_id); 3430 pm8001_dbg(pm8001_ha, MSG, 3431 " Last phy Down and port invalid\n"); 3432 port->port_attached = 0; 3433 pm8001_hw_event_ack_req(pm8001_ha, 0, HW_EVENT_PHY_DOWN, 3434 port_id, phy_id, 0, 0); 3435 break; 3436 case PORT_IN_RESET: 3437 pm8001_dbg(pm8001_ha, MSG, " Port In Reset portID %d\n", 3438 port_id); 3439 break; 3440 case PORT_NOT_ESTABLISHED: 3441 pm8001_dbg(pm8001_ha, MSG, 3442 " phy Down and PORT_NOT_ESTABLISHED\n"); 3443 port->port_attached = 0; 3444 break; 3445 case PORT_LOSTCOMM: 3446 pm8001_dbg(pm8001_ha, MSG, " phy Down and PORT_LOSTCOMM\n"); 3447 pm8001_dbg(pm8001_ha, MSG, 3448 " Last phy Down and port invalid\n"); 3449 port->port_attached = 0; 3450 pm8001_hw_event_ack_req(pm8001_ha, 0, HW_EVENT_PHY_DOWN, 3451 port_id, phy_id, 0, 0); 3452 break; 3453 default: 3454 port->port_attached = 0; 3455 pm8001_dbg(pm8001_ha, DEVIO, " phy Down and(default) = %x\n", 3456 portstate); 3457 break; 3458 3459 } 3460 } 3461 3462 /** 3463 * pm8001_mpi_reg_resp -process register device ID response. 3464 * @pm8001_ha: our hba card information 3465 * @piomb: IO message buffer 3466 * 3467 * when sas layer find a device it will notify LLDD, then the driver register 3468 * the domain device to FW, this event is the return device ID which the FW 3469 * has assigned, from now,inter-communication with FW is no longer using the 3470 * SAS address, use device ID which FW assigned. 3471 */ 3472 int pm8001_mpi_reg_resp(struct pm8001_hba_info *pm8001_ha, void *piomb) 3473 { 3474 u32 status; 3475 u32 device_id; 3476 u32 htag; 3477 struct pm8001_ccb_info *ccb; 3478 struct pm8001_device *pm8001_dev; 3479 struct dev_reg_resp *registerRespPayload = 3480 (struct dev_reg_resp *)(piomb + 4); 3481 3482 htag = le32_to_cpu(registerRespPayload->tag); 3483 ccb = &pm8001_ha->ccb_info[htag]; 3484 pm8001_dev = ccb->device; 3485 status = le32_to_cpu(registerRespPayload->status); 3486 device_id = le32_to_cpu(registerRespPayload->device_id); 3487 pm8001_dbg(pm8001_ha, MSG, " register device is status = %d\n", 3488 status); 3489 switch (status) { 3490 case DEVREG_SUCCESS: 3491 pm8001_dbg(pm8001_ha, MSG, "DEVREG_SUCCESS\n"); 3492 pm8001_dev->device_id = device_id; 3493 break; 3494 case DEVREG_FAILURE_OUT_OF_RESOURCE: 3495 pm8001_dbg(pm8001_ha, MSG, "DEVREG_FAILURE_OUT_OF_RESOURCE\n"); 3496 break; 3497 case DEVREG_FAILURE_DEVICE_ALREADY_REGISTERED: 3498 pm8001_dbg(pm8001_ha, MSG, 3499 "DEVREG_FAILURE_DEVICE_ALREADY_REGISTERED\n"); 3500 break; 3501 case DEVREG_FAILURE_INVALID_PHY_ID: 3502 pm8001_dbg(pm8001_ha, MSG, "DEVREG_FAILURE_INVALID_PHY_ID\n"); 3503 break; 3504 case DEVREG_FAILURE_PHY_ID_ALREADY_REGISTERED: 3505 pm8001_dbg(pm8001_ha, MSG, 3506 "DEVREG_FAILURE_PHY_ID_ALREADY_REGISTERED\n"); 3507 break; 3508 case DEVREG_FAILURE_PORT_ID_OUT_OF_RANGE: 3509 pm8001_dbg(pm8001_ha, MSG, 3510 "DEVREG_FAILURE_PORT_ID_OUT_OF_RANGE\n"); 3511 break; 3512 case DEVREG_FAILURE_PORT_NOT_VALID_STATE: 3513 pm8001_dbg(pm8001_ha, MSG, 3514 "DEVREG_FAILURE_PORT_NOT_VALID_STATE\n"); 3515 break; 3516 case DEVREG_FAILURE_DEVICE_TYPE_NOT_VALID: 3517 pm8001_dbg(pm8001_ha, MSG, 3518 "DEVREG_FAILURE_DEVICE_TYPE_NOT_VALID\n"); 3519 break; 3520 default: 3521 pm8001_dbg(pm8001_ha, MSG, 3522 "DEVREG_FAILURE_DEVICE_TYPE_NOT_SUPPORTED\n"); 3523 break; 3524 } 3525 complete(pm8001_dev->dcompletion); 3526 ccb->task = NULL; 3527 ccb->ccb_tag = 0xFFFFFFFF; 3528 pm8001_tag_free(pm8001_ha, htag); 3529 return 0; 3530 } 3531 3532 int pm8001_mpi_dereg_resp(struct pm8001_hba_info *pm8001_ha, void *piomb) 3533 { 3534 u32 status; 3535 u32 device_id; 3536 struct dev_reg_resp *registerRespPayload = 3537 (struct dev_reg_resp *)(piomb + 4); 3538 3539 status = le32_to_cpu(registerRespPayload->status); 3540 device_id = le32_to_cpu(registerRespPayload->device_id); 3541 if (status != 0) 3542 pm8001_dbg(pm8001_ha, MSG, 3543 " deregister device failed ,status = %x, device_id = %x\n", 3544 status, device_id); 3545 return 0; 3546 } 3547 3548 /** 3549 * fw_flash_update_resp - Response from FW for flash update command. 3550 * @pm8001_ha: our hba card information 3551 * @piomb: IO message buffer 3552 */ 3553 int pm8001_mpi_fw_flash_update_resp(struct pm8001_hba_info *pm8001_ha, 3554 void *piomb) 3555 { 3556 u32 status; 3557 struct fw_flash_Update_resp *ppayload = 3558 (struct fw_flash_Update_resp *)(piomb + 4); 3559 u32 tag = le32_to_cpu(ppayload->tag); 3560 struct pm8001_ccb_info *ccb = &pm8001_ha->ccb_info[tag]; 3561 status = le32_to_cpu(ppayload->status); 3562 switch (status) { 3563 case FLASH_UPDATE_COMPLETE_PENDING_REBOOT: 3564 pm8001_dbg(pm8001_ha, MSG, 3565 ": FLASH_UPDATE_COMPLETE_PENDING_REBOOT\n"); 3566 break; 3567 case FLASH_UPDATE_IN_PROGRESS: 3568 pm8001_dbg(pm8001_ha, MSG, ": FLASH_UPDATE_IN_PROGRESS\n"); 3569 break; 3570 case FLASH_UPDATE_HDR_ERR: 3571 pm8001_dbg(pm8001_ha, MSG, ": FLASH_UPDATE_HDR_ERR\n"); 3572 break; 3573 case FLASH_UPDATE_OFFSET_ERR: 3574 pm8001_dbg(pm8001_ha, MSG, ": FLASH_UPDATE_OFFSET_ERR\n"); 3575 break; 3576 case FLASH_UPDATE_CRC_ERR: 3577 pm8001_dbg(pm8001_ha, MSG, ": FLASH_UPDATE_CRC_ERR\n"); 3578 break; 3579 case FLASH_UPDATE_LENGTH_ERR: 3580 pm8001_dbg(pm8001_ha, MSG, ": FLASH_UPDATE_LENGTH_ERR\n"); 3581 break; 3582 case FLASH_UPDATE_HW_ERR: 3583 pm8001_dbg(pm8001_ha, MSG, ": FLASH_UPDATE_HW_ERR\n"); 3584 break; 3585 case FLASH_UPDATE_DNLD_NOT_SUPPORTED: 3586 pm8001_dbg(pm8001_ha, MSG, 3587 ": FLASH_UPDATE_DNLD_NOT_SUPPORTED\n"); 3588 break; 3589 case FLASH_UPDATE_DISABLED: 3590 pm8001_dbg(pm8001_ha, MSG, ": FLASH_UPDATE_DISABLED\n"); 3591 break; 3592 default: 3593 pm8001_dbg(pm8001_ha, DEVIO, "No matched status = %d\n", 3594 status); 3595 break; 3596 } 3597 kfree(ccb->fw_control_context); 3598 ccb->task = NULL; 3599 ccb->ccb_tag = 0xFFFFFFFF; 3600 pm8001_tag_free(pm8001_ha, tag); 3601 complete(pm8001_ha->nvmd_completion); 3602 return 0; 3603 } 3604 3605 int pm8001_mpi_general_event(struct pm8001_hba_info *pm8001_ha , void *piomb) 3606 { 3607 u32 status; 3608 int i; 3609 struct general_event_resp *pPayload = 3610 (struct general_event_resp *)(piomb + 4); 3611 status = le32_to_cpu(pPayload->status); 3612 pm8001_dbg(pm8001_ha, MSG, " status = 0x%x\n", status); 3613 for (i = 0; i < GENERAL_EVENT_PAYLOAD; i++) 3614 pm8001_dbg(pm8001_ha, MSG, "inb_IOMB_payload[0x%x] 0x%x,\n", 3615 i, 3616 pPayload->inb_IOMB_payload[i]); 3617 return 0; 3618 } 3619 3620 int pm8001_mpi_task_abort_resp(struct pm8001_hba_info *pm8001_ha, void *piomb) 3621 { 3622 struct sas_task *t; 3623 struct pm8001_ccb_info *ccb; 3624 unsigned long flags; 3625 u32 status ; 3626 u32 tag, scp; 3627 struct task_status_struct *ts; 3628 struct pm8001_device *pm8001_dev; 3629 3630 struct task_abort_resp *pPayload = 3631 (struct task_abort_resp *)(piomb + 4); 3632 3633 status = le32_to_cpu(pPayload->status); 3634 tag = le32_to_cpu(pPayload->tag); 3635 if (!tag) { 3636 pm8001_dbg(pm8001_ha, FAIL, " TAG NULL. RETURNING !!!\n"); 3637 return -1; 3638 } 3639 3640 scp = le32_to_cpu(pPayload->scp); 3641 ccb = &pm8001_ha->ccb_info[tag]; 3642 t = ccb->task; 3643 pm8001_dev = ccb->device; /* retrieve device */ 3644 3645 if (!t) { 3646 pm8001_dbg(pm8001_ha, FAIL, " TASK NULL. RETURNING !!!\n"); 3647 return -1; 3648 } 3649 ts = &t->task_status; 3650 if (status != 0) 3651 pm8001_dbg(pm8001_ha, FAIL, "task abort failed status 0x%x ,tag = 0x%x, scp= 0x%x\n", 3652 status, tag, scp); 3653 switch (status) { 3654 case IO_SUCCESS: 3655 pm8001_dbg(pm8001_ha, EH, "IO_SUCCESS\n"); 3656 ts->resp = SAS_TASK_COMPLETE; 3657 ts->stat = SAM_STAT_GOOD; 3658 break; 3659 case IO_NOT_VALID: 3660 pm8001_dbg(pm8001_ha, EH, "IO_NOT_VALID\n"); 3661 ts->resp = TMF_RESP_FUNC_FAILED; 3662 break; 3663 } 3664 spin_lock_irqsave(&t->task_state_lock, flags); 3665 t->task_state_flags &= ~SAS_TASK_STATE_PENDING; 3666 t->task_state_flags &= ~SAS_TASK_AT_INITIATOR; 3667 t->task_state_flags |= SAS_TASK_STATE_DONE; 3668 spin_unlock_irqrestore(&t->task_state_lock, flags); 3669 pm8001_ccb_task_free(pm8001_ha, t, ccb, tag); 3670 mb(); 3671 3672 if (pm8001_dev->id & NCQ_ABORT_ALL_FLAG) { 3673 pm8001_tag_free(pm8001_ha, tag); 3674 sas_free_task(t); 3675 /* clear the flag */ 3676 pm8001_dev->id &= 0xBFFFFFFF; 3677 } else 3678 t->task_done(t); 3679 3680 return 0; 3681 } 3682 3683 /** 3684 * mpi_hw_event -The hw event has come. 3685 * @pm8001_ha: our hba card information 3686 * @piomb: IO message buffer 3687 */ 3688 static int mpi_hw_event(struct pm8001_hba_info *pm8001_ha, void* piomb) 3689 { 3690 unsigned long flags; 3691 struct hw_event_resp *pPayload = 3692 (struct hw_event_resp *)(piomb + 4); 3693 u32 lr_evt_status_phyid_portid = 3694 le32_to_cpu(pPayload->lr_evt_status_phyid_portid); 3695 u8 port_id = (u8)(lr_evt_status_phyid_portid & 0x0000000F); 3696 u8 phy_id = 3697 (u8)((lr_evt_status_phyid_portid & 0x000000F0) >> 4); 3698 u16 eventType = 3699 (u16)((lr_evt_status_phyid_portid & 0x00FFFF00) >> 8); 3700 u8 status = 3701 (u8)((lr_evt_status_phyid_portid & 0x0F000000) >> 24); 3702 struct sas_ha_struct *sas_ha = pm8001_ha->sas; 3703 struct pm8001_phy *phy = &pm8001_ha->phy[phy_id]; 3704 struct asd_sas_phy *sas_phy = sas_ha->sas_phy[phy_id]; 3705 pm8001_dbg(pm8001_ha, DEVIO, 3706 "SPC HW event for portid:%d, phyid:%d, event:%x, status:%x\n", 3707 port_id, phy_id, eventType, status); 3708 switch (eventType) { 3709 case HW_EVENT_PHY_START_STATUS: 3710 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_PHY_START_STATUS status = %x\n", 3711 status); 3712 if (status == 0) { 3713 phy->phy_state = 1; 3714 if (pm8001_ha->flags == PM8001F_RUN_TIME && 3715 phy->enable_completion != NULL) 3716 complete(phy->enable_completion); 3717 } 3718 break; 3719 case HW_EVENT_SAS_PHY_UP: 3720 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_PHY_START_STATUS\n"); 3721 hw_event_sas_phy_up(pm8001_ha, piomb); 3722 break; 3723 case HW_EVENT_SATA_PHY_UP: 3724 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_SATA_PHY_UP\n"); 3725 hw_event_sata_phy_up(pm8001_ha, piomb); 3726 break; 3727 case HW_EVENT_PHY_STOP_STATUS: 3728 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_PHY_STOP_STATUS status = %x\n", 3729 status); 3730 if (status == 0) 3731 phy->phy_state = 0; 3732 break; 3733 case HW_EVENT_SATA_SPINUP_HOLD: 3734 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_SATA_SPINUP_HOLD\n"); 3735 sas_notify_phy_event(&phy->sas_phy, PHYE_SPINUP_HOLD, 3736 GFP_ATOMIC); 3737 break; 3738 case HW_EVENT_PHY_DOWN: 3739 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_PHY_DOWN\n"); 3740 sas_notify_phy_event(&phy->sas_phy, PHYE_LOSS_OF_SIGNAL, 3741 GFP_ATOMIC); 3742 phy->phy_attached = 0; 3743 phy->phy_state = 0; 3744 hw_event_phy_down(pm8001_ha, piomb); 3745 break; 3746 case HW_EVENT_PORT_INVALID: 3747 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_PORT_INVALID\n"); 3748 sas_phy_disconnected(sas_phy); 3749 phy->phy_attached = 0; 3750 sas_notify_port_event(sas_phy, PORTE_LINK_RESET_ERR, 3751 GFP_ATOMIC); 3752 break; 3753 /* the broadcast change primitive received, tell the LIBSAS this event 3754 to revalidate the sas domain*/ 3755 case HW_EVENT_BROADCAST_CHANGE: 3756 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_BROADCAST_CHANGE\n"); 3757 pm8001_hw_event_ack_req(pm8001_ha, 0, HW_EVENT_BROADCAST_CHANGE, 3758 port_id, phy_id, 1, 0); 3759 spin_lock_irqsave(&sas_phy->sas_prim_lock, flags); 3760 sas_phy->sas_prim = HW_EVENT_BROADCAST_CHANGE; 3761 spin_unlock_irqrestore(&sas_phy->sas_prim_lock, flags); 3762 sas_notify_port_event(sas_phy, PORTE_BROADCAST_RCVD, 3763 GFP_ATOMIC); 3764 break; 3765 case HW_EVENT_PHY_ERROR: 3766 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_PHY_ERROR\n"); 3767 sas_phy_disconnected(&phy->sas_phy); 3768 phy->phy_attached = 0; 3769 sas_notify_phy_event(&phy->sas_phy, PHYE_OOB_ERROR, GFP_ATOMIC); 3770 break; 3771 case HW_EVENT_BROADCAST_EXP: 3772 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_BROADCAST_EXP\n"); 3773 spin_lock_irqsave(&sas_phy->sas_prim_lock, flags); 3774 sas_phy->sas_prim = HW_EVENT_BROADCAST_EXP; 3775 spin_unlock_irqrestore(&sas_phy->sas_prim_lock, flags); 3776 sas_notify_port_event(sas_phy, PORTE_BROADCAST_RCVD, 3777 GFP_ATOMIC); 3778 break; 3779 case HW_EVENT_LINK_ERR_INVALID_DWORD: 3780 pm8001_dbg(pm8001_ha, MSG, 3781 "HW_EVENT_LINK_ERR_INVALID_DWORD\n"); 3782 pm8001_hw_event_ack_req(pm8001_ha, 0, 3783 HW_EVENT_LINK_ERR_INVALID_DWORD, port_id, phy_id, 0, 0); 3784 sas_phy_disconnected(sas_phy); 3785 phy->phy_attached = 0; 3786 sas_notify_port_event(sas_phy, PORTE_LINK_RESET_ERR, 3787 GFP_ATOMIC); 3788 break; 3789 case HW_EVENT_LINK_ERR_DISPARITY_ERROR: 3790 pm8001_dbg(pm8001_ha, MSG, 3791 "HW_EVENT_LINK_ERR_DISPARITY_ERROR\n"); 3792 pm8001_hw_event_ack_req(pm8001_ha, 0, 3793 HW_EVENT_LINK_ERR_DISPARITY_ERROR, 3794 port_id, phy_id, 0, 0); 3795 sas_phy_disconnected(sas_phy); 3796 phy->phy_attached = 0; 3797 sas_notify_port_event(sas_phy, PORTE_LINK_RESET_ERR, 3798 GFP_ATOMIC); 3799 break; 3800 case HW_EVENT_LINK_ERR_CODE_VIOLATION: 3801 pm8001_dbg(pm8001_ha, MSG, 3802 "HW_EVENT_LINK_ERR_CODE_VIOLATION\n"); 3803 pm8001_hw_event_ack_req(pm8001_ha, 0, 3804 HW_EVENT_LINK_ERR_CODE_VIOLATION, 3805 port_id, phy_id, 0, 0); 3806 sas_phy_disconnected(sas_phy); 3807 phy->phy_attached = 0; 3808 sas_notify_port_event(sas_phy, PORTE_LINK_RESET_ERR, 3809 GFP_ATOMIC); 3810 break; 3811 case HW_EVENT_LINK_ERR_LOSS_OF_DWORD_SYNCH: 3812 pm8001_dbg(pm8001_ha, MSG, 3813 "HW_EVENT_LINK_ERR_LOSS_OF_DWORD_SYNCH\n"); 3814 pm8001_hw_event_ack_req(pm8001_ha, 0, 3815 HW_EVENT_LINK_ERR_LOSS_OF_DWORD_SYNCH, 3816 port_id, phy_id, 0, 0); 3817 sas_phy_disconnected(sas_phy); 3818 phy->phy_attached = 0; 3819 sas_notify_port_event(sas_phy, PORTE_LINK_RESET_ERR, 3820 GFP_ATOMIC); 3821 break; 3822 case HW_EVENT_MALFUNCTION: 3823 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_MALFUNCTION\n"); 3824 break; 3825 case HW_EVENT_BROADCAST_SES: 3826 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_BROADCAST_SES\n"); 3827 spin_lock_irqsave(&sas_phy->sas_prim_lock, flags); 3828 sas_phy->sas_prim = HW_EVENT_BROADCAST_SES; 3829 spin_unlock_irqrestore(&sas_phy->sas_prim_lock, flags); 3830 sas_notify_port_event(sas_phy, PORTE_BROADCAST_RCVD, 3831 GFP_ATOMIC); 3832 break; 3833 case HW_EVENT_INBOUND_CRC_ERROR: 3834 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_INBOUND_CRC_ERROR\n"); 3835 pm8001_hw_event_ack_req(pm8001_ha, 0, 3836 HW_EVENT_INBOUND_CRC_ERROR, 3837 port_id, phy_id, 0, 0); 3838 break; 3839 case HW_EVENT_HARD_RESET_RECEIVED: 3840 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_HARD_RESET_RECEIVED\n"); 3841 sas_notify_port_event(sas_phy, PORTE_HARD_RESET, GFP_ATOMIC); 3842 break; 3843 case HW_EVENT_ID_FRAME_TIMEOUT: 3844 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_ID_FRAME_TIMEOUT\n"); 3845 sas_phy_disconnected(sas_phy); 3846 phy->phy_attached = 0; 3847 sas_notify_port_event(sas_phy, PORTE_LINK_RESET_ERR, 3848 GFP_ATOMIC); 3849 break; 3850 case HW_EVENT_LINK_ERR_PHY_RESET_FAILED: 3851 pm8001_dbg(pm8001_ha, MSG, 3852 "HW_EVENT_LINK_ERR_PHY_RESET_FAILED\n"); 3853 pm8001_hw_event_ack_req(pm8001_ha, 0, 3854 HW_EVENT_LINK_ERR_PHY_RESET_FAILED, 3855 port_id, phy_id, 0, 0); 3856 sas_phy_disconnected(sas_phy); 3857 phy->phy_attached = 0; 3858 sas_notify_port_event(sas_phy, PORTE_LINK_RESET_ERR, 3859 GFP_ATOMIC); 3860 break; 3861 case HW_EVENT_PORT_RESET_TIMER_TMO: 3862 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_PORT_RESET_TIMER_TMO\n"); 3863 sas_phy_disconnected(sas_phy); 3864 phy->phy_attached = 0; 3865 sas_notify_port_event(sas_phy, PORTE_LINK_RESET_ERR, 3866 GFP_ATOMIC); 3867 break; 3868 case HW_EVENT_PORT_RECOVERY_TIMER_TMO: 3869 pm8001_dbg(pm8001_ha, MSG, 3870 "HW_EVENT_PORT_RECOVERY_TIMER_TMO\n"); 3871 sas_phy_disconnected(sas_phy); 3872 phy->phy_attached = 0; 3873 sas_notify_port_event(sas_phy, PORTE_LINK_RESET_ERR, 3874 GFP_ATOMIC); 3875 break; 3876 case HW_EVENT_PORT_RECOVER: 3877 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_PORT_RECOVER\n"); 3878 break; 3879 case HW_EVENT_PORT_RESET_COMPLETE: 3880 pm8001_dbg(pm8001_ha, MSG, "HW_EVENT_PORT_RESET_COMPLETE\n"); 3881 break; 3882 case EVENT_BROADCAST_ASYNCH_EVENT: 3883 pm8001_dbg(pm8001_ha, MSG, "EVENT_BROADCAST_ASYNCH_EVENT\n"); 3884 break; 3885 default: 3886 pm8001_dbg(pm8001_ha, DEVIO, "Unknown event type = %x\n", 3887 eventType); 3888 break; 3889 } 3890 return 0; 3891 } 3892 3893 /** 3894 * process_one_iomb - process one outbound Queue memory block 3895 * @pm8001_ha: our hba card information 3896 * @piomb: IO message buffer 3897 */ 3898 static void process_one_iomb(struct pm8001_hba_info *pm8001_ha, void *piomb) 3899 { 3900 __le32 pHeader = *(__le32 *)piomb; 3901 u8 opc = (u8)((le32_to_cpu(pHeader)) & 0xFFF); 3902 3903 pm8001_dbg(pm8001_ha, MSG, "process_one_iomb:\n"); 3904 3905 switch (opc) { 3906 case OPC_OUB_ECHO: 3907 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_ECHO\n"); 3908 break; 3909 case OPC_OUB_HW_EVENT: 3910 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_HW_EVENT\n"); 3911 mpi_hw_event(pm8001_ha, piomb); 3912 break; 3913 case OPC_OUB_SSP_COMP: 3914 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SSP_COMP\n"); 3915 mpi_ssp_completion(pm8001_ha, piomb); 3916 break; 3917 case OPC_OUB_SMP_COMP: 3918 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SMP_COMP\n"); 3919 mpi_smp_completion(pm8001_ha, piomb); 3920 break; 3921 case OPC_OUB_LOCAL_PHY_CNTRL: 3922 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_LOCAL_PHY_CNTRL\n"); 3923 pm8001_mpi_local_phy_ctl(pm8001_ha, piomb); 3924 break; 3925 case OPC_OUB_DEV_REGIST: 3926 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_DEV_REGIST\n"); 3927 pm8001_mpi_reg_resp(pm8001_ha, piomb); 3928 break; 3929 case OPC_OUB_DEREG_DEV: 3930 pm8001_dbg(pm8001_ha, MSG, "unregister the device\n"); 3931 pm8001_mpi_dereg_resp(pm8001_ha, piomb); 3932 break; 3933 case OPC_OUB_GET_DEV_HANDLE: 3934 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_GET_DEV_HANDLE\n"); 3935 break; 3936 case OPC_OUB_SATA_COMP: 3937 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SATA_COMP\n"); 3938 mpi_sata_completion(pm8001_ha, piomb); 3939 break; 3940 case OPC_OUB_SATA_EVENT: 3941 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SATA_EVENT\n"); 3942 mpi_sata_event(pm8001_ha, piomb); 3943 break; 3944 case OPC_OUB_SSP_EVENT: 3945 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SSP_EVENT\n"); 3946 mpi_ssp_event(pm8001_ha, piomb); 3947 break; 3948 case OPC_OUB_DEV_HANDLE_ARRIV: 3949 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_DEV_HANDLE_ARRIV\n"); 3950 /*This is for target*/ 3951 break; 3952 case OPC_OUB_SSP_RECV_EVENT: 3953 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SSP_RECV_EVENT\n"); 3954 /*This is for target*/ 3955 break; 3956 case OPC_OUB_DEV_INFO: 3957 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_DEV_INFO\n"); 3958 break; 3959 case OPC_OUB_FW_FLASH_UPDATE: 3960 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_FW_FLASH_UPDATE\n"); 3961 pm8001_mpi_fw_flash_update_resp(pm8001_ha, piomb); 3962 break; 3963 case OPC_OUB_GPIO_RESPONSE: 3964 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_GPIO_RESPONSE\n"); 3965 break; 3966 case OPC_OUB_GPIO_EVENT: 3967 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_GPIO_EVENT\n"); 3968 break; 3969 case OPC_OUB_GENERAL_EVENT: 3970 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_GENERAL_EVENT\n"); 3971 pm8001_mpi_general_event(pm8001_ha, piomb); 3972 break; 3973 case OPC_OUB_SSP_ABORT_RSP: 3974 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SSP_ABORT_RSP\n"); 3975 pm8001_mpi_task_abort_resp(pm8001_ha, piomb); 3976 break; 3977 case OPC_OUB_SATA_ABORT_RSP: 3978 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SATA_ABORT_RSP\n"); 3979 pm8001_mpi_task_abort_resp(pm8001_ha, piomb); 3980 break; 3981 case OPC_OUB_SAS_DIAG_MODE_START_END: 3982 pm8001_dbg(pm8001_ha, MSG, 3983 "OPC_OUB_SAS_DIAG_MODE_START_END\n"); 3984 break; 3985 case OPC_OUB_SAS_DIAG_EXECUTE: 3986 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SAS_DIAG_EXECUTE\n"); 3987 break; 3988 case OPC_OUB_GET_TIME_STAMP: 3989 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_GET_TIME_STAMP\n"); 3990 break; 3991 case OPC_OUB_SAS_HW_EVENT_ACK: 3992 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SAS_HW_EVENT_ACK\n"); 3993 break; 3994 case OPC_OUB_PORT_CONTROL: 3995 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_PORT_CONTROL\n"); 3996 break; 3997 case OPC_OUB_SMP_ABORT_RSP: 3998 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SMP_ABORT_RSP\n"); 3999 pm8001_mpi_task_abort_resp(pm8001_ha, piomb); 4000 break; 4001 case OPC_OUB_GET_NVMD_DATA: 4002 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_GET_NVMD_DATA\n"); 4003 pm8001_mpi_get_nvmd_resp(pm8001_ha, piomb); 4004 break; 4005 case OPC_OUB_SET_NVMD_DATA: 4006 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SET_NVMD_DATA\n"); 4007 pm8001_mpi_set_nvmd_resp(pm8001_ha, piomb); 4008 break; 4009 case OPC_OUB_DEVICE_HANDLE_REMOVAL: 4010 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_DEVICE_HANDLE_REMOVAL\n"); 4011 break; 4012 case OPC_OUB_SET_DEVICE_STATE: 4013 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SET_DEVICE_STATE\n"); 4014 pm8001_mpi_set_dev_state_resp(pm8001_ha, piomb); 4015 break; 4016 case OPC_OUB_GET_DEVICE_STATE: 4017 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_GET_DEVICE_STATE\n"); 4018 break; 4019 case OPC_OUB_SET_DEV_INFO: 4020 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SET_DEV_INFO\n"); 4021 break; 4022 case OPC_OUB_SAS_RE_INITIALIZE: 4023 pm8001_dbg(pm8001_ha, MSG, "OPC_OUB_SAS_RE_INITIALIZE\n"); 4024 break; 4025 default: 4026 pm8001_dbg(pm8001_ha, DEVIO, 4027 "Unknown outbound Queue IOMB OPC = %x\n", 4028 opc); 4029 break; 4030 } 4031 } 4032 4033 static int process_oq(struct pm8001_hba_info *pm8001_ha, u8 vec) 4034 { 4035 struct outbound_queue_table *circularQ; 4036 void *pMsg1 = NULL; 4037 u8 bc; 4038 u32 ret = MPI_IO_STATUS_FAIL; 4039 unsigned long flags; 4040 4041 spin_lock_irqsave(&pm8001_ha->lock, flags); 4042 circularQ = &pm8001_ha->outbnd_q_tbl[vec]; 4043 do { 4044 ret = pm8001_mpi_msg_consume(pm8001_ha, circularQ, &pMsg1, &bc); 4045 if (MPI_IO_STATUS_SUCCESS == ret) { 4046 /* process the outbound message */ 4047 process_one_iomb(pm8001_ha, (void *)(pMsg1 - 4)); 4048 /* free the message from the outbound circular buffer */ 4049 pm8001_mpi_msg_free_set(pm8001_ha, pMsg1, 4050 circularQ, bc); 4051 } 4052 if (MPI_IO_STATUS_BUSY == ret) { 4053 /* Update the producer index from SPC */ 4054 circularQ->producer_index = 4055 cpu_to_le32(pm8001_read_32(circularQ->pi_virt)); 4056 if (le32_to_cpu(circularQ->producer_index) == 4057 circularQ->consumer_idx) 4058 /* OQ is empty */ 4059 break; 4060 } 4061 } while (1); 4062 spin_unlock_irqrestore(&pm8001_ha->lock, flags); 4063 return ret; 4064 } 4065 4066 /* DMA_... to our direction translation. */ 4067 static const u8 data_dir_flags[] = { 4068 [DMA_BIDIRECTIONAL] = DATA_DIR_BYRECIPIENT, /* UNSPECIFIED */ 4069 [DMA_TO_DEVICE] = DATA_DIR_OUT, /* OUTBOUND */ 4070 [DMA_FROM_DEVICE] = DATA_DIR_IN, /* INBOUND */ 4071 [DMA_NONE] = DATA_DIR_NONE, /* NO TRANSFER */ 4072 }; 4073 void 4074 pm8001_chip_make_sg(struct scatterlist *scatter, int nr, void *prd) 4075 { 4076 int i; 4077 struct scatterlist *sg; 4078 struct pm8001_prd *buf_prd = prd; 4079 4080 for_each_sg(scatter, sg, nr, i) { 4081 buf_prd->addr = cpu_to_le64(sg_dma_address(sg)); 4082 buf_prd->im_len.len = cpu_to_le32(sg_dma_len(sg)); 4083 buf_prd->im_len.e = 0; 4084 buf_prd++; 4085 } 4086 } 4087 4088 static void build_smp_cmd(u32 deviceID, __le32 hTag, struct smp_req *psmp_cmd) 4089 { 4090 psmp_cmd->tag = hTag; 4091 psmp_cmd->device_id = cpu_to_le32(deviceID); 4092 psmp_cmd->len_ip_ir = cpu_to_le32(1|(1 << 1)); 4093 } 4094 4095 /** 4096 * pm8001_chip_smp_req - send a SMP task to FW 4097 * @pm8001_ha: our hba card information. 4098 * @ccb: the ccb information this request used. 4099 */ 4100 static int pm8001_chip_smp_req(struct pm8001_hba_info *pm8001_ha, 4101 struct pm8001_ccb_info *ccb) 4102 { 4103 int elem, rc; 4104 struct sas_task *task = ccb->task; 4105 struct domain_device *dev = task->dev; 4106 struct pm8001_device *pm8001_dev = dev->lldd_dev; 4107 struct scatterlist *sg_req, *sg_resp; 4108 u32 req_len, resp_len; 4109 struct smp_req smp_cmd; 4110 u32 opc; 4111 struct inbound_queue_table *circularQ; 4112 4113 memset(&smp_cmd, 0, sizeof(smp_cmd)); 4114 /* 4115 * DMA-map SMP request, response buffers 4116 */ 4117 sg_req = &task->smp_task.smp_req; 4118 elem = dma_map_sg(pm8001_ha->dev, sg_req, 1, DMA_TO_DEVICE); 4119 if (!elem) 4120 return -ENOMEM; 4121 req_len = sg_dma_len(sg_req); 4122 4123 sg_resp = &task->smp_task.smp_resp; 4124 elem = dma_map_sg(pm8001_ha->dev, sg_resp, 1, DMA_FROM_DEVICE); 4125 if (!elem) { 4126 rc = -ENOMEM; 4127 goto err_out; 4128 } 4129 resp_len = sg_dma_len(sg_resp); 4130 /* must be in dwords */ 4131 if ((req_len & 0x3) || (resp_len & 0x3)) { 4132 rc = -EINVAL; 4133 goto err_out_2; 4134 } 4135 4136 opc = OPC_INB_SMP_REQUEST; 4137 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4138 smp_cmd.tag = cpu_to_le32(ccb->ccb_tag); 4139 smp_cmd.long_smp_req.long_req_addr = 4140 cpu_to_le64((u64)sg_dma_address(&task->smp_task.smp_req)); 4141 smp_cmd.long_smp_req.long_req_size = 4142 cpu_to_le32((u32)sg_dma_len(&task->smp_task.smp_req)-4); 4143 smp_cmd.long_smp_req.long_resp_addr = 4144 cpu_to_le64((u64)sg_dma_address(&task->smp_task.smp_resp)); 4145 smp_cmd.long_smp_req.long_resp_size = 4146 cpu_to_le32((u32)sg_dma_len(&task->smp_task.smp_resp)-4); 4147 build_smp_cmd(pm8001_dev->device_id, smp_cmd.tag, &smp_cmd); 4148 rc = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, 4149 &smp_cmd, sizeof(smp_cmd), 0); 4150 if (rc) 4151 goto err_out_2; 4152 4153 return 0; 4154 4155 err_out_2: 4156 dma_unmap_sg(pm8001_ha->dev, &ccb->task->smp_task.smp_resp, 1, 4157 DMA_FROM_DEVICE); 4158 err_out: 4159 dma_unmap_sg(pm8001_ha->dev, &ccb->task->smp_task.smp_req, 1, 4160 DMA_TO_DEVICE); 4161 return rc; 4162 } 4163 4164 /** 4165 * pm8001_chip_ssp_io_req - send a SSP task to FW 4166 * @pm8001_ha: our hba card information. 4167 * @ccb: the ccb information this request used. 4168 */ 4169 static int pm8001_chip_ssp_io_req(struct pm8001_hba_info *pm8001_ha, 4170 struct pm8001_ccb_info *ccb) 4171 { 4172 struct sas_task *task = ccb->task; 4173 struct domain_device *dev = task->dev; 4174 struct pm8001_device *pm8001_dev = dev->lldd_dev; 4175 struct ssp_ini_io_start_req ssp_cmd; 4176 u32 tag = ccb->ccb_tag; 4177 int ret; 4178 u64 phys_addr; 4179 struct inbound_queue_table *circularQ; 4180 u32 opc = OPC_INB_SSPINIIOSTART; 4181 memset(&ssp_cmd, 0, sizeof(ssp_cmd)); 4182 memcpy(ssp_cmd.ssp_iu.lun, task->ssp_task.LUN, 8); 4183 ssp_cmd.dir_m_tlr = 4184 cpu_to_le32(data_dir_flags[task->data_dir] << 8 | 0x0);/*0 for 4185 SAS 1.1 compatible TLR*/ 4186 ssp_cmd.data_len = cpu_to_le32(task->total_xfer_len); 4187 ssp_cmd.device_id = cpu_to_le32(pm8001_dev->device_id); 4188 ssp_cmd.tag = cpu_to_le32(tag); 4189 if (task->ssp_task.enable_first_burst) 4190 ssp_cmd.ssp_iu.efb_prio_attr |= 0x80; 4191 ssp_cmd.ssp_iu.efb_prio_attr |= (task->ssp_task.task_prio << 3); 4192 ssp_cmd.ssp_iu.efb_prio_attr |= (task->ssp_task.task_attr & 7); 4193 memcpy(ssp_cmd.ssp_iu.cdb, task->ssp_task.cmd->cmnd, 4194 task->ssp_task.cmd->cmd_len); 4195 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4196 4197 /* fill in PRD (scatter/gather) table, if any */ 4198 if (task->num_scatter > 1) { 4199 pm8001_chip_make_sg(task->scatter, ccb->n_elem, ccb->buf_prd); 4200 phys_addr = ccb->ccb_dma_handle; 4201 ssp_cmd.addr_low = cpu_to_le32(lower_32_bits(phys_addr)); 4202 ssp_cmd.addr_high = cpu_to_le32(upper_32_bits(phys_addr)); 4203 ssp_cmd.esgl = cpu_to_le32(1<<31); 4204 } else if (task->num_scatter == 1) { 4205 u64 dma_addr = sg_dma_address(task->scatter); 4206 ssp_cmd.addr_low = cpu_to_le32(lower_32_bits(dma_addr)); 4207 ssp_cmd.addr_high = cpu_to_le32(upper_32_bits(dma_addr)); 4208 ssp_cmd.len = cpu_to_le32(task->total_xfer_len); 4209 ssp_cmd.esgl = 0; 4210 } else if (task->num_scatter == 0) { 4211 ssp_cmd.addr_low = 0; 4212 ssp_cmd.addr_high = 0; 4213 ssp_cmd.len = cpu_to_le32(task->total_xfer_len); 4214 ssp_cmd.esgl = 0; 4215 } 4216 ret = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &ssp_cmd, 4217 sizeof(ssp_cmd), 0); 4218 return ret; 4219 } 4220 4221 static int pm8001_chip_sata_req(struct pm8001_hba_info *pm8001_ha, 4222 struct pm8001_ccb_info *ccb) 4223 { 4224 struct sas_task *task = ccb->task; 4225 struct domain_device *dev = task->dev; 4226 struct pm8001_device *pm8001_ha_dev = dev->lldd_dev; 4227 u32 tag = ccb->ccb_tag; 4228 int ret; 4229 struct sata_start_req sata_cmd; 4230 u32 hdr_tag, ncg_tag = 0; 4231 u64 phys_addr; 4232 u32 ATAP = 0x0; 4233 u32 dir; 4234 struct inbound_queue_table *circularQ; 4235 unsigned long flags; 4236 u32 opc = OPC_INB_SATA_HOST_OPSTART; 4237 memset(&sata_cmd, 0, sizeof(sata_cmd)); 4238 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4239 if (task->data_dir == DMA_NONE) { 4240 ATAP = 0x04; /* no data*/ 4241 pm8001_dbg(pm8001_ha, IO, "no data\n"); 4242 } else if (likely(!task->ata_task.device_control_reg_update)) { 4243 if (task->ata_task.dma_xfer) { 4244 ATAP = 0x06; /* DMA */ 4245 pm8001_dbg(pm8001_ha, IO, "DMA\n"); 4246 } else { 4247 ATAP = 0x05; /* PIO*/ 4248 pm8001_dbg(pm8001_ha, IO, "PIO\n"); 4249 } 4250 if (task->ata_task.use_ncq && 4251 dev->sata_dev.class != ATA_DEV_ATAPI) { 4252 ATAP = 0x07; /* FPDMA */ 4253 pm8001_dbg(pm8001_ha, IO, "FPDMA\n"); 4254 } 4255 } 4256 if (task->ata_task.use_ncq && pm8001_get_ncq_tag(task, &hdr_tag)) { 4257 task->ata_task.fis.sector_count |= (u8) (hdr_tag << 3); 4258 ncg_tag = hdr_tag; 4259 } 4260 dir = data_dir_flags[task->data_dir] << 8; 4261 sata_cmd.tag = cpu_to_le32(tag); 4262 sata_cmd.device_id = cpu_to_le32(pm8001_ha_dev->device_id); 4263 sata_cmd.data_len = cpu_to_le32(task->total_xfer_len); 4264 sata_cmd.ncqtag_atap_dir_m = 4265 cpu_to_le32(((ncg_tag & 0xff)<<16)|((ATAP & 0x3f) << 10) | dir); 4266 sata_cmd.sata_fis = task->ata_task.fis; 4267 if (likely(!task->ata_task.device_control_reg_update)) 4268 sata_cmd.sata_fis.flags |= 0x80;/* C=1: update ATA cmd reg */ 4269 sata_cmd.sata_fis.flags &= 0xF0;/* PM_PORT field shall be 0 */ 4270 /* fill in PRD (scatter/gather) table, if any */ 4271 if (task->num_scatter > 1) { 4272 pm8001_chip_make_sg(task->scatter, ccb->n_elem, ccb->buf_prd); 4273 phys_addr = ccb->ccb_dma_handle; 4274 sata_cmd.addr_low = lower_32_bits(phys_addr); 4275 sata_cmd.addr_high = upper_32_bits(phys_addr); 4276 sata_cmd.esgl = cpu_to_le32(1 << 31); 4277 } else if (task->num_scatter == 1) { 4278 u64 dma_addr = sg_dma_address(task->scatter); 4279 sata_cmd.addr_low = lower_32_bits(dma_addr); 4280 sata_cmd.addr_high = upper_32_bits(dma_addr); 4281 sata_cmd.len = cpu_to_le32(task->total_xfer_len); 4282 sata_cmd.esgl = 0; 4283 } else if (task->num_scatter == 0) { 4284 sata_cmd.addr_low = 0; 4285 sata_cmd.addr_high = 0; 4286 sata_cmd.len = cpu_to_le32(task->total_xfer_len); 4287 sata_cmd.esgl = 0; 4288 } 4289 4290 /* Check for read log for failed drive and return */ 4291 if (sata_cmd.sata_fis.command == 0x2f) { 4292 if (((pm8001_ha_dev->id & NCQ_READ_LOG_FLAG) || 4293 (pm8001_ha_dev->id & NCQ_ABORT_ALL_FLAG) || 4294 (pm8001_ha_dev->id & NCQ_2ND_RLE_FLAG))) { 4295 struct task_status_struct *ts; 4296 4297 pm8001_ha_dev->id &= 0xDFFFFFFF; 4298 ts = &task->task_status; 4299 4300 spin_lock_irqsave(&task->task_state_lock, flags); 4301 ts->resp = SAS_TASK_COMPLETE; 4302 ts->stat = SAM_STAT_GOOD; 4303 task->task_state_flags &= ~SAS_TASK_STATE_PENDING; 4304 task->task_state_flags &= ~SAS_TASK_AT_INITIATOR; 4305 task->task_state_flags |= SAS_TASK_STATE_DONE; 4306 if (unlikely((task->task_state_flags & 4307 SAS_TASK_STATE_ABORTED))) { 4308 spin_unlock_irqrestore(&task->task_state_lock, 4309 flags); 4310 pm8001_dbg(pm8001_ha, FAIL, 4311 "task 0x%p resp 0x%x stat 0x%x but aborted by upper layer\n", 4312 task, ts->resp, 4313 ts->stat); 4314 pm8001_ccb_task_free(pm8001_ha, task, ccb, tag); 4315 } else { 4316 spin_unlock_irqrestore(&task->task_state_lock, 4317 flags); 4318 pm8001_ccb_task_free_done(pm8001_ha, task, 4319 ccb, tag); 4320 return 0; 4321 } 4322 } 4323 } 4324 4325 ret = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &sata_cmd, 4326 sizeof(sata_cmd), 0); 4327 return ret; 4328 } 4329 4330 /** 4331 * pm8001_chip_phy_start_req - start phy via PHY_START COMMAND 4332 * @pm8001_ha: our hba card information. 4333 * @phy_id: the phy id which we wanted to start up. 4334 */ 4335 static int 4336 pm8001_chip_phy_start_req(struct pm8001_hba_info *pm8001_ha, u8 phy_id) 4337 { 4338 struct phy_start_req payload; 4339 struct inbound_queue_table *circularQ; 4340 int ret; 4341 u32 tag = 0x01; 4342 u32 opcode = OPC_INB_PHYSTART; 4343 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4344 memset(&payload, 0, sizeof(payload)); 4345 payload.tag = cpu_to_le32(tag); 4346 /* 4347 ** [0:7] PHY Identifier 4348 ** [8:11] link rate 1.5G, 3G, 6G 4349 ** [12:13] link mode 01b SAS mode; 10b SATA mode; 11b both 4350 ** [14] 0b disable spin up hold; 1b enable spin up hold 4351 */ 4352 payload.ase_sh_lm_slr_phyid = cpu_to_le32(SPINHOLD_DISABLE | 4353 LINKMODE_AUTO | LINKRATE_15 | 4354 LINKRATE_30 | LINKRATE_60 | phy_id); 4355 payload.sas_identify.dev_type = SAS_END_DEVICE; 4356 payload.sas_identify.initiator_bits = SAS_PROTOCOL_ALL; 4357 memcpy(payload.sas_identify.sas_addr, 4358 pm8001_ha->sas_addr, SAS_ADDR_SIZE); 4359 payload.sas_identify.phy_id = phy_id; 4360 ret = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opcode, &payload, 4361 sizeof(payload), 0); 4362 return ret; 4363 } 4364 4365 /** 4366 * pm8001_chip_phy_stop_req - start phy via PHY_STOP COMMAND 4367 * @pm8001_ha: our hba card information. 4368 * @phy_id: the phy id which we wanted to start up. 4369 */ 4370 static int pm8001_chip_phy_stop_req(struct pm8001_hba_info *pm8001_ha, 4371 u8 phy_id) 4372 { 4373 struct phy_stop_req payload; 4374 struct inbound_queue_table *circularQ; 4375 int ret; 4376 u32 tag = 0x01; 4377 u32 opcode = OPC_INB_PHYSTOP; 4378 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4379 memset(&payload, 0, sizeof(payload)); 4380 payload.tag = cpu_to_le32(tag); 4381 payload.phy_id = cpu_to_le32(phy_id); 4382 ret = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opcode, &payload, 4383 sizeof(payload), 0); 4384 return ret; 4385 } 4386 4387 /* 4388 * see comments on pm8001_mpi_reg_resp. 4389 */ 4390 static int pm8001_chip_reg_dev_req(struct pm8001_hba_info *pm8001_ha, 4391 struct pm8001_device *pm8001_dev, u32 flag) 4392 { 4393 struct reg_dev_req payload; 4394 u32 opc; 4395 u32 stp_sspsmp_sata = 0x4; 4396 struct inbound_queue_table *circularQ; 4397 u32 linkrate, phy_id; 4398 int rc, tag = 0xdeadbeef; 4399 struct pm8001_ccb_info *ccb; 4400 u8 retryFlag = 0x1; 4401 u16 firstBurstSize = 0; 4402 u16 ITNT = 2000; 4403 struct domain_device *dev = pm8001_dev->sas_device; 4404 struct domain_device *parent_dev = dev->parent; 4405 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4406 4407 memset(&payload, 0, sizeof(payload)); 4408 rc = pm8001_tag_alloc(pm8001_ha, &tag); 4409 if (rc) 4410 return rc; 4411 ccb = &pm8001_ha->ccb_info[tag]; 4412 ccb->device = pm8001_dev; 4413 ccb->ccb_tag = tag; 4414 payload.tag = cpu_to_le32(tag); 4415 if (flag == 1) 4416 stp_sspsmp_sata = 0x02; /*direct attached sata */ 4417 else { 4418 if (pm8001_dev->dev_type == SAS_SATA_DEV) 4419 stp_sspsmp_sata = 0x00; /* stp*/ 4420 else if (pm8001_dev->dev_type == SAS_END_DEVICE || 4421 pm8001_dev->dev_type == SAS_EDGE_EXPANDER_DEVICE || 4422 pm8001_dev->dev_type == SAS_FANOUT_EXPANDER_DEVICE) 4423 stp_sspsmp_sata = 0x01; /*ssp or smp*/ 4424 } 4425 if (parent_dev && dev_is_expander(parent_dev->dev_type)) 4426 phy_id = parent_dev->ex_dev.ex_phy->phy_id; 4427 else 4428 phy_id = pm8001_dev->attached_phy; 4429 opc = OPC_INB_REG_DEV; 4430 linkrate = (pm8001_dev->sas_device->linkrate < dev->port->linkrate) ? 4431 pm8001_dev->sas_device->linkrate : dev->port->linkrate; 4432 payload.phyid_portid = 4433 cpu_to_le32(((pm8001_dev->sas_device->port->id) & 0x0F) | 4434 ((phy_id & 0x0F) << 4)); 4435 payload.dtype_dlr_retry = cpu_to_le32((retryFlag & 0x01) | 4436 ((linkrate & 0x0F) * 0x1000000) | 4437 ((stp_sspsmp_sata & 0x03) * 0x10000000)); 4438 payload.firstburstsize_ITNexustimeout = 4439 cpu_to_le32(ITNT | (firstBurstSize * 0x10000)); 4440 memcpy(payload.sas_addr, pm8001_dev->sas_device->sas_addr, 4441 SAS_ADDR_SIZE); 4442 rc = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &payload, 4443 sizeof(payload), 0); 4444 return rc; 4445 } 4446 4447 /* 4448 * see comments on pm8001_mpi_reg_resp. 4449 */ 4450 int pm8001_chip_dereg_dev_req(struct pm8001_hba_info *pm8001_ha, 4451 u32 device_id) 4452 { 4453 struct dereg_dev_req payload; 4454 u32 opc = OPC_INB_DEREG_DEV_HANDLE; 4455 int ret; 4456 struct inbound_queue_table *circularQ; 4457 4458 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4459 memset(&payload, 0, sizeof(payload)); 4460 payload.tag = cpu_to_le32(1); 4461 payload.device_id = cpu_to_le32(device_id); 4462 pm8001_dbg(pm8001_ha, MSG, "unregister device device_id = %d\n", 4463 device_id); 4464 ret = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &payload, 4465 sizeof(payload), 0); 4466 return ret; 4467 } 4468 4469 /** 4470 * pm8001_chip_phy_ctl_req - support the local phy operation 4471 * @pm8001_ha: our hba card information. 4472 * @phyId: the phy id which we wanted to operate 4473 * @phy_op: the phy operation to request 4474 */ 4475 static int pm8001_chip_phy_ctl_req(struct pm8001_hba_info *pm8001_ha, 4476 u32 phyId, u32 phy_op) 4477 { 4478 struct local_phy_ctl_req payload; 4479 struct inbound_queue_table *circularQ; 4480 int ret; 4481 u32 opc = OPC_INB_LOCAL_PHY_CONTROL; 4482 memset(&payload, 0, sizeof(payload)); 4483 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4484 payload.tag = cpu_to_le32(1); 4485 payload.phyop_phyid = 4486 cpu_to_le32(((phy_op & 0xff) << 8) | (phyId & 0x0F)); 4487 ret = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &payload, 4488 sizeof(payload), 0); 4489 return ret; 4490 } 4491 4492 static u32 pm8001_chip_is_our_interrupt(struct pm8001_hba_info *pm8001_ha) 4493 { 4494 #ifdef PM8001_USE_MSIX 4495 return 1; 4496 #else 4497 u32 value; 4498 4499 value = pm8001_cr32(pm8001_ha, 0, MSGU_ODR); 4500 if (value) 4501 return 1; 4502 return 0; 4503 #endif 4504 } 4505 4506 /** 4507 * pm8001_chip_isr - PM8001 isr handler. 4508 * @pm8001_ha: our hba card information. 4509 * @vec: IRQ number 4510 */ 4511 static irqreturn_t 4512 pm8001_chip_isr(struct pm8001_hba_info *pm8001_ha, u8 vec) 4513 { 4514 pm8001_chip_interrupt_disable(pm8001_ha, vec); 4515 pm8001_dbg(pm8001_ha, DEVIO, 4516 "irq vec %d, ODMR:0x%x\n", 4517 vec, pm8001_cr32(pm8001_ha, 0, 0x30)); 4518 process_oq(pm8001_ha, vec); 4519 pm8001_chip_interrupt_enable(pm8001_ha, vec); 4520 return IRQ_HANDLED; 4521 } 4522 4523 static int send_task_abort(struct pm8001_hba_info *pm8001_ha, u32 opc, 4524 u32 dev_id, u8 flag, u32 task_tag, u32 cmd_tag) 4525 { 4526 struct task_abort_req task_abort; 4527 struct inbound_queue_table *circularQ; 4528 int ret; 4529 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4530 memset(&task_abort, 0, sizeof(task_abort)); 4531 if (ABORT_SINGLE == (flag & ABORT_MASK)) { 4532 task_abort.abort_all = 0; 4533 task_abort.device_id = cpu_to_le32(dev_id); 4534 task_abort.tag_to_abort = cpu_to_le32(task_tag); 4535 task_abort.tag = cpu_to_le32(cmd_tag); 4536 } else if (ABORT_ALL == (flag & ABORT_MASK)) { 4537 task_abort.abort_all = cpu_to_le32(1); 4538 task_abort.device_id = cpu_to_le32(dev_id); 4539 task_abort.tag = cpu_to_le32(cmd_tag); 4540 } 4541 ret = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &task_abort, 4542 sizeof(task_abort), 0); 4543 return ret; 4544 } 4545 4546 /* 4547 * pm8001_chip_abort_task - SAS abort task when error or exception happened. 4548 */ 4549 int pm8001_chip_abort_task(struct pm8001_hba_info *pm8001_ha, 4550 struct pm8001_device *pm8001_dev, u8 flag, u32 task_tag, u32 cmd_tag) 4551 { 4552 u32 opc, device_id; 4553 int rc = TMF_RESP_FUNC_FAILED; 4554 pm8001_dbg(pm8001_ha, EH, "cmd_tag = %x, abort task tag = 0x%x\n", 4555 cmd_tag, task_tag); 4556 if (pm8001_dev->dev_type == SAS_END_DEVICE) 4557 opc = OPC_INB_SSP_ABORT; 4558 else if (pm8001_dev->dev_type == SAS_SATA_DEV) 4559 opc = OPC_INB_SATA_ABORT; 4560 else 4561 opc = OPC_INB_SMP_ABORT;/* SMP */ 4562 device_id = pm8001_dev->device_id; 4563 rc = send_task_abort(pm8001_ha, opc, device_id, flag, 4564 task_tag, cmd_tag); 4565 if (rc != TMF_RESP_FUNC_COMPLETE) 4566 pm8001_dbg(pm8001_ha, EH, "rc= %d\n", rc); 4567 return rc; 4568 } 4569 4570 /** 4571 * pm8001_chip_ssp_tm_req - built the task management command. 4572 * @pm8001_ha: our hba card information. 4573 * @ccb: the ccb information. 4574 * @tmf: task management function. 4575 */ 4576 int pm8001_chip_ssp_tm_req(struct pm8001_hba_info *pm8001_ha, 4577 struct pm8001_ccb_info *ccb, struct pm8001_tmf_task *tmf) 4578 { 4579 struct sas_task *task = ccb->task; 4580 struct domain_device *dev = task->dev; 4581 struct pm8001_device *pm8001_dev = dev->lldd_dev; 4582 u32 opc = OPC_INB_SSPINITMSTART; 4583 struct inbound_queue_table *circularQ; 4584 struct ssp_ini_tm_start_req sspTMCmd; 4585 int ret; 4586 4587 memset(&sspTMCmd, 0, sizeof(sspTMCmd)); 4588 sspTMCmd.device_id = cpu_to_le32(pm8001_dev->device_id); 4589 sspTMCmd.relate_tag = cpu_to_le32(tmf->tag_of_task_to_be_managed); 4590 sspTMCmd.tmf = cpu_to_le32(tmf->tmf); 4591 memcpy(sspTMCmd.lun, task->ssp_task.LUN, 8); 4592 sspTMCmd.tag = cpu_to_le32(ccb->ccb_tag); 4593 if (pm8001_ha->chip_id != chip_8001) 4594 sspTMCmd.ds_ads_m = 0x08; 4595 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4596 ret = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &sspTMCmd, 4597 sizeof(sspTMCmd), 0); 4598 return ret; 4599 } 4600 4601 int pm8001_chip_get_nvmd_req(struct pm8001_hba_info *pm8001_ha, 4602 void *payload) 4603 { 4604 u32 opc = OPC_INB_GET_NVMD_DATA; 4605 u32 nvmd_type; 4606 int rc; 4607 u32 tag; 4608 struct pm8001_ccb_info *ccb; 4609 struct inbound_queue_table *circularQ; 4610 struct get_nvm_data_req nvmd_req; 4611 struct fw_control_ex *fw_control_context; 4612 struct pm8001_ioctl_payload *ioctl_payload = payload; 4613 4614 nvmd_type = ioctl_payload->minor_function; 4615 fw_control_context = kzalloc(sizeof(struct fw_control_ex), GFP_KERNEL); 4616 if (!fw_control_context) 4617 return -ENOMEM; 4618 fw_control_context->usrAddr = (u8 *)ioctl_payload->func_specific; 4619 fw_control_context->len = ioctl_payload->rd_length; 4620 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4621 memset(&nvmd_req, 0, sizeof(nvmd_req)); 4622 rc = pm8001_tag_alloc(pm8001_ha, &tag); 4623 if (rc) { 4624 kfree(fw_control_context); 4625 return rc; 4626 } 4627 ccb = &pm8001_ha->ccb_info[tag]; 4628 ccb->ccb_tag = tag; 4629 ccb->fw_control_context = fw_control_context; 4630 nvmd_req.tag = cpu_to_le32(tag); 4631 4632 switch (nvmd_type) { 4633 case TWI_DEVICE: { 4634 u32 twi_addr, twi_page_size; 4635 twi_addr = 0xa8; 4636 twi_page_size = 2; 4637 4638 nvmd_req.len_ir_vpdd = cpu_to_le32(IPMode | twi_addr << 16 | 4639 twi_page_size << 8 | TWI_DEVICE); 4640 nvmd_req.resp_len = cpu_to_le32(ioctl_payload->rd_length); 4641 nvmd_req.resp_addr_hi = 4642 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi); 4643 nvmd_req.resp_addr_lo = 4644 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo); 4645 break; 4646 } 4647 case C_SEEPROM: { 4648 nvmd_req.len_ir_vpdd = cpu_to_le32(IPMode | C_SEEPROM); 4649 nvmd_req.resp_len = cpu_to_le32(ioctl_payload->rd_length); 4650 nvmd_req.resp_addr_hi = 4651 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi); 4652 nvmd_req.resp_addr_lo = 4653 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo); 4654 break; 4655 } 4656 case VPD_FLASH: { 4657 nvmd_req.len_ir_vpdd = cpu_to_le32(IPMode | VPD_FLASH); 4658 nvmd_req.resp_len = cpu_to_le32(ioctl_payload->rd_length); 4659 nvmd_req.resp_addr_hi = 4660 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi); 4661 nvmd_req.resp_addr_lo = 4662 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo); 4663 break; 4664 } 4665 case EXPAN_ROM: { 4666 nvmd_req.len_ir_vpdd = cpu_to_le32(IPMode | EXPAN_ROM); 4667 nvmd_req.resp_len = cpu_to_le32(ioctl_payload->rd_length); 4668 nvmd_req.resp_addr_hi = 4669 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi); 4670 nvmd_req.resp_addr_lo = 4671 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo); 4672 break; 4673 } 4674 case IOP_RDUMP: { 4675 nvmd_req.len_ir_vpdd = cpu_to_le32(IPMode | IOP_RDUMP); 4676 nvmd_req.resp_len = cpu_to_le32(ioctl_payload->rd_length); 4677 nvmd_req.vpd_offset = cpu_to_le32(ioctl_payload->offset); 4678 nvmd_req.resp_addr_hi = 4679 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi); 4680 nvmd_req.resp_addr_lo = 4681 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo); 4682 break; 4683 } 4684 default: 4685 break; 4686 } 4687 rc = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &nvmd_req, 4688 sizeof(nvmd_req), 0); 4689 if (rc) { 4690 kfree(fw_control_context); 4691 pm8001_tag_free(pm8001_ha, tag); 4692 } 4693 return rc; 4694 } 4695 4696 int pm8001_chip_set_nvmd_req(struct pm8001_hba_info *pm8001_ha, 4697 void *payload) 4698 { 4699 u32 opc = OPC_INB_SET_NVMD_DATA; 4700 u32 nvmd_type; 4701 int rc; 4702 u32 tag; 4703 struct pm8001_ccb_info *ccb; 4704 struct inbound_queue_table *circularQ; 4705 struct set_nvm_data_req nvmd_req; 4706 struct fw_control_ex *fw_control_context; 4707 struct pm8001_ioctl_payload *ioctl_payload = payload; 4708 4709 nvmd_type = ioctl_payload->minor_function; 4710 fw_control_context = kzalloc(sizeof(struct fw_control_ex), GFP_KERNEL); 4711 if (!fw_control_context) 4712 return -ENOMEM; 4713 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4714 memcpy(pm8001_ha->memoryMap.region[NVMD].virt_ptr, 4715 &ioctl_payload->func_specific, 4716 ioctl_payload->wr_length); 4717 memset(&nvmd_req, 0, sizeof(nvmd_req)); 4718 rc = pm8001_tag_alloc(pm8001_ha, &tag); 4719 if (rc) { 4720 kfree(fw_control_context); 4721 return -EBUSY; 4722 } 4723 ccb = &pm8001_ha->ccb_info[tag]; 4724 ccb->fw_control_context = fw_control_context; 4725 ccb->ccb_tag = tag; 4726 nvmd_req.tag = cpu_to_le32(tag); 4727 switch (nvmd_type) { 4728 case TWI_DEVICE: { 4729 u32 twi_addr, twi_page_size; 4730 twi_addr = 0xa8; 4731 twi_page_size = 2; 4732 nvmd_req.reserved[0] = cpu_to_le32(0xFEDCBA98); 4733 nvmd_req.len_ir_vpdd = cpu_to_le32(IPMode | twi_addr << 16 | 4734 twi_page_size << 8 | TWI_DEVICE); 4735 nvmd_req.resp_len = cpu_to_le32(ioctl_payload->wr_length); 4736 nvmd_req.resp_addr_hi = 4737 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi); 4738 nvmd_req.resp_addr_lo = 4739 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo); 4740 break; 4741 } 4742 case C_SEEPROM: 4743 nvmd_req.len_ir_vpdd = cpu_to_le32(IPMode | C_SEEPROM); 4744 nvmd_req.resp_len = cpu_to_le32(ioctl_payload->wr_length); 4745 nvmd_req.reserved[0] = cpu_to_le32(0xFEDCBA98); 4746 nvmd_req.resp_addr_hi = 4747 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi); 4748 nvmd_req.resp_addr_lo = 4749 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo); 4750 break; 4751 case VPD_FLASH: 4752 nvmd_req.len_ir_vpdd = cpu_to_le32(IPMode | VPD_FLASH); 4753 nvmd_req.resp_len = cpu_to_le32(ioctl_payload->wr_length); 4754 nvmd_req.reserved[0] = cpu_to_le32(0xFEDCBA98); 4755 nvmd_req.resp_addr_hi = 4756 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi); 4757 nvmd_req.resp_addr_lo = 4758 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo); 4759 break; 4760 case EXPAN_ROM: 4761 nvmd_req.len_ir_vpdd = cpu_to_le32(IPMode | EXPAN_ROM); 4762 nvmd_req.resp_len = cpu_to_le32(ioctl_payload->wr_length); 4763 nvmd_req.reserved[0] = cpu_to_le32(0xFEDCBA98); 4764 nvmd_req.resp_addr_hi = 4765 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_hi); 4766 nvmd_req.resp_addr_lo = 4767 cpu_to_le32(pm8001_ha->memoryMap.region[NVMD].phys_addr_lo); 4768 break; 4769 default: 4770 break; 4771 } 4772 rc = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &nvmd_req, 4773 sizeof(nvmd_req), 0); 4774 if (rc) { 4775 kfree(fw_control_context); 4776 pm8001_tag_free(pm8001_ha, tag); 4777 } 4778 return rc; 4779 } 4780 4781 /** 4782 * pm8001_chip_fw_flash_update_build - support the firmware update operation 4783 * @pm8001_ha: our hba card information. 4784 * @fw_flash_updata_info: firmware flash update param 4785 * @tag: Tag to apply to the payload 4786 */ 4787 int 4788 pm8001_chip_fw_flash_update_build(struct pm8001_hba_info *pm8001_ha, 4789 void *fw_flash_updata_info, u32 tag) 4790 { 4791 struct fw_flash_Update_req payload; 4792 struct fw_flash_updata_info *info; 4793 struct inbound_queue_table *circularQ; 4794 int ret; 4795 u32 opc = OPC_INB_FW_FLASH_UPDATE; 4796 4797 memset(&payload, 0, sizeof(struct fw_flash_Update_req)); 4798 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4799 info = fw_flash_updata_info; 4800 payload.tag = cpu_to_le32(tag); 4801 payload.cur_image_len = cpu_to_le32(info->cur_image_len); 4802 payload.cur_image_offset = cpu_to_le32(info->cur_image_offset); 4803 payload.total_image_len = cpu_to_le32(info->total_image_len); 4804 payload.len = info->sgl.im_len.len ; 4805 payload.sgl_addr_lo = 4806 cpu_to_le32(lower_32_bits(le64_to_cpu(info->sgl.addr))); 4807 payload.sgl_addr_hi = 4808 cpu_to_le32(upper_32_bits(le64_to_cpu(info->sgl.addr))); 4809 ret = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &payload, 4810 sizeof(payload), 0); 4811 return ret; 4812 } 4813 4814 int 4815 pm8001_chip_fw_flash_update_req(struct pm8001_hba_info *pm8001_ha, 4816 void *payload) 4817 { 4818 struct fw_flash_updata_info flash_update_info; 4819 struct fw_control_info *fw_control; 4820 struct fw_control_ex *fw_control_context; 4821 int rc; 4822 u32 tag; 4823 struct pm8001_ccb_info *ccb; 4824 void *buffer = pm8001_ha->memoryMap.region[FW_FLASH].virt_ptr; 4825 dma_addr_t phys_addr = pm8001_ha->memoryMap.region[FW_FLASH].phys_addr; 4826 struct pm8001_ioctl_payload *ioctl_payload = payload; 4827 4828 fw_control_context = kzalloc(sizeof(struct fw_control_ex), GFP_KERNEL); 4829 if (!fw_control_context) 4830 return -ENOMEM; 4831 fw_control = (struct fw_control_info *)&ioctl_payload->func_specific; 4832 pm8001_dbg(pm8001_ha, DEVIO, 4833 "dma fw_control context input length :%x\n", 4834 fw_control->len); 4835 memcpy(buffer, fw_control->buffer, fw_control->len); 4836 flash_update_info.sgl.addr = cpu_to_le64(phys_addr); 4837 flash_update_info.sgl.im_len.len = cpu_to_le32(fw_control->len); 4838 flash_update_info.sgl.im_len.e = 0; 4839 flash_update_info.cur_image_offset = fw_control->offset; 4840 flash_update_info.cur_image_len = fw_control->len; 4841 flash_update_info.total_image_len = fw_control->size; 4842 fw_control_context->fw_control = fw_control; 4843 fw_control_context->virtAddr = buffer; 4844 fw_control_context->phys_addr = phys_addr; 4845 fw_control_context->len = fw_control->len; 4846 rc = pm8001_tag_alloc(pm8001_ha, &tag); 4847 if (rc) { 4848 kfree(fw_control_context); 4849 return -EBUSY; 4850 } 4851 ccb = &pm8001_ha->ccb_info[tag]; 4852 ccb->fw_control_context = fw_control_context; 4853 ccb->ccb_tag = tag; 4854 rc = pm8001_chip_fw_flash_update_build(pm8001_ha, &flash_update_info, 4855 tag); 4856 return rc; 4857 } 4858 4859 ssize_t 4860 pm8001_get_gsm_dump(struct device *cdev, u32 length, char *buf) 4861 { 4862 u32 value, rem, offset = 0, bar = 0; 4863 u32 index, work_offset, dw_length; 4864 u32 shift_value, gsm_base, gsm_dump_offset; 4865 char *direct_data; 4866 struct Scsi_Host *shost = class_to_shost(cdev); 4867 struct sas_ha_struct *sha = SHOST_TO_SAS_HA(shost); 4868 struct pm8001_hba_info *pm8001_ha = sha->lldd_ha; 4869 4870 direct_data = buf; 4871 gsm_dump_offset = pm8001_ha->fatal_forensic_shift_offset; 4872 4873 /* check max is 1 Mbytes */ 4874 if ((length > 0x100000) || (gsm_dump_offset & 3) || 4875 ((gsm_dump_offset + length) > 0x1000000)) 4876 return -EINVAL; 4877 4878 if (pm8001_ha->chip_id == chip_8001) 4879 bar = 2; 4880 else 4881 bar = 1; 4882 4883 work_offset = gsm_dump_offset & 0xFFFF0000; 4884 offset = gsm_dump_offset & 0x0000FFFF; 4885 gsm_dump_offset = work_offset; 4886 /* adjust length to dword boundary */ 4887 rem = length & 3; 4888 dw_length = length >> 2; 4889 4890 for (index = 0; index < dw_length; index++) { 4891 if ((work_offset + offset) & 0xFFFF0000) { 4892 if (pm8001_ha->chip_id == chip_8001) 4893 shift_value = ((gsm_dump_offset + offset) & 4894 SHIFT_REG_64K_MASK); 4895 else 4896 shift_value = (((gsm_dump_offset + offset) & 4897 SHIFT_REG_64K_MASK) >> 4898 SHIFT_REG_BIT_SHIFT); 4899 4900 if (pm8001_ha->chip_id == chip_8001) { 4901 gsm_base = GSM_BASE; 4902 if (-1 == pm8001_bar4_shift(pm8001_ha, 4903 (gsm_base + shift_value))) 4904 return -EIO; 4905 } else { 4906 gsm_base = 0; 4907 if (-1 == pm80xx_bar4_shift(pm8001_ha, 4908 (gsm_base + shift_value))) 4909 return -EIO; 4910 } 4911 gsm_dump_offset = (gsm_dump_offset + offset) & 4912 0xFFFF0000; 4913 work_offset = 0; 4914 offset = offset & 0x0000FFFF; 4915 } 4916 value = pm8001_cr32(pm8001_ha, bar, (work_offset + offset) & 4917 0x0000FFFF); 4918 direct_data += sprintf(direct_data, "%08x ", value); 4919 offset += 4; 4920 } 4921 if (rem != 0) { 4922 value = pm8001_cr32(pm8001_ha, bar, (work_offset + offset) & 4923 0x0000FFFF); 4924 /* xfr for non_dw */ 4925 direct_data += sprintf(direct_data, "%08x ", value); 4926 } 4927 /* Shift back to BAR4 original address */ 4928 if (-1 == pm8001_bar4_shift(pm8001_ha, 0)) 4929 return -EIO; 4930 pm8001_ha->fatal_forensic_shift_offset += 1024; 4931 4932 if (pm8001_ha->fatal_forensic_shift_offset >= 0x100000) 4933 pm8001_ha->fatal_forensic_shift_offset = 0; 4934 return direct_data - buf; 4935 } 4936 4937 int 4938 pm8001_chip_set_dev_state_req(struct pm8001_hba_info *pm8001_ha, 4939 struct pm8001_device *pm8001_dev, u32 state) 4940 { 4941 struct set_dev_state_req payload; 4942 struct inbound_queue_table *circularQ; 4943 struct pm8001_ccb_info *ccb; 4944 int rc; 4945 u32 tag; 4946 u32 opc = OPC_INB_SET_DEVICE_STATE; 4947 memset(&payload, 0, sizeof(payload)); 4948 rc = pm8001_tag_alloc(pm8001_ha, &tag); 4949 if (rc) 4950 return -1; 4951 ccb = &pm8001_ha->ccb_info[tag]; 4952 ccb->ccb_tag = tag; 4953 ccb->device = pm8001_dev; 4954 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4955 payload.tag = cpu_to_le32(tag); 4956 payload.device_id = cpu_to_le32(pm8001_dev->device_id); 4957 payload.nds = cpu_to_le32(state); 4958 rc = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &payload, 4959 sizeof(payload), 0); 4960 return rc; 4961 4962 } 4963 4964 static int 4965 pm8001_chip_sas_re_initialization(struct pm8001_hba_info *pm8001_ha) 4966 { 4967 struct sas_re_initialization_req payload; 4968 struct inbound_queue_table *circularQ; 4969 struct pm8001_ccb_info *ccb; 4970 int rc; 4971 u32 tag; 4972 u32 opc = OPC_INB_SAS_RE_INITIALIZE; 4973 memset(&payload, 0, sizeof(payload)); 4974 rc = pm8001_tag_alloc(pm8001_ha, &tag); 4975 if (rc) 4976 return -ENOMEM; 4977 ccb = &pm8001_ha->ccb_info[tag]; 4978 ccb->ccb_tag = tag; 4979 circularQ = &pm8001_ha->inbnd_q_tbl[0]; 4980 payload.tag = cpu_to_le32(tag); 4981 payload.SSAHOLT = cpu_to_le32(0xd << 25); 4982 payload.sata_hol_tmo = cpu_to_le32(80); 4983 payload.open_reject_cmdretries_data_retries = cpu_to_le32(0xff00ff); 4984 rc = pm8001_mpi_build_cmd(pm8001_ha, circularQ, opc, &payload, 4985 sizeof(payload), 0); 4986 if (rc) 4987 pm8001_tag_free(pm8001_ha, tag); 4988 return rc; 4989 4990 } 4991 4992 const struct pm8001_dispatch pm8001_8001_dispatch = { 4993 .name = "pmc8001", 4994 .chip_init = pm8001_chip_init, 4995 .chip_soft_rst = pm8001_chip_soft_rst, 4996 .chip_rst = pm8001_hw_chip_rst, 4997 .chip_iounmap = pm8001_chip_iounmap, 4998 .isr = pm8001_chip_isr, 4999 .is_our_interrupt = pm8001_chip_is_our_interrupt, 5000 .isr_process_oq = process_oq, 5001 .interrupt_enable = pm8001_chip_interrupt_enable, 5002 .interrupt_disable = pm8001_chip_interrupt_disable, 5003 .make_prd = pm8001_chip_make_sg, 5004 .smp_req = pm8001_chip_smp_req, 5005 .ssp_io_req = pm8001_chip_ssp_io_req, 5006 .sata_req = pm8001_chip_sata_req, 5007 .phy_start_req = pm8001_chip_phy_start_req, 5008 .phy_stop_req = pm8001_chip_phy_stop_req, 5009 .reg_dev_req = pm8001_chip_reg_dev_req, 5010 .dereg_dev_req = pm8001_chip_dereg_dev_req, 5011 .phy_ctl_req = pm8001_chip_phy_ctl_req, 5012 .task_abort = pm8001_chip_abort_task, 5013 .ssp_tm_req = pm8001_chip_ssp_tm_req, 5014 .get_nvmd_req = pm8001_chip_get_nvmd_req, 5015 .set_nvmd_req = pm8001_chip_set_nvmd_req, 5016 .fw_flash_update_req = pm8001_chip_fw_flash_update_req, 5017 .set_dev_state_req = pm8001_chip_set_dev_state_req, 5018 .sas_re_init_req = pm8001_chip_sas_re_initialization, 5019 .fatal_errors = pm80xx_fatal_errors, 5020 }; 5021