1 /* 2 * CDDL HEADER START 3 * 4 * The contents of this file are subject to the terms of the 5 * Common Development and Distribution License (the "License"). 6 * You may not use this file except in compliance with the License. 7 * 8 * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 9 * or http://www.opensolaris.org/os/licensing. 10 * See the License for the specific language governing permissions 11 * and limitations under the License. 12 * 13 * When distributing Covered Code, include this CDDL HEADER in each 14 * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 15 * If applicable, add the following below this CDDL HEADER, with the 16 * fields enclosed by brackets "[]" replaced with your own identifying 17 * information: Portions Copyright [yyyy] [name of copyright owner] 18 * 19 * CDDL HEADER END 20 */ 21 22 /* 23 * Copyright 2006 Sun Microsystems, Inc. All rights reserved. 24 * Use is subject to license terms. 25 */ 26 27 #pragma ident "%Z%%M% %I% %E% SMI" 28 29 /* 30 * Page retirement can be an extended process due to the fact that a retirement 31 * may not be possible when the original request is made. The kernel will 32 * repeatedly attempt to retire a given page, but will not let us know when the 33 * page has been retired. We therefore have to poll to see if the retirement 34 * has been completed. This poll is implemented with a bounded exponential 35 * backoff to reduce the burden which we impose upon the system. 36 * 37 * To reduce the burden on fmd in the face of retirement storms, we schedule 38 * all retries as a group. In the simplest case, we attempt to retire a single 39 * page. When forced to retry, we initially schedule a retry at a configurable 40 * interval t. If the retry fails, we schedule another at 2 * t, and so on, 41 * until t reaches the maximum interval (also configurable). Future retries 42 * for that page will occur with t equal to the maximum interval value. We 43 * will never give up on a retirement. 44 * 45 * With multiple retirements, the situation gets slightly more complicated. As 46 * indicated above, we schedule retries as a group. We don't want to deny new 47 * pages their short retry intervals, so we'll (re)set the retry interval to the 48 * value appropriate for the newest page. 49 */ 50 51 #include <cma.h> 52 53 #include <time.h> 54 #include <fcntl.h> 55 #include <errno.h> 56 #include <unistd.h> 57 #include <strings.h> 58 #include <fm/fmd_api.h> 59 #include <fm/libtopo.h> 60 #include <sys/fm/protocol.h> 61 #include <sys/mem.h> 62 63 static int 64 cma_page_cmd(fmd_hdl_t *hdl, int cmd, nvlist_t *nvl) 65 { 66 mem_page_t mpage; 67 char *fmribuf; 68 size_t fmrisz; 69 int fd, rc, err; 70 71 if ((fd = open("/dev/mem", O_RDONLY)) < 0) 72 return (-1); /* errno is set for us */ 73 74 if ((errno = nvlist_size(nvl, &fmrisz, NV_ENCODE_NATIVE)) != 0 || 75 fmrisz > MEM_FMRI_MAX_BUFSIZE || 76 (fmribuf = fmd_hdl_alloc(hdl, fmrisz, FMD_SLEEP)) == NULL) { 77 (void) close(fd); 78 return (-1); /* errno is set for us */ 79 } 80 81 if ((errno = nvlist_pack(nvl, &fmribuf, &fmrisz, 82 NV_ENCODE_NATIVE, 0)) != 0) { 83 fmd_hdl_free(hdl, fmribuf, fmrisz); 84 (void) close(fd); 85 return (-1); /* errno is set for us */ 86 } 87 88 mpage.m_fmri = fmribuf; 89 mpage.m_fmrisz = fmrisz; 90 91 if ((rc = ioctl(fd, cmd, &mpage)) < 0) 92 err = errno; 93 94 fmd_hdl_free(hdl, fmribuf, fmrisz); 95 96 (void) close(fd); 97 98 if (rc < 0) { 99 errno = err; 100 return (-1); 101 } 102 103 return (0); 104 } 105 106 static void 107 cma_page_free(fmd_hdl_t *hdl, cma_page_t *page) 108 { 109 if (page->pg_fmri != NULL) 110 nvlist_free(page->pg_fmri); 111 fmd_hdl_free(hdl, page, sizeof (cma_page_t)); 112 } 113 114 /* 115 * Retire the specified ASRU, referring to a memory page by PA or by DIMM 116 * offset (i.e. the encoded coordinates internal bank, row, and column). 117 * In the initial FMA implementation, fault.memory.page exported an ASRU 118 * with an explicit physical address, which is valid at the initial time of 119 * diagnosis but may not be later following DR, DIMM removal, or interleave 120 * changes. On SPARC, this issue was solved by exporting the DIMM offset 121 * and pushing the entire FMRI to the platform memory controller through 122 * /dev/mem so it can derive the current PA from the DIMM and offset. 123 * On x64, we also use DIMM and offset, but the mem:/// unum string is an 124 * encoded hc:/// FMRI that is then used by the x64 memory controller driver. 125 * At some point these three approaches need to be rationalized: all platforms 126 * should use the same scheme, either with decoding in the kernel or decoding 127 * in userland (i.e. with a libtopo method to compute and update the PA). 128 */ 129 /*ARGSUSED*/ 130 int 131 cma_page_retire(fmd_hdl_t *hdl, nvlist_t *nvl, nvlist_t *asru, const char *uuid) 132 { 133 cma_page_t *page; 134 uint64_t pageaddr; 135 char *unumstr; 136 nvlist_t *asrucp = NULL; 137 138 /* It should already be expanded, but we'll do it again anyway */ 139 if (fmd_nvl_fmri_expand(hdl, asru) < 0) { 140 fmd_hdl_debug(hdl, "failed to expand page asru\n"); 141 cma_stats.bad_flts.fmds_value.ui64++; 142 return (CMA_RA_FAILURE); 143 } 144 145 if (nvlist_lookup_uint64(asru, FM_FMRI_MEM_PHYSADDR, &pageaddr) != 0) { 146 fmd_hdl_debug(hdl, "mem fault missing '%s'\n", 147 FM_FMRI_MEM_PHYSADDR); 148 cma_stats.bad_flts.fmds_value.ui64++; 149 return (CMA_RA_FAILURE); 150 } 151 152 if (!cma.cma_page_doretire) { 153 fmd_hdl_debug(hdl, "suppressed retire of page %llx\n", 154 (u_longlong_t)pageaddr); 155 cma_stats.page_supp.fmds_value.ui64++; 156 return (CMA_RA_FAILURE); 157 } 158 159 if (!fmd_nvl_fmri_present(hdl, asru)) { 160 fmd_hdl_debug(hdl, "page retire overtaken by events\n"); 161 cma_stats.page_nonent.fmds_value.ui64++; 162 return (CMA_RA_SUCCESS); 163 } 164 165 /* 166 * If the unum is an hc fmri string expand it to an fmri and include 167 * that in a modified asru nvlist. 168 */ 169 if (nvlist_lookup_string(asru, FM_FMRI_MEM_UNUM, &unumstr) == 0 && 170 strncmp(unumstr, "hc:/", 4) == 0) { 171 int err; 172 nvlist_t *unumfmri; 173 struct topo_hdl *thp = fmd_hdl_topology(hdl, TOPO_VERSION); 174 175 if (topo_fmri_str2nvl(thp, unumstr, &unumfmri, &err) != 0) { 176 fmd_hdl_debug(hdl, "page retire str2nvl failed: %s\n", 177 topo_strerror(err)); 178 return (CMA_RA_FAILURE); 179 } 180 181 if (nvlist_dup(asru, &asrucp, 0) != 0) { 182 fmd_hdl_debug(hdl, "page retire nvlist dup failed\n"); 183 nvlist_free(unumfmri); 184 return (CMA_RA_FAILURE); 185 } 186 187 if (nvlist_add_nvlist(asrucp, FM_FMRI_MEM_UNUM "-fmri", 188 unumfmri) != 0) { 189 fmd_hdl_debug(hdl, "page retire failed to add " 190 "unumfmri to modified asru"); 191 nvlist_free(unumfmri); 192 nvlist_free(asrucp); 193 return (CMA_RA_FAILURE); 194 } 195 nvlist_free(unumfmri); 196 } 197 198 if (cma_page_cmd(hdl, MEM_PAGE_FMRI_RETIRE, 199 asrucp ? asrucp : asru) == 0) { 200 fmd_hdl_debug(hdl, "retired page 0x%llx\n", 201 (u_longlong_t)pageaddr); 202 cma_stats.page_flts.fmds_value.ui64++; 203 if (asrucp) 204 nvlist_free(asrucp); 205 return (CMA_RA_SUCCESS); 206 } else if (errno != EAGAIN) { 207 fmd_hdl_debug(hdl, "retire of page 0x%llx failed, will not " 208 "retry: %s\n", (u_longlong_t)pageaddr, strerror(errno)); 209 if (asrucp) 210 nvlist_free(asrucp); 211 if (uuid != NULL && cma.cma_page_maxretries != 0) 212 return (CMA_RA_SUCCESS); 213 return (CMA_RA_FAILURE); 214 } 215 216 /* 217 * The page didn't immediately retire. We'll need to periodically 218 * check to see if it has been retired. 219 */ 220 fmd_hdl_debug(hdl, "page didn't retire - sleeping\n"); 221 222 page = fmd_hdl_zalloc(hdl, sizeof (cma_page_t), FMD_SLEEP); 223 page->pg_addr = pageaddr; 224 if (asrucp) { 225 page->pg_fmri = asrucp; 226 } else { 227 (void) nvlist_dup(asru, &page->pg_fmri, 0); 228 } 229 if (uuid != NULL) 230 page->pg_uuid = fmd_hdl_strdup(hdl, uuid, FMD_SLEEP); 231 232 page->pg_next = cma.cma_pages; 233 cma.cma_pages = page; 234 235 if (cma.cma_page_timerid != 0) 236 fmd_timer_remove(hdl, cma.cma_page_timerid); 237 238 cma.cma_page_curdelay = cma.cma_page_mindelay; 239 240 cma.cma_page_timerid = 241 fmd_timer_install(hdl, NULL, NULL, cma.cma_page_curdelay); 242 243 return (CMA_RA_FAILURE); 244 } 245 246 static int 247 page_retry(fmd_hdl_t *hdl, cma_page_t *page) 248 { 249 if (page->pg_fmri != NULL && !fmd_nvl_fmri_present(hdl, 250 page->pg_fmri)) { 251 fmd_hdl_debug(hdl, "page retire overtaken by events"); 252 cma_stats.page_nonent.fmds_value.ui64++; 253 254 if (page->pg_uuid != NULL) 255 fmd_case_uuclose(hdl, page->pg_uuid); 256 return (1); /* no longer a page to retire */ 257 } 258 259 if (cma_page_cmd(hdl, MEM_PAGE_FMRI_ISRETIRED, page->pg_fmri) == 0) { 260 fmd_hdl_debug(hdl, "retired page 0x%llx on retry %u\n", 261 page->pg_addr, page->pg_nretries); 262 cma_stats.page_flts.fmds_value.ui64++; 263 264 if (page->pg_uuid != NULL) 265 fmd_case_uuclose(hdl, page->pg_uuid); 266 return (1); /* page retired */ 267 } 268 269 if (errno == EAGAIN) { 270 fmd_hdl_debug(hdl, "scheduling another retry for 0x%llx\n", 271 page->pg_addr); 272 return (0); /* schedule another retry */ 273 } else { 274 if (errno == EIO) { 275 fmd_hdl_debug(hdl, "failed to retry page 0x%llx " 276 "retirement: page isn't scheduled for retirement" 277 "(request made beyond page_retire limit?)\n", 278 page->pg_addr); 279 } else { 280 fmd_hdl_debug(hdl, "failed to retry page 0x%llx " 281 "retirement: %s\n", page->pg_addr, 282 strerror(errno)); 283 } 284 285 if (page->pg_uuid != NULL && cma.cma_page_maxretries != 0) 286 fmd_case_uuclose(hdl, page->pg_uuid); 287 288 cma_stats.page_fails.fmds_value.ui64++; 289 return (1); /* give up */ 290 } 291 } 292 293 void 294 cma_page_retry(fmd_hdl_t *hdl) 295 { 296 cma_page_t **pagep; 297 298 cma.cma_page_timerid = 0; 299 300 fmd_hdl_debug(hdl, "page_retry: timer fired\n"); 301 302 pagep = &cma.cma_pages; 303 while (*pagep != NULL) { 304 cma_page_t *page = *pagep; 305 306 if (page_retry(hdl, page)) { 307 /* 308 * Successful retry or we're giving up - remove from 309 * the list 310 */ 311 *pagep = page->pg_next; 312 313 if (page->pg_uuid != NULL) 314 fmd_hdl_strfree(hdl, page->pg_uuid); 315 316 cma_page_free(hdl, page); 317 } else if (cma.cma_page_maxretries == 0 || 318 page->pg_nretries < cma.cma_page_maxretries) { 319 page->pg_nretries++; 320 pagep = &page->pg_next; 321 } else { 322 /* 323 * Tunable maxretries was set and we reached 324 * the max, so just close the case. 325 */ 326 fmd_hdl_debug(hdl, 327 "giving up page retire 0x%llx on retry %u\n", 328 page->pg_addr, page->pg_nretries); 329 cma_stats.page_retmax.fmds_value.ui64++; 330 331 if (page->pg_uuid != NULL) { 332 fmd_case_uuclose(hdl, page->pg_uuid); 333 fmd_hdl_strfree(hdl, page->pg_uuid); 334 } 335 336 *pagep = page->pg_next; 337 338 cma_page_free(hdl, page); 339 } 340 } 341 342 if (cma.cma_pages == NULL) 343 return; /* no more retirements */ 344 345 /* 346 * We still have retirements that haven't completed. Back the delay 347 * off, and schedule a retry. 348 */ 349 cma.cma_page_curdelay = MIN(cma.cma_page_curdelay * 2, 350 cma.cma_page_maxdelay); 351 352 fmd_hdl_debug(hdl, "scheduled page retirement retry for %llu secs\n", 353 (u_longlong_t)(cma.cma_page_curdelay / NANOSEC)); 354 355 cma.cma_page_timerid = 356 fmd_timer_install(hdl, NULL, NULL, cma.cma_page_curdelay); 357 } 358 359 void 360 cma_page_fini(fmd_hdl_t *hdl) 361 { 362 cma_page_t *page; 363 364 while ((page = cma.cma_pages) != NULL) { 365 cma.cma_pages = page->pg_next; 366 cma_page_free(hdl, page); 367 } 368 } 369