Lines Matching full:lsb
89 * Zero buffer, sufficient for padding LSB entries, that does not span a page
130 * Given a queue and a reserved LSB entry index, compute the LSB *entry id* of
131 * that entry for the queue's private LSB region.
140 * Given a queue and a reserved LSB entry index, compute the LSB *address* of
141 * that entry for the queue's private LSB region.
152 * LSB - Local Storage Block
175 * - Read/write of LSB
231 /* Ignore queues that do not have any LSB access. */ in ccp_hw_attach_queue()
233 device_printf(dev, "Ignoring queue %u with no LSB access\n", in ccp_hw_attach_queue()
408 "mapping to unsupported LSB segment id" },
411 "Data buffer located in a LSB location disallowed by the LSB protection masks; or\n"
420 "A Zlib command attmpted to use the LSB." },
423 "Context located in a LSB location disallowed by the LSB protection masks; or\n"
427 "Key located in a LSB location disallowed by the LSB protection masks; or\n"
451 "A read from the LSB encountered a parity error" },
483 device_printf(sc->dev, "Error: %s (%u) Source: %u Faulting LSB block: %u\n", in ccp_intr_handle_error()
794 * Copy private LSB mask to public registers to enable access to LSB in ccp_hw_attach()
1090 /* Load initial SHA state into LSB */ in ccp_sha()
1152 * Translate CCP internal LSB hash format into a standard hash ouput.
1513 /* Set up passthrough op(s) to copy IV into LSB */ in ccp_do_blkcipher()
1554 /* Store key material into LSB to avoid page boundaries */ in ccp_do_blkcipher()
1894 /* Send IV + initial zero GHASH, key data, and lengths buffer to LSB */ in ccp_gcm()
1967 /* Send just initial IV (not GHASH!) to LSB again */ in ccp_gcm()
2039 db_printf(" lsb public mask lo: 0x%x\n", in db_ccp_show_hw()
2041 db_printf(" lsb public mask hi: 0x%x\n", in db_ccp_show_hw()
2043 db_printf(" lsb private mask lo: 0x%x\n", in db_ccp_show_hw()
2045 db_printf(" lsb private mask hi: 0x%x\n", in db_ccp_show_hw()
2094 db_printf(" Error: %s (%u) Source: %u Faulting LSB block: %u\n", in db_ccp_show_queue_hw()