112f080e7Smrj /* 212f080e7Smrj * CDDL HEADER START 312f080e7Smrj * 412f080e7Smrj * The contents of this file are subject to the terms of the 512f080e7Smrj * Common Development and Distribution License, Version 1.0 only 612f080e7Smrj * (the "License"). You may not use this file except in compliance 712f080e7Smrj * with the License. 812f080e7Smrj * 912f080e7Smrj * You can obtain a copy of the license at usr/src/OPENSOLARIS.LICENSE 1012f080e7Smrj * or http://www.opensolaris.org/os/licensing. 1112f080e7Smrj * See the License for the specific language governing permissions 1212f080e7Smrj * and limitations under the License. 1312f080e7Smrj * 1412f080e7Smrj * When distributing Covered Code, include this CDDL HEADER in each 1512f080e7Smrj * file and include the License file at usr/src/OPENSOLARIS.LICENSE. 1612f080e7Smrj * If applicable, add the following below this CDDL HEADER, with the 1712f080e7Smrj * fields enclosed by brackets "[]" replaced with your own identifying 1812f080e7Smrj * information: Portions Copyright [yyyy] [name of copyright owner] 1912f080e7Smrj * 2012f080e7Smrj * CDDL HEADER END 2112f080e7Smrj */ 2212f080e7Smrj /* 23*7aec1d6eScindi * Copyright 2006 Sun Microsystems, Inc. All rights reserved. 2412f080e7Smrj * Use is subject to license terms. 2512f080e7Smrj */ 2612f080e7Smrj 2712f080e7Smrj #ifndef _SYS_ROOTNEX_H 2812f080e7Smrj #define _SYS_ROOTNEX_H 2912f080e7Smrj 3012f080e7Smrj #pragma ident "%Z%%M% %I% %E% SMI" 3112f080e7Smrj 3212f080e7Smrj /* 3312f080e7Smrj * x86 root nexus implementation specific state 3412f080e7Smrj */ 3512f080e7Smrj 3612f080e7Smrj #include <sys/types.h> 3712f080e7Smrj #include <sys/conf.h> 3812f080e7Smrj #include <sys/modctl.h> 3912f080e7Smrj #include <sys/sunddi.h> 4012f080e7Smrj 4112f080e7Smrj #ifdef __cplusplus 4212f080e7Smrj extern "C" { 4312f080e7Smrj #endif 4412f080e7Smrj 4512f080e7Smrj 4612f080e7Smrj /* size of buffer used for ctlop reportdev */ 4712f080e7Smrj #define REPORTDEV_BUFSIZE 1024 4812f080e7Smrj 4912f080e7Smrj /* min and max interrupt vectors */ 5012f080e7Smrj #define VEC_MIN 1 5112f080e7Smrj #define VEC_MAX 255 5212f080e7Smrj 5312f080e7Smrj /* atomic increment/decrement to keep track of outstanding binds, etc */ 5412f080e7Smrj #define ROOTNEX_PROF_INC(addr) atomic_inc_64(addr) 5512f080e7Smrj #define ROOTNEX_PROF_DEC(addr) atomic_add_64(addr, -1) 5612f080e7Smrj 5712f080e7Smrj /* set in dmac_type to signify that this cookie uses the copy buffer */ 5812f080e7Smrj #define ROOTNEX_USES_COPYBUF 0x80000000 5912f080e7Smrj 6012f080e7Smrj /* 6112f080e7Smrj * integer or boolean property name and value. A few static rootnex properties 6212f080e7Smrj * are created during rootnex attach from an array of rootnex_intprop_t.. 6312f080e7Smrj */ 6412f080e7Smrj typedef struct rootnex_intprop_s { 6512f080e7Smrj char *prop_name; 6612f080e7Smrj int prop_value; 6712f080e7Smrj } rootnex_intprop_t; 6812f080e7Smrj 6912f080e7Smrj /* 7012f080e7Smrj * sgl related information which is visible to rootnex_get_sgl(). Trying to 7112f080e7Smrj * isolate get_sgl() as much as possible so it can be easily replaced. 7212f080e7Smrj */ 7312f080e7Smrj typedef struct rootnex_sglinfo_s { 7412f080e7Smrj /* 7512f080e7Smrj * These are passed into rootnex_get_sgl(). 7612f080e7Smrj * 7712f080e7Smrj * si_min_addr - the minimum physical address 7812f080e7Smrj * si_max_addr - the maximum physical address 7912f080e7Smrj * si_max_cookie_size - the maximum size of a physically contiguous 8012f080e7Smrj * piece of memory that we can handle in a sgl. 8112f080e7Smrj * si_segmask - segment mask to determine if we cross a segment boundary 8212f080e7Smrj * si_max_pages - max number of pages this sgl could occupy (which 8312f080e7Smrj * is also the maximum number of cookies we might see. 8412f080e7Smrj */ 8512f080e7Smrj uint64_t si_min_addr; 8612f080e7Smrj uint64_t si_max_addr; 8712f080e7Smrj uint64_t si_max_cookie_size; 8812f080e7Smrj uint64_t si_segmask; 8912f080e7Smrj uint_t si_max_pages; 9012f080e7Smrj 9112f080e7Smrj /* 9212f080e7Smrj * these are returned by rootnex_get_sgl() 9312f080e7Smrj * 9412f080e7Smrj * si_copybuf_req - amount of copy buffer needed by the buffer. 9512f080e7Smrj * si_buf_offset - The initial offset into the first page of the buffer. 9612f080e7Smrj * It's set in get sgl and used in the bind slow path to help 9712f080e7Smrj * calculate the current page index & offset from the current offset 9812f080e7Smrj * which is relative to the start of the buffer. 9912f080e7Smrj * si_asp - address space of buffer passed in. 10012f080e7Smrj * si_sgl_size - The actual number of cookies in the sgl. This does 10112f080e7Smrj * not reflect and sharing that we might do on window boundaries. 10212f080e7Smrj */ 10312f080e7Smrj size_t si_copybuf_req; 10412f080e7Smrj off_t si_buf_offset; 10512f080e7Smrj struct as *si_asp; 10612f080e7Smrj uint_t si_sgl_size; 10712f080e7Smrj } rootnex_sglinfo_t; 10812f080e7Smrj 10912f080e7Smrj /* 11012f080e7Smrj * When we have to use the copy buffer, we allocate one of these structures per 11112f080e7Smrj * buffer page to track which pages need the copy buffer, what the kernel 11212f080e7Smrj * virtual address is (which the device can't reach), and what the copy buffer 11312f080e7Smrj * virtual address is (where the device dma's to/from). For 32-bit kernels, 11412f080e7Smrj * since we can't use seg kpm, we also need to keep the page_t around and state 11512f080e7Smrj * if we've currently mapped in the page into KVA space for buffers which don't 11612f080e7Smrj * have kva already and when we have multiple windows because we used up all our 11712f080e7Smrj * copy buffer space. 11812f080e7Smrj */ 11912f080e7Smrj typedef struct rootnex_pgmap_s { 12012f080e7Smrj boolean_t pm_uses_copybuf; 12112f080e7Smrj #if !defined(__amd64) 12212f080e7Smrj boolean_t pm_mapped; 12312f080e7Smrj page_t *pm_pp; 12412f080e7Smrj caddr_t pm_vaddr; 12512f080e7Smrj #endif 12612f080e7Smrj caddr_t pm_kaddr; 12712f080e7Smrj caddr_t pm_cbaddr; 12812f080e7Smrj } rootnex_pgmap_t; 12912f080e7Smrj 13012f080e7Smrj /* 13112f080e7Smrj * We only need to trim a buffer when we have multiple windows. Each window has 13212f080e7Smrj * trim state. We might have trimmed the end of the previous window, leaving the 13312f080e7Smrj * first cookie of this window trimmed[tr_trim_first] (which basically means we 13412f080e7Smrj * won't start with a new cookie), or we might need to trim the end of the 13512f080e7Smrj * current window [tr_trim_last] (which basically means we won't end with a 13612f080e7Smrj * complete cookie). We keep the same state for the first & last cookie in a 13712f080e7Smrj * window (a window can have one or more cookies). However, when we trim the 13812f080e7Smrj * last cookie, we keep a pointer to the last cookie in the trim state since we 13912f080e7Smrj * only need this info when we trim. The pointer to the first cookie in the 14012f080e7Smrj * window is in the window state since we need to know what the first cookie in 14112f080e7Smrj * the window is in various places. 14212f080e7Smrj * 14312f080e7Smrj * If we do trim a cookie, we save away the physical address and size of the 14412f080e7Smrj * cookie so that we can over write the cookie when we switch windows (the 14512f080e7Smrj * space for a cookie which is in two windows is shared between the windows. 14612f080e7Smrj * We keep around the same information for the last page in a window. 14712f080e7Smrj * 14812f080e7Smrj * if we happened to trim on a page that uses the copy buffer, and that page 14912f080e7Smrj * is also in the middle of a window boundary because we have filled up the 15012f080e7Smrj * copy buffer, we need to remember the copy buffer address for both windows 15112f080e7Smrj * since the same page will have different copy buffer addresses in the two 15212f080e7Smrj * windows. We need to due the same for kaddr in the 32-bit kernel since we 15312f080e7Smrj * have a limited kva space which we map to. 15412f080e7Smrj */ 15512f080e7Smrj typedef struct rootnex_trim_s { 15612f080e7Smrj boolean_t tr_trim_first; 15712f080e7Smrj boolean_t tr_trim_last; 15812f080e7Smrj ddi_dma_cookie_t *tr_last_cookie; 15912f080e7Smrj uint64_t tr_first_paddr; 16012f080e7Smrj uint64_t tr_last_paddr; 16112f080e7Smrj size_t tr_first_size; 16212f080e7Smrj size_t tr_last_size; 16312f080e7Smrj 16412f080e7Smrj boolean_t tr_first_copybuf_win; 16512f080e7Smrj boolean_t tr_last_copybuf_win; 16612f080e7Smrj uint_t tr_first_pidx; 16712f080e7Smrj uint_t tr_last_pidx; 16812f080e7Smrj caddr_t tr_first_cbaddr; 16912f080e7Smrj caddr_t tr_last_cbaddr; 17012f080e7Smrj #if !defined(__amd64) 17112f080e7Smrj caddr_t tr_first_kaddr; 17212f080e7Smrj caddr_t tr_last_kaddr; 17312f080e7Smrj #endif 17412f080e7Smrj } rootnex_trim_t; 17512f080e7Smrj 17612f080e7Smrj /* 17712f080e7Smrj * per window state. A bound DMA handle can have multiple windows. Each window 17812f080e7Smrj * will have the following state. We track if this window needs to sync, 17912f080e7Smrj * the offset into the buffer where the window starts, the size of the window. 18012f080e7Smrj * a pointer to the first cookie in the window, the number of cookies in the 18112f080e7Smrj * window, and the trim state for the window. For the 32-bit kernel, we keep 18212f080e7Smrj * track of if we need to remap the copy buffer when we switch to a this window 18312f080e7Smrj */ 18412f080e7Smrj typedef struct rootnex_window_s { 18512f080e7Smrj boolean_t wd_dosync; 18612f080e7Smrj uint_t wd_cookie_cnt; 18712f080e7Smrj off_t wd_offset; 18812f080e7Smrj size_t wd_size; 18912f080e7Smrj ddi_dma_cookie_t *wd_first_cookie; 19012f080e7Smrj rootnex_trim_t wd_trim; 19112f080e7Smrj #if !defined(__amd64) 19212f080e7Smrj boolean_t wd_remap_copybuf; 19312f080e7Smrj #endif 19412f080e7Smrj } rootnex_window_t; 19512f080e7Smrj 19612f080e7Smrj /* per dma handle private state */ 19712f080e7Smrj typedef struct rootnex_dma_s { 19812f080e7Smrj /* 19912f080e7Smrj * sgl related state used to build and describe the sgl. 20012f080e7Smrj * 20112f080e7Smrj * dp_partial_required - used in the bind slow path to identify if we 20212f080e7Smrj * need to do a partial mapping or not. 20312f080e7Smrj * dp_trim_required - used in the bind slow path to identify if we 20412f080e7Smrj * need to trim when switching to a new window. This should only be 20512f080e7Smrj * set when partial is set. 20612f080e7Smrj * dp_granularity_power_2 - set in alloc handle and used in bind slow 20712f080e7Smrj * path to determine if we & or % to calculate the trim. 20812f080e7Smrj * dp_dma - copy of dma "object" passed in during bind 20912f080e7Smrj * dp_maxxfer - trimmed dma_attr_maxxfer so that it is a whole 21012f080e7Smrj * multiple of granularity 21112f080e7Smrj * dp_sglinfo - See rootnex_sglinfo_t above. 21212f080e7Smrj */ 21312f080e7Smrj boolean_t dp_partial_required; 21412f080e7Smrj boolean_t dp_trim_required; 21512f080e7Smrj boolean_t dp_granularity_power_2; 21612f080e7Smrj uint64_t dp_maxxfer; 21712f080e7Smrj ddi_dma_obj_t dp_dma; 21812f080e7Smrj rootnex_sglinfo_t dp_sglinfo; 21912f080e7Smrj 22012f080e7Smrj /* 22112f080e7Smrj * Copy buffer related state 22212f080e7Smrj * 22312f080e7Smrj * dp_copybuf_size - the actual size of the copy buffer that we are 22412f080e7Smrj * using. This can be smaller that dp_copybuf_req, i.e. bind size > 22512f080e7Smrj * max copy buffer size. 22612f080e7Smrj * dp_cbaddr - kernel address of copy buffer. Used to determine where 22712f080e7Smrj * where to copy to/from. 22812f080e7Smrj * dp_cbsize - the "real" size returned from the copy buffer alloc. 22912f080e7Smrj * Set in the copybuf alloc and used to free copybuf. 23012f080e7Smrj * dp_pgmap - page map used in sync to determine which pages in the 23112f080e7Smrj * buffer use the copy buffer and what addresses to use to copy to/ 23212f080e7Smrj * from. 23312f080e7Smrj * dp_cb_remaping - status if this bind causes us to have to remap 23412f080e7Smrj * the copybuf when switching to new windows. This is only used in 23512f080e7Smrj * the 32-bit kernel since we use seg kpm in the 64-bit kernel for 23612f080e7Smrj * this case. 23712f080e7Smrj * dp_kva - kernel heap arena vmem space for mapping to buffers which 23812f080e7Smrj * we don't have a kernel VA to bcopy to/from. This is only used in 23912f080e7Smrj * the 32-bit kernel since we use seg kpm in the 64-bit kernel for 24012f080e7Smrj * this case. 24112f080e7Smrj */ 24212f080e7Smrj size_t dp_copybuf_size; 24312f080e7Smrj caddr_t dp_cbaddr; 24412f080e7Smrj size_t dp_cbsize; 24512f080e7Smrj rootnex_pgmap_t *dp_pgmap; 24612f080e7Smrj #if !defined(__amd64) 24712f080e7Smrj boolean_t dp_cb_remaping; 24812f080e7Smrj caddr_t dp_kva; 24912f080e7Smrj #endif 25012f080e7Smrj 25112f080e7Smrj /* 25212f080e7Smrj * window related state. The pointer to the window state array which may 25312f080e7Smrj * be a pointer into the pre allocated state, or we may have had to 25412f080e7Smrj * allocate the window array on the fly because it wouldn't fit. If 25512f080e7Smrj * we allocate it, we'll use dp_need_to_free_window and dp_window_size 25612f080e7Smrj * during cleanup. dp_current_win keeps track of the current window. 25712f080e7Smrj * dp_max_win is the maximum number of windows we could have. 25812f080e7Smrj */ 25912f080e7Smrj uint_t dp_current_win; 26012f080e7Smrj rootnex_window_t *dp_window; 26112f080e7Smrj boolean_t dp_need_to_free_window; 26212f080e7Smrj uint_t dp_window_size; 26312f080e7Smrj uint_t dp_max_win; 26412f080e7Smrj 26512f080e7Smrj /* dip of driver which "owns" handle. set to rdip in alloc_handle() */ 26612f080e7Smrj dev_info_t *dp_dip; 26712f080e7Smrj 26812f080e7Smrj /* 26912f080e7Smrj * dp_mutex and dp_inuse are only used to see if a driver is trying to 27012f080e7Smrj * bind to an already bound dma handle. dp_mutex only used for dp_inuse 27112f080e7Smrj */ 27212f080e7Smrj kmutex_t dp_mutex; 27312f080e7Smrj boolean_t dp_inuse; 27412f080e7Smrj 27512f080e7Smrj /* 27612f080e7Smrj * cookie related state. The pointer to the cookies (dp_cookies) may 27712f080e7Smrj * be a pointer into the pre allocated state, or we may have had to 27812f080e7Smrj * allocate the cookie array on the fly because it wouldn't fit. If 27912f080e7Smrj * we allocate it, we'll use dp_need_to_free_cookie and dp_cookie_size 28012f080e7Smrj * during cleanup. dp_current_cookie is only used in the obsoleted 28112f080e7Smrj * interfaces to determine when we've used up all the cookies in a 28212f080e7Smrj * window during nextseg().. 28312f080e7Smrj */ 28412f080e7Smrj size_t dp_cookie_size; 28512f080e7Smrj ddi_dma_cookie_t *dp_cookies; 28612f080e7Smrj boolean_t dp_need_to_free_cookie; 28712f080e7Smrj uint_t dp_current_cookie; /* for obsoleted I/Fs */ 28812f080e7Smrj 28912f080e7Smrj /* 29012f080e7Smrj * pre allocated space for the bind state, allocated during alloc 29112f080e7Smrj * handle. For a lot of devices, this will save us from having to do 29212f080e7Smrj * kmem_alloc's during the bind most of the time. kmem_alloc's can be 29312f080e7Smrj * expensive on x86 when the cpu count goes up since xcalls are 29412f080e7Smrj * expensive on x86. 29512f080e7Smrj */ 29612f080e7Smrj uchar_t *dp_prealloc_buffer; 29712f080e7Smrj } rootnex_dma_t; 29812f080e7Smrj 29912f080e7Smrj /* 30012f080e7Smrj * profile/performance counters. Most things will be dtrace probes, but there 30112f080e7Smrj * are a couple of things we want to keep track all the time. We track the 30212f080e7Smrj * total number of active handles and binds (i.e. an alloc without a free or 30312f080e7Smrj * a bind without an unbind) since rootnex attach. We also track the total 30412f080e7Smrj * number of binds which have failed since rootnex attach. 30512f080e7Smrj */ 30612f080e7Smrj typedef enum { 30712f080e7Smrj ROOTNEX_CNT_ACTIVE_HDLS = 0, 30812f080e7Smrj ROOTNEX_CNT_ACTIVE_BINDS = 1, 30912f080e7Smrj ROOTNEX_CNT_ALLOC_FAIL = 2, 31012f080e7Smrj ROOTNEX_CNT_BIND_FAIL = 3, 31112f080e7Smrj ROOTNEX_CNT_SYNC_FAIL = 4, 31212f080e7Smrj ROOTNEX_CNT_GETWIN_FAIL = 5, 31312f080e7Smrj 31412f080e7Smrj /* This one must be last */ 31512f080e7Smrj ROOTNEX_CNT_LAST 31612f080e7Smrj } rootnex_cnt_t; 31712f080e7Smrj 31812f080e7Smrj /* 31912f080e7Smrj * global driver state. 32012f080e7Smrj * r_dmahdl_cache - dma_handle kmem_cache 32112f080e7Smrj * r_dvma_call_list_id - ddi_set_callback() id 32212f080e7Smrj * r_peekpoke_mutex - serialize peeks and pokes. 32312f080e7Smrj * r_dip - rootnex dip 32412f080e7Smrj * r_reserved_msg_printed - ctlops reserve message threshold 32512f080e7Smrj * r_counters - profile/performance counters 32612f080e7Smrj */ 32712f080e7Smrj typedef struct rootnex_state_s { 32812f080e7Smrj uint_t r_prealloc_cookies; 32912f080e7Smrj uint_t r_prealloc_size; 33012f080e7Smrj kmem_cache_t *r_dmahdl_cache; 33112f080e7Smrj uintptr_t r_dvma_call_list_id; 33212f080e7Smrj kmutex_t r_peekpoke_mutex; 33312f080e7Smrj dev_info_t *r_dip; 334*7aec1d6eScindi ddi_iblock_cookie_t r_err_ibc; 33512f080e7Smrj boolean_t r_reserved_msg_printed; 33612f080e7Smrj uint64_t r_counters[ROOTNEX_CNT_LAST]; 33712f080e7Smrj } rootnex_state_t; 33812f080e7Smrj 33912f080e7Smrj 34012f080e7Smrj #ifdef __cplusplus 34112f080e7Smrj } 34212f080e7Smrj #endif 34312f080e7Smrj 34412f080e7Smrj #endif /* _SYS_ROOTNEX_H */ 345