Lines Matching +full:access +full:- +full:ns

1 // SPDX-License-Identifier: GPL-2.0
14 #include <linux/io-64-nonatomic-lo-hi.h>
36 return !!(ivdev->quirks & VSEC_QUIRK_EARLY_HW); in intel_pmt_is_early_client_hw()
47 return -EFAULT; in pmt_memcpy64_fromio()
66 if (cb && cb->read_telem) in pmt_telem_read_mmio()
67 return cb->read_telem(pdev, guid, buf, off, count); in pmt_telem_read_mmio()
72 /* PUNIT on SPR only supports aligned 64-bit read */ in pmt_telem_read_mmio()
94 return -EINVAL; in intel_pmt_read()
96 if (off >= entry->size) in intel_pmt_read()
99 if (count > entry->size - off) in intel_pmt_read()
100 count = entry->size - off; in intel_pmt_read()
102 count = pmt_telem_read_mmio(entry->pcidev, entry->cb, entry->header.guid, buf, in intel_pmt_read()
103 entry->base, off, count); in intel_pmt_read()
115 unsigned long vsize = vma->vm_end - vma->vm_start; in intel_pmt_mmap()
117 unsigned long phys = entry->base_addr; in intel_pmt_mmap()
121 if (vma->vm_flags & (VM_WRITE | VM_MAYWRITE)) in intel_pmt_mmap()
122 return -EROFS; in intel_pmt_mmap()
124 psize = (PFN_UP(entry->base_addr + entry->size) - pfn) * PAGE_SIZE; in intel_pmt_mmap()
127 return -EINVAL; in intel_pmt_mmap()
130 vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot); in intel_pmt_mmap()
131 if (io_remap_pfn_range(vma, vma->vm_start, pfn, in intel_pmt_mmap()
132 vsize, vma->vm_page_prot)) in intel_pmt_mmap()
133 return -EAGAIN; in intel_pmt_mmap()
143 return sprintf(buf, "0x%x\n", entry->guid); in guid_show()
152 return sprintf(buf, "%zu\n", entry->size); in size_show()
161 return sprintf(buf, "%lu\n", offset_in_page(entry->base_addr)); in offset_show()
176 struct auxiliary_device *auxdev = to_auxiliary_dev(dev->parent); in intel_pmt_attr_visible()
183 if (ivdev->cap_id == ilog2(VSEC_CAP_DISCOVERY)) in intel_pmt_attr_visible()
186 return attr->mode; in intel_pmt_attr_visible()
211 struct pci_dev *pci_dev = ivdev->pcidev; in intel_pmt_populate_entry()
212 struct device *dev = &ivdev->auxdev.dev; in intel_pmt_populate_entry()
213 struct intel_pmt_header *header = &entry->header; in intel_pmt_populate_entry()
219 * For non-local access types the lower 3 bits of base offset in intel_pmt_populate_entry()
223 bir = GET_BIR(header->base_offset); in intel_pmt_populate_entry()
225 /* Local access and BARID only for now */ in intel_pmt_populate_entry()
226 switch (header->access_type) { in intel_pmt_populate_entry()
230 "Unsupported BAR index %d for access type %d\n", in intel_pmt_populate_entry()
231 bir, header->access_type); in intel_pmt_populate_entry()
232 return -EINVAL; in intel_pmt_populate_entry()
238 entry->base_addr = disc_res->end + 1 + header->base_offset; in intel_pmt_populate_entry()
251 entry->base_addr = 0; in intel_pmt_populate_entry()
253 if (disc_res->start >= pci_resource_start(pci_dev, i) && in intel_pmt_populate_entry()
254 (disc_res->start <= pci_resource_end(pci_dev, i))) { in intel_pmt_populate_entry()
255 entry->base_addr = pci_resource_start(pci_dev, i) + in intel_pmt_populate_entry()
256 header->base_offset; in intel_pmt_populate_entry()
259 if (!entry->base_addr) in intel_pmt_populate_entry()
260 return -EINVAL; in intel_pmt_populate_entry()
266 if (ivdev->base_addr) { in intel_pmt_populate_entry()
267 entry->base_addr = ivdev->base_addr + in intel_pmt_populate_entry()
268 GET_ADDRESS(header->base_offset); in intel_pmt_populate_entry()
277 entry->base_addr = pci_resource_start(pci_dev, bir) + in intel_pmt_populate_entry()
278 GET_ADDRESS(header->base_offset); in intel_pmt_populate_entry()
281 dev_err(dev, "Unsupported access type %d\n", in intel_pmt_populate_entry()
282 header->access_type); in intel_pmt_populate_entry()
283 return -EINVAL; in intel_pmt_populate_entry()
286 entry->pcidev = pci_dev; in intel_pmt_populate_entry()
287 entry->guid = header->guid; in intel_pmt_populate_entry()
288 entry->size = header->size; in intel_pmt_populate_entry()
289 entry->cb = ivdev->priv_data; in intel_pmt_populate_entry()
295 struct intel_pmt_namespace *ns, in intel_pmt_dev_register() argument
303 ret = xa_alloc(ns->xa, &entry->devid, entry, PMT_XA_LIMIT, GFP_KERNEL); in intel_pmt_dev_register()
308 "%s%d", ns->name, entry->devid); in intel_pmt_dev_register()
312 ns->name, entry->devid); in intel_pmt_dev_register()
317 entry->kobj = &dev->kobj; in intel_pmt_dev_register()
319 if (entry->attr_grp) { in intel_pmt_dev_register()
320 ret = sysfs_create_group(entry->kobj, entry->attr_grp); in intel_pmt_dev_register()
326 if (!entry->size) in intel_pmt_dev_register()
329 res.start = entry->base_addr; in intel_pmt_dev_register()
330 res.end = res.start + entry->size - 1; in intel_pmt_dev_register()
333 entry->base = devm_ioremap_resource(dev, &res); in intel_pmt_dev_register()
334 if (IS_ERR(entry->base)) { in intel_pmt_dev_register()
335 ret = PTR_ERR(entry->base); in intel_pmt_dev_register()
339 sysfs_bin_attr_init(&entry->pmt_bin_attr); in intel_pmt_dev_register()
340 entry->pmt_bin_attr.attr.name = ns->name; in intel_pmt_dev_register()
341 entry->pmt_bin_attr.attr.mode = 0440; in intel_pmt_dev_register()
342 entry->pmt_bin_attr.mmap = intel_pmt_mmap; in intel_pmt_dev_register()
343 entry->pmt_bin_attr.read = intel_pmt_read; in intel_pmt_dev_register()
344 entry->pmt_bin_attr.size = entry->size; in intel_pmt_dev_register()
346 ret = sysfs_create_bin_file(&dev->kobj, &entry->pmt_bin_attr); in intel_pmt_dev_register()
350 if (ns->pmt_add_endpoint) { in intel_pmt_dev_register()
351 ret = ns->pmt_add_endpoint(ivdev, entry); in intel_pmt_dev_register()
359 sysfs_remove_bin_file(entry->kobj, &entry->pmt_bin_attr); in intel_pmt_dev_register()
361 if (entry->attr_grp) in intel_pmt_dev_register()
362 sysfs_remove_group(entry->kobj, entry->attr_grp); in intel_pmt_dev_register()
366 xa_erase(ns->xa, entry->devid); in intel_pmt_dev_register()
371 int intel_pmt_dev_create(struct intel_pmt_entry *entry, struct intel_pmt_namespace *ns, in intel_pmt_dev_create() argument
374 struct device *dev = &intel_vsec_dev->auxdev.dev; in intel_pmt_dev_create()
378 disc_res = &intel_vsec_dev->resource[idx]; in intel_pmt_dev_create()
380 entry->disc_table = devm_ioremap_resource(dev, disc_res); in intel_pmt_dev_create()
381 if (IS_ERR(entry->disc_table)) in intel_pmt_dev_create()
382 return PTR_ERR(entry->disc_table); in intel_pmt_dev_create()
384 ret = ns->pmt_header_decode(entry, dev); in intel_pmt_dev_create()
392 return intel_pmt_dev_register(entry, ns, dev); in intel_pmt_dev_create()
397 struct intel_pmt_namespace *ns) in intel_pmt_dev_destroy() argument
399 struct device *dev = kobj_to_dev(entry->kobj); in intel_pmt_dev_destroy()
401 if (entry->size) in intel_pmt_dev_destroy()
402 sysfs_remove_bin_file(entry->kobj, &entry->pmt_bin_attr); in intel_pmt_dev_destroy()
404 if (entry->attr_grp) in intel_pmt_dev_destroy()
405 sysfs_remove_group(entry->kobj, entry->attr_grp); in intel_pmt_dev_destroy()
408 xa_erase(ns->xa, entry->devid); in intel_pmt_dev_destroy()