blob: b7a09b21ab36f026317546c1850e283e951f88a5 [file] [log] [blame]
// SPDX-License-Identifier: GPL-2.0+
// Copyright 2019 IBM Corp.
#include <linux/idr.h>
#include "ocxl_internal.h"
static struct ocxl_fn *ocxl_fn_get(struct ocxl_fn *fn)
{
return (get_device(&fn->dev) == NULL) ? NULL : fn;
}
static void ocxl_fn_put(struct ocxl_fn *fn)
{
put_device(&fn->dev);
}
static struct ocxl_afu *alloc_afu(struct ocxl_fn *fn)
{
struct ocxl_afu *afu;
afu = kzalloc(sizeof(struct ocxl_afu), GFP_KERNEL);
if (!afu)
return NULL;
kref_init(&afu->kref);
mutex_init(&afu->contexts_lock);
mutex_init(&afu->afu_control_lock);
idr_init(&afu->contexts_idr);
afu->fn = fn;
ocxl_fn_get(fn);
return afu;
}
static void free_afu(struct kref *kref)
{
struct ocxl_afu *afu = container_of(kref, struct ocxl_afu, kref);
idr_destroy(&afu->contexts_idr);
ocxl_fn_put(afu->fn);
kfree(afu);
}
void ocxl_afu_get(struct ocxl_afu *afu)
{
kref_get(&afu->kref);
}
EXPORT_SYMBOL_GPL(ocxl_afu_get);
void ocxl_afu_put(struct ocxl_afu *afu)
{
kref_put(&afu->kref, free_afu);
}
EXPORT_SYMBOL_GPL(ocxl_afu_put);
static int assign_afu_actag(struct ocxl_afu *afu)
{
struct ocxl_fn *fn = afu->fn;
int actag_count, actag_offset;
struct pci_dev *pci_dev = to_pci_dev(fn->dev.parent);
/*
* if there were not enough actags for the function, each afu
* reduces its count as well
*/
actag_count = afu->config.actag_supported *
fn->actag_enabled / fn->actag_supported;
actag_offset = ocxl_actag_afu_alloc(fn, actag_count);
if (actag_offset < 0) {
dev_err(&pci_dev->dev, "Can't allocate %d actags for AFU: %d\n",
actag_count, actag_offset);
return actag_offset;
}
afu->actag_base = fn->actag_base + actag_offset;
afu->actag_enabled = actag_count;
ocxl_config_set_afu_actag(pci_dev, afu->config.dvsec_afu_control_pos,
afu->actag_base, afu->actag_enabled);
dev_dbg(&pci_dev->dev, "actag base=%d enabled=%d\n",
afu->actag_base, afu->actag_enabled);
return 0;
}
static void reclaim_afu_actag(struct ocxl_afu *afu)
{
struct ocxl_fn *fn = afu->fn;
int start_offset, size;
start_offset = afu->actag_base - fn->actag_base;
size = afu->actag_enabled;
ocxl_actag_afu_free(afu->fn, start_offset, size);
}
static int assign_afu_pasid(struct ocxl_afu *afu)
{
struct ocxl_fn *fn = afu->fn;
int pasid_count, pasid_offset;
struct pci_dev *pci_dev = to_pci_dev(fn->dev.parent);
/*
* We only support the case where the function configuration
* requested enough PASIDs to cover all AFUs.
*/
pasid_count = 1 << afu->config.pasid_supported_log;
pasid_offset = ocxl_pasid_afu_alloc(fn, pasid_count);
if (pasid_offset < 0) {
dev_err(&pci_dev->dev, "Can't allocate %d PASIDs for AFU: %d\n",
pasid_count, pasid_offset);
return pasid_offset;
}
afu->pasid_base = fn->pasid_base + pasid_offset;
afu->pasid_count = 0;
afu->pasid_max = pasid_count;
ocxl_config_set_afu_pasid(pci_dev, afu->config.dvsec_afu_control_pos,
afu->pasid_base,
afu->config.pasid_supported_log);
dev_dbg(&pci_dev->dev, "PASID base=%d, enabled=%d\n",
afu->pasid_base, pasid_count);
return 0;
}
static void reclaim_afu_pasid(struct ocxl_afu *afu)
{
struct ocxl_fn *fn = afu->fn;
int start_offset, size;
start_offset = afu->pasid_base - fn->pasid_base;
size = 1 << afu->config.pasid_supported_log;
ocxl_pasid_afu_free(afu->fn, start_offset, size);
}
static int reserve_fn_bar(struct ocxl_fn *fn, int bar)
{
struct pci_dev *dev = to_pci_dev(fn->dev.parent);
int rc, idx;
if (bar != 0 && bar != 2 && bar != 4)
return -EINVAL;
idx = bar >> 1;
if (fn->bar_used[idx]++ == 0) {
rc = pci_request_region(dev, bar, "ocxl");
if (rc)
return rc;
}
return 0;
}
static void release_fn_bar(struct ocxl_fn *fn, int bar)
{
struct pci_dev *dev = to_pci_dev(fn->dev.parent);
int idx;
if (bar != 0 && bar != 2 && bar != 4)
return;
idx = bar >> 1;
if (--fn->bar_used[idx] == 0)
pci_release_region(dev, bar);
WARN_ON(fn->bar_used[idx] < 0);
}
static int map_mmio_areas(struct ocxl_afu *afu)
{
int rc;
struct pci_dev *pci_dev = to_pci_dev(afu->fn->dev.parent);
rc = reserve_fn_bar(afu->fn, afu->config.global_mmio_bar);
if (rc)
return rc;
rc = reserve_fn_bar(afu->fn, afu->config.pp_mmio_bar);
if (rc) {
release_fn_bar(afu->fn, afu->config.global_mmio_bar);
return rc;
}
afu->global_mmio_start =
pci_resource_start(pci_dev, afu->config.global_mmio_bar) +
afu->config.global_mmio_offset;
afu->pp_mmio_start =
pci_resource_start(pci_dev, afu->config.pp_mmio_bar) +
afu->config.pp_mmio_offset;
afu->global_mmio_ptr = ioremap(afu->global_mmio_start,
afu->config.global_mmio_size);
if (!afu->global_mmio_ptr) {
release_fn_bar(afu->fn, afu->config.pp_mmio_bar);
release_fn_bar(afu->fn, afu->config.global_mmio_bar);
dev_err(&pci_dev->dev, "Error mapping global mmio area\n");
return -ENOMEM;
}
/*
* Leave an empty page between the per-process mmio area and
* the AFU interrupt mappings
*/
afu->irq_base_offset = afu->config.pp_mmio_stride + PAGE_SIZE;
return 0;
}
static void unmap_mmio_areas(struct ocxl_afu *afu)
{
if (afu->global_mmio_ptr) {
iounmap(afu->global_mmio_ptr);
afu->global_mmio_ptr = NULL;
}
afu->global_mmio_start = 0;
afu->pp_mmio_start = 0;
release_fn_bar(afu->fn, afu->config.pp_mmio_bar);
release_fn_bar(afu->fn, afu->config.global_mmio_bar);
}
static int configure_afu(struct ocxl_afu *afu, u8 afu_idx, struct pci_dev *dev)
{
int rc;
rc = ocxl_config_read_afu(dev, &afu->fn->config, &afu->config, afu_idx);
if (rc)
return rc;
rc = assign_afu_actag(afu);
if (rc)
return rc;
rc = assign_afu_pasid(afu);
if (rc)
goto err_free_actag;
rc = map_mmio_areas(afu);
if (rc)
goto err_free_pasid;
return 0;
err_free_pasid:
reclaim_afu_pasid(afu);
err_free_actag:
reclaim_afu_actag(afu);
return rc;
}
static void deconfigure_afu(struct ocxl_afu *afu)
{
unmap_mmio_areas(afu);
reclaim_afu_pasid(afu);
reclaim_afu_actag(afu);
}
static int activate_afu(struct pci_dev *dev, struct ocxl_afu *afu)
{
ocxl_config_set_afu_state(dev, afu->config.dvsec_afu_control_pos, 1);
return 0;
}
static void deactivate_afu(struct ocxl_afu *afu)
{
struct pci_dev *dev = to_pci_dev(afu->fn->dev.parent);
ocxl_config_set_afu_state(dev, afu->config.dvsec_afu_control_pos, 0);
}
static int init_afu(struct pci_dev *dev, struct ocxl_fn *fn, u8 afu_idx)
{
int rc;
struct ocxl_afu *afu;
afu = alloc_afu(fn);
if (!afu)
return -ENOMEM;
rc = configure_afu(afu, afu_idx, dev);
if (rc) {
ocxl_afu_put(afu);
return rc;
}
rc = activate_afu(dev, afu);
if (rc) {
deconfigure_afu(afu);
ocxl_afu_put(afu);
return rc;
}
list_add_tail(&afu->list, &fn->afu_list);
return 0;
}
static void remove_afu(struct ocxl_afu *afu)
{
list_del(&afu->list);
ocxl_context_detach_all(afu);
deactivate_afu(afu);
deconfigure_afu(afu);
ocxl_afu_put(afu); // matches the implicit get in alloc_afu
}
static struct ocxl_fn *alloc_function(void)
{
struct ocxl_fn *fn;
fn = kzalloc(sizeof(struct ocxl_fn), GFP_KERNEL);
if (!fn)
return NULL;
INIT_LIST_HEAD(&fn->afu_list);
INIT_LIST_HEAD(&fn->pasid_list);
INIT_LIST_HEAD(&fn->actag_list);
return fn;
}
static void free_function(struct ocxl_fn *fn)
{
WARN_ON(!list_empty(&fn->afu_list));
WARN_ON(!list_empty(&fn->pasid_list));
kfree(fn);
}
static void free_function_dev(struct device *dev)
{
struct ocxl_fn *fn = container_of(dev, struct ocxl_fn, dev);
free_function(fn);
}
static int set_function_device(struct ocxl_fn *fn, struct pci_dev *dev)
{
int rc;
fn->dev.parent = &dev->dev;
fn->dev.release = free_function_dev;
rc = dev_set_name(&fn->dev, "ocxlfn.%s", dev_name(&dev->dev));
if (rc)
return rc;
return 0;
}
static int assign_function_actag(struct ocxl_fn *fn)
{
struct pci_dev *dev = to_pci_dev(fn->dev.parent);
u16 base, enabled, supported;
int rc;
rc = ocxl_config_get_actag_info(dev, &base, &enabled, &supported);
if (rc)
return rc;
fn->actag_base = base;
fn->actag_enabled = enabled;
fn->actag_supported = supported;
ocxl_config_set_actag(dev, fn->config.dvsec_function_pos,
fn->actag_base, fn->actag_enabled);
dev_dbg(&fn->dev, "actag range starting at %d, enabled %d\n",
fn->actag_base, fn->actag_enabled);
return 0;
}
static int set_function_pasid(struct ocxl_fn *fn)
{
struct pci_dev *dev = to_pci_dev(fn->dev.parent);
int rc, desired_count, max_count;
/* A function may not require any PASID */
if (fn->config.max_pasid_log < 0)
return 0;
rc = ocxl_config_get_pasid_info(dev, &max_count);
if (rc)
return rc;
desired_count = 1 << fn->config.max_pasid_log;
if (desired_count > max_count) {
dev_err(&fn->dev,
"Function requires more PASIDs than is available (%d vs. %d)\n",
desired_count, max_count);
return -ENOSPC;
}
fn->pasid_base = 0;
return 0;
}
static int configure_function(struct ocxl_fn *fn, struct pci_dev *dev)
{
int rc;
rc = pci_enable_device(dev);
if (rc) {
dev_err(&dev->dev, "pci_enable_device failed: %d\n", rc);
return rc;
}
/*
* Once it has been confirmed to work on our hardware, we
* should reset the function, to force the adapter to restart
* from scratch.
* A function reset would also reset all its AFUs.
*
* Some hints for implementation:
*
* - there's not status bit to know when the reset is done. We
* should try reading the config space to know when it's
* done.
* - probably something like:
* Reset
* wait 100ms
* issue config read
* allow device up to 1 sec to return success on config
* read before declaring it broken
*
* Some shared logic on the card (CFG, TLX) won't be reset, so
* there's no guarantee that it will be enough.
*/
rc = ocxl_config_read_function(dev, &fn->config);
if (rc)
return rc;
rc = set_function_device(fn, dev);
if (rc)
return rc;
rc = assign_function_actag(fn);
if (rc)
return rc;
rc = set_function_pasid(fn);
if (rc)
return rc;
rc = ocxl_link_setup(dev, 0, &fn->link);
if (rc)
return rc;
rc = ocxl_config_set_TL(dev, fn->config.dvsec_tl_pos);
if (rc) {
ocxl_link_release(dev, fn->link);
return rc;
}
return 0;
}
static void deconfigure_function(struct ocxl_fn *fn)
{
struct pci_dev *dev = to_pci_dev(fn->dev.parent);
ocxl_link_release(dev, fn->link);
pci_disable_device(dev);
}
static struct ocxl_fn *init_function(struct pci_dev *dev)
{
struct ocxl_fn *fn;
int rc;
fn = alloc_function();
if (!fn)
return ERR_PTR(-ENOMEM);
rc = configure_function(fn, dev);
if (rc) {
free_function(fn);
return ERR_PTR(rc);
}
rc = device_register(&fn->dev);
if (rc) {
deconfigure_function(fn);
put_device(&fn->dev);
return ERR_PTR(rc);
}
return fn;
}
// Device detection & initialisation
struct ocxl_fn *ocxl_function_open(struct pci_dev *dev)
{
int rc, afu_count = 0;
u8 afu;
struct ocxl_fn *fn;
if (!radix_enabled()) {
dev_err(&dev->dev, "Unsupported memory model (hash)\n");
return ERR_PTR(-ENODEV);
}
fn = init_function(dev);
if (IS_ERR(fn)) {
dev_err(&dev->dev, "function init failed: %li\n",
PTR_ERR(fn));
return fn;
}
for (afu = 0; afu <= fn->config.max_afu_index; afu++) {
rc = ocxl_config_check_afu_index(dev, &fn->config, afu);
if (rc > 0) {
rc = init_afu(dev, fn, afu);
if (rc) {
dev_err(&dev->dev,
"Can't initialize AFU index %d\n", afu);
continue;
}
afu_count++;
}
}
dev_info(&dev->dev, "%d AFU(s) configured\n", afu_count);
return fn;
}
EXPORT_SYMBOL_GPL(ocxl_function_open);
struct list_head *ocxl_function_afu_list(struct ocxl_fn *fn)
{
return &fn->afu_list;
}
EXPORT_SYMBOL_GPL(ocxl_function_afu_list);
struct ocxl_afu *ocxl_function_fetch_afu(struct ocxl_fn *fn, u8 afu_idx)
{
struct ocxl_afu *afu;
list_for_each_entry(afu, &fn->afu_list, list) {
if (afu->config.idx == afu_idx)
return afu;
}
return NULL;
}
EXPORT_SYMBOL_GPL(ocxl_function_fetch_afu);
const struct ocxl_fn_config *ocxl_function_config(struct ocxl_fn *fn)
{
return &fn->config;
}
EXPORT_SYMBOL_GPL(ocxl_function_config);
void ocxl_function_close(struct ocxl_fn *fn)
{
struct ocxl_afu *afu, *tmp;
list_for_each_entry_safe(afu, tmp, &fn->afu_list, list) {
remove_afu(afu);
}
deconfigure_function(fn);
device_unregister(&fn->dev);
}
EXPORT_SYMBOL_GPL(ocxl_function_close);
// AFU Metadata
struct ocxl_afu_config *ocxl_afu_config(struct ocxl_afu *afu)
{
return &afu->config;
}
EXPORT_SYMBOL_GPL(ocxl_afu_config);
void ocxl_afu_set_private(struct ocxl_afu *afu, void *private)
{
afu->private = private;
}
EXPORT_SYMBOL_GPL(ocxl_afu_set_private);
void *ocxl_afu_get_private(struct ocxl_afu *afu)
{
if (afu)
return afu->private;
return NULL;
}
EXPORT_SYMBOL_GPL(ocxl_afu_get_private);