mirror of
https://github.com/torvalds/linux.git
synced 2024-11-10 14:11:52 +00:00
8c6eabae38
In nested translation, the stage-1 page table is user-managed but cached by the IOMMU hardware, so an update on present page table entries in the stage-1 page table should be followed with a cache invalidation. Add an IOMMU_HWPT_INVALIDATE ioctl to support such a cache invalidation. It takes hwpt_id to specify the iommu_domain, and a multi-entry array to support multiple invalidation data in one ioctl. enum iommu_hwpt_invalidate_data_type is defined to tag the data type of the entries in the multi-entry array. Link: https://lore.kernel.org/r/20240111041015.47920-3-yi.l.liu@intel.com Reviewed-by: Kevin Tian <kevin.tian@intel.com> Co-developed-by: Nicolin Chen <nicolinc@nvidia.com> Signed-off-by: Nicolin Chen <nicolinc@nvidia.com> Signed-off-by: Yi Liu <yi.l.liu@intel.com> Signed-off-by: Jason Gunthorpe <jgg@nvidia.com>
585 lines
16 KiB
C
585 lines
16 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/* Copyright (C) 2021 Intel Corporation
|
|
* Copyright (c) 2021-2022, NVIDIA CORPORATION & AFFILIATES
|
|
*
|
|
* iommufd provides control over the IOMMU HW objects created by IOMMU kernel
|
|
* drivers. IOMMU HW objects revolve around IO page tables that map incoming DMA
|
|
* addresses (IOVA) to CPU addresses.
|
|
*/
|
|
#define pr_fmt(fmt) "iommufd: " fmt
|
|
|
|
#include <linux/file.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/module.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/miscdevice.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/bug.h>
|
|
#include <uapi/linux/iommufd.h>
|
|
#include <linux/iommufd.h>
|
|
|
|
#include "io_pagetable.h"
|
|
#include "iommufd_private.h"
|
|
#include "iommufd_test.h"
|
|
|
|
struct iommufd_object_ops {
|
|
void (*destroy)(struct iommufd_object *obj);
|
|
void (*abort)(struct iommufd_object *obj);
|
|
};
|
|
static const struct iommufd_object_ops iommufd_object_ops[];
|
|
static struct miscdevice vfio_misc_dev;
|
|
|
|
struct iommufd_object *_iommufd_object_alloc(struct iommufd_ctx *ictx,
|
|
size_t size,
|
|
enum iommufd_object_type type)
|
|
{
|
|
struct iommufd_object *obj;
|
|
int rc;
|
|
|
|
obj = kzalloc(size, GFP_KERNEL_ACCOUNT);
|
|
if (!obj)
|
|
return ERR_PTR(-ENOMEM);
|
|
obj->type = type;
|
|
/* Starts out bias'd by 1 until it is removed from the xarray */
|
|
refcount_set(&obj->shortterm_users, 1);
|
|
refcount_set(&obj->users, 1);
|
|
|
|
/*
|
|
* Reserve an ID in the xarray but do not publish the pointer yet since
|
|
* the caller hasn't initialized it yet. Once the pointer is published
|
|
* in the xarray and visible to other threads we can't reliably destroy
|
|
* it anymore, so the caller must complete all errorable operations
|
|
* before calling iommufd_object_finalize().
|
|
*/
|
|
rc = xa_alloc(&ictx->objects, &obj->id, XA_ZERO_ENTRY,
|
|
xa_limit_31b, GFP_KERNEL_ACCOUNT);
|
|
if (rc)
|
|
goto out_free;
|
|
return obj;
|
|
out_free:
|
|
kfree(obj);
|
|
return ERR_PTR(rc);
|
|
}
|
|
|
|
/*
|
|
* Allow concurrent access to the object.
|
|
*
|
|
* Once another thread can see the object pointer it can prevent object
|
|
* destruction. Expect for special kernel-only objects there is no in-kernel way
|
|
* to reliably destroy a single object. Thus all APIs that are creating objects
|
|
* must use iommufd_object_abort() to handle their errors and only call
|
|
* iommufd_object_finalize() once object creation cannot fail.
|
|
*/
|
|
void iommufd_object_finalize(struct iommufd_ctx *ictx,
|
|
struct iommufd_object *obj)
|
|
{
|
|
void *old;
|
|
|
|
old = xa_store(&ictx->objects, obj->id, obj, GFP_KERNEL);
|
|
/* obj->id was returned from xa_alloc() so the xa_store() cannot fail */
|
|
WARN_ON(old);
|
|
}
|
|
|
|
/* Undo _iommufd_object_alloc() if iommufd_object_finalize() was not called */
|
|
void iommufd_object_abort(struct iommufd_ctx *ictx, struct iommufd_object *obj)
|
|
{
|
|
void *old;
|
|
|
|
old = xa_erase(&ictx->objects, obj->id);
|
|
WARN_ON(old);
|
|
kfree(obj);
|
|
}
|
|
|
|
/*
|
|
* Abort an object that has been fully initialized and needs destroy, but has
|
|
* not been finalized.
|
|
*/
|
|
void iommufd_object_abort_and_destroy(struct iommufd_ctx *ictx,
|
|
struct iommufd_object *obj)
|
|
{
|
|
if (iommufd_object_ops[obj->type].abort)
|
|
iommufd_object_ops[obj->type].abort(obj);
|
|
else
|
|
iommufd_object_ops[obj->type].destroy(obj);
|
|
iommufd_object_abort(ictx, obj);
|
|
}
|
|
|
|
struct iommufd_object *iommufd_get_object(struct iommufd_ctx *ictx, u32 id,
|
|
enum iommufd_object_type type)
|
|
{
|
|
struct iommufd_object *obj;
|
|
|
|
if (iommufd_should_fail())
|
|
return ERR_PTR(-ENOENT);
|
|
|
|
xa_lock(&ictx->objects);
|
|
obj = xa_load(&ictx->objects, id);
|
|
if (!obj || (type != IOMMUFD_OBJ_ANY && obj->type != type) ||
|
|
!iommufd_lock_obj(obj))
|
|
obj = ERR_PTR(-ENOENT);
|
|
xa_unlock(&ictx->objects);
|
|
return obj;
|
|
}
|
|
|
|
static int iommufd_object_dec_wait_shortterm(struct iommufd_ctx *ictx,
|
|
struct iommufd_object *to_destroy)
|
|
{
|
|
if (refcount_dec_and_test(&to_destroy->shortterm_users))
|
|
return 0;
|
|
|
|
if (wait_event_timeout(ictx->destroy_wait,
|
|
refcount_read(&to_destroy->shortterm_users) ==
|
|
0,
|
|
msecs_to_jiffies(10000)))
|
|
return 0;
|
|
|
|
pr_crit("Time out waiting for iommufd object to become free\n");
|
|
refcount_inc(&to_destroy->shortterm_users);
|
|
return -EBUSY;
|
|
}
|
|
|
|
/*
|
|
* Remove the given object id from the xarray if the only reference to the
|
|
* object is held by the xarray.
|
|
*/
|
|
int iommufd_object_remove(struct iommufd_ctx *ictx,
|
|
struct iommufd_object *to_destroy, u32 id,
|
|
unsigned int flags)
|
|
{
|
|
struct iommufd_object *obj;
|
|
XA_STATE(xas, &ictx->objects, id);
|
|
bool zerod_shortterm = false;
|
|
int ret;
|
|
|
|
/*
|
|
* The purpose of the shortterm_users is to ensure deterministic
|
|
* destruction of objects used by external drivers and destroyed by this
|
|
* function. Any temporary increment of the refcount must increment
|
|
* shortterm_users, such as during ioctl execution.
|
|
*/
|
|
if (flags & REMOVE_WAIT_SHORTTERM) {
|
|
ret = iommufd_object_dec_wait_shortterm(ictx, to_destroy);
|
|
if (ret) {
|
|
/*
|
|
* We have a bug. Put back the callers reference and
|
|
* defer cleaning this object until close.
|
|
*/
|
|
refcount_dec(&to_destroy->users);
|
|
return ret;
|
|
}
|
|
zerod_shortterm = true;
|
|
}
|
|
|
|
xa_lock(&ictx->objects);
|
|
obj = xas_load(&xas);
|
|
if (to_destroy) {
|
|
/*
|
|
* If the caller is holding a ref on obj we put it here under
|
|
* the spinlock.
|
|
*/
|
|
refcount_dec(&obj->users);
|
|
|
|
if (WARN_ON(obj != to_destroy)) {
|
|
ret = -ENOENT;
|
|
goto err_xa;
|
|
}
|
|
} else if (xa_is_zero(obj) || !obj) {
|
|
ret = -ENOENT;
|
|
goto err_xa;
|
|
}
|
|
|
|
if (!refcount_dec_if_one(&obj->users)) {
|
|
ret = -EBUSY;
|
|
goto err_xa;
|
|
}
|
|
|
|
xas_store(&xas, NULL);
|
|
if (ictx->vfio_ioas == container_of(obj, struct iommufd_ioas, obj))
|
|
ictx->vfio_ioas = NULL;
|
|
xa_unlock(&ictx->objects);
|
|
|
|
/*
|
|
* Since users is zero any positive users_shortterm must be racing
|
|
* iommufd_put_object(), or we have a bug.
|
|
*/
|
|
if (!zerod_shortterm) {
|
|
ret = iommufd_object_dec_wait_shortterm(ictx, obj);
|
|
if (WARN_ON(ret))
|
|
return ret;
|
|
}
|
|
|
|
iommufd_object_ops[obj->type].destroy(obj);
|
|
kfree(obj);
|
|
return 0;
|
|
|
|
err_xa:
|
|
if (zerod_shortterm) {
|
|
/* Restore the xarray owned reference */
|
|
refcount_set(&obj->shortterm_users, 1);
|
|
}
|
|
xa_unlock(&ictx->objects);
|
|
|
|
/* The returned object reference count is zero */
|
|
return ret;
|
|
}
|
|
|
|
static int iommufd_destroy(struct iommufd_ucmd *ucmd)
|
|
{
|
|
struct iommu_destroy *cmd = ucmd->cmd;
|
|
|
|
return iommufd_object_remove(ucmd->ictx, NULL, cmd->id, 0);
|
|
}
|
|
|
|
static int iommufd_fops_open(struct inode *inode, struct file *filp)
|
|
{
|
|
struct iommufd_ctx *ictx;
|
|
|
|
ictx = kzalloc(sizeof(*ictx), GFP_KERNEL_ACCOUNT);
|
|
if (!ictx)
|
|
return -ENOMEM;
|
|
|
|
/*
|
|
* For compatibility with VFIO when /dev/vfio/vfio is opened we default
|
|
* to the same rlimit accounting as vfio uses.
|
|
*/
|
|
if (IS_ENABLED(CONFIG_IOMMUFD_VFIO_CONTAINER) &&
|
|
filp->private_data == &vfio_misc_dev) {
|
|
ictx->account_mode = IOPT_PAGES_ACCOUNT_MM;
|
|
pr_info_once("IOMMUFD is providing /dev/vfio/vfio, not VFIO.\n");
|
|
}
|
|
|
|
xa_init_flags(&ictx->objects, XA_FLAGS_ALLOC1 | XA_FLAGS_ACCOUNT);
|
|
xa_init(&ictx->groups);
|
|
ictx->file = filp;
|
|
init_waitqueue_head(&ictx->destroy_wait);
|
|
filp->private_data = ictx;
|
|
return 0;
|
|
}
|
|
|
|
static int iommufd_fops_release(struct inode *inode, struct file *filp)
|
|
{
|
|
struct iommufd_ctx *ictx = filp->private_data;
|
|
struct iommufd_object *obj;
|
|
|
|
/*
|
|
* The objects in the xarray form a graph of "users" counts, and we have
|
|
* to destroy them in a depth first manner. Leaf objects will reduce the
|
|
* users count of interior objects when they are destroyed.
|
|
*
|
|
* Repeatedly destroying all the "1 users" leaf objects will progress
|
|
* until the entire list is destroyed. If this can't progress then there
|
|
* is some bug related to object refcounting.
|
|
*/
|
|
while (!xa_empty(&ictx->objects)) {
|
|
unsigned int destroyed = 0;
|
|
unsigned long index;
|
|
|
|
xa_for_each(&ictx->objects, index, obj) {
|
|
if (!refcount_dec_if_one(&obj->users))
|
|
continue;
|
|
destroyed++;
|
|
xa_erase(&ictx->objects, index);
|
|
iommufd_object_ops[obj->type].destroy(obj);
|
|
kfree(obj);
|
|
}
|
|
/* Bug related to users refcount */
|
|
if (WARN_ON(!destroyed))
|
|
break;
|
|
}
|
|
WARN_ON(!xa_empty(&ictx->groups));
|
|
kfree(ictx);
|
|
return 0;
|
|
}
|
|
|
|
static int iommufd_option(struct iommufd_ucmd *ucmd)
|
|
{
|
|
struct iommu_option *cmd = ucmd->cmd;
|
|
int rc;
|
|
|
|
if (cmd->__reserved)
|
|
return -EOPNOTSUPP;
|
|
|
|
switch (cmd->option_id) {
|
|
case IOMMU_OPTION_RLIMIT_MODE:
|
|
rc = iommufd_option_rlimit_mode(cmd, ucmd->ictx);
|
|
break;
|
|
case IOMMU_OPTION_HUGE_PAGES:
|
|
rc = iommufd_ioas_option(ucmd);
|
|
break;
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
if (rc)
|
|
return rc;
|
|
if (copy_to_user(&((struct iommu_option __user *)ucmd->ubuffer)->val64,
|
|
&cmd->val64, sizeof(cmd->val64)))
|
|
return -EFAULT;
|
|
return 0;
|
|
}
|
|
|
|
union ucmd_buffer {
|
|
struct iommu_destroy destroy;
|
|
struct iommu_hw_info info;
|
|
struct iommu_hwpt_alloc hwpt;
|
|
struct iommu_hwpt_get_dirty_bitmap get_dirty_bitmap;
|
|
struct iommu_hwpt_invalidate cache;
|
|
struct iommu_hwpt_set_dirty_tracking set_dirty_tracking;
|
|
struct iommu_ioas_alloc alloc;
|
|
struct iommu_ioas_allow_iovas allow_iovas;
|
|
struct iommu_ioas_copy ioas_copy;
|
|
struct iommu_ioas_iova_ranges iova_ranges;
|
|
struct iommu_ioas_map map;
|
|
struct iommu_ioas_unmap unmap;
|
|
struct iommu_option option;
|
|
struct iommu_vfio_ioas vfio_ioas;
|
|
#ifdef CONFIG_IOMMUFD_TEST
|
|
struct iommu_test_cmd test;
|
|
#endif
|
|
};
|
|
|
|
struct iommufd_ioctl_op {
|
|
unsigned int size;
|
|
unsigned int min_size;
|
|
unsigned int ioctl_num;
|
|
int (*execute)(struct iommufd_ucmd *ucmd);
|
|
};
|
|
|
|
#define IOCTL_OP(_ioctl, _fn, _struct, _last) \
|
|
[_IOC_NR(_ioctl) - IOMMUFD_CMD_BASE] = { \
|
|
.size = sizeof(_struct) + \
|
|
BUILD_BUG_ON_ZERO(sizeof(union ucmd_buffer) < \
|
|
sizeof(_struct)), \
|
|
.min_size = offsetofend(_struct, _last), \
|
|
.ioctl_num = _ioctl, \
|
|
.execute = _fn, \
|
|
}
|
|
static const struct iommufd_ioctl_op iommufd_ioctl_ops[] = {
|
|
IOCTL_OP(IOMMU_DESTROY, iommufd_destroy, struct iommu_destroy, id),
|
|
IOCTL_OP(IOMMU_GET_HW_INFO, iommufd_get_hw_info, struct iommu_hw_info,
|
|
__reserved),
|
|
IOCTL_OP(IOMMU_HWPT_ALLOC, iommufd_hwpt_alloc, struct iommu_hwpt_alloc,
|
|
__reserved),
|
|
IOCTL_OP(IOMMU_HWPT_GET_DIRTY_BITMAP, iommufd_hwpt_get_dirty_bitmap,
|
|
struct iommu_hwpt_get_dirty_bitmap, data),
|
|
IOCTL_OP(IOMMU_HWPT_INVALIDATE, iommufd_hwpt_invalidate,
|
|
struct iommu_hwpt_invalidate, __reserved),
|
|
IOCTL_OP(IOMMU_HWPT_SET_DIRTY_TRACKING, iommufd_hwpt_set_dirty_tracking,
|
|
struct iommu_hwpt_set_dirty_tracking, __reserved),
|
|
IOCTL_OP(IOMMU_IOAS_ALLOC, iommufd_ioas_alloc_ioctl,
|
|
struct iommu_ioas_alloc, out_ioas_id),
|
|
IOCTL_OP(IOMMU_IOAS_ALLOW_IOVAS, iommufd_ioas_allow_iovas,
|
|
struct iommu_ioas_allow_iovas, allowed_iovas),
|
|
IOCTL_OP(IOMMU_IOAS_COPY, iommufd_ioas_copy, struct iommu_ioas_copy,
|
|
src_iova),
|
|
IOCTL_OP(IOMMU_IOAS_IOVA_RANGES, iommufd_ioas_iova_ranges,
|
|
struct iommu_ioas_iova_ranges, out_iova_alignment),
|
|
IOCTL_OP(IOMMU_IOAS_MAP, iommufd_ioas_map, struct iommu_ioas_map,
|
|
iova),
|
|
IOCTL_OP(IOMMU_IOAS_UNMAP, iommufd_ioas_unmap, struct iommu_ioas_unmap,
|
|
length),
|
|
IOCTL_OP(IOMMU_OPTION, iommufd_option, struct iommu_option,
|
|
val64),
|
|
IOCTL_OP(IOMMU_VFIO_IOAS, iommufd_vfio_ioas, struct iommu_vfio_ioas,
|
|
__reserved),
|
|
#ifdef CONFIG_IOMMUFD_TEST
|
|
IOCTL_OP(IOMMU_TEST_CMD, iommufd_test, struct iommu_test_cmd, last),
|
|
#endif
|
|
};
|
|
|
|
static long iommufd_fops_ioctl(struct file *filp, unsigned int cmd,
|
|
unsigned long arg)
|
|
{
|
|
struct iommufd_ctx *ictx = filp->private_data;
|
|
const struct iommufd_ioctl_op *op;
|
|
struct iommufd_ucmd ucmd = {};
|
|
union ucmd_buffer buf;
|
|
unsigned int nr;
|
|
int ret;
|
|
|
|
nr = _IOC_NR(cmd);
|
|
if (nr < IOMMUFD_CMD_BASE ||
|
|
(nr - IOMMUFD_CMD_BASE) >= ARRAY_SIZE(iommufd_ioctl_ops))
|
|
return iommufd_vfio_ioctl(ictx, cmd, arg);
|
|
|
|
ucmd.ictx = ictx;
|
|
ucmd.ubuffer = (void __user *)arg;
|
|
ret = get_user(ucmd.user_size, (u32 __user *)ucmd.ubuffer);
|
|
if (ret)
|
|
return ret;
|
|
|
|
op = &iommufd_ioctl_ops[nr - IOMMUFD_CMD_BASE];
|
|
if (op->ioctl_num != cmd)
|
|
return -ENOIOCTLCMD;
|
|
if (ucmd.user_size < op->min_size)
|
|
return -EINVAL;
|
|
|
|
ucmd.cmd = &buf;
|
|
ret = copy_struct_from_user(ucmd.cmd, op->size, ucmd.ubuffer,
|
|
ucmd.user_size);
|
|
if (ret)
|
|
return ret;
|
|
ret = op->execute(&ucmd);
|
|
return ret;
|
|
}
|
|
|
|
static const struct file_operations iommufd_fops = {
|
|
.owner = THIS_MODULE,
|
|
.open = iommufd_fops_open,
|
|
.release = iommufd_fops_release,
|
|
.unlocked_ioctl = iommufd_fops_ioctl,
|
|
};
|
|
|
|
/**
|
|
* iommufd_ctx_get - Get a context reference
|
|
* @ictx: Context to get
|
|
*
|
|
* The caller must already hold a valid reference to ictx.
|
|
*/
|
|
void iommufd_ctx_get(struct iommufd_ctx *ictx)
|
|
{
|
|
get_file(ictx->file);
|
|
}
|
|
EXPORT_SYMBOL_NS_GPL(iommufd_ctx_get, IOMMUFD);
|
|
|
|
/**
|
|
* iommufd_ctx_from_file - Acquires a reference to the iommufd context
|
|
* @file: File to obtain the reference from
|
|
*
|
|
* Returns a pointer to the iommufd_ctx, otherwise ERR_PTR. The struct file
|
|
* remains owned by the caller and the caller must still do fput. On success
|
|
* the caller is responsible to call iommufd_ctx_put().
|
|
*/
|
|
struct iommufd_ctx *iommufd_ctx_from_file(struct file *file)
|
|
{
|
|
struct iommufd_ctx *ictx;
|
|
|
|
if (file->f_op != &iommufd_fops)
|
|
return ERR_PTR(-EBADFD);
|
|
ictx = file->private_data;
|
|
iommufd_ctx_get(ictx);
|
|
return ictx;
|
|
}
|
|
EXPORT_SYMBOL_NS_GPL(iommufd_ctx_from_file, IOMMUFD);
|
|
|
|
/**
|
|
* iommufd_ctx_from_fd - Acquires a reference to the iommufd context
|
|
* @fd: File descriptor to obtain the reference from
|
|
*
|
|
* Returns a pointer to the iommufd_ctx, otherwise ERR_PTR. On success
|
|
* the caller is responsible to call iommufd_ctx_put().
|
|
*/
|
|
struct iommufd_ctx *iommufd_ctx_from_fd(int fd)
|
|
{
|
|
struct file *file;
|
|
|
|
file = fget(fd);
|
|
if (!file)
|
|
return ERR_PTR(-EBADF);
|
|
|
|
if (file->f_op != &iommufd_fops) {
|
|
fput(file);
|
|
return ERR_PTR(-EBADFD);
|
|
}
|
|
/* fget is the same as iommufd_ctx_get() */
|
|
return file->private_data;
|
|
}
|
|
EXPORT_SYMBOL_NS_GPL(iommufd_ctx_from_fd, IOMMUFD);
|
|
|
|
/**
|
|
* iommufd_ctx_put - Put back a reference
|
|
* @ictx: Context to put back
|
|
*/
|
|
void iommufd_ctx_put(struct iommufd_ctx *ictx)
|
|
{
|
|
fput(ictx->file);
|
|
}
|
|
EXPORT_SYMBOL_NS_GPL(iommufd_ctx_put, IOMMUFD);
|
|
|
|
static const struct iommufd_object_ops iommufd_object_ops[] = {
|
|
[IOMMUFD_OBJ_ACCESS] = {
|
|
.destroy = iommufd_access_destroy_object,
|
|
},
|
|
[IOMMUFD_OBJ_DEVICE] = {
|
|
.destroy = iommufd_device_destroy,
|
|
},
|
|
[IOMMUFD_OBJ_IOAS] = {
|
|
.destroy = iommufd_ioas_destroy,
|
|
},
|
|
[IOMMUFD_OBJ_HWPT_PAGING] = {
|
|
.destroy = iommufd_hwpt_paging_destroy,
|
|
.abort = iommufd_hwpt_paging_abort,
|
|
},
|
|
[IOMMUFD_OBJ_HWPT_NESTED] = {
|
|
.destroy = iommufd_hwpt_nested_destroy,
|
|
.abort = iommufd_hwpt_nested_abort,
|
|
},
|
|
#ifdef CONFIG_IOMMUFD_TEST
|
|
[IOMMUFD_OBJ_SELFTEST] = {
|
|
.destroy = iommufd_selftest_destroy,
|
|
},
|
|
#endif
|
|
};
|
|
|
|
static struct miscdevice iommu_misc_dev = {
|
|
.minor = MISC_DYNAMIC_MINOR,
|
|
.name = "iommu",
|
|
.fops = &iommufd_fops,
|
|
.nodename = "iommu",
|
|
.mode = 0660,
|
|
};
|
|
|
|
|
|
static struct miscdevice vfio_misc_dev = {
|
|
.minor = VFIO_MINOR,
|
|
.name = "vfio",
|
|
.fops = &iommufd_fops,
|
|
.nodename = "vfio/vfio",
|
|
.mode = 0666,
|
|
};
|
|
|
|
static int __init iommufd_init(void)
|
|
{
|
|
int ret;
|
|
|
|
ret = misc_register(&iommu_misc_dev);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (IS_ENABLED(CONFIG_IOMMUFD_VFIO_CONTAINER)) {
|
|
ret = misc_register(&vfio_misc_dev);
|
|
if (ret)
|
|
goto err_misc;
|
|
}
|
|
ret = iommufd_test_init();
|
|
if (ret)
|
|
goto err_vfio_misc;
|
|
return 0;
|
|
|
|
err_vfio_misc:
|
|
if (IS_ENABLED(CONFIG_IOMMUFD_VFIO_CONTAINER))
|
|
misc_deregister(&vfio_misc_dev);
|
|
err_misc:
|
|
misc_deregister(&iommu_misc_dev);
|
|
return ret;
|
|
}
|
|
|
|
static void __exit iommufd_exit(void)
|
|
{
|
|
iommufd_test_exit();
|
|
if (IS_ENABLED(CONFIG_IOMMUFD_VFIO_CONTAINER))
|
|
misc_deregister(&vfio_misc_dev);
|
|
misc_deregister(&iommu_misc_dev);
|
|
}
|
|
|
|
module_init(iommufd_init);
|
|
module_exit(iommufd_exit);
|
|
|
|
#if IS_ENABLED(CONFIG_IOMMUFD_VFIO_CONTAINER)
|
|
MODULE_ALIAS_MISCDEV(VFIO_MINOR);
|
|
MODULE_ALIAS("devname:vfio/vfio");
|
|
#endif
|
|
MODULE_IMPORT_NS(IOMMUFD_INTERNAL);
|
|
MODULE_IMPORT_NS(IOMMUFD);
|
|
MODULE_DESCRIPTION("I/O Address Space Management for passthrough devices");
|
|
MODULE_LICENSE("GPL");
|