mirror of
https://github.com/torvalds/linux.git
synced 2024-11-19 10:31:48 +00:00
06a45a93e7
This moves the usermode helper locks into only code paths that use the usermode helper API from the kernel. The usermode helper locks were originally added to prevent stalling suspend, later the firmware cache was added to help with this, and further later direct filesystem lookup was added by Linus to completely bypass udev due to the amount of issues the umh approach had. The usermode helper locks were kept even when the direct filesystem lookup mechanism is used though. A lot has changed since the original usermode helper locks were added but the recent commit which added the code for firmware_enabled() are intended to address any possible races cured only as collateral by using the locks as though side consequence of code evolution and this not being addressed any time sooner. With the firmware_enabled() code in place we are a bit more sure to move the usermode helper locks to UMH only code. There is a bit of history here so let's recap a bit of it to ensure nothing is lost and things are clear. The direct filesystem approach to loading firmware is rather new, it was added via commitabb139e75c
("firmware: teach the kernel to load firmware files directly from the filesystem") by Linus merged on the v3.7 release, to enable to bypass udev. usermodehelper_read_lock_wait() was added earlier via commit9b78c1da60
("firmware_class: Do not warn that system is not ready from async loads") merged on v3.4, after Rafael noted that the async firmware API call request_firmware_nowait() should not be penalized to fail if userspace is not available yet or frozen, it'd allow for a timeout grace period before giving up. The WARN_ON() was kept for the sync firmware API call though on request_firmware(). At this time there was no direct filesystem lookup for firmware though. The original usermode helper lock came from commita144c6a6c9
("PM: Print a warning if firmware is requested when tasks are frozen") merged on the v3.0 kernel by Rafael to print a warning back when firmware requests were used on resume(), thaw() or restore() callbacks and there was no direct fs lookups or the firmware cache. Signed-off-by: Luis R. Rodriguez <mcgrof@kernel.org> Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
1911 lines
45 KiB
C
1911 lines
45 KiB
C
/*
|
|
* firmware_class.c - Multi purpose firmware loading support
|
|
*
|
|
* Copyright (c) 2003 Manuel Estrada Sainz
|
|
*
|
|
* Please see Documentation/firmware_class/ for more information.
|
|
*
|
|
*/
|
|
|
|
#include <linux/capability.h>
|
|
#include <linux/device.h>
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
#include <linux/timer.h>
|
|
#include <linux/vmalloc.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/bitops.h>
|
|
#include <linux/mutex.h>
|
|
#include <linux/workqueue.h>
|
|
#include <linux/highmem.h>
|
|
#include <linux/firmware.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/file.h>
|
|
#include <linux/list.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/async.h>
|
|
#include <linux/pm.h>
|
|
#include <linux/suspend.h>
|
|
#include <linux/syscore_ops.h>
|
|
#include <linux/reboot.h>
|
|
#include <linux/security.h>
|
|
#include <linux/swait.h>
|
|
|
|
#include <generated/utsrelease.h>
|
|
|
|
#include "base.h"
|
|
|
|
MODULE_AUTHOR("Manuel Estrada Sainz");
|
|
MODULE_DESCRIPTION("Multi purpose firmware loading support");
|
|
MODULE_LICENSE("GPL");
|
|
|
|
/* Builtin firmware support */
|
|
|
|
#ifdef CONFIG_FW_LOADER
|
|
|
|
extern struct builtin_fw __start_builtin_fw[];
|
|
extern struct builtin_fw __end_builtin_fw[];
|
|
|
|
static bool fw_get_builtin_firmware(struct firmware *fw, const char *name,
|
|
void *buf, size_t size)
|
|
{
|
|
struct builtin_fw *b_fw;
|
|
|
|
for (b_fw = __start_builtin_fw; b_fw != __end_builtin_fw; b_fw++) {
|
|
if (strcmp(name, b_fw->name) == 0) {
|
|
fw->size = b_fw->size;
|
|
fw->data = b_fw->data;
|
|
|
|
if (buf && fw->size <= size)
|
|
memcpy(buf, fw->data, fw->size);
|
|
return true;
|
|
}
|
|
}
|
|
|
|
return false;
|
|
}
|
|
|
|
static bool fw_is_builtin_firmware(const struct firmware *fw)
|
|
{
|
|
struct builtin_fw *b_fw;
|
|
|
|
for (b_fw = __start_builtin_fw; b_fw != __end_builtin_fw; b_fw++)
|
|
if (fw->data == b_fw->data)
|
|
return true;
|
|
|
|
return false;
|
|
}
|
|
|
|
#else /* Module case - no builtin firmware support */
|
|
|
|
static inline bool fw_get_builtin_firmware(struct firmware *fw,
|
|
const char *name, void *buf,
|
|
size_t size)
|
|
{
|
|
return false;
|
|
}
|
|
|
|
static inline bool fw_is_builtin_firmware(const struct firmware *fw)
|
|
{
|
|
return false;
|
|
}
|
|
#endif
|
|
|
|
enum fw_status {
|
|
FW_STATUS_UNKNOWN,
|
|
FW_STATUS_LOADING,
|
|
FW_STATUS_DONE,
|
|
FW_STATUS_ABORTED,
|
|
};
|
|
|
|
static int loading_timeout = 60; /* In seconds */
|
|
|
|
static inline long firmware_loading_timeout(void)
|
|
{
|
|
return loading_timeout > 0 ? loading_timeout * HZ : MAX_JIFFY_OFFSET;
|
|
}
|
|
|
|
/*
|
|
* Concurrent request_firmware() for the same firmware need to be
|
|
* serialized. struct fw_state is simple state machine which hold the
|
|
* state of the firmware loading.
|
|
*/
|
|
struct fw_state {
|
|
struct swait_queue_head wq;
|
|
enum fw_status status;
|
|
};
|
|
|
|
static void fw_state_init(struct fw_state *fw_st)
|
|
{
|
|
init_swait_queue_head(&fw_st->wq);
|
|
fw_st->status = FW_STATUS_UNKNOWN;
|
|
}
|
|
|
|
static inline bool __fw_state_is_done(enum fw_status status)
|
|
{
|
|
return status == FW_STATUS_DONE || status == FW_STATUS_ABORTED;
|
|
}
|
|
|
|
static int __fw_state_wait_common(struct fw_state *fw_st, long timeout)
|
|
{
|
|
long ret;
|
|
|
|
ret = swait_event_interruptible_timeout(fw_st->wq,
|
|
__fw_state_is_done(READ_ONCE(fw_st->status)),
|
|
timeout);
|
|
if (ret != 0 && fw_st->status == FW_STATUS_ABORTED)
|
|
return -ENOENT;
|
|
if (!ret)
|
|
return -ETIMEDOUT;
|
|
|
|
return ret < 0 ? ret : 0;
|
|
}
|
|
|
|
static void __fw_state_set(struct fw_state *fw_st,
|
|
enum fw_status status)
|
|
{
|
|
WRITE_ONCE(fw_st->status, status);
|
|
|
|
if (status == FW_STATUS_DONE || status == FW_STATUS_ABORTED)
|
|
swake_up(&fw_st->wq);
|
|
}
|
|
|
|
#define fw_state_start(fw_st) \
|
|
__fw_state_set(fw_st, FW_STATUS_LOADING)
|
|
#define fw_state_done(fw_st) \
|
|
__fw_state_set(fw_st, FW_STATUS_DONE)
|
|
#define fw_state_wait(fw_st) \
|
|
__fw_state_wait_common(fw_st, MAX_SCHEDULE_TIMEOUT)
|
|
|
|
#ifndef CONFIG_FW_LOADER_USER_HELPER
|
|
|
|
#define fw_state_is_aborted(fw_st) false
|
|
|
|
#else /* CONFIG_FW_LOADER_USER_HELPER */
|
|
|
|
static int __fw_state_check(struct fw_state *fw_st, enum fw_status status)
|
|
{
|
|
return fw_st->status == status;
|
|
}
|
|
|
|
#define fw_state_aborted(fw_st) \
|
|
__fw_state_set(fw_st, FW_STATUS_ABORTED)
|
|
#define fw_state_is_done(fw_st) \
|
|
__fw_state_check(fw_st, FW_STATUS_DONE)
|
|
#define fw_state_is_loading(fw_st) \
|
|
__fw_state_check(fw_st, FW_STATUS_LOADING)
|
|
#define fw_state_is_aborted(fw_st) \
|
|
__fw_state_check(fw_st, FW_STATUS_ABORTED)
|
|
#define fw_state_wait_timeout(fw_st, timeout) \
|
|
__fw_state_wait_common(fw_st, timeout)
|
|
|
|
#endif /* CONFIG_FW_LOADER_USER_HELPER */
|
|
|
|
/* firmware behavior options */
|
|
#define FW_OPT_UEVENT (1U << 0)
|
|
#define FW_OPT_NOWAIT (1U << 1)
|
|
#ifdef CONFIG_FW_LOADER_USER_HELPER
|
|
#define FW_OPT_USERHELPER (1U << 2)
|
|
#else
|
|
#define FW_OPT_USERHELPER 0
|
|
#endif
|
|
#ifdef CONFIG_FW_LOADER_USER_HELPER_FALLBACK
|
|
#define FW_OPT_FALLBACK FW_OPT_USERHELPER
|
|
#else
|
|
#define FW_OPT_FALLBACK 0
|
|
#endif
|
|
#define FW_OPT_NO_WARN (1U << 3)
|
|
#define FW_OPT_NOCACHE (1U << 4)
|
|
|
|
struct firmware_cache {
|
|
/* firmware_buf instance will be added into the below list */
|
|
spinlock_t lock;
|
|
struct list_head head;
|
|
int state;
|
|
|
|
#ifdef CONFIG_PM_SLEEP
|
|
/*
|
|
* Names of firmware images which have been cached successfully
|
|
* will be added into the below list so that device uncache
|
|
* helper can trace which firmware images have been cached
|
|
* before.
|
|
*/
|
|
spinlock_t name_lock;
|
|
struct list_head fw_names;
|
|
|
|
struct delayed_work work;
|
|
|
|
struct notifier_block pm_notify;
|
|
#endif
|
|
};
|
|
|
|
struct firmware_buf {
|
|
struct kref ref;
|
|
struct list_head list;
|
|
struct firmware_cache *fwc;
|
|
struct fw_state fw_st;
|
|
void *data;
|
|
size_t size;
|
|
size_t allocated_size;
|
|
#ifdef CONFIG_FW_LOADER_USER_HELPER
|
|
bool is_paged_buf;
|
|
bool need_uevent;
|
|
struct page **pages;
|
|
int nr_pages;
|
|
int page_array_size;
|
|
struct list_head pending_list;
|
|
#endif
|
|
const char *fw_id;
|
|
};
|
|
|
|
struct fw_cache_entry {
|
|
struct list_head list;
|
|
const char *name;
|
|
};
|
|
|
|
struct fw_name_devm {
|
|
unsigned long magic;
|
|
const char *name;
|
|
};
|
|
|
|
#define to_fwbuf(d) container_of(d, struct firmware_buf, ref)
|
|
|
|
#define FW_LOADER_NO_CACHE 0
|
|
#define FW_LOADER_START_CACHE 1
|
|
|
|
static int fw_cache_piggyback_on_request(const char *name);
|
|
|
|
/* fw_lock could be moved to 'struct firmware_priv' but since it is just
|
|
* guarding for corner cases a global lock should be OK */
|
|
static DEFINE_MUTEX(fw_lock);
|
|
|
|
static bool __enable_firmware = false;
|
|
|
|
static void enable_firmware(void)
|
|
{
|
|
mutex_lock(&fw_lock);
|
|
__enable_firmware = true;
|
|
mutex_unlock(&fw_lock);
|
|
}
|
|
|
|
static void disable_firmware(void)
|
|
{
|
|
mutex_lock(&fw_lock);
|
|
__enable_firmware = false;
|
|
mutex_unlock(&fw_lock);
|
|
}
|
|
|
|
/*
|
|
* When disabled only the built-in firmware and the firmware cache will be
|
|
* used to look for firmware.
|
|
*/
|
|
static bool firmware_enabled(void)
|
|
{
|
|
bool enabled = false;
|
|
|
|
mutex_lock(&fw_lock);
|
|
if (__enable_firmware)
|
|
enabled = true;
|
|
mutex_unlock(&fw_lock);
|
|
|
|
return enabled;
|
|
}
|
|
|
|
static struct firmware_cache fw_cache;
|
|
|
|
static struct firmware_buf *__allocate_fw_buf(const char *fw_name,
|
|
struct firmware_cache *fwc,
|
|
void *dbuf, size_t size)
|
|
{
|
|
struct firmware_buf *buf;
|
|
|
|
buf = kzalloc(sizeof(*buf), GFP_ATOMIC);
|
|
if (!buf)
|
|
return NULL;
|
|
|
|
buf->fw_id = kstrdup_const(fw_name, GFP_ATOMIC);
|
|
if (!buf->fw_id) {
|
|
kfree(buf);
|
|
return NULL;
|
|
}
|
|
|
|
kref_init(&buf->ref);
|
|
buf->fwc = fwc;
|
|
buf->data = dbuf;
|
|
buf->allocated_size = size;
|
|
fw_state_init(&buf->fw_st);
|
|
#ifdef CONFIG_FW_LOADER_USER_HELPER
|
|
INIT_LIST_HEAD(&buf->pending_list);
|
|
#endif
|
|
|
|
pr_debug("%s: fw-%s buf=%p\n", __func__, fw_name, buf);
|
|
|
|
return buf;
|
|
}
|
|
|
|
static struct firmware_buf *__fw_lookup_buf(const char *fw_name)
|
|
{
|
|
struct firmware_buf *tmp;
|
|
struct firmware_cache *fwc = &fw_cache;
|
|
|
|
list_for_each_entry(tmp, &fwc->head, list)
|
|
if (!strcmp(tmp->fw_id, fw_name))
|
|
return tmp;
|
|
return NULL;
|
|
}
|
|
|
|
static int fw_lookup_and_allocate_buf(const char *fw_name,
|
|
struct firmware_cache *fwc,
|
|
struct firmware_buf **buf, void *dbuf,
|
|
size_t size)
|
|
{
|
|
struct firmware_buf *tmp;
|
|
|
|
spin_lock(&fwc->lock);
|
|
tmp = __fw_lookup_buf(fw_name);
|
|
if (tmp) {
|
|
kref_get(&tmp->ref);
|
|
spin_unlock(&fwc->lock);
|
|
*buf = tmp;
|
|
return 1;
|
|
}
|
|
tmp = __allocate_fw_buf(fw_name, fwc, dbuf, size);
|
|
if (tmp)
|
|
list_add(&tmp->list, &fwc->head);
|
|
spin_unlock(&fwc->lock);
|
|
|
|
*buf = tmp;
|
|
|
|
return tmp ? 0 : -ENOMEM;
|
|
}
|
|
|
|
static void __fw_free_buf(struct kref *ref)
|
|
__releases(&fwc->lock)
|
|
{
|
|
struct firmware_buf *buf = to_fwbuf(ref);
|
|
struct firmware_cache *fwc = buf->fwc;
|
|
|
|
pr_debug("%s: fw-%s buf=%p data=%p size=%u\n",
|
|
__func__, buf->fw_id, buf, buf->data,
|
|
(unsigned int)buf->size);
|
|
|
|
list_del(&buf->list);
|
|
spin_unlock(&fwc->lock);
|
|
|
|
#ifdef CONFIG_FW_LOADER_USER_HELPER
|
|
if (buf->is_paged_buf) {
|
|
int i;
|
|
vunmap(buf->data);
|
|
for (i = 0; i < buf->nr_pages; i++)
|
|
__free_page(buf->pages[i]);
|
|
vfree(buf->pages);
|
|
} else
|
|
#endif
|
|
if (!buf->allocated_size)
|
|
vfree(buf->data);
|
|
kfree_const(buf->fw_id);
|
|
kfree(buf);
|
|
}
|
|
|
|
static void fw_free_buf(struct firmware_buf *buf)
|
|
{
|
|
struct firmware_cache *fwc = buf->fwc;
|
|
spin_lock(&fwc->lock);
|
|
if (!kref_put(&buf->ref, __fw_free_buf))
|
|
spin_unlock(&fwc->lock);
|
|
}
|
|
|
|
/* direct firmware loading support */
|
|
static char fw_path_para[256];
|
|
static const char * const fw_path[] = {
|
|
fw_path_para,
|
|
"/lib/firmware/updates/" UTS_RELEASE,
|
|
"/lib/firmware/updates",
|
|
"/lib/firmware/" UTS_RELEASE,
|
|
"/lib/firmware"
|
|
};
|
|
|
|
/*
|
|
* Typical usage is that passing 'firmware_class.path=$CUSTOMIZED_PATH'
|
|
* from kernel command line because firmware_class is generally built in
|
|
* kernel instead of module.
|
|
*/
|
|
module_param_string(path, fw_path_para, sizeof(fw_path_para), 0644);
|
|
MODULE_PARM_DESC(path, "customized firmware image search path with a higher priority than default path");
|
|
|
|
static int
|
|
fw_get_filesystem_firmware(struct device *device, struct firmware_buf *buf)
|
|
{
|
|
loff_t size;
|
|
int i, len;
|
|
int rc = -ENOENT;
|
|
char *path;
|
|
enum kernel_read_file_id id = READING_FIRMWARE;
|
|
size_t msize = INT_MAX;
|
|
|
|
/* Already populated data member means we're loading into a buffer */
|
|
if (buf->data) {
|
|
id = READING_FIRMWARE_PREALLOC_BUFFER;
|
|
msize = buf->allocated_size;
|
|
}
|
|
|
|
path = __getname();
|
|
if (!path)
|
|
return -ENOMEM;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(fw_path); i++) {
|
|
/* skip the unset customized path */
|
|
if (!fw_path[i][0])
|
|
continue;
|
|
|
|
len = snprintf(path, PATH_MAX, "%s/%s",
|
|
fw_path[i], buf->fw_id);
|
|
if (len >= PATH_MAX) {
|
|
rc = -ENAMETOOLONG;
|
|
break;
|
|
}
|
|
|
|
buf->size = 0;
|
|
rc = kernel_read_file_from_path(path, &buf->data, &size, msize,
|
|
id);
|
|
if (rc) {
|
|
if (rc == -ENOENT)
|
|
dev_dbg(device, "loading %s failed with error %d\n",
|
|
path, rc);
|
|
else
|
|
dev_warn(device, "loading %s failed with error %d\n",
|
|
path, rc);
|
|
continue;
|
|
}
|
|
dev_dbg(device, "direct-loading %s\n", buf->fw_id);
|
|
buf->size = size;
|
|
fw_state_done(&buf->fw_st);
|
|
break;
|
|
}
|
|
__putname(path);
|
|
|
|
return rc;
|
|
}
|
|
|
|
/* firmware holds the ownership of pages */
|
|
static void firmware_free_data(const struct firmware *fw)
|
|
{
|
|
/* Loaded directly? */
|
|
if (!fw->priv) {
|
|
vfree(fw->data);
|
|
return;
|
|
}
|
|
fw_free_buf(fw->priv);
|
|
}
|
|
|
|
/* store the pages buffer info firmware from buf */
|
|
static void fw_set_page_data(struct firmware_buf *buf, struct firmware *fw)
|
|
{
|
|
fw->priv = buf;
|
|
#ifdef CONFIG_FW_LOADER_USER_HELPER
|
|
fw->pages = buf->pages;
|
|
#endif
|
|
fw->size = buf->size;
|
|
fw->data = buf->data;
|
|
|
|
pr_debug("%s: fw-%s buf=%p data=%p size=%u\n",
|
|
__func__, buf->fw_id, buf, buf->data,
|
|
(unsigned int)buf->size);
|
|
}
|
|
|
|
#ifdef CONFIG_PM_SLEEP
|
|
static void fw_name_devm_release(struct device *dev, void *res)
|
|
{
|
|
struct fw_name_devm *fwn = res;
|
|
|
|
if (fwn->magic == (unsigned long)&fw_cache)
|
|
pr_debug("%s: fw_name-%s devm-%p released\n",
|
|
__func__, fwn->name, res);
|
|
kfree_const(fwn->name);
|
|
}
|
|
|
|
static int fw_devm_match(struct device *dev, void *res,
|
|
void *match_data)
|
|
{
|
|
struct fw_name_devm *fwn = res;
|
|
|
|
return (fwn->magic == (unsigned long)&fw_cache) &&
|
|
!strcmp(fwn->name, match_data);
|
|
}
|
|
|
|
static struct fw_name_devm *fw_find_devm_name(struct device *dev,
|
|
const char *name)
|
|
{
|
|
struct fw_name_devm *fwn;
|
|
|
|
fwn = devres_find(dev, fw_name_devm_release,
|
|
fw_devm_match, (void *)name);
|
|
return fwn;
|
|
}
|
|
|
|
/* add firmware name into devres list */
|
|
static int fw_add_devm_name(struct device *dev, const char *name)
|
|
{
|
|
struct fw_name_devm *fwn;
|
|
|
|
fwn = fw_find_devm_name(dev, name);
|
|
if (fwn)
|
|
return 1;
|
|
|
|
fwn = devres_alloc(fw_name_devm_release, sizeof(struct fw_name_devm),
|
|
GFP_KERNEL);
|
|
if (!fwn)
|
|
return -ENOMEM;
|
|
fwn->name = kstrdup_const(name, GFP_KERNEL);
|
|
if (!fwn->name) {
|
|
devres_free(fwn);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
fwn->magic = (unsigned long)&fw_cache;
|
|
devres_add(dev, fwn);
|
|
|
|
return 0;
|
|
}
|
|
#else
|
|
static int fw_add_devm_name(struct device *dev, const char *name)
|
|
{
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
static int assign_firmware_buf(struct firmware *fw, struct device *device,
|
|
unsigned int opt_flags)
|
|
{
|
|
struct firmware_buf *buf = fw->priv;
|
|
|
|
mutex_lock(&fw_lock);
|
|
if (!buf->size || fw_state_is_aborted(&buf->fw_st)) {
|
|
mutex_unlock(&fw_lock);
|
|
return -ENOENT;
|
|
}
|
|
|
|
/*
|
|
* add firmware name into devres list so that we can auto cache
|
|
* and uncache firmware for device.
|
|
*
|
|
* device may has been deleted already, but the problem
|
|
* should be fixed in devres or driver core.
|
|
*/
|
|
/* don't cache firmware handled without uevent */
|
|
if (device && (opt_flags & FW_OPT_UEVENT) &&
|
|
!(opt_flags & FW_OPT_NOCACHE))
|
|
fw_add_devm_name(device, buf->fw_id);
|
|
|
|
/*
|
|
* After caching firmware image is started, let it piggyback
|
|
* on request firmware.
|
|
*/
|
|
if (!(opt_flags & FW_OPT_NOCACHE) &&
|
|
buf->fwc->state == FW_LOADER_START_CACHE) {
|
|
if (fw_cache_piggyback_on_request(buf->fw_id))
|
|
kref_get(&buf->ref);
|
|
}
|
|
|
|
/* pass the pages buffer to driver at the last minute */
|
|
fw_set_page_data(buf, fw);
|
|
mutex_unlock(&fw_lock);
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* user-mode helper code
|
|
*/
|
|
#ifdef CONFIG_FW_LOADER_USER_HELPER
|
|
struct firmware_priv {
|
|
bool nowait;
|
|
struct device dev;
|
|
struct firmware_buf *buf;
|
|
struct firmware *fw;
|
|
};
|
|
|
|
static struct firmware_priv *to_firmware_priv(struct device *dev)
|
|
{
|
|
return container_of(dev, struct firmware_priv, dev);
|
|
}
|
|
|
|
static void __fw_load_abort(struct firmware_buf *buf)
|
|
{
|
|
/*
|
|
* There is a small window in which user can write to 'loading'
|
|
* between loading done and disappearance of 'loading'
|
|
*/
|
|
if (fw_state_is_done(&buf->fw_st))
|
|
return;
|
|
|
|
list_del_init(&buf->pending_list);
|
|
fw_state_aborted(&buf->fw_st);
|
|
}
|
|
|
|
static void fw_load_abort(struct firmware_priv *fw_priv)
|
|
{
|
|
struct firmware_buf *buf = fw_priv->buf;
|
|
|
|
__fw_load_abort(buf);
|
|
}
|
|
|
|
static LIST_HEAD(pending_fw_head);
|
|
|
|
static void kill_pending_fw_fallback_reqs(bool only_kill_custom)
|
|
{
|
|
struct firmware_buf *buf;
|
|
struct firmware_buf *next;
|
|
|
|
mutex_lock(&fw_lock);
|
|
list_for_each_entry_safe(buf, next, &pending_fw_head, pending_list) {
|
|
if (!buf->need_uevent || !only_kill_custom)
|
|
__fw_load_abort(buf);
|
|
}
|
|
mutex_unlock(&fw_lock);
|
|
}
|
|
|
|
static ssize_t timeout_show(struct class *class, struct class_attribute *attr,
|
|
char *buf)
|
|
{
|
|
return sprintf(buf, "%d\n", loading_timeout);
|
|
}
|
|
|
|
/**
|
|
* firmware_timeout_store - set number of seconds to wait for firmware
|
|
* @class: device class pointer
|
|
* @attr: device attribute pointer
|
|
* @buf: buffer to scan for timeout value
|
|
* @count: number of bytes in @buf
|
|
*
|
|
* Sets the number of seconds to wait for the firmware. Once
|
|
* this expires an error will be returned to the driver and no
|
|
* firmware will be provided.
|
|
*
|
|
* Note: zero means 'wait forever'.
|
|
**/
|
|
static ssize_t timeout_store(struct class *class, struct class_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
loading_timeout = simple_strtol(buf, NULL, 10);
|
|
if (loading_timeout < 0)
|
|
loading_timeout = 0;
|
|
|
|
return count;
|
|
}
|
|
static CLASS_ATTR_RW(timeout);
|
|
|
|
static struct attribute *firmware_class_attrs[] = {
|
|
&class_attr_timeout.attr,
|
|
NULL,
|
|
};
|
|
ATTRIBUTE_GROUPS(firmware_class);
|
|
|
|
static void fw_dev_release(struct device *dev)
|
|
{
|
|
struct firmware_priv *fw_priv = to_firmware_priv(dev);
|
|
|
|
kfree(fw_priv);
|
|
}
|
|
|
|
static int do_firmware_uevent(struct firmware_priv *fw_priv, struct kobj_uevent_env *env)
|
|
{
|
|
if (add_uevent_var(env, "FIRMWARE=%s", fw_priv->buf->fw_id))
|
|
return -ENOMEM;
|
|
if (add_uevent_var(env, "TIMEOUT=%i", loading_timeout))
|
|
return -ENOMEM;
|
|
if (add_uevent_var(env, "ASYNC=%d", fw_priv->nowait))
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int firmware_uevent(struct device *dev, struct kobj_uevent_env *env)
|
|
{
|
|
struct firmware_priv *fw_priv = to_firmware_priv(dev);
|
|
int err = 0;
|
|
|
|
mutex_lock(&fw_lock);
|
|
if (fw_priv->buf)
|
|
err = do_firmware_uevent(fw_priv, env);
|
|
mutex_unlock(&fw_lock);
|
|
return err;
|
|
}
|
|
|
|
static struct class firmware_class = {
|
|
.name = "firmware",
|
|
.class_groups = firmware_class_groups,
|
|
.dev_uevent = firmware_uevent,
|
|
.dev_release = fw_dev_release,
|
|
};
|
|
|
|
static ssize_t firmware_loading_show(struct device *dev,
|
|
struct device_attribute *attr, char *buf)
|
|
{
|
|
struct firmware_priv *fw_priv = to_firmware_priv(dev);
|
|
int loading = 0;
|
|
|
|
mutex_lock(&fw_lock);
|
|
if (fw_priv->buf)
|
|
loading = fw_state_is_loading(&fw_priv->buf->fw_st);
|
|
mutex_unlock(&fw_lock);
|
|
|
|
return sprintf(buf, "%d\n", loading);
|
|
}
|
|
|
|
/* Some architectures don't have PAGE_KERNEL_RO */
|
|
#ifndef PAGE_KERNEL_RO
|
|
#define PAGE_KERNEL_RO PAGE_KERNEL
|
|
#endif
|
|
|
|
/* one pages buffer should be mapped/unmapped only once */
|
|
static int fw_map_pages_buf(struct firmware_buf *buf)
|
|
{
|
|
if (!buf->is_paged_buf)
|
|
return 0;
|
|
|
|
vunmap(buf->data);
|
|
buf->data = vmap(buf->pages, buf->nr_pages, 0, PAGE_KERNEL_RO);
|
|
if (!buf->data)
|
|
return -ENOMEM;
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* firmware_loading_store - set value in the 'loading' control file
|
|
* @dev: device pointer
|
|
* @attr: device attribute pointer
|
|
* @buf: buffer to scan for loading control value
|
|
* @count: number of bytes in @buf
|
|
*
|
|
* The relevant values are:
|
|
*
|
|
* 1: Start a load, discarding any previous partial load.
|
|
* 0: Conclude the load and hand the data to the driver code.
|
|
* -1: Conclude the load with an error and discard any written data.
|
|
**/
|
|
static ssize_t firmware_loading_store(struct device *dev,
|
|
struct device_attribute *attr,
|
|
const char *buf, size_t count)
|
|
{
|
|
struct firmware_priv *fw_priv = to_firmware_priv(dev);
|
|
struct firmware_buf *fw_buf;
|
|
ssize_t written = count;
|
|
int loading = simple_strtol(buf, NULL, 10);
|
|
int i;
|
|
|
|
mutex_lock(&fw_lock);
|
|
fw_buf = fw_priv->buf;
|
|
if (fw_state_is_aborted(&fw_buf->fw_st))
|
|
goto out;
|
|
|
|
switch (loading) {
|
|
case 1:
|
|
/* discarding any previous partial load */
|
|
if (!fw_state_is_done(&fw_buf->fw_st)) {
|
|
for (i = 0; i < fw_buf->nr_pages; i++)
|
|
__free_page(fw_buf->pages[i]);
|
|
vfree(fw_buf->pages);
|
|
fw_buf->pages = NULL;
|
|
fw_buf->page_array_size = 0;
|
|
fw_buf->nr_pages = 0;
|
|
fw_state_start(&fw_buf->fw_st);
|
|
}
|
|
break;
|
|
case 0:
|
|
if (fw_state_is_loading(&fw_buf->fw_st)) {
|
|
int rc;
|
|
|
|
/*
|
|
* Several loading requests may be pending on
|
|
* one same firmware buf, so let all requests
|
|
* see the mapped 'buf->data' once the loading
|
|
* is completed.
|
|
* */
|
|
rc = fw_map_pages_buf(fw_buf);
|
|
if (rc)
|
|
dev_err(dev, "%s: map pages failed\n",
|
|
__func__);
|
|
else
|
|
rc = security_kernel_post_read_file(NULL,
|
|
fw_buf->data, fw_buf->size,
|
|
READING_FIRMWARE);
|
|
|
|
/*
|
|
* Same logic as fw_load_abort, only the DONE bit
|
|
* is ignored and we set ABORT only on failure.
|
|
*/
|
|
list_del_init(&fw_buf->pending_list);
|
|
if (rc) {
|
|
fw_state_aborted(&fw_buf->fw_st);
|
|
written = rc;
|
|
} else {
|
|
fw_state_done(&fw_buf->fw_st);
|
|
}
|
|
break;
|
|
}
|
|
/* fallthrough */
|
|
default:
|
|
dev_err(dev, "%s: unexpected value (%d)\n", __func__, loading);
|
|
/* fallthrough */
|
|
case -1:
|
|
fw_load_abort(fw_priv);
|
|
break;
|
|
}
|
|
out:
|
|
mutex_unlock(&fw_lock);
|
|
return written;
|
|
}
|
|
|
|
static DEVICE_ATTR(loading, 0644, firmware_loading_show, firmware_loading_store);
|
|
|
|
static void firmware_rw_buf(struct firmware_buf *buf, char *buffer,
|
|
loff_t offset, size_t count, bool read)
|
|
{
|
|
if (read)
|
|
memcpy(buffer, buf->data + offset, count);
|
|
else
|
|
memcpy(buf->data + offset, buffer, count);
|
|
}
|
|
|
|
static void firmware_rw(struct firmware_buf *buf, char *buffer,
|
|
loff_t offset, size_t count, bool read)
|
|
{
|
|
while (count) {
|
|
void *page_data;
|
|
int page_nr = offset >> PAGE_SHIFT;
|
|
int page_ofs = offset & (PAGE_SIZE-1);
|
|
int page_cnt = min_t(size_t, PAGE_SIZE - page_ofs, count);
|
|
|
|
page_data = kmap(buf->pages[page_nr]);
|
|
|
|
if (read)
|
|
memcpy(buffer, page_data + page_ofs, page_cnt);
|
|
else
|
|
memcpy(page_data + page_ofs, buffer, page_cnt);
|
|
|
|
kunmap(buf->pages[page_nr]);
|
|
buffer += page_cnt;
|
|
offset += page_cnt;
|
|
count -= page_cnt;
|
|
}
|
|
}
|
|
|
|
static ssize_t firmware_data_read(struct file *filp, struct kobject *kobj,
|
|
struct bin_attribute *bin_attr,
|
|
char *buffer, loff_t offset, size_t count)
|
|
{
|
|
struct device *dev = kobj_to_dev(kobj);
|
|
struct firmware_priv *fw_priv = to_firmware_priv(dev);
|
|
struct firmware_buf *buf;
|
|
ssize_t ret_count;
|
|
|
|
mutex_lock(&fw_lock);
|
|
buf = fw_priv->buf;
|
|
if (!buf || fw_state_is_done(&buf->fw_st)) {
|
|
ret_count = -ENODEV;
|
|
goto out;
|
|
}
|
|
if (offset > buf->size) {
|
|
ret_count = 0;
|
|
goto out;
|
|
}
|
|
if (count > buf->size - offset)
|
|
count = buf->size - offset;
|
|
|
|
ret_count = count;
|
|
|
|
if (buf->data)
|
|
firmware_rw_buf(buf, buffer, offset, count, true);
|
|
else
|
|
firmware_rw(buf, buffer, offset, count, true);
|
|
|
|
out:
|
|
mutex_unlock(&fw_lock);
|
|
return ret_count;
|
|
}
|
|
|
|
static int fw_realloc_buffer(struct firmware_priv *fw_priv, int min_size)
|
|
{
|
|
struct firmware_buf *buf = fw_priv->buf;
|
|
int pages_needed = PAGE_ALIGN(min_size) >> PAGE_SHIFT;
|
|
|
|
/* If the array of pages is too small, grow it... */
|
|
if (buf->page_array_size < pages_needed) {
|
|
int new_array_size = max(pages_needed,
|
|
buf->page_array_size * 2);
|
|
struct page **new_pages;
|
|
|
|
new_pages = vmalloc(new_array_size * sizeof(void *));
|
|
if (!new_pages) {
|
|
fw_load_abort(fw_priv);
|
|
return -ENOMEM;
|
|
}
|
|
memcpy(new_pages, buf->pages,
|
|
buf->page_array_size * sizeof(void *));
|
|
memset(&new_pages[buf->page_array_size], 0, sizeof(void *) *
|
|
(new_array_size - buf->page_array_size));
|
|
vfree(buf->pages);
|
|
buf->pages = new_pages;
|
|
buf->page_array_size = new_array_size;
|
|
}
|
|
|
|
while (buf->nr_pages < pages_needed) {
|
|
buf->pages[buf->nr_pages] =
|
|
alloc_page(GFP_KERNEL | __GFP_HIGHMEM);
|
|
|
|
if (!buf->pages[buf->nr_pages]) {
|
|
fw_load_abort(fw_priv);
|
|
return -ENOMEM;
|
|
}
|
|
buf->nr_pages++;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* firmware_data_write - write method for firmware
|
|
* @filp: open sysfs file
|
|
* @kobj: kobject for the device
|
|
* @bin_attr: bin_attr structure
|
|
* @buffer: buffer being written
|
|
* @offset: buffer offset for write in total data store area
|
|
* @count: buffer size
|
|
*
|
|
* Data written to the 'data' attribute will be later handed to
|
|
* the driver as a firmware image.
|
|
**/
|
|
static ssize_t firmware_data_write(struct file *filp, struct kobject *kobj,
|
|
struct bin_attribute *bin_attr,
|
|
char *buffer, loff_t offset, size_t count)
|
|
{
|
|
struct device *dev = kobj_to_dev(kobj);
|
|
struct firmware_priv *fw_priv = to_firmware_priv(dev);
|
|
struct firmware_buf *buf;
|
|
ssize_t retval;
|
|
|
|
if (!capable(CAP_SYS_RAWIO))
|
|
return -EPERM;
|
|
|
|
mutex_lock(&fw_lock);
|
|
buf = fw_priv->buf;
|
|
if (!buf || fw_state_is_done(&buf->fw_st)) {
|
|
retval = -ENODEV;
|
|
goto out;
|
|
}
|
|
|
|
if (buf->data) {
|
|
if (offset + count > buf->allocated_size) {
|
|
retval = -ENOMEM;
|
|
goto out;
|
|
}
|
|
firmware_rw_buf(buf, buffer, offset, count, false);
|
|
retval = count;
|
|
} else {
|
|
retval = fw_realloc_buffer(fw_priv, offset + count);
|
|
if (retval)
|
|
goto out;
|
|
|
|
retval = count;
|
|
firmware_rw(buf, buffer, offset, count, false);
|
|
}
|
|
|
|
buf->size = max_t(size_t, offset + count, buf->size);
|
|
out:
|
|
mutex_unlock(&fw_lock);
|
|
return retval;
|
|
}
|
|
|
|
static struct bin_attribute firmware_attr_data = {
|
|
.attr = { .name = "data", .mode = 0644 },
|
|
.size = 0,
|
|
.read = firmware_data_read,
|
|
.write = firmware_data_write,
|
|
};
|
|
|
|
static struct attribute *fw_dev_attrs[] = {
|
|
&dev_attr_loading.attr,
|
|
NULL
|
|
};
|
|
|
|
static struct bin_attribute *fw_dev_bin_attrs[] = {
|
|
&firmware_attr_data,
|
|
NULL
|
|
};
|
|
|
|
static const struct attribute_group fw_dev_attr_group = {
|
|
.attrs = fw_dev_attrs,
|
|
.bin_attrs = fw_dev_bin_attrs,
|
|
};
|
|
|
|
static const struct attribute_group *fw_dev_attr_groups[] = {
|
|
&fw_dev_attr_group,
|
|
NULL
|
|
};
|
|
|
|
static struct firmware_priv *
|
|
fw_create_instance(struct firmware *firmware, const char *fw_name,
|
|
struct device *device, unsigned int opt_flags)
|
|
{
|
|
struct firmware_priv *fw_priv;
|
|
struct device *f_dev;
|
|
|
|
fw_priv = kzalloc(sizeof(*fw_priv), GFP_KERNEL);
|
|
if (!fw_priv) {
|
|
fw_priv = ERR_PTR(-ENOMEM);
|
|
goto exit;
|
|
}
|
|
|
|
fw_priv->nowait = !!(opt_flags & FW_OPT_NOWAIT);
|
|
fw_priv->fw = firmware;
|
|
f_dev = &fw_priv->dev;
|
|
|
|
device_initialize(f_dev);
|
|
dev_set_name(f_dev, "%s", fw_name);
|
|
f_dev->parent = device;
|
|
f_dev->class = &firmware_class;
|
|
f_dev->groups = fw_dev_attr_groups;
|
|
exit:
|
|
return fw_priv;
|
|
}
|
|
|
|
/* load a firmware via user helper */
|
|
static int _request_firmware_load(struct firmware_priv *fw_priv,
|
|
unsigned int opt_flags, long timeout)
|
|
{
|
|
int retval = 0;
|
|
struct device *f_dev = &fw_priv->dev;
|
|
struct firmware_buf *buf = fw_priv->buf;
|
|
|
|
/* fall back on userspace loading */
|
|
if (!buf->data)
|
|
buf->is_paged_buf = true;
|
|
|
|
dev_set_uevent_suppress(f_dev, true);
|
|
|
|
retval = device_add(f_dev);
|
|
if (retval) {
|
|
dev_err(f_dev, "%s: device_register failed\n", __func__);
|
|
goto err_put_dev;
|
|
}
|
|
|
|
mutex_lock(&fw_lock);
|
|
list_add(&buf->pending_list, &pending_fw_head);
|
|
mutex_unlock(&fw_lock);
|
|
|
|
if (opt_flags & FW_OPT_UEVENT) {
|
|
buf->need_uevent = true;
|
|
dev_set_uevent_suppress(f_dev, false);
|
|
dev_dbg(f_dev, "firmware: requesting %s\n", buf->fw_id);
|
|
kobject_uevent(&fw_priv->dev.kobj, KOBJ_ADD);
|
|
} else {
|
|
timeout = MAX_JIFFY_OFFSET;
|
|
}
|
|
|
|
retval = fw_state_wait_timeout(&buf->fw_st, timeout);
|
|
if (retval < 0) {
|
|
mutex_lock(&fw_lock);
|
|
fw_load_abort(fw_priv);
|
|
mutex_unlock(&fw_lock);
|
|
}
|
|
|
|
if (fw_state_is_aborted(&buf->fw_st))
|
|
retval = -EAGAIN;
|
|
else if (buf->is_paged_buf && !buf->data)
|
|
retval = -ENOMEM;
|
|
|
|
device_del(f_dev);
|
|
err_put_dev:
|
|
put_device(f_dev);
|
|
return retval;
|
|
}
|
|
|
|
static int fw_load_from_user_helper(struct firmware *firmware,
|
|
const char *name, struct device *device,
|
|
unsigned int opt_flags)
|
|
{
|
|
struct firmware_priv *fw_priv;
|
|
long timeout;
|
|
int ret;
|
|
|
|
timeout = firmware_loading_timeout();
|
|
if (opt_flags & FW_OPT_NOWAIT) {
|
|
timeout = usermodehelper_read_lock_wait(timeout);
|
|
if (!timeout) {
|
|
dev_dbg(device, "firmware: %s loading timed out\n",
|
|
name);
|
|
return -EBUSY;
|
|
}
|
|
} else {
|
|
ret = usermodehelper_read_trylock();
|
|
if (WARN_ON(ret)) {
|
|
dev_err(device, "firmware: %s will not be loaded\n",
|
|
name);
|
|
return ret;
|
|
}
|
|
}
|
|
|
|
fw_priv = fw_create_instance(firmware, name, device, opt_flags);
|
|
if (IS_ERR(fw_priv)) {
|
|
ret = PTR_ERR(fw_priv);
|
|
goto out_unlock;
|
|
}
|
|
|
|
fw_priv->buf = firmware->priv;
|
|
ret = _request_firmware_load(fw_priv, opt_flags, timeout);
|
|
|
|
if (!ret)
|
|
ret = assign_firmware_buf(firmware, device, opt_flags);
|
|
|
|
out_unlock:
|
|
usermodehelper_read_unlock();
|
|
|
|
return ret;
|
|
}
|
|
|
|
#else /* CONFIG_FW_LOADER_USER_HELPER */
|
|
static inline int
|
|
fw_load_from_user_helper(struct firmware *firmware, const char *name,
|
|
struct device *device, unsigned int opt_flags)
|
|
{
|
|
return -ENOENT;
|
|
}
|
|
|
|
static inline void kill_pending_fw_fallback_reqs(bool only_kill_custom) { }
|
|
|
|
#endif /* CONFIG_FW_LOADER_USER_HELPER */
|
|
|
|
/* prepare firmware and firmware_buf structs;
|
|
* return 0 if a firmware is already assigned, 1 if need to load one,
|
|
* or a negative error code
|
|
*/
|
|
static int
|
|
_request_firmware_prepare(struct firmware **firmware_p, const char *name,
|
|
struct device *device, void *dbuf, size_t size)
|
|
{
|
|
struct firmware *firmware;
|
|
struct firmware_buf *buf;
|
|
int ret;
|
|
|
|
*firmware_p = firmware = kzalloc(sizeof(*firmware), GFP_KERNEL);
|
|
if (!firmware) {
|
|
dev_err(device, "%s: kmalloc(struct firmware) failed\n",
|
|
__func__);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
if (fw_get_builtin_firmware(firmware, name, dbuf, size)) {
|
|
dev_dbg(device, "using built-in %s\n", name);
|
|
return 0; /* assigned */
|
|
}
|
|
|
|
ret = fw_lookup_and_allocate_buf(name, &fw_cache, &buf, dbuf, size);
|
|
|
|
/*
|
|
* bind with 'buf' now to avoid warning in failure path
|
|
* of requesting firmware.
|
|
*/
|
|
firmware->priv = buf;
|
|
|
|
if (ret > 0) {
|
|
ret = fw_state_wait(&buf->fw_st);
|
|
if (!ret) {
|
|
fw_set_page_data(buf, firmware);
|
|
return 0; /* assigned */
|
|
}
|
|
}
|
|
|
|
if (ret < 0)
|
|
return ret;
|
|
return 1; /* need to load */
|
|
}
|
|
|
|
/* called from request_firmware() and request_firmware_work_func() */
|
|
static int
|
|
_request_firmware(const struct firmware **firmware_p, const char *name,
|
|
struct device *device, void *buf, size_t size,
|
|
unsigned int opt_flags)
|
|
{
|
|
struct firmware *fw = NULL;
|
|
int ret;
|
|
|
|
if (!firmware_p)
|
|
return -EINVAL;
|
|
|
|
if (!name || name[0] == '\0') {
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
ret = _request_firmware_prepare(&fw, name, device, buf, size);
|
|
if (ret <= 0) /* error or already assigned */
|
|
goto out;
|
|
|
|
if (!firmware_enabled()) {
|
|
WARN(1, "firmware request while host is not available\n");
|
|
ret = -EHOSTDOWN;
|
|
goto out;
|
|
}
|
|
|
|
ret = fw_get_filesystem_firmware(device, fw->priv);
|
|
if (ret) {
|
|
if (!(opt_flags & FW_OPT_NO_WARN))
|
|
dev_warn(device,
|
|
"Direct firmware load for %s failed with error %d\n",
|
|
name, ret);
|
|
if (opt_flags & FW_OPT_USERHELPER) {
|
|
dev_warn(device, "Falling back to user helper\n");
|
|
ret = fw_load_from_user_helper(fw, name, device,
|
|
opt_flags);
|
|
}
|
|
} else
|
|
ret = assign_firmware_buf(fw, device, opt_flags);
|
|
|
|
out:
|
|
if (ret < 0) {
|
|
release_firmware(fw);
|
|
fw = NULL;
|
|
}
|
|
|
|
*firmware_p = fw;
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* request_firmware: - send firmware request and wait for it
|
|
* @firmware_p: pointer to firmware image
|
|
* @name: name of firmware file
|
|
* @device: device for which firmware is being loaded
|
|
*
|
|
* @firmware_p will be used to return a firmware image by the name
|
|
* of @name for device @device.
|
|
*
|
|
* Should be called from user context where sleeping is allowed.
|
|
*
|
|
* @name will be used as $FIRMWARE in the uevent environment and
|
|
* should be distinctive enough not to be confused with any other
|
|
* firmware image for this or any other device.
|
|
*
|
|
* Caller must hold the reference count of @device.
|
|
*
|
|
* The function can be called safely inside device's suspend and
|
|
* resume callback.
|
|
**/
|
|
int
|
|
request_firmware(const struct firmware **firmware_p, const char *name,
|
|
struct device *device)
|
|
{
|
|
int ret;
|
|
|
|
/* Need to pin this module until return */
|
|
__module_get(THIS_MODULE);
|
|
ret = _request_firmware(firmware_p, name, device, NULL, 0,
|
|
FW_OPT_UEVENT | FW_OPT_FALLBACK);
|
|
module_put(THIS_MODULE);
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(request_firmware);
|
|
|
|
/**
|
|
* request_firmware_direct: - load firmware directly without usermode helper
|
|
* @firmware_p: pointer to firmware image
|
|
* @name: name of firmware file
|
|
* @device: device for which firmware is being loaded
|
|
*
|
|
* This function works pretty much like request_firmware(), but this doesn't
|
|
* fall back to usermode helper even if the firmware couldn't be loaded
|
|
* directly from fs. Hence it's useful for loading optional firmwares, which
|
|
* aren't always present, without extra long timeouts of udev.
|
|
**/
|
|
int request_firmware_direct(const struct firmware **firmware_p,
|
|
const char *name, struct device *device)
|
|
{
|
|
int ret;
|
|
|
|
__module_get(THIS_MODULE);
|
|
ret = _request_firmware(firmware_p, name, device, NULL, 0,
|
|
FW_OPT_UEVENT | FW_OPT_NO_WARN);
|
|
module_put(THIS_MODULE);
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL_GPL(request_firmware_direct);
|
|
|
|
/**
|
|
* request_firmware_into_buf - load firmware into a previously allocated buffer
|
|
* @firmware_p: pointer to firmware image
|
|
* @name: name of firmware file
|
|
* @device: device for which firmware is being loaded and DMA region allocated
|
|
* @buf: address of buffer to load firmware into
|
|
* @size: size of buffer
|
|
*
|
|
* This function works pretty much like request_firmware(), but it doesn't
|
|
* allocate a buffer to hold the firmware data. Instead, the firmware
|
|
* is loaded directly into the buffer pointed to by @buf and the @firmware_p
|
|
* data member is pointed at @buf.
|
|
*
|
|
* This function doesn't cache firmware either.
|
|
*/
|
|
int
|
|
request_firmware_into_buf(const struct firmware **firmware_p, const char *name,
|
|
struct device *device, void *buf, size_t size)
|
|
{
|
|
int ret;
|
|
|
|
__module_get(THIS_MODULE);
|
|
ret = _request_firmware(firmware_p, name, device, buf, size,
|
|
FW_OPT_UEVENT | FW_OPT_FALLBACK |
|
|
FW_OPT_NOCACHE);
|
|
module_put(THIS_MODULE);
|
|
return ret;
|
|
}
|
|
EXPORT_SYMBOL(request_firmware_into_buf);
|
|
|
|
/**
|
|
* release_firmware: - release the resource associated with a firmware image
|
|
* @fw: firmware resource to release
|
|
**/
|
|
void release_firmware(const struct firmware *fw)
|
|
{
|
|
if (fw) {
|
|
if (!fw_is_builtin_firmware(fw))
|
|
firmware_free_data(fw);
|
|
kfree(fw);
|
|
}
|
|
}
|
|
EXPORT_SYMBOL(release_firmware);
|
|
|
|
/* Async support */
|
|
struct firmware_work {
|
|
struct work_struct work;
|
|
struct module *module;
|
|
const char *name;
|
|
struct device *device;
|
|
void *context;
|
|
void (*cont)(const struct firmware *fw, void *context);
|
|
unsigned int opt_flags;
|
|
};
|
|
|
|
static void request_firmware_work_func(struct work_struct *work)
|
|
{
|
|
struct firmware_work *fw_work;
|
|
const struct firmware *fw;
|
|
|
|
fw_work = container_of(work, struct firmware_work, work);
|
|
|
|
_request_firmware(&fw, fw_work->name, fw_work->device, NULL, 0,
|
|
fw_work->opt_flags);
|
|
fw_work->cont(fw, fw_work->context);
|
|
put_device(fw_work->device); /* taken in request_firmware_nowait() */
|
|
|
|
module_put(fw_work->module);
|
|
kfree_const(fw_work->name);
|
|
kfree(fw_work);
|
|
}
|
|
|
|
/**
|
|
* request_firmware_nowait - asynchronous version of request_firmware
|
|
* @module: module requesting the firmware
|
|
* @uevent: sends uevent to copy the firmware image if this flag
|
|
* is non-zero else the firmware copy must be done manually.
|
|
* @name: name of firmware file
|
|
* @device: device for which firmware is being loaded
|
|
* @gfp: allocation flags
|
|
* @context: will be passed over to @cont, and
|
|
* @fw may be %NULL if firmware request fails.
|
|
* @cont: function will be called asynchronously when the firmware
|
|
* request is over.
|
|
*
|
|
* Caller must hold the reference count of @device.
|
|
*
|
|
* Asynchronous variant of request_firmware() for user contexts:
|
|
* - sleep for as small periods as possible since it may
|
|
* increase kernel boot time of built-in device drivers
|
|
* requesting firmware in their ->probe() methods, if
|
|
* @gfp is GFP_KERNEL.
|
|
*
|
|
* - can't sleep at all if @gfp is GFP_ATOMIC.
|
|
**/
|
|
int
|
|
request_firmware_nowait(
|
|
struct module *module, bool uevent,
|
|
const char *name, struct device *device, gfp_t gfp, void *context,
|
|
void (*cont)(const struct firmware *fw, void *context))
|
|
{
|
|
struct firmware_work *fw_work;
|
|
|
|
fw_work = kzalloc(sizeof(struct firmware_work), gfp);
|
|
if (!fw_work)
|
|
return -ENOMEM;
|
|
|
|
fw_work->module = module;
|
|
fw_work->name = kstrdup_const(name, gfp);
|
|
if (!fw_work->name) {
|
|
kfree(fw_work);
|
|
return -ENOMEM;
|
|
}
|
|
fw_work->device = device;
|
|
fw_work->context = context;
|
|
fw_work->cont = cont;
|
|
fw_work->opt_flags = FW_OPT_NOWAIT | FW_OPT_FALLBACK |
|
|
(uevent ? FW_OPT_UEVENT : FW_OPT_USERHELPER);
|
|
|
|
if (!try_module_get(module)) {
|
|
kfree_const(fw_work->name);
|
|
kfree(fw_work);
|
|
return -EFAULT;
|
|
}
|
|
|
|
get_device(fw_work->device);
|
|
INIT_WORK(&fw_work->work, request_firmware_work_func);
|
|
schedule_work(&fw_work->work);
|
|
return 0;
|
|
}
|
|
EXPORT_SYMBOL(request_firmware_nowait);
|
|
|
|
#ifdef CONFIG_PM_SLEEP
|
|
static ASYNC_DOMAIN_EXCLUSIVE(fw_cache_domain);
|
|
|
|
/**
|
|
* cache_firmware - cache one firmware image in kernel memory space
|
|
* @fw_name: the firmware image name
|
|
*
|
|
* Cache firmware in kernel memory so that drivers can use it when
|
|
* system isn't ready for them to request firmware image from userspace.
|
|
* Once it returns successfully, driver can use request_firmware or its
|
|
* nowait version to get the cached firmware without any interacting
|
|
* with userspace
|
|
*
|
|
* Return 0 if the firmware image has been cached successfully
|
|
* Return !0 otherwise
|
|
*
|
|
*/
|
|
static int cache_firmware(const char *fw_name)
|
|
{
|
|
int ret;
|
|
const struct firmware *fw;
|
|
|
|
pr_debug("%s: %s\n", __func__, fw_name);
|
|
|
|
ret = request_firmware(&fw, fw_name, NULL);
|
|
if (!ret)
|
|
kfree(fw);
|
|
|
|
pr_debug("%s: %s ret=%d\n", __func__, fw_name, ret);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static struct firmware_buf *fw_lookup_buf(const char *fw_name)
|
|
{
|
|
struct firmware_buf *tmp;
|
|
struct firmware_cache *fwc = &fw_cache;
|
|
|
|
spin_lock(&fwc->lock);
|
|
tmp = __fw_lookup_buf(fw_name);
|
|
spin_unlock(&fwc->lock);
|
|
|
|
return tmp;
|
|
}
|
|
|
|
/**
|
|
* uncache_firmware - remove one cached firmware image
|
|
* @fw_name: the firmware image name
|
|
*
|
|
* Uncache one firmware image which has been cached successfully
|
|
* before.
|
|
*
|
|
* Return 0 if the firmware cache has been removed successfully
|
|
* Return !0 otherwise
|
|
*
|
|
*/
|
|
static int uncache_firmware(const char *fw_name)
|
|
{
|
|
struct firmware_buf *buf;
|
|
struct firmware fw;
|
|
|
|
pr_debug("%s: %s\n", __func__, fw_name);
|
|
|
|
if (fw_get_builtin_firmware(&fw, fw_name, NULL, 0))
|
|
return 0;
|
|
|
|
buf = fw_lookup_buf(fw_name);
|
|
if (buf) {
|
|
fw_free_buf(buf);
|
|
return 0;
|
|
}
|
|
|
|
return -EINVAL;
|
|
}
|
|
|
|
static struct fw_cache_entry *alloc_fw_cache_entry(const char *name)
|
|
{
|
|
struct fw_cache_entry *fce;
|
|
|
|
fce = kzalloc(sizeof(*fce), GFP_ATOMIC);
|
|
if (!fce)
|
|
goto exit;
|
|
|
|
fce->name = kstrdup_const(name, GFP_ATOMIC);
|
|
if (!fce->name) {
|
|
kfree(fce);
|
|
fce = NULL;
|
|
goto exit;
|
|
}
|
|
exit:
|
|
return fce;
|
|
}
|
|
|
|
static int __fw_entry_found(const char *name)
|
|
{
|
|
struct firmware_cache *fwc = &fw_cache;
|
|
struct fw_cache_entry *fce;
|
|
|
|
list_for_each_entry(fce, &fwc->fw_names, list) {
|
|
if (!strcmp(fce->name, name))
|
|
return 1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int fw_cache_piggyback_on_request(const char *name)
|
|
{
|
|
struct firmware_cache *fwc = &fw_cache;
|
|
struct fw_cache_entry *fce;
|
|
int ret = 0;
|
|
|
|
spin_lock(&fwc->name_lock);
|
|
if (__fw_entry_found(name))
|
|
goto found;
|
|
|
|
fce = alloc_fw_cache_entry(name);
|
|
if (fce) {
|
|
ret = 1;
|
|
list_add(&fce->list, &fwc->fw_names);
|
|
pr_debug("%s: fw: %s\n", __func__, name);
|
|
}
|
|
found:
|
|
spin_unlock(&fwc->name_lock);
|
|
return ret;
|
|
}
|
|
|
|
static void free_fw_cache_entry(struct fw_cache_entry *fce)
|
|
{
|
|
kfree_const(fce->name);
|
|
kfree(fce);
|
|
}
|
|
|
|
static void __async_dev_cache_fw_image(void *fw_entry,
|
|
async_cookie_t cookie)
|
|
{
|
|
struct fw_cache_entry *fce = fw_entry;
|
|
struct firmware_cache *fwc = &fw_cache;
|
|
int ret;
|
|
|
|
ret = cache_firmware(fce->name);
|
|
if (ret) {
|
|
spin_lock(&fwc->name_lock);
|
|
list_del(&fce->list);
|
|
spin_unlock(&fwc->name_lock);
|
|
|
|
free_fw_cache_entry(fce);
|
|
}
|
|
}
|
|
|
|
/* called with dev->devres_lock held */
|
|
static void dev_create_fw_entry(struct device *dev, void *res,
|
|
void *data)
|
|
{
|
|
struct fw_name_devm *fwn = res;
|
|
const char *fw_name = fwn->name;
|
|
struct list_head *head = data;
|
|
struct fw_cache_entry *fce;
|
|
|
|
fce = alloc_fw_cache_entry(fw_name);
|
|
if (fce)
|
|
list_add(&fce->list, head);
|
|
}
|
|
|
|
static int devm_name_match(struct device *dev, void *res,
|
|
void *match_data)
|
|
{
|
|
struct fw_name_devm *fwn = res;
|
|
return (fwn->magic == (unsigned long)match_data);
|
|
}
|
|
|
|
static void dev_cache_fw_image(struct device *dev, void *data)
|
|
{
|
|
LIST_HEAD(todo);
|
|
struct fw_cache_entry *fce;
|
|
struct fw_cache_entry *fce_next;
|
|
struct firmware_cache *fwc = &fw_cache;
|
|
|
|
devres_for_each_res(dev, fw_name_devm_release,
|
|
devm_name_match, &fw_cache,
|
|
dev_create_fw_entry, &todo);
|
|
|
|
list_for_each_entry_safe(fce, fce_next, &todo, list) {
|
|
list_del(&fce->list);
|
|
|
|
spin_lock(&fwc->name_lock);
|
|
/* only one cache entry for one firmware */
|
|
if (!__fw_entry_found(fce->name)) {
|
|
list_add(&fce->list, &fwc->fw_names);
|
|
} else {
|
|
free_fw_cache_entry(fce);
|
|
fce = NULL;
|
|
}
|
|
spin_unlock(&fwc->name_lock);
|
|
|
|
if (fce)
|
|
async_schedule_domain(__async_dev_cache_fw_image,
|
|
(void *)fce,
|
|
&fw_cache_domain);
|
|
}
|
|
}
|
|
|
|
static void __device_uncache_fw_images(void)
|
|
{
|
|
struct firmware_cache *fwc = &fw_cache;
|
|
struct fw_cache_entry *fce;
|
|
|
|
spin_lock(&fwc->name_lock);
|
|
while (!list_empty(&fwc->fw_names)) {
|
|
fce = list_entry(fwc->fw_names.next,
|
|
struct fw_cache_entry, list);
|
|
list_del(&fce->list);
|
|
spin_unlock(&fwc->name_lock);
|
|
|
|
uncache_firmware(fce->name);
|
|
free_fw_cache_entry(fce);
|
|
|
|
spin_lock(&fwc->name_lock);
|
|
}
|
|
spin_unlock(&fwc->name_lock);
|
|
}
|
|
|
|
/**
|
|
* device_cache_fw_images - cache devices' firmware
|
|
*
|
|
* If one device called request_firmware or its nowait version
|
|
* successfully before, the firmware names are recored into the
|
|
* device's devres link list, so device_cache_fw_images can call
|
|
* cache_firmware() to cache these firmwares for the device,
|
|
* then the device driver can load its firmwares easily at
|
|
* time when system is not ready to complete loading firmware.
|
|
*/
|
|
static void device_cache_fw_images(void)
|
|
{
|
|
struct firmware_cache *fwc = &fw_cache;
|
|
int old_timeout;
|
|
DEFINE_WAIT(wait);
|
|
|
|
pr_debug("%s\n", __func__);
|
|
|
|
/* cancel uncache work */
|
|
cancel_delayed_work_sync(&fwc->work);
|
|
|
|
/*
|
|
* use small loading timeout for caching devices' firmware
|
|
* because all these firmware images have been loaded
|
|
* successfully at lease once, also system is ready for
|
|
* completing firmware loading now. The maximum size of
|
|
* firmware in current distributions is about 2M bytes,
|
|
* so 10 secs should be enough.
|
|
*/
|
|
old_timeout = loading_timeout;
|
|
loading_timeout = 10;
|
|
|
|
mutex_lock(&fw_lock);
|
|
fwc->state = FW_LOADER_START_CACHE;
|
|
dpm_for_each_dev(NULL, dev_cache_fw_image);
|
|
mutex_unlock(&fw_lock);
|
|
|
|
/* wait for completion of caching firmware for all devices */
|
|
async_synchronize_full_domain(&fw_cache_domain);
|
|
|
|
loading_timeout = old_timeout;
|
|
}
|
|
|
|
/**
|
|
* device_uncache_fw_images - uncache devices' firmware
|
|
*
|
|
* uncache all firmwares which have been cached successfully
|
|
* by device_uncache_fw_images earlier
|
|
*/
|
|
static void device_uncache_fw_images(void)
|
|
{
|
|
pr_debug("%s\n", __func__);
|
|
__device_uncache_fw_images();
|
|
}
|
|
|
|
static void device_uncache_fw_images_work(struct work_struct *work)
|
|
{
|
|
device_uncache_fw_images();
|
|
}
|
|
|
|
/**
|
|
* device_uncache_fw_images_delay - uncache devices firmwares
|
|
* @delay: number of milliseconds to delay uncache device firmwares
|
|
*
|
|
* uncache all devices's firmwares which has been cached successfully
|
|
* by device_cache_fw_images after @delay milliseconds.
|
|
*/
|
|
static void device_uncache_fw_images_delay(unsigned long delay)
|
|
{
|
|
queue_delayed_work(system_power_efficient_wq, &fw_cache.work,
|
|
msecs_to_jiffies(delay));
|
|
}
|
|
|
|
/**
|
|
* fw_pm_notify - notifier for suspend/resume
|
|
* @notify_block: unused
|
|
* @mode: mode we are switching to
|
|
* @unused: unused
|
|
*
|
|
* Used to modify the firmware_class state as we move in between states.
|
|
* The firmware_class implements a firmware cache to enable device driver
|
|
* to fetch firmware upon resume before the root filesystem is ready. We
|
|
* disable API calls which do not use the built-in firmware or the firmware
|
|
* cache when we know these calls will not work.
|
|
*
|
|
* The inner logic behind all this is a bit complex so it is worth summarizing
|
|
* the kernel's own suspend/resume process with context and focus on how this
|
|
* can impact the firmware API.
|
|
*
|
|
* First a review on how we go to suspend::
|
|
*
|
|
* pm_suspend() --> enter_state() -->
|
|
* sys_sync()
|
|
* suspend_prepare() -->
|
|
* __pm_notifier_call_chain(PM_SUSPEND_PREPARE, ...);
|
|
* suspend_freeze_processes() -->
|
|
* freeze_processes() -->
|
|
* __usermodehelper_set_disable_depth(UMH_DISABLED);
|
|
* freeze all tasks ...
|
|
* freeze_kernel_threads()
|
|
* suspend_devices_and_enter() -->
|
|
* dpm_suspend_start() -->
|
|
* dpm_prepare()
|
|
* dpm_suspend()
|
|
* suspend_enter() -->
|
|
* platform_suspend_prepare()
|
|
* dpm_suspend_late()
|
|
* freeze_enter()
|
|
* syscore_suspend()
|
|
*
|
|
* When we resume we bail out of a loop from suspend_devices_and_enter() and
|
|
* unwind back out to the caller enter_state() where we were before as follows::
|
|
*
|
|
* enter_state() -->
|
|
* suspend_devices_and_enter() --> (bail from loop)
|
|
* dpm_resume_end() -->
|
|
* dpm_resume()
|
|
* dpm_complete()
|
|
* suspend_finish() -->
|
|
* suspend_thaw_processes() -->
|
|
* thaw_processes() -->
|
|
* __usermodehelper_set_disable_depth(UMH_FREEZING);
|
|
* thaw_workqueues();
|
|
* thaw all processes ...
|
|
* usermodehelper_enable();
|
|
* pm_notifier_call_chain(PM_POST_SUSPEND);
|
|
*
|
|
* fw_pm_notify() works through pm_notifier_call_chain().
|
|
*/
|
|
static int fw_pm_notify(struct notifier_block *notify_block,
|
|
unsigned long mode, void *unused)
|
|
{
|
|
switch (mode) {
|
|
case PM_HIBERNATION_PREPARE:
|
|
case PM_SUSPEND_PREPARE:
|
|
case PM_RESTORE_PREPARE:
|
|
/*
|
|
* kill pending fallback requests with a custom fallback
|
|
* to avoid stalling suspend.
|
|
*/
|
|
kill_pending_fw_fallback_reqs(true);
|
|
device_cache_fw_images();
|
|
disable_firmware();
|
|
break;
|
|
|
|
case PM_POST_SUSPEND:
|
|
case PM_POST_HIBERNATION:
|
|
case PM_POST_RESTORE:
|
|
/*
|
|
* In case that system sleep failed and syscore_suspend is
|
|
* not called.
|
|
*/
|
|
mutex_lock(&fw_lock);
|
|
fw_cache.state = FW_LOADER_NO_CACHE;
|
|
mutex_unlock(&fw_lock);
|
|
enable_firmware();
|
|
|
|
device_uncache_fw_images_delay(10 * MSEC_PER_SEC);
|
|
break;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* stop caching firmware once syscore_suspend is reached */
|
|
static int fw_suspend(void)
|
|
{
|
|
fw_cache.state = FW_LOADER_NO_CACHE;
|
|
return 0;
|
|
}
|
|
|
|
static struct syscore_ops fw_syscore_ops = {
|
|
.suspend = fw_suspend,
|
|
};
|
|
#else
|
|
static int fw_cache_piggyback_on_request(const char *name)
|
|
{
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
static void __init fw_cache_init(void)
|
|
{
|
|
spin_lock_init(&fw_cache.lock);
|
|
INIT_LIST_HEAD(&fw_cache.head);
|
|
fw_cache.state = FW_LOADER_NO_CACHE;
|
|
|
|
#ifdef CONFIG_PM_SLEEP
|
|
spin_lock_init(&fw_cache.name_lock);
|
|
INIT_LIST_HEAD(&fw_cache.fw_names);
|
|
|
|
INIT_DELAYED_WORK(&fw_cache.work,
|
|
device_uncache_fw_images_work);
|
|
|
|
fw_cache.pm_notify.notifier_call = fw_pm_notify;
|
|
register_pm_notifier(&fw_cache.pm_notify);
|
|
|
|
register_syscore_ops(&fw_syscore_ops);
|
|
#endif
|
|
}
|
|
|
|
static int fw_shutdown_notify(struct notifier_block *unused1,
|
|
unsigned long unused2, void *unused3)
|
|
{
|
|
disable_firmware();
|
|
/*
|
|
* Kill all pending fallback requests to avoid both stalling shutdown,
|
|
* and avoid a deadlock with the usermode_lock.
|
|
*/
|
|
kill_pending_fw_fallback_reqs(false);
|
|
|
|
return NOTIFY_DONE;
|
|
}
|
|
|
|
static struct notifier_block fw_shutdown_nb = {
|
|
.notifier_call = fw_shutdown_notify,
|
|
};
|
|
|
|
static int __init firmware_class_init(void)
|
|
{
|
|
enable_firmware();
|
|
fw_cache_init();
|
|
register_reboot_notifier(&fw_shutdown_nb);
|
|
#ifdef CONFIG_FW_LOADER_USER_HELPER
|
|
return class_register(&firmware_class);
|
|
#else
|
|
return 0;
|
|
#endif
|
|
}
|
|
|
|
static void __exit firmware_class_exit(void)
|
|
{
|
|
disable_firmware();
|
|
#ifdef CONFIG_PM_SLEEP
|
|
unregister_syscore_ops(&fw_syscore_ops);
|
|
unregister_pm_notifier(&fw_cache.pm_notify);
|
|
#endif
|
|
unregister_reboot_notifier(&fw_shutdown_nb);
|
|
#ifdef CONFIG_FW_LOADER_USER_HELPER
|
|
class_unregister(&firmware_class);
|
|
#endif
|
|
}
|
|
|
|
fs_initcall(firmware_class_init);
|
|
module_exit(firmware_class_exit);
|