mirror of
https://github.com/torvalds/linux.git
synced 2024-11-14 16:12:02 +00:00
d25282d1c9
Pull module signing support from Rusty Russell: "module signing is the highlight, but it's an all-over David Howells frenzy..." Hmm "Magrathea: Glacier signing key". Somebody has been reading too much HHGTTG. * 'modules-next' of git://git.kernel.org/pub/scm/linux/kernel/git/rusty/linux: (37 commits) X.509: Fix indefinite length element skip error handling X.509: Convert some printk calls to pr_devel asymmetric keys: fix printk format warning MODSIGN: Fix 32-bit overflow in X.509 certificate validity date checking MODSIGN: Make mrproper should remove generated files. MODSIGN: Use utf8 strings in signer's name in autogenerated X.509 certs MODSIGN: Use the same digest for the autogen key sig as for the module sig MODSIGN: Sign modules during the build process MODSIGN: Provide a script for generating a key ID from an X.509 cert MODSIGN: Implement module signature checking MODSIGN: Provide module signing public keys to the kernel MODSIGN: Automatically generate module signing keys if missing MODSIGN: Provide Kconfig options MODSIGN: Provide gitignore and make clean rules for extra files MODSIGN: Add FIPS policy module: signature checking hook X.509: Add a crypto key parser for binary (DER) X.509 certificates MPILIB: Provide a function to read raw data into an MPI X.509: Add an ASN.1 decoder X.509: Add simple ASN.1 grammar compiler ...
1358 lines
36 KiB
C
1358 lines
36 KiB
C
/*
|
|
* fs/cifs/cifsacl.c
|
|
*
|
|
* Copyright (C) International Business Machines Corp., 2007,2008
|
|
* Author(s): Steve French (sfrench@us.ibm.com)
|
|
*
|
|
* Contains the routines for mapping CIFS/NTFS ACLs
|
|
*
|
|
* This library is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU Lesser General Public License as published
|
|
* by the Free Software Foundation; either version 2.1 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This library is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See
|
|
* the GNU Lesser General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU Lesser General Public License
|
|
* along with this library; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA
|
|
*/
|
|
|
|
#include <linux/fs.h>
|
|
#include <linux/slab.h>
|
|
#include <linux/string.h>
|
|
#include <linux/keyctl.h>
|
|
#include <linux/key-type.h>
|
|
#include <keys/user-type.h>
|
|
#include "cifspdu.h"
|
|
#include "cifsglob.h"
|
|
#include "cifsacl.h"
|
|
#include "cifsproto.h"
|
|
#include "cifs_debug.h"
|
|
|
|
/* security id for everyone/world system group */
|
|
static const struct cifs_sid sid_everyone = {
|
|
1, 1, {0, 0, 0, 0, 0, 1}, {0} };
|
|
/* security id for Authenticated Users system group */
|
|
static const struct cifs_sid sid_authusers = {
|
|
1, 1, {0, 0, 0, 0, 0, 5}, {__constant_cpu_to_le32(11)} };
|
|
/* group users */
|
|
static const struct cifs_sid sid_user = {1, 2 , {0, 0, 0, 0, 0, 5}, {} };
|
|
|
|
const struct cred *root_cred;
|
|
|
|
static void
|
|
shrink_idmap_tree(struct rb_root *root, int nr_to_scan, int *nr_rem,
|
|
int *nr_del)
|
|
{
|
|
struct rb_node *node;
|
|
struct rb_node *tmp;
|
|
struct cifs_sid_id *psidid;
|
|
|
|
node = rb_first(root);
|
|
while (node) {
|
|
tmp = node;
|
|
node = rb_next(tmp);
|
|
psidid = rb_entry(tmp, struct cifs_sid_id, rbnode);
|
|
if (nr_to_scan == 0 || *nr_del == nr_to_scan)
|
|
++(*nr_rem);
|
|
else {
|
|
if (time_after(jiffies, psidid->time + SID_MAP_EXPIRE)
|
|
&& psidid->refcount == 0) {
|
|
rb_erase(tmp, root);
|
|
++(*nr_del);
|
|
} else
|
|
++(*nr_rem);
|
|
}
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Run idmap cache shrinker.
|
|
*/
|
|
static int
|
|
cifs_idmap_shrinker(struct shrinker *shrink, struct shrink_control *sc)
|
|
{
|
|
int nr_to_scan = sc->nr_to_scan;
|
|
int nr_del = 0;
|
|
int nr_rem = 0;
|
|
struct rb_root *root;
|
|
|
|
root = &uidtree;
|
|
spin_lock(&siduidlock);
|
|
shrink_idmap_tree(root, nr_to_scan, &nr_rem, &nr_del);
|
|
spin_unlock(&siduidlock);
|
|
|
|
root = &gidtree;
|
|
spin_lock(&sidgidlock);
|
|
shrink_idmap_tree(root, nr_to_scan, &nr_rem, &nr_del);
|
|
spin_unlock(&sidgidlock);
|
|
|
|
root = &siduidtree;
|
|
spin_lock(&uidsidlock);
|
|
shrink_idmap_tree(root, nr_to_scan, &nr_rem, &nr_del);
|
|
spin_unlock(&uidsidlock);
|
|
|
|
root = &sidgidtree;
|
|
spin_lock(&gidsidlock);
|
|
shrink_idmap_tree(root, nr_to_scan, &nr_rem, &nr_del);
|
|
spin_unlock(&gidsidlock);
|
|
|
|
return nr_rem;
|
|
}
|
|
|
|
static void
|
|
sid_rb_insert(struct rb_root *root, unsigned long cid,
|
|
struct cifs_sid_id **psidid, char *typestr)
|
|
{
|
|
char *strptr;
|
|
struct rb_node *node = root->rb_node;
|
|
struct rb_node *parent = NULL;
|
|
struct rb_node **linkto = &(root->rb_node);
|
|
struct cifs_sid_id *lsidid;
|
|
|
|
while (node) {
|
|
lsidid = rb_entry(node, struct cifs_sid_id, rbnode);
|
|
parent = node;
|
|
if (cid > lsidid->id) {
|
|
linkto = &(node->rb_left);
|
|
node = node->rb_left;
|
|
}
|
|
if (cid < lsidid->id) {
|
|
linkto = &(node->rb_right);
|
|
node = node->rb_right;
|
|
}
|
|
}
|
|
|
|
(*psidid)->id = cid;
|
|
(*psidid)->time = jiffies - (SID_MAP_RETRY + 1);
|
|
(*psidid)->refcount = 0;
|
|
|
|
sprintf((*psidid)->sidstr, "%s", typestr);
|
|
strptr = (*psidid)->sidstr + strlen((*psidid)->sidstr);
|
|
sprintf(strptr, "%ld", cid);
|
|
|
|
clear_bit(SID_ID_PENDING, &(*psidid)->state);
|
|
clear_bit(SID_ID_MAPPED, &(*psidid)->state);
|
|
|
|
rb_link_node(&(*psidid)->rbnode, parent, linkto);
|
|
rb_insert_color(&(*psidid)->rbnode, root);
|
|
}
|
|
|
|
static struct cifs_sid_id *
|
|
sid_rb_search(struct rb_root *root, unsigned long cid)
|
|
{
|
|
struct rb_node *node = root->rb_node;
|
|
struct cifs_sid_id *lsidid;
|
|
|
|
while (node) {
|
|
lsidid = rb_entry(node, struct cifs_sid_id, rbnode);
|
|
if (cid > lsidid->id)
|
|
node = node->rb_left;
|
|
else if (cid < lsidid->id)
|
|
node = node->rb_right;
|
|
else /* node found */
|
|
return lsidid;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static struct shrinker cifs_shrinker = {
|
|
.shrink = cifs_idmap_shrinker,
|
|
.seeks = DEFAULT_SEEKS,
|
|
};
|
|
|
|
static int
|
|
cifs_idmap_key_instantiate(struct key *key, struct key_preparsed_payload *prep)
|
|
{
|
|
char *payload;
|
|
|
|
payload = kmalloc(prep->datalen, GFP_KERNEL);
|
|
if (!payload)
|
|
return -ENOMEM;
|
|
|
|
memcpy(payload, prep->data, prep->datalen);
|
|
key->payload.data = payload;
|
|
key->datalen = prep->datalen;
|
|
return 0;
|
|
}
|
|
|
|
static inline void
|
|
cifs_idmap_key_destroy(struct key *key)
|
|
{
|
|
kfree(key->payload.data);
|
|
}
|
|
|
|
struct key_type cifs_idmap_key_type = {
|
|
.name = "cifs.idmap",
|
|
.instantiate = cifs_idmap_key_instantiate,
|
|
.destroy = cifs_idmap_key_destroy,
|
|
.describe = user_describe,
|
|
.match = user_match,
|
|
};
|
|
|
|
static void
|
|
sid_to_str(struct cifs_sid *sidptr, char *sidstr)
|
|
{
|
|
int i;
|
|
unsigned long saval;
|
|
char *strptr;
|
|
|
|
strptr = sidstr;
|
|
|
|
sprintf(strptr, "%s", "S");
|
|
strptr = sidstr + strlen(sidstr);
|
|
|
|
sprintf(strptr, "-%d", sidptr->revision);
|
|
strptr = sidstr + strlen(sidstr);
|
|
|
|
for (i = 0; i < 6; ++i) {
|
|
if (sidptr->authority[i]) {
|
|
sprintf(strptr, "-%d", sidptr->authority[i]);
|
|
strptr = sidstr + strlen(sidstr);
|
|
}
|
|
}
|
|
|
|
for (i = 0; i < sidptr->num_subauth; ++i) {
|
|
saval = le32_to_cpu(sidptr->sub_auth[i]);
|
|
sprintf(strptr, "-%ld", saval);
|
|
strptr = sidstr + strlen(sidstr);
|
|
}
|
|
}
|
|
|
|
static void
|
|
id_rb_insert(struct rb_root *root, struct cifs_sid *sidptr,
|
|
struct cifs_sid_id **psidid, char *typestr)
|
|
{
|
|
int rc;
|
|
char *strptr;
|
|
struct rb_node *node = root->rb_node;
|
|
struct rb_node *parent = NULL;
|
|
struct rb_node **linkto = &(root->rb_node);
|
|
struct cifs_sid_id *lsidid;
|
|
|
|
while (node) {
|
|
lsidid = rb_entry(node, struct cifs_sid_id, rbnode);
|
|
parent = node;
|
|
rc = compare_sids(sidptr, &((lsidid)->sid));
|
|
if (rc > 0) {
|
|
linkto = &(node->rb_left);
|
|
node = node->rb_left;
|
|
} else if (rc < 0) {
|
|
linkto = &(node->rb_right);
|
|
node = node->rb_right;
|
|
}
|
|
}
|
|
|
|
memcpy(&(*psidid)->sid, sidptr, sizeof(struct cifs_sid));
|
|
(*psidid)->time = jiffies - (SID_MAP_RETRY + 1);
|
|
(*psidid)->refcount = 0;
|
|
|
|
sprintf((*psidid)->sidstr, "%s", typestr);
|
|
strptr = (*psidid)->sidstr + strlen((*psidid)->sidstr);
|
|
sid_to_str(&(*psidid)->sid, strptr);
|
|
|
|
clear_bit(SID_ID_PENDING, &(*psidid)->state);
|
|
clear_bit(SID_ID_MAPPED, &(*psidid)->state);
|
|
|
|
rb_link_node(&(*psidid)->rbnode, parent, linkto);
|
|
rb_insert_color(&(*psidid)->rbnode, root);
|
|
}
|
|
|
|
static struct cifs_sid_id *
|
|
id_rb_search(struct rb_root *root, struct cifs_sid *sidptr)
|
|
{
|
|
int rc;
|
|
struct rb_node *node = root->rb_node;
|
|
struct cifs_sid_id *lsidid;
|
|
|
|
while (node) {
|
|
lsidid = rb_entry(node, struct cifs_sid_id, rbnode);
|
|
rc = compare_sids(sidptr, &((lsidid)->sid));
|
|
if (rc > 0) {
|
|
node = node->rb_left;
|
|
} else if (rc < 0) {
|
|
node = node->rb_right;
|
|
} else /* node found */
|
|
return lsidid;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static int
|
|
sidid_pending_wait(void *unused)
|
|
{
|
|
schedule();
|
|
return signal_pending(current) ? -ERESTARTSYS : 0;
|
|
}
|
|
|
|
static int
|
|
id_to_sid(unsigned long cid, uint sidtype, struct cifs_sid *ssid)
|
|
{
|
|
int rc = 0;
|
|
struct key *sidkey;
|
|
const struct cred *saved_cred;
|
|
struct cifs_sid *lsid;
|
|
struct cifs_sid_id *psidid, *npsidid;
|
|
struct rb_root *cidtree;
|
|
spinlock_t *cidlock;
|
|
|
|
if (sidtype == SIDOWNER) {
|
|
cidlock = &siduidlock;
|
|
cidtree = &uidtree;
|
|
} else if (sidtype == SIDGROUP) {
|
|
cidlock = &sidgidlock;
|
|
cidtree = &gidtree;
|
|
} else
|
|
return -EINVAL;
|
|
|
|
spin_lock(cidlock);
|
|
psidid = sid_rb_search(cidtree, cid);
|
|
|
|
if (!psidid) { /* node does not exist, allocate one & attempt adding */
|
|
spin_unlock(cidlock);
|
|
npsidid = kzalloc(sizeof(struct cifs_sid_id), GFP_KERNEL);
|
|
if (!npsidid)
|
|
return -ENOMEM;
|
|
|
|
npsidid->sidstr = kmalloc(SIDLEN, GFP_KERNEL);
|
|
if (!npsidid->sidstr) {
|
|
kfree(npsidid);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
spin_lock(cidlock);
|
|
psidid = sid_rb_search(cidtree, cid);
|
|
if (psidid) { /* node happened to get inserted meanwhile */
|
|
++psidid->refcount;
|
|
spin_unlock(cidlock);
|
|
kfree(npsidid->sidstr);
|
|
kfree(npsidid);
|
|
} else {
|
|
psidid = npsidid;
|
|
sid_rb_insert(cidtree, cid, &psidid,
|
|
sidtype == SIDOWNER ? "oi:" : "gi:");
|
|
++psidid->refcount;
|
|
spin_unlock(cidlock);
|
|
}
|
|
} else {
|
|
++psidid->refcount;
|
|
spin_unlock(cidlock);
|
|
}
|
|
|
|
/*
|
|
* If we are here, it is safe to access psidid and its fields
|
|
* since a reference was taken earlier while holding the spinlock.
|
|
* A reference on the node is put without holding the spinlock
|
|
* and it is OK to do so in this case, shrinker will not erase
|
|
* this node until all references are put and we do not access
|
|
* any fields of the node after a reference is put .
|
|
*/
|
|
if (test_bit(SID_ID_MAPPED, &psidid->state)) {
|
|
memcpy(ssid, &psidid->sid, sizeof(struct cifs_sid));
|
|
psidid->time = jiffies; /* update ts for accessing */
|
|
goto id_sid_out;
|
|
}
|
|
|
|
if (time_after(psidid->time + SID_MAP_RETRY, jiffies)) {
|
|
rc = -EINVAL;
|
|
goto id_sid_out;
|
|
}
|
|
|
|
if (!test_and_set_bit(SID_ID_PENDING, &psidid->state)) {
|
|
saved_cred = override_creds(root_cred);
|
|
sidkey = request_key(&cifs_idmap_key_type, psidid->sidstr, "");
|
|
if (IS_ERR(sidkey)) {
|
|
rc = -EINVAL;
|
|
cFYI(1, "%s: Can't map and id to a SID", __func__);
|
|
} else {
|
|
lsid = (struct cifs_sid *)sidkey->payload.data;
|
|
memcpy(&psidid->sid, lsid,
|
|
sidkey->datalen < sizeof(struct cifs_sid) ?
|
|
sidkey->datalen : sizeof(struct cifs_sid));
|
|
memcpy(ssid, &psidid->sid,
|
|
sidkey->datalen < sizeof(struct cifs_sid) ?
|
|
sidkey->datalen : sizeof(struct cifs_sid));
|
|
set_bit(SID_ID_MAPPED, &psidid->state);
|
|
key_put(sidkey);
|
|
kfree(psidid->sidstr);
|
|
}
|
|
psidid->time = jiffies; /* update ts for accessing */
|
|
revert_creds(saved_cred);
|
|
clear_bit(SID_ID_PENDING, &psidid->state);
|
|
wake_up_bit(&psidid->state, SID_ID_PENDING);
|
|
} else {
|
|
rc = wait_on_bit(&psidid->state, SID_ID_PENDING,
|
|
sidid_pending_wait, TASK_INTERRUPTIBLE);
|
|
if (rc) {
|
|
cFYI(1, "%s: sidid_pending_wait interrupted %d",
|
|
__func__, rc);
|
|
--psidid->refcount;
|
|
return rc;
|
|
}
|
|
if (test_bit(SID_ID_MAPPED, &psidid->state))
|
|
memcpy(ssid, &psidid->sid, sizeof(struct cifs_sid));
|
|
else
|
|
rc = -EINVAL;
|
|
}
|
|
id_sid_out:
|
|
--psidid->refcount;
|
|
return rc;
|
|
}
|
|
|
|
static int
|
|
sid_to_id(struct cifs_sb_info *cifs_sb, struct cifs_sid *psid,
|
|
struct cifs_fattr *fattr, uint sidtype)
|
|
{
|
|
int rc;
|
|
unsigned long cid;
|
|
struct key *idkey;
|
|
const struct cred *saved_cred;
|
|
struct cifs_sid_id *psidid, *npsidid;
|
|
struct rb_root *cidtree;
|
|
spinlock_t *cidlock;
|
|
|
|
if (sidtype == SIDOWNER) {
|
|
cid = cifs_sb->mnt_uid; /* default uid, in case upcall fails */
|
|
cidlock = &siduidlock;
|
|
cidtree = &uidtree;
|
|
} else if (sidtype == SIDGROUP) {
|
|
cid = cifs_sb->mnt_gid; /* default gid, in case upcall fails */
|
|
cidlock = &sidgidlock;
|
|
cidtree = &gidtree;
|
|
} else
|
|
return -ENOENT;
|
|
|
|
spin_lock(cidlock);
|
|
psidid = id_rb_search(cidtree, psid);
|
|
|
|
if (!psidid) { /* node does not exist, allocate one & attempt adding */
|
|
spin_unlock(cidlock);
|
|
npsidid = kzalloc(sizeof(struct cifs_sid_id), GFP_KERNEL);
|
|
if (!npsidid)
|
|
return -ENOMEM;
|
|
|
|
npsidid->sidstr = kmalloc(SIDLEN, GFP_KERNEL);
|
|
if (!npsidid->sidstr) {
|
|
kfree(npsidid);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
spin_lock(cidlock);
|
|
psidid = id_rb_search(cidtree, psid);
|
|
if (psidid) { /* node happened to get inserted meanwhile */
|
|
++psidid->refcount;
|
|
spin_unlock(cidlock);
|
|
kfree(npsidid->sidstr);
|
|
kfree(npsidid);
|
|
} else {
|
|
psidid = npsidid;
|
|
id_rb_insert(cidtree, psid, &psidid,
|
|
sidtype == SIDOWNER ? "os:" : "gs:");
|
|
++psidid->refcount;
|
|
spin_unlock(cidlock);
|
|
}
|
|
} else {
|
|
++psidid->refcount;
|
|
spin_unlock(cidlock);
|
|
}
|
|
|
|
/*
|
|
* If we are here, it is safe to access psidid and its fields
|
|
* since a reference was taken earlier while holding the spinlock.
|
|
* A reference on the node is put without holding the spinlock
|
|
* and it is OK to do so in this case, shrinker will not erase
|
|
* this node until all references are put and we do not access
|
|
* any fields of the node after a reference is put .
|
|
*/
|
|
if (test_bit(SID_ID_MAPPED, &psidid->state)) {
|
|
cid = psidid->id;
|
|
psidid->time = jiffies; /* update ts for accessing */
|
|
goto sid_to_id_out;
|
|
}
|
|
|
|
if (time_after(psidid->time + SID_MAP_RETRY, jiffies))
|
|
goto sid_to_id_out;
|
|
|
|
if (!test_and_set_bit(SID_ID_PENDING, &psidid->state)) {
|
|
saved_cred = override_creds(root_cred);
|
|
idkey = request_key(&cifs_idmap_key_type, psidid->sidstr, "");
|
|
if (IS_ERR(idkey))
|
|
cFYI(1, "%s: Can't map SID to an id", __func__);
|
|
else {
|
|
cid = *(unsigned long *)idkey->payload.value;
|
|
psidid->id = cid;
|
|
set_bit(SID_ID_MAPPED, &psidid->state);
|
|
key_put(idkey);
|
|
kfree(psidid->sidstr);
|
|
}
|
|
revert_creds(saved_cred);
|
|
psidid->time = jiffies; /* update ts for accessing */
|
|
clear_bit(SID_ID_PENDING, &psidid->state);
|
|
wake_up_bit(&psidid->state, SID_ID_PENDING);
|
|
} else {
|
|
rc = wait_on_bit(&psidid->state, SID_ID_PENDING,
|
|
sidid_pending_wait, TASK_INTERRUPTIBLE);
|
|
if (rc) {
|
|
cFYI(1, "%s: sidid_pending_wait interrupted %d",
|
|
__func__, rc);
|
|
--psidid->refcount; /* decremented without spinlock */
|
|
return rc;
|
|
}
|
|
if (test_bit(SID_ID_MAPPED, &psidid->state))
|
|
cid = psidid->id;
|
|
}
|
|
|
|
sid_to_id_out:
|
|
--psidid->refcount; /* decremented without spinlock */
|
|
if (sidtype == SIDOWNER)
|
|
fattr->cf_uid = cid;
|
|
else
|
|
fattr->cf_gid = cid;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int
|
|
init_cifs_idmap(void)
|
|
{
|
|
struct cred *cred;
|
|
struct key *keyring;
|
|
int ret;
|
|
|
|
cFYI(1, "Registering the %s key type", cifs_idmap_key_type.name);
|
|
|
|
/* create an override credential set with a special thread keyring in
|
|
* which requests are cached
|
|
*
|
|
* this is used to prevent malicious redirections from being installed
|
|
* with add_key().
|
|
*/
|
|
cred = prepare_kernel_cred(NULL);
|
|
if (!cred)
|
|
return -ENOMEM;
|
|
|
|
keyring = key_alloc(&key_type_keyring, ".cifs_idmap", 0, 0, cred,
|
|
(KEY_POS_ALL & ~KEY_POS_SETATTR) |
|
|
KEY_USR_VIEW | KEY_USR_READ,
|
|
KEY_ALLOC_NOT_IN_QUOTA);
|
|
if (IS_ERR(keyring)) {
|
|
ret = PTR_ERR(keyring);
|
|
goto failed_put_cred;
|
|
}
|
|
|
|
ret = key_instantiate_and_link(keyring, NULL, 0, NULL, NULL);
|
|
if (ret < 0)
|
|
goto failed_put_key;
|
|
|
|
ret = register_key_type(&cifs_idmap_key_type);
|
|
if (ret < 0)
|
|
goto failed_put_key;
|
|
|
|
/* instruct request_key() to use this special keyring as a cache for
|
|
* the results it looks up */
|
|
set_bit(KEY_FLAG_ROOT_CAN_CLEAR, &keyring->flags);
|
|
cred->thread_keyring = keyring;
|
|
cred->jit_keyring = KEY_REQKEY_DEFL_THREAD_KEYRING;
|
|
root_cred = cred;
|
|
|
|
spin_lock_init(&siduidlock);
|
|
uidtree = RB_ROOT;
|
|
spin_lock_init(&sidgidlock);
|
|
gidtree = RB_ROOT;
|
|
|
|
spin_lock_init(&uidsidlock);
|
|
siduidtree = RB_ROOT;
|
|
spin_lock_init(&gidsidlock);
|
|
sidgidtree = RB_ROOT;
|
|
register_shrinker(&cifs_shrinker);
|
|
|
|
cFYI(1, "cifs idmap keyring: %d", key_serial(keyring));
|
|
return 0;
|
|
|
|
failed_put_key:
|
|
key_put(keyring);
|
|
failed_put_cred:
|
|
put_cred(cred);
|
|
return ret;
|
|
}
|
|
|
|
void
|
|
exit_cifs_idmap(void)
|
|
{
|
|
key_revoke(root_cred->thread_keyring);
|
|
unregister_key_type(&cifs_idmap_key_type);
|
|
put_cred(root_cred);
|
|
unregister_shrinker(&cifs_shrinker);
|
|
cFYI(1, "Unregistered %s key type", cifs_idmap_key_type.name);
|
|
}
|
|
|
|
void
|
|
cifs_destroy_idmaptrees(void)
|
|
{
|
|
struct rb_root *root;
|
|
struct rb_node *node;
|
|
|
|
root = &uidtree;
|
|
spin_lock(&siduidlock);
|
|
while ((node = rb_first(root)))
|
|
rb_erase(node, root);
|
|
spin_unlock(&siduidlock);
|
|
|
|
root = &gidtree;
|
|
spin_lock(&sidgidlock);
|
|
while ((node = rb_first(root)))
|
|
rb_erase(node, root);
|
|
spin_unlock(&sidgidlock);
|
|
|
|
root = &siduidtree;
|
|
spin_lock(&uidsidlock);
|
|
while ((node = rb_first(root)))
|
|
rb_erase(node, root);
|
|
spin_unlock(&uidsidlock);
|
|
|
|
root = &sidgidtree;
|
|
spin_lock(&gidsidlock);
|
|
while ((node = rb_first(root)))
|
|
rb_erase(node, root);
|
|
spin_unlock(&gidsidlock);
|
|
}
|
|
|
|
/* if the two SIDs (roughly equivalent to a UUID for a user or group) are
|
|
the same returns 1, if they do not match returns 0 */
|
|
int compare_sids(const struct cifs_sid *ctsid, const struct cifs_sid *cwsid)
|
|
{
|
|
int i;
|
|
int num_subauth, num_sat, num_saw;
|
|
|
|
if ((!ctsid) || (!cwsid))
|
|
return 1;
|
|
|
|
/* compare the revision */
|
|
if (ctsid->revision != cwsid->revision) {
|
|
if (ctsid->revision > cwsid->revision)
|
|
return 1;
|
|
else
|
|
return -1;
|
|
}
|
|
|
|
/* compare all of the six auth values */
|
|
for (i = 0; i < 6; ++i) {
|
|
if (ctsid->authority[i] != cwsid->authority[i]) {
|
|
if (ctsid->authority[i] > cwsid->authority[i])
|
|
return 1;
|
|
else
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
/* compare all of the subauth values if any */
|
|
num_sat = ctsid->num_subauth;
|
|
num_saw = cwsid->num_subauth;
|
|
num_subauth = num_sat < num_saw ? num_sat : num_saw;
|
|
if (num_subauth) {
|
|
for (i = 0; i < num_subauth; ++i) {
|
|
if (ctsid->sub_auth[i] != cwsid->sub_auth[i]) {
|
|
if (le32_to_cpu(ctsid->sub_auth[i]) >
|
|
le32_to_cpu(cwsid->sub_auth[i]))
|
|
return 1;
|
|
else
|
|
return -1;
|
|
}
|
|
}
|
|
}
|
|
|
|
return 0; /* sids compare/match */
|
|
}
|
|
|
|
|
|
/* copy ntsd, owner sid, and group sid from a security descriptor to another */
|
|
static void copy_sec_desc(const struct cifs_ntsd *pntsd,
|
|
struct cifs_ntsd *pnntsd, __u32 sidsoffset)
|
|
{
|
|
int i;
|
|
|
|
struct cifs_sid *owner_sid_ptr, *group_sid_ptr;
|
|
struct cifs_sid *nowner_sid_ptr, *ngroup_sid_ptr;
|
|
|
|
/* copy security descriptor control portion */
|
|
pnntsd->revision = pntsd->revision;
|
|
pnntsd->type = pntsd->type;
|
|
pnntsd->dacloffset = cpu_to_le32(sizeof(struct cifs_ntsd));
|
|
pnntsd->sacloffset = 0;
|
|
pnntsd->osidoffset = cpu_to_le32(sidsoffset);
|
|
pnntsd->gsidoffset = cpu_to_le32(sidsoffset + sizeof(struct cifs_sid));
|
|
|
|
/* copy owner sid */
|
|
owner_sid_ptr = (struct cifs_sid *)((char *)pntsd +
|
|
le32_to_cpu(pntsd->osidoffset));
|
|
nowner_sid_ptr = (struct cifs_sid *)((char *)pnntsd + sidsoffset);
|
|
|
|
nowner_sid_ptr->revision = owner_sid_ptr->revision;
|
|
nowner_sid_ptr->num_subauth = owner_sid_ptr->num_subauth;
|
|
for (i = 0; i < 6; i++)
|
|
nowner_sid_ptr->authority[i] = owner_sid_ptr->authority[i];
|
|
for (i = 0; i < 5; i++)
|
|
nowner_sid_ptr->sub_auth[i] = owner_sid_ptr->sub_auth[i];
|
|
|
|
/* copy group sid */
|
|
group_sid_ptr = (struct cifs_sid *)((char *)pntsd +
|
|
le32_to_cpu(pntsd->gsidoffset));
|
|
ngroup_sid_ptr = (struct cifs_sid *)((char *)pnntsd + sidsoffset +
|
|
sizeof(struct cifs_sid));
|
|
|
|
ngroup_sid_ptr->revision = group_sid_ptr->revision;
|
|
ngroup_sid_ptr->num_subauth = group_sid_ptr->num_subauth;
|
|
for (i = 0; i < 6; i++)
|
|
ngroup_sid_ptr->authority[i] = group_sid_ptr->authority[i];
|
|
for (i = 0; i < 5; i++)
|
|
ngroup_sid_ptr->sub_auth[i] = group_sid_ptr->sub_auth[i];
|
|
|
|
return;
|
|
}
|
|
|
|
|
|
/*
|
|
change posix mode to reflect permissions
|
|
pmode is the existing mode (we only want to overwrite part of this
|
|
bits to set can be: S_IRWXU, S_IRWXG or S_IRWXO ie 00700 or 00070 or 00007
|
|
*/
|
|
static void access_flags_to_mode(__le32 ace_flags, int type, umode_t *pmode,
|
|
umode_t *pbits_to_set)
|
|
{
|
|
__u32 flags = le32_to_cpu(ace_flags);
|
|
/* the order of ACEs is important. The canonical order is to begin with
|
|
DENY entries followed by ALLOW, otherwise an allow entry could be
|
|
encountered first, making the subsequent deny entry like "dead code"
|
|
which would be superflous since Windows stops when a match is made
|
|
for the operation you are trying to perform for your user */
|
|
|
|
/* For deny ACEs we change the mask so that subsequent allow access
|
|
control entries do not turn on the bits we are denying */
|
|
if (type == ACCESS_DENIED) {
|
|
if (flags & GENERIC_ALL)
|
|
*pbits_to_set &= ~S_IRWXUGO;
|
|
|
|
if ((flags & GENERIC_WRITE) ||
|
|
((flags & FILE_WRITE_RIGHTS) == FILE_WRITE_RIGHTS))
|
|
*pbits_to_set &= ~S_IWUGO;
|
|
if ((flags & GENERIC_READ) ||
|
|
((flags & FILE_READ_RIGHTS) == FILE_READ_RIGHTS))
|
|
*pbits_to_set &= ~S_IRUGO;
|
|
if ((flags & GENERIC_EXECUTE) ||
|
|
((flags & FILE_EXEC_RIGHTS) == FILE_EXEC_RIGHTS))
|
|
*pbits_to_set &= ~S_IXUGO;
|
|
return;
|
|
} else if (type != ACCESS_ALLOWED) {
|
|
cERROR(1, "unknown access control type %d", type);
|
|
return;
|
|
}
|
|
/* else ACCESS_ALLOWED type */
|
|
|
|
if (flags & GENERIC_ALL) {
|
|
*pmode |= (S_IRWXUGO & (*pbits_to_set));
|
|
cFYI(DBG2, "all perms");
|
|
return;
|
|
}
|
|
if ((flags & GENERIC_WRITE) ||
|
|
((flags & FILE_WRITE_RIGHTS) == FILE_WRITE_RIGHTS))
|
|
*pmode |= (S_IWUGO & (*pbits_to_set));
|
|
if ((flags & GENERIC_READ) ||
|
|
((flags & FILE_READ_RIGHTS) == FILE_READ_RIGHTS))
|
|
*pmode |= (S_IRUGO & (*pbits_to_set));
|
|
if ((flags & GENERIC_EXECUTE) ||
|
|
((flags & FILE_EXEC_RIGHTS) == FILE_EXEC_RIGHTS))
|
|
*pmode |= (S_IXUGO & (*pbits_to_set));
|
|
|
|
cFYI(DBG2, "access flags 0x%x mode now 0x%x", flags, *pmode);
|
|
return;
|
|
}
|
|
|
|
/*
|
|
Generate access flags to reflect permissions mode is the existing mode.
|
|
This function is called for every ACE in the DACL whose SID matches
|
|
with either owner or group or everyone.
|
|
*/
|
|
|
|
static void mode_to_access_flags(umode_t mode, umode_t bits_to_use,
|
|
__u32 *pace_flags)
|
|
{
|
|
/* reset access mask */
|
|
*pace_flags = 0x0;
|
|
|
|
/* bits to use are either S_IRWXU or S_IRWXG or S_IRWXO */
|
|
mode &= bits_to_use;
|
|
|
|
/* check for R/W/X UGO since we do not know whose flags
|
|
is this but we have cleared all the bits sans RWX for
|
|
either user or group or other as per bits_to_use */
|
|
if (mode & S_IRUGO)
|
|
*pace_flags |= SET_FILE_READ_RIGHTS;
|
|
if (mode & S_IWUGO)
|
|
*pace_flags |= SET_FILE_WRITE_RIGHTS;
|
|
if (mode & S_IXUGO)
|
|
*pace_flags |= SET_FILE_EXEC_RIGHTS;
|
|
|
|
cFYI(DBG2, "mode: 0x%x, access flags now 0x%x", mode, *pace_flags);
|
|
return;
|
|
}
|
|
|
|
static __u16 fill_ace_for_sid(struct cifs_ace *pntace,
|
|
const struct cifs_sid *psid, __u64 nmode, umode_t bits)
|
|
{
|
|
int i;
|
|
__u16 size = 0;
|
|
__u32 access_req = 0;
|
|
|
|
pntace->type = ACCESS_ALLOWED;
|
|
pntace->flags = 0x0;
|
|
mode_to_access_flags(nmode, bits, &access_req);
|
|
if (!access_req)
|
|
access_req = SET_MINIMUM_RIGHTS;
|
|
pntace->access_req = cpu_to_le32(access_req);
|
|
|
|
pntace->sid.revision = psid->revision;
|
|
pntace->sid.num_subauth = psid->num_subauth;
|
|
for (i = 0; i < 6; i++)
|
|
pntace->sid.authority[i] = psid->authority[i];
|
|
for (i = 0; i < psid->num_subauth; i++)
|
|
pntace->sid.sub_auth[i] = psid->sub_auth[i];
|
|
|
|
size = 1 + 1 + 2 + 4 + 1 + 1 + 6 + (psid->num_subauth * 4);
|
|
pntace->size = cpu_to_le16(size);
|
|
|
|
return size;
|
|
}
|
|
|
|
|
|
#ifdef CONFIG_CIFS_DEBUG2
|
|
static void dump_ace(struct cifs_ace *pace, char *end_of_acl)
|
|
{
|
|
int num_subauth;
|
|
|
|
/* validate that we do not go past end of acl */
|
|
|
|
if (le16_to_cpu(pace->size) < 16) {
|
|
cERROR(1, "ACE too small %d", le16_to_cpu(pace->size));
|
|
return;
|
|
}
|
|
|
|
if (end_of_acl < (char *)pace + le16_to_cpu(pace->size)) {
|
|
cERROR(1, "ACL too small to parse ACE");
|
|
return;
|
|
}
|
|
|
|
num_subauth = pace->sid.num_subauth;
|
|
if (num_subauth) {
|
|
int i;
|
|
cFYI(1, "ACE revision %d num_auth %d type %d flags %d size %d",
|
|
pace->sid.revision, pace->sid.num_subauth, pace->type,
|
|
pace->flags, le16_to_cpu(pace->size));
|
|
for (i = 0; i < num_subauth; ++i) {
|
|
cFYI(1, "ACE sub_auth[%d]: 0x%x", i,
|
|
le32_to_cpu(pace->sid.sub_auth[i]));
|
|
}
|
|
|
|
/* BB add length check to make sure that we do not have huge
|
|
num auths and therefore go off the end */
|
|
}
|
|
|
|
return;
|
|
}
|
|
#endif
|
|
|
|
|
|
static void parse_dacl(struct cifs_acl *pdacl, char *end_of_acl,
|
|
struct cifs_sid *pownersid, struct cifs_sid *pgrpsid,
|
|
struct cifs_fattr *fattr)
|
|
{
|
|
int i;
|
|
int num_aces = 0;
|
|
int acl_size;
|
|
char *acl_base;
|
|
struct cifs_ace **ppace;
|
|
|
|
/* BB need to add parm so we can store the SID BB */
|
|
|
|
if (!pdacl) {
|
|
/* no DACL in the security descriptor, set
|
|
all the permissions for user/group/other */
|
|
fattr->cf_mode |= S_IRWXUGO;
|
|
return;
|
|
}
|
|
|
|
/* validate that we do not go past end of acl */
|
|
if (end_of_acl < (char *)pdacl + le16_to_cpu(pdacl->size)) {
|
|
cERROR(1, "ACL too small to parse DACL");
|
|
return;
|
|
}
|
|
|
|
cFYI(DBG2, "DACL revision %d size %d num aces %d",
|
|
le16_to_cpu(pdacl->revision), le16_to_cpu(pdacl->size),
|
|
le32_to_cpu(pdacl->num_aces));
|
|
|
|
/* reset rwx permissions for user/group/other.
|
|
Also, if num_aces is 0 i.e. DACL has no ACEs,
|
|
user/group/other have no permissions */
|
|
fattr->cf_mode &= ~(S_IRWXUGO);
|
|
|
|
acl_base = (char *)pdacl;
|
|
acl_size = sizeof(struct cifs_acl);
|
|
|
|
num_aces = le32_to_cpu(pdacl->num_aces);
|
|
if (num_aces > 0) {
|
|
umode_t user_mask = S_IRWXU;
|
|
umode_t group_mask = S_IRWXG;
|
|
umode_t other_mask = S_IRWXU | S_IRWXG | S_IRWXO;
|
|
|
|
if (num_aces > ULONG_MAX / sizeof(struct cifs_ace *))
|
|
return;
|
|
ppace = kmalloc(num_aces * sizeof(struct cifs_ace *),
|
|
GFP_KERNEL);
|
|
if (!ppace) {
|
|
cERROR(1, "DACL memory allocation error");
|
|
return;
|
|
}
|
|
|
|
for (i = 0; i < num_aces; ++i) {
|
|
ppace[i] = (struct cifs_ace *) (acl_base + acl_size);
|
|
#ifdef CONFIG_CIFS_DEBUG2
|
|
dump_ace(ppace[i], end_of_acl);
|
|
#endif
|
|
if (compare_sids(&(ppace[i]->sid), pownersid) == 0)
|
|
access_flags_to_mode(ppace[i]->access_req,
|
|
ppace[i]->type,
|
|
&fattr->cf_mode,
|
|
&user_mask);
|
|
if (compare_sids(&(ppace[i]->sid), pgrpsid) == 0)
|
|
access_flags_to_mode(ppace[i]->access_req,
|
|
ppace[i]->type,
|
|
&fattr->cf_mode,
|
|
&group_mask);
|
|
if (compare_sids(&(ppace[i]->sid), &sid_everyone) == 0)
|
|
access_flags_to_mode(ppace[i]->access_req,
|
|
ppace[i]->type,
|
|
&fattr->cf_mode,
|
|
&other_mask);
|
|
if (compare_sids(&(ppace[i]->sid), &sid_authusers) == 0)
|
|
access_flags_to_mode(ppace[i]->access_req,
|
|
ppace[i]->type,
|
|
&fattr->cf_mode,
|
|
&other_mask);
|
|
|
|
|
|
/* memcpy((void *)(&(cifscred->aces[i])),
|
|
(void *)ppace[i],
|
|
sizeof(struct cifs_ace)); */
|
|
|
|
acl_base = (char *)ppace[i];
|
|
acl_size = le16_to_cpu(ppace[i]->size);
|
|
}
|
|
|
|
kfree(ppace);
|
|
}
|
|
|
|
return;
|
|
}
|
|
|
|
|
|
static int set_chmod_dacl(struct cifs_acl *pndacl, struct cifs_sid *pownersid,
|
|
struct cifs_sid *pgrpsid, __u64 nmode)
|
|
{
|
|
u16 size = 0;
|
|
struct cifs_acl *pnndacl;
|
|
|
|
pnndacl = (struct cifs_acl *)((char *)pndacl + sizeof(struct cifs_acl));
|
|
|
|
size += fill_ace_for_sid((struct cifs_ace *) ((char *)pnndacl + size),
|
|
pownersid, nmode, S_IRWXU);
|
|
size += fill_ace_for_sid((struct cifs_ace *)((char *)pnndacl + size),
|
|
pgrpsid, nmode, S_IRWXG);
|
|
size += fill_ace_for_sid((struct cifs_ace *)((char *)pnndacl + size),
|
|
&sid_everyone, nmode, S_IRWXO);
|
|
|
|
pndacl->size = cpu_to_le16(size + sizeof(struct cifs_acl));
|
|
pndacl->num_aces = cpu_to_le32(3);
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
static int parse_sid(struct cifs_sid *psid, char *end_of_acl)
|
|
{
|
|
/* BB need to add parm so we can store the SID BB */
|
|
|
|
/* validate that we do not go past end of ACL - sid must be at least 8
|
|
bytes long (assuming no sub-auths - e.g. the null SID */
|
|
if (end_of_acl < (char *)psid + 8) {
|
|
cERROR(1, "ACL too small to parse SID %p", psid);
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (psid->num_subauth) {
|
|
#ifdef CONFIG_CIFS_DEBUG2
|
|
int i;
|
|
cFYI(1, "SID revision %d num_auth %d",
|
|
psid->revision, psid->num_subauth);
|
|
|
|
for (i = 0; i < psid->num_subauth; i++) {
|
|
cFYI(1, "SID sub_auth[%d]: 0x%x ", i,
|
|
le32_to_cpu(psid->sub_auth[i]));
|
|
}
|
|
|
|
/* BB add length check to make sure that we do not have huge
|
|
num auths and therefore go off the end */
|
|
cFYI(1, "RID 0x%x",
|
|
le32_to_cpu(psid->sub_auth[psid->num_subauth-1]));
|
|
#endif
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
/* Convert CIFS ACL to POSIX form */
|
|
static int parse_sec_desc(struct cifs_sb_info *cifs_sb,
|
|
struct cifs_ntsd *pntsd, int acl_len, struct cifs_fattr *fattr)
|
|
{
|
|
int rc = 0;
|
|
struct cifs_sid *owner_sid_ptr, *group_sid_ptr;
|
|
struct cifs_acl *dacl_ptr; /* no need for SACL ptr */
|
|
char *end_of_acl = ((char *)pntsd) + acl_len;
|
|
__u32 dacloffset;
|
|
|
|
if (pntsd == NULL)
|
|
return -EIO;
|
|
|
|
owner_sid_ptr = (struct cifs_sid *)((char *)pntsd +
|
|
le32_to_cpu(pntsd->osidoffset));
|
|
group_sid_ptr = (struct cifs_sid *)((char *)pntsd +
|
|
le32_to_cpu(pntsd->gsidoffset));
|
|
dacloffset = le32_to_cpu(pntsd->dacloffset);
|
|
dacl_ptr = (struct cifs_acl *)((char *)pntsd + dacloffset);
|
|
cFYI(DBG2, "revision %d type 0x%x ooffset 0x%x goffset 0x%x "
|
|
"sacloffset 0x%x dacloffset 0x%x",
|
|
pntsd->revision, pntsd->type, le32_to_cpu(pntsd->osidoffset),
|
|
le32_to_cpu(pntsd->gsidoffset),
|
|
le32_to_cpu(pntsd->sacloffset), dacloffset);
|
|
/* cifs_dump_mem("owner_sid: ", owner_sid_ptr, 64); */
|
|
rc = parse_sid(owner_sid_ptr, end_of_acl);
|
|
if (rc) {
|
|
cFYI(1, "%s: Error %d parsing Owner SID", __func__, rc);
|
|
return rc;
|
|
}
|
|
rc = sid_to_id(cifs_sb, owner_sid_ptr, fattr, SIDOWNER);
|
|
if (rc) {
|
|
cFYI(1, "%s: Error %d mapping Owner SID to uid", __func__, rc);
|
|
return rc;
|
|
}
|
|
|
|
rc = parse_sid(group_sid_ptr, end_of_acl);
|
|
if (rc) {
|
|
cFYI(1, "%s: Error %d mapping Owner SID to gid", __func__, rc);
|
|
return rc;
|
|
}
|
|
rc = sid_to_id(cifs_sb, group_sid_ptr, fattr, SIDGROUP);
|
|
if (rc) {
|
|
cFYI(1, "%s: Error %d mapping Group SID to gid", __func__, rc);
|
|
return rc;
|
|
}
|
|
|
|
if (dacloffset)
|
|
parse_dacl(dacl_ptr, end_of_acl, owner_sid_ptr,
|
|
group_sid_ptr, fattr);
|
|
else
|
|
cFYI(1, "no ACL"); /* BB grant all or default perms? */
|
|
|
|
return rc;
|
|
}
|
|
|
|
/* Convert permission bits from mode to equivalent CIFS ACL */
|
|
static int build_sec_desc(struct cifs_ntsd *pntsd, struct cifs_ntsd *pnntsd,
|
|
__u32 secdesclen, __u64 nmode, uid_t uid, gid_t gid, int *aclflag)
|
|
{
|
|
int rc = 0;
|
|
__u32 dacloffset;
|
|
__u32 ndacloffset;
|
|
__u32 sidsoffset;
|
|
struct cifs_sid *owner_sid_ptr, *group_sid_ptr;
|
|
struct cifs_sid *nowner_sid_ptr, *ngroup_sid_ptr;
|
|
struct cifs_acl *dacl_ptr = NULL; /* no need for SACL ptr */
|
|
struct cifs_acl *ndacl_ptr = NULL; /* no need for SACL ptr */
|
|
|
|
if (nmode != NO_CHANGE_64) { /* chmod */
|
|
owner_sid_ptr = (struct cifs_sid *)((char *)pntsd +
|
|
le32_to_cpu(pntsd->osidoffset));
|
|
group_sid_ptr = (struct cifs_sid *)((char *)pntsd +
|
|
le32_to_cpu(pntsd->gsidoffset));
|
|
dacloffset = le32_to_cpu(pntsd->dacloffset);
|
|
dacl_ptr = (struct cifs_acl *)((char *)pntsd + dacloffset);
|
|
ndacloffset = sizeof(struct cifs_ntsd);
|
|
ndacl_ptr = (struct cifs_acl *)((char *)pnntsd + ndacloffset);
|
|
ndacl_ptr->revision = dacl_ptr->revision;
|
|
ndacl_ptr->size = 0;
|
|
ndacl_ptr->num_aces = 0;
|
|
|
|
rc = set_chmod_dacl(ndacl_ptr, owner_sid_ptr, group_sid_ptr,
|
|
nmode);
|
|
sidsoffset = ndacloffset + le16_to_cpu(ndacl_ptr->size);
|
|
/* copy sec desc control portion & owner and group sids */
|
|
copy_sec_desc(pntsd, pnntsd, sidsoffset);
|
|
*aclflag = CIFS_ACL_DACL;
|
|
} else {
|
|
memcpy(pnntsd, pntsd, secdesclen);
|
|
if (uid != NO_CHANGE_32) { /* chown */
|
|
owner_sid_ptr = (struct cifs_sid *)((char *)pnntsd +
|
|
le32_to_cpu(pnntsd->osidoffset));
|
|
nowner_sid_ptr = kmalloc(sizeof(struct cifs_sid),
|
|
GFP_KERNEL);
|
|
if (!nowner_sid_ptr)
|
|
return -ENOMEM;
|
|
rc = id_to_sid(uid, SIDOWNER, nowner_sid_ptr);
|
|
if (rc) {
|
|
cFYI(1, "%s: Mapping error %d for owner id %d",
|
|
__func__, rc, uid);
|
|
kfree(nowner_sid_ptr);
|
|
return rc;
|
|
}
|
|
memcpy(owner_sid_ptr, nowner_sid_ptr,
|
|
sizeof(struct cifs_sid));
|
|
kfree(nowner_sid_ptr);
|
|
*aclflag = CIFS_ACL_OWNER;
|
|
}
|
|
if (gid != NO_CHANGE_32) { /* chgrp */
|
|
group_sid_ptr = (struct cifs_sid *)((char *)pnntsd +
|
|
le32_to_cpu(pnntsd->gsidoffset));
|
|
ngroup_sid_ptr = kmalloc(sizeof(struct cifs_sid),
|
|
GFP_KERNEL);
|
|
if (!ngroup_sid_ptr)
|
|
return -ENOMEM;
|
|
rc = id_to_sid(gid, SIDGROUP, ngroup_sid_ptr);
|
|
if (rc) {
|
|
cFYI(1, "%s: Mapping error %d for group id %d",
|
|
__func__, rc, gid);
|
|
kfree(ngroup_sid_ptr);
|
|
return rc;
|
|
}
|
|
memcpy(group_sid_ptr, ngroup_sid_ptr,
|
|
sizeof(struct cifs_sid));
|
|
kfree(ngroup_sid_ptr);
|
|
*aclflag = CIFS_ACL_GROUP;
|
|
}
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static struct cifs_ntsd *get_cifs_acl_by_fid(struct cifs_sb_info *cifs_sb,
|
|
__u16 fid, u32 *pacllen)
|
|
{
|
|
struct cifs_ntsd *pntsd = NULL;
|
|
unsigned int xid;
|
|
int rc;
|
|
struct tcon_link *tlink = cifs_sb_tlink(cifs_sb);
|
|
|
|
if (IS_ERR(tlink))
|
|
return ERR_CAST(tlink);
|
|
|
|
xid = get_xid();
|
|
rc = CIFSSMBGetCIFSACL(xid, tlink_tcon(tlink), fid, &pntsd, pacllen);
|
|
free_xid(xid);
|
|
|
|
cifs_put_tlink(tlink);
|
|
|
|
cFYI(1, "%s: rc = %d ACL len %d", __func__, rc, *pacllen);
|
|
if (rc)
|
|
return ERR_PTR(rc);
|
|
return pntsd;
|
|
}
|
|
|
|
static struct cifs_ntsd *get_cifs_acl_by_path(struct cifs_sb_info *cifs_sb,
|
|
const char *path, u32 *pacllen)
|
|
{
|
|
struct cifs_ntsd *pntsd = NULL;
|
|
int oplock = 0;
|
|
unsigned int xid;
|
|
int rc, create_options = 0;
|
|
__u16 fid;
|
|
struct cifs_tcon *tcon;
|
|
struct tcon_link *tlink = cifs_sb_tlink(cifs_sb);
|
|
|
|
if (IS_ERR(tlink))
|
|
return ERR_CAST(tlink);
|
|
|
|
tcon = tlink_tcon(tlink);
|
|
xid = get_xid();
|
|
|
|
if (backup_cred(cifs_sb))
|
|
create_options |= CREATE_OPEN_BACKUP_INTENT;
|
|
|
|
rc = CIFSSMBOpen(xid, tcon, path, FILE_OPEN, READ_CONTROL,
|
|
create_options, &fid, &oplock, NULL, cifs_sb->local_nls,
|
|
cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MAP_SPECIAL_CHR);
|
|
if (!rc) {
|
|
rc = CIFSSMBGetCIFSACL(xid, tcon, fid, &pntsd, pacllen);
|
|
CIFSSMBClose(xid, tcon, fid);
|
|
}
|
|
|
|
cifs_put_tlink(tlink);
|
|
free_xid(xid);
|
|
|
|
cFYI(1, "%s: rc = %d ACL len %d", __func__, rc, *pacllen);
|
|
if (rc)
|
|
return ERR_PTR(rc);
|
|
return pntsd;
|
|
}
|
|
|
|
/* Retrieve an ACL from the server */
|
|
struct cifs_ntsd *get_cifs_acl(struct cifs_sb_info *cifs_sb,
|
|
struct inode *inode, const char *path,
|
|
u32 *pacllen)
|
|
{
|
|
struct cifs_ntsd *pntsd = NULL;
|
|
struct cifsFileInfo *open_file = NULL;
|
|
|
|
if (inode)
|
|
open_file = find_readable_file(CIFS_I(inode), true);
|
|
if (!open_file)
|
|
return get_cifs_acl_by_path(cifs_sb, path, pacllen);
|
|
|
|
pntsd = get_cifs_acl_by_fid(cifs_sb, open_file->fid.netfid, pacllen);
|
|
cifsFileInfo_put(open_file);
|
|
return pntsd;
|
|
}
|
|
|
|
/* Set an ACL on the server */
|
|
int set_cifs_acl(struct cifs_ntsd *pnntsd, __u32 acllen,
|
|
struct inode *inode, const char *path, int aclflag)
|
|
{
|
|
int oplock = 0;
|
|
unsigned int xid;
|
|
int rc, access_flags, create_options = 0;
|
|
__u16 fid;
|
|
struct cifs_tcon *tcon;
|
|
struct cifs_sb_info *cifs_sb = CIFS_SB(inode->i_sb);
|
|
struct tcon_link *tlink = cifs_sb_tlink(cifs_sb);
|
|
|
|
if (IS_ERR(tlink))
|
|
return PTR_ERR(tlink);
|
|
|
|
tcon = tlink_tcon(tlink);
|
|
xid = get_xid();
|
|
|
|
if (backup_cred(cifs_sb))
|
|
create_options |= CREATE_OPEN_BACKUP_INTENT;
|
|
|
|
if (aclflag == CIFS_ACL_OWNER || aclflag == CIFS_ACL_GROUP)
|
|
access_flags = WRITE_OWNER;
|
|
else
|
|
access_flags = WRITE_DAC;
|
|
|
|
rc = CIFSSMBOpen(xid, tcon, path, FILE_OPEN, access_flags,
|
|
create_options, &fid, &oplock, NULL, cifs_sb->local_nls,
|
|
cifs_sb->mnt_cifs_flags & CIFS_MOUNT_MAP_SPECIAL_CHR);
|
|
if (rc) {
|
|
cERROR(1, "Unable to open file to set ACL");
|
|
goto out;
|
|
}
|
|
|
|
rc = CIFSSMBSetCIFSACL(xid, tcon, fid, pnntsd, acllen, aclflag);
|
|
cFYI(DBG2, "SetCIFSACL rc = %d", rc);
|
|
|
|
CIFSSMBClose(xid, tcon, fid);
|
|
out:
|
|
free_xid(xid);
|
|
cifs_put_tlink(tlink);
|
|
return rc;
|
|
}
|
|
|
|
/* Translate the CIFS ACL (simlar to NTFS ACL) for a file into mode bits */
|
|
int
|
|
cifs_acl_to_fattr(struct cifs_sb_info *cifs_sb, struct cifs_fattr *fattr,
|
|
struct inode *inode, const char *path, const __u16 *pfid)
|
|
{
|
|
struct cifs_ntsd *pntsd = NULL;
|
|
u32 acllen = 0;
|
|
int rc = 0;
|
|
|
|
cFYI(DBG2, "converting ACL to mode for %s", path);
|
|
|
|
if (pfid)
|
|
pntsd = get_cifs_acl_by_fid(cifs_sb, *pfid, &acllen);
|
|
else
|
|
pntsd = get_cifs_acl(cifs_sb, inode, path, &acllen);
|
|
|
|
/* if we can retrieve the ACL, now parse Access Control Entries, ACEs */
|
|
if (IS_ERR(pntsd)) {
|
|
rc = PTR_ERR(pntsd);
|
|
cERROR(1, "%s: error %d getting sec desc", __func__, rc);
|
|
} else {
|
|
rc = parse_sec_desc(cifs_sb, pntsd, acllen, fattr);
|
|
kfree(pntsd);
|
|
if (rc)
|
|
cERROR(1, "parse sec desc failed rc = %d", rc);
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
/* Convert mode bits to an ACL so we can update the ACL on the server */
|
|
int
|
|
id_mode_to_cifs_acl(struct inode *inode, const char *path, __u64 nmode,
|
|
uid_t uid, gid_t gid)
|
|
{
|
|
int rc = 0;
|
|
int aclflag = CIFS_ACL_DACL; /* default flag to set */
|
|
__u32 secdesclen = 0;
|
|
struct cifs_ntsd *pntsd = NULL; /* acl obtained from server */
|
|
struct cifs_ntsd *pnntsd = NULL; /* modified acl to be sent to server */
|
|
|
|
cFYI(DBG2, "set ACL from mode for %s", path);
|
|
|
|
/* Get the security descriptor */
|
|
pntsd = get_cifs_acl(CIFS_SB(inode->i_sb), inode, path, &secdesclen);
|
|
|
|
/* Add three ACEs for owner, group, everyone getting rid of
|
|
other ACEs as chmod disables ACEs and set the security descriptor */
|
|
|
|
if (IS_ERR(pntsd)) {
|
|
rc = PTR_ERR(pntsd);
|
|
cERROR(1, "%s: error %d getting sec desc", __func__, rc);
|
|
} else {
|
|
/* allocate memory for the smb header,
|
|
set security descriptor request security descriptor
|
|
parameters, and secuirty descriptor itself */
|
|
|
|
secdesclen = secdesclen < DEFSECDESCLEN ?
|
|
DEFSECDESCLEN : secdesclen;
|
|
pnntsd = kmalloc(secdesclen, GFP_KERNEL);
|
|
if (!pnntsd) {
|
|
cERROR(1, "Unable to allocate security descriptor");
|
|
kfree(pntsd);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
rc = build_sec_desc(pntsd, pnntsd, secdesclen, nmode, uid, gid,
|
|
&aclflag);
|
|
|
|
cFYI(DBG2, "build_sec_desc rc: %d", rc);
|
|
|
|
if (!rc) {
|
|
/* Set the security descriptor */
|
|
rc = set_cifs_acl(pnntsd, secdesclen, inode,
|
|
path, aclflag);
|
|
cFYI(DBG2, "set_cifs_acl rc: %d", rc);
|
|
}
|
|
|
|
kfree(pnntsd);
|
|
kfree(pntsd);
|
|
}
|
|
|
|
return rc;
|
|
}
|