ae6d35ed0a
Allow the xindex to have 2^24 entries. Signed-off-by: John Johansen <john.johansen@canonical.com>
1540 lines
38 KiB
C
1540 lines
38 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* AppArmor security module
|
|
*
|
|
* This file contains AppArmor functions for unpacking policy loaded from
|
|
* userspace.
|
|
*
|
|
* Copyright (C) 1998-2008 Novell/SUSE
|
|
* Copyright 2009-2010 Canonical Ltd.
|
|
*
|
|
* AppArmor uses a serialized binary format for loading policy. To find
|
|
* policy format documentation see Documentation/admin-guide/LSM/apparmor.rst
|
|
* All policy is validated before it is used.
|
|
*/
|
|
|
|
#include <asm/unaligned.h>
|
|
#include <linux/ctype.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/zstd.h>
|
|
|
|
#include "include/apparmor.h"
|
|
#include "include/audit.h"
|
|
#include "include/cred.h"
|
|
#include "include/crypto.h"
|
|
#include "include/file.h"
|
|
#include "include/match.h"
|
|
#include "include/path.h"
|
|
#include "include/policy.h"
|
|
#include "include/policy_unpack.h"
|
|
|
|
#define K_ABI_MASK 0x3ff
|
|
#define FORCE_COMPLAIN_FLAG 0x800
|
|
#define VERSION_LT(X, Y) (((X) & K_ABI_MASK) < ((Y) & K_ABI_MASK))
|
|
#define VERSION_LE(X, Y) (((X) & K_ABI_MASK) <= ((Y) & K_ABI_MASK))
|
|
#define VERSION_GT(X, Y) (((X) & K_ABI_MASK) > ((Y) & K_ABI_MASK))
|
|
|
|
#define v5 5 /* base version */
|
|
#define v6 6 /* per entry policydb mediation check */
|
|
#define v7 7
|
|
#define v8 8 /* full network masking */
|
|
#define v9 9 /* xbits are used as permission bits in policydb */
|
|
|
|
/*
|
|
* The AppArmor interface treats data as a type byte followed by the
|
|
* actual data. The interface has the notion of a named entry
|
|
* which has a name (AA_NAME typecode followed by name string) followed by
|
|
* the entries typecode and data. Named types allow for optional
|
|
* elements and extensions to be added and tested for without breaking
|
|
* backwards compatibility.
|
|
*/
|
|
|
|
enum aa_code {
|
|
AA_U8,
|
|
AA_U16,
|
|
AA_U32,
|
|
AA_U64,
|
|
AA_NAME, /* same as string except it is items name */
|
|
AA_STRING,
|
|
AA_BLOB,
|
|
AA_STRUCT,
|
|
AA_STRUCTEND,
|
|
AA_LIST,
|
|
AA_LISTEND,
|
|
AA_ARRAY,
|
|
AA_ARRAYEND,
|
|
};
|
|
|
|
/*
|
|
* aa_ext is the read of the buffer containing the serialized profile. The
|
|
* data is copied into a kernel buffer in apparmorfs and then handed off to
|
|
* the unpack routines.
|
|
*/
|
|
struct aa_ext {
|
|
void *start;
|
|
void *end;
|
|
void *pos; /* pointer to current position in the buffer */
|
|
u32 version;
|
|
};
|
|
|
|
/* audit callback for unpack fields */
|
|
static void audit_cb(struct audit_buffer *ab, void *va)
|
|
{
|
|
struct common_audit_data *sa = va;
|
|
|
|
if (aad(sa)->iface.ns) {
|
|
audit_log_format(ab, " ns=");
|
|
audit_log_untrustedstring(ab, aad(sa)->iface.ns);
|
|
}
|
|
if (aad(sa)->name) {
|
|
audit_log_format(ab, " name=");
|
|
audit_log_untrustedstring(ab, aad(sa)->name);
|
|
}
|
|
if (aad(sa)->iface.pos)
|
|
audit_log_format(ab, " offset=%ld", aad(sa)->iface.pos);
|
|
}
|
|
|
|
/**
|
|
* audit_iface - do audit message for policy unpacking/load/replace/remove
|
|
* @new: profile if it has been allocated (MAYBE NULL)
|
|
* @ns_name: name of the ns the profile is to be loaded to (MAY BE NULL)
|
|
* @name: name of the profile being manipulated (MAYBE NULL)
|
|
* @info: any extra info about the failure (MAYBE NULL)
|
|
* @e: buffer position info
|
|
* @error: error code
|
|
*
|
|
* Returns: %0 or error
|
|
*/
|
|
static int audit_iface(struct aa_profile *new, const char *ns_name,
|
|
const char *name, const char *info, struct aa_ext *e,
|
|
int error)
|
|
{
|
|
struct aa_profile *profile = labels_profile(aa_current_raw_label());
|
|
DEFINE_AUDIT_DATA(sa, LSM_AUDIT_DATA_NONE, NULL);
|
|
if (e)
|
|
aad(&sa)->iface.pos = e->pos - e->start;
|
|
aad(&sa)->iface.ns = ns_name;
|
|
if (new)
|
|
aad(&sa)->name = new->base.hname;
|
|
else
|
|
aad(&sa)->name = name;
|
|
aad(&sa)->info = info;
|
|
aad(&sa)->error = error;
|
|
|
|
return aa_audit(AUDIT_APPARMOR_STATUS, profile, &sa, audit_cb);
|
|
}
|
|
|
|
void __aa_loaddata_update(struct aa_loaddata *data, long revision)
|
|
{
|
|
AA_BUG(!data);
|
|
AA_BUG(!data->ns);
|
|
AA_BUG(!mutex_is_locked(&data->ns->lock));
|
|
AA_BUG(data->revision > revision);
|
|
|
|
data->revision = revision;
|
|
if ((data->dents[AAFS_LOADDATA_REVISION])) {
|
|
d_inode(data->dents[AAFS_LOADDATA_DIR])->i_mtime =
|
|
current_time(d_inode(data->dents[AAFS_LOADDATA_DIR]));
|
|
d_inode(data->dents[AAFS_LOADDATA_REVISION])->i_mtime =
|
|
current_time(d_inode(data->dents[AAFS_LOADDATA_REVISION]));
|
|
}
|
|
}
|
|
|
|
bool aa_rawdata_eq(struct aa_loaddata *l, struct aa_loaddata *r)
|
|
{
|
|
if (l->size != r->size)
|
|
return false;
|
|
if (l->compressed_size != r->compressed_size)
|
|
return false;
|
|
if (aa_g_hash_policy && memcmp(l->hash, r->hash, aa_hash_size()) != 0)
|
|
return false;
|
|
return memcmp(l->data, r->data, r->compressed_size ?: r->size) == 0;
|
|
}
|
|
|
|
/*
|
|
* need to take the ns mutex lock which is NOT safe most places that
|
|
* put_loaddata is called, so we have to delay freeing it
|
|
*/
|
|
static void do_loaddata_free(struct work_struct *work)
|
|
{
|
|
struct aa_loaddata *d = container_of(work, struct aa_loaddata, work);
|
|
struct aa_ns *ns = aa_get_ns(d->ns);
|
|
|
|
if (ns) {
|
|
mutex_lock_nested(&ns->lock, ns->level);
|
|
__aa_fs_remove_rawdata(d);
|
|
mutex_unlock(&ns->lock);
|
|
aa_put_ns(ns);
|
|
}
|
|
|
|
kfree_sensitive(d->hash);
|
|
kfree_sensitive(d->name);
|
|
kvfree(d->data);
|
|
kfree_sensitive(d);
|
|
}
|
|
|
|
void aa_loaddata_kref(struct kref *kref)
|
|
{
|
|
struct aa_loaddata *d = container_of(kref, struct aa_loaddata, count);
|
|
|
|
if (d) {
|
|
INIT_WORK(&d->work, do_loaddata_free);
|
|
schedule_work(&d->work);
|
|
}
|
|
}
|
|
|
|
struct aa_loaddata *aa_loaddata_alloc(size_t size)
|
|
{
|
|
struct aa_loaddata *d;
|
|
|
|
d = kzalloc(sizeof(*d), GFP_KERNEL);
|
|
if (d == NULL)
|
|
return ERR_PTR(-ENOMEM);
|
|
d->data = kvzalloc(size, GFP_KERNEL);
|
|
if (!d->data) {
|
|
kfree(d);
|
|
return ERR_PTR(-ENOMEM);
|
|
}
|
|
kref_init(&d->count);
|
|
INIT_LIST_HEAD(&d->list);
|
|
|
|
return d;
|
|
}
|
|
|
|
/* test if read will be in packed data bounds */
|
|
static bool inbounds(struct aa_ext *e, size_t size)
|
|
{
|
|
return (size <= e->end - e->pos);
|
|
}
|
|
|
|
static void *kvmemdup(const void *src, size_t len)
|
|
{
|
|
void *p = kvmalloc(len, GFP_KERNEL);
|
|
|
|
if (p)
|
|
memcpy(p, src, len);
|
|
return p;
|
|
}
|
|
|
|
/**
|
|
* unpack_u16_chunk - test and do bounds checking for a u16 size based chunk
|
|
* @e: serialized data read head (NOT NULL)
|
|
* @chunk: start address for chunk of data (NOT NULL)
|
|
*
|
|
* Returns: the size of chunk found with the read head at the end of the chunk.
|
|
*/
|
|
static size_t unpack_u16_chunk(struct aa_ext *e, char **chunk)
|
|
{
|
|
size_t size = 0;
|
|
void *pos = e->pos;
|
|
|
|
if (!inbounds(e, sizeof(u16)))
|
|
goto fail;
|
|
size = le16_to_cpu(get_unaligned((__le16 *) e->pos));
|
|
e->pos += sizeof(__le16);
|
|
if (!inbounds(e, size))
|
|
goto fail;
|
|
*chunk = e->pos;
|
|
e->pos += size;
|
|
return size;
|
|
|
|
fail:
|
|
e->pos = pos;
|
|
return 0;
|
|
}
|
|
|
|
/* unpack control byte */
|
|
static bool unpack_X(struct aa_ext *e, enum aa_code code)
|
|
{
|
|
if (!inbounds(e, 1))
|
|
return false;
|
|
if (*(u8 *) e->pos != code)
|
|
return false;
|
|
e->pos++;
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* unpack_nameX - check is the next element is of type X with a name of @name
|
|
* @e: serialized data extent information (NOT NULL)
|
|
* @code: type code
|
|
* @name: name to match to the serialized element. (MAYBE NULL)
|
|
*
|
|
* check that the next serialized data element is of type X and has a tag
|
|
* name @name. If @name is specified then there must be a matching
|
|
* name element in the stream. If @name is NULL any name element will be
|
|
* skipped and only the typecode will be tested.
|
|
*
|
|
* Returns true on success (both type code and name tests match) and the read
|
|
* head is advanced past the headers
|
|
*
|
|
* Returns: false if either match fails, the read head does not move
|
|
*/
|
|
static bool unpack_nameX(struct aa_ext *e, enum aa_code code, const char *name)
|
|
{
|
|
/*
|
|
* May need to reset pos if name or type doesn't match
|
|
*/
|
|
void *pos = e->pos;
|
|
/*
|
|
* Check for presence of a tagname, and if present name size
|
|
* AA_NAME tag value is a u16.
|
|
*/
|
|
if (unpack_X(e, AA_NAME)) {
|
|
char *tag = NULL;
|
|
size_t size = unpack_u16_chunk(e, &tag);
|
|
/* if a name is specified it must match. otherwise skip tag */
|
|
if (name && (!size || tag[size-1] != '\0' || strcmp(name, tag)))
|
|
goto fail;
|
|
} else if (name) {
|
|
/* if a name is specified and there is no name tag fail */
|
|
goto fail;
|
|
}
|
|
|
|
/* now check if type code matches */
|
|
if (unpack_X(e, code))
|
|
return true;
|
|
|
|
fail:
|
|
e->pos = pos;
|
|
return false;
|
|
}
|
|
|
|
static bool unpack_u8(struct aa_ext *e, u8 *data, const char *name)
|
|
{
|
|
void *pos = e->pos;
|
|
|
|
if (unpack_nameX(e, AA_U8, name)) {
|
|
if (!inbounds(e, sizeof(u8)))
|
|
goto fail;
|
|
if (data)
|
|
*data = *((u8 *)e->pos);
|
|
e->pos += sizeof(u8);
|
|
return true;
|
|
}
|
|
|
|
fail:
|
|
e->pos = pos;
|
|
return false;
|
|
}
|
|
|
|
static bool unpack_u32(struct aa_ext *e, u32 *data, const char *name)
|
|
{
|
|
void *pos = e->pos;
|
|
|
|
if (unpack_nameX(e, AA_U32, name)) {
|
|
if (!inbounds(e, sizeof(u32)))
|
|
goto fail;
|
|
if (data)
|
|
*data = le32_to_cpu(get_unaligned((__le32 *) e->pos));
|
|
e->pos += sizeof(u32);
|
|
return true;
|
|
}
|
|
|
|
fail:
|
|
e->pos = pos;
|
|
return false;
|
|
}
|
|
|
|
static bool unpack_u64(struct aa_ext *e, u64 *data, const char *name)
|
|
{
|
|
void *pos = e->pos;
|
|
|
|
if (unpack_nameX(e, AA_U64, name)) {
|
|
if (!inbounds(e, sizeof(u64)))
|
|
goto fail;
|
|
if (data)
|
|
*data = le64_to_cpu(get_unaligned((__le64 *) e->pos));
|
|
e->pos += sizeof(u64);
|
|
return true;
|
|
}
|
|
|
|
fail:
|
|
e->pos = pos;
|
|
return false;
|
|
}
|
|
|
|
static size_t unpack_array(struct aa_ext *e, const char *name)
|
|
{
|
|
void *pos = e->pos;
|
|
|
|
if (unpack_nameX(e, AA_ARRAY, name)) {
|
|
int size;
|
|
if (!inbounds(e, sizeof(u16)))
|
|
goto fail;
|
|
size = (int)le16_to_cpu(get_unaligned((__le16 *) e->pos));
|
|
e->pos += sizeof(u16);
|
|
return size;
|
|
}
|
|
|
|
fail:
|
|
e->pos = pos;
|
|
return 0;
|
|
}
|
|
|
|
static size_t unpack_blob(struct aa_ext *e, char **blob, const char *name)
|
|
{
|
|
void *pos = e->pos;
|
|
|
|
if (unpack_nameX(e, AA_BLOB, name)) {
|
|
u32 size;
|
|
if (!inbounds(e, sizeof(u32)))
|
|
goto fail;
|
|
size = le32_to_cpu(get_unaligned((__le32 *) e->pos));
|
|
e->pos += sizeof(u32);
|
|
if (inbounds(e, (size_t) size)) {
|
|
*blob = e->pos;
|
|
e->pos += size;
|
|
return size;
|
|
}
|
|
}
|
|
|
|
fail:
|
|
e->pos = pos;
|
|
return 0;
|
|
}
|
|
|
|
static int unpack_str(struct aa_ext *e, const char **string, const char *name)
|
|
{
|
|
char *src_str;
|
|
size_t size = 0;
|
|
void *pos = e->pos;
|
|
*string = NULL;
|
|
if (unpack_nameX(e, AA_STRING, name)) {
|
|
size = unpack_u16_chunk(e, &src_str);
|
|
if (size) {
|
|
/* strings are null terminated, length is size - 1 */
|
|
if (src_str[size - 1] != 0)
|
|
goto fail;
|
|
*string = src_str;
|
|
|
|
return size;
|
|
}
|
|
}
|
|
|
|
fail:
|
|
e->pos = pos;
|
|
return 0;
|
|
}
|
|
|
|
static int unpack_strdup(struct aa_ext *e, char **string, const char *name)
|
|
{
|
|
const char *tmp;
|
|
void *pos = e->pos;
|
|
int res = unpack_str(e, &tmp, name);
|
|
*string = NULL;
|
|
|
|
if (!res)
|
|
return 0;
|
|
|
|
*string = kmemdup(tmp, res, GFP_KERNEL);
|
|
if (!*string) {
|
|
e->pos = pos;
|
|
return 0;
|
|
}
|
|
|
|
return res;
|
|
}
|
|
|
|
|
|
/**
|
|
* unpack_dfa - unpack a file rule dfa
|
|
* @e: serialized data extent information (NOT NULL)
|
|
*
|
|
* returns dfa or ERR_PTR or NULL if no dfa
|
|
*/
|
|
static struct aa_dfa *unpack_dfa(struct aa_ext *e)
|
|
{
|
|
char *blob = NULL;
|
|
size_t size;
|
|
struct aa_dfa *dfa = NULL;
|
|
|
|
size = unpack_blob(e, &blob, "aadfa");
|
|
if (size) {
|
|
/*
|
|
* The dfa is aligned with in the blob to 8 bytes
|
|
* from the beginning of the stream.
|
|
* alignment adjust needed by dfa unpack
|
|
*/
|
|
size_t sz = blob - (char *) e->start -
|
|
((e->pos - e->start) & 7);
|
|
size_t pad = ALIGN(sz, 8) - sz;
|
|
int flags = TO_ACCEPT1_FLAG(YYTD_DATA32) |
|
|
TO_ACCEPT2_FLAG(YYTD_DATA32);
|
|
if (aa_g_paranoid_load)
|
|
flags |= DFA_FLAG_VERIFY_STATES;
|
|
dfa = aa_dfa_unpack(blob + pad, size - pad, flags);
|
|
|
|
if (IS_ERR(dfa))
|
|
return dfa;
|
|
|
|
}
|
|
|
|
return dfa;
|
|
}
|
|
|
|
/**
|
|
* unpack_trans_table - unpack a profile transition table
|
|
* @e: serialized data extent information (NOT NULL)
|
|
* @profile: profile to add the accept table to (NOT NULL)
|
|
*
|
|
* Returns: true if table successfully unpacked
|
|
*/
|
|
static bool unpack_trans_table(struct aa_ext *e, struct aa_profile *profile)
|
|
{
|
|
void *saved_pos = e->pos;
|
|
|
|
/* exec table is optional */
|
|
if (unpack_nameX(e, AA_STRUCT, "xtable")) {
|
|
int i, size;
|
|
|
|
size = unpack_array(e, NULL);
|
|
/* currently 2^24 bits entries 0-3 */
|
|
if (size > (1 << 24))
|
|
goto fail;
|
|
profile->file.trans.table = kcalloc(size, sizeof(char *),
|
|
GFP_KERNEL);
|
|
if (!profile->file.trans.table)
|
|
goto fail;
|
|
|
|
profile->file.trans.size = size;
|
|
for (i = 0; i < size; i++) {
|
|
char *str;
|
|
int c, j, pos, size2 = unpack_strdup(e, &str, NULL);
|
|
/* unpack_strdup verifies that the last character is
|
|
* null termination byte.
|
|
*/
|
|
if (!size2)
|
|
goto fail;
|
|
profile->file.trans.table[i] = str;
|
|
/* verify that name doesn't start with space */
|
|
if (isspace(*str))
|
|
goto fail;
|
|
|
|
/* count internal # of internal \0 */
|
|
for (c = j = 0; j < size2 - 1; j++) {
|
|
if (!str[j]) {
|
|
pos = j;
|
|
c++;
|
|
}
|
|
}
|
|
if (*str == ':') {
|
|
/* first character after : must be valid */
|
|
if (!str[1])
|
|
goto fail;
|
|
/* beginning with : requires an embedded \0,
|
|
* verify that exactly 1 internal \0 exists
|
|
* trailing \0 already verified by unpack_strdup
|
|
*
|
|
* convert \0 back to : for label_parse
|
|
*/
|
|
if (c == 1)
|
|
str[pos] = ':';
|
|
else if (c > 1)
|
|
goto fail;
|
|
} else if (c)
|
|
/* fail - all other cases with embedded \0 */
|
|
goto fail;
|
|
}
|
|
if (!unpack_nameX(e, AA_ARRAYEND, NULL))
|
|
goto fail;
|
|
if (!unpack_nameX(e, AA_STRUCTEND, NULL))
|
|
goto fail;
|
|
}
|
|
return true;
|
|
|
|
fail:
|
|
aa_free_domain_entries(&profile->file.trans);
|
|
e->pos = saved_pos;
|
|
return false;
|
|
}
|
|
|
|
static bool unpack_xattrs(struct aa_ext *e, struct aa_profile *profile)
|
|
{
|
|
void *pos = e->pos;
|
|
|
|
if (unpack_nameX(e, AA_STRUCT, "xattrs")) {
|
|
int i, size;
|
|
|
|
size = unpack_array(e, NULL);
|
|
profile->xattr_count = size;
|
|
profile->xattrs = kcalloc(size, sizeof(char *), GFP_KERNEL);
|
|
if (!profile->xattrs)
|
|
goto fail;
|
|
for (i = 0; i < size; i++) {
|
|
if (!unpack_strdup(e, &profile->xattrs[i], NULL))
|
|
goto fail;
|
|
}
|
|
if (!unpack_nameX(e, AA_ARRAYEND, NULL))
|
|
goto fail;
|
|
if (!unpack_nameX(e, AA_STRUCTEND, NULL))
|
|
goto fail;
|
|
}
|
|
|
|
return true;
|
|
|
|
fail:
|
|
e->pos = pos;
|
|
return false;
|
|
}
|
|
|
|
static bool unpack_secmark(struct aa_ext *e, struct aa_profile *profile)
|
|
{
|
|
void *pos = e->pos;
|
|
int i, size;
|
|
|
|
if (unpack_nameX(e, AA_STRUCT, "secmark")) {
|
|
size = unpack_array(e, NULL);
|
|
|
|
profile->secmark = kcalloc(size, sizeof(struct aa_secmark),
|
|
GFP_KERNEL);
|
|
if (!profile->secmark)
|
|
goto fail;
|
|
|
|
profile->secmark_count = size;
|
|
|
|
for (i = 0; i < size; i++) {
|
|
if (!unpack_u8(e, &profile->secmark[i].audit, NULL))
|
|
goto fail;
|
|
if (!unpack_u8(e, &profile->secmark[i].deny, NULL))
|
|
goto fail;
|
|
if (!unpack_strdup(e, &profile->secmark[i].label, NULL))
|
|
goto fail;
|
|
}
|
|
if (!unpack_nameX(e, AA_ARRAYEND, NULL))
|
|
goto fail;
|
|
if (!unpack_nameX(e, AA_STRUCTEND, NULL))
|
|
goto fail;
|
|
}
|
|
|
|
return true;
|
|
|
|
fail:
|
|
if (profile->secmark) {
|
|
for (i = 0; i < size; i++)
|
|
kfree(profile->secmark[i].label);
|
|
kfree(profile->secmark);
|
|
profile->secmark_count = 0;
|
|
profile->secmark = NULL;
|
|
}
|
|
|
|
e->pos = pos;
|
|
return false;
|
|
}
|
|
|
|
static bool unpack_rlimits(struct aa_ext *e, struct aa_profile *profile)
|
|
{
|
|
void *pos = e->pos;
|
|
|
|
/* rlimits are optional */
|
|
if (unpack_nameX(e, AA_STRUCT, "rlimits")) {
|
|
int i, size;
|
|
u32 tmp = 0;
|
|
if (!unpack_u32(e, &tmp, NULL))
|
|
goto fail;
|
|
profile->rlimits.mask = tmp;
|
|
|
|
size = unpack_array(e, NULL);
|
|
if (size > RLIM_NLIMITS)
|
|
goto fail;
|
|
for (i = 0; i < size; i++) {
|
|
u64 tmp2 = 0;
|
|
int a = aa_map_resource(i);
|
|
if (!unpack_u64(e, &tmp2, NULL))
|
|
goto fail;
|
|
profile->rlimits.limits[a].rlim_max = tmp2;
|
|
}
|
|
if (!unpack_nameX(e, AA_ARRAYEND, NULL))
|
|
goto fail;
|
|
if (!unpack_nameX(e, AA_STRUCTEND, NULL))
|
|
goto fail;
|
|
}
|
|
return true;
|
|
|
|
fail:
|
|
e->pos = pos;
|
|
return false;
|
|
}
|
|
|
|
static u32 strhash(const void *data, u32 len, u32 seed)
|
|
{
|
|
const char * const *key = data;
|
|
|
|
return jhash(*key, strlen(*key), seed);
|
|
}
|
|
|
|
static int datacmp(struct rhashtable_compare_arg *arg, const void *obj)
|
|
{
|
|
const struct aa_data *data = obj;
|
|
const char * const *key = arg->key;
|
|
|
|
return strcmp(data->key, *key);
|
|
}
|
|
|
|
/* remap old accept table embedded permissions to separate permission table */
|
|
static u32 dfa_map_xindex(u16 mask)
|
|
{
|
|
u16 old_index = (mask >> 10) & 0xf;
|
|
u32 index = 0;
|
|
|
|
if (mask & 0x100)
|
|
index |= AA_X_UNSAFE;
|
|
if (mask & 0x200)
|
|
index |= AA_X_INHERIT;
|
|
if (mask & 0x80)
|
|
index |= AA_X_UNCONFINED;
|
|
|
|
if (old_index == 1) {
|
|
index |= AA_X_UNCONFINED;
|
|
} else if (old_index == 2) {
|
|
index |= AA_X_NAME;
|
|
} else if (old_index == 3) {
|
|
index |= AA_X_NAME | AA_X_CHILD;
|
|
} else if (old_index) {
|
|
index |= AA_X_TABLE;
|
|
index |= old_index - 4;
|
|
}
|
|
|
|
return index;
|
|
}
|
|
|
|
/*
|
|
* map old dfa inline permissions to new format
|
|
*/
|
|
#define dfa_user_allow(dfa, state) (((ACCEPT_TABLE(dfa)[state]) & 0x7f) | \
|
|
((ACCEPT_TABLE(dfa)[state]) & 0x80000000))
|
|
#define dfa_user_xbits(dfa, state) (((ACCEPT_TABLE(dfa)[state]) >> 7) & 0x7f)
|
|
#define dfa_user_audit(dfa, state) ((ACCEPT_TABLE2(dfa)[state]) & 0x7f)
|
|
#define dfa_user_quiet(dfa, state) (((ACCEPT_TABLE2(dfa)[state]) >> 7) & 0x7f)
|
|
#define dfa_user_xindex(dfa, state) \
|
|
(dfa_map_xindex(ACCEPT_TABLE(dfa)[state] & 0x3fff))
|
|
|
|
#define dfa_other_allow(dfa, state) ((((ACCEPT_TABLE(dfa)[state]) >> 14) & \
|
|
0x7f) | \
|
|
((ACCEPT_TABLE(dfa)[state]) & 0x80000000))
|
|
#define dfa_other_xbits(dfa, state) \
|
|
((((ACCEPT_TABLE(dfa)[state]) >> 7) >> 14) & 0x7f)
|
|
#define dfa_other_audit(dfa, state) (((ACCEPT_TABLE2(dfa)[state]) >> 14) & 0x7f)
|
|
#define dfa_other_quiet(dfa, state) \
|
|
((((ACCEPT_TABLE2(dfa)[state]) >> 7) >> 14) & 0x7f)
|
|
#define dfa_other_xindex(dfa, state) \
|
|
dfa_map_xindex((ACCEPT_TABLE(dfa)[state] >> 14) & 0x3fff)
|
|
|
|
/**
|
|
* map_old_perms - map old file perms layout to the new layout
|
|
* @old: permission set in old mapping
|
|
*
|
|
* Returns: new permission mapping
|
|
*/
|
|
static u32 map_old_perms(u32 old)
|
|
{
|
|
u32 new = old & 0xf;
|
|
|
|
if (old & MAY_READ)
|
|
new |= AA_MAY_GETATTR | AA_MAY_OPEN;
|
|
if (old & MAY_WRITE)
|
|
new |= AA_MAY_SETATTR | AA_MAY_CREATE | AA_MAY_DELETE |
|
|
AA_MAY_CHMOD | AA_MAY_CHOWN | AA_MAY_OPEN;
|
|
if (old & 0x10)
|
|
new |= AA_MAY_LINK;
|
|
/* the old mapping lock and link_subset flags where overlaid
|
|
* and use was determined by part of a pair that they were in
|
|
*/
|
|
if (old & 0x20)
|
|
new |= AA_MAY_LOCK | AA_LINK_SUBSET;
|
|
if (old & 0x40) /* AA_EXEC_MMAP */
|
|
new |= AA_EXEC_MMAP;
|
|
|
|
return new;
|
|
}
|
|
|
|
static void compute_fperms_allow(struct aa_perms *perms, struct aa_dfa *dfa,
|
|
aa_state_t state)
|
|
{
|
|
perms->allow |= AA_MAY_GETATTR;
|
|
|
|
/* change_profile wasn't determined by ownership in old mapping */
|
|
if (ACCEPT_TABLE(dfa)[state] & 0x80000000)
|
|
perms->allow |= AA_MAY_CHANGE_PROFILE;
|
|
if (ACCEPT_TABLE(dfa)[state] & 0x40000000)
|
|
perms->allow |= AA_MAY_ONEXEC;
|
|
}
|
|
|
|
static struct aa_perms compute_fperms_user(struct aa_dfa *dfa,
|
|
aa_state_t state)
|
|
{
|
|
struct aa_perms perms = { };
|
|
|
|
perms.allow = map_old_perms(dfa_user_allow(dfa, state));
|
|
perms.audit = map_old_perms(dfa_user_audit(dfa, state));
|
|
perms.quiet = map_old_perms(dfa_user_quiet(dfa, state));
|
|
perms.xindex = dfa_user_xindex(dfa, state);
|
|
|
|
compute_fperms_allow(&perms, dfa, state);
|
|
|
|
return perms;
|
|
}
|
|
|
|
static struct aa_perms compute_fperms_other(struct aa_dfa *dfa,
|
|
aa_state_t state)
|
|
{
|
|
struct aa_perms perms = { };
|
|
|
|
perms.allow = map_old_perms(dfa_other_allow(dfa, state));
|
|
perms.audit = map_old_perms(dfa_other_audit(dfa, state));
|
|
perms.quiet = map_old_perms(dfa_other_quiet(dfa, state));
|
|
perms.xindex = dfa_other_xindex(dfa, state);
|
|
|
|
compute_fperms_allow(&perms, dfa, state);
|
|
|
|
return perms;
|
|
}
|
|
|
|
/**
|
|
* aa_compute_fperms - convert dfa compressed perms to internal perms and store
|
|
* them so they can be retrieved later.
|
|
* @dfa: a dfa using fperms to remap to internal permissions
|
|
*
|
|
* Returns: remapped perm table
|
|
*/
|
|
static struct aa_perms *compute_fperms(struct aa_dfa *dfa)
|
|
{
|
|
aa_state_t state;
|
|
unsigned int state_count;
|
|
struct aa_perms *table;
|
|
|
|
AA_BUG(!dfa);
|
|
|
|
state_count = dfa->tables[YYTD_ID_BASE]->td_lolen;
|
|
/* DFAs are restricted from having a state_count of less than 2 */
|
|
table = kvcalloc(state_count * 2, sizeof(struct aa_perms), GFP_KERNEL);
|
|
if (!table)
|
|
return NULL;
|
|
|
|
/* zero init so skip the trap state (state == 0) */
|
|
for (state = 1; state < state_count; state++) {
|
|
table[state * 2] = compute_fperms_user(dfa, state);
|
|
table[state * 2 + 1] = compute_fperms_other(dfa, state);
|
|
}
|
|
|
|
return table;
|
|
}
|
|
|
|
static struct aa_perms *compute_xmatch_perms(struct aa_dfa *xmatch)
|
|
{
|
|
struct aa_perms *perms;
|
|
int state;
|
|
int state_count;
|
|
|
|
AA_BUG(!xmatch);
|
|
|
|
state_count = xmatch->tables[YYTD_ID_BASE]->td_lolen;
|
|
/* DFAs are restricted from having a state_count of less than 2 */
|
|
perms = kvcalloc(state_count, sizeof(struct aa_perms), GFP_KERNEL);
|
|
|
|
/* zero init so skip the trap state (state == 0) */
|
|
for (state = 1; state < state_count; state++)
|
|
perms[state].allow = dfa_user_allow(xmatch, state);
|
|
|
|
return perms;
|
|
}
|
|
|
|
static u32 map_other(u32 x)
|
|
{
|
|
return ((x & 0x3) << 8) | /* SETATTR/GETATTR */
|
|
((x & 0x1c) << 18) | /* ACCEPT/BIND/LISTEN */
|
|
((x & 0x60) << 19); /* SETOPT/GETOPT */
|
|
}
|
|
|
|
static u32 map_xbits(u32 x)
|
|
{
|
|
return ((x & 0x1) << 7) |
|
|
((x & 0x7e) << 9);
|
|
}
|
|
|
|
static struct aa_perms compute_perms_entry(struct aa_dfa *dfa,
|
|
aa_state_t state,
|
|
u32 version)
|
|
{
|
|
struct aa_perms perms = { };
|
|
|
|
perms.allow = dfa_user_allow(dfa, state);
|
|
perms.audit = dfa_user_audit(dfa, state);
|
|
perms.quiet = dfa_user_quiet(dfa, state);
|
|
|
|
/*
|
|
* This mapping is convulated due to history.
|
|
* v1-v4: only file perms, which are handled by compute_fperms
|
|
* v5: added policydb which dropped user conditional to gain new
|
|
* perm bits, but had to map around the xbits because the
|
|
* userspace compiler was still munging them.
|
|
* v9: adds using the xbits in policydb because the compiler now
|
|
* supports treating policydb permission bits different.
|
|
* Unfortunately there is no way to force auditing on the
|
|
* perms represented by the xbits
|
|
*/
|
|
perms.allow |= map_other(dfa_other_allow(dfa, state));
|
|
if (VERSION_LE(version, v8))
|
|
perms.allow |= AA_MAY_LOCK;
|
|
else
|
|
perms.allow |= map_xbits(dfa_user_xbits(dfa, state));
|
|
|
|
/*
|
|
* for v5-v9 perm mapping in the policydb, the other set is used
|
|
* to extend the general perm set
|
|
*/
|
|
perms.audit |= map_other(dfa_other_audit(dfa, state));
|
|
perms.quiet |= map_other(dfa_other_quiet(dfa, state));
|
|
if (VERSION_GT(version, v8))
|
|
perms.quiet |= map_xbits(dfa_other_xbits(dfa, state));
|
|
|
|
return perms;
|
|
}
|
|
|
|
static struct aa_perms *compute_perms(struct aa_dfa *dfa, u32 version)
|
|
{
|
|
unsigned int state;
|
|
unsigned int state_count;
|
|
struct aa_perms *table;
|
|
|
|
AA_BUG(!dfa);
|
|
|
|
state_count = dfa->tables[YYTD_ID_BASE]->td_lolen;
|
|
/* DFAs are restricted from having a state_count of less than 2 */
|
|
table = kvcalloc(state_count, sizeof(struct aa_perms), GFP_KERNEL);
|
|
if (!table)
|
|
return NULL;
|
|
|
|
/* zero init so skip the trap state (state == 0) */
|
|
for (state = 1; state < state_count; state++)
|
|
table[state] = compute_perms_entry(dfa, state, version);
|
|
|
|
return table;
|
|
}
|
|
|
|
/**
|
|
* remap_dfa_accept - remap old dfa accept table to be an index
|
|
* @dfa: dfa to do the remapping on
|
|
* @factor: scaling factor for the index conversion.
|
|
*
|
|
* Used in conjunction with compute_Xperms, it converts old style perms
|
|
* that are encoded in the dfa accept tables to the new style where
|
|
* there is a permission table and the accept table is an index into
|
|
* the permission table.
|
|
*/
|
|
static void remap_dfa_accept(struct aa_dfa *dfa, unsigned int factor)
|
|
{
|
|
unsigned int state;
|
|
unsigned int state_count = dfa->tables[YYTD_ID_BASE]->td_lolen;
|
|
|
|
AA_BUG(!dfa);
|
|
|
|
for (state = 0; state < state_count; state++)
|
|
ACCEPT_TABLE(dfa)[state] = state * factor;
|
|
kvfree(dfa->tables[YYTD_ID_ACCEPT2]);
|
|
dfa->tables[YYTD_ID_ACCEPT2] = NULL;
|
|
}
|
|
|
|
/**
|
|
* unpack_profile - unpack a serialized profile
|
|
* @e: serialized data extent information (NOT NULL)
|
|
* @ns_name: pointer of newly allocated copy of %NULL in case of error
|
|
*
|
|
* NOTE: unpack profile sets audit struct if there is a failure
|
|
*/
|
|
static struct aa_profile *unpack_profile(struct aa_ext *e, char **ns_name)
|
|
{
|
|
struct aa_profile *profile = NULL;
|
|
const char *tmpname, *tmpns = NULL, *name = NULL;
|
|
const char *info = "failed to unpack profile";
|
|
size_t ns_len;
|
|
struct rhashtable_params params = { 0 };
|
|
char *key = NULL;
|
|
struct aa_data *data;
|
|
int i, error = -EPROTO;
|
|
kernel_cap_t tmpcap;
|
|
u32 tmp;
|
|
|
|
*ns_name = NULL;
|
|
|
|
/* check that we have the right struct being passed */
|
|
if (!unpack_nameX(e, AA_STRUCT, "profile"))
|
|
goto fail;
|
|
if (!unpack_str(e, &name, NULL))
|
|
goto fail;
|
|
if (*name == '\0')
|
|
goto fail;
|
|
|
|
tmpname = aa_splitn_fqname(name, strlen(name), &tmpns, &ns_len);
|
|
if (tmpns) {
|
|
*ns_name = kstrndup(tmpns, ns_len, GFP_KERNEL);
|
|
if (!*ns_name) {
|
|
info = "out of memory";
|
|
goto fail;
|
|
}
|
|
name = tmpname;
|
|
}
|
|
|
|
profile = aa_alloc_profile(name, NULL, GFP_KERNEL);
|
|
if (!profile)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
/* profile renaming is optional */
|
|
(void) unpack_str(e, &profile->rename, "rename");
|
|
|
|
/* attachment string is optional */
|
|
(void) unpack_str(e, &profile->attach, "attach");
|
|
|
|
/* xmatch is optional and may be NULL */
|
|
profile->xmatch.dfa = unpack_dfa(e);
|
|
if (IS_ERR(profile->xmatch.dfa)) {
|
|
error = PTR_ERR(profile->xmatch.dfa);
|
|
profile->xmatch.dfa = NULL;
|
|
info = "bad xmatch";
|
|
goto fail;
|
|
}
|
|
/* neither xmatch_len not xmatch_perms are optional if xmatch is set */
|
|
if (profile->xmatch.dfa) {
|
|
if (!unpack_u32(e, &tmp, NULL)) {
|
|
info = "missing xmatch len";
|
|
goto fail;
|
|
}
|
|
profile->xmatch_len = tmp;
|
|
profile->xmatch.start[AA_CLASS_XMATCH] = DFA_START;
|
|
profile->xmatch.perms = compute_xmatch_perms(profile->xmatch.dfa);
|
|
if (!profile->xmatch.perms) {
|
|
info = "failed to convert xmatch permission table";
|
|
goto fail;
|
|
}
|
|
remap_dfa_accept(profile->xmatch.dfa, 1);
|
|
}
|
|
|
|
/* disconnected attachment string is optional */
|
|
(void) unpack_str(e, &profile->disconnected, "disconnected");
|
|
|
|
/* per profile debug flags (complain, audit) */
|
|
if (!unpack_nameX(e, AA_STRUCT, "flags")) {
|
|
info = "profile missing flags";
|
|
goto fail;
|
|
}
|
|
info = "failed to unpack profile flags";
|
|
if (!unpack_u32(e, &tmp, NULL))
|
|
goto fail;
|
|
if (tmp & PACKED_FLAG_HAT)
|
|
profile->label.flags |= FLAG_HAT;
|
|
if (tmp & PACKED_FLAG_DEBUG1)
|
|
profile->label.flags |= FLAG_DEBUG1;
|
|
if (tmp & PACKED_FLAG_DEBUG2)
|
|
profile->label.flags |= FLAG_DEBUG2;
|
|
if (!unpack_u32(e, &tmp, NULL))
|
|
goto fail;
|
|
if (tmp == PACKED_MODE_COMPLAIN || (e->version & FORCE_COMPLAIN_FLAG)) {
|
|
profile->mode = APPARMOR_COMPLAIN;
|
|
} else if (tmp == PACKED_MODE_ENFORCE) {
|
|
profile->mode = APPARMOR_ENFORCE;
|
|
} else if (tmp == PACKED_MODE_KILL) {
|
|
profile->mode = APPARMOR_KILL;
|
|
} else if (tmp == PACKED_MODE_UNCONFINED) {
|
|
profile->mode = APPARMOR_UNCONFINED;
|
|
profile->label.flags |= FLAG_UNCONFINED;
|
|
} else {
|
|
goto fail;
|
|
}
|
|
if (!unpack_u32(e, &tmp, NULL))
|
|
goto fail;
|
|
if (tmp)
|
|
profile->audit = AUDIT_ALL;
|
|
|
|
if (!unpack_nameX(e, AA_STRUCTEND, NULL))
|
|
goto fail;
|
|
|
|
/* path_flags is optional */
|
|
if (unpack_u32(e, &profile->path_flags, "path_flags"))
|
|
profile->path_flags |= profile->label.flags &
|
|
PATH_MEDIATE_DELETED;
|
|
else
|
|
/* set a default value if path_flags field is not present */
|
|
profile->path_flags = PATH_MEDIATE_DELETED;
|
|
|
|
info = "failed to unpack profile capabilities";
|
|
if (!unpack_u32(e, &(profile->caps.allow.cap[0]), NULL))
|
|
goto fail;
|
|
if (!unpack_u32(e, &(profile->caps.audit.cap[0]), NULL))
|
|
goto fail;
|
|
if (!unpack_u32(e, &(profile->caps.quiet.cap[0]), NULL))
|
|
goto fail;
|
|
if (!unpack_u32(e, &tmpcap.cap[0], NULL))
|
|
goto fail;
|
|
|
|
info = "failed to unpack upper profile capabilities";
|
|
if (unpack_nameX(e, AA_STRUCT, "caps64")) {
|
|
/* optional upper half of 64 bit caps */
|
|
if (!unpack_u32(e, &(profile->caps.allow.cap[1]), NULL))
|
|
goto fail;
|
|
if (!unpack_u32(e, &(profile->caps.audit.cap[1]), NULL))
|
|
goto fail;
|
|
if (!unpack_u32(e, &(profile->caps.quiet.cap[1]), NULL))
|
|
goto fail;
|
|
if (!unpack_u32(e, &(tmpcap.cap[1]), NULL))
|
|
goto fail;
|
|
if (!unpack_nameX(e, AA_STRUCTEND, NULL))
|
|
goto fail;
|
|
}
|
|
|
|
info = "failed to unpack extended profile capabilities";
|
|
if (unpack_nameX(e, AA_STRUCT, "capsx")) {
|
|
/* optional extended caps mediation mask */
|
|
if (!unpack_u32(e, &(profile->caps.extended.cap[0]), NULL))
|
|
goto fail;
|
|
if (!unpack_u32(e, &(profile->caps.extended.cap[1]), NULL))
|
|
goto fail;
|
|
if (!unpack_nameX(e, AA_STRUCTEND, NULL))
|
|
goto fail;
|
|
}
|
|
|
|
if (!unpack_xattrs(e, profile)) {
|
|
info = "failed to unpack profile xattrs";
|
|
goto fail;
|
|
}
|
|
|
|
if (!unpack_rlimits(e, profile)) {
|
|
info = "failed to unpack profile rlimits";
|
|
goto fail;
|
|
}
|
|
|
|
if (!unpack_secmark(e, profile)) {
|
|
info = "failed to unpack profile secmark rules";
|
|
goto fail;
|
|
}
|
|
|
|
if (unpack_nameX(e, AA_STRUCT, "policydb")) {
|
|
/* generic policy dfa - optional and may be NULL */
|
|
info = "failed to unpack policydb";
|
|
profile->policy.dfa = unpack_dfa(e);
|
|
if (IS_ERR(profile->policy.dfa)) {
|
|
error = PTR_ERR(profile->policy.dfa);
|
|
profile->policy.dfa = NULL;
|
|
goto fail;
|
|
} else if (!profile->policy.dfa) {
|
|
error = -EPROTO;
|
|
goto fail;
|
|
}
|
|
if (!unpack_u32(e, &profile->policy.start[0], "start"))
|
|
/* default start state */
|
|
profile->policy.start[0] = DFA_START;
|
|
/* setup class index */
|
|
for (i = AA_CLASS_FILE; i <= AA_CLASS_LAST; i++) {
|
|
profile->policy.start[i] =
|
|
aa_dfa_next(profile->policy.dfa,
|
|
profile->policy.start[0],
|
|
i);
|
|
}
|
|
if (!unpack_nameX(e, AA_STRUCTEND, NULL))
|
|
goto fail;
|
|
profile->policy.perms = compute_perms(profile->policy.dfa,
|
|
e->version);
|
|
if (!profile->policy.perms) {
|
|
info = "failed to remap policydb permission table";
|
|
goto fail;
|
|
}
|
|
/* Do not remap internal dfas */
|
|
remap_dfa_accept(profile->policy.dfa, 1);
|
|
} else
|
|
profile->policy.dfa = aa_get_dfa(nulldfa);
|
|
|
|
/* get file rules */
|
|
profile->file.dfa = unpack_dfa(e);
|
|
if (IS_ERR(profile->file.dfa)) {
|
|
error = PTR_ERR(profile->file.dfa);
|
|
profile->file.dfa = NULL;
|
|
info = "failed to unpack profile file rules";
|
|
goto fail;
|
|
} else if (profile->file.dfa) {
|
|
if (!unpack_u32(e, &profile->file.start[AA_CLASS_FILE],
|
|
"dfa_start"))
|
|
/* default start state */
|
|
profile->file.start[AA_CLASS_FILE] = DFA_START;
|
|
profile->file.perms = compute_fperms(profile->file.dfa);
|
|
if (!profile->file.perms) {
|
|
info = "failed to remap file permission table";
|
|
goto fail;
|
|
}
|
|
remap_dfa_accept(profile->file.dfa, 2);
|
|
if (!unpack_trans_table(e, profile)) {
|
|
info = "failed to unpack profile transition table";
|
|
goto fail;
|
|
}
|
|
} else if (profile->policy.dfa &&
|
|
profile->policy.start[AA_CLASS_FILE]) {
|
|
profile->file.dfa = aa_get_dfa(profile->policy.dfa);
|
|
profile->file.start[AA_CLASS_FILE] = profile->policy.start[AA_CLASS_FILE];
|
|
} else
|
|
profile->file.dfa = aa_get_dfa(nulldfa);
|
|
|
|
if (unpack_nameX(e, AA_STRUCT, "data")) {
|
|
info = "out of memory";
|
|
profile->data = kzalloc(sizeof(*profile->data), GFP_KERNEL);
|
|
if (!profile->data)
|
|
goto fail;
|
|
|
|
params.nelem_hint = 3;
|
|
params.key_len = sizeof(void *);
|
|
params.key_offset = offsetof(struct aa_data, key);
|
|
params.head_offset = offsetof(struct aa_data, head);
|
|
params.hashfn = strhash;
|
|
params.obj_cmpfn = datacmp;
|
|
|
|
if (rhashtable_init(profile->data, ¶ms)) {
|
|
info = "failed to init key, value hash table";
|
|
goto fail;
|
|
}
|
|
|
|
while (unpack_strdup(e, &key, NULL)) {
|
|
data = kzalloc(sizeof(*data), GFP_KERNEL);
|
|
if (!data) {
|
|
kfree_sensitive(key);
|
|
goto fail;
|
|
}
|
|
|
|
data->key = key;
|
|
data->size = unpack_blob(e, &data->data, NULL);
|
|
data->data = kvmemdup(data->data, data->size);
|
|
if (data->size && !data->data) {
|
|
kfree_sensitive(data->key);
|
|
kfree_sensitive(data);
|
|
goto fail;
|
|
}
|
|
|
|
rhashtable_insert_fast(profile->data, &data->head,
|
|
profile->data->p);
|
|
}
|
|
|
|
if (!unpack_nameX(e, AA_STRUCTEND, NULL)) {
|
|
info = "failed to unpack end of key, value data table";
|
|
goto fail;
|
|
}
|
|
}
|
|
|
|
if (!unpack_nameX(e, AA_STRUCTEND, NULL)) {
|
|
info = "failed to unpack end of profile";
|
|
goto fail;
|
|
}
|
|
|
|
return profile;
|
|
|
|
fail:
|
|
if (profile)
|
|
name = NULL;
|
|
else if (!name)
|
|
name = "unknown";
|
|
audit_iface(profile, NULL, name, info, e, error);
|
|
aa_free_profile(profile);
|
|
|
|
return ERR_PTR(error);
|
|
}
|
|
|
|
/**
|
|
* verify_header - unpack serialized stream header
|
|
* @e: serialized data read head (NOT NULL)
|
|
* @required: whether the header is required or optional
|
|
* @ns: Returns - namespace if one is specified else NULL (NOT NULL)
|
|
*
|
|
* Returns: error or 0 if header is good
|
|
*/
|
|
static int verify_header(struct aa_ext *e, int required, const char **ns)
|
|
{
|
|
int error = -EPROTONOSUPPORT;
|
|
const char *name = NULL;
|
|
*ns = NULL;
|
|
|
|
/* get the interface version */
|
|
if (!unpack_u32(e, &e->version, "version")) {
|
|
if (required) {
|
|
audit_iface(NULL, NULL, NULL, "invalid profile format",
|
|
e, error);
|
|
return error;
|
|
}
|
|
}
|
|
|
|
/* Check that the interface version is currently supported.
|
|
* if not specified use previous version
|
|
* Mask off everything that is not kernel abi version
|
|
*/
|
|
if (VERSION_LT(e->version, v5) || VERSION_GT(e->version, v9)) {
|
|
audit_iface(NULL, NULL, NULL, "unsupported interface version",
|
|
e, error);
|
|
return error;
|
|
}
|
|
|
|
/* read the namespace if present */
|
|
if (unpack_str(e, &name, "namespace")) {
|
|
if (*name == '\0') {
|
|
audit_iface(NULL, NULL, NULL, "invalid namespace name",
|
|
e, error);
|
|
return error;
|
|
}
|
|
if (*ns && strcmp(*ns, name)) {
|
|
audit_iface(NULL, NULL, NULL, "invalid ns change", e,
|
|
error);
|
|
} else if (!*ns) {
|
|
*ns = kstrdup(name, GFP_KERNEL);
|
|
if (!*ns)
|
|
return -ENOMEM;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static bool verify_xindex(int xindex, int table_size)
|
|
{
|
|
int index, xtype;
|
|
xtype = xindex & AA_X_TYPE_MASK;
|
|
index = xindex & AA_X_INDEX_MASK;
|
|
if (xtype == AA_X_TABLE && index >= table_size)
|
|
return false;
|
|
return true;
|
|
}
|
|
|
|
/* verify dfa xindexes are in range of transition tables */
|
|
static bool verify_dfa_xindex(struct aa_dfa *dfa, int table_size)
|
|
{
|
|
int i;
|
|
for (i = 0; i < dfa->tables[YYTD_ID_ACCEPT]->td_lolen; i++) {
|
|
if (!verify_xindex(ACCEPT_TABLE(dfa)[i], table_size))
|
|
return false;
|
|
}
|
|
return true;
|
|
}
|
|
|
|
/**
|
|
* verify_profile - Do post unpack analysis to verify profile consistency
|
|
* @profile: profile to verify (NOT NULL)
|
|
*
|
|
* Returns: 0 if passes verification else error
|
|
*
|
|
* This verification is post any unpack mapping or changes
|
|
*/
|
|
static int verify_profile(struct aa_profile *profile)
|
|
{
|
|
if ((profile->file.dfa &&
|
|
!verify_dfa_xindex(profile->file.dfa,
|
|
profile->file.trans.size)) ||
|
|
(profile->policy.dfa &&
|
|
!verify_dfa_xindex(profile->policy.dfa,
|
|
profile->policy.trans.size))) {
|
|
audit_iface(profile, NULL, NULL,
|
|
"Unpack: Invalid named transition", NULL, -EPROTO);
|
|
return -EPROTO;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
void aa_load_ent_free(struct aa_load_ent *ent)
|
|
{
|
|
if (ent) {
|
|
aa_put_profile(ent->rename);
|
|
aa_put_profile(ent->old);
|
|
aa_put_profile(ent->new);
|
|
kfree(ent->ns_name);
|
|
kfree_sensitive(ent);
|
|
}
|
|
}
|
|
|
|
struct aa_load_ent *aa_load_ent_alloc(void)
|
|
{
|
|
struct aa_load_ent *ent = kzalloc(sizeof(*ent), GFP_KERNEL);
|
|
if (ent)
|
|
INIT_LIST_HEAD(&ent->list);
|
|
return ent;
|
|
}
|
|
|
|
static int compress_zstd(const char *src, size_t slen, char **dst, size_t *dlen)
|
|
{
|
|
#ifdef CONFIG_SECURITY_APPARMOR_EXPORT_BINARY
|
|
const zstd_parameters params =
|
|
zstd_get_params(aa_g_rawdata_compression_level, slen);
|
|
const size_t wksp_len = zstd_cctx_workspace_bound(¶ms.cParams);
|
|
void *wksp = NULL;
|
|
zstd_cctx *ctx = NULL;
|
|
size_t out_len = zstd_compress_bound(slen);
|
|
void *out = NULL;
|
|
int ret = 0;
|
|
|
|
out = kvzalloc(out_len, GFP_KERNEL);
|
|
if (!out) {
|
|
ret = -ENOMEM;
|
|
goto cleanup;
|
|
}
|
|
|
|
wksp = kvzalloc(wksp_len, GFP_KERNEL);
|
|
if (!wksp) {
|
|
ret = -ENOMEM;
|
|
goto cleanup;
|
|
}
|
|
|
|
ctx = zstd_init_cctx(wksp, wksp_len);
|
|
if (!ctx) {
|
|
ret = -EINVAL;
|
|
goto cleanup;
|
|
}
|
|
|
|
out_len = zstd_compress_cctx(ctx, out, out_len, src, slen, ¶ms);
|
|
if (zstd_is_error(out_len)) {
|
|
ret = -EINVAL;
|
|
goto cleanup;
|
|
}
|
|
|
|
if (is_vmalloc_addr(out)) {
|
|
*dst = kvzalloc(out_len, GFP_KERNEL);
|
|
if (*dst) {
|
|
memcpy(*dst, out, out_len);
|
|
kvfree(out);
|
|
out = NULL;
|
|
}
|
|
} else {
|
|
/*
|
|
* If the staging buffer was kmalloc'd, then using krealloc is
|
|
* probably going to be faster. The destination buffer will
|
|
* always be smaller, so it's just shrunk, avoiding a memcpy
|
|
*/
|
|
*dst = krealloc(out, out_len, GFP_KERNEL);
|
|
}
|
|
|
|
if (!*dst) {
|
|
ret = -ENOMEM;
|
|
goto cleanup;
|
|
}
|
|
|
|
*dlen = out_len;
|
|
|
|
cleanup:
|
|
if (ret) {
|
|
kvfree(out);
|
|
*dst = NULL;
|
|
}
|
|
|
|
kvfree(wksp);
|
|
return ret;
|
|
#else
|
|
*dlen = slen;
|
|
return 0;
|
|
#endif
|
|
}
|
|
|
|
static int compress_loaddata(struct aa_loaddata *data)
|
|
{
|
|
AA_BUG(data->compressed_size > 0);
|
|
|
|
/*
|
|
* Shortcut the no compression case, else we increase the amount of
|
|
* storage required by a small amount
|
|
*/
|
|
if (aa_g_rawdata_compression_level != 0) {
|
|
void *udata = data->data;
|
|
int error = compress_zstd(udata, data->size, &data->data,
|
|
&data->compressed_size);
|
|
if (error)
|
|
return error;
|
|
|
|
if (udata != data->data)
|
|
kvfree(udata);
|
|
} else
|
|
data->compressed_size = data->size;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* aa_unpack - unpack packed binary profile(s) data loaded from user space
|
|
* @udata: user data copied to kmem (NOT NULL)
|
|
* @lh: list to place unpacked profiles in a aa_repl_ws
|
|
* @ns: Returns namespace profile is in if specified else NULL (NOT NULL)
|
|
*
|
|
* Unpack user data and return refcounted allocated profile(s) stored in
|
|
* @lh in order of discovery, with the list chain stored in base.list
|
|
* or error
|
|
*
|
|
* Returns: profile(s) on @lh else error pointer if fails to unpack
|
|
*/
|
|
int aa_unpack(struct aa_loaddata *udata, struct list_head *lh,
|
|
const char **ns)
|
|
{
|
|
struct aa_load_ent *tmp, *ent;
|
|
struct aa_profile *profile = NULL;
|
|
int error;
|
|
struct aa_ext e = {
|
|
.start = udata->data,
|
|
.end = udata->data + udata->size,
|
|
.pos = udata->data,
|
|
};
|
|
|
|
*ns = NULL;
|
|
while (e.pos < e.end) {
|
|
char *ns_name = NULL;
|
|
void *start;
|
|
error = verify_header(&e, e.pos == e.start, ns);
|
|
if (error)
|
|
goto fail;
|
|
|
|
start = e.pos;
|
|
profile = unpack_profile(&e, &ns_name);
|
|
if (IS_ERR(profile)) {
|
|
error = PTR_ERR(profile);
|
|
goto fail;
|
|
}
|
|
|
|
error = verify_profile(profile);
|
|
if (error)
|
|
goto fail_profile;
|
|
|
|
if (aa_g_hash_policy)
|
|
error = aa_calc_profile_hash(profile, e.version, start,
|
|
e.pos - start);
|
|
if (error)
|
|
goto fail_profile;
|
|
|
|
ent = aa_load_ent_alloc();
|
|
if (!ent) {
|
|
error = -ENOMEM;
|
|
goto fail_profile;
|
|
}
|
|
|
|
ent->new = profile;
|
|
ent->ns_name = ns_name;
|
|
list_add_tail(&ent->list, lh);
|
|
}
|
|
udata->abi = e.version & K_ABI_MASK;
|
|
if (aa_g_hash_policy) {
|
|
udata->hash = aa_calc_hash(udata->data, udata->size);
|
|
if (IS_ERR(udata->hash)) {
|
|
error = PTR_ERR(udata->hash);
|
|
udata->hash = NULL;
|
|
goto fail;
|
|
}
|
|
}
|
|
|
|
if (aa_g_export_binary) {
|
|
error = compress_loaddata(udata);
|
|
if (error)
|
|
goto fail;
|
|
}
|
|
return 0;
|
|
|
|
fail_profile:
|
|
aa_put_profile(profile);
|
|
|
|
fail:
|
|
list_for_each_entry_safe(ent, tmp, lh, list) {
|
|
list_del_init(&ent->list);
|
|
aa_load_ent_free(ent);
|
|
}
|
|
|
|
return error;
|
|
}
|
|
|
|
#ifdef CONFIG_SECURITY_APPARMOR_KUNIT_TEST
|
|
#include "policy_unpack_test.c"
|
|
#endif /* CONFIG_SECURITY_APPARMOR_KUNIT_TEST */
|