Merge master.kernel.org:/pub/scm/linux/kernel/git/davem/net-2.6
* master.kernel.org:/pub/scm/linux/kernel/git/davem/net-2.6: [PKT_SCHED] sch_htb: use rb_first() cleanup [RTNETLINK]: Fix use of wrong skb in do_getlink() [DECNET]: Fix sfuzz hanging on 2.6.18 [NET]: Do not memcmp() over pad bytes of struct flowi. [NET]: Introduce protocol-specific destructor for time-wait sockets. [NET]: Use typesafe inet_twsk() inline function instead of cast. [NET]: Use hton{l,s}() for non-initializers. [TCP]: Use TCPOLEN_TSTAMP_ALIGNED macro instead of magic number. [IPV6]: Seperate sit driver to extra module (addrconf.c changes) [IPV6]: Seperate sit driver to extra module [NET]: File descriptor loss while receiving SCM_RIGHTS [SCTP]: Fix the RX queue size shown in /proc/net/sctp/assocs output. [SCTP]: Fix receive buffer accounting. SELinux: Bug fix in polidydb_destroy IPsec: fix handling of errors for socket policies IPsec: correct semantics for SELinux policy matching IPsec: propagate security module errors up from flow_cache_lookup NetLabel: use SECINITSID_UNLABELED for a base SID NetLabel: fix a cache race condition
This commit is contained in:
commit
9ff4680e99
43 changed files with 409 additions and 217 deletions
|
@ -882,7 +882,8 @@ struct request_sock;
|
||||||
* Check permission when a flow selects a xfrm_policy for processing
|
* Check permission when a flow selects a xfrm_policy for processing
|
||||||
* XFRMs on a packet. The hook is called when selecting either a
|
* XFRMs on a packet. The hook is called when selecting either a
|
||||||
* per-socket policy or a generic xfrm policy.
|
* per-socket policy or a generic xfrm policy.
|
||||||
* Return 0 if permission is granted.
|
* Return 0 if permission is granted, -ESRCH otherwise, or -errno
|
||||||
|
* on other errors.
|
||||||
* @xfrm_state_pol_flow_match:
|
* @xfrm_state_pol_flow_match:
|
||||||
* @x contains the state to match.
|
* @x contains the state to match.
|
||||||
* @xp contains the policy to check for a match.
|
* @xp contains the policy to check for a match.
|
||||||
|
@ -891,6 +892,7 @@ struct request_sock;
|
||||||
* @xfrm_flow_state_match:
|
* @xfrm_flow_state_match:
|
||||||
* @fl contains the flow key to match.
|
* @fl contains the flow key to match.
|
||||||
* @xfrm points to the xfrm_state to match.
|
* @xfrm points to the xfrm_state to match.
|
||||||
|
* @xp points to the xfrm_policy to match.
|
||||||
* Return 1 if there is a match.
|
* Return 1 if there is a match.
|
||||||
* @xfrm_decode_session:
|
* @xfrm_decode_session:
|
||||||
* @skb points to skb to decode.
|
* @skb points to skb to decode.
|
||||||
|
@ -1388,7 +1390,8 @@ struct security_operations {
|
||||||
int (*xfrm_policy_lookup)(struct xfrm_policy *xp, u32 fl_secid, u8 dir);
|
int (*xfrm_policy_lookup)(struct xfrm_policy *xp, u32 fl_secid, u8 dir);
|
||||||
int (*xfrm_state_pol_flow_match)(struct xfrm_state *x,
|
int (*xfrm_state_pol_flow_match)(struct xfrm_state *x,
|
||||||
struct xfrm_policy *xp, struct flowi *fl);
|
struct xfrm_policy *xp, struct flowi *fl);
|
||||||
int (*xfrm_flow_state_match)(struct flowi *fl, struct xfrm_state *xfrm);
|
int (*xfrm_flow_state_match)(struct flowi *fl, struct xfrm_state *xfrm,
|
||||||
|
struct xfrm_policy *xp);
|
||||||
int (*xfrm_decode_session)(struct sk_buff *skb, u32 *secid, int ckall);
|
int (*xfrm_decode_session)(struct sk_buff *skb, u32 *secid, int ckall);
|
||||||
#endif /* CONFIG_SECURITY_NETWORK_XFRM */
|
#endif /* CONFIG_SECURITY_NETWORK_XFRM */
|
||||||
|
|
||||||
|
@ -3120,11 +3123,6 @@ static inline int security_xfrm_policy_alloc(struct xfrm_policy *xp, struct xfrm
|
||||||
return security_ops->xfrm_policy_alloc_security(xp, sec_ctx, NULL);
|
return security_ops->xfrm_policy_alloc_security(xp, sec_ctx, NULL);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline int security_xfrm_sock_policy_alloc(struct xfrm_policy *xp, struct sock *sk)
|
|
||||||
{
|
|
||||||
return security_ops->xfrm_policy_alloc_security(xp, NULL, sk);
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int security_xfrm_policy_clone(struct xfrm_policy *old, struct xfrm_policy *new)
|
static inline int security_xfrm_policy_clone(struct xfrm_policy *old, struct xfrm_policy *new)
|
||||||
{
|
{
|
||||||
return security_ops->xfrm_policy_clone_security(old, new);
|
return security_ops->xfrm_policy_clone_security(old, new);
|
||||||
|
@ -3175,9 +3173,10 @@ static inline int security_xfrm_state_pol_flow_match(struct xfrm_state *x,
|
||||||
return security_ops->xfrm_state_pol_flow_match(x, xp, fl);
|
return security_ops->xfrm_state_pol_flow_match(x, xp, fl);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline int security_xfrm_flow_state_match(struct flowi *fl, struct xfrm_state *xfrm)
|
static inline int security_xfrm_flow_state_match(struct flowi *fl,
|
||||||
|
struct xfrm_state *xfrm, struct xfrm_policy *xp)
|
||||||
{
|
{
|
||||||
return security_ops->xfrm_flow_state_match(fl, xfrm);
|
return security_ops->xfrm_flow_state_match(fl, xfrm, xp);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline int security_xfrm_decode_session(struct sk_buff *skb, u32 *secid)
|
static inline int security_xfrm_decode_session(struct sk_buff *skb, u32 *secid)
|
||||||
|
@ -3197,11 +3196,6 @@ static inline int security_xfrm_policy_alloc(struct xfrm_policy *xp, struct xfrm
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline int security_xfrm_sock_policy_alloc(struct xfrm_policy *xp, struct sock *sk)
|
|
||||||
{
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
static inline int security_xfrm_policy_clone(struct xfrm_policy *old, struct xfrm_policy *new)
|
static inline int security_xfrm_policy_clone(struct xfrm_policy *old, struct xfrm_policy *new)
|
||||||
{
|
{
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -3249,7 +3243,7 @@ static inline int security_xfrm_state_pol_flow_match(struct xfrm_state *x,
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline int security_xfrm_flow_state_match(struct flowi *fl,
|
static inline int security_xfrm_flow_state_match(struct flowi *fl,
|
||||||
struct xfrm_state *xfrm)
|
struct xfrm_state *xfrm, struct xfrm_policy *xp)
|
||||||
{
|
{
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
|
@ -97,7 +97,7 @@ struct flowi {
|
||||||
#define FLOW_DIR_FWD 2
|
#define FLOW_DIR_FWD 2
|
||||||
|
|
||||||
struct sock;
|
struct sock;
|
||||||
typedef void (*flow_resolve_t)(struct flowi *key, u16 family, u8 dir,
|
typedef int (*flow_resolve_t)(struct flowi *key, u16 family, u8 dir,
|
||||||
void **objp, atomic_t **obj_refp);
|
void **objp, atomic_t **obj_refp);
|
||||||
|
|
||||||
extern void *flow_cache_lookup(struct flowi *key, u16 family, u8 dir,
|
extern void *flow_cache_lookup(struct flowi *key, u16 family, u8 dir,
|
||||||
|
|
|
@ -196,6 +196,7 @@ static inline void inet_twsk_put(struct inet_timewait_sock *tw)
|
||||||
{
|
{
|
||||||
if (atomic_dec_and_test(&tw->tw_refcnt)) {
|
if (atomic_dec_and_test(&tw->tw_refcnt)) {
|
||||||
struct module *owner = tw->tw_prot->owner;
|
struct module *owner = tw->tw_prot->owner;
|
||||||
|
twsk_destructor((struct sock *)tw);
|
||||||
#ifdef SOCK_REFCNT_DEBUG
|
#ifdef SOCK_REFCNT_DEBUG
|
||||||
printk(KERN_DEBUG "%s timewait_sock %p released\n",
|
printk(KERN_DEBUG "%s timewait_sock %p released\n",
|
||||||
tw->tw_prot->name, tw);
|
tw->tw_prot->name, tw);
|
||||||
|
|
|
@ -34,6 +34,7 @@
|
||||||
#include <linux/net.h>
|
#include <linux/net.h>
|
||||||
#include <linux/skbuff.h>
|
#include <linux/skbuff.h>
|
||||||
#include <net/netlink.h>
|
#include <net/netlink.h>
|
||||||
|
#include <asm/atomic.h>
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* NetLabel - A management interface for maintaining network packet label
|
* NetLabel - A management interface for maintaining network packet label
|
||||||
|
@ -106,6 +107,7 @@ int netlbl_domhsh_remove(const char *domain, struct netlbl_audit *audit_info);
|
||||||
|
|
||||||
/* LSM security attributes */
|
/* LSM security attributes */
|
||||||
struct netlbl_lsm_cache {
|
struct netlbl_lsm_cache {
|
||||||
|
atomic_t refcount;
|
||||||
void (*free) (const void *data);
|
void (*free) (const void *data);
|
||||||
void *data;
|
void *data;
|
||||||
};
|
};
|
||||||
|
@ -117,7 +119,7 @@ struct netlbl_lsm_secattr {
|
||||||
unsigned char *mls_cat;
|
unsigned char *mls_cat;
|
||||||
size_t mls_cat_len;
|
size_t mls_cat_len;
|
||||||
|
|
||||||
struct netlbl_lsm_cache cache;
|
struct netlbl_lsm_cache *cache;
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -125,6 +127,43 @@ struct netlbl_lsm_secattr {
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
|
||||||
|
/**
|
||||||
|
* netlbl_secattr_cache_alloc - Allocate and initialize a secattr cache
|
||||||
|
* @flags: the memory allocation flags
|
||||||
|
*
|
||||||
|
* Description:
|
||||||
|
* Allocate and initialize a netlbl_lsm_cache structure. Returns a pointer
|
||||||
|
* on success, NULL on failure.
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
static inline struct netlbl_lsm_cache *netlbl_secattr_cache_alloc(int flags)
|
||||||
|
{
|
||||||
|
struct netlbl_lsm_cache *cache;
|
||||||
|
|
||||||
|
cache = kzalloc(sizeof(*cache), flags);
|
||||||
|
if (cache)
|
||||||
|
atomic_set(&cache->refcount, 1);
|
||||||
|
return cache;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* netlbl_secattr_cache_free - Frees a netlbl_lsm_cache struct
|
||||||
|
* @cache: the struct to free
|
||||||
|
*
|
||||||
|
* Description:
|
||||||
|
* Frees @secattr including all of the internal buffers.
|
||||||
|
*
|
||||||
|
*/
|
||||||
|
static inline void netlbl_secattr_cache_free(struct netlbl_lsm_cache *cache)
|
||||||
|
{
|
||||||
|
if (!atomic_dec_and_test(&cache->refcount))
|
||||||
|
return;
|
||||||
|
|
||||||
|
if (cache->free)
|
||||||
|
cache->free(cache->data);
|
||||||
|
kfree(cache);
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* netlbl_secattr_init - Initialize a netlbl_lsm_secattr struct
|
* netlbl_secattr_init - Initialize a netlbl_lsm_secattr struct
|
||||||
* @secattr: the struct to initialize
|
* @secattr: the struct to initialize
|
||||||
|
@ -143,20 +182,16 @@ static inline int netlbl_secattr_init(struct netlbl_lsm_secattr *secattr)
|
||||||
/**
|
/**
|
||||||
* netlbl_secattr_destroy - Clears a netlbl_lsm_secattr struct
|
* netlbl_secattr_destroy - Clears a netlbl_lsm_secattr struct
|
||||||
* @secattr: the struct to clear
|
* @secattr: the struct to clear
|
||||||
* @clear_cache: cache clear flag
|
|
||||||
*
|
*
|
||||||
* Description:
|
* Description:
|
||||||
* Destroys the @secattr struct, including freeing all of the internal buffers.
|
* Destroys the @secattr struct, including freeing all of the internal buffers.
|
||||||
* If @clear_cache is true then free the cache fields, otherwise leave them
|
* The struct must be reset with a call to netlbl_secattr_init() before reuse.
|
||||||
* intact. The struct must be reset with a call to netlbl_secattr_init()
|
|
||||||
* before reuse.
|
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
static inline void netlbl_secattr_destroy(struct netlbl_lsm_secattr *secattr,
|
static inline void netlbl_secattr_destroy(struct netlbl_lsm_secattr *secattr)
|
||||||
u32 clear_cache)
|
|
||||||
{
|
{
|
||||||
if (clear_cache && secattr->cache.data != NULL && secattr->cache.free)
|
if (secattr->cache)
|
||||||
secattr->cache.free(secattr->cache.data);
|
netlbl_secattr_cache_free(secattr->cache);
|
||||||
kfree(secattr->domain);
|
kfree(secattr->domain);
|
||||||
kfree(secattr->mls_cat);
|
kfree(secattr->mls_cat);
|
||||||
}
|
}
|
||||||
|
@ -178,17 +213,14 @@ static inline struct netlbl_lsm_secattr *netlbl_secattr_alloc(int flags)
|
||||||
/**
|
/**
|
||||||
* netlbl_secattr_free - Frees a netlbl_lsm_secattr struct
|
* netlbl_secattr_free - Frees a netlbl_lsm_secattr struct
|
||||||
* @secattr: the struct to free
|
* @secattr: the struct to free
|
||||||
* @clear_cache: cache clear flag
|
|
||||||
*
|
*
|
||||||
* Description:
|
* Description:
|
||||||
* Frees @secattr including all of the internal buffers. If @clear_cache is
|
* Frees @secattr including all of the internal buffers.
|
||||||
* true then free the cache fields, otherwise leave them intact.
|
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
static inline void netlbl_secattr_free(struct netlbl_lsm_secattr *secattr,
|
static inline void netlbl_secattr_free(struct netlbl_lsm_secattr *secattr)
|
||||||
u32 clear_cache)
|
|
||||||
{
|
{
|
||||||
netlbl_secattr_destroy(secattr, clear_cache);
|
netlbl_secattr_destroy(secattr);
|
||||||
kfree(secattr);
|
kfree(secattr);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -139,6 +139,7 @@ int sctp_inet_listen(struct socket *sock, int backlog);
|
||||||
void sctp_write_space(struct sock *sk);
|
void sctp_write_space(struct sock *sk);
|
||||||
unsigned int sctp_poll(struct file *file, struct socket *sock,
|
unsigned int sctp_poll(struct file *file, struct socket *sock,
|
||||||
poll_table *wait);
|
poll_table *wait);
|
||||||
|
void sctp_sock_rfree(struct sk_buff *skb);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* sctp/primitive.c
|
* sctp/primitive.c
|
||||||
|
@ -444,6 +445,19 @@ static inline struct list_head *sctp_list_dequeue(struct list_head *list)
|
||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* SCTP version of skb_set_owner_r. We need this one because
|
||||||
|
* of the way we have to do receive buffer accounting on bundled
|
||||||
|
* chunks.
|
||||||
|
*/
|
||||||
|
static inline void sctp_skb_set_owner_r(struct sk_buff *skb, struct sock *sk)
|
||||||
|
{
|
||||||
|
struct sctp_ulpevent *event = sctp_skb2event(skb);
|
||||||
|
|
||||||
|
skb->sk = sk;
|
||||||
|
skb->destructor = sctp_sock_rfree;
|
||||||
|
atomic_add(event->rmem_len, &sk->sk_rmem_alloc);
|
||||||
|
}
|
||||||
|
|
||||||
/* Tests if the list has one and only one entry. */
|
/* Tests if the list has one and only one entry. */
|
||||||
static inline int sctp_list_single_entry(struct list_head *head)
|
static inline int sctp_list_single_entry(struct list_head *head)
|
||||||
{
|
{
|
||||||
|
|
|
@ -63,6 +63,7 @@ struct sctp_ulpevent {
|
||||||
__u32 cumtsn;
|
__u32 cumtsn;
|
||||||
int msg_flags;
|
int msg_flags;
|
||||||
int iif;
|
int iif;
|
||||||
|
unsigned int rmem_len;
|
||||||
};
|
};
|
||||||
|
|
||||||
/* Retrieve the skb this event sits inside of. */
|
/* Retrieve the skb this event sits inside of. */
|
||||||
|
|
|
@ -19,6 +19,7 @@ struct timewait_sock_ops {
|
||||||
unsigned int twsk_obj_size;
|
unsigned int twsk_obj_size;
|
||||||
int (*twsk_unique)(struct sock *sk,
|
int (*twsk_unique)(struct sock *sk,
|
||||||
struct sock *sktw, void *twp);
|
struct sock *sktw, void *twp);
|
||||||
|
void (*twsk_destructor)(struct sock *sk);
|
||||||
};
|
};
|
||||||
|
|
||||||
static inline int twsk_unique(struct sock *sk, struct sock *sktw, void *twp)
|
static inline int twsk_unique(struct sock *sk, struct sock *sktw, void *twp)
|
||||||
|
@ -28,4 +29,10 @@ static inline int twsk_unique(struct sock *sk, struct sock *sktw, void *twp)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static inline void twsk_destructor(struct sock *sk)
|
||||||
|
{
|
||||||
|
if (sk->sk_prot->twsk_prot->twsk_destructor != NULL)
|
||||||
|
sk->sk_prot->twsk_prot->twsk_destructor(sk);
|
||||||
|
}
|
||||||
|
|
||||||
#endif /* _TIMEWAIT_SOCK_H */
|
#endif /* _TIMEWAIT_SOCK_H */
|
||||||
|
|
|
@ -995,7 +995,8 @@ struct xfrm_state * xfrm_find_acq(u8 mode, u32 reqid, u8 proto,
|
||||||
int create, unsigned short family);
|
int create, unsigned short family);
|
||||||
extern void xfrm_policy_flush(u8 type);
|
extern void xfrm_policy_flush(u8 type);
|
||||||
extern int xfrm_sk_policy_insert(struct sock *sk, int dir, struct xfrm_policy *pol);
|
extern int xfrm_sk_policy_insert(struct sock *sk, int dir, struct xfrm_policy *pol);
|
||||||
extern int xfrm_bundle_ok(struct xfrm_dst *xdst, struct flowi *fl, int family, int strict);
|
extern int xfrm_bundle_ok(struct xfrm_policy *pol, struct xfrm_dst *xdst,
|
||||||
|
struct flowi *fl, int family, int strict);
|
||||||
extern void xfrm_init_pmtu(struct dst_entry *dst);
|
extern void xfrm_init_pmtu(struct dst_entry *dst);
|
||||||
|
|
||||||
extern wait_queue_head_t km_waitq;
|
extern wait_queue_head_t km_waitq;
|
||||||
|
|
|
@ -285,8 +285,7 @@ void scm_detach_fds_compat(struct msghdr *kmsg, struct scm_cookie *scm)
|
||||||
|
|
||||||
if (i > 0) {
|
if (i > 0) {
|
||||||
int cmlen = CMSG_COMPAT_LEN(i * sizeof(int));
|
int cmlen = CMSG_COMPAT_LEN(i * sizeof(int));
|
||||||
if (!err)
|
err = put_user(SOL_SOCKET, &cm->cmsg_level);
|
||||||
err = put_user(SOL_SOCKET, &cm->cmsg_level);
|
|
||||||
if (!err)
|
if (!err)
|
||||||
err = put_user(SCM_RIGHTS, &cm->cmsg_type);
|
err = put_user(SCM_RIGHTS, &cm->cmsg_type);
|
||||||
if (!err)
|
if (!err)
|
||||||
|
|
|
@ -85,6 +85,14 @@ static void flow_cache_new_hashrnd(unsigned long arg)
|
||||||
add_timer(&flow_hash_rnd_timer);
|
add_timer(&flow_hash_rnd_timer);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void flow_entry_kill(int cpu, struct flow_cache_entry *fle)
|
||||||
|
{
|
||||||
|
if (fle->object)
|
||||||
|
atomic_dec(fle->object_ref);
|
||||||
|
kmem_cache_free(flow_cachep, fle);
|
||||||
|
flow_count(cpu)--;
|
||||||
|
}
|
||||||
|
|
||||||
static void __flow_cache_shrink(int cpu, int shrink_to)
|
static void __flow_cache_shrink(int cpu, int shrink_to)
|
||||||
{
|
{
|
||||||
struct flow_cache_entry *fle, **flp;
|
struct flow_cache_entry *fle, **flp;
|
||||||
|
@ -100,10 +108,7 @@ static void __flow_cache_shrink(int cpu, int shrink_to)
|
||||||
}
|
}
|
||||||
while ((fle = *flp) != NULL) {
|
while ((fle = *flp) != NULL) {
|
||||||
*flp = fle->next;
|
*flp = fle->next;
|
||||||
if (fle->object)
|
flow_entry_kill(cpu, fle);
|
||||||
atomic_dec(fle->object_ref);
|
|
||||||
kmem_cache_free(flow_cachep, fle);
|
|
||||||
flow_count(cpu)--;
|
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
@ -220,24 +225,33 @@ void *flow_cache_lookup(struct flowi *key, u16 family, u8 dir,
|
||||||
|
|
||||||
nocache:
|
nocache:
|
||||||
{
|
{
|
||||||
|
int err;
|
||||||
void *obj;
|
void *obj;
|
||||||
atomic_t *obj_ref;
|
atomic_t *obj_ref;
|
||||||
|
|
||||||
resolver(key, family, dir, &obj, &obj_ref);
|
err = resolver(key, family, dir, &obj, &obj_ref);
|
||||||
|
|
||||||
if (fle) {
|
if (fle) {
|
||||||
fle->genid = atomic_read(&flow_cache_genid);
|
if (err) {
|
||||||
|
/* Force security policy check on next lookup */
|
||||||
|
*head = fle->next;
|
||||||
|
flow_entry_kill(cpu, fle);
|
||||||
|
} else {
|
||||||
|
fle->genid = atomic_read(&flow_cache_genid);
|
||||||
|
|
||||||
if (fle->object)
|
if (fle->object)
|
||||||
atomic_dec(fle->object_ref);
|
atomic_dec(fle->object_ref);
|
||||||
|
|
||||||
fle->object = obj;
|
fle->object = obj;
|
||||||
fle->object_ref = obj_ref;
|
fle->object_ref = obj_ref;
|
||||||
if (obj)
|
if (obj)
|
||||||
atomic_inc(fle->object_ref);
|
atomic_inc(fle->object_ref);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
local_bh_enable();
|
local_bh_enable();
|
||||||
|
|
||||||
|
if (err)
|
||||||
|
obj = ERR_PTR(err);
|
||||||
return obj;
|
return obj;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
|
@ -602,7 +602,7 @@ static int rtnl_getlink(struct sk_buff *skb, struct nlmsghdr* nlh, void *arg)
|
||||||
goto errout;
|
goto errout;
|
||||||
}
|
}
|
||||||
|
|
||||||
err = rtnl_unicast(skb, NETLINK_CB(skb).pid);
|
err = rtnl_unicast(nskb, NETLINK_CB(skb).pid);
|
||||||
errout:
|
errout:
|
||||||
kfree(iw_buf);
|
kfree(iw_buf);
|
||||||
dev_put(dev);
|
dev_put(dev);
|
||||||
|
|
|
@ -245,8 +245,7 @@ void scm_detach_fds(struct msghdr *msg, struct scm_cookie *scm)
|
||||||
if (i > 0)
|
if (i > 0)
|
||||||
{
|
{
|
||||||
int cmlen = CMSG_LEN(i*sizeof(int));
|
int cmlen = CMSG_LEN(i*sizeof(int));
|
||||||
if (!err)
|
err = put_user(SOL_SOCKET, &cm->cmsg_level);
|
||||||
err = put_user(SOL_SOCKET, &cm->cmsg_level);
|
|
||||||
if (!err)
|
if (!err)
|
||||||
err = put_user(SCM_RIGHTS, &cm->cmsg_type);
|
err = put_user(SCM_RIGHTS, &cm->cmsg_type);
|
||||||
if (!err)
|
if (!err)
|
||||||
|
|
|
@ -311,7 +311,7 @@ static void dccp_v4_err(struct sk_buff *skb, u32 info)
|
||||||
}
|
}
|
||||||
|
|
||||||
if (sk->sk_state == DCCP_TIME_WAIT) {
|
if (sk->sk_state == DCCP_TIME_WAIT) {
|
||||||
inet_twsk_put((struct inet_timewait_sock *)sk);
|
inet_twsk_put(inet_twsk(sk));
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -614,7 +614,7 @@ static struct sock *dccp_v4_hnd_req(struct sock *sk, struct sk_buff *skb)
|
||||||
bh_lock_sock(nsk);
|
bh_lock_sock(nsk);
|
||||||
return nsk;
|
return nsk;
|
||||||
}
|
}
|
||||||
inet_twsk_put((struct inet_timewait_sock *)nsk);
|
inet_twsk_put(inet_twsk(nsk));
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -980,7 +980,7 @@ discard_and_relse:
|
||||||
goto discard_it;
|
goto discard_it;
|
||||||
|
|
||||||
do_time_wait:
|
do_time_wait:
|
||||||
inet_twsk_put((struct inet_timewait_sock *)sk);
|
inet_twsk_put(inet_twsk(sk));
|
||||||
goto no_dccp_socket;
|
goto no_dccp_socket;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -285,7 +285,7 @@ static void dccp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt,
|
||||||
}
|
}
|
||||||
|
|
||||||
if (sk->sk_state == DCCP_TIME_WAIT) {
|
if (sk->sk_state == DCCP_TIME_WAIT) {
|
||||||
inet_twsk_put((struct inet_timewait_sock *)sk);
|
inet_twsk_put(inet_twsk(sk));
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -663,7 +663,7 @@ static struct sock *dccp_v6_hnd_req(struct sock *sk,struct sk_buff *skb)
|
||||||
bh_lock_sock(nsk);
|
bh_lock_sock(nsk);
|
||||||
return nsk;
|
return nsk;
|
||||||
}
|
}
|
||||||
inet_twsk_put((struct inet_timewait_sock *)nsk);
|
inet_twsk_put(inet_twsk(nsk));
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1109,7 +1109,7 @@ discard_and_relse:
|
||||||
goto discard_it;
|
goto discard_it;
|
||||||
|
|
||||||
do_time_wait:
|
do_time_wait:
|
||||||
inet_twsk_put((struct inet_timewait_sock *)sk);
|
inet_twsk_put(inet_twsk(sk));
|
||||||
goto no_dccp_socket;
|
goto no_dccp_socket;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1178,8 +1178,10 @@ static int dn_getname(struct socket *sock, struct sockaddr *uaddr,int *uaddr_len
|
||||||
if (peer) {
|
if (peer) {
|
||||||
if ((sock->state != SS_CONNECTED &&
|
if ((sock->state != SS_CONNECTED &&
|
||||||
sock->state != SS_CONNECTING) &&
|
sock->state != SS_CONNECTING) &&
|
||||||
scp->accept_mode == ACC_IMMED)
|
scp->accept_mode == ACC_IMMED) {
|
||||||
|
release_sock(sk);
|
||||||
return -ENOTCONN;
|
return -ENOTCONN;
|
||||||
|
}
|
||||||
|
|
||||||
memcpy(sa, &scp->peer, sizeof(struct sockaddr_dn));
|
memcpy(sa, &scp->peer, sizeof(struct sockaddr_dn));
|
||||||
} else {
|
} else {
|
||||||
|
|
|
@ -267,9 +267,14 @@ static void dn_dst_link_failure(struct sk_buff *skb)
|
||||||
|
|
||||||
static inline int compare_keys(struct flowi *fl1, struct flowi *fl2)
|
static inline int compare_keys(struct flowi *fl1, struct flowi *fl2)
|
||||||
{
|
{
|
||||||
return memcmp(&fl1->nl_u.dn_u, &fl2->nl_u.dn_u, sizeof(fl1->nl_u.dn_u)) == 0 &&
|
return ((fl1->nl_u.dn_u.daddr ^ fl2->nl_u.dn_u.daddr) |
|
||||||
fl1->oif == fl2->oif &&
|
(fl1->nl_u.dn_u.saddr ^ fl2->nl_u.dn_u.saddr) |
|
||||||
fl1->iif == fl2->iif;
|
#ifdef CONFIG_IP_ROUTE_FWMARK
|
||||||
|
(fl1->nl_u.dn_u.fwmark ^ fl2->nl_u.dn_u.fwmark) |
|
||||||
|
#endif
|
||||||
|
(fl1->nl_u.dn_u.scope ^ fl2->nl_u.dn_u.scope) |
|
||||||
|
(fl1->oif ^ fl2->oif) |
|
||||||
|
(fl1->iif ^ fl2->iif)) == 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int dn_insert_route(struct dn_route *rt, unsigned hash, struct dn_route **rp)
|
static int dn_insert_route(struct dn_route *rt, unsigned hash, struct dn_route **rp)
|
||||||
|
|
|
@ -43,6 +43,7 @@
|
||||||
#include <net/tcp.h>
|
#include <net/tcp.h>
|
||||||
#include <net/netlabel.h>
|
#include <net/netlabel.h>
|
||||||
#include <net/cipso_ipv4.h>
|
#include <net/cipso_ipv4.h>
|
||||||
|
#include <asm/atomic.h>
|
||||||
#include <asm/bug.h>
|
#include <asm/bug.h>
|
||||||
|
|
||||||
struct cipso_v4_domhsh_entry {
|
struct cipso_v4_domhsh_entry {
|
||||||
|
@ -79,7 +80,7 @@ struct cipso_v4_map_cache_entry {
|
||||||
unsigned char *key;
|
unsigned char *key;
|
||||||
size_t key_len;
|
size_t key_len;
|
||||||
|
|
||||||
struct netlbl_lsm_cache lsm_data;
|
struct netlbl_lsm_cache *lsm_data;
|
||||||
|
|
||||||
u32 activity;
|
u32 activity;
|
||||||
struct list_head list;
|
struct list_head list;
|
||||||
|
@ -188,13 +189,14 @@ static void cipso_v4_doi_domhsh_free(struct rcu_head *entry)
|
||||||
* @entry: the entry to free
|
* @entry: the entry to free
|
||||||
*
|
*
|
||||||
* Description:
|
* Description:
|
||||||
* This function frees the memory associated with a cache entry.
|
* This function frees the memory associated with a cache entry including the
|
||||||
|
* LSM cache data if there are no longer any users, i.e. reference count == 0.
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
static void cipso_v4_cache_entry_free(struct cipso_v4_map_cache_entry *entry)
|
static void cipso_v4_cache_entry_free(struct cipso_v4_map_cache_entry *entry)
|
||||||
{
|
{
|
||||||
if (entry->lsm_data.free)
|
if (entry->lsm_data)
|
||||||
entry->lsm_data.free(entry->lsm_data.data);
|
netlbl_secattr_cache_free(entry->lsm_data);
|
||||||
kfree(entry->key);
|
kfree(entry->key);
|
||||||
kfree(entry);
|
kfree(entry);
|
||||||
}
|
}
|
||||||
|
@ -315,8 +317,8 @@ static int cipso_v4_cache_check(const unsigned char *key,
|
||||||
entry->key_len == key_len &&
|
entry->key_len == key_len &&
|
||||||
memcmp(entry->key, key, key_len) == 0) {
|
memcmp(entry->key, key, key_len) == 0) {
|
||||||
entry->activity += 1;
|
entry->activity += 1;
|
||||||
secattr->cache.free = entry->lsm_data.free;
|
atomic_inc(&entry->lsm_data->refcount);
|
||||||
secattr->cache.data = entry->lsm_data.data;
|
secattr->cache = entry->lsm_data;
|
||||||
if (prev_entry == NULL) {
|
if (prev_entry == NULL) {
|
||||||
spin_unlock_bh(&cipso_v4_cache[bkt].lock);
|
spin_unlock_bh(&cipso_v4_cache[bkt].lock);
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -383,8 +385,8 @@ int cipso_v4_cache_add(const struct sk_buff *skb,
|
||||||
memcpy(entry->key, cipso_ptr, cipso_ptr_len);
|
memcpy(entry->key, cipso_ptr, cipso_ptr_len);
|
||||||
entry->key_len = cipso_ptr_len;
|
entry->key_len = cipso_ptr_len;
|
||||||
entry->hash = cipso_v4_map_cache_hash(cipso_ptr, cipso_ptr_len);
|
entry->hash = cipso_v4_map_cache_hash(cipso_ptr, cipso_ptr_len);
|
||||||
entry->lsm_data.free = secattr->cache.free;
|
atomic_inc(&secattr->cache->refcount);
|
||||||
entry->lsm_data.data = secattr->cache.data;
|
entry->lsm_data = secattr->cache;
|
||||||
|
|
||||||
bkt = entry->hash & (CIPSO_V4_CACHE_BUCKETBITS - 1);
|
bkt = entry->hash & (CIPSO_V4_CACHE_BUCKETBITS - 1);
|
||||||
spin_lock_bh(&cipso_v4_cache[bkt].lock);
|
spin_lock_bh(&cipso_v4_cache[bkt].lock);
|
||||||
|
|
|
@ -611,8 +611,8 @@ static int ipgre_rcv(struct sk_buff *skb)
|
||||||
* - When dealing with WCCPv2, Skip extra 4 bytes in GRE header
|
* - When dealing with WCCPv2, Skip extra 4 bytes in GRE header
|
||||||
*/
|
*/
|
||||||
if (flags == 0 &&
|
if (flags == 0 &&
|
||||||
skb->protocol == __constant_htons(ETH_P_WCCP)) {
|
skb->protocol == htons(ETH_P_WCCP)) {
|
||||||
skb->protocol = __constant_htons(ETH_P_IP);
|
skb->protocol = htons(ETH_P_IP);
|
||||||
if ((*(h + offset) & 0xF0) != 0x40)
|
if ((*(h + offset) & 0xF0) != 0x40)
|
||||||
offset += 4;
|
offset += 4;
|
||||||
}
|
}
|
||||||
|
|
|
@ -566,9 +566,15 @@ static inline u32 rt_score(struct rtable *rt)
|
||||||
|
|
||||||
static inline int compare_keys(struct flowi *fl1, struct flowi *fl2)
|
static inline int compare_keys(struct flowi *fl1, struct flowi *fl2)
|
||||||
{
|
{
|
||||||
return memcmp(&fl1->nl_u.ip4_u, &fl2->nl_u.ip4_u, sizeof(fl1->nl_u.ip4_u)) == 0 &&
|
return ((fl1->nl_u.ip4_u.daddr ^ fl2->nl_u.ip4_u.daddr) |
|
||||||
fl1->oif == fl2->oif &&
|
(fl1->nl_u.ip4_u.saddr ^ fl2->nl_u.ip4_u.saddr) |
|
||||||
fl1->iif == fl2->iif;
|
#ifdef CONFIG_IP_ROUTE_FWMARK
|
||||||
|
(fl1->nl_u.ip4_u.fwmark ^ fl2->nl_u.ip4_u.fwmark) |
|
||||||
|
#endif
|
||||||
|
(*(u16 *)&fl1->nl_u.ip4_u.tos ^
|
||||||
|
*(u16 *)&fl2->nl_u.ip4_u.tos) |
|
||||||
|
(fl1->oif ^ fl2->oif) |
|
||||||
|
(fl1->iif ^ fl2->iif)) == 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_IP_ROUTE_MULTIPATH_CACHED
|
#ifdef CONFIG_IP_ROUTE_MULTIPATH_CACHED
|
||||||
|
|
|
@ -355,7 +355,7 @@ void tcp_v4_err(struct sk_buff *skb, u32 info)
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
if (sk->sk_state == TCP_TIME_WAIT) {
|
if (sk->sk_state == TCP_TIME_WAIT) {
|
||||||
inet_twsk_put((struct inet_timewait_sock *)sk);
|
inet_twsk_put(inet_twsk(sk));
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -578,7 +578,7 @@ static void tcp_v4_send_ack(struct sk_buff *skb, u32 seq, u32 ack,
|
||||||
struct tcphdr *th = skb->h.th;
|
struct tcphdr *th = skb->h.th;
|
||||||
struct {
|
struct {
|
||||||
struct tcphdr th;
|
struct tcphdr th;
|
||||||
u32 tsopt[3];
|
u32 tsopt[TCPOLEN_TSTAMP_ALIGNED >> 2];
|
||||||
} rep;
|
} rep;
|
||||||
struct ip_reply_arg arg;
|
struct ip_reply_arg arg;
|
||||||
|
|
||||||
|
@ -960,7 +960,7 @@ static struct sock *tcp_v4_hnd_req(struct sock *sk, struct sk_buff *skb)
|
||||||
bh_lock_sock(nsk);
|
bh_lock_sock(nsk);
|
||||||
return nsk;
|
return nsk;
|
||||||
}
|
}
|
||||||
inet_twsk_put((struct inet_timewait_sock *)nsk);
|
inet_twsk_put(inet_twsk(nsk));
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1154,26 +1154,24 @@ discard_and_relse:
|
||||||
|
|
||||||
do_time_wait:
|
do_time_wait:
|
||||||
if (!xfrm4_policy_check(NULL, XFRM_POLICY_IN, skb)) {
|
if (!xfrm4_policy_check(NULL, XFRM_POLICY_IN, skb)) {
|
||||||
inet_twsk_put((struct inet_timewait_sock *) sk);
|
inet_twsk_put(inet_twsk(sk));
|
||||||
goto discard_it;
|
goto discard_it;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (skb->len < (th->doff << 2) || tcp_checksum_complete(skb)) {
|
if (skb->len < (th->doff << 2) || tcp_checksum_complete(skb)) {
|
||||||
TCP_INC_STATS_BH(TCP_MIB_INERRS);
|
TCP_INC_STATS_BH(TCP_MIB_INERRS);
|
||||||
inet_twsk_put((struct inet_timewait_sock *) sk);
|
inet_twsk_put(inet_twsk(sk));
|
||||||
goto discard_it;
|
goto discard_it;
|
||||||
}
|
}
|
||||||
switch (tcp_timewait_state_process((struct inet_timewait_sock *)sk,
|
switch (tcp_timewait_state_process(inet_twsk(sk), skb, th)) {
|
||||||
skb, th)) {
|
|
||||||
case TCP_TW_SYN: {
|
case TCP_TW_SYN: {
|
||||||
struct sock *sk2 = inet_lookup_listener(&tcp_hashinfo,
|
struct sock *sk2 = inet_lookup_listener(&tcp_hashinfo,
|
||||||
skb->nh.iph->daddr,
|
skb->nh.iph->daddr,
|
||||||
th->dest,
|
th->dest,
|
||||||
inet_iif(skb));
|
inet_iif(skb));
|
||||||
if (sk2) {
|
if (sk2) {
|
||||||
inet_twsk_deschedule((struct inet_timewait_sock *)sk,
|
inet_twsk_deschedule(inet_twsk(sk), &tcp_death_row);
|
||||||
&tcp_death_row);
|
inet_twsk_put(inet_twsk(sk));
|
||||||
inet_twsk_put((struct inet_timewait_sock *)sk);
|
|
||||||
sk = sk2;
|
sk = sk2;
|
||||||
goto process;
|
goto process;
|
||||||
}
|
}
|
||||||
|
|
|
@ -273,10 +273,10 @@ static void tcp_build_and_update_options(__be32 *ptr, struct tcp_sock *tp,
|
||||||
__u32 tstamp)
|
__u32 tstamp)
|
||||||
{
|
{
|
||||||
if (tp->rx_opt.tstamp_ok) {
|
if (tp->rx_opt.tstamp_ok) {
|
||||||
*ptr++ = __constant_htonl((TCPOPT_NOP << 24) |
|
*ptr++ = htonl((TCPOPT_NOP << 24) |
|
||||||
(TCPOPT_NOP << 16) |
|
(TCPOPT_NOP << 16) |
|
||||||
(TCPOPT_TIMESTAMP << 8) |
|
(TCPOPT_TIMESTAMP << 8) |
|
||||||
TCPOLEN_TIMESTAMP);
|
TCPOLEN_TIMESTAMP);
|
||||||
*ptr++ = htonl(tstamp);
|
*ptr++ = htonl(tstamp);
|
||||||
*ptr++ = htonl(tp->rx_opt.ts_recent);
|
*ptr++ = htonl(tp->rx_opt.ts_recent);
|
||||||
}
|
}
|
||||||
|
@ -325,18 +325,27 @@ static void tcp_syn_build_options(__be32 *ptr, int mss, int ts, int sack,
|
||||||
*ptr++ = htonl((TCPOPT_MSS << 24) | (TCPOLEN_MSS << 16) | mss);
|
*ptr++ = htonl((TCPOPT_MSS << 24) | (TCPOLEN_MSS << 16) | mss);
|
||||||
if (ts) {
|
if (ts) {
|
||||||
if(sack)
|
if(sack)
|
||||||
*ptr++ = __constant_htonl((TCPOPT_SACK_PERM << 24) | (TCPOLEN_SACK_PERM << 16) |
|
*ptr++ = htonl((TCPOPT_SACK_PERM << 24) |
|
||||||
(TCPOPT_TIMESTAMP << 8) | TCPOLEN_TIMESTAMP);
|
(TCPOLEN_SACK_PERM << 16) |
|
||||||
|
(TCPOPT_TIMESTAMP << 8) |
|
||||||
|
TCPOLEN_TIMESTAMP);
|
||||||
else
|
else
|
||||||
*ptr++ = __constant_htonl((TCPOPT_NOP << 24) | (TCPOPT_NOP << 16) |
|
*ptr++ = htonl((TCPOPT_NOP << 24) |
|
||||||
(TCPOPT_TIMESTAMP << 8) | TCPOLEN_TIMESTAMP);
|
(TCPOPT_NOP << 16) |
|
||||||
|
(TCPOPT_TIMESTAMP << 8) |
|
||||||
|
TCPOLEN_TIMESTAMP);
|
||||||
*ptr++ = htonl(tstamp); /* TSVAL */
|
*ptr++ = htonl(tstamp); /* TSVAL */
|
||||||
*ptr++ = htonl(ts_recent); /* TSECR */
|
*ptr++ = htonl(ts_recent); /* TSECR */
|
||||||
} else if(sack)
|
} else if(sack)
|
||||||
*ptr++ = __constant_htonl((TCPOPT_NOP << 24) | (TCPOPT_NOP << 16) |
|
*ptr++ = htonl((TCPOPT_NOP << 24) |
|
||||||
(TCPOPT_SACK_PERM << 8) | TCPOLEN_SACK_PERM);
|
(TCPOPT_NOP << 16) |
|
||||||
|
(TCPOPT_SACK_PERM << 8) |
|
||||||
|
TCPOLEN_SACK_PERM);
|
||||||
if (offer_wscale)
|
if (offer_wscale)
|
||||||
*ptr++ = htonl((TCPOPT_NOP << 24) | (TCPOPT_WINDOW << 16) | (TCPOLEN_WINDOW << 8) | (wscale));
|
*ptr++ = htonl((TCPOPT_NOP << 24) |
|
||||||
|
(TCPOPT_WINDOW << 16) |
|
||||||
|
(TCPOLEN_WINDOW << 8) |
|
||||||
|
(wscale));
|
||||||
}
|
}
|
||||||
|
|
||||||
/* This routine actually transmits TCP packets queued in by
|
/* This routine actually transmits TCP packets queued in by
|
||||||
|
|
|
@ -52,7 +52,7 @@ __xfrm4_find_bundle(struct flowi *fl, struct xfrm_policy *policy)
|
||||||
xdst->u.rt.fl.fl4_dst == fl->fl4_dst &&
|
xdst->u.rt.fl.fl4_dst == fl->fl4_dst &&
|
||||||
xdst->u.rt.fl.fl4_src == fl->fl4_src &&
|
xdst->u.rt.fl.fl4_src == fl->fl4_src &&
|
||||||
xdst->u.rt.fl.fl4_tos == fl->fl4_tos &&
|
xdst->u.rt.fl.fl4_tos == fl->fl4_tos &&
|
||||||
xfrm_bundle_ok(xdst, fl, AF_INET, 0)) {
|
xfrm_bundle_ok(policy, xdst, fl, AF_INET, 0)) {
|
||||||
dst_clone(dst);
|
dst_clone(dst);
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
|
|
|
@ -153,6 +153,19 @@ config INET6_XFRM_MODE_ROUTEOPTIMIZATION
|
||||||
---help---
|
---help---
|
||||||
Support for MIPv6 route optimization mode.
|
Support for MIPv6 route optimization mode.
|
||||||
|
|
||||||
|
config IPV6_SIT
|
||||||
|
tristate "IPv6: IPv6-in-IPv4 tunnel (SIT driver)"
|
||||||
|
depends on IPV6
|
||||||
|
default y
|
||||||
|
---help---
|
||||||
|
Tunneling means encapsulating data of one protocol type within
|
||||||
|
another protocol and sending it over a channel that understands the
|
||||||
|
encapsulating protocol. This driver implements encapsulation of IPv6
|
||||||
|
into IPv4 packets. This is useful if you want to connect two IPv6
|
||||||
|
networks over an IPv4-only path.
|
||||||
|
|
||||||
|
Saying M here will produce a module called sit.ko. If unsure, say Y.
|
||||||
|
|
||||||
config IPV6_TUNNEL
|
config IPV6_TUNNEL
|
||||||
tristate "IPv6: IPv6-in-IPv6 tunnel"
|
tristate "IPv6: IPv6-in-IPv6 tunnel"
|
||||||
select INET6_TUNNEL
|
select INET6_TUNNEL
|
||||||
|
|
|
@ -4,7 +4,7 @@
|
||||||
|
|
||||||
obj-$(CONFIG_IPV6) += ipv6.o
|
obj-$(CONFIG_IPV6) += ipv6.o
|
||||||
|
|
||||||
ipv6-objs := af_inet6.o anycast.o ip6_output.o ip6_input.o addrconf.o sit.o \
|
ipv6-objs := af_inet6.o anycast.o ip6_output.o ip6_input.o addrconf.o \
|
||||||
route.o ip6_fib.o ipv6_sockglue.o ndisc.o udp.o raw.o \
|
route.o ip6_fib.o ipv6_sockglue.o ndisc.o udp.o raw.o \
|
||||||
protocol.o icmp.o mcast.o reassembly.o tcp_ipv6.o \
|
protocol.o icmp.o mcast.o reassembly.o tcp_ipv6.o \
|
||||||
exthdrs.o sysctl_net_ipv6.o datagram.o proc.o \
|
exthdrs.o sysctl_net_ipv6.o datagram.o proc.o \
|
||||||
|
@ -29,6 +29,7 @@ obj-$(CONFIG_INET6_XFRM_MODE_ROUTEOPTIMIZATION) += xfrm6_mode_ro.o
|
||||||
obj-$(CONFIG_INET6_XFRM_MODE_BEET) += xfrm6_mode_beet.o
|
obj-$(CONFIG_INET6_XFRM_MODE_BEET) += xfrm6_mode_beet.o
|
||||||
obj-$(CONFIG_NETFILTER) += netfilter/
|
obj-$(CONFIG_NETFILTER) += netfilter/
|
||||||
|
|
||||||
|
obj-$(CONFIG_IPV6_SIT) += sit.o
|
||||||
obj-$(CONFIG_IPV6_TUNNEL) += ip6_tunnel.o
|
obj-$(CONFIG_IPV6_TUNNEL) += ip6_tunnel.o
|
||||||
|
|
||||||
obj-y += exthdrs_core.o
|
obj-y += exthdrs_core.o
|
||||||
|
|
|
@ -396,8 +396,10 @@ static struct inet6_dev * ipv6_add_dev(struct net_device *dev)
|
||||||
ndev->regen_timer.data = (unsigned long) ndev;
|
ndev->regen_timer.data = (unsigned long) ndev;
|
||||||
if ((dev->flags&IFF_LOOPBACK) ||
|
if ((dev->flags&IFF_LOOPBACK) ||
|
||||||
dev->type == ARPHRD_TUNNEL ||
|
dev->type == ARPHRD_TUNNEL ||
|
||||||
dev->type == ARPHRD_NONE ||
|
#if defined(CONFIG_IPV6_SIT) || defined(CONFIG_IPV6_SIT_MODULE)
|
||||||
dev->type == ARPHRD_SIT) {
|
dev->type == ARPHRD_SIT ||
|
||||||
|
#endif
|
||||||
|
dev->type == ARPHRD_NONE) {
|
||||||
printk(KERN_INFO
|
printk(KERN_INFO
|
||||||
"%s: Disabled Privacy Extensions\n",
|
"%s: Disabled Privacy Extensions\n",
|
||||||
dev->name);
|
dev->name);
|
||||||
|
@ -1546,8 +1548,10 @@ addrconf_prefix_route(struct in6_addr *pfx, int plen, struct net_device *dev,
|
||||||
This thing is done here expecting that the whole
|
This thing is done here expecting that the whole
|
||||||
class of non-broadcast devices need not cloning.
|
class of non-broadcast devices need not cloning.
|
||||||
*/
|
*/
|
||||||
|
#if defined(CONFIG_IPV6_SIT) || defined(CONFIG_IPV6_SIT_MODULE)
|
||||||
if (dev->type == ARPHRD_SIT && (dev->flags & IFF_POINTOPOINT))
|
if (dev->type == ARPHRD_SIT && (dev->flags & IFF_POINTOPOINT))
|
||||||
cfg.fc_flags |= RTF_NONEXTHOP;
|
cfg.fc_flags |= RTF_NONEXTHOP;
|
||||||
|
#endif
|
||||||
|
|
||||||
ip6_route_add(&cfg);
|
ip6_route_add(&cfg);
|
||||||
}
|
}
|
||||||
|
@ -1569,6 +1573,7 @@ static void addrconf_add_mroute(struct net_device *dev)
|
||||||
ip6_route_add(&cfg);
|
ip6_route_add(&cfg);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#if defined(CONFIG_IPV6_SIT) || defined(CONFIG_IPV6_SIT_MODULE)
|
||||||
static void sit_route_add(struct net_device *dev)
|
static void sit_route_add(struct net_device *dev)
|
||||||
{
|
{
|
||||||
struct fib6_config cfg = {
|
struct fib6_config cfg = {
|
||||||
|
@ -1582,6 +1587,7 @@ static void sit_route_add(struct net_device *dev)
|
||||||
/* prefix length - 96 bits "::d.d.d.d" */
|
/* prefix length - 96 bits "::d.d.d.d" */
|
||||||
ip6_route_add(&cfg);
|
ip6_route_add(&cfg);
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
static void addrconf_add_lroute(struct net_device *dev)
|
static void addrconf_add_lroute(struct net_device *dev)
|
||||||
{
|
{
|
||||||
|
@ -1852,6 +1858,7 @@ int addrconf_set_dstaddr(void __user *arg)
|
||||||
if (dev == NULL)
|
if (dev == NULL)
|
||||||
goto err_exit;
|
goto err_exit;
|
||||||
|
|
||||||
|
#if defined(CONFIG_IPV6_SIT) || defined(CONFIG_IPV6_SIT_MODULE)
|
||||||
if (dev->type == ARPHRD_SIT) {
|
if (dev->type == ARPHRD_SIT) {
|
||||||
struct ifreq ifr;
|
struct ifreq ifr;
|
||||||
mm_segment_t oldfs;
|
mm_segment_t oldfs;
|
||||||
|
@ -1881,6 +1888,7 @@ int addrconf_set_dstaddr(void __user *arg)
|
||||||
err = dev_open(dev);
|
err = dev_open(dev);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
err_exit:
|
err_exit:
|
||||||
rtnl_unlock();
|
rtnl_unlock();
|
||||||
|
@ -2010,6 +2018,7 @@ int addrconf_del_ifaddr(void __user *arg)
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#if defined(CONFIG_IPV6_SIT) || defined(CONFIG_IPV6_SIT_MODULE)
|
||||||
static void sit_add_v4_addrs(struct inet6_dev *idev)
|
static void sit_add_v4_addrs(struct inet6_dev *idev)
|
||||||
{
|
{
|
||||||
struct inet6_ifaddr * ifp;
|
struct inet6_ifaddr * ifp;
|
||||||
|
@ -2078,6 +2087,7 @@ static void sit_add_v4_addrs(struct inet6_dev *idev)
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
static void init_loopback(struct net_device *dev)
|
static void init_loopback(struct net_device *dev)
|
||||||
{
|
{
|
||||||
|
@ -2141,6 +2151,7 @@ static void addrconf_dev_config(struct net_device *dev)
|
||||||
addrconf_add_linklocal(idev, &addr);
|
addrconf_add_linklocal(idev, &addr);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
#if defined(CONFIG_IPV6_SIT) || defined(CONFIG_IPV6_SIT_MODULE)
|
||||||
static void addrconf_sit_config(struct net_device *dev)
|
static void addrconf_sit_config(struct net_device *dev)
|
||||||
{
|
{
|
||||||
struct inet6_dev *idev;
|
struct inet6_dev *idev;
|
||||||
|
@ -2166,6 +2177,7 @@ static void addrconf_sit_config(struct net_device *dev)
|
||||||
} else
|
} else
|
||||||
sit_route_add(dev);
|
sit_route_add(dev);
|
||||||
}
|
}
|
||||||
|
#endif
|
||||||
|
|
||||||
static inline int
|
static inline int
|
||||||
ipv6_inherit_linklocal(struct inet6_dev *idev, struct net_device *link_dev)
|
ipv6_inherit_linklocal(struct inet6_dev *idev, struct net_device *link_dev)
|
||||||
|
@ -2260,9 +2272,11 @@ static int addrconf_notify(struct notifier_block *this, unsigned long event,
|
||||||
}
|
}
|
||||||
|
|
||||||
switch(dev->type) {
|
switch(dev->type) {
|
||||||
|
#if defined(CONFIG_IPV6_SIT) || defined(CONFIG_IPV6_SIT_MODULE)
|
||||||
case ARPHRD_SIT:
|
case ARPHRD_SIT:
|
||||||
addrconf_sit_config(dev);
|
addrconf_sit_config(dev);
|
||||||
break;
|
break;
|
||||||
|
#endif
|
||||||
case ARPHRD_TUNNEL6:
|
case ARPHRD_TUNNEL6:
|
||||||
addrconf_ip6_tnl_config(dev);
|
addrconf_ip6_tnl_config(dev);
|
||||||
break;
|
break;
|
||||||
|
|
|
@ -850,7 +850,6 @@ static int __init inet6_init(void)
|
||||||
err = addrconf_init();
|
err = addrconf_init();
|
||||||
if (err)
|
if (err)
|
||||||
goto addrconf_fail;
|
goto addrconf_fail;
|
||||||
sit_init();
|
|
||||||
|
|
||||||
/* Init v6 extension headers. */
|
/* Init v6 extension headers. */
|
||||||
ipv6_rthdr_init();
|
ipv6_rthdr_init();
|
||||||
|
@ -927,7 +926,6 @@ static void __exit inet6_exit(void)
|
||||||
mip6_fini();
|
mip6_fini();
|
||||||
#endif
|
#endif
|
||||||
/* Cleanup code parts. */
|
/* Cleanup code parts. */
|
||||||
sit_cleanup();
|
|
||||||
ip6_flowlabel_cleanup();
|
ip6_flowlabel_cleanup();
|
||||||
addrconf_cleanup();
|
addrconf_cleanup();
|
||||||
ip6_route_cleanup();
|
ip6_route_cleanup();
|
||||||
|
|
|
@ -850,3 +850,6 @@ int __init sit_init(void)
|
||||||
inet_del_protocol(&sit_protocol, IPPROTO_IPV6);
|
inet_del_protocol(&sit_protocol, IPPROTO_IPV6);
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
module_init(sit_init);
|
||||||
|
module_exit(sit_cleanup);
|
||||||
|
|
|
@ -329,7 +329,7 @@ static void tcp_v6_err(struct sk_buff *skb, struct inet6_skb_parm *opt,
|
||||||
}
|
}
|
||||||
|
|
||||||
if (sk->sk_state == TCP_TIME_WAIT) {
|
if (sk->sk_state == TCP_TIME_WAIT) {
|
||||||
inet_twsk_put((struct inet_timewait_sock *)sk);
|
inet_twsk_put(inet_twsk(sk));
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -653,7 +653,7 @@ static void tcp_v6_send_ack(struct sk_buff *skb, u32 seq, u32 ack, u32 win, u32
|
||||||
int tot_len = sizeof(struct tcphdr);
|
int tot_len = sizeof(struct tcphdr);
|
||||||
|
|
||||||
if (ts)
|
if (ts)
|
||||||
tot_len += 3*4;
|
tot_len += TCPOLEN_TSTAMP_ALIGNED;
|
||||||
|
|
||||||
buff = alloc_skb(MAX_HEADER + sizeof(struct ipv6hdr) + tot_len,
|
buff = alloc_skb(MAX_HEADER + sizeof(struct ipv6hdr) + tot_len,
|
||||||
GFP_ATOMIC);
|
GFP_ATOMIC);
|
||||||
|
@ -749,7 +749,7 @@ static struct sock *tcp_v6_hnd_req(struct sock *sk,struct sk_buff *skb)
|
||||||
bh_lock_sock(nsk);
|
bh_lock_sock(nsk);
|
||||||
return nsk;
|
return nsk;
|
||||||
}
|
}
|
||||||
inet_twsk_put((struct inet_timewait_sock *)nsk);
|
inet_twsk_put(inet_twsk(nsk));
|
||||||
return NULL;
|
return NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1283,18 +1283,17 @@ discard_and_relse:
|
||||||
|
|
||||||
do_time_wait:
|
do_time_wait:
|
||||||
if (!xfrm6_policy_check(NULL, XFRM_POLICY_IN, skb)) {
|
if (!xfrm6_policy_check(NULL, XFRM_POLICY_IN, skb)) {
|
||||||
inet_twsk_put((struct inet_timewait_sock *)sk);
|
inet_twsk_put(inet_twsk(sk));
|
||||||
goto discard_it;
|
goto discard_it;
|
||||||
}
|
}
|
||||||
|
|
||||||
if (skb->len < (th->doff<<2) || tcp_checksum_complete(skb)) {
|
if (skb->len < (th->doff<<2) || tcp_checksum_complete(skb)) {
|
||||||
TCP_INC_STATS_BH(TCP_MIB_INERRS);
|
TCP_INC_STATS_BH(TCP_MIB_INERRS);
|
||||||
inet_twsk_put((struct inet_timewait_sock *)sk);
|
inet_twsk_put(inet_twsk(sk));
|
||||||
goto discard_it;
|
goto discard_it;
|
||||||
}
|
}
|
||||||
|
|
||||||
switch (tcp_timewait_state_process((struct inet_timewait_sock *)sk,
|
switch (tcp_timewait_state_process(inet_twsk(sk), skb, th)) {
|
||||||
skb, th)) {
|
|
||||||
case TCP_TW_SYN:
|
case TCP_TW_SYN:
|
||||||
{
|
{
|
||||||
struct sock *sk2;
|
struct sock *sk2;
|
||||||
|
|
|
@ -73,7 +73,7 @@ __xfrm6_find_bundle(struct flowi *fl, struct xfrm_policy *policy)
|
||||||
xdst->u.rt6.rt6i_src.plen);
|
xdst->u.rt6.rt6i_src.plen);
|
||||||
if (ipv6_addr_equal(&xdst->u.rt6.rt6i_dst.addr, &fl_dst_prefix) &&
|
if (ipv6_addr_equal(&xdst->u.rt6.rt6i_dst.addr, &fl_dst_prefix) &&
|
||||||
ipv6_addr_equal(&xdst->u.rt6.rt6i_src.addr, &fl_src_prefix) &&
|
ipv6_addr_equal(&xdst->u.rt6.rt6i_src.addr, &fl_src_prefix) &&
|
||||||
xfrm_bundle_ok(xdst, fl, AF_INET6,
|
xfrm_bundle_ok(policy, xdst, fl, AF_INET6,
|
||||||
(xdst->u.rt6.rt6i_dst.plen != 128 ||
|
(xdst->u.rt6.rt6i_dst.plen != 128 ||
|
||||||
xdst->u.rt6.rt6i_src.plen != 128))) {
|
xdst->u.rt6.rt6i_src.plen != 128))) {
|
||||||
dst_clone(dst);
|
dst_clone(dst);
|
||||||
|
|
|
@ -2928,11 +2928,6 @@ static struct xfrm_policy *pfkey_compile_policy(struct sock *sk, int opt,
|
||||||
if (*dir)
|
if (*dir)
|
||||||
goto out;
|
goto out;
|
||||||
}
|
}
|
||||||
else {
|
|
||||||
*dir = security_xfrm_sock_policy_alloc(xp, sk);
|
|
||||||
if (*dir)
|
|
||||||
goto out;
|
|
||||||
}
|
|
||||||
|
|
||||||
*dir = pol->sadb_x_policy_dir-1;
|
*dir = pol->sadb_x_policy_dir-1;
|
||||||
return xp;
|
return xp;
|
||||||
|
|
|
@ -200,7 +200,7 @@ void netlbl_cache_invalidate(void)
|
||||||
int netlbl_cache_add(const struct sk_buff *skb,
|
int netlbl_cache_add(const struct sk_buff *skb,
|
||||||
const struct netlbl_lsm_secattr *secattr)
|
const struct netlbl_lsm_secattr *secattr)
|
||||||
{
|
{
|
||||||
if (secattr->cache.data == NULL)
|
if (secattr->cache == NULL)
|
||||||
return -ENOMSG;
|
return -ENOMSG;
|
||||||
|
|
||||||
if (CIPSO_V4_OPTEXIST(skb))
|
if (CIPSO_V4_OPTEXIST(skb))
|
||||||
|
|
|
@ -786,11 +786,10 @@ static long htb_do_events(struct htb_sched *q, int level)
|
||||||
for (i = 0; i < 500; i++) {
|
for (i = 0; i < 500; i++) {
|
||||||
struct htb_class *cl;
|
struct htb_class *cl;
|
||||||
long diff;
|
long diff;
|
||||||
struct rb_node *p = q->wait_pq[level].rb_node;
|
struct rb_node *p = rb_first(&q->wait_pq[level]);
|
||||||
|
|
||||||
if (!p)
|
if (!p)
|
||||||
return 0;
|
return 0;
|
||||||
while (p->rb_left)
|
|
||||||
p = p->rb_left;
|
|
||||||
|
|
||||||
cl = rb_entry(p, struct htb_class, pq_node);
|
cl = rb_entry(p, struct htb_class, pq_node);
|
||||||
if (time_after(cl->pq_key, q->jiffies)) {
|
if (time_after(cl->pq_key, q->jiffies)) {
|
||||||
|
|
|
@ -344,7 +344,7 @@ static int sctp_assocs_seq_show(struct seq_file *seq, void *v)
|
||||||
assoc, sk, sctp_sk(sk)->type, sk->sk_state,
|
assoc, sk, sctp_sk(sk)->type, sk->sk_state,
|
||||||
assoc->state, hash, assoc->assoc_id,
|
assoc->state, hash, assoc->assoc_id,
|
||||||
assoc->sndbuf_used,
|
assoc->sndbuf_used,
|
||||||
(sk->sk_rcvbuf - assoc->rwnd),
|
atomic_read(&assoc->rmem_alloc),
|
||||||
sock_i_uid(sk), sock_i_ino(sk),
|
sock_i_uid(sk), sock_i_ino(sk),
|
||||||
epb->bind_addr.port,
|
epb->bind_addr.port,
|
||||||
assoc->peer.port);
|
assoc->peer.port);
|
||||||
|
|
|
@ -5362,6 +5362,20 @@ static void sctp_wfree(struct sk_buff *skb)
|
||||||
sctp_association_put(asoc);
|
sctp_association_put(asoc);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Do accounting for the receive space on the socket.
|
||||||
|
* Accounting for the association is done in ulpevent.c
|
||||||
|
* We set this as a destructor for the cloned data skbs so that
|
||||||
|
* accounting is done at the correct time.
|
||||||
|
*/
|
||||||
|
void sctp_sock_rfree(struct sk_buff *skb)
|
||||||
|
{
|
||||||
|
struct sock *sk = skb->sk;
|
||||||
|
struct sctp_ulpevent *event = sctp_skb2event(skb);
|
||||||
|
|
||||||
|
atomic_sub(event->rmem_len, &sk->sk_rmem_alloc);
|
||||||
|
}
|
||||||
|
|
||||||
|
|
||||||
/* Helper function to wait for space in the sndbuf. */
|
/* Helper function to wait for space in the sndbuf. */
|
||||||
static int sctp_wait_for_sndbuf(struct sctp_association *asoc, long *timeo_p,
|
static int sctp_wait_for_sndbuf(struct sctp_association *asoc, long *timeo_p,
|
||||||
size_t msg_len)
|
size_t msg_len)
|
||||||
|
@ -5634,10 +5648,10 @@ static void sctp_sock_migrate(struct sock *oldsk, struct sock *newsk,
|
||||||
sctp_skb_for_each(skb, &oldsk->sk_receive_queue, tmp) {
|
sctp_skb_for_each(skb, &oldsk->sk_receive_queue, tmp) {
|
||||||
event = sctp_skb2event(skb);
|
event = sctp_skb2event(skb);
|
||||||
if (event->asoc == assoc) {
|
if (event->asoc == assoc) {
|
||||||
sock_rfree(skb);
|
sctp_sock_rfree(skb);
|
||||||
__skb_unlink(skb, &oldsk->sk_receive_queue);
|
__skb_unlink(skb, &oldsk->sk_receive_queue);
|
||||||
__skb_queue_tail(&newsk->sk_receive_queue, skb);
|
__skb_queue_tail(&newsk->sk_receive_queue, skb);
|
||||||
skb_set_owner_r(skb, newsk);
|
sctp_skb_set_owner_r(skb, newsk);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -5665,10 +5679,10 @@ static void sctp_sock_migrate(struct sock *oldsk, struct sock *newsk,
|
||||||
sctp_skb_for_each(skb, &oldsp->pd_lobby, tmp) {
|
sctp_skb_for_each(skb, &oldsp->pd_lobby, tmp) {
|
||||||
event = sctp_skb2event(skb);
|
event = sctp_skb2event(skb);
|
||||||
if (event->asoc == assoc) {
|
if (event->asoc == assoc) {
|
||||||
sock_rfree(skb);
|
sctp_sock_rfree(skb);
|
||||||
__skb_unlink(skb, &oldsp->pd_lobby);
|
__skb_unlink(skb, &oldsp->pd_lobby);
|
||||||
__skb_queue_tail(queue, skb);
|
__skb_queue_tail(queue, skb);
|
||||||
skb_set_owner_r(skb, newsk);
|
sctp_skb_set_owner_r(skb, newsk);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -55,10 +55,13 @@ static void sctp_ulpevent_release_frag_data(struct sctp_ulpevent *event);
|
||||||
|
|
||||||
|
|
||||||
/* Initialize an ULP event from an given skb. */
|
/* Initialize an ULP event from an given skb. */
|
||||||
SCTP_STATIC void sctp_ulpevent_init(struct sctp_ulpevent *event, int msg_flags)
|
SCTP_STATIC void sctp_ulpevent_init(struct sctp_ulpevent *event,
|
||||||
|
int msg_flags,
|
||||||
|
unsigned int len)
|
||||||
{
|
{
|
||||||
memset(event, 0, sizeof(struct sctp_ulpevent));
|
memset(event, 0, sizeof(struct sctp_ulpevent));
|
||||||
event->msg_flags = msg_flags;
|
event->msg_flags = msg_flags;
|
||||||
|
event->rmem_len = len;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Create a new sctp_ulpevent. */
|
/* Create a new sctp_ulpevent. */
|
||||||
|
@ -73,7 +76,7 @@ SCTP_STATIC struct sctp_ulpevent *sctp_ulpevent_new(int size, int msg_flags,
|
||||||
goto fail;
|
goto fail;
|
||||||
|
|
||||||
event = sctp_skb2event(skb);
|
event = sctp_skb2event(skb);
|
||||||
sctp_ulpevent_init(event, msg_flags);
|
sctp_ulpevent_init(event, msg_flags, skb->truesize);
|
||||||
|
|
||||||
return event;
|
return event;
|
||||||
|
|
||||||
|
@ -101,17 +104,16 @@ static inline void sctp_ulpevent_set_owner(struct sctp_ulpevent *event,
|
||||||
sctp_association_hold((struct sctp_association *)asoc);
|
sctp_association_hold((struct sctp_association *)asoc);
|
||||||
skb = sctp_event2skb(event);
|
skb = sctp_event2skb(event);
|
||||||
event->asoc = (struct sctp_association *)asoc;
|
event->asoc = (struct sctp_association *)asoc;
|
||||||
atomic_add(skb->truesize, &event->asoc->rmem_alloc);
|
atomic_add(event->rmem_len, &event->asoc->rmem_alloc);
|
||||||
skb_set_owner_r(skb, asoc->base.sk);
|
sctp_skb_set_owner_r(skb, asoc->base.sk);
|
||||||
}
|
}
|
||||||
|
|
||||||
/* A simple destructor to give up the reference to the association. */
|
/* A simple destructor to give up the reference to the association. */
|
||||||
static inline void sctp_ulpevent_release_owner(struct sctp_ulpevent *event)
|
static inline void sctp_ulpevent_release_owner(struct sctp_ulpevent *event)
|
||||||
{
|
{
|
||||||
struct sctp_association *asoc = event->asoc;
|
struct sctp_association *asoc = event->asoc;
|
||||||
struct sk_buff *skb = sctp_event2skb(event);
|
|
||||||
|
|
||||||
atomic_sub(skb->truesize, &asoc->rmem_alloc);
|
atomic_sub(event->rmem_len, &asoc->rmem_alloc);
|
||||||
sctp_association_put(asoc);
|
sctp_association_put(asoc);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -372,7 +374,7 @@ struct sctp_ulpevent *sctp_ulpevent_make_remote_error(
|
||||||
|
|
||||||
/* Embed the event fields inside the cloned skb. */
|
/* Embed the event fields inside the cloned skb. */
|
||||||
event = sctp_skb2event(skb);
|
event = sctp_skb2event(skb);
|
||||||
sctp_ulpevent_init(event, MSG_NOTIFICATION);
|
sctp_ulpevent_init(event, MSG_NOTIFICATION, skb->truesize);
|
||||||
|
|
||||||
sre = (struct sctp_remote_error *)
|
sre = (struct sctp_remote_error *)
|
||||||
skb_push(skb, sizeof(struct sctp_remote_error));
|
skb_push(skb, sizeof(struct sctp_remote_error));
|
||||||
|
@ -464,7 +466,7 @@ struct sctp_ulpevent *sctp_ulpevent_make_send_failed(
|
||||||
|
|
||||||
/* Embed the event fields inside the cloned skb. */
|
/* Embed the event fields inside the cloned skb. */
|
||||||
event = sctp_skb2event(skb);
|
event = sctp_skb2event(skb);
|
||||||
sctp_ulpevent_init(event, MSG_NOTIFICATION);
|
sctp_ulpevent_init(event, MSG_NOTIFICATION, skb->truesize);
|
||||||
|
|
||||||
ssf = (struct sctp_send_failed *)
|
ssf = (struct sctp_send_failed *)
|
||||||
skb_push(skb, sizeof(struct sctp_send_failed));
|
skb_push(skb, sizeof(struct sctp_send_failed));
|
||||||
|
@ -682,8 +684,11 @@ struct sctp_ulpevent *sctp_ulpevent_make_rcvmsg(struct sctp_association *asoc,
|
||||||
/* Embed the event fields inside the cloned skb. */
|
/* Embed the event fields inside the cloned skb. */
|
||||||
event = sctp_skb2event(skb);
|
event = sctp_skb2event(skb);
|
||||||
|
|
||||||
/* Initialize event with flags 0. */
|
/* Initialize event with flags 0 and correct length
|
||||||
sctp_ulpevent_init(event, 0);
|
* Since this is a clone of the original skb, only account for
|
||||||
|
* the data of this chunk as other chunks will be accounted separately.
|
||||||
|
*/
|
||||||
|
sctp_ulpevent_init(event, 0, skb->len + sizeof(struct sk_buff));
|
||||||
|
|
||||||
sctp_ulpevent_receive_data(event, asoc);
|
sctp_ulpevent_receive_data(event, asoc);
|
||||||
|
|
||||||
|
|
|
@ -309,7 +309,7 @@ static struct sctp_ulpevent *sctp_make_reassembled_event(struct sk_buff_head *qu
|
||||||
if (!new)
|
if (!new)
|
||||||
return NULL; /* try again later */
|
return NULL; /* try again later */
|
||||||
|
|
||||||
new->sk = f_frag->sk;
|
sctp_skb_set_owner_r(new, f_frag->sk);
|
||||||
|
|
||||||
skb_shinfo(new)->frag_list = pos;
|
skb_shinfo(new)->frag_list = pos;
|
||||||
} else
|
} else
|
||||||
|
|
|
@ -883,30 +883,32 @@ out:
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(xfrm_policy_walk);
|
EXPORT_SYMBOL(xfrm_policy_walk);
|
||||||
|
|
||||||
/* Find policy to apply to this flow. */
|
/*
|
||||||
|
* Find policy to apply to this flow.
|
||||||
|
*
|
||||||
|
* Returns 0 if policy found, else an -errno.
|
||||||
|
*/
|
||||||
static int xfrm_policy_match(struct xfrm_policy *pol, struct flowi *fl,
|
static int xfrm_policy_match(struct xfrm_policy *pol, struct flowi *fl,
|
||||||
u8 type, u16 family, int dir)
|
u8 type, u16 family, int dir)
|
||||||
{
|
{
|
||||||
struct xfrm_selector *sel = &pol->selector;
|
struct xfrm_selector *sel = &pol->selector;
|
||||||
int match;
|
int match, ret = -ESRCH;
|
||||||
|
|
||||||
if (pol->family != family ||
|
if (pol->family != family ||
|
||||||
pol->type != type)
|
pol->type != type)
|
||||||
return 0;
|
return ret;
|
||||||
|
|
||||||
match = xfrm_selector_match(sel, fl, family);
|
match = xfrm_selector_match(sel, fl, family);
|
||||||
if (match) {
|
if (match)
|
||||||
if (!security_xfrm_policy_lookup(pol, fl->secid, dir))
|
ret = security_xfrm_policy_lookup(pol, fl->secid, dir);
|
||||||
return 1;
|
|
||||||
}
|
|
||||||
|
|
||||||
return 0;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct xfrm_policy *xfrm_policy_lookup_bytype(u8 type, struct flowi *fl,
|
static struct xfrm_policy *xfrm_policy_lookup_bytype(u8 type, struct flowi *fl,
|
||||||
u16 family, u8 dir)
|
u16 family, u8 dir)
|
||||||
{
|
{
|
||||||
|
int err;
|
||||||
struct xfrm_policy *pol, *ret;
|
struct xfrm_policy *pol, *ret;
|
||||||
xfrm_address_t *daddr, *saddr;
|
xfrm_address_t *daddr, *saddr;
|
||||||
struct hlist_node *entry;
|
struct hlist_node *entry;
|
||||||
|
@ -922,7 +924,15 @@ static struct xfrm_policy *xfrm_policy_lookup_bytype(u8 type, struct flowi *fl,
|
||||||
chain = policy_hash_direct(daddr, saddr, family, dir);
|
chain = policy_hash_direct(daddr, saddr, family, dir);
|
||||||
ret = NULL;
|
ret = NULL;
|
||||||
hlist_for_each_entry(pol, entry, chain, bydst) {
|
hlist_for_each_entry(pol, entry, chain, bydst) {
|
||||||
if (xfrm_policy_match(pol, fl, type, family, dir)) {
|
err = xfrm_policy_match(pol, fl, type, family, dir);
|
||||||
|
if (err) {
|
||||||
|
if (err == -ESRCH)
|
||||||
|
continue;
|
||||||
|
else {
|
||||||
|
ret = ERR_PTR(err);
|
||||||
|
goto fail;
|
||||||
|
}
|
||||||
|
} else {
|
||||||
ret = pol;
|
ret = pol;
|
||||||
priority = ret->priority;
|
priority = ret->priority;
|
||||||
break;
|
break;
|
||||||
|
@ -930,36 +940,53 @@ static struct xfrm_policy *xfrm_policy_lookup_bytype(u8 type, struct flowi *fl,
|
||||||
}
|
}
|
||||||
chain = &xfrm_policy_inexact[dir];
|
chain = &xfrm_policy_inexact[dir];
|
||||||
hlist_for_each_entry(pol, entry, chain, bydst) {
|
hlist_for_each_entry(pol, entry, chain, bydst) {
|
||||||
if (xfrm_policy_match(pol, fl, type, family, dir) &&
|
err = xfrm_policy_match(pol, fl, type, family, dir);
|
||||||
pol->priority < priority) {
|
if (err) {
|
||||||
|
if (err == -ESRCH)
|
||||||
|
continue;
|
||||||
|
else {
|
||||||
|
ret = ERR_PTR(err);
|
||||||
|
goto fail;
|
||||||
|
}
|
||||||
|
} else if (pol->priority < priority) {
|
||||||
ret = pol;
|
ret = pol;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
if (ret)
|
if (ret)
|
||||||
xfrm_pol_hold(ret);
|
xfrm_pol_hold(ret);
|
||||||
|
fail:
|
||||||
read_unlock_bh(&xfrm_policy_lock);
|
read_unlock_bh(&xfrm_policy_lock);
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void xfrm_policy_lookup(struct flowi *fl, u16 family, u8 dir,
|
static int xfrm_policy_lookup(struct flowi *fl, u16 family, u8 dir,
|
||||||
void **objp, atomic_t **obj_refp)
|
void **objp, atomic_t **obj_refp)
|
||||||
{
|
{
|
||||||
struct xfrm_policy *pol;
|
struct xfrm_policy *pol;
|
||||||
|
int err = 0;
|
||||||
|
|
||||||
#ifdef CONFIG_XFRM_SUB_POLICY
|
#ifdef CONFIG_XFRM_SUB_POLICY
|
||||||
pol = xfrm_policy_lookup_bytype(XFRM_POLICY_TYPE_SUB, fl, family, dir);
|
pol = xfrm_policy_lookup_bytype(XFRM_POLICY_TYPE_SUB, fl, family, dir);
|
||||||
if (pol)
|
if (IS_ERR(pol)) {
|
||||||
|
err = PTR_ERR(pol);
|
||||||
|
pol = NULL;
|
||||||
|
}
|
||||||
|
if (pol || err)
|
||||||
goto end;
|
goto end;
|
||||||
#endif
|
#endif
|
||||||
pol = xfrm_policy_lookup_bytype(XFRM_POLICY_TYPE_MAIN, fl, family, dir);
|
pol = xfrm_policy_lookup_bytype(XFRM_POLICY_TYPE_MAIN, fl, family, dir);
|
||||||
|
if (IS_ERR(pol)) {
|
||||||
|
err = PTR_ERR(pol);
|
||||||
|
pol = NULL;
|
||||||
|
}
|
||||||
#ifdef CONFIG_XFRM_SUB_POLICY
|
#ifdef CONFIG_XFRM_SUB_POLICY
|
||||||
end:
|
end:
|
||||||
#endif
|
#endif
|
||||||
if ((*objp = (void *) pol) != NULL)
|
if ((*objp = (void *) pol) != NULL)
|
||||||
*obj_refp = &pol->refcnt;
|
*obj_refp = &pol->refcnt;
|
||||||
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline int policy_to_flow_dir(int dir)
|
static inline int policy_to_flow_dir(int dir)
|
||||||
|
@ -989,12 +1016,16 @@ static struct xfrm_policy *xfrm_sk_policy_lookup(struct sock *sk, int dir, struc
|
||||||
sk->sk_family);
|
sk->sk_family);
|
||||||
int err = 0;
|
int err = 0;
|
||||||
|
|
||||||
if (match)
|
if (match) {
|
||||||
err = security_xfrm_policy_lookup(pol, fl->secid, policy_to_flow_dir(dir));
|
err = security_xfrm_policy_lookup(pol, fl->secid,
|
||||||
|
policy_to_flow_dir(dir));
|
||||||
if (match && !err)
|
if (!err)
|
||||||
xfrm_pol_hold(pol);
|
xfrm_pol_hold(pol);
|
||||||
else
|
else if (err == -ESRCH)
|
||||||
|
pol = NULL;
|
||||||
|
else
|
||||||
|
pol = ERR_PTR(err);
|
||||||
|
} else
|
||||||
pol = NULL;
|
pol = NULL;
|
||||||
}
|
}
|
||||||
read_unlock_bh(&xfrm_policy_lock);
|
read_unlock_bh(&xfrm_policy_lock);
|
||||||
|
@ -1286,8 +1317,11 @@ restart:
|
||||||
pol_dead = 0;
|
pol_dead = 0;
|
||||||
xfrm_nr = 0;
|
xfrm_nr = 0;
|
||||||
|
|
||||||
if (sk && sk->sk_policy[1])
|
if (sk && sk->sk_policy[1]) {
|
||||||
policy = xfrm_sk_policy_lookup(sk, XFRM_POLICY_OUT, fl);
|
policy = xfrm_sk_policy_lookup(sk, XFRM_POLICY_OUT, fl);
|
||||||
|
if (IS_ERR(policy))
|
||||||
|
return PTR_ERR(policy);
|
||||||
|
}
|
||||||
|
|
||||||
if (!policy) {
|
if (!policy) {
|
||||||
/* To accelerate a bit... */
|
/* To accelerate a bit... */
|
||||||
|
@ -1297,6 +1331,8 @@ restart:
|
||||||
|
|
||||||
policy = flow_cache_lookup(fl, dst_orig->ops->family,
|
policy = flow_cache_lookup(fl, dst_orig->ops->family,
|
||||||
dir, xfrm_policy_lookup);
|
dir, xfrm_policy_lookup);
|
||||||
|
if (IS_ERR(policy))
|
||||||
|
return PTR_ERR(policy);
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!policy)
|
if (!policy)
|
||||||
|
@ -1343,6 +1379,10 @@ restart:
|
||||||
fl, family,
|
fl, family,
|
||||||
XFRM_POLICY_OUT);
|
XFRM_POLICY_OUT);
|
||||||
if (pols[1]) {
|
if (pols[1]) {
|
||||||
|
if (IS_ERR(pols[1])) {
|
||||||
|
err = PTR_ERR(pols[1]);
|
||||||
|
goto error;
|
||||||
|
}
|
||||||
if (pols[1]->action == XFRM_POLICY_BLOCK) {
|
if (pols[1]->action == XFRM_POLICY_BLOCK) {
|
||||||
err = -EPERM;
|
err = -EPERM;
|
||||||
goto error;
|
goto error;
|
||||||
|
@ -1574,13 +1614,19 @@ int __xfrm_policy_check(struct sock *sk, int dir, struct sk_buff *skb,
|
||||||
}
|
}
|
||||||
|
|
||||||
pol = NULL;
|
pol = NULL;
|
||||||
if (sk && sk->sk_policy[dir])
|
if (sk && sk->sk_policy[dir]) {
|
||||||
pol = xfrm_sk_policy_lookup(sk, dir, &fl);
|
pol = xfrm_sk_policy_lookup(sk, dir, &fl);
|
||||||
|
if (IS_ERR(pol))
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
if (!pol)
|
if (!pol)
|
||||||
pol = flow_cache_lookup(&fl, family, fl_dir,
|
pol = flow_cache_lookup(&fl, family, fl_dir,
|
||||||
xfrm_policy_lookup);
|
xfrm_policy_lookup);
|
||||||
|
|
||||||
|
if (IS_ERR(pol))
|
||||||
|
return 0;
|
||||||
|
|
||||||
if (!pol) {
|
if (!pol) {
|
||||||
if (skb->sp && secpath_has_nontransport(skb->sp, 0, &xerr_idx)) {
|
if (skb->sp && secpath_has_nontransport(skb->sp, 0, &xerr_idx)) {
|
||||||
xfrm_secpath_reject(xerr_idx, skb, &fl);
|
xfrm_secpath_reject(xerr_idx, skb, &fl);
|
||||||
|
@ -1599,6 +1645,8 @@ int __xfrm_policy_check(struct sock *sk, int dir, struct sk_buff *skb,
|
||||||
&fl, family,
|
&fl, family,
|
||||||
XFRM_POLICY_IN);
|
XFRM_POLICY_IN);
|
||||||
if (pols[1]) {
|
if (pols[1]) {
|
||||||
|
if (IS_ERR(pols[1]))
|
||||||
|
return 0;
|
||||||
pols[1]->curlft.use_time = (unsigned long)xtime.tv_sec;
|
pols[1]->curlft.use_time = (unsigned long)xtime.tv_sec;
|
||||||
npols ++;
|
npols ++;
|
||||||
}
|
}
|
||||||
|
@ -1706,7 +1754,7 @@ static struct dst_entry *xfrm_dst_check(struct dst_entry *dst, u32 cookie)
|
||||||
|
|
||||||
static int stale_bundle(struct dst_entry *dst)
|
static int stale_bundle(struct dst_entry *dst)
|
||||||
{
|
{
|
||||||
return !xfrm_bundle_ok((struct xfrm_dst *)dst, NULL, AF_UNSPEC, 0);
|
return !xfrm_bundle_ok(NULL, (struct xfrm_dst *)dst, NULL, AF_UNSPEC, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
void xfrm_dst_ifdown(struct dst_entry *dst, struct net_device *dev)
|
void xfrm_dst_ifdown(struct dst_entry *dst, struct net_device *dev)
|
||||||
|
@ -1828,7 +1876,8 @@ EXPORT_SYMBOL(xfrm_init_pmtu);
|
||||||
* still valid.
|
* still valid.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
int xfrm_bundle_ok(struct xfrm_dst *first, struct flowi *fl, int family, int strict)
|
int xfrm_bundle_ok(struct xfrm_policy *pol, struct xfrm_dst *first,
|
||||||
|
struct flowi *fl, int family, int strict)
|
||||||
{
|
{
|
||||||
struct dst_entry *dst = &first->u.dst;
|
struct dst_entry *dst = &first->u.dst;
|
||||||
struct xfrm_dst *last;
|
struct xfrm_dst *last;
|
||||||
|
@ -1845,7 +1894,7 @@ int xfrm_bundle_ok(struct xfrm_dst *first, struct flowi *fl, int family, int str
|
||||||
|
|
||||||
if (fl && !xfrm_selector_match(&dst->xfrm->sel, fl, family))
|
if (fl && !xfrm_selector_match(&dst->xfrm->sel, fl, family))
|
||||||
return 0;
|
return 0;
|
||||||
if (fl && !security_xfrm_flow_state_match(fl, dst->xfrm))
|
if (fl && !security_xfrm_flow_state_match(fl, dst->xfrm, pol))
|
||||||
return 0;
|
return 0;
|
||||||
if (dst->xfrm->km.state != XFRM_STATE_VALID)
|
if (dst->xfrm->km.state != XFRM_STATE_VALID)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
|
@ -1992,15 +1992,6 @@ static struct xfrm_policy *xfrm_compile_policy(struct sock *sk, int opt,
|
||||||
xp->type = XFRM_POLICY_TYPE_MAIN;
|
xp->type = XFRM_POLICY_TYPE_MAIN;
|
||||||
copy_templates(xp, ut, nr);
|
copy_templates(xp, ut, nr);
|
||||||
|
|
||||||
if (!xp->security) {
|
|
||||||
int err = security_xfrm_sock_policy_alloc(xp, sk);
|
|
||||||
if (err) {
|
|
||||||
kfree(xp);
|
|
||||||
*dir = err;
|
|
||||||
return NULL;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
*dir = p->dir;
|
*dir = p->dir;
|
||||||
|
|
||||||
return xp;
|
return xp;
|
||||||
|
|
|
@ -881,7 +881,8 @@ static int dummy_xfrm_state_pol_flow_match(struct xfrm_state *x,
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int dummy_xfrm_flow_state_match(struct flowi *fl, struct xfrm_state *xfrm)
|
static int dummy_xfrm_flow_state_match(struct flowi *fl, struct xfrm_state *xfrm,
|
||||||
|
struct xfrm_policy *xp)
|
||||||
{
|
{
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
|
@ -19,7 +19,8 @@ int selinux_xfrm_state_delete(struct xfrm_state *x);
|
||||||
int selinux_xfrm_policy_lookup(struct xfrm_policy *xp, u32 fl_secid, u8 dir);
|
int selinux_xfrm_policy_lookup(struct xfrm_policy *xp, u32 fl_secid, u8 dir);
|
||||||
int selinux_xfrm_state_pol_flow_match(struct xfrm_state *x,
|
int selinux_xfrm_state_pol_flow_match(struct xfrm_state *x,
|
||||||
struct xfrm_policy *xp, struct flowi *fl);
|
struct xfrm_policy *xp, struct flowi *fl);
|
||||||
int selinux_xfrm_flow_state_match(struct flowi *fl, struct xfrm_state *xfrm);
|
int selinux_xfrm_flow_state_match(struct flowi *fl, struct xfrm_state *xfrm,
|
||||||
|
struct xfrm_policy *xp);
|
||||||
|
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
|
|
@ -618,6 +618,7 @@ void policydb_destroy(struct policydb *p)
|
||||||
c = c->next;
|
c = c->next;
|
||||||
ocontext_destroy(ctmp,i);
|
ocontext_destroy(ctmp,i);
|
||||||
}
|
}
|
||||||
|
p->ocontexts[i] = NULL;
|
||||||
}
|
}
|
||||||
|
|
||||||
g = p->genfs;
|
g = p->genfs;
|
||||||
|
@ -633,6 +634,7 @@ void policydb_destroy(struct policydb *p)
|
||||||
g = g->next;
|
g = g->next;
|
||||||
kfree(gtmp);
|
kfree(gtmp);
|
||||||
}
|
}
|
||||||
|
p->genfs = NULL;
|
||||||
|
|
||||||
cond_policydb_destroy(p);
|
cond_policydb_destroy(p);
|
||||||
|
|
||||||
|
|
|
@ -2172,7 +2172,12 @@ struct netlbl_cache {
|
||||||
*/
|
*/
|
||||||
static void selinux_netlbl_cache_free(const void *data)
|
static void selinux_netlbl_cache_free(const void *data)
|
||||||
{
|
{
|
||||||
struct netlbl_cache *cache = NETLBL_CACHE(data);
|
struct netlbl_cache *cache;
|
||||||
|
|
||||||
|
if (data == NULL)
|
||||||
|
return;
|
||||||
|
|
||||||
|
cache = NETLBL_CACHE(data);
|
||||||
switch (cache->type) {
|
switch (cache->type) {
|
||||||
case NETLBL_CACHE_T_MLS:
|
case NETLBL_CACHE_T_MLS:
|
||||||
ebitmap_destroy(&cache->data.mls_label.level[0].cat);
|
ebitmap_destroy(&cache->data.mls_label.level[0].cat);
|
||||||
|
@ -2197,17 +2202,20 @@ static void selinux_netlbl_cache_add(struct sk_buff *skb, struct context *ctx)
|
||||||
struct netlbl_lsm_secattr secattr;
|
struct netlbl_lsm_secattr secattr;
|
||||||
|
|
||||||
netlbl_secattr_init(&secattr);
|
netlbl_secattr_init(&secattr);
|
||||||
|
secattr.cache = netlbl_secattr_cache_alloc(GFP_ATOMIC);
|
||||||
|
if (secattr.cache == NULL)
|
||||||
|
goto netlbl_cache_add_return;
|
||||||
|
|
||||||
cache = kzalloc(sizeof(*cache), GFP_ATOMIC);
|
cache = kzalloc(sizeof(*cache), GFP_ATOMIC);
|
||||||
if (cache == NULL)
|
if (cache == NULL)
|
||||||
goto netlbl_cache_add_failure;
|
goto netlbl_cache_add_return;
|
||||||
secattr.cache.free = selinux_netlbl_cache_free;
|
secattr.cache->free = selinux_netlbl_cache_free;
|
||||||
secattr.cache.data = (void *)cache;
|
secattr.cache->data = (void *)cache;
|
||||||
|
|
||||||
cache->type = NETLBL_CACHE_T_MLS;
|
cache->type = NETLBL_CACHE_T_MLS;
|
||||||
if (ebitmap_cpy(&cache->data.mls_label.level[0].cat,
|
if (ebitmap_cpy(&cache->data.mls_label.level[0].cat,
|
||||||
&ctx->range.level[0].cat) != 0)
|
&ctx->range.level[0].cat) != 0)
|
||||||
goto netlbl_cache_add_failure;
|
goto netlbl_cache_add_return;
|
||||||
cache->data.mls_label.level[1].cat.highbit =
|
cache->data.mls_label.level[1].cat.highbit =
|
||||||
cache->data.mls_label.level[0].cat.highbit;
|
cache->data.mls_label.level[0].cat.highbit;
|
||||||
cache->data.mls_label.level[1].cat.node =
|
cache->data.mls_label.level[1].cat.node =
|
||||||
|
@ -2215,13 +2223,10 @@ static void selinux_netlbl_cache_add(struct sk_buff *skb, struct context *ctx)
|
||||||
cache->data.mls_label.level[0].sens = ctx->range.level[0].sens;
|
cache->data.mls_label.level[0].sens = ctx->range.level[0].sens;
|
||||||
cache->data.mls_label.level[1].sens = ctx->range.level[0].sens;
|
cache->data.mls_label.level[1].sens = ctx->range.level[0].sens;
|
||||||
|
|
||||||
if (netlbl_cache_add(skb, &secattr) != 0)
|
netlbl_cache_add(skb, &secattr);
|
||||||
goto netlbl_cache_add_failure;
|
|
||||||
|
|
||||||
return;
|
netlbl_cache_add_return:
|
||||||
|
netlbl_secattr_destroy(&secattr);
|
||||||
netlbl_cache_add_failure:
|
|
||||||
netlbl_secattr_destroy(&secattr, 1);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
|
@ -2263,8 +2268,8 @@ static int selinux_netlbl_secattr_to_sid(struct sk_buff *skb,
|
||||||
|
|
||||||
POLICY_RDLOCK;
|
POLICY_RDLOCK;
|
||||||
|
|
||||||
if (secattr->cache.data) {
|
if (secattr->cache) {
|
||||||
cache = NETLBL_CACHE(secattr->cache.data);
|
cache = NETLBL_CACHE(secattr->cache->data);
|
||||||
switch (cache->type) {
|
switch (cache->type) {
|
||||||
case NETLBL_CACHE_T_SID:
|
case NETLBL_CACHE_T_SID:
|
||||||
*sid = cache->data.sid;
|
*sid = cache->data.sid;
|
||||||
|
@ -2331,7 +2336,7 @@ static int selinux_netlbl_secattr_to_sid(struct sk_buff *skb,
|
||||||
selinux_netlbl_cache_add(skb, &ctx_new);
|
selinux_netlbl_cache_add(skb, &ctx_new);
|
||||||
ebitmap_destroy(&ctx_new.range.level[0].cat);
|
ebitmap_destroy(&ctx_new.range.level[0].cat);
|
||||||
} else {
|
} else {
|
||||||
*sid = SECINITSID_UNLABELED;
|
*sid = SECSID_NULL;
|
||||||
rc = 0;
|
rc = 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2369,7 +2374,7 @@ static int selinux_netlbl_skbuff_getsid(struct sk_buff *skb,
|
||||||
&secattr,
|
&secattr,
|
||||||
base_sid,
|
base_sid,
|
||||||
sid);
|
sid);
|
||||||
netlbl_secattr_destroy(&secattr, 0);
|
netlbl_secattr_destroy(&secattr);
|
||||||
|
|
||||||
return rc;
|
return rc;
|
||||||
}
|
}
|
||||||
|
@ -2415,7 +2420,7 @@ static int selinux_netlbl_socket_setsid(struct socket *sock, u32 sid)
|
||||||
if (rc == 0)
|
if (rc == 0)
|
||||||
sksec->nlbl_state = NLBL_LABELED;
|
sksec->nlbl_state = NLBL_LABELED;
|
||||||
|
|
||||||
netlbl_secattr_destroy(&secattr, 0);
|
netlbl_secattr_destroy(&secattr);
|
||||||
|
|
||||||
netlbl_socket_setsid_return:
|
netlbl_socket_setsid_return:
|
||||||
POLICY_RDUNLOCK;
|
POLICY_RDUNLOCK;
|
||||||
|
@ -2514,10 +2519,10 @@ void selinux_netlbl_sock_graft(struct sock *sk, struct socket *sock)
|
||||||
if (netlbl_sock_getattr(sk, &secattr) == 0 &&
|
if (netlbl_sock_getattr(sk, &secattr) == 0 &&
|
||||||
selinux_netlbl_secattr_to_sid(NULL,
|
selinux_netlbl_secattr_to_sid(NULL,
|
||||||
&secattr,
|
&secattr,
|
||||||
sksec->sid,
|
SECINITSID_UNLABELED,
|
||||||
&nlbl_peer_sid) == 0)
|
&nlbl_peer_sid) == 0)
|
||||||
sksec->peer_sid = nlbl_peer_sid;
|
sksec->peer_sid = nlbl_peer_sid;
|
||||||
netlbl_secattr_destroy(&secattr, 0);
|
netlbl_secattr_destroy(&secattr);
|
||||||
|
|
||||||
sksec->nlbl_state = NLBL_REQUIRE;
|
sksec->nlbl_state = NLBL_REQUIRE;
|
||||||
|
|
||||||
|
@ -2547,9 +2552,6 @@ u32 selinux_netlbl_inet_conn_request(struct sk_buff *skb, u32 sock_sid)
|
||||||
if (rc != 0)
|
if (rc != 0)
|
||||||
return SECSID_NULL;
|
return SECSID_NULL;
|
||||||
|
|
||||||
if (peer_sid == SECINITSID_UNLABELED)
|
|
||||||
return SECSID_NULL;
|
|
||||||
|
|
||||||
return peer_sid;
|
return peer_sid;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2611,11 +2613,13 @@ int selinux_netlbl_sock_rcv_skb(struct sk_security_struct *sksec,
|
||||||
u32 netlbl_sid;
|
u32 netlbl_sid;
|
||||||
u32 recv_perm;
|
u32 recv_perm;
|
||||||
|
|
||||||
rc = selinux_netlbl_skbuff_getsid(skb, SECINITSID_NETMSG, &netlbl_sid);
|
rc = selinux_netlbl_skbuff_getsid(skb,
|
||||||
|
SECINITSID_UNLABELED,
|
||||||
|
&netlbl_sid);
|
||||||
if (rc != 0)
|
if (rc != 0)
|
||||||
return rc;
|
return rc;
|
||||||
|
|
||||||
if (netlbl_sid == SECINITSID_UNLABELED)
|
if (netlbl_sid == SECSID_NULL)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
switch (sksec->sclass) {
|
switch (sksec->sclass) {
|
||||||
|
@ -2653,10 +2657,6 @@ int selinux_netlbl_sock_rcv_skb(struct sk_security_struct *sksec,
|
||||||
u32 selinux_netlbl_socket_getpeersec_stream(struct socket *sock)
|
u32 selinux_netlbl_socket_getpeersec_stream(struct socket *sock)
|
||||||
{
|
{
|
||||||
struct sk_security_struct *sksec = sock->sk->sk_security;
|
struct sk_security_struct *sksec = sock->sk->sk_security;
|
||||||
|
|
||||||
if (sksec->peer_sid == SECINITSID_UNLABELED)
|
|
||||||
return SECSID_NULL;
|
|
||||||
|
|
||||||
return sksec->peer_sid;
|
return sksec->peer_sid;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2672,16 +2672,10 @@ u32 selinux_netlbl_socket_getpeersec_stream(struct socket *sock)
|
||||||
u32 selinux_netlbl_socket_getpeersec_dgram(struct sk_buff *skb)
|
u32 selinux_netlbl_socket_getpeersec_dgram(struct sk_buff *skb)
|
||||||
{
|
{
|
||||||
int peer_sid;
|
int peer_sid;
|
||||||
struct sock *sk = skb->sk;
|
|
||||||
struct inode_security_struct *isec;
|
|
||||||
|
|
||||||
if (sk == NULL || sk->sk_socket == NULL)
|
if (selinux_netlbl_skbuff_getsid(skb,
|
||||||
return SECSID_NULL;
|
SECINITSID_UNLABELED,
|
||||||
|
&peer_sid) != 0)
|
||||||
isec = SOCK_INODE(sk->sk_socket)->i_security;
|
|
||||||
if (selinux_netlbl_skbuff_getsid(skb, isec->sid, &peer_sid) != 0)
|
|
||||||
return SECSID_NULL;
|
|
||||||
if (peer_sid == SECINITSID_UNLABELED)
|
|
||||||
return SECSID_NULL;
|
return SECSID_NULL;
|
||||||
|
|
||||||
return peer_sid;
|
return peer_sid;
|
||||||
|
|
|
@ -77,8 +77,8 @@ static inline int selinux_authorizable_xfrm(struct xfrm_state *x)
|
||||||
*/
|
*/
|
||||||
int selinux_xfrm_policy_lookup(struct xfrm_policy *xp, u32 fl_secid, u8 dir)
|
int selinux_xfrm_policy_lookup(struct xfrm_policy *xp, u32 fl_secid, u8 dir)
|
||||||
{
|
{
|
||||||
int rc = 0;
|
int rc;
|
||||||
u32 sel_sid = SECINITSID_UNLABELED;
|
u32 sel_sid;
|
||||||
struct xfrm_sec_ctx *ctx;
|
struct xfrm_sec_ctx *ctx;
|
||||||
|
|
||||||
/* Context sid is either set to label or ANY_ASSOC */
|
/* Context sid is either set to label or ANY_ASSOC */
|
||||||
|
@ -88,11 +88,21 @@ int selinux_xfrm_policy_lookup(struct xfrm_policy *xp, u32 fl_secid, u8 dir)
|
||||||
|
|
||||||
sel_sid = ctx->ctx_sid;
|
sel_sid = ctx->ctx_sid;
|
||||||
}
|
}
|
||||||
|
else
|
||||||
|
/*
|
||||||
|
* All flows should be treated as polmatch'ing an
|
||||||
|
* otherwise applicable "non-labeled" policy. This
|
||||||
|
* would prevent inadvertent "leaks".
|
||||||
|
*/
|
||||||
|
return 0;
|
||||||
|
|
||||||
rc = avc_has_perm(fl_secid, sel_sid, SECCLASS_ASSOCIATION,
|
rc = avc_has_perm(fl_secid, sel_sid, SECCLASS_ASSOCIATION,
|
||||||
ASSOCIATION__POLMATCH,
|
ASSOCIATION__POLMATCH,
|
||||||
NULL);
|
NULL);
|
||||||
|
|
||||||
|
if (rc == -EACCES)
|
||||||
|
rc = -ESRCH;
|
||||||
|
|
||||||
return rc;
|
return rc;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -108,15 +118,20 @@ int selinux_xfrm_state_pol_flow_match(struct xfrm_state *x, struct xfrm_policy *
|
||||||
u32 pol_sid;
|
u32 pol_sid;
|
||||||
int err;
|
int err;
|
||||||
|
|
||||||
if (x->security)
|
if (xp->security) {
|
||||||
state_sid = x->security->ctx_sid;
|
if (!x->security)
|
||||||
else
|
/* unlabeled SA and labeled policy can't match */
|
||||||
state_sid = SECINITSID_UNLABELED;
|
return 0;
|
||||||
|
else
|
||||||
if (xp->security)
|
state_sid = x->security->ctx_sid;
|
||||||
pol_sid = xp->security->ctx_sid;
|
pol_sid = xp->security->ctx_sid;
|
||||||
else
|
} else
|
||||||
pol_sid = SECINITSID_UNLABELED;
|
if (x->security)
|
||||||
|
/* unlabeled policy and labeled SA can't match */
|
||||||
|
return 0;
|
||||||
|
else
|
||||||
|
/* unlabeled policy and unlabeled SA match all flows */
|
||||||
|
return 1;
|
||||||
|
|
||||||
err = avc_has_perm(state_sid, pol_sid, SECCLASS_ASSOCIATION,
|
err = avc_has_perm(state_sid, pol_sid, SECCLASS_ASSOCIATION,
|
||||||
ASSOCIATION__POLMATCH,
|
ASSOCIATION__POLMATCH,
|
||||||
|
@ -125,7 +140,11 @@ int selinux_xfrm_state_pol_flow_match(struct xfrm_state *x, struct xfrm_policy *
|
||||||
if (err)
|
if (err)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
return selinux_xfrm_flow_state_match(fl, x);
|
err = avc_has_perm(fl->secid, state_sid, SECCLASS_ASSOCIATION,
|
||||||
|
ASSOCIATION__SENDTO,
|
||||||
|
NULL)? 0:1;
|
||||||
|
|
||||||
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -133,12 +152,22 @@ int selinux_xfrm_state_pol_flow_match(struct xfrm_state *x, struct xfrm_policy *
|
||||||
* can use a given security association.
|
* can use a given security association.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
int selinux_xfrm_flow_state_match(struct flowi *fl, struct xfrm_state *xfrm)
|
int selinux_xfrm_flow_state_match(struct flowi *fl, struct xfrm_state *xfrm,
|
||||||
|
struct xfrm_policy *xp)
|
||||||
{
|
{
|
||||||
int rc = 0;
|
int rc = 0;
|
||||||
u32 sel_sid = SECINITSID_UNLABELED;
|
u32 sel_sid = SECINITSID_UNLABELED;
|
||||||
struct xfrm_sec_ctx *ctx;
|
struct xfrm_sec_ctx *ctx;
|
||||||
|
|
||||||
|
if (!xp->security)
|
||||||
|
if (!xfrm->security)
|
||||||
|
return 1;
|
||||||
|
else
|
||||||
|
return 0;
|
||||||
|
else
|
||||||
|
if (!xfrm->security)
|
||||||
|
return 0;
|
||||||
|
|
||||||
/* Context sid is either set to label or ANY_ASSOC */
|
/* Context sid is either set to label or ANY_ASSOC */
|
||||||
if ((ctx = xfrm->security)) {
|
if ((ctx = xfrm->security)) {
|
||||||
if (!selinux_authorizable_ctx(ctx))
|
if (!selinux_authorizable_ctx(ctx))
|
||||||
|
|
Loading…
Add table
Reference in a new issue