264 lines
		
	
	
		
			8.3 KiB
		
	
	
	
		
			Diff
		
	
	
	
	
	
			
		
		
	
	
			264 lines
		
	
	
		
			8.3 KiB
		
	
	
	
		
			Diff
		
	
	
	
	
	
| --- a/ath/if_ath.c
 | |
| +++ b/ath/if_ath.c
 | |
| @@ -3239,7 +3239,6 @@ ath_hardstart(struct sk_buff *skb, struc
 | |
|  	struct ath_softc *sc = dev->priv;
 | |
|  	struct ieee80211_node *ni = NULL;
 | |
|  	struct ath_buf *bf = NULL;
 | |
| -	struct ether_header *eh;
 | |
|  	ath_bufhead bf_head;
 | |
|  	struct ath_buf *tbf, *tempbf;
 | |
|  	struct sk_buff *tskb;
 | |
| @@ -3251,6 +3250,7 @@ ath_hardstart(struct sk_buff *skb, struc
 | |
|  	*/
 | |
|  	int requeue = 0;
 | |
|  #ifdef ATH_SUPERG_FF
 | |
| +	struct ether_header *eh;
 | |
|  	unsigned int pktlen;
 | |
|  	struct ieee80211com *ic = &sc->sc_ic;
 | |
|  	struct ath_node *an;
 | |
| @@ -3316,27 +3316,9 @@ ath_hardstart(struct sk_buff *skb, struc
 | |
|  		requeue = 1;
 | |
|  		goto hardstart_fail;
 | |
|  	}
 | |
| -#endif
 | |
|  
 | |
| -	/* If the skb data is shared, we will copy it so we can strip padding
 | |
| -	 * without affecting any other bridge ports. */
 | |
| -	if (skb_cloned(skb)) {
 | |
| -		/* Remember the original SKB so we can free up our references */
 | |
| -		struct sk_buff *skb_new;
 | |
| -		skb_new = skb_copy(skb, GFP_ATOMIC);
 | |
| -		if (skb_new == NULL) {
 | |
| -			DPRINTF(sc, ATH_DEBUG_XMIT,
 | |
| -				"Dropping; skb_copy failure.\n");
 | |
| -			/* No free RAM, do not requeue! */
 | |
| -			goto hardstart_fail;
 | |
| -		}
 | |
| -		ieee80211_skb_copy_noderef(skb, skb_new);
 | |
| -		ieee80211_dev_kfree_skb(&skb);
 | |
| -		skb = skb_new;
 | |
| -	}
 | |
|  	eh = (struct ether_header *)skb->data;
 | |
|  
 | |
| -#ifdef ATH_SUPERG_FF
 | |
|  	/* NB: use this lock to protect an->an_tx_ffbuf (and txq->axq_stageq)
 | |
|  	 *     in athff_can_aggregate() call too. */
 | |
|  	ATH_TXQ_LOCK_IRQ(txq);
 | |
| --- a/net80211/ieee80211_output.c
 | |
| +++ b/net80211/ieee80211_output.c
 | |
| @@ -283,7 +283,7 @@ ieee80211_hardstart(struct sk_buff *skb,
 | |
|  	 * normal vap. */
 | |
|  	if (vap->iv_xrvap && (ni == vap->iv_bss) &&
 | |
|  	    vap->iv_xrvap->iv_sta_assoc) {
 | |
| -		struct sk_buff *skb1 = skb_copy(skb, GFP_ATOMIC);
 | |
| +		struct sk_buff *skb1 = skb_clone(skb, GFP_ATOMIC);
 | |
|  		if (skb1) {
 | |
|  			memset(SKB_CB(skb1), 0, sizeof(struct ieee80211_cb));
 | |
|  #ifdef IEEE80211_DEBUG_REFCNT
 | |
| @@ -566,7 +566,7 @@ ieee80211_skbhdr_adjust(struct ieee80211
 | |
|  	struct ieee80211_key *key, struct sk_buff *skb, int ismulticast)
 | |
|  {
 | |
|  	/* XXX pre-calculate per node? */
 | |
| -	int need_headroom = LLC_SNAPFRAMELEN + hdrsize + IEEE80211_ADDR_LEN;
 | |
| +	int need_headroom = LLC_SNAPFRAMELEN + hdrsize;
 | |
|  	int need_tailroom = 0;
 | |
|  #ifdef ATH_SUPERG_FF
 | |
|  	int isff = ATH_FF_MAGIC_PRESENT(skb);
 | |
| @@ -608,109 +608,56 @@ ieee80211_skbhdr_adjust(struct ieee80211
 | |
|  				need_tailroom += cip->ic_miclen;
 | |
|  	}
 | |
|  
 | |
| -	if (skb_shared(skb)) {
 | |
| -		/* Take our own reference to the node in the clone */
 | |
| -		ieee80211_ref_node(SKB_CB(skb)->ni);
 | |
| -		/* Unshare the node, decrementing users in the old skb */
 | |
| -		skb = skb_unshare(skb, GFP_ATOMIC);
 | |
| +	need_headroom -= skb_headroom(skb);
 | |
| +	if (isff)
 | |
| +		need_tailroom -= skb_tailroom(skb2);
 | |
| +	else
 | |
| +		need_tailroom -= skb_tailroom(skb);
 | |
| +
 | |
| +	if (need_headroom < 0)
 | |
| +		need_headroom = 0;
 | |
| +	if (need_tailroom < 0)
 | |
| +		need_tailroom = 0;
 | |
| +
 | |
| +	if (skb_cloned(skb) || (need_headroom > 0) ||
 | |
| +		(!isff && (need_tailroom > 0))) {
 | |
| +
 | |
| +		if (pskb_expand_head(skb, need_headroom, need_tailroom, GFP_ATOMIC)) {
 | |
| +			IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
 | |
| +				"%s: cannot expand storage (tail)\n", __func__);
 | |
| +			goto error;
 | |
| +		}
 | |
|  	}
 | |
|  
 | |
|  #ifdef ATH_SUPERG_FF
 | |
|  	if (isff) {
 | |
| -		if (skb == NULL) {
 | |
| -			IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
 | |
| -				"%s: cannot unshare for encapsulation\n",
 | |
| -				__func__);
 | |
| -			vap->iv_stats.is_tx_nobuf++;
 | |
| -			ieee80211_dev_kfree_skb(&skb2);
 | |
| -
 | |
| -			return NULL;
 | |
| -		}
 | |
| +		inter_headroom -= skb_headroom(skb2);
 | |
| +		if (inter_headroom < 0)
 | |
| +			inter_headroom = 0;
 | |
| +		if ((skb_cloned(skb2) ||
 | |
| +			(inter_headroom > 0) || (need_tailroom > 0))) {
 | |
|  
 | |
| -		/* first skb header */
 | |
| -		if (skb_headroom(skb) < need_headroom) {
 | |
| -			struct sk_buff *tmp = skb;
 | |
| -			skb = skb_realloc_headroom(skb, need_headroom);
 | |
| -			if (skb == NULL) {
 | |
| +			if (pskb_expand_head(skb2, inter_headroom,
 | |
| +				need_tailroom, GFP_ATOMIC)) {
 | |
|  				IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
 | |
| -					"%s: cannot expand storage (head1)\n",
 | |
| -					__func__);
 | |
| -				vap->iv_stats.is_tx_nobuf++;
 | |
| -				ieee80211_dev_kfree_skb(&skb2);
 | |
| -				return NULL;
 | |
| -			} else
 | |
| -				ieee80211_skb_copy_noderef(tmp, skb);
 | |
| -			ieee80211_dev_kfree_skb(&tmp);
 | |
| -			/* NB: cb[] area was copied, but not next ptr. must do that
 | |
| -			 *     prior to return on success. */
 | |
| -		}
 | |
| -
 | |
| -		/* second skb with header and tail adjustments possible */
 | |
| -		if (skb_tailroom(skb2) < need_tailroom) {
 | |
| -			int n = 0;
 | |
| -			if (inter_headroom > skb_headroom(skb2))
 | |
| -				n = inter_headroom - skb_headroom(skb2);
 | |
| -			if (pskb_expand_head(skb2, n,
 | |
| -			    need_tailroom - skb_tailroom(skb2), GFP_ATOMIC)) {
 | |
| -				ieee80211_dev_kfree_skb(&skb2);
 | |
| -				IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
 | |
| -					"%s: cannot expand storage (tail2)\n",
 | |
| -					__func__);
 | |
| -				vap->iv_stats.is_tx_nobuf++;
 | |
| -				/* this shouldn't happen, but don't send first ff either */
 | |
| -				ieee80211_dev_kfree_skb(&skb);
 | |
| +					"%s: cannot expand storage (tail)\n", __func__);
 | |
| +				goto error;
 | |
|  			}
 | |
| -		} else if (skb_headroom(skb2) < inter_headroom) {
 | |
| -			struct sk_buff *tmp = skb2;
 | |
| -
 | |
| -			skb2 = skb_realloc_headroom(skb2, inter_headroom);
 | |
| -			if (skb2 == NULL) {
 | |
| -				IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
 | |
| -					"%s: cannot expand storage (head2)\n",
 | |
| -					__func__);
 | |
| -				vap->iv_stats.is_tx_nobuf++;
 | |
| -				/* this shouldn't happen, but don't send first ff either */
 | |
| -				ieee80211_dev_kfree_skb(&skb);
 | |
| -				skb = NULL;
 | |
| -			} else
 | |
| -				ieee80211_skb_copy_noderef(tmp, skb);
 | |
| -			ieee80211_dev_kfree_skb(&tmp);
 | |
|  		}
 | |
| -		if (skb) {
 | |
| -			skb->next = skb2;
 | |
| -		}
 | |
| -		return skb;
 | |
| +		skb->next = skb2;
 | |
|  	}
 | |
|  #endif /* ATH_SUPERG_FF */
 | |
| -	if (skb == NULL) {
 | |
| -		IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
 | |
| -			"%s: cannot unshare for encapsulation\n", __func__);
 | |
| -		vap->iv_stats.is_tx_nobuf++;
 | |
| -	} else if (skb_tailroom(skb) < need_tailroom) {
 | |
| -		int n = 0;
 | |
| -		if (need_headroom > skb_headroom(skb))
 | |
| -			n = need_headroom - skb_headroom(skb);
 | |
| -		if (pskb_expand_head(skb, n, need_tailroom - 
 | |
| -					skb_tailroom(skb), GFP_ATOMIC)) {
 | |
| -			IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
 | |
| -				"%s: cannot expand storage (tail)\n", __func__);
 | |
| -			vap->iv_stats.is_tx_nobuf++;
 | |
| -			ieee80211_dev_kfree_skb(&skb);
 | |
| -		}
 | |
| -	} else if (skb_headroom(skb) < need_headroom) {
 | |
| -		struct sk_buff *tmp = skb;
 | |
| -		skb = skb_realloc_headroom(skb, need_headroom);
 | |
| -		/* Increment reference count after copy */
 | |
| -		if (skb == NULL) {
 | |
| -			IEEE80211_DPRINTF(vap, IEEE80211_MSG_OUTPUT,
 | |
| -				"%s: cannot expand storage (head)\n", __func__);
 | |
| -			vap->iv_stats.is_tx_nobuf++;
 | |
| -		} else
 | |
| -			ieee80211_skb_copy_noderef(tmp, skb);
 | |
| -		ieee80211_dev_kfree_skb(&tmp);
 | |
| -	}
 | |
|  
 | |
|  	return skb;
 | |
| +
 | |
| +error:
 | |
| +	vap->iv_stats.is_tx_nobuf++;
 | |
| +	ieee80211_dev_kfree_skb(&skb);
 | |
| +#ifdef ATH_SUPERG_FF
 | |
| +	if (skb2)
 | |
| +		ieee80211_dev_kfree_skb(&skb2);
 | |
| +#endif
 | |
| +	return NULL;
 | |
|  }
 | |
|  
 | |
|  #define	KEY_UNDEFINED(k)	((k).wk_cipher == &ieee80211_cipher_none)
 | |
| --- a/net80211/ieee80211_input.c
 | |
| +++ b/net80211/ieee80211_input.c
 | |
| @@ -204,7 +204,6 @@ ieee80211_input(struct ieee80211vap * va
 | |
|  	struct ieee80211_frame *wh;
 | |
|  	struct ieee80211_key *key;
 | |
|  	struct ether_header *eh;
 | |
| -	struct sk_buff *skb2;
 | |
|  #ifdef ATH_SUPERG_FF
 | |
|  	struct llc *llc;
 | |
|  #endif
 | |
| @@ -244,20 +243,6 @@ ieee80211_input(struct ieee80211vap * va
 | |
|  		vap->iv_stats.is_rx_tooshort++;
 | |
|  		goto out;
 | |
|  	}
 | |
| -	/* Clone the SKB... we assume somewhere in this driver that we 'own'
 | |
| -	 * the skbuff passed into hard start and we do a lot of messing with it
 | |
| -	 * but bridges under some cases will not clone for the first pass of skb
 | |
| -	 * to a bridge port, but will then clone for subsequent ones.  This is 
 | |
| -	 * odd behavior but it means that if we have trashed the skb we are given
 | |
| -	 * then other ports get clones of the residual garbage.
 | |
| -	 */
 | |
| -	if ((skb2 = skb_copy(skb, GFP_ATOMIC)) == NULL) {
 | |
| -		vap->iv_devstats.tx_dropped++;
 | |
| -		goto out;
 | |
| -	}
 | |
| -	ieee80211_skb_copy_noderef(skb, skb2);
 | |
| -	ieee80211_dev_kfree_skb(&skb);
 | |
| -	skb = skb2;
 | |
|  
 | |
|  	/*
 | |
|  	 * Bit of a cheat here, we use a pointer for a 3-address
 | |
| @@ -738,7 +723,7 @@ ieee80211_input(struct ieee80211vap * va
 | |
|  			/* ether_type must be length as FF frames are always LLC/SNAP encap'd */ 
 | |
|  			frame_len = ntohs(eh_tmp->ether_type); 
 | |
|  
 | |
| -			skb1 = skb_copy(skb, GFP_ATOMIC);
 | |
| +			skb1 = skb_clone(skb, GFP_ATOMIC);
 | |
|  			if (skb1 == NULL)
 | |
|  				goto err;
 | |
|  			ieee80211_skb_copy_noderef(skb, skb1);
 | |
| @@ -1137,7 +1122,7 @@ ieee80211_deliver_data(struct ieee80211_
 | |
|  
 | |
|  		if (ETHER_IS_MULTICAST(eh->ether_dhost) && !netif_queue_stopped(dev)) {
 | |
|  			/* Create a SKB for the BSS to send out. */
 | |
| -			skb1 = skb_copy(skb, GFP_ATOMIC);
 | |
| +			skb1 = skb_clone(skb, GFP_ATOMIC);
 | |
|  			if (skb1)
 | |
|  				SKB_CB(skb1)->ni = ieee80211_ref_node(vap->iv_bss); 
 | |
|  		}
 | 
