]> pilppa.org Git - linux-2.6-omap-h63xx.git/blobdiff - net/dccp/output.c
mfd: twl4030-gpio driver
[linux-2.6-omap-h63xx.git] / net / dccp / output.c
index 9afd58e39e239136cbb926af965a90925e1f735a..809d803d5006ec6614821b9d0d48b63cf131a31f 100644 (file)
@@ -161,27 +161,21 @@ unsigned int dccp_sync_mss(struct sock *sk, u32 pmtu)
        struct inet_connection_sock *icsk = inet_csk(sk);
        struct dccp_sock *dp = dccp_sk(sk);
        u32 ccmps = dccp_determine_ccmps(dp);
-       u32 cur_mps = ccmps ? min(pmtu, ccmps) : pmtu;
+       int cur_mps = ccmps ? min(pmtu, ccmps) : pmtu;
 
        /* Account for header lengths and IPv4/v6 option overhead */
        cur_mps -= (icsk->icsk_af_ops->net_header_len + icsk->icsk_ext_hdr_len +
                    sizeof(struct dccp_hdr) + sizeof(struct dccp_hdr_ext));
 
        /*
-        * Leave enough headroom for common DCCP header options.
-        * This only considers options which may appear on DCCP-Data packets, as
-        * per table 3 in RFC 4340, 5.8. When running out of space for other
-        * options (eg. Ack Vector which can take up to 255 bytes), it is better
-        * to schedule a separate Ack. Thus we leave headroom for the following:
-        *  - 1 byte for Slow Receiver (11.6)
-        *  - 6 bytes for Timestamp (13.1)
-        *  - 10 bytes for Timestamp Echo (13.3)
-        *  - 8 bytes for NDP count (7.7, when activated)
-        *  - 6 bytes for Data Checksum (9.3)
-        *  - %DCCPAV_MIN_OPTLEN bytes for Ack Vector size (11.4, when enabled)
+        * FIXME: this should come from the CCID infrastructure, where, say,
+        * TFRC will say it wants TIMESTAMPS, ELAPSED time, etc, for now lets
+        * put a rough estimate for NDP + TIMESTAMP + TIMESTAMP_ECHO + ELAPSED
+        * TIME + TFRC_OPT_LOSS_EVENT_RATE + TFRC_OPT_RECEIVE_RATE + padding to
+        * make it a multiple of 4
         */
-       cur_mps -= roundup(1 + 6 + 10 + dp->dccps_send_ndp_count * 8 + 6 +
-                          (dp->dccps_hc_rx_ackvec ? DCCPAV_MIN_OPTLEN : 0), 4);
+
+       cur_mps -= ((5 + 6 + 10 + 6 + 6 + 6 + 3) / 4) * 4;
 
        /* And store cached results */
        icsk->icsk_pmtu_cookie = pmtu;
@@ -251,98 +245,50 @@ do_interrupted:
        goto out;
 }
 
-/**
- * dccp_xmit_packet  -  Send data packet under control of CCID
- * Transmits next-queued payload and informs CCID to account for the packet.
- */
-static void dccp_xmit_packet(struct sock *sk)
-{
-       int err, len;
-       struct dccp_sock *dp = dccp_sk(sk);
-       struct sk_buff *skb = skb_dequeue(&sk->sk_write_queue);
-
-       if (unlikely(skb == NULL))
-               return;
-       len = skb->len;
-
-       if (sk->sk_state == DCCP_PARTOPEN) {
-               const u32 cur_mps = dp->dccps_mss_cache - DCCP_FEATNEG_OVERHEAD;
-               /*
-                * See 8.1.5 - Handshake Completion.
-                *
-                * For robustness we resend Confirm options until the client has
-                * entered OPEN. During the initial feature negotiation, the MPS
-                * is smaller than usual, reduced by the Change/Confirm options.
-                */
-               if (!list_empty(&dp->dccps_featneg) && len > cur_mps) {
-                       DCCP_WARN("Payload too large (%d) for featneg.\n", len);
-                       dccp_send_ack(sk);
-                       dccp_feat_list_purge(&dp->dccps_featneg);
-               }
-
-               inet_csk_schedule_ack(sk);
-               inet_csk_reset_xmit_timer(sk, ICSK_TIME_DACK,
-                                             inet_csk(sk)->icsk_rto,
-                                             DCCP_RTO_MAX);
-               DCCP_SKB_CB(skb)->dccpd_type = DCCP_PKT_DATAACK;
-       } else if (dccp_ack_pending(sk)) {
-               DCCP_SKB_CB(skb)->dccpd_type = DCCP_PKT_DATAACK;
-       } else {
-               DCCP_SKB_CB(skb)->dccpd_type = DCCP_PKT_DATA;
-       }
-
-       err = dccp_transmit_skb(sk, skb);
-       if (err)
-               dccp_pr_debug("transmit_skb() returned err=%d\n", err);
-       /*
-        * Register this one as sent even if an error occurred. To the remote
-        * end a local packet drop is indistinguishable from network loss, i.e.
-        * any local drop will eventually be reported via receiver feedback.
-        */
-       ccid_hc_tx_packet_sent(dp->dccps_hc_tx_ccid, sk, len);
-
-       /*
-        * If the CCID needs to transfer additional header options out-of-band
-        * (e.g. Ack Vectors or feature-negotiation options), it activates this
-        * flag to schedule a Sync. The Sync will automatically incorporate all
-        * currently pending header options, thus clearing the backlog.
-        */
-       if (dp->dccps_sync_scheduled)
-               dccp_send_sync(sk, dp->dccps_gsr, DCCP_PKT_SYNC);
-}
-
 void dccp_write_xmit(struct sock *sk, int block)
 {
        struct dccp_sock *dp = dccp_sk(sk);
        struct sk_buff *skb;
 
        while ((skb = skb_peek(&sk->sk_write_queue))) {
-               int rc = ccid_hc_tx_send_packet(dp->dccps_hc_tx_ccid, sk, skb);
+               int err = ccid_hc_tx_send_packet(dp->dccps_hc_tx_ccid, sk, skb);
 
-               switch (ccid_packet_dequeue_eval(rc)) {
-               case CCID_PACKET_WILL_DEQUEUE_LATER:
-                       return;
-               case CCID_PACKET_DELAY:
+               if (err > 0) {
                        if (!block) {
                                sk_reset_timer(sk, &dp->dccps_xmit_timer,
-                                               msecs_to_jiffies(rc)+jiffies);
-                               return;
-                       }
-                       rc = dccp_wait_for_ccid(sk, skb, rc);
-                       if (rc && rc != -EINTR) {
-                               DCCP_BUG("err=%d after dccp_wait_for_ccid", rc);
-                               skb_dequeue(&sk->sk_write_queue);
-                               kfree_skb(skb);
+                                               msecs_to_jiffies(err)+jiffies);
                                break;
-                       }
-                       /* fall through */
-               case CCID_PACKET_SEND_AT_ONCE:
-                       dccp_xmit_packet(sk);
-                       break;
-               case CCID_PACKET_ERR:
-                       skb_dequeue(&sk->sk_write_queue);
+                       } else
+                               err = dccp_wait_for_ccid(sk, skb, err);
+                       if (err && err != -EINTR)
+                               DCCP_BUG("err=%d after dccp_wait_for_ccid", err);
+               }
+
+               skb_dequeue(&sk->sk_write_queue);
+               if (err == 0) {
+                       struct dccp_skb_cb *dcb = DCCP_SKB_CB(skb);
+                       const int len = skb->len;
+
+                       if (sk->sk_state == DCCP_PARTOPEN) {
+                               /* See 8.1.5.  Handshake Completion */
+                               inet_csk_schedule_ack(sk);
+                               inet_csk_reset_xmit_timer(sk, ICSK_TIME_DACK,
+                                                 inet_csk(sk)->icsk_rto,
+                                                 DCCP_RTO_MAX);
+                               dcb->dccpd_type = DCCP_PKT_DATAACK;
+                       } else if (dccp_ack_pending(sk))
+                               dcb->dccpd_type = DCCP_PKT_DATAACK;
+                       else
+                               dcb->dccpd_type = DCCP_PKT_DATA;
+
+                       err = dccp_transmit_skb(sk, skb);
+                       ccid_hc_tx_packet_sent(dp->dccps_hc_tx_ccid, sk, 0, len);
+                       if (err)
+                               DCCP_BUG("err=%d after ccid_hc_tx_packet_sent",
+                                        err);
+               } else {
+                       dccp_pr_debug("packet discarded due to err=%d\n", err);
                        kfree_skb(skb);
-                       dccp_pr_debug("packet discarded due to err=%d\n", rc);
                }
        }
 }
@@ -393,17 +339,15 @@ struct sk_buff *dccp_make_response(struct sock *sk, struct dst_entry *dst,
        DCCP_SKB_CB(skb)->dccpd_type = DCCP_PKT_RESPONSE;
        DCCP_SKB_CB(skb)->dccpd_seq  = dreq->dreq_iss;
 
-       /* Resolve feature dependencies resulting from choice of CCID */
-       if (dccp_feat_server_ccid_dependencies(dreq))
-               goto response_failed;
-
-       if (dccp_insert_options_rsk(dreq, skb))
-               goto response_failed;
+       if (dccp_insert_options_rsk(dreq, skb)) {
+               kfree_skb(skb);
+               return NULL;
+       }
 
        /* Build and checksum header */
        dh = dccp_zeroed_hdr(skb, dccp_header_size);
 
-       dh->dccph_sport = inet_sk(sk)->sport;
+       dh->dccph_sport = inet_rsk(req)->loc_port;
        dh->dccph_dport = inet_rsk(req)->rmt_port;
        dh->dccph_doff  = (dccp_header_size +
                           DCCP_SKB_CB(skb)->dccpd_opt_len) / 4;
@@ -419,9 +363,6 @@ struct sk_buff *dccp_make_response(struct sock *sk, struct dst_entry *dst,
        inet_rsk(req)->acked = 1;
        DCCP_INC_STATS(DCCP_MIB_OUTSEGS);
        return skb;
-response_failed:
-       kfree_skb(skb);
-       return NULL;
 }
 
 EXPORT_SYMBOL_GPL(dccp_make_response);
@@ -506,9 +447,8 @@ int dccp_send_reset(struct sock *sk, enum dccp_reset_codes code)
 /*
  * Do all connect socket setups that can be done AF independent.
  */
-int dccp_connect(struct sock *sk)
+static inline void dccp_connect_init(struct sock *sk)
 {
-       struct sk_buff *skb;
        struct dccp_sock *dp = dccp_sk(sk);
        struct dst_entry *dst = __sk_dst_get(sk);
        struct inet_connection_sock *icsk = inet_csk(sk);
@@ -518,13 +458,19 @@ int dccp_connect(struct sock *sk)
 
        dccp_sync_mss(sk, dst_mtu(dst));
 
-       /* do not connect if feature negotiation setup fails */
-       if (dccp_feat_finalise_settings(dccp_sk(sk)))
-               return -EPROTO;
-
        /* Initialise GAR as per 8.5; AWL/AWH are set in dccp_transmit_skb() */
        dp->dccps_gar = dp->dccps_iss;
 
+       icsk->icsk_retransmits = 0;
+}
+
+int dccp_connect(struct sock *sk)
+{
+       struct sk_buff *skb;
+       struct inet_connection_sock *icsk = inet_csk(sk);
+
+       dccp_connect_init(sk);
+
        skb = alloc_skb(sk->sk_prot->max_header, sk->sk_allocation);
        if (unlikely(skb == NULL))
                return -ENOBUFS;
@@ -539,7 +485,6 @@ int dccp_connect(struct sock *sk)
        DCCP_INC_STATS(DCCP_MIB_ACTIVEOPENS);
 
        /* Timer for repeating the REQUEST until an answer. */
-       icsk->icsk_retransmits = 0;
        inet_csk_reset_xmit_timer(sk, ICSK_TIME_RETRANS,
                                  icsk->icsk_rto, DCCP_RTO_MAX);
        return 0;
@@ -626,12 +571,6 @@ void dccp_send_sync(struct sock *sk, const u64 ackno,
        DCCP_SKB_CB(skb)->dccpd_type = pkt_type;
        DCCP_SKB_CB(skb)->dccpd_ack_seq = ackno;
 
-       /*
-        * Clear the flag in case the Sync was scheduled for out-of-band data,
-        * such as carrying a long Ack Vector.
-        */
-       dccp_sk(sk)->dccps_sync_scheduled = 0;
-
        dccp_transmit_skb(sk, skb);
 }