Chapter III: Transport Layer

Similar documents
Chapter 3- parte B outline

Chapter III: Transport Layer

Chapter 3 outline. 3.5 connection-oriented transport: TCP segment structure reliable data transfer flow control connection management

TCP reliable data transfer. Chapter 3 outline. TCP sender events: TCP sender (simplified) TCP: retransmission scenarios. TCP: retransmission scenarios

CS 4390 Computer Networks. Pointers to Corresponding Section of Textbook

TCP: Overview RFCs: 793,1122,1323, 2018, 2581

Computer Networking Introduction

TCP (Part 2) Session 10 INST 346 Technologies, Infrastructure and Architecture

Chapter 3 Transport Layer

Chapter 3 Transport Layer

Chapter 3 Transport Layer

CMPE 150/L : Introduction to Computer Networks. Chen Qian Computer Engineering UCSC Baskin Engineering Lecture 9

CSC 401 Data and Computer Communications Networks

10 minutes survey (anonymous)

Chapter III: Transport Layer

CMPE 150/L : Introduction to Computer Networks. Chen Qian Computer Engineering UCSC Baskin Engineering Lecture 10

Chapter 3 Transport Layer

CNT 6885 Network Review on Transport Layer

Correcting mistakes. TCP: Overview RFCs: 793, 1122, 1323, 2018, TCP seq. # s and ACKs. GBN in action. TCP segment structure

Lecture 8. TCP/IP Transport Layer (2)

CC451 Computer Networks

CSC 8560 Computer Networks: TCP

Chapter 3 outline. 3.5 Connection-oriented transport: TCP. 3.6 Principles of congestion control 3.7 TCP congestion control

Chapter 3 outline. 3.5 Connection-oriented transport: TCP. 3.6 Principles of congestion control 3.7 TCP congestion control

TCP: Overview RFCs: 793, 1122, 1323, 2018, 2581

CSC358 Week 5. Adapted from slides by J.F. Kurose and K. W. Ross. All material copyright J.F Kurose and K.W. Ross, All Rights Reserved

Outline. TCP: Overview RFCs: 793, 1122, 1323, 2018, steam: r Development of reliable protocol r Sliding window protocols

Outline. TCP: Overview RFCs: 793, 1122, 1323, 2018, Development of reliable protocol Sliding window protocols

32 bits. source port # dest port # sequence number acknowledgement number not used. checksum. Options (variable length)

Lecture 12: Transport Layer TCP again

rdt3.0: channels with errors and loss

Transport Layer: outline

Chapter 3 Transport Layer

Course on Computer Communication and Networks. Lecture 5 Chapter 3; Transport Layer, Part B

Pipelined protocols: overview

Chapter 3 Transport Layer

Lecture 08: The Transport Layer (Part 2) The Transport Layer Protocol (TCP) Dr. Anis Koubaa

Transport Layer: Outline

Suprakash Datta. Office: CSEB 3043 Phone: ext Course page:

Computer Communication Networks Midterm Review

CSC 401 Data and Computer Communications Networks

CSE 4213: Computer Networks II

Chapter 3 Transport Layer

CSC 4900 Computer Networks: TCP

Chapter 3 outline. 3.5 Connection-oriented transport: TCP. 3.6 Principles of congestion control 3.7 TCP congestion control

Go-Back-N. Pipelining: increased utilization. Pipelined protocols. GBN: sender extended FSM

Computer Networking Introduction

RSC Part III: Transport Layer 3. TCP

Transport layer. Review principles: Instantiation in the Internet UDP TCP. Reliable data transfer Flow control Congestion control

COMP 431 Internet Services & Protocols. Transport Layer Protocols & Services Outline. The Transport Layer Reliable data delivery & flow control in TCP

Chapter 3 Transport Layer

Chapter 3 Transport Layer

Transport layer. UDP: User Datagram Protocol [RFC 768] Review principles: Instantiation in the Internet UDP TCP

Chapter 3 Transport Layer

CS450 Introduc0on to Networking Lecture 14 TCP. Phu Phung Feb 13, 2015

CSC 4900 Computer Networks: TCP

CS 43: Computer Networks. 19: TCP Flow and Congestion Control October 31, Nov 2, 2018

Chapter 3 outline. Chapter 3: Transport Layer. Transport vs. network layer. Transport services and protocols. Internet transport-layer protocols

TCP. TCP: Overview. TCP Segment Structure. Maximum Segment Size (MSS) Computer Networks 10/19/2009. CSC 257/457 - Fall

Fall 2012: FCM 708 Bridge Foundation I

Master Course Computer Networks IN2097

Lecture 11. Transport Layer (cont d) Transport Layer 1

Chapter 3: Transport Layer

Chapter 3: Transport Layer

internet technologies and standards

LECTURE 3 - TRANSPORT LAYER

Computer Networking: A Top Down Approach

internet technologies and standards

CSCE 463/612 Networks and Distributed Processing Spring 2017

Mid Term Exam Results

Application. Transport. Network. Link. Physical

Chapter 3 Transport Layer

Chapter 3 Transport Layer

Chapter 3 Transport Layer

The Transport Layer Reliable data delivery & flow control in TCP. Transport Layer Protocols & Services Outline

Chapter 3 Transport Layer

By Ossi Mokryn, Based also on slides from: the Computer Networking: A Top Down Approach Featuring the Internet by Kurose and Ross

Lecture 15: Transport Layer Congestion Control

Chapter 3 Transport Layer

Transport Layer. Chapter 3. Computer Networking: A Top Down Approach

Transport Layer. CMPS 4750/6750: Computer Networks

The Transport Layer Reliable data delivery & flow control in TCP. Transport Layer Protocols & Services Outline

Routers. Session 12 INST 346 Technologies, Infrastructure and Architecture

Chapter 3 Transport Layer

Architettura di Reti

Chapter 3 Transport Layer

Chapter 3 Transport Layer

CS Lecture 1 Review of Basic Protocols

Chapter 3: Transport Layer

CSCD 330 Network Programming

EC441 Fall 2018 Introduction to Computer Networking Chapter 3: Transport Layer

Chapter 3 Transport Layer

Chapter 6 Transport Layer

Transport Layer PREPARED BY AHMED ABDEL-RAOUF

Chapter 3 Transport Layer

Transmission Control Protocol

Chapter 3: Transport Layer

CS/ECE 438: Communication Networks Fall Transport Layer

Lecture 5. Transport Layer. Transport Layer 1-1

Chapter 3 Transport Layer

Transcription:

Chapter III: Transport Layer UG3 Computer Communications & Networks (COMN) Myungjin Lee myungjin.lee@ed.ac.uk Slides copyright of Kurose and Ross

TCP: Overview RFCs: 793,1122,1323, 2018, 2581 point-to-point: one sender, one receiver reliable, in-order byte steam: no message boundaries pipelined: TCP congestion and flow control set window size full duplex data: bi-directional data flow in same connection MSS: maximum segment size connection-oriented: handshaking (exchange of control msgs) inits sender, receiver state before data exchange flow controlled: sender will not overwhelm receiver 68

TCP segment structure URG: urgent data (generally not used) ACK: ACK # valid PSH: push data now RST, SYN, FIN: connection estab (setup, teardown commands) Internet checksum (as in UDP) head len 32 bits source port # dest port # sequence number acknowledgement number not used UAP R S F checksum receive window application data (variable length) Urg data pointer options (variable length) counting by bytes of data (not segments!) # bytes rcvr willing to accept 69

TCP seq. numbers, ACKs sequence numbers: byte stream number of first byte in segment s data acknowledgements: seq # of next byte expected from other side cumulative ACK Q: how receiver handles outof-order segments A: TCP spec doesn t say, - up to implementor outgoing segment from sender source port # dest port # sequence number acknowledgement number rwnd checksum sent ACKed urg pointer window size N sender sequence number space sent, notyet ACKed ( inflight ) incoming segment to sender source port # dest port # sequence number acknowledgement number rwnd A checksum usable but not yet sent urg pointer not usable 70

Byte stream in TCP Cannot be transmitted now 100th byte HTTP Get Message (K bytes) M bytes Window: N bytes TCP header (seq no. = 100) HTTP Get Message (K bytes) 71

TCP seq. numbers, ACKs Host A Host B User types C host ACKs receipt of echoed C Seq=42, ACK=79, data = C Seq=79, ACK=43, data = C Seq=43, ACK=80 host ACKs receipt of C, echoes back C simple telnet scenario 72

TCP round trip time, timeout Q: how to set TCP timeout value? longer than RTT but RTT varies too short: premature timeout, unnecessary retransmissions too long: slow reaction to segment loss Q: how to estimate RTT? SampleRTT: measured time from segment transmission until ACK receipt ignore retransmissions SampleRTT will vary, want estimated RTT smoother average several recent measurements, not just current SampleRTT 73

TCP round trip time, timeout EstimatedRTT = (1- a)*estimatedrtt + a*samplertt v exponential weighted moving average v influence of past sample decreases exponentially fast v typical value: a = RTT: 0.125 gaia.cs.umass.edu to fantasia.eurecom.fr 350 RTT: gaia.cs.umass.edu to fantasia.eurecom.fr RTT (milliseconds) RTT (milliseconds) 300 250 200 150 samplertt EstimatedRTT 100 1 8 15 22 29 36 43 50 57 64 71 78 85 92 99 106 time (seconnds) time (seconds) SampleRTT Estimated RTT 74

TCP round trip time, timeout timeout interval: EstimatedRTT plus safety margin large variation in EstimatedRTT -> larger safety margin estimate SampleRTT deviation from EstimatedRTT: DevRTT = (1-b)*DevRTT + b* SampleRTT-EstimatedRTT (typically, b = 0.25) TimeoutInterval = EstimatedRTT + 4*DevRTT estimated RTT safety margin 75

TCP reliable data transfer TCP creates rdt service on top of IP s unreliable service pipelined segments cumulative acks single retransmission timer retransmissions triggered by: timeout events duplicate acks let s initially consider simplified TCP sender: ignore duplicate acks ignore flow control, congestion control 76

TCP sender events: data rcvd from app: create segment with seq # seq # is byte-stream number of first data byte in segment start timer if not already running think of timer as for oldest unacked segment expiration interval: TimeOutInterval timeout: retransmit segment that caused timeout restart timer ack rcvd: if ack acknowledges previously unacked segments update what is known to be ACKed start timer if there are still unacked segments 77

TCP sender (simplified) L NextSeqNum = InitialSeqNum SendBase = InitialSeqNum wait for event ACK received, with ACK field value y data received from application above create segment, seq. #: NextSeqNum pass segment to IP (i.e., send ) NextSeqNum = NextSeqNum + length(data) if (timer currently not running) start timer timeout retransmit not-yet-acked segment with smallest seq. # start timer if (y > SendBase) { SendBase = y /* SendBase 1: last cumulatively ACKed byte */ if (there are currently not-yet-acked segments) start timer else stop timer } 78

TCP: retransmission scenarios Host A Host B Host A Host B Seq=92, 8 bytes of data SendBase=92 Seq=92, 8 bytes of data timeout X ACK=100 timeout Seq=100, 20 bytes of data ACK=100 ACK=120 Seq=92, 8 bytes of data ACK=100 lost ACK scenario SendBase=100 SendBase=120 SendBase=120 Seq=92, 8 bytes of data ACK=120 premature timeout 79

TCP: retransmission scenarios Host A Host B Seq=92, 8 bytes of data Seq=100, 20 bytes of data timeout X ACK=100 ACK=120 Seq=120, 15 bytes of data cumulative ACK 80

TCP ACK generation [RFC 5861] event at receiver arrival of in-order segment with expected seq #. All data up to expected seq # already ACKed arrival of in-order segment with expected seq #. One other segment has ACK pending arrival of out-of-order segment higher-than-expect seq. #. Gap detected arrival of segment that partially or completely fills gap TCP receiver action delayed ACK. Wait up to 500ms for next segment. If no next segment, send ACK immediately send single cumulative ACK, ACKing both in-order segments immediately send duplicate ACK, indicating seq. # of next expected byte immediate send ACK, provided that segment starts at lower end of gap 81

TCP fast retransmit time-out period often relatively long: long delay before resending lost packet detect lost segments via duplicate ACKs sender often sends many segments back-to-back if segment is lost, there will likely be many duplicate ACKs TCP fast retransmit if sender receives 3 ACKs for same data ( triple duplicate ACKs ), resend unacked segment with smallest seq # likely that unacked segment lost, so don t wait for timeout 82

TCP fast retransmit Host A Host B Seq=92, 8 bytes of data Seq=100, 20 bytes of data X ACK=100 3 DUP ACKs timeout ACK=100 ACK=100 ACK=100 Seq=100, 20 bytes of data fast retransmit after sender receipt of triple duplicate ACK 83

TCP flow control application may remove data from TCP socket buffers. slower than TCP receiver is delivering (sender is sending) application process TCP socket receiver buffers TCP code application OS flow control receiver controls sender, so sender won t overflow receiver s buffer by transmitting too much, too fast from sender IP code receiver protocol stack 84

TCP flow control receiver advertises free buffer space by including rwnd value in TCP header of receiver-to-sender segments RcvBuffer size set via socket options (typical default is 4096 bytes) many operating systems autoadjust RcvBuffer sender limits amount of unacked ( in-flight ) data to receiver s rwnd value guarantees receive buffer will not overflow RcvBuffer rwnd to application process buffered data free buffer space TCP segment payloads receiver-side buffering 85

Connection Management before exchanging data, sender/receiver handshake : agree to establish connection (each knowing the other willing to establish connection) agree on connection parameters application connection state: ESTAB connection variables: seq # client-to-server server-to-client rcvbuffer size at server,client network application connection state: ESTAB connection Variables: seq # client-to-server server-to-client rcvbuffer size at server,client network Socket clientsocket = newsocket("hostname","port number"); Socket connectionsocket = welcomesocket.accept(); 86

TCP 3-way handshake client state CLOSED SYNSENT ESTAB choose init seq num, x send TCP SYN msg received SYNACK(x) indicates server is live; send ACK for SYNACK; this segment may contain client-to-server data SYNbit=1, Seq=x SYNbit=1, Seq=y ACKbit=1; ACKnum=x+1 ACKbit=1, ACKnum=y+1 choose init seq num, y send TCP SYNACK msg, acking SYN received ACK(y) indicates client is live server state LISTEN SYN RCVD ESTAB 87

TCP 3-way handshake: FSM closed Socket connectionsocket = welcomesocket.accept(); SYN(x) SYNACK(seq=y,ACKnum=x+1) create new socket for communication back to client L listen Socket clientsocket = newsocket("hostname","port number"); SYN(seq=x) SYN rcvd SYN sent ACK(ACKnum=y+1) L ESTAB SYNACK(seq=y,ACKnum=x+1) ACK(ACKnum=y+1) 88

TCP: closing a connection client, server each close their side of connection send TCP segment with FIN bit = 1 respond to received FIN with ACK on receiving FIN, ACK can be combined with own FIN simultaneous FIN exchanges can be handled 89

TCP: closing a connection client state server state ESTAB ESTAB clientsocket.close() FIN_WAIT_1 FIN_WAIT_2 can no longer send but can receive data wait for server close FINbit=1, seq=x ACKbit=1; ACKnum=x+1 can still send data CLOSE_WAIT TIMED_WAIT timed wait for 2*max segment lifetime FINbit=1, seq=y ACKbit=1; ACKnum=y+1 can no longer send data LAST_ACK CLOSED CLOSED 90

Principles of congestion control congestion: informally: too many sources sending too much data too fast for network to handle different from flow control! manifestations: lost packets (buffer overflow at routers) long delays (queueing in router buffers) a top-10 problem! 91

Causes/costs of congestion: scenario 1 two senders, two receivers one router, infinite buffers output link capacity: R no retransmission Host B original data: l in Host A unlimited shared output link buffers throughput: l out R/2 l out delay maximum per-connection throughput: R/2 l in R/2 l in R/2 v large delays as arrival rate, l in, approaches capacity 92

Causes/costs of congestion: scenario 2 one router, finite buffers sender retransmission of timed-out packet application-layer input = application-layer output: l in = l out transport-layer input includes retransmissions : l in l in l in : original data l' in : original data, plus retransmitted data l out Host A Host B finite shared output link buffers 93

Causes/costs of congestion: scenario 2 idealization: perfect knowledge sender sends only when router buffers available R/2 l out l in R/2 copy l in : original data l' in : original data, plus retransmitted data l out A free buffer space! Host B finite shared output link buffers 94

Causes/costs of congestion: scenario 2 Idealization: known loss packets can be lost, dropped at router due to full buffers sender only resends if packet known to be lost copy l in : original data l' in : original data, plus retransmitted data l out A no buffer space! Host B 95

Causes/costs of congestion: scenario 2 Idealization: known loss packets can be lost, dropped at router due to full buffers sender only resends if packet known to be lost R/2 l out l in R/2 when sending at R/2, some packets are retransmissions but asymptotic goodput is still R/2 (why?) l in : original data l' in : original data, plus retransmitted data l out A free buffer space! Host B 96

Realistic: duplicates v v Causes/costs of congestion: scenario 2 packets can be lost, dropped at router due to full buffers sender times out prematurely, sending two copies, both of which are delivered R/2 l out l in R/2 when sending at R/2, some packets are retransmissions including duplicated that are delivered! copy timeout l in l' in l out A free buffer space! Host B 97

Causes/costs of congestion: scenario 2 Realistic: duplicates v v packets can be lost, dropped at router due to full buffers sender times out prematurely, sending two copies, both of which are delivered l out R/2 l in R/2 when sending at R/2, some packets are retransmissions including duplicated that are delivered! costs of congestion: v v more work (retrans) for given goodput unneeded retransmissions: link carries multiple copies of pkt decreasing goodput 98

Causes/costs of congestion: scenario 3 four senders multihop paths timeout/retransmit Q: what happens as l in and l in increase? A: as red l in increases, all arriving blue pkts at upper queue are dropped, blue throughput g 0 Host A l out Host B l in : original data l' in : original data, plus retransmitted data finite shared output link buffers Host D Host C 99

Throughput by blue traffic R/2 l out Causes/costs of congestion: scenario 3 l in R/2 Offered load by Host A Bandwidth wastage for packets dropped at the 2nd router another cost of congestion: v when packet dropped, any upstream transmission capacity used for that packet was wasted! 100

Approaches towards congestion control two broad approaches towards congestion control: end-end congestion control: no explicit feedback from network congestion inferred from end-system observed loss, delay approach taken by TCP network-assisted congestion control: routers provide feedback to end systems single bit indicating congestion (SNA, DECbit, TCP/IP ECN, ATM) explicit rate for sender to send at 101

TCP congestion control: additive increase multiplicative decrease (AIMD) v approach: sender increases transmission rate (window size), probing for usable bandwidth, until loss occurs additive increase: increase cwnd by 1 MSS every RTT until loss detected multiplicative decrease: cut cwnd in half after loss AIMD saw tooth behavior: probing for bandwidth cwnd: TCP sender congestion window size additively increase window size. until loss occurs (then cut window in half) time 102

TCP Congestion Control: details sender sequence number space cwnd last byte ACKed last byte sent sender limits transmission: sent, notyet ACKed ( inflight ) LastByteSent- LastByteAcked < cwnd cwnd is dynamic, function of perceived network congestion TCP sending rate: roughly: send cwnd bytes, wait RTT for ACKS, then send more bytes rate ~ cwnd RTT bytes/sec 103

TCP Slow Start when connection begins, increase rate exponentially until first loss event: initially cwnd = 1 MSS double cwnd every RTT done by incrementing cwnd for every ACK received summary: initial rate is slow but ramps up exponentially fast Host A RTT Host B time 104

TCP: detecting, reacting to loss loss indicated by timeout: cwnd set to 1 MSS; window then grows exponentially (as in slow start) to threshold, then grows linearly loss indicated by 3 duplicate ACKs: TCP RENO dup ACKs indicate network capable of delivering some segments cwnd is cut in half window then grows linearly TCP Tahoe always sets cwnd to 1 (timeout or 3 duplicate acks) 105

TCP: switching from slow start to CA Q: when should the exponential increase switch to linear? A: when cwnd gets to 1/2 of its value before timeout. Implementation: variable ssthresh on loss event, ssthresh is set to 1/2 of cwnd just before loss event 106

Summary: TCP Congestion Control L cwnd = 1 MSS ssthresh = 64 KB dupackcount = 0 timeout ssthresh = cwnd/2 cwnd = 1 MSS dupackcount = 0 retransmit missing segment dupackcount == 3 ssthresh= cwnd/2 cwnd = ssthresh + 3 retransmit missing segment duplicate ACK dupackcount++ slow start New ACK! new ACK cwnd = cwnd+mss dupackcount = 0 transmit new segment(s), as allowed cwnd > ssthresh L timeout ssthresh = cwnd/2 cwnd = 1 MSS dupackcount = 0 retransmit missing segment timeout ssthresh = cwnd/2 cwnd = 1 dupackcount = 0 retransmit missing segment fast recovery duplicate ACK new ACK cwnd = cwnd + MSS (MSS/cwnd) dupackcount = 0 transmit new segment(s), as allowed cwnd = ssthresh dupackcount = 0 congestion avoidance New ACK! New ACK cwnd = cwnd + MSS transmit new segment(s), as allowed. New ACK! duplicate ACK dupackcount++ dupackcount == 3 ssthresh= cwnd/2 cwnd = ssthresh + 3 retransmit missing segment 107

TCP throughput avg. TCP throuput as function of window size, RTT? ignore slow start, assume always data to send W: window size (measured in bytes) where loss occurs avg. window size (# in-flight bytes) is ¾ W avg. throuput is 3/4W per RTT W avg TCP throuput = 3 4 W RTT bytes/sec W/2 3/4W 108

TCP Fairness fairness goal: if K TCP sessions share same bottleneck link of bandwidth R, each should have average rate of R/K TCP connection 1 TCP connection 2 bottleneck router capacity R 109

two competing sessions: Why is TCP fair? additive increase gives slope of 1, as throughout increases multiplicative decrease decreases throughput proportionally Full bandwidth utilization line R equal bandwidth share (X 2, Y 2 ) where X 2 = Y 2 loss: decrease window by factor of 2 congestion avoidance: additive increase loss: decrease window by factor of 2 congestion avoidance: additive increase (X 1, Y 1 ) where X 1 +Y 1 = R Connection 1 throughput R 110

Fairness (more) Fairness and UDP multimedia apps often do not use TCP do not want rate throttled by congestion control instead use UDP: send audio/video at constant rate, tolerate packet loss Fairness, parallel TCP connections application can open multiple parallel connections between two hosts web browsers do this e.g., link of rate R with 9 existing connections: new app asks for 1 TCP, gets rate R/10 new app asks for 11 TCPs, gets R/2 111