The CSIRO ATNF Gigabit Wide-Area Network Shaun W Amy CSIRO Australia Telescope
Download
Report
Transcript The CSIRO ATNF Gigabit Wide-Area Network Shaun W Amy CSIRO Australia Telescope
The CSIRO ATNF Gigabit
Wide-Area Network
Shaun W Amy
<[email protected]>
CSIRO Australia Telescope
National Facility
Project Aim
• Initially provide a 1Gbit/s link to each of the three ATNF
observatories:
– Parkes,
– Mopra (cost/fibre sharing with ANU’s Siding Springs Observatory),
– ATCA Narrabri (and then to CSIRO Plant Industry, Myall Vale).
• Uses the AARNet Regional Transmission Service (RTS).
• Connect each observatory (without aggregation) back to
CSIRO ATNF headquarters at Marsfield in Sydney.
• Require network performance and stability that can
enable “real” e-VLBI.
• Production and Research traffic will share the same link.
• Extend the network to WASP at UWA (via
CeNTIE/GrangeNet) and also to Swinburne University
(via the southern leg) of the AARNet Regional Network.
An ATNF Telescope “Refresher”
• Parkes:
–
–
–
–
–
64m prime-focus antenna,
range of receivers and backends,
celebrates its 45th birthday in October,
VLBI (including Mk III for geodesy),
the star of The Dish.
• Mopra:
–
–
–
–
22m wheel-on-track, cassegrain antenna,
primary use is for mm observations during winter,
new spectrometer: MOPS,
almost always used for VLBI (including the first Mopra observations).
• Narrabri:
–
–
–
–
–
6 x 22m (with 5 movable), cassegrain antennas,
frequency agility,
various antenna configurations (5km E-W track with N-S spur),
CABB wide-band backend scheduled for 2007-8,
VLBI: tied-array mode.
The Parkes Telescope
The Mopra Telescope
The Australia Telescope Compact
Array, Narrabri
AARNet Regional Transmission
Service
• Implemented using Nextgen fibre infrastructure.
• The RTS provides a connection between the
Nextgen connection point on the regional network
and the Nextgen POP located in the capital city.
• Point-to-point Ethernet service (Layer 2):
– the customer can use this however they wish.
• Service delivered via CWDM MUX or direct fibre
depending on the connection model (see later)
• Regional fibre tail builds are the responsibility of
the customer not AARNet.
• “Last mile” in the capital city is the responsibility
of the customer.
The Nextgen Network
Source: AARNet Pty Ltd
Backbone Design
• Nextgen:
– AARNet have access to two fibre pairs (not the whole Nextgen
network),
– Pair 1: AARNet DWDM 10Gbit/s service (provides inter-capital city
AARNet3 service),
– Pair 2: Physical connections to tail sites.
• Implemented using Cisco Carrier-class (ONS 15454)
optical transmission systems:
– initially supports 16 x 10Gbit/s wavelengths,
– can be upgraded to 32 x 10Gbit/s with no chassis changes.
– each customer 1Gbit/s service is full-line rate (i.e. no
oversubscription).
• Requires amplification/regeneration every 80-100km:
– optical-optical and optical-electrical-optical,
– housed in a Controlled Environment Vault (CEV) but these are not
always located at ideal locations for site connections!
Part of the Nextgen SB2 Segment
Source: AARNet Pty Ltd
Connecting to the AARNet Regional
Transmission Network
Source: AARNet Pty Ltd
Cost Considerations (for a 1Gbit/s
transmission service)
• Setup/Install/Construction:
–
–
–
–
–
fibre build (approx $2m),
break-out equipment ($32k per location)
initial connection charge within a network segment ($60k per circuit),
active equipment (switches, routers, optical transceivers, patch leads),
travel/labour.
• Recurrent:
–
–
–
–
access charge per circuit ($34k p.a.),
fibre maintenance charges (about $40k p.a.),
CSIRO equipment maintenance/self-sparing,
labour.
• Other:
– depreciation,
– whole-of-life costing and equipment rollover/upgrades.
Network Design
•
•
•
•
•
•
Implemented by ATNF using mid-range equipment at the observatories
capable of 1Gbit/s but NOT 10Gbit/s.
Can support “jumbo” frames at Layer 2 but NOT at Layer 3.
CSIRO’s corporate IT group are interested in upgrading/exchanging this
equipment for high-end hardware that is modular and capable of 10Gbit/s
but…
Combined Layer 2 (Ethernet) and Layer 3 (IP) network.
For e-VLBI, a layer 2 Virtual Local Area Network (VLAN) has been
implemented across all ATNF sites (with extensions to UWA and Swinburne),
primarily for performance reasons.
The e-VLBI VLAN uses so-called CSIRO “untrusted” address space and thus
can be accessible from hosts that aren’t connected to this VLAN (e.g.
University of Tasmania, JIVE etc):
– this external connectivity is via a standard routed IP connection,
– this traffic transits a CSIRO firewall appliance.
•
For production traffic, layer 3 point-to-point links are used which allows for
rapid failover to a backup link (via the existing Layer 3 routing protocols):
– a recent science-related use of the production network has been to implement
Mopra remote observing from Narrabri.
A Hybrid Switched/Routed Network
Network Protocol and Performance
Considerations
• TCP or UDP?
– currently using TCP.
• Data recorders currently running kernel 2.6.16.
• What about Ethernet “jumbo” frames:
– not currently being used by the ATNF disk/network-based
recorders.
• A number of TCP variants were tested, including Reno,
BIC, highspeed, htcp) and settled on BIC.
• Default TCP configuration is not tuned for high-bandwidth,
long-haul networks:
– TCP window is the amount of un-acknowledged data in the
network,
– Optimise buffers (TCP window size) using:
window = bandwidth x RTT
Performance between e-VLBI data
recorders (memory-memory)
Source: Dr Chris Phillips
Future Developments (1)
• Additional three 1Gbit/s links to be commissioned:
– location of endpoints,
– ensure ATNF production and research (e-VLBI) network requirements are
met,
– satisifying the requirements of CSIRO’s corporate IT group,
– load balancing/link sharing considerations.
• “Lighting up” the Swinburne connection:
– ATNF have an agreement with Swinburne to provide a software
correllation facility starting 1 October 2006,
– technically easy but legal issues causing the delay.
• Is there a simple mechanism to guarantee that e-VLBI gets the
required bandwidth when needed (e.g. some form of policing/ratelimiting)?
• Is Quality of Service (QoS) required?
• Should we use “jumbo” frames as the default even though good
results are being obtained with a 1500byte MTU on 1Gbit/s links.
Future Developments (2)
• The (not-so-wild) West:
– Currently uses GrangeNet and CeNTIE (at no direct cost) to provide a
dedicated 1Gbit/s path to the hosts at WASP at UWA,
– GrangeNet due to close before the end of 2006,
– the 10Gbit/s (multiple 1Gbit/s circuits) CeNTIE Melbourne-Perth path will
be shutdown in December 2006
– AARNet3 production service is a possible alternative but need to
consider the following:
• cost (traffic charges and setup),
• layer 3 (i.e. IPv4/v6 routed traffic) only,
• currently provides 1Gbit/s connections via a somewhat restrictive connection
mechanism,
– Need to factor in xNTD (and LFD) network requirements.
• EXPReS:
– engineering of overseas links (will AARNet look to providing UCLP or
some sort of hybrid optical-packet technology on one of the two SX
Transport 10Gbit/s links between Australia and the USA),
• 10Gbit/s and beyond…
• Do we need to consider non-Ethernet based services?
First Mopra Remote Observing by Dion
Lewis (Operations Scientist) on Saturday
29 July 2006