SlideShare ist ein Scribd-Unternehmen logo
1 von 68
Downloaden Sie, um offline zu lesen
The CERN Network

                        Openlab Summer 2012
                        CERN, 6th August 2012
                       edoardo.martelli@cern.ch



CERN IT Department
CH-1211 Genève 23
   Switzerland
www.cern.ch/it
                                                  1
Summary

 - IT-CS
 - CERN networks
 - LHC Data Challenge
 - WLCG
 - LHCOPN and LHCONE
 - Openlab
 - Conclusions

                        2
IT-CS
Communication systems




                        3
IT-CS


      The IT-CS group is responsible for all
    communication services in use at CERN for
             data, voice and video

        http://it-cs.web.cern.ch/it-cs/




                                                4
IT-CS organization




                     5
Networks at CERN



                   6
CERN accelerator complex




                           7
High Energy Physics over IP



    Most of the CERN infrastructure is
     controlled and managed over a
         pervasive IP network




                                         8
Cryogenics

       27Km of pipes at
       -271.11° C by means of
       700.000 litres of Helium:
       controlled over IP




 Source: http://te-dep-crg-oa.web.cern.ch/te-dep-crg-oa/te-crg-oa_fichiers/cryolhc/LHC%20Cryo_BEOP_lectures2009.pdf

                                                                                                                      9
Access control
Safety and Security: made over IP




        Source:https://edms.cern.ch/file/931641/1/LASS-LACS_IHM.pdf

                                                                      10
Remote inspections

                Remote inspection of dangerous areas: robots
                controlled and giving feedback over WiFi and
                GSM IP networks




                                                               11
DAQ: Data Acquisition
A constant stream of data from the four Detectors to disk storage




  Source: http://aliceinfo.cern.ch/Public/Objects/Chapter2/DetectorComponents/daq_architecture.pdf

                                                                                                     12
CCC: CERN Control Centre
The neuralgic centre of the particle accelerator: over IP




                                                            13
CERN data network




              - 150 routers
              - 2200 Switches
              - 50000 connected devices
              - 5000km of optical fibres




                                           14
Network Provisioning and
Management System
 - 250 Database
   tables
 - 100,000
   Registered devices
 - 50,000 hits/day on
   web user interface
 - 1,000,000 lines of
   codes
 - 11 years of
   development




                           15
Monitoring and Operations
The whole network is monitored and operated by the CERN NOC
(Network Operation Centre)




                                                              16
IPv6


  IPv6 dual stack network deployment on going:
  ready in 2013

  Already available: dual-stack testbed


  More information:
   http://cern.ch/ipv6
                                 almost




                                                 17
LHC Data Challenge



                     18
Collisions in the LHC




                        19
Comparing theory...




     Simulated production of a Higgs event in ATLAS

                                                      20
.. to real events




 Higgs event in CMS
                      21
Data flow
                                   3 PBytes/s

     4 Experiments
                                                                        Filter and first selection




                                                2 GBytes/s
                                                to the CERN computer center



                               Create sub-samples

                                                                                       World-Wide Analysis
                                                       1 TByte/s ?
                                                  Distributed + local                                 Physics
                                                                                                Explanation of nature
   10 GBytes/s   4 GBytes/s
                                                                                          sΓZ2
  Store on disk and tape      Export copies                   σ    _   ≈σ _×
                                                                           0
                                                                                                          with
                                                                  ff     ff  ( s - m 2 ) 2 + s 2 ΓZ / m 2
                                                                                     Z
                                                                                                  2
                                                                                                        z

                                                            12π Γ Γ                         GF m 3
                                                  σ   0
                                                       _   = 2 ee 2 ff         and    Γff =      Z
                                                                                                   × ( v f2 + a f2 ) × N col
                                                      ff    m Z ΓZ                          6π 2


                                                                                                                               22
Data Challenge


 - 40 million collisions per second

 - After filtering, 100 collisions of interest per
    second

 - 1010 collisions recorded each year
    = 15 Petabytes/year of data



                                                     23
Computing model




                  24
Last months data transfers




                             25
WLCG
Worldwide LHC Computing Grid




                               26
WLCG

 Distributed Computing Infrastructure for
 LHC experiments


 Collaborative effort of the HEP
 community




                                            27
WLCG resources


  WLCG sites:
  - 1 Tier0 (CERN)
  - 11 Tier1s
  - ~140 Tier2s
  - >300 Tier3s worldwide
  - ~250,000 CPUs
  - ~ 150PB of disk space




                            28
CERN Tier0 resources


 Servers              11000   Disks                    64000

 Processors           15000   Raw Disk Capacity (TB)   63000

 Cores                64000   Memory Modules           56000

 HEPspec06           480000   RAID controllers          3750



 Tape drives           160    High Speed routers         23

 Tape cartridges     45000    Ethernet switches         500

 Tape slots          56000    10Gbps ports             3000

 Tape capacity(TB)   34000    100Gbps ports              48




 March 2012
                                                               29
CERN Tier0 LCG network


                       Tier2/3s       Tier1s                        CERN Campus


                                               170G aggregated

         Border routers                                                                 LHC Experiments


                                                                                                   40G links

                       Core routers


                                                                                      100G links


              Distribution routers


                                                                   10G or 40G links


      LCG access switches
        Access switches                                      ...            x892 (max)

                                                          1G or 10G links
    Servers


                                                                                                               30
Trends


 Virtualization mobility (Software Defined
 Networks)

 Commodity Servers with 10G NICs

 High-end Servers with 40G NICs

 40G and 100G interfaces on switches and
 routers


                                             31
LHCOPN
LHC Optical Private Network




                              32
Tier0-Tier1s network




                       33
A collaborative effort


 Designed, built and operated by the Tier0-Tier1s
 community

 Links provided by the Research and Education
 network providers: Geant, USLHCnet, Esnet,
 Canarie, ASnet, Nordunet, Surfnet, GARR,
 Renater, JANET.UK, Rediris, DFN, SWITCH




                                                    34
Technology

  - Single and bundled long distance 10G
    ethernet links
  - Multiple redundant paths. Star+PartialMesh
    topology
  - BGP routing: communities for traffic
    engineering, load balancing.
  - Security: only declared IP prefixes can
    exchange traffic.



                                                 35
Traffic to the Tier1s




                        36
Monitoring




             37
LHCONE
LHC Open Network Environment




                               38
Driving the change

  “The Network infrastructure is the most
  reliable service we have”

   “Network Bandwidth (rather than disk)
    will need to scale more with users and
                 data volume”

     “Data placement will be driven by
     demand for analysis and not pre-
                placement”

                 Ian Bird, WLCG project leader
                                                 39
Change of computing model (ATLAS)




                                    40
New computing model


 - Better and more dynamic use of storage
 - Reduce the load on the Tier1s for data serving
 - Increase the speed to populate analysis
   facilities

 Needs for a faster, predictable, pervasive
  network connecting Tier1s and Tier2s



                                                    41
Requirements from the Experiments

  - Connecting any pair of sites, regardless of the
    continent they reside
  - Bandwidth ranging from 1Gbps (Minimal), 5Gbps
    (Nominal), 10G and above (Leadership)
  - Scalability: sites are expected to grow
  - Flexibility: sites may join and leave at any time
  - Predictable cost: well defined cost, and not too
    high


                                                        42
Needs for a better network

  - more bandwidth by federating (existing)
    resources
  - sharing cost of expensive resources
  - accessible to any TierX site

                         =


        LHC Open Network Environment
                                              43
LHCONE concepts

  - Serves any LHC sites according to their needs and
  allowing them to grow

  - A collaborative effort among Research & Education
  Network Providers

  - Based on Open Exchange Points: easy to join,
  neutral

  - Multiple services: one cannot fit all

  - Traffic separation: no clash with other data transfer,
  resource allocated for and funded by HEP community
                                                             44
LHCONE architecture




                      45
LHCONE building blocks


 - Single node Exchange Points
 - Continental/regional Distributed Exchange Points
 - Interconnect circuits between Exchange Points

        These exchange points and the links in between
  collectively provide LHCONE services and operate under a
                    common LHCONE policy




                                                             46
The underlying infrastructure




                                47
LHCONE services


  - Layer3 VPN
  - Point-to-Point links
  - Monitoring




                           48
Openlab and IT-CS



                    49
Openlab project:
  CINBAD



                   50
CINBAD
 CERN Investigation of Network Behaviour and Anomaly Detection

 Project Goals:
  Understand the behaviour of large computer networks (10’000+
  nodes) in High Performance Computing or large Campus
  installations to be able to:
   ●
     detect traffic anomalies in the system
   ●
     perform trend analysis
   ●
     automatically take counter measures
   ●
     provide post-mortem analysis facilities


 Resources:
 - In collaboration with HP Networking
 - Two Engineers in IT-CS
                                                                 51
Results

  Project completed in 2010

  For CERN:
  Designed and deployed a complete framework
  (hardware and software) to detect anomalies in
  the Campus Network (GPN)

  For HP:
  Intellectual properties of new technologies used
  in commercial products


                                                     52
CINBAD Architecture

             data
            sources




                                   storage

                                             analysis


                      collectors


                                                        53
Openlab project:
    WIND



                   54
WIND

 Wireless Infrastructure Network Deployment

 Project Goals
 - Analyze the problems of large scale wireless deployments and
   understand the constraint
 - Simulate behaviour of WLAN
 - Develop new optimisation algorithms

 Resources:
 - In collaboration with HP Networking
 - Two Engineers in IT-CS
 - Started in 2010

                                                                  55
Needs

  Wireless LAN (WLAN) deployments are
   problematic:
  ●
      Radio propagation is very difficult to predict

  ●
      Interference is an ever present danger

  ●
      WLANs are difficult to properly deploy

  ●
      Monitoring was not an issue when the first standards
      were developed

  ●
      When administrators are struggling just to operate the
      WLAN, performance optimisation is often forgotten
                                                               56
Example: Radio interferences
  Max data rate in 0031-S: The APs work on the same channel




   Max data rate in 0031-S: The APs work on 3 independent channels




                                                                     57
Expected results


   Extend monitoring and analysis tools

   Act on the network
   - smart load balancing
   - isolating misbehaving clients
   - intelligent minimum data rates

   More accurate troubleshooting

   Streamline WLAN design

                                          58
Openlab project:
   ViSION



                   59
ViSION




 Project Goals:
 - Develop a SDN traffic orchestrator using OpenFlow

 Resources:
 - In collaboration with HP Networking
 - Two Engineers in IT-CS
 - Started in 2012

                                                       60
Goals
   SDN traffic orchestrator using OpenFlow:
    ●
      distribute traffic over a set of network resources
    ●
      perform classification (different types of applications and
      resources)
    ●
      perform load sharing (similar resources).

   Benefits:
    ●
      improved scalability and control than traditional networking
      technologies




                                                                     61
From traditional networks...

Closed boxes, fully distributed protocols


                                     App   App         App



                                           Operating
                                            System
                                                                                     App   App         App
                                      Specialized Packet
                                     Forwarding Hardware
                                                                                           Operating
                                                                                            System

 App   App         App
                                                                                      Specialized Packet
                                                                                     Forwarding Hardware
       Operating
        System
                                                             App   App         App

  Specialized Packet
 Forwarding Hardware                                               Operating
                                                                    System

                                                              Specialized Packet
                         App   App         App               Forwarding Hardware


                               Operating
                                System

                          Specialized Packet
                         Forwarding Hardware


                                                                                                             62
.. to Software Defined Networks (SDN)
                                                               2. At least one good operating system
 3. Well-defined open API                                         Extensible, possibly open-source

                            App         App         App

                              Network Operating System


                                           1. Open interface to hardware
                                          (OpenFlow)

                                  Simple Packet
                                   Forwarding
                                    Hardware
                                                                             Simple Packet
                                                                              Forwarding
                                                                               Hardware


                                                    Simple Packet
                                                     Forwarding
                                                      Hardware
     Simple Packet
      Forwarding
       Hardware



                                    Simple Packet
                                     Forwarding
                                      Hardware
                                                                                                   63
OpenFlow example                                                                                  Controller


                                                                                                           PC
Software
Layer
               OpenFlow Client

                               Flow Table
           MAC       MAC   IP          IP        TCP     TCP
                                                                 Action            Hardware Forwarding table remotely
           src       dst   Src         Dst       sport   dport                      Hardware Forwarding table remotely
Hardware                                                                                       controlled
                                                                                                controlled
           *        *      *           5.6.7.8   *       *       port 1
Layer




                 port 1          port 2              port 3               port 4




                                                                                                                   64
Conclusions



              65
Conclusions

  - The Data Network is an essential component
    of the LHC instrument
  - The Data Network is a key part of the LHC
    data processing and will become even more
    important
  - More and more security and design
    challenges to come




                                                 66
Credits

  Artur Barczyk (LHCONE)
  Dan Savu (VISION)
  Milosz Hulboj (WIND and CINBAD)
  Ryszrard Jurga (CINBAD)
  Sebastien Ceuterickx (WIND)
  Stefan Stancu (VISION)
  Vlad Lapadatescu (WIND)



                                    67
What's next

    SWAN: Space Wide Area Network :-)




                                        68

Weitere ähnliche Inhalte

Was ist angesagt?

Direct Code Execution - LinuxCon Japan 2014
Direct Code Execution - LinuxCon Japan 2014Direct Code Execution - LinuxCon Japan 2014
Direct Code Execution - LinuxCon Japan 2014Hajime Tazaki
 
Netmap presentation
Netmap presentationNetmap presentation
Netmap presentationAmir Razmjou
 
LibOS as a regression test framework for Linux networking #netdev1.1
LibOS as a regression test framework for Linux networking #netdev1.1LibOS as a regression test framework for Linux networking #netdev1.1
LibOS as a regression test framework for Linux networking #netdev1.1Hajime Tazaki
 
Linux Kernel Library - Reusing Monolithic Kernel
Linux Kernel Library - Reusing Monolithic KernelLinux Kernel Library - Reusing Monolithic Kernel
Linux Kernel Library - Reusing Monolithic KernelHajime Tazaki
 
Network Stack in Userspace (NUSE)
Network Stack in Userspace (NUSE)Network Stack in Userspace (NUSE)
Network Stack in Userspace (NUSE)Hajime Tazaki
 
Adaptive Linear Solvers and Eigensolvers
Adaptive Linear Solvers and EigensolversAdaptive Linear Solvers and Eigensolvers
Adaptive Linear Solvers and Eigensolversinside-BigData.com
 
NUSE (Network Stack in Userspace) at #osio
NUSE (Network Stack in Userspace) at #osioNUSE (Network Stack in Userspace) at #osio
NUSE (Network Stack in Userspace) at #osioHajime Tazaki
 
Description of Open-E Snapshot
Description of Open-E SnapshotDescription of Open-E Snapshot
Description of Open-E Snapshotopen-e
 
Linux rumpkernel - ABC2018 (AsiaBSDCon 2018)
Linux rumpkernel - ABC2018 (AsiaBSDCon 2018)Linux rumpkernel - ABC2018 (AsiaBSDCon 2018)
Linux rumpkernel - ABC2018 (AsiaBSDCon 2018)Hajime Tazaki
 
Library Operating System for Linux #netdev01
Library Operating System for Linux #netdev01Library Operating System for Linux #netdev01
Library Operating System for Linux #netdev01Hajime Tazaki
 
HTCC poster for CERN Openlab opendays 2015
HTCC poster for CERN Openlab opendays 2015HTCC poster for CERN Openlab opendays 2015
HTCC poster for CERN Openlab opendays 2015Karel Ha
 
Recent advance in netmap/VALE(mSwitch)
Recent advance in netmap/VALE(mSwitch)Recent advance in netmap/VALE(mSwitch)
Recent advance in netmap/VALE(mSwitch)micchie
 
VLANs in the Linux Kernel
VLANs in the Linux KernelVLANs in the Linux Kernel
VLANs in the Linux KernelKernel TLV
 

Was ist angesagt? (14)

Direct Code Execution - LinuxCon Japan 2014
Direct Code Execution - LinuxCon Japan 2014Direct Code Execution - LinuxCon Japan 2014
Direct Code Execution - LinuxCon Japan 2014
 
Netmap presentation
Netmap presentationNetmap presentation
Netmap presentation
 
LibOS as a regression test framework for Linux networking #netdev1.1
LibOS as a regression test framework for Linux networking #netdev1.1LibOS as a regression test framework for Linux networking #netdev1.1
LibOS as a regression test framework for Linux networking #netdev1.1
 
Linux Kernel Library - Reusing Monolithic Kernel
Linux Kernel Library - Reusing Monolithic KernelLinux Kernel Library - Reusing Monolithic Kernel
Linux Kernel Library - Reusing Monolithic Kernel
 
Network Stack in Userspace (NUSE)
Network Stack in Userspace (NUSE)Network Stack in Userspace (NUSE)
Network Stack in Userspace (NUSE)
 
Adaptive Linear Solvers and Eigensolvers
Adaptive Linear Solvers and EigensolversAdaptive Linear Solvers and Eigensolvers
Adaptive Linear Solvers and Eigensolvers
 
NUSE (Network Stack in Userspace) at #osio
NUSE (Network Stack in Userspace) at #osioNUSE (Network Stack in Userspace) at #osio
NUSE (Network Stack in Userspace) at #osio
 
Description of Open-E Snapshot
Description of Open-E SnapshotDescription of Open-E Snapshot
Description of Open-E Snapshot
 
Linux rumpkernel - ABC2018 (AsiaBSDCon 2018)
Linux rumpkernel - ABC2018 (AsiaBSDCon 2018)Linux rumpkernel - ABC2018 (AsiaBSDCon 2018)
Linux rumpkernel - ABC2018 (AsiaBSDCon 2018)
 
Library Operating System for Linux #netdev01
Library Operating System for Linux #netdev01Library Operating System for Linux #netdev01
Library Operating System for Linux #netdev01
 
HTCC poster for CERN Openlab opendays 2015
HTCC poster for CERN Openlab opendays 2015HTCC poster for CERN Openlab opendays 2015
HTCC poster for CERN Openlab opendays 2015
 
Lrz kurs: big data analysis
Lrz kurs: big data analysisLrz kurs: big data analysis
Lrz kurs: big data analysis
 
Recent advance in netmap/VALE(mSwitch)
Recent advance in netmap/VALE(mSwitch)Recent advance in netmap/VALE(mSwitch)
Recent advance in netmap/VALE(mSwitch)
 
VLANs in the Linux Kernel
VLANs in the Linux KernelVLANs in the Linux Kernel
VLANs in the Linux Kernel
 

Andere mochten auch (6)

Network Virtualization for the Enterprise Data Center
Network Virtualization for the Enterprise Data CenterNetwork Virtualization for the Enterprise Data Center
Network Virtualization for the Enterprise Data Center
 
Career Opportunities at CERN
Career Opportunities at CERNCareer Opportunities at CERN
Career Opportunities at CERN
 
NFV & SDN ( Efficiency & Automation )
NFV & SDN ( Efficiency & Automation )NFV & SDN ( Efficiency & Automation )
NFV & SDN ( Efficiency & Automation )
 
Onderzoekstages CERN
Onderzoekstages CERNOnderzoekstages CERN
Onderzoekstages CERN
 
Summer Student Programme
Summer Student ProgrammeSummer Student Programme
Summer Student Programme
 
Summer @CERN
Summer @CERNSummer @CERN
Summer @CERN
 

Ähnlich wie The size and complexity of the CERN network

Using Photonics to Prototype the Research Campus Infrastructure of the Future...
Using Photonics to Prototype the Research Campus Infrastructure of the Future...Using Photonics to Prototype the Research Campus Infrastructure of the Future...
Using Photonics to Prototype the Research Campus Infrastructure of the Future...Larry Smarr
 
大強子計算網格與OSS
大強子計算網格與OSS大強子計算網格與OSS
大強子計算網格與OSSYuan CHAO
 
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...Larry Smarr
 
High Performance Cyberinfrastructure is Needed to Enable Data-Intensive Scien...
High Performance Cyberinfrastructure is Needed to Enable Data-Intensive Scien...High Performance Cyberinfrastructure is Needed to Enable Data-Intensive Scien...
High Performance Cyberinfrastructure is Needed to Enable Data-Intensive Scien...Larry Smarr
 
20121205 open stack_accelerating_science_v3
20121205 open stack_accelerating_science_v320121205 open stack_accelerating_science_v3
20121205 open stack_accelerating_science_v3Tim Bell
 
The OptIPuter as a Prototype for CalREN-XD
The OptIPuter as a Prototype for CalREN-XDThe OptIPuter as a Prototype for CalREN-XD
The OptIPuter as a Prototype for CalREN-XDLarry Smarr
 
A High-Performance Campus-Scale Cyberinfrastructure for Effectively Bridging ...
A High-Performance Campus-Scale Cyberinfrastructure for Effectively Bridging ...A High-Performance Campus-Scale Cyberinfrastructure for Effectively Bridging ...
A High-Performance Campus-Scale Cyberinfrastructure for Effectively Bridging ...Larry Smarr
 
The Transformation of Systems Biology Into A Large Data Science
The Transformation of Systems Biology Into A Large Data ScienceThe Transformation of Systems Biology Into A Large Data Science
The Transformation of Systems Biology Into A Large Data ScienceRobert Grossman
 
Accelerating science with Puppet
Accelerating science with PuppetAccelerating science with Puppet
Accelerating science with PuppetTim Bell
 
NRP Engagement webinar - Running a 51k GPU multi-cloud burst for MMA with Ic...
 NRP Engagement webinar - Running a 51k GPU multi-cloud burst for MMA with Ic... NRP Engagement webinar - Running a 51k GPU multi-cloud burst for MMA with Ic...
NRP Engagement webinar - Running a 51k GPU multi-cloud burst for MMA with Ic...Igor Sfiligoi
 
Overview of Photonics Research at Calit2: Scaling from Nanometers to the Earth
Overview of Photonics Research at Calit2: Scaling from Nanometers to the EarthOverview of Photonics Research at Calit2: Scaling from Nanometers to the Earth
Overview of Photonics Research at Calit2: Scaling from Nanometers to the EarthLarry Smarr
 
Hong kongopenstack2013 sdn_bluehost
Hong kongopenstack2013 sdn_bluehostHong kongopenstack2013 sdn_bluehost
Hong kongopenstack2013 sdn_bluehostJun Park
 
Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...
Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...
Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...Larry Smarr
 
The Search for Gravitational Waves
The Search for Gravitational WavesThe Search for Gravitational Waves
The Search for Gravitational Wavesinside-BigData.com
 
A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...
A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...
A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...Larry Smarr
 
Scratch to Supercomputers: Bottoms-up Build of Large-scale Computational Lens...
Scratch to Supercomputers: Bottoms-up Build of Large-scale Computational Lens...Scratch to Supercomputers: Bottoms-up Build of Large-scale Computational Lens...
Scratch to Supercomputers: Bottoms-up Build of Large-scale Computational Lens...inside-BigData.com
 
High Performance Cyberinfrastructure Enabling Data-Driven Science Supporting ...
High Performance Cyberinfrastructure Enabling Data-Driven Science Supporting ...High Performance Cyberinfrastructure Enabling Data-Driven Science Supporting ...
High Performance Cyberinfrastructure Enabling Data-Driven Science Supporting ...Larry Smarr
 

Ähnlich wie The size and complexity of the CERN network (20)

Using Photonics to Prototype the Research Campus Infrastructure of the Future...
Using Photonics to Prototype the Research Campus Infrastructure of the Future...Using Photonics to Prototype the Research Campus Infrastructure of the Future...
Using Photonics to Prototype the Research Campus Infrastructure of the Future...
 
大強子計算網格與OSS
大強子計算網格與OSS大強子計算網格與OSS
大強子計算網格與OSS
 
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...
High Performance Cyberinfrastructure Enabling Data-Driven Science in the Biom...
 
High Performance Cyberinfrastructure is Needed to Enable Data-Intensive Scien...
High Performance Cyberinfrastructure is Needed to Enable Data-Intensive Scien...High Performance Cyberinfrastructure is Needed to Enable Data-Intensive Scien...
High Performance Cyberinfrastructure is Needed to Enable Data-Intensive Scien...
 
20121205 open stack_accelerating_science_v3
20121205 open stack_accelerating_science_v320121205 open stack_accelerating_science_v3
20121205 open stack_accelerating_science_v3
 
The OptIPuter as a Prototype for CalREN-XD
The OptIPuter as a Prototype for CalREN-XDThe OptIPuter as a Prototype for CalREN-XD
The OptIPuter as a Prototype for CalREN-XD
 
A High-Performance Campus-Scale Cyberinfrastructure for Effectively Bridging ...
A High-Performance Campus-Scale Cyberinfrastructure for Effectively Bridging ...A High-Performance Campus-Scale Cyberinfrastructure for Effectively Bridging ...
A High-Performance Campus-Scale Cyberinfrastructure for Effectively Bridging ...
 
The Transformation of Systems Biology Into A Large Data Science
The Transformation of Systems Biology Into A Large Data ScienceThe Transformation of Systems Biology Into A Large Data Science
The Transformation of Systems Biology Into A Large Data Science
 
Lxcloud
LxcloudLxcloud
Lxcloud
 
Accelerating science with Puppet
Accelerating science with PuppetAccelerating science with Puppet
Accelerating science with Puppet
 
NRP Engagement webinar - Running a 51k GPU multi-cloud burst for MMA with Ic...
 NRP Engagement webinar - Running a 51k GPU multi-cloud burst for MMA with Ic... NRP Engagement webinar - Running a 51k GPU multi-cloud burst for MMA with Ic...
NRP Engagement webinar - Running a 51k GPU multi-cloud burst for MMA with Ic...
 
Overview of Photonics Research at Calit2: Scaling from Nanometers to the Earth
Overview of Photonics Research at Calit2: Scaling from Nanometers to the EarthOverview of Photonics Research at Calit2: Scaling from Nanometers to the Earth
Overview of Photonics Research at Calit2: Scaling from Nanometers to the Earth
 
Hong kongopenstack2013 sdn_bluehost
Hong kongopenstack2013 sdn_bluehostHong kongopenstack2013 sdn_bluehost
Hong kongopenstack2013 sdn_bluehost
 
Grid computing & its applications
Grid computing & its applicationsGrid computing & its applications
Grid computing & its applications
 
Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...
Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...
Project StarGate An End-to-End 10Gbps HPC to User Cyberinfrastructure ANL * C...
 
The Search for Gravitational Waves
The Search for Gravitational WavesThe Search for Gravitational Waves
The Search for Gravitational Waves
 
A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...
A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...
A Campus-Scale High Performance Cyberinfrastructure is Required for Data-Int...
 
Ns fundamentals 1
Ns fundamentals 1Ns fundamentals 1
Ns fundamentals 1
 
Scratch to Supercomputers: Bottoms-up Build of Large-scale Computational Lens...
Scratch to Supercomputers: Bottoms-up Build of Large-scale Computational Lens...Scratch to Supercomputers: Bottoms-up Build of Large-scale Computational Lens...
Scratch to Supercomputers: Bottoms-up Build of Large-scale Computational Lens...
 
High Performance Cyberinfrastructure Enabling Data-Driven Science Supporting ...
High Performance Cyberinfrastructure Enabling Data-Driven Science Supporting ...High Performance Cyberinfrastructure Enabling Data-Driven Science Supporting ...
High Performance Cyberinfrastructure Enabling Data-Driven Science Supporting ...
 

The size and complexity of the CERN network

  • 1. The CERN Network Openlab Summer 2012 CERN, 6th August 2012 edoardo.martelli@cern.ch CERN IT Department CH-1211 Genève 23 Switzerland www.cern.ch/it 1
  • 2. Summary - IT-CS - CERN networks - LHC Data Challenge - WLCG - LHCOPN and LHCONE - Openlab - Conclusions 2
  • 4. IT-CS The IT-CS group is responsible for all communication services in use at CERN for data, voice and video http://it-cs.web.cern.ch/it-cs/ 4
  • 8. High Energy Physics over IP Most of the CERN infrastructure is controlled and managed over a pervasive IP network 8
  • 9. Cryogenics 27Km of pipes at -271.11° C by means of 700.000 litres of Helium: controlled over IP Source: http://te-dep-crg-oa.web.cern.ch/te-dep-crg-oa/te-crg-oa_fichiers/cryolhc/LHC%20Cryo_BEOP_lectures2009.pdf 9
  • 10. Access control Safety and Security: made over IP Source:https://edms.cern.ch/file/931641/1/LASS-LACS_IHM.pdf 10
  • 11. Remote inspections Remote inspection of dangerous areas: robots controlled and giving feedback over WiFi and GSM IP networks 11
  • 12. DAQ: Data Acquisition A constant stream of data from the four Detectors to disk storage Source: http://aliceinfo.cern.ch/Public/Objects/Chapter2/DetectorComponents/daq_architecture.pdf 12
  • 13. CCC: CERN Control Centre The neuralgic centre of the particle accelerator: over IP 13
  • 14. CERN data network - 150 routers - 2200 Switches - 50000 connected devices - 5000km of optical fibres 14
  • 15. Network Provisioning and Management System - 250 Database tables - 100,000 Registered devices - 50,000 hits/day on web user interface - 1,000,000 lines of codes - 11 years of development 15
  • 16. Monitoring and Operations The whole network is monitored and operated by the CERN NOC (Network Operation Centre) 16
  • 17. IPv6 IPv6 dual stack network deployment on going: ready in 2013 Already available: dual-stack testbed More information: http://cern.ch/ipv6 almost 17
  • 20. Comparing theory... Simulated production of a Higgs event in ATLAS 20
  • 21. .. to real events Higgs event in CMS 21
  • 22. Data flow 3 PBytes/s 4 Experiments Filter and first selection 2 GBytes/s to the CERN computer center Create sub-samples World-Wide Analysis 1 TByte/s ? Distributed + local Physics Explanation of nature 10 GBytes/s 4 GBytes/s sΓZ2 Store on disk and tape Export copies σ _ ≈σ _× 0 with ff ff ( s - m 2 ) 2 + s 2 ΓZ / m 2 Z 2 z 12π Γ Γ GF m 3 σ 0 _ = 2 ee 2 ff and Γff = Z × ( v f2 + a f2 ) × N col ff m Z ΓZ 6π 2 22
  • 23. Data Challenge - 40 million collisions per second - After filtering, 100 collisions of interest per second - 1010 collisions recorded each year = 15 Petabytes/year of data 23
  • 25. Last months data transfers 25
  • 27. WLCG Distributed Computing Infrastructure for LHC experiments Collaborative effort of the HEP community 27
  • 28. WLCG resources WLCG sites: - 1 Tier0 (CERN) - 11 Tier1s - ~140 Tier2s - >300 Tier3s worldwide - ~250,000 CPUs - ~ 150PB of disk space 28
  • 29. CERN Tier0 resources Servers 11000 Disks 64000 Processors 15000 Raw Disk Capacity (TB) 63000 Cores 64000 Memory Modules 56000 HEPspec06 480000 RAID controllers 3750 Tape drives 160 High Speed routers 23 Tape cartridges 45000 Ethernet switches 500 Tape slots 56000 10Gbps ports 3000 Tape capacity(TB) 34000 100Gbps ports 48 March 2012 29
  • 30. CERN Tier0 LCG network Tier2/3s Tier1s CERN Campus 170G aggregated Border routers LHC Experiments 40G links Core routers 100G links Distribution routers 10G or 40G links LCG access switches Access switches ... x892 (max) 1G or 10G links Servers 30
  • 31. Trends Virtualization mobility (Software Defined Networks) Commodity Servers with 10G NICs High-end Servers with 40G NICs 40G and 100G interfaces on switches and routers 31
  • 34. A collaborative effort Designed, built and operated by the Tier0-Tier1s community Links provided by the Research and Education network providers: Geant, USLHCnet, Esnet, Canarie, ASnet, Nordunet, Surfnet, GARR, Renater, JANET.UK, Rediris, DFN, SWITCH 34
  • 35. Technology - Single and bundled long distance 10G ethernet links - Multiple redundant paths. Star+PartialMesh topology - BGP routing: communities for traffic engineering, load balancing. - Security: only declared IP prefixes can exchange traffic. 35
  • 36. Traffic to the Tier1s 36
  • 38. LHCONE LHC Open Network Environment 38
  • 39. Driving the change “The Network infrastructure is the most reliable service we have” “Network Bandwidth (rather than disk) will need to scale more with users and data volume” “Data placement will be driven by demand for analysis and not pre- placement” Ian Bird, WLCG project leader 39
  • 40. Change of computing model (ATLAS) 40
  • 41. New computing model - Better and more dynamic use of storage - Reduce the load on the Tier1s for data serving - Increase the speed to populate analysis facilities Needs for a faster, predictable, pervasive network connecting Tier1s and Tier2s 41
  • 42. Requirements from the Experiments - Connecting any pair of sites, regardless of the continent they reside - Bandwidth ranging from 1Gbps (Minimal), 5Gbps (Nominal), 10G and above (Leadership) - Scalability: sites are expected to grow - Flexibility: sites may join and leave at any time - Predictable cost: well defined cost, and not too high 42
  • 43. Needs for a better network - more bandwidth by federating (existing) resources - sharing cost of expensive resources - accessible to any TierX site = LHC Open Network Environment 43
  • 44. LHCONE concepts - Serves any LHC sites according to their needs and allowing them to grow - A collaborative effort among Research & Education Network Providers - Based on Open Exchange Points: easy to join, neutral - Multiple services: one cannot fit all - Traffic separation: no clash with other data transfer, resource allocated for and funded by HEP community 44
  • 46. LHCONE building blocks - Single node Exchange Points - Continental/regional Distributed Exchange Points - Interconnect circuits between Exchange Points These exchange points and the links in between collectively provide LHCONE services and operate under a common LHCONE policy 46
  • 48. LHCONE services - Layer3 VPN - Point-to-Point links - Monitoring 48
  • 50. Openlab project: CINBAD 50
  • 51. CINBAD CERN Investigation of Network Behaviour and Anomaly Detection Project Goals: Understand the behaviour of large computer networks (10’000+ nodes) in High Performance Computing or large Campus installations to be able to: ● detect traffic anomalies in the system ● perform trend analysis ● automatically take counter measures ● provide post-mortem analysis facilities Resources: - In collaboration with HP Networking - Two Engineers in IT-CS 51
  • 52. Results Project completed in 2010 For CERN: Designed and deployed a complete framework (hardware and software) to detect anomalies in the Campus Network (GPN) For HP: Intellectual properties of new technologies used in commercial products 52
  • 53. CINBAD Architecture data sources storage analysis collectors 53
  • 54. Openlab project: WIND 54
  • 55. WIND Wireless Infrastructure Network Deployment Project Goals - Analyze the problems of large scale wireless deployments and understand the constraint - Simulate behaviour of WLAN - Develop new optimisation algorithms Resources: - In collaboration with HP Networking - Two Engineers in IT-CS - Started in 2010 55
  • 56. Needs Wireless LAN (WLAN) deployments are problematic: ● Radio propagation is very difficult to predict ● Interference is an ever present danger ● WLANs are difficult to properly deploy ● Monitoring was not an issue when the first standards were developed ● When administrators are struggling just to operate the WLAN, performance optimisation is often forgotten 56
  • 57. Example: Radio interferences Max data rate in 0031-S: The APs work on the same channel Max data rate in 0031-S: The APs work on 3 independent channels 57
  • 58. Expected results Extend monitoring and analysis tools Act on the network - smart load balancing - isolating misbehaving clients - intelligent minimum data rates More accurate troubleshooting Streamline WLAN design 58
  • 59. Openlab project: ViSION 59
  • 60. ViSION Project Goals: - Develop a SDN traffic orchestrator using OpenFlow Resources: - In collaboration with HP Networking - Two Engineers in IT-CS - Started in 2012 60
  • 61. Goals SDN traffic orchestrator using OpenFlow: ● distribute traffic over a set of network resources ● perform classification (different types of applications and resources) ● perform load sharing (similar resources). Benefits: ● improved scalability and control than traditional networking technologies 61
  • 62. From traditional networks... Closed boxes, fully distributed protocols App App App Operating System App App App Specialized Packet Forwarding Hardware Operating System App App App Specialized Packet Forwarding Hardware Operating System App App App Specialized Packet Forwarding Hardware Operating System Specialized Packet App App App Forwarding Hardware Operating System Specialized Packet Forwarding Hardware 62
  • 63. .. to Software Defined Networks (SDN) 2. At least one good operating system 3. Well-defined open API Extensible, possibly open-source App App App Network Operating System 1. Open interface to hardware (OpenFlow) Simple Packet Forwarding Hardware Simple Packet Forwarding Hardware Simple Packet Forwarding Hardware Simple Packet Forwarding Hardware Simple Packet Forwarding Hardware 63
  • 64. OpenFlow example Controller PC Software Layer OpenFlow Client Flow Table MAC MAC IP IP TCP TCP Action Hardware Forwarding table remotely src dst Src Dst sport dport Hardware Forwarding table remotely Hardware controlled controlled * * * 5.6.7.8 * * port 1 Layer port 1 port 2 port 3 port 4 64
  • 66. Conclusions - The Data Network is an essential component of the LHC instrument - The Data Network is a key part of the LHC data processing and will become even more important - More and more security and design challenges to come 66
  • 67. Credits Artur Barczyk (LHCONE) Dan Savu (VISION) Milosz Hulboj (WIND and CINBAD) Ryszrard Jurga (CINBAD) Sebastien Ceuterickx (WIND) Stefan Stancu (VISION) Vlad Lapadatescu (WIND) 67
  • 68. What's next SWAN: Space Wide Area Network :-) 68