The document discusses Netvisor, a network hypervisor developed by Pluribus Networks. Netvisor aims to simplify network management by running as a software layer on top of switches and providing visibility into traffic flows across the entire network. It addresses challenges with traditional networks, which are complex to manage and secure due to multiple protocols, devices, and lack of end-to-end visibility. Netvisor provides an open and interoperable approach intended to improve agility, security, and troubleshooting without increasing costs exponentially as networks scale physically and virtually.
1. Confiden'al
Netvisor - “The” Network Hypervisor
Pluribus Networks NetvisorOS
Open Fabrics, Analytics, and
Virtualizations
Robert Drost, Ph.D.
Founder and COO
2. Pluribus
Networks
–
What
Do
We
Do?
A
Network
(Hypervisor)
OS:
Netvisor !
Designed to build the easiest
to
use
networks
Leaf
and
Spine
Switches
Turn-‐key
HW+SW ! SW on ODM
switches!
3. Is
The
Life
of
Netops
GeIng
Any
Easier?
! complex
155. Netvisor
Server-‐Style
Distributed
Architecture
“A
computer
cluster
consists
of
a
set
of
loosely
or
'ghtly
connected
computers
that
work
together
so
that,
in
many
respects,
they
can
be
viewed
as
a
single
system”
Servers
323. !
!
Netvisor Fabric-Cluster For Simplified Mgmt!
!
!
L2!
L3!
HA block!
Distributed peer-to-peer
HA cluster. No single point
of failure. Node hot-plug.!
Classic database !
3-phase commit for
config changes!
One logical switch
to manage!
–!
Every Node shares the
same view of the Fabric:
MAC, IP, connections
and app flows!
A distributed architecture based on a collection of
compute clustering techniques to present !
an open, standard-based Ethernet fabric as !
one logical switch.!
Fabric-wide ARP Suppression!
324. Customer
Example:
US
Manufacturing
Company
Fabric
VirtualizaWon
For
Private
Cloud
Pluribus
Architectural
Value:
! Leaf-‐Spine
Fabric
simplificaWon
! Virtual
network
with
HW
vRouters
! Analy'cs
w/
forensic
recording
! Enable
inser'on
of
new
services
w/o
HW
sprawl
(e.g.
For'net
FW)
Requirements:
! Consolida'on
of
legacy
mul'-‐layer
design
to
leaf-‐spine
! Segmenta'on/mul'-‐tenancy
w/o
dedicated
HW
(routers,
services)
! E-‐W
flow
visibility,
compliance,
audi'ng
383. Tracing
CongesWon
Along
The
Path
Of
A
Flow
./flowtrace_new.py
-‐n
aquila02
-‐c
10.9.18.249
-‐s
10.9.9.73
==================================vport
info===============================
The
packet
enters
fabric
through
pn-‐dev01
port
15
The
packet
leaves
pn-‐dev01
through
port
63
and
enters
aquila02
through
port
10
The
packet
leaves
aquila02
through
port
129(36,40)
and
enters
spine02
through
ports
4,5
The
packet
leaves
fabric
through
spine02
port
17
pn-‐dev01
aquila02
spine02
15
63
Visibility
into
conges'on
sta's'cs
on
each
port
along
the
path
of
the
applica'on
flows
36,40
4,5
17
10
384. Top
Talkers
App
Client
Server
Loca'on
Latency
Top
Talkers
Server
Dura'on
Device
Manufacturer
385. Built-‐in
PCAP
Capture
Of
Any
Flow
In
The
Fabric
admin@S68-‐Leaf1:~#
snoop
-‐v
-‐d
igb0
Using
device
igb0
(promiscuous
mode)
ETHER:
-‐-‐-‐-‐-‐
Ether
Header
-‐-‐-‐-‐-‐
ETHER:
ETHER:
Packet
1
arrived
at
18:25:15.03827
ETHER:
Packet
size
=
66
bytes
ETHER:
Destination
=
64:e:94:30:2:4b,
ETHER:
Source
=
78:da:6e:65:1d:58,
ETHER:
Ethertype
=
0800
(IP)
ETHER:
IP:
-‐-‐-‐-‐-‐
IP
Header
-‐-‐-‐-‐-‐
IP:
IP:
Version
=
4
IP:
Header
length
=
20
bytes
IP:
Type
of
service
=
0x00
IP:
xxx.
....
=
0
(precedence)
snip
IP:
Header
checksum
=
2ba6
IP:
Source
address
=
50.203.11.18,
50-‐203-‐11-‐18-‐
static.hfc.comcastbusiness.net
IP:
Destination
address
=
192.168.1.53,
EBC-‐Leaf-‐1
IP:
No
options
IP:
TCP:
-‐-‐-‐-‐-‐
TCP
Header
-‐-‐-‐-‐-‐
TCP:
TCP:
Source
port
=
63469
TCP:
Destination
port
=
22
TCP:
Sequence
number
=
3984394763
TCP:
Acknowledgement
number
=
3169545619
TCP:
Data
offset
=
32
bytes
TCP:
Flags
=
0x10
TCP:
0...
....
=
No
ECN
congestion
window
reduced
TCP:
.0..
....
=
No
ECN
echo
TCP:
..0.
....
=
No
urgent
pointer
TCP:
...1
....
=
Acknowledgement
TCP:
....
0...
=
No
push
Snip
395. Customer
Example:
US
Insurance
Company
Large
Scale
BGP
Fabric
For
Big
Data
ApplicaWon
Pluribus
Architectural
Value:
! Fabric
management
simplificaWon
! Monitoring
and
visibility
of
Hadoop
cluster
performance
Requirements:
! BGP
fabric
with
144
racks
! Economics
! Interop
w/
third
party
modular
spine
switch
with
rapid
failover
396. Integrated
AnalyWcs
Use
Cases:
Financial
Company
Fabric
AnalyWcs
For
AudiWng,
Compliance
Pluribus
Architectural
Value:
! Fabric
management
simplificaWon
! E-‐W
flow
analy'cs
w/
forensic
recording
and
inline
PCAP
capture
! Granular
flow
control
(vflow)
Requirements:
! Three
Data
Centers
Chicago-‐New
York-‐London
! Leaf-‐Spine
architecture
! Audi'ng,
compliance,
forensic
flow
visibility
421. Unified
Overlay
and
Underlay
w/
Switch
VTEP
Off-‐load
VXLAN
over
BGP
fabric
or
VXLAN
for
L2
POD
extension
! Netvisor
SDN
Fabric
to
orchestrate
and
automate
VXLAN
tunnel
offload
on
Leaf
switches
! Keep
server
networking
simple
and
eliminate
VXLAN
encap/decap
performance
tax
! ONE
fabric
with
seamless
overlay-‐
underlay
(VXLAN)
visibility
!
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
L3#
L2#$#VLAG# L2#$#VLAG# L2#$#VLAG# L2#$#VLAG#
IP/ECMP'Fabric'
!
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
L3#
L2#$#VLAG# L2#$#VLAG# L2#$#VLAG# L2#$#VLAG#
IP/ECMP'Fabric'
VXLAN
Tunnel
POD
#1
POD
#2
L3
CORE
!
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
L3#
L2#$#VLAG# L2#$#VLAG# L2#$#VLAG# L2#$#VLAG#
IP/ECMP'Fabric'
VXLAN
Tunnel
100K
VMs
• 95%
line
rate@10Gbps
• 0%
packet
drop
• Convergence:
• Cold:
25.54s
• Warm:
7.35s
422. L3
CORE
route
route
route
route
route
route
route
route
route
route
route
route
route
route
route
POD
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
POD
POD
POD
TradiWonal
Server
Overlays
Centralized
Controller
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
L3
CORE
423. POD
POD
POD
Netvisor
Dynamic
Switch
Overlays
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
VM
① Dynamic
switch-‐based
VXLAN
encap/decap
② Reduce
#
of
tunnels
(e.g.
~3K+!12)
and
remove
servers
encap/decap
Distributed
Cluster
w/
Overlay
OrchestraOon
L3
CORE
POD
424. Netvisor
OFV
Architecture
Unify
Overlay
Underlay
!
!
!
L3!
VXLAN !
Tunnel endpoint!
VXLAN! Eliminate host perf. penalty!
(as high as 65% drop) by off-loading
tunneling to switch ASIC @ line rate!
0!
1!
2!
3!
4!
5!
6!
7!
8!
9!
10!
64Bytes! 512Bytes! 1450Bytes!
Gbps!
OVS to OVS performance!
Baseline!
Switch Overlays!
Server Overlays!
Fabric-based Overlay tunnel
orchestration and segmentation
one fabric for physical and virtual!
Uncompromised physical/virtual
visibility!
425. Netvisor
Fabric
MulW-‐tenancy
And
SegmentaWon
! Rapid
provisioning
of
secure
virtual
networks
(aka
VNETs)
with
management,
control
and
data
plane
isola'on
! Provision
in
minutes
per
tenant
virtual
routers
with
management,
control,
data
plane
isola'on
! Per
tenant
virtual
services:
vSLB,
vFW,
vCLI,
DNS,
DHCP,
Pixie,
NTP…
! Per
tenant
visibility
of
flows,
services,
VMs…
VLAN%10(12% VLAN%20(22% VLAN%30(32%
426. App
1
App
100
Customer
Example:
US
Financial
Company
Fabric
VirtualizaWon
For
Cloud
App.
Delivery
PN
Arch.
Value:
! Consolidate
100
vRouters
under
netops
control
on
2
devices
! VNET
for
app
developers
control
! Analy'cs
to
monitor
applicaWon
performance
Requirements:
! Per
tenant/app
virtual
routers
for
BGP
peering
with
AWS
vRouters
! Visibility
and
ease
of
troubleshoo'ng
cloud
connec'on
! Allow
each
tenant
to
manage
and
monitor
its
own
router/virtual
network
App
2
445. High-‐Availability
! Peer-‐to-‐peer
highly
available
cluster
technology
for
the
fabric
! Distributed,
no
central
controller,
single
point
of
failure
! Most
fabric
opera'ons
as
atomic
transac'ons,
either
they
succeed
or
fail
across
the
en're
fabric
cluster
! Networking:
! Mul'-‐chassis
LAG
with
sub
200
ms
failover
! Ac've-‐Ac've
VRRP
with
sub
200
ms
failover
(no
VRRP
'mer
dependency)
! BFD
for
BGP
for
sub-‐second
failover
! ISSU
across
the
en're
fabric
of
switches
(rolling
upgrade)
446. ! Netvisor
drivers
support
ML2
based
neutron
plugin
qualified
for
Juno/Icehouse.
The
plugin
repo
is
publicly
available
at
hkps://github.com/PluribusNetworks/pluribus_neutron/tree/master/neutron-‐plugin-‐pluribus
! The
ML2
plugin
supports:
! Na've
neutron
APIs
for
logical
networks/subnets/port
management
! L3
agent
APIs
for
logical
router
management
! NAT
capability
on
the
switches
! Loadbalancer
API
support
! In
addi'on
the
ML2
plugin
supports
Netvisor
specific
advanced
extensions:
! HW
based
rouWng
! High
availability
(VRRP)
for
logical
routers
! NFS
backed
store
for
glance
to
deliver
sta'c
image
content
! Host
analyWcs
/visibility
into
tenant
traffic
! Flow
programming
(vflow)
across
the
fabric
(vflow)
! Full
PCAP
vflow
packet
capture
! Per-‐tenant
QOS/SLA
! Tracking
VM
lifecycle
using
virtual
ports
concept
on
Netvisor
Netvisor
OpenStack
Highlights
447. Summary – Pluribus Netvisor Differentiation
Fabric-wide !
Multi-box!
management!
Simple to Manage! Simple to Monitor!
Fabric-wide!
Application and VM
Visibility/Analytics!
w/ Time Machine!
Fabric-wide!
Programmability,!
NFV, Flow !
Programmability!
Simple to Program!
And Secure!
Simple to Virtualize!
Fabric-wide!
Unified Overlay/
Underlay +!
Penalty-free VXLAN
Tunnel Orchestration!!
Netvisor Server-Style Distributed Cluster ! ONE “Touch Point” For The Entire Fabric!