nanog mailing list archives

Re: 400G forwarding - how does it work?


From: Jeff Tantsura <jefftant.ietf () gmail com>
Date: Wed, 10 Aug 2022 11:38:48 -0700

Sharada’s answers:

a) Yes, the run-to-completion model of Trio is superior to FP5/Nokia model when it comes to flexible processing 
engines. In Trio, the same engines can do either ingress or egress processing. Traditionally, there is more processing 
on ingress than on egress. When that happens, by design, less number of processing engines get used for egress, and 
more engines are available for ingress processing. Trio gives full flexibility. Unless Nokia is optimizing the engines 
(not all engines are identical, and some are area optimized for specific processing) to save overall area, I do not see 
any other advantage.  

b) Trio provides on-chip shallow buffering on ingress for fabric queues. We share this buffer between the slices on the 
same die. This gives us the flexibility to go easy on the size of SRAM we want to support for buffering. 

c) I didn't completely follow the question. Shallow ingress buffers are for fabric-facing queues, and we do not expect 
sustained fabric congestion. This, combined with the fact that we have some speed up over fabric, ensures that all WAN 
packets do reach the egress PFE buffer. On ingress, if packet processing is oversubscribed, we have line rate 
pre-classifiers do proper drops based on WAN queue priority.

Cheers,
Jeff

On Aug 9, 2022, at 16:34, Jeff Tantsura <jefftant.ietf () gmail com> wrote:


Saku,
 
I have forwarded your questions to Sharada.
 
All,
 
For this week – at 11:00am PST, Thursday 08/11, we will be joined by Guy Caspary (co-founder of Leaba Semiconductor 
(acquired by Cisco -> SiliconOne)
https://m.youtube.com/watch?v=GDthnCj31_Y
 
For the next week, I’m planning to get one of main architects of Broadcom DNX  (Jericho/Qumran/Ramon).
 
Cheers,
Jeff
 
From: Saku Ytti
Sent: Friday, August 5, 2022 12:15 AM
To: Jeff Tantsura
Cc: NANOG; Jeff Doyle
Subject: Re: 400G forwarding - how does it work?
 
Thank you for this.
 
I wish there would have been a deeper dive to the lookup side. My open questions
 
a) Trio model of packet stays in single PPE until done vs. FP model of
line-of-PPE (identical cores). I don't understand the advantages of
the FP model, the Trio model advantages are clear to me. Obviously the
FP model has to have some advantages, what are they?
 
b) What exactly are the gains of putting two trios on-package in
Trio6, there is no local-switching between WANs of trios in-package,
they are, as far as I can tell, ships in the night, packets between
trios go via fabric, as they would with separate Trios. I can
understand the benefit of putting trio and HBM2 on the same package,
to reduce distance so wattage goes down or frequency goes up.
 
c) What evolution they are thinking for the shallow ingress buffers
for Trio6. The collateral damage potential is significant, because WAN
which asks most, gets most, instead each having their fair share, thus
potentially arbitrarily low rate WAN ingress might not get access to
ingress buffer causing drop. Would it be practical in terms of
wattage/area to add some sort of preQoS towards the shallow ingress
buffer, so each WAN ingress has a fair guaranteed-rate to shallow
buffers?
 
On Fri, 5 Aug 2022 at 02:18, Jeff Tantsura <jefftant.ietf () gmail com> wrote:

Apologies for garbage/HTMLed email, not sure what happened (thanks
Brian F for letting me know).
Anyway, the podcast with Juniper (mostly around Trio/Express) has been
broadcasted today and is available at
https://www.youtube.com/watch?v=1he8GjDBq9g
Next in the pipeline are:
Cisco SiliconOne
Broadcom DNX (Jericho/Qumran/Ramon)
For both - the guests are main architects of the silicon

Enjoy


On Wed, Aug 3, 2022 at 5:06 PM Jeff Tantsura <jefftant.ietf () gmail com> wrote:

Hey,



This is not an advertisement but an attempt to help folks to better understand networking HW.



Some of you might know (and love 😊) “between 0x2 nerds” podcast Jeff Doyle and I have been hosting for a couple 
of years.



Following up the discussion we have decided to dedicate a number of upcoming podcasts to networking HW, the topic 
where more information and better education is very much needed (no, you won’t have to sign NDA before joining 
😊), we have lined up a number of great guests, people who design and build ASICs and can talk firsthand about 
evolution of networking HW, complexity of the process, differences between fixed and programmable pipelines, 
memories and databases. This Thursday (08/04) at 11:00PST we are joined by one and only Sharada Yeluri - Sr. 
Director ASIC at Juniper. Other vendors will be joining in the later episodes, usual rules apply – no marketing, 
no BS.

More to come, stay tuned.

Live feed: https://lnkd.in/gk2x2ezZ

Between 0x2 nerds playlist, videos will be published to: 
https://www.youtube.com/playlist?list=PLMYH1xDLIabuZCr1Yeoo39enogPA2yJB7



Cheers,

Jeff



From: James Bensley
Sent: Wednesday, July 27, 2022 12:53 PM
To: Lawrence Wobker; NANOG
Subject: Re: 400G forwarding - how does it work?



On Tue, 26 Jul 2022 at 21:39, Lawrence Wobker <ljwobker () gmail com> wrote:

So if this pipeline can do 1.25 billion PPS and I want to be able to forward 10BPPS, I can build a chip that 
has 8 of these pipelines and get my performance target that way.  I could also build a "pipeline" that 
processes multiple packets per clock, if I have one that does 2 packets/clock then I only need 4 of said 
pipelines.. and so on and so forth.



Thanks for the response Lawrence.



The Broadcom BCM16K KBP has a clock speed of 1.2Ghz, so I expect the

J2 to have something similar (as someone already mentioned, most chips

I've seen are in the 1-1.5Ghz range), so in this case "only" 2

pipelines would be needed to maintain the headline 2Bpps rate of the

J2, or even just 1 if they have managed to squeeze out two packets per

cycle through parallelisation within the pipeline.



Cheers,

James.


 
 
 
--
  ++ytti
 

Current thread: