Difference between revisions of "EJFAT EPSCI Meeting Jan. 17, 2024"

From epsciwiki
Jump to navigation Jump to search
(Created page with "The meeting time is 2:30pm. === Connection Info: === <div class="toccolours mw-collapsible mw-collapsed"> You can connect using [https://teams.microsoft.com/l/meetup-join/19%...")
 
 
(One intermediate revision by the same user not shown)
Line 62: Line 62:
 
# EJFAT Phase II
 
# EJFAT Phase II
 
##  '''Priority - work EJFAT Reconfig Design / PR'''
 
##  '''Priority - work EJFAT Reconfig Design / PR'''
 +
### [https://docs.nvidia.com/networking/display/connectx6vpi PO for PCIe NICS]
 +
### [https://network.nvidia.com/sites/default/files/doc-2020/ocp-vpi-adapter-cards-brochure.pdf OCP NIC]
 
##  Implementation details in the DAOS gateway.   
 
##  Implementation details in the DAOS gateway.   
 
### [https://app.slack.com/client/T4RUE2FDH/C4SM0RZ54 Intel standing up special slack channel to discuss DAOS]
 
### [https://app.slack.com/client/T4RUE2FDH/C4SM0RZ54 Intel standing up special slack channel to discuss DAOS]

Latest revision as of 13:43, 18 January 2024

The meeting time is 2:30pm.

Connection Info:


Agenda:

  1. Previous meeting
  2. Announcements:
    1. 24th IEEE Real Time Conference – Quy Nhon, Vietnam 22-26 April 2024 - Abstract Submissitted
    2. The first 100Gig circuit passed testing and ready for traffic - need to sync up with ESNet
  3. NERSC Test Development:
    1. Data Source:
      1. JLAB, CLAS12, pre-triggered events - 1 channel
      2. Front End Packetizer pending mods for Tick-sync msg to CP - UDP packet to port on CP Host
    2. Data Sink:
      1. Perlmutter
      2. ERSAP
    3. Networking for Test
      1. Currently 2 x 10 Gbps for JLab/L3 VPN - will expand to 200Gbps in near fututure
    4. JLab Preps
    5. Test Plans - JLab, ESnet, NERSC:
  4. Test with Oak Ridge (similar to NERSC) - Shankar, Mallikarjun (Arjun) <shankarm@ornl.gov>
  5. Hall B CLAS12 detector streaming test
    1. Switch 7050 is expected to arrive some time around October; we have already transceivers, short cables and patch panel to connect up to 32 VTPs to it using two 10GBit links per VTP
    2. Fiber installation between hallb forward carriage and hallb counting room should be done this summer, will be enough for 24 VTPs using two 10GBit links per VTP
    3. We have only one fiber between hallb counting room and counting house second floor available right now, will order more fibers installation, may take several months
    4. There are several available fibers between counting house second floor and computer center (like 6), we can use a couple of them for our test
    5. Summary: sometime in October, we should have 48 10GBit links from 24 VTPs connected to the switch in hallb counting room, with that switch connected to computer center by 2x100GBit links
    6. Need to develop CONOPS with Streaming group (Abbott)
    7. December 2023 Testing Activity
    8. SRO RTDP LDRD - need configuration for Spring '24 test with EJFAT
    9. Data Compressibility Studies using Hall B/D sample data
    10. Ready to supply up to 200 Gbps to EJFAT switch
  6. Demo Ready EJFAT Instance
    1. Live Session?
      1. ESNet had a demo at SC23 on EJFAT LB. (Yatish)
    2. Emulator ?
    3. Recorded Session
  7. EJFAT Operational Status Board -> Prometheus Reporting
  8. XDP sockets working (ejfat-4) - 50% less cpu, 3500 MTU limit
    1. Need to spec JLab/HPDF DAOS Use Cases
  9. EJFAT Phase II
    1. Priority - work EJFAT Reconfig Design / PR
      1. PO for PCIe NICS
      2. OCP NIC
    2. Implementation details in the DAOS gateway.
      1. Intel standing up special slack channel to discuss DAOS
      2. Connection Strategy to DAOS
      3. Specially when to keep track of how the FPGA would DMA event data cells in the future if it was a SmartNIC card. ( Cissie )
      4. daosfs01​ has 2 physical IB cards and can run 2 true engines with each CPU socket hosting one engine.
    3. Progress of multi FPGA and multi virtual LB control plane sw. ( Derek )
    4. Progress of FPGA architecture ( Peter and Jonathan )
      1. LB FW currently limited to 100 Gbps
      2. Reassembly work commencing soon
    5. Progress of finalizing a reassembly frame format (subordinate to 4.) ( Carl / Stacey )
    6. Progress on software development for NVIDIA Bluefield2 DPU data steering from NIC to GPU memory ( Amitoj/Cissie )
    7. GPU purchase (A100) for EJFAT Test stand servers under IRIAD funds. The servers are capable of hosting 2 GPUs per server. ( Amitoj )
      1. In a pinch one can use the 2x A100 GPUs in the NVIDIA Bluefield2 DPU server (hostname: nvidarm)
      2. Initially purchase one of each GPU: NVIDIA, INTEL, AMD so we can compare performance across all 3 flavors.
      3. Tracking Code using GPUs from Available from Hall-B
      4. GPU PCIe board requires freeing PCIe slot - looking at OTP NIC
  10. Resources:
    1. HPDF
  11. AOT