LIVE from GTC: DGX Spark Insides First Look

00:12:38
https://www.youtube.com/watch?v=AOL0RIZxJF0

Zusammenfassung

TLDRIntervjuet diskuterer DJX Spark fra Nvidia, en mini AI supercomputer designet for utviklere, som starter på 3000 dollar. Modellen fungerer som et utviklingsverktøy som tillater programmering og testing av AI-programmerering i en kraftig, men kompakt form. Forskjellene mellom modellene inkluderer lagringskapasitet, der dyreste modell leveres med 4 TB, og en nettverksløsning av høy kvalitet med 200 Gbit/s Ethernet porter. DJX Spark tilbyr også god minneytelse med 128 GB DDR5X minne. Den er beregnet for mer seriøs utvikling og lar brukere kjøre komplekse AI-modeller.

Mitbringsel

  • 💻 DJX Spark er en mini AI supercomputer from Nvidia.
  • 💡 Startpris er 3000 dollar, med oppgraderinger opp til 4000 dollar.
  • 📊 Den dyrere modellen har 4 TB lagringsplass.
  • 🔧 Utviklersenter for AI som lar programmer sofrie uten justering.
  • 🌐 Utstyret har 200 Gbit/s nettverksport.
  • 👨‍💻 Primært for profesjonelle utviklere, ikke hobbyister.
  • 📦 Leveres med DJXOS basert på Ubuntu 22.04 LTS.
  • 🔌 128 GB DDR5X minne er standard for alle modeller.
  • 🤖 Kan håndtere komplekse AI-prosjekter som finjustering av språkmodeller.
  • 🏭 Partnerselskap inkluderer Asus, HP, Dell, og Lenovo.

Zeitleiste

  • 00:00:00 - 00:05:00

    Israel, som jobber med teknisk markedsføring for DJX Spark, forklarer at produktet er en mini AI-superdatamaskin som er designet for utviklere. Den er tilgjengelig i to modeller, med en grunnpris på 3000 dollar og en oppgradering til 4000 dollar, med forskjeller i lagringskapasitet. 3000-modellen har 1 TB lagringsminne, mens 4000-modellen kommer med 4 TB. Hovedmålet er å gi utviklere en plattform som reflekterer de større data sentrene, slik at koden kan overføres enkelt til produksjon i større miljøer.

  • 00:05:00 - 00:12:38

    Produktene er designet for seriøst AI-arbeid, og ikke bare hobbybrukere. Det er et bredt spekter av bruksområder, inkludert ynsning av 32 milliard store LLM-er. En klar verdi bak produktet er evnen til å utvikle og teste kode lokalt, og deretter implementere den uten endringer i strømmen til storskala data sentre. Israel forklarer også at produktet har betydelige teknologiske spesifikasjoner, inkludert kraftige nettverksmuligheter, som 200 Gbit Ethernet, og er basert på Ubuntu, tilpasset DriveXOS, med høy ytelse for utviklere av AI-løsninger.

Mind Map

Video-Fragen und Antworten

  • Hva er DJX Spark?

    Det er en mini AI supercomputer designet for utviklere.

  • Hva er prisene for DJX Spark?

    Den starter på 3000 dollar og går opp til 4000 dollar.

  • Hva er forskjellen mellom modellene?

    Den billigere modellen har 1TB lagring, mens den dyrere modellen har 4TB lagring.

  • Hvilken programvare leveres med DJX Spark?

    Den leveres med DJXOS, som er basert på Ubuntu 22.04 LTS.

  • Hvilket nettverksutstyr har DJX Spark?

    Den har ConnectX-7 med 200 Gbit/s per port.

  • Kan man oppgradere minnet?

    Det er 128 GB DDR5X minne som følger med alle modeller, men oppgraderbare alternativer er ikke spesifisert.

  • Hvem kan bruke DJX Spark?

    Primært utviklere som ønsker å lage AI-modeller og ikke hobbyister.

  • Når blir DJX Spark tilgjengelig?

    Den er tilgjengelig nå for venteliste.

  • Hvilke selskaper tilbyr DJX Spark?

    Asus, HP, Dell og Lenovo er partnere som tilbyr modellen.

  • Hvilken type prosjekter kan man bruke DJX Spark til?

    Den kan brukes til tunge AI-prosjekter, som finjustering av store språkmodeller.

Weitere Video-Zusammenfassungen anzeigen

Erhalten Sie sofortigen Zugang zu kostenlosen YouTube-Videozusammenfassungen, die von AI unterstützt werden!
Untertitel
en
Automatisches Blättern:
  • 00:00:03
    uh hi i'm here with uh Israel from
  • 00:00:05
    Nvidia welcome um what is your role on
  • 00:00:08
    the DJX Spark well I'm I'm doing the
  • 00:00:10
    tech marketing for it so I'm the person
  • 00:00:13
    that bridges the technical teams and the
  • 00:00:16
    marketing teams so I know enough about
  • 00:00:20
    the hardware but also in a way that it
  • 00:00:22
    can be uh digested I guess by Yeah by
  • 00:00:26
    the non- tech people yeah so maybe let's
  • 00:00:29
    talk about that right this is a $3,000
  • 00:00:31
    device uh starting starting starting at
  • 00:00:33
    3,000 wait what does it go up to 4,000
  • 00:00:36
    okay so like just extra upgrades of
  • 00:00:37
    course that might be a good first
  • 00:00:38
    question for you right what what's the
  • 00:00:40
    difference between the 3000 and the
  • 00:00:42
    4,000 model sure i don't I don't know
  • 00:00:44
    actually just extra memory storage
  • 00:00:46
    four four terabytes on the four uh on
  • 00:00:49
    the $4,000 one and one TB on the $3,000
  • 00:00:53
    and what uh part of the market is this
  • 00:00:55
    designed for like do you have a name for
  • 00:00:57
    like I guess proumer yeah it's we call
  • 00:01:00
    it a a mini AI supercomputer right it is
  • 00:01:05
    it is really in many ways a miniaturized
  • 00:01:08
    version of these things you see behind
  • 00:01:11
    me here yeah the real data center
  • 00:01:12
    equipment so the idea is you're taking a
  • 00:01:15
    a nibble of the that data center to your
  • 00:01:18
    home so you can develop uh this is
  • 00:01:21
    really a developer box to focus on on
  • 00:01:25
    individual developers so they can do
  • 00:01:28
    work on a platform that is really
  • 00:01:32
    100% derived from from these right the
  • 00:01:36
    architecture the CPU instruction set
  • 00:01:38
    software stack yeah the drivers the
  • 00:01:41
    network
  • 00:01:42
    component networking accelerations
  • 00:01:46
    um even the cross memory right that that
  • 00:01:49
    we have here the shared memory uh layout
  • 00:01:52
    that that's the machine that you would
  • 00:01:54
    find on a Grace Blackwell data center
  • 00:01:57
    machine
  • 00:01:58
    and you can once you you learn how to
  • 00:02:02
    develop here you're and you write code
  • 00:02:06
    here this code is ready to go there into
  • 00:02:09
    production to large scale deployment
  • 00:02:12
    without having to worry about is this
  • 00:02:14
    going to work you know is is my stack
  • 00:02:16
    ready for for that no it's the same
  • 00:02:18
    stack so is this for how much is is
  • 00:02:21
    hobbyist usage and how much is that kind
  • 00:02:24
    of workflow where they're actually
  • 00:02:25
    prototyping for the big data centers uh
  • 00:02:27
    honestly the the amount of things you
  • 00:02:29
    can do here I wouldn't call it a
  • 00:02:31
    hobbyist anymore because you you can do
  • 00:02:33
    uh some really serious AI uh work
  • 00:02:37
    finetuning which is very demanding right
  • 00:02:39
    and this is one of our key use cases
  • 00:02:42
    when we developed this product so uh
  • 00:02:46
    this is yeah I would I would say this is
  • 00:02:48
    one step further from a hobus and and
  • 00:02:51
    today there's so much you can do in in
  • 00:02:54
    terms of bit real business right um over
  • 00:02:57
    say a 32 billion LLM right just by
  • 00:03:00
    adjusting that to a certain demand using
  • 00:03:03
    fine-tuning that's already a product in
  • 00:03:05
    many ways now the only thing that you
  • 00:03:07
    cannot do here compared to these things
  • 00:03:09
    is yeah this is not designed to serve I
  • 00:03:12
    bet yeah 500,000 client at the same time
  • 00:03:15
    right so um you do this at home you test
  • 00:03:18
    and you and you run or even on a small
  • 00:03:21
    company and once you your stack is good
  • 00:03:24
    this can be pushed out to a a larger reg
  • 00:03:27
    either on prem or or a cloud resource
  • 00:03:30
    that will run and like I said the the
  • 00:03:34
    beauty is you don't have to tweak your
  • 00:03:37
    code it's it's just going to move over
  • 00:03:40
    to the next step right whether that is
  • 00:03:42
    the workstation that we have there or a
  • 00:03:45
    real data center uh grade equipment yeah
  • 00:03:48
    I was actually kind of wondering so I
  • 00:03:50
    know about the Jetsson Nano which you
  • 00:03:52
    released a few months ago um and then we
  • 00:03:54
    have the 5090s over there um what is the
  • 00:03:57
    sort of per the the the product suite
  • 00:04:00
    like how would you sort of grade it from
  • 00:04:03
    like the small end to the high end yeah
  • 00:04:05
    this is a little hard to compare with
  • 00:04:06
    Jetson because it's all new right this
  • 00:04:10
    is a Grace Black wall so all this all
  • 00:04:14
    the silicon here is uh newer than than
  • 00:04:16
    what you would find um but the the idea
  • 00:04:21
    is that this would be your starting
  • 00:04:23
    point for
  • 00:04:25
    Go ahead oh okay all right he's he's
  • 00:04:26
    doing it right okay cool sorry keep
  • 00:04:29
    going yeah the idea is Yeah you this is
  • 00:04:32
    really your starting point as a as a
  • 00:04:35
    developer for for the Grace Blackwell
  • 00:04:37
    architecture right
  • 00:04:39
    um we we will continue to support
  • 00:04:43
    different ranges of products for
  • 00:04:46
    different types of development right so
  • 00:04:48
    we have the IGX is more oriented towards
  • 00:04:51
    enterprise work and
  • 00:04:53
    um it's it's more of a professional tool
  • 00:04:56
    right than than this and this is really
  • 00:04:58
    more oriented towards a home user one
  • 00:05:01
    feature that you find here that
  • 00:05:04
    it's here we'll I'm I cannot wait for
  • 00:05:07
    this to reach homes and people start
  • 00:05:10
    playing with but this is also a a 5000
  • 00:05:14
    series Blackwell in many ways right it
  • 00:05:16
    has DLSS it has RTX and it has a very
  • 00:05:19
    capable display output subsystem and
  • 00:05:22
    right over here right so there are
  • 00:05:24
    things you can do here that we're not
  • 00:05:26
    even touching just yet um in terms of of
  • 00:05:29
    marketing right because for now we're
  • 00:05:30
    focusing more like ML and AI right but
  • 00:05:35
    um this is still a very capable home
  • 00:05:38
    computer well yeah gaming comes with
  • 00:05:41
    with you know we're this ships with
  • 00:05:43
    Linux right so that's your you start
  • 00:05:46
    from there and it's an ARM machine so
  • 00:05:49
    code right it has to run and it it it's
  • 00:05:53
    not today you know the the Linux game
  • 00:05:55
    and on ARM is a still like a this moving
  • 00:05:59
    piece but uh we've seen effort from
  • 00:06:01
    Valve right there are companies that are
  • 00:06:03
    working on on on that but I mean this
  • 00:06:06
    was not designed for it but sure still
  • 00:06:08
    you know capable telling you what's in
  • 00:06:10
    here silicon wise and I'm pretty sure
  • 00:06:13
    people will find interesting uses for
  • 00:06:14
    that as well yeah so um I mean this is
  • 00:06:17
    the first time like usually we see these
  • 00:06:18
    boxes this is the first time you sort of
  • 00:06:19
    open up the that I've seen sort of this
  • 00:06:22
    open up anything interesting that we
  • 00:06:23
    should look at this is literally the
  • 00:06:25
    first time we're showing this board yeah
  • 00:06:27
    so the board the board layout is
  • 00:06:28
    actually very simple okay so on the top
  • 00:06:30
    here we have the GB10 SOC
  • 00:06:33
    uh that we built in partnership with
  • 00:06:34
    MediaTek mhm around it you have the
  • 00:06:38
    LPDDR5X memory modules for your shared
  • 00:06:41
    128 gigabit gigabyte memory space that
  • 00:06:45
    these two chips can reach out to without
  • 00:06:48
    having to create copies of the memory
  • 00:06:50
    content right so that alone should give
  • 00:06:53
    you an enormous advantage on certain
  • 00:06:55
    workloads that you have a lot of transit
  • 00:06:57
    between your your RAM and your
  • 00:07:00
    VRAM compared to say even a powerful
  • 00:07:03
    workstation but that you're still using
  • 00:07:06
    a regular PCI Express card right so it's
  • 00:07:08
    DDR5
  • 00:07:10
    uh yeah it's LP LP because uh DDR this
  • 00:07:13
    is more like a laptop memory in the
  • 00:07:15
    electrical sense than uh so it's low
  • 00:07:17
    power DDR
  • 00:07:18
    but the the performance is great it's
  • 00:07:20
    actually a little faster than a socketed
  • 00:07:22
    um dim comparable with like using the
  • 00:07:25
    same type of uh chip right right right
  • 00:07:28
    and we have the
  • 00:07:29
    C2C uh interconnect that's the tech this
  • 00:07:32
    is Nvidia right it's our design that
  • 00:07:37
    connects the GPU to the CPU and allows
  • 00:07:40
    for we're estimating there's somewhere
  • 00:07:42
    in five times faster than PC express
  • 00:07:45
    communications here between these two
  • 00:07:46
    chips so they share the same memory
  • 00:07:48
    controller and everything right yeah uh
  • 00:07:51
    the there there's the the memory
  • 00:07:53
    controller is basically provides access
  • 00:07:55
    to both GPU and CPU
  • 00:07:58
    yeah this this is um it's actually very
  • 00:08:00
    interesting topic but uh it's very deep
  • 00:08:03
    as well so I I I can only get this this
  • 00:08:07
    far because I'm a marketing guy so what
  • 00:08:10
    is the networking on it is that so yeah
  • 00:08:12
    oh that's another crazy part okay so
  • 00:08:14
    yeah we were just talking about this
  • 00:08:15
    right so far we're talking about this
  • 00:08:18
    so let's move over this part of the PCB
  • 00:08:20
    here so we have a this is
  • 00:08:22
    enterprisegrade connect X7 dual port 200
  • 00:08:26
    Gbits per second Ethernet 200 Gbit per
  • 00:08:29
    port per port okay
  • 00:08:31
    which is wild right i don't from top of
  • 00:08:35
    my memory I can't think of a box this
  • 00:08:36
    small with a network this fast now of
  • 00:08:40
    course we don't expect people to have
  • 00:08:41
    200 gigabit ports on their own that'll
  • 00:08:44
    be actually nice but uh yeah that's not
  • 00:08:46
    happening but this is designed for the
  • 00:08:48
    scale out option so using um this is
  • 00:08:53
    basically a cluster to network them yeah
  • 00:08:55
    yeah okay so
  • 00:08:57
    using distributed workload balancing
  • 00:08:59
    with TRTLM
  • 00:09:01
    and all our software stack for that
  • 00:09:03
    which is the same enterprise software
  • 00:09:06
    you can develop something at home that
  • 00:09:08
    would run on say the NVL72
  • 00:09:12
    so is it shipping with like Kuntu is
  • 00:09:14
    that yeah so uh we call it DJXOS but
  • 00:09:17
    this is Ubuntu 244 LTS okay the only
  • 00:09:20
    thing is yeah we we add a few extras to
  • 00:09:23
    it so we add um I don't know from top I
  • 00:09:27
    have everything that we add to it but it
  • 00:09:29
    has some performance optimizations it
  • 00:09:31
    has our repositories and our software
  • 00:09:33
    preloaded so uh driver driver for the
  • 00:09:36
    connect 7 all of these pieces are there
  • 00:09:39
    for you those modules are so did you
  • 00:09:41
    guys announce when it's going to be
  • 00:09:42
    available or price i Yeah so it is
  • 00:09:45
    available right now for a weight list
  • 00:09:47
    right so people that want to get on the
  • 00:09:49
    wait list they they can get in uh our
  • 00:09:52
    partners are also enabling their weight
  • 00:09:54
    lists one thing that I want all of you
  • 00:09:56
    uh media to be aware is the what what
  • 00:09:59
    our partners Asus HP Dell and Lenovo
  • 00:10:03
    these are the four partners we have
  • 00:10:04
    enabled right now we're selling the same
  • 00:10:07
    thing okay the the only difference is
  • 00:10:09
    they they're going to have different
  • 00:10:11
    case designs maybe different cooling
  • 00:10:13
    solutions
  • 00:10:14
    But uh this board is the same the
  • 00:10:16
    feature set is the same the only
  • 00:10:18
    variable that we're we have between
  • 00:10:20
    models today is uh we on the bottom of
  • 00:10:23
    the the board that you can't see cuz
  • 00:10:25
    it's glued in the base here um there's
  • 00:10:27
    an M.2 for NVME right it's PCI Express
  • 00:10:31
    um there's a one TB disc for the entry
  • 00:10:35
    model and a 4 TB version for the top
  • 00:10:38
    model that's only the only difference
  • 00:10:40
    how much memory are upgradeable
  • 00:10:43
    uh I don't I don't know if we're going
  • 00:10:45
    to state it that way okay but it's it's
  • 00:10:47
    it's M.2 yeah how much memory does it
  • 00:10:50
    come with the base model 1 TB and the
  • 00:10:54
    one TB of D RAM no how much DRAM sorry
  • 00:10:57
    oh okay no yeah it's 128 for all trims
  • 00:11:00
    all trims 128
  • 00:11:02
    that's what I'm saying here this is this
  • 00:11:03
    is something that I really want people
  • 00:11:04
    to be clear about so what's the largest
  • 00:11:07
    model that will run on that well we're
  • 00:11:10
    expecting uh for the single unit
  • 00:11:13
    somewhere in the 200 uh billion in FP4
  • 00:11:16
    which is yeah try doing that on a laptop
  • 00:11:19
    it's going to be really hard and for the
  • 00:11:22
    stackup it's it's 400 right with with
  • 00:11:24
    two and all like the Python software for
  • 00:11:26
    like the data scientist that works like
  • 00:11:29
    the QDF stuff oh I'm I'm going to have a
  • 00:11:31
    guy in the booth later today afternoon
  • 00:11:34
    uh when we're done with the press that
  • 00:11:36
    he is working on the Python optimization
  • 00:11:38
    you can okay he can give you a much
  • 00:11:40
    better answer than I why is that a
  • 00:11:42
    question for you why is that a question
  • 00:11:44
    yeah why is just the I mean typically
  • 00:11:47
    x86 is the platform for a lot of yeah
  • 00:11:50
    stuff so just making sure that like if
  • 00:11:52
    I'm writing Python code and I need to
  • 00:11:56
    move it to this it just works right
  • 00:11:58
    awesome how many can you run in parallel
  • 00:12:00
    uh well so far we're supporting two
  • 00:12:02
    right but um again this is Ethernet and
  • 00:12:05
    and the scale out is using is software
  • 00:12:08
    based okay we're offering you the the
  • 00:12:11
    same software that we use for very large
  • 00:12:14
    clustered distributed workloads on on
  • 00:12:17
    systems like the NVL72 behind us here to
  • 00:12:21
    use here so that that's the interesting
  • 00:12:24
    part you have a mini lab at your home
  • 00:12:25
    that's using basically the best scale
  • 00:12:29
    out technology available today if you
  • 00:12:32
    learn here you can just go there and do
  • 00:12:34
    the same with the same code with the
  • 00:12:36
    same
Tags
  • Nvidia
  • DJX Spark
  • AI supercomputer
  • Utviklingsverktøy
  • Teknologi
  • Deep learning
  • AI utvikling
  • Nettverksløsning
  • Hardware
  • Maskinvare