Watch CoreWeave Chief Revenue Officer, John Jones, and NVIDIA’s Director of Accelerated Computing Products, Dave Salvatore, break down what leaders need to know about the next wave of AI. They explore how to think about modern infrastructure decisions, what today’s benchmarks really signal, and how emerging trends—from agentic AI to new system architectures—are reshaping what’s possible for teams building AI at scale.
Here’s what you’ll learn:
- How to think about AI readiness, skills, and the pace of change
- Why benchmark-proven performance matters for real-world infrastructure planning
- How mixture-of-experts architectures are impacting agents and reasoning
- How market disruptions actually move AI forward
- What’s coming next in accelerated computing, including new architectures and interconnects
1
00:00:04,371 --> 00:00:05,722
So, without further
2
00:00:05,722 --> 00:00:06,923
ado, let's keep the momentum going.
3
00:00:06,923 --> 00:00:09,392
Let's please welcome to the stage
4
00:00:09,392 --> 00:00:12,595
Jon Jones, the new right, just onboarded,
5
00:00:12,812 --> 00:00:15,715
new Chief Revenue Officer at CoreWeave.
6
00:00:15,715 --> 00:00:17,100
And Dave Salvator,
7
00:00:17,100 --> 00:00:18,418
Director of Accelerated
8
00:00:18,418 --> 00:00:20,603
Computing Products at Nvidia.
9
00:00:20,603 --> 00:00:21,921
They're going to talk about accelerating
10
00:00:21,921 --> 00:00:23,006
AI infrastructure
11
00:00:23,006 --> 00:00:24,858
and balancing responsible leadership
12
00:00:24,858 --> 00:00:26,393
and relentless innovation.
13
00:00:26,393 --> 00:00:27,560
Let's give them a round of applause.
14
00:00:27,560 --> 00:00:30,563
Come on up.
15
00:00:36,469 --> 00:00:37,170
Okay.
16
00:00:37,170 --> 00:00:38,938
Thank you for that introduction, Dave.
17
00:00:38,938 --> 00:00:40,306
Thank you for joining us.
18
00:00:40,306 --> 00:00:40,840
Yeah. Good.
19
00:00:40,840 --> 00:00:43,076
For everyone who,
20
00:00:43,076 --> 00:00:44,127
got to eat lunch with Mike
21
00:00:44,127 --> 00:00:46,496
if you didn't get to eat lunch with Mike,
22
00:00:46,496 --> 00:00:47,964
this is about the only time
23
00:00:47,964 --> 00:00:48,531
you'll see him
24
00:00:48,531 --> 00:00:49,849
without his security detail.
25
00:00:49,849 --> 00:00:51,451
So go get a selfie
26
00:00:51,451 --> 00:00:54,287
and ambush him on the way out.
27
00:00:54,287 --> 00:00:55,038
All right.
28
00:00:55,038 --> 00:00:57,057
So thank you for the introduction.
29
00:00:57,057 --> 00:00:59,059
Dave, that was kind of a quick one.
30
00:00:59,059 --> 00:00:59,909
Why don't you tell us a little bit
31
00:00:59,909 --> 00:01:00,527
about what you do?
32
00:01:00,527 --> 00:01:00,977
Sure.
33
00:01:00,977 --> 00:01:02,629
So, my team at Nvidia,
34
00:01:02,629 --> 00:01:04,397
we work with our cloud providers,
35
00:01:04,397 --> 00:01:06,950
and work very closely with CoreWeave,
36
00:01:06,950 --> 00:01:08,101
on all things
37
00:01:08,101 --> 00:01:09,319
related to bring the Nvidia
38
00:01:09,319 --> 00:01:10,670
sort of technology to the cloud
39
00:01:10,670 --> 00:01:12,355
and helping to raise awareness
40
00:01:12,355 --> 00:01:14,040
and make sure that customers
41
00:01:14,040 --> 00:01:15,141
and the market in general
42
00:01:15,141 --> 00:01:16,192
just really understands
43
00:01:16,192 --> 00:01:17,627
the full scope of our partnership,
44
00:01:17,627 --> 00:01:19,112
but also the full scope
45
00:01:19,112 --> 00:01:21,347
of the capabilities of the platform.
46
00:01:21,347 --> 00:01:22,465
There is a bit of a tendency
47
00:01:22,465 --> 00:01:23,433
to think of Nvidia as,
48
00:01:23,433 --> 00:01:24,951
oh, those GPU guys,
49
00:01:24,951 --> 00:01:26,119
but the GPU really is just the beginning
50
00:01:26,119 --> 00:01:28,104
CoreWeave and Nvidia
51
00:01:28,104 --> 00:01:29,189
have a lot of what I'll call
52
00:01:29,189 --> 00:01:30,406
cultural simpatico.
53
00:01:30,406 --> 00:01:31,741
You know, Jean talked earlier
54
00:01:31,741 --> 00:01:34,010
about your three P's, you know, being,
55
00:01:34,010 --> 00:01:37,464
performance, partnership and pace.
56
00:01:37,464 --> 00:01:38,148
Right.
57
00:01:38,148 --> 00:01:39,449
And I may have gotten the order wrong,
58
00:01:39,449 --> 00:01:40,517
but those are the three,
59
00:01:40,517 --> 00:01:42,018
we're very much wired
60
00:01:42,018 --> 00:01:42,952
the same way at Nvidia.
61
00:01:42,952 --> 00:01:43,553
You know, at Nvidia,
62
00:01:43,553 --> 00:01:45,054
we talk about speed of light, right?
63
00:01:45,054 --> 00:01:46,039
Which is how quickly
64
00:01:46,039 --> 00:01:48,024
can we move to make something happen.
65
00:01:48,024 --> 00:01:48,925
And we have done
66
00:01:48,925 --> 00:01:50,543
some great collaborations with Kevin.
67
00:01:50,543 --> 00:01:51,361
Continue to do them.
68
00:01:51,361 --> 00:01:52,662
They're ongoing of course.
69
00:01:52,662 --> 00:01:54,030
But particularly around things
70
00:01:54,030 --> 00:01:55,315
like some of the work you've done
71
00:01:55,315 --> 00:01:56,549
with, your submissions
72
00:01:56,549 --> 00:01:58,017
to industry standard benchmarks
73
00:01:58,017 --> 00:01:58,668
like Mlperf
74
00:01:58,668 --> 00:01:59,002
is something
75
00:01:59,002 --> 00:02:00,537
I've been involved in myself,
76
00:02:00,537 --> 00:02:02,388
but you guys actually hold the record
77
00:02:02,388 --> 00:02:03,807
for the largest scale
78
00:02:03,807 --> 00:02:04,257
submission
79
00:02:04,257 --> 00:02:05,492
the benchmarks ever seen at over
80
00:02:05,492 --> 00:02:07,827
11,600 GPUs.
81
00:02:07,827 --> 00:02:08,611
And along with that,
82
00:02:08,611 --> 00:02:09,813
you have the training performance
83
00:02:09,813 --> 00:02:11,281
record to go with it.
84
00:02:11,281 --> 00:02:12,565
And at the end of that
85
00:02:12,565 --> 00:02:14,517
CoreWeave has a pretty good track record of,
86
00:02:14,517 --> 00:02:15,568
phenomenal track record.
87
00:02:15,568 --> 00:02:18,721
Yeah I think, again, that's the agility piece,
88
00:02:18,721 --> 00:02:20,039
that sort of that bias towards
89
00:02:20,039 --> 00:02:22,442
just go, go, go and being able
90
00:02:22,442 --> 00:02:23,393
to also work closely
91
00:02:23,393 --> 00:02:24,377
with our engineering teams.
92
00:02:24,377 --> 00:02:24,828
You know,
93
00:02:24,828 --> 00:02:25,979
we have reference designs
94
00:02:25,979 --> 00:02:28,014
we make available to our partners called
95
00:02:28,014 --> 00:02:29,332
Reference Designs.
96
00:02:29,332 --> 00:02:32,302
And I believe CoreWeave doesn't quite use it 1 to 1,
97
00:02:32,302 --> 00:02:33,153
but they do
98
00:02:33,153 --> 00:02:34,020
sort of take
99
00:02:34,020 --> 00:02:35,188
or you do take a fair
100
00:02:35,188 --> 00:02:36,422
amount of what's
101
00:02:36,422 --> 00:02:37,373
in our reference designs.
102
00:02:37,373 --> 00:02:37,941
You take it,
103
00:02:37,941 --> 00:02:38,842
maybe tweak it a little bit
104
00:02:38,842 --> 00:02:39,926
for how you want to deploy it,
105
00:02:39,926 --> 00:02:41,327
and then you guys deploy
106
00:02:41,327 --> 00:02:43,346
and that helps speed your time to market.
107
00:02:43,346 --> 00:02:43,730
Yeah.
108
00:02:43,730 --> 00:02:45,748
And and as part of that process,
109
00:02:45,748 --> 00:02:46,466
there's a lot of talk
110
00:02:46,466 --> 00:02:48,501
about GPUs and GPU availability,
111
00:02:48,501 --> 00:02:51,504
but that it's not it's not really a chip
112
00:02:51,621 --> 00:02:52,705
constraint alone
113
00:02:52,705 --> 00:02:54,290
or chip deployment alone. Right.
114
00:02:54,290 --> 00:02:55,341
There's a lot of hard work
115
00:02:55,341 --> 00:02:56,709
at the software layer.
116
00:02:56,709 --> 00:02:58,061
But the interconnect
117
00:02:58,061 --> 00:02:58,761
layer, building
118
00:02:58,761 --> 00:02:59,846
clusters of these things
119
00:02:59,846 --> 00:03:00,480
is very different
120
00:03:00,480 --> 00:03:02,065
than building a few of these things
121
00:03:02,065 --> 00:03:03,349
very much talk a little bit.
122
00:03:03,349 --> 00:03:05,118
And the software stack, of course,
123
00:03:05,118 --> 00:03:06,753
the Nvidia built is world class
124
00:03:06,753 --> 00:03:08,338
and is by far the market leader.
125
00:03:08,338 --> 00:03:08,872
Talk a little bit
126
00:03:08,872 --> 00:03:10,373
about the importance of software
127
00:03:10,373 --> 00:03:11,658
as part of these deployments.
128
00:03:11,658 --> 00:03:13,459
Software plays a huge role.
129
00:03:13,459 --> 00:03:13,877
I mean,
130
00:03:13,877 --> 00:03:15,411
without the software to really unlock
131
00:03:15,411 --> 00:03:17,180
the full potential of the hardware,
132
00:03:17,180 --> 00:03:18,648
the hardware is a glorified space
133
00:03:18,648 --> 00:03:19,315
heater, right?
134
00:03:19,315 --> 00:03:20,600
I mean, you really need
135
00:03:20,600 --> 00:03:22,669
software to be a performant,
136
00:03:22,669 --> 00:03:24,470
but it involves so much work
137
00:03:24,470 --> 00:03:25,405
with the ecosystem.
138
00:03:25,405 --> 00:03:27,390
As many of your audience members know,
139
00:03:28,675 --> 00:03:29,209
open source
140
00:03:29,209 --> 00:03:31,194
software is a hugely important component
141
00:03:31,194 --> 00:03:33,079
to driving AI innovation, right?
142
00:03:33,079 --> 00:03:33,680
It's one of the things
143
00:03:33,680 --> 00:03:34,647
that allows developers
144
00:03:34,647 --> 00:03:36,266
to kind of share, work with each other
145
00:03:36,266 --> 00:03:36,933
and be able to take
146
00:03:36,933 --> 00:03:38,484
that work, build upon that work,
147
00:03:38,484 --> 00:03:40,370
and sequentially drive innovation
148
00:03:40,370 --> 00:03:41,004
into the market.
149
00:03:41,004 --> 00:03:41,671
Right.
150
00:03:41,671 --> 00:03:43,439
CoreWeave makes a lot of contributions
151
00:03:43,439 --> 00:03:44,190
into the open source
152
00:03:44,190 --> 00:03:45,475
world, as does Nvidia.
153
00:03:45,475 --> 00:03:45,959
We you know,
154
00:03:45,959 --> 00:03:46,559
we contribute
155
00:03:46,559 --> 00:03:49,762
on models, on frameworks, on things
156
00:03:49,762 --> 00:03:50,847
like our dynamo,
157
00:03:50,847 --> 00:03:51,864
inference serving software,
158
00:03:51,864 --> 00:03:53,299
which is also open source.
159
00:03:53,299 --> 00:03:54,717
And we do a huge amount of work
160
00:03:54,717 --> 00:03:55,551
in terms of submissions
161
00:03:55,551 --> 00:03:57,237
we make into the Linux project,
162
00:03:57,237 --> 00:03:59,188
which ultimately get rolled into,
163
00:03:59,188 --> 00:03:59,939
you know, the main
164
00:03:59,939 --> 00:04:01,591
code base of it,
165
00:04:01,591 --> 00:04:03,476
to help enable accelerated computing.
166
00:04:03,476 --> 00:04:04,978
And then a number of our own libraries
167
00:04:04,978 --> 00:04:06,679
as well are also open source.
168
00:04:06,679 --> 00:04:07,580
So,
169
00:04:07,580 --> 00:04:09,115
you know, the software piece
170
00:04:09,115 --> 00:04:10,650
is just incredibly important.
171
00:04:10,650 --> 00:04:13,019
And because the early, early work on
172
00:04:13,019 --> 00:04:13,770
deep learning was done
173
00:04:13,770 --> 00:04:14,637
on Nvidia products,
174
00:04:14,637 --> 00:04:16,322
if you go all the way back to AlexNet,
175
00:04:16,322 --> 00:04:18,358
back in 2012,
176
00:04:18,358 --> 00:04:19,592
we saw that opportunity
177
00:04:19,592 --> 00:04:20,760
and we already had something in place
178
00:04:20,760 --> 00:04:21,411
called Cuda,
179
00:04:21,411 --> 00:04:22,312
which had been in the market
180
00:04:22,312 --> 00:04:24,564
for a, you know, about five or so years.
181
00:04:24,564 --> 00:04:25,648
And we were trying to find
182
00:04:25,648 --> 00:04:26,733
we were starting to do things with it
183
00:04:26,733 --> 00:04:28,101
around scientific simulation,
184
00:04:28,101 --> 00:04:29,252
you know, whether it's CFD,
185
00:04:29,252 --> 00:04:31,187
molecular dynamics, what have you.
186
00:04:31,187 --> 00:04:31,738
And it turned out
187
00:04:31,738 --> 00:04:33,539
GPUs are really good at that.
188
00:04:33,539 --> 00:04:33,890
In fact,
189
00:04:33,890 --> 00:04:35,458
GPUs are pretty much really good
190
00:04:35,458 --> 00:04:37,160
at anything that can be parallelized
191
00:04:37,160 --> 00:04:38,278
because it's a massively
192
00:04:38,278 --> 00:04:40,079
parallel machine. Right?
193
00:04:40,079 --> 00:04:41,297
Alex comes along, built
194
00:04:41,297 --> 00:04:43,099
AlexNet was done on our product
195
00:04:43,099 --> 00:04:44,484
and then we saw an opportunity for Cuda.
196
00:04:44,484 --> 00:04:45,535
So then we started building
197
00:04:45,535 --> 00:04:46,552
more libraries to help
198
00:04:46,552 --> 00:04:47,236
support
199
00:04:47,236 --> 00:04:48,621
this notion of deep learning because,
200
00:04:48,621 --> 00:04:50,456
I mean, I it's been with us for decades.
201
00:04:50,456 --> 00:04:50,990
Right.
202
00:04:50,990 --> 00:04:52,225
But the notion of deep learning,
203
00:04:52,225 --> 00:04:53,159
of building networks
204
00:04:53,159 --> 00:04:55,194
that have multiple, multiple layers
205
00:04:55,194 --> 00:04:56,129
and massive amounts
206
00:04:56,129 --> 00:04:58,431
of compute necessary to drive them
207
00:04:58,431 --> 00:05:00,116
wasn't was simply infeasible
208
00:05:00,116 --> 00:05:01,317
in the past, right?
209
00:05:01,317 --> 00:05:02,969
GPUs made that possible.
210
00:05:02,969 --> 00:05:03,770
So we've been driving
211
00:05:03,770 --> 00:05:06,773
and innovating software really for over,
212
00:05:06,956 --> 00:05:08,574
you know, over a dozen years
213
00:05:08,574 --> 00:05:09,709
to make AI
214
00:05:09,709 --> 00:05:11,661
and help it drive, drive it forward.
215
00:05:11,661 --> 00:05:12,028
Right.
216
00:05:12,028 --> 00:05:13,579
And then of course, CoreWeave
217
00:05:13,579 --> 00:05:14,130
you know, is doing
218
00:05:14,130 --> 00:05:15,264
a lot of that as well,
219
00:05:15,264 --> 00:05:15,782
through work
220
00:05:15,782 --> 00:05:16,949
you do certainly internally
221
00:05:16,949 --> 00:05:17,817
and some of your recent
222
00:05:17,817 --> 00:05:19,202
acquisitions as well.
223
00:05:19,202 --> 00:05:20,603
You guys are also building out
224
00:05:20,603 --> 00:05:21,504
a pretty robust software
225
00:05:21,504 --> 00:05:22,939
stack of your own.
226
00:05:22,939 --> 00:05:24,023
A lot of the workloads
227
00:05:24,023 --> 00:05:25,224
that you talked about,
228
00:05:25,224 --> 00:05:26,909
simulation and rendering
229
00:05:26,909 --> 00:05:28,728
are very high performance,
230
00:05:28,728 --> 00:05:29,729
classical, high performance
231
00:05:29,729 --> 00:05:30,530
computing workloads.
232
00:05:30,530 --> 00:05:30,863
By the way,
233
00:05:30,863 --> 00:05:31,547
is anybody in here
234
00:05:31,547 --> 00:05:32,749
willing to self-identify as
235
00:05:32,749 --> 00:05:34,250
having HPC expertise?
236
00:05:35,418 --> 00:05:36,919
No, it
237
00:05:36,919 --> 00:05:37,704
just means you're all
238
00:05:37,704 --> 00:05:40,673
much younger than I am. Six, 16 months.
239
00:05:40,673 --> 00:05:41,190
Okay.
240
00:05:41,190 --> 00:05:44,477
So, once you get into, just up to
241
00:05:44,477 --> 00:05:45,178
and through the GPU
242
00:05:45,178 --> 00:05:46,429
in the software stack,
243
00:05:46,429 --> 00:05:47,847
we were talking in the back of
244
00:05:47,847 --> 00:05:48,564
you have a lot of long
245
00:05:48,564 --> 00:05:49,499
history of benchmarking.
246
00:05:49,499 --> 00:05:51,734
Two going going, way back.
247
00:05:51,734 --> 00:05:53,603
Maybe you can talk a little bit
248
00:05:53,603 --> 00:05:54,987
about the arc of,
249
00:05:54,987 --> 00:05:56,839
of the benchmarking progress
250
00:05:56,839 --> 00:05:59,342
as it, as it, relates to pre
251
00:05:59,342 --> 00:06:01,677
AI and then post AI because ultimately
252
00:06:01,677 --> 00:06:03,813
AI clusters
253
00:06:03,813 --> 00:06:06,799
are just big supercomputers.
254
00:06:07,650 --> 00:06:07,917
Yeah.
255
00:06:07,917 --> 00:06:09,052
I guess we just had a pick up
256
00:06:09,052 --> 00:06:10,353
in the volume there. That's fine.
257
00:06:10,353 --> 00:06:11,387
You know lunch.
258
00:06:11,387 --> 00:06:12,388
Come on. Might be setting in.
259
00:06:12,388 --> 00:06:13,790
So this will help keep them awake.
260
00:06:13,790 --> 00:06:16,642
No. They are similar workloads.
261
00:06:16,642 --> 00:06:17,276
They both have
262
00:06:17,276 --> 00:06:19,112
parallelizable aspects, but
263
00:06:19,112 --> 00:06:20,430
but AI in particular
264
00:06:20,430 --> 00:06:22,532
is just embarrassingly parallelizable,
265
00:06:22,532 --> 00:06:23,449
which is what makes it such
266
00:06:23,449 --> 00:06:24,767
a great fit for GPUs.
267
00:06:26,069 --> 00:06:26,602
But, you know,
268
00:06:26,602 --> 00:06:29,005
again, the benchmarking space of
269
00:06:29,005 --> 00:06:30,773
AI has been pretty crazy.
270
00:06:30,773 --> 00:06:32,725
I would say it's been kind of Wild West,
271
00:06:32,725 --> 00:06:34,110
for some time,
272
00:06:34,110 --> 00:06:35,027
which is why,
273
00:06:35,027 --> 00:06:35,328
you know,
274
00:06:35,328 --> 00:06:36,362
we're a member of what's called
275
00:06:36,362 --> 00:06:37,880
the ML Commons Consortium,
276
00:06:37,880 --> 00:06:38,431
which builds
277
00:06:38,431 --> 00:06:39,715
the mlperf benchmarks both
278
00:06:39,715 --> 00:06:41,451
for training and inference.
279
00:06:41,451 --> 00:06:42,235
And they build a number
280
00:06:42,235 --> 00:06:43,202
of other benchmarks as well.
281
00:06:43,202 --> 00:06:44,337
But the data center benchmarks
282
00:06:44,337 --> 00:06:46,639
is our primary focus at Nvidia.
283
00:06:46,639 --> 00:06:48,324
And we have done very,
284
00:06:48,324 --> 00:06:49,409
very well on those benchmarks.
285
00:06:49,409 --> 00:06:50,777
And we try and, you know, bring out
286
00:06:50,777 --> 00:06:52,011
new versions of those benchmarks
287
00:06:52,011 --> 00:06:52,979
about every six months,
288
00:06:52,979 --> 00:06:54,480
which has industry
289
00:06:54,480 --> 00:06:55,765
consortium driven benchmarks.
290
00:06:55,765 --> 00:06:58,050
Go is a very fast pace.
291
00:06:58,050 --> 00:06:59,368
That said, we're trying
292
00:06:59,368 --> 00:07:00,887
to even make that pace faster
293
00:07:00,887 --> 00:07:01,587
because of the rate at
294
00:07:01,587 --> 00:07:02,722
which AI is evolving.
295
00:07:04,307 --> 00:07:05,274
And so this is an
296
00:07:05,274 --> 00:07:05,942
attempt, frankly,
297
00:07:05,942 --> 00:07:06,692
to kind of bring some order
298
00:07:06,692 --> 00:07:07,677
to chaos to kind of say,
299
00:07:07,677 --> 00:07:08,911
let's build a benchmark
300
00:07:08,911 --> 00:07:09,595
where we have directly
301
00:07:09,595 --> 00:07:10,847
comparable results.
302
00:07:10,847 --> 00:07:12,348
So we can actually look at two systems
303
00:07:12,348 --> 00:07:13,850
and go, okay, they did the same work.
304
00:07:13,850 --> 00:07:15,618
Who's better? Right.
305
00:07:15,618 --> 00:07:16,586
And they're fairly
306
00:07:16,586 --> 00:07:18,988
stringent requirements around accuracy,
307
00:07:18,988 --> 00:07:19,822
around the ability
308
00:07:19,822 --> 00:07:21,107
that the models have to converge.
309
00:07:21,107 --> 00:07:22,742
If you're training as an example.
310
00:07:22,742 --> 00:07:24,260
And these are peer reviewed benchmarks.
311
00:07:24,260 --> 00:07:25,978
So once the results are submitted,
312
00:07:25,978 --> 00:07:27,530
there's about a month long period
313
00:07:27,530 --> 00:07:28,998
where every submitter and, you know,
314
00:07:28,998 --> 00:07:30,683
a lot of these are competitors
315
00:07:30,683 --> 00:07:31,734
can basically scrutinize
316
00:07:31,734 --> 00:07:32,752
everyone else's submission.
317
00:07:32,752 --> 00:07:33,870
And if they see something
318
00:07:33,870 --> 00:07:35,972
that they think might be questionable,
319
00:07:35,972 --> 00:07:37,140
they can raise a concern.
320
00:07:37,140 --> 00:07:38,541
And that concern has to be addressed
321
00:07:38,541 --> 00:07:40,009
before the benchmark results
322
00:07:40,009 --> 00:07:41,093
will see light of day.
323
00:07:41,093 --> 00:07:42,411
So what you're getting is results
324
00:07:42,411 --> 00:07:42,962
that have kind of,
325
00:07:42,962 --> 00:07:45,331
you know, had held the fire, if you will,
326
00:07:45,331 --> 00:07:46,382
that frankly,
327
00:07:46,382 --> 00:07:46,999
are more meaningful
328
00:07:46,999 --> 00:07:48,351
and directly comparable.
329
00:07:48,351 --> 00:07:48,718
Right.
330
00:07:48,718 --> 00:07:49,469
You know, that said,
331
00:07:49,469 --> 00:07:50,486
there are other benchmarking
332
00:07:50,486 --> 00:07:51,654
efforts underway.
333
00:07:51,654 --> 00:07:52,705
Most recently,
334
00:07:52,705 --> 00:07:53,539
some of the analysis did
335
00:07:53,539 --> 00:07:54,690
a thing called inference. Max.
336
00:07:54,690 --> 00:07:55,741
We did very well with that
337
00:07:55,741 --> 00:07:57,677
with our Blackwell architecture.
338
00:07:57,677 --> 00:07:59,061
Very pleased to see those results.
339
00:07:59,061 --> 00:08:00,480
And it speaks to the
340
00:08:00,480 --> 00:08:01,831
the growing importance of inference.
341
00:08:01,831 --> 00:08:02,698
And Peter spoke to this
342
00:08:02,698 --> 00:08:04,317
a little bit earlier this morning
343
00:08:04,317 --> 00:08:06,802
about about the growth of this scale
344
00:08:06,802 --> 00:08:08,704
and the demands of inference.
345
00:08:08,704 --> 00:08:10,256
Really just kind of,
346
00:08:10,256 --> 00:08:12,325
you know, on this, on this incredible,
347
00:08:12,325 --> 00:08:14,310
basically exponential trajectory.
348
00:08:14,310 --> 00:08:15,361
So, so you get to see
349
00:08:15,361 --> 00:08:16,829
most of the benchmarking work.
350
00:08:16,829 --> 00:08:18,297
I know our teams
351
00:08:18,297 --> 00:08:19,649
collaborate closely together on that.
352
00:08:19,649 --> 00:08:20,967
I'll ask you about that in a minute.
353
00:08:20,967 --> 00:08:23,903
But, from from that objective viewpoint, how
354
00:08:23,903 --> 00:08:24,470
How’s CoreWeave doing?
355
00:08:25,521 --> 00:08:27,123
CoreWeave, again, does very well.
356
00:08:27,123 --> 00:08:28,441
I think because of the close
357
00:08:28,441 --> 00:08:28,975
collaboration
358
00:08:28,975 --> 00:08:30,593
we have on the engineering side,
359
00:08:30,593 --> 00:08:31,527
and because you guys
360
00:08:31,527 --> 00:08:32,879
are also very focused on
361
00:08:32,879 --> 00:08:33,629
making sure you get
362
00:08:33,629 --> 00:08:34,680
phenomenal performance
363
00:08:34,680 --> 00:08:36,482
from from our technologies.
364
00:08:36,482 --> 00:08:38,267
The result comes out in places
365
00:08:38,267 --> 00:08:39,051
like Mlperf.
366
00:08:39,051 --> 00:08:39,619
And, you know,
367
00:08:39,619 --> 00:08:40,436
you also, of course,
368
00:08:40,436 --> 00:08:41,571
on the recent cluster
369
00:08:41,571 --> 00:08:43,339
Max evaluation that SA,
370
00:08:43,339 --> 00:08:45,541
did you guys also were the only,
371
00:08:45,541 --> 00:08:46,792
submitter,
372
00:08:46,792 --> 00:08:47,510
I think, who achieved,
373
00:08:47,510 --> 00:08:48,628
I guess they call it the platinum level,
374
00:08:48,628 --> 00:08:50,096
which is the highest level.
375
00:08:50,096 --> 00:08:52,315
So no, CoreWeave has definitely proven itself.
376
00:08:52,315 --> 00:08:54,617
I mean, track record matters, right?
377
00:08:54,617 --> 00:08:56,419
And it's one thing to sort of
378
00:08:56,419 --> 00:08:58,321
basically do some kind of a performance
379
00:08:58,321 --> 00:08:58,821
stunt and go
380
00:08:58,821 --> 00:08:59,572
look at how fast we are
381
00:08:59,572 --> 00:09:01,407
on this one model. In this one condition,
382
00:09:02,658 --> 00:09:03,109
it can be a
383
00:09:03,109 --> 00:09:03,859
valid data point,
384
00:09:03,859 --> 00:09:05,177
but we all know that I,
385
00:09:05,177 --> 00:09:06,212
particularly on the inference
386
00:09:06,212 --> 00:09:06,979
side,
387
00:09:06,979 --> 00:09:08,748
has so many parametric knobs
388
00:09:08,748 --> 00:09:10,099
that can be played with,
389
00:09:10,099 --> 00:09:11,701
that getting to directly
390
00:09:11,701 --> 00:09:12,652
comparable performance
391
00:09:12,652 --> 00:09:14,036
can be a real challenge. Right?
392
00:09:14,036 --> 00:09:16,172
Whereas when you're willing to basically
393
00:09:16,172 --> 00:09:16,739
be weighed
394
00:09:16,739 --> 00:09:17,607
and measured in a place
395
00:09:17,607 --> 00:09:18,658
like, say, mlperf,
396
00:09:18,658 --> 00:09:19,942
where you're going to be scrutinized,
397
00:09:19,942 --> 00:09:21,928
if you do anything that's at. All.
398
00:09:21,928 --> 00:09:23,062
Odd looking.
399
00:09:23,062 --> 00:09:25,047
It means that the results
400
00:09:25,047 --> 00:09:26,732
that ultimately come out of that,
401
00:09:26,732 --> 00:09:28,484
frankly, are more meaningful.
402
00:09:28,484 --> 00:09:29,368
They're more credible,
403
00:09:29,368 --> 00:09:30,319
and I think they're ultimately
404
00:09:30,319 --> 00:09:31,437
a better guide to make,
405
00:09:31,437 --> 00:09:32,989
you know, informed investment decisions
406
00:09:32,989 --> 00:09:33,723
about where you want
407
00:09:33,723 --> 00:09:35,541
to build out your infrastructure.
408
00:09:35,541 --> 00:09:37,260
And again, we were talking in the back,
409
00:09:37,260 --> 00:09:40,313
but pretty consistently, CoreWeave’s shown that
410
00:09:40,313 --> 00:09:41,547
the company has the expertise
411
00:09:41,547 --> 00:09:42,748
and the engineering collaboration
412
00:09:42,748 --> 00:09:44,216
is showing up in those benchmarks.
413
00:09:44,216 --> 00:09:45,418
And,
414
00:09:45,418 --> 00:09:46,319
and pretty consistently
415
00:09:46,319 --> 00:09:47,687
Cory's coming out on top.
416
00:09:47,687 --> 00:09:48,871
Yeah, CoreWeave has done
417
00:09:48,871 --> 00:09:49,739
very well with those.
418
00:09:49,739 --> 00:09:50,373
And again,
419
00:09:50,373 --> 00:09:51,240
I think that's a testament
420
00:09:51,240 --> 00:09:52,141
to the partnership.
421
00:09:52,141 --> 00:09:53,643
And it's also a testament to,
422
00:09:53,643 --> 00:09:55,077
the level of engineering skill
423
00:09:55,077 --> 00:09:56,429
that, you know, Peter and his team have.
424
00:09:56,429 --> 00:09:58,698
I mean, they, they do terrific work.
425
00:09:58,698 --> 00:09:59,515
Okay. All right. Great.
426
00:09:59,515 --> 00:10:00,049
Well, let's move
427
00:10:00,049 --> 00:10:00,716
a little bit past
428
00:10:00,716 --> 00:10:01,917
the technical conversation.
429
00:10:01,917 --> 00:10:03,569
There are a lot of leaders in the room
430
00:10:03,569 --> 00:10:05,004
thinking about some of the hard decisions
431
00:10:05,004 --> 00:10:06,505
they have to make as it relates to AI,
432
00:10:06,505 --> 00:10:08,124
as it relates to infrastructure.
433
00:10:08,124 --> 00:10:08,941
I know
434
00:10:08,941 --> 00:10:09,508
one of the things
435
00:10:09,508 --> 00:10:10,910
you wanted to talk about
436
00:10:10,910 --> 00:10:12,778
is mental models around leaders
437
00:10:12,778 --> 00:10:14,447
making decisions about the future
438
00:10:14,447 --> 00:10:15,931
and how that impacts their business.
439
00:10:15,931 --> 00:10:17,750
Maybe maybe you can share a little bit,
440
00:10:17,750 --> 00:10:19,018
with the crowd
441
00:10:19,018 --> 00:10:20,102
here on how
442
00:10:20,102 --> 00:10:20,786
they should think about
443
00:10:20,786 --> 00:10:22,038
some of those decisions.
444
00:10:22,038 --> 00:10:22,888
Yeah, it's
445
00:10:22,888 --> 00:10:24,173
it's an interesting time
446
00:10:24,173 --> 00:10:26,542
because AI is really
447
00:10:26,542 --> 00:10:27,543
is getting going now.
448
00:10:27,543 --> 00:10:29,128
And we've seen some of the initial things
449
00:10:29,128 --> 00:10:30,863
that agenda guy can do for us. Right.
450
00:10:30,863 --> 00:10:32,148
But I'd say we're still very much
451
00:10:32,148 --> 00:10:34,233
in the early innings on a genetic AI.
452
00:10:34,233 --> 00:10:35,751
So as you think about,
453
00:10:35,751 --> 00:10:36,686
I mean, in a lot of ways
454
00:10:36,686 --> 00:10:38,037
for business leaders,
455
00:10:38,037 --> 00:10:39,171
AI is no different
456
00:10:39,171 --> 00:10:40,790
than any other technology or tool
457
00:10:40,790 --> 00:10:42,058
you consider for your business.
458
00:10:42,058 --> 00:10:42,842
And the first question
459
00:10:42,842 --> 00:10:43,592
you want to ask yourself
460
00:10:43,592 --> 00:10:44,910
is, what business problem
461
00:10:44,910 --> 00:10:46,829
am I going to solve with it, right?
462
00:10:46,829 --> 00:10:47,163
In other words,
463
00:10:47,163 --> 00:10:48,047
what are your biggest pain
464
00:10:48,047 --> 00:10:49,765
points in your business? Is it?
465
00:10:49,765 --> 00:10:51,651
It could be as simple as our wait
466
00:10:51,651 --> 00:10:52,401
times are too long