Richard Hsu commited on
Commit
df0d3a5
·
1 Parent(s): f59bbd9
app.py CHANGED
@@ -16,8 +16,8 @@ def extract_text_from_pdf(pdf_file):
16
  # Create a Gradio interface
17
  interface = gr.Interface(
18
  fn=extract_text_from_pdf,
19
- inputs=gr.inputs.File(label="Upload PDF"),
20
- outputs=gr.outputs.Textbox(label="Extracted Text"),
21
  title="PDF Text Extractor",
22
  description="Upload a PDF file to extract and display its text content."
23
  )
 
16
  # Create a Gradio interface
17
  interface = gr.Interface(
18
  fn=extract_text_from_pdf,
19
+ inputs=gr.File(label="Upload PDF"),
20
+ outputs=gr.Textbox(label="Extracted Text"),
21
  title="PDF Text Extractor",
22
  description="Upload a PDF file to extract and display its text content."
23
  )
flagged/Upload PDF/df016a9cf3c724acbaf36497283ca4669fa5629f/2024Q2.pdf ADDED
Binary file (96.7 kB). View file
 
flagged/log.csv ADDED
@@ -0,0 +1,682 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Upload PDF,Extracted Text,flag,username,timestamp
2
+ /Users/richardhsu/pdf_upload/flagged/Upload PDF/df016a9cf3c724acbaf36497283ca4669fa5629f/2024Q2.pdf,"'Broadcom Inc. (NASDAQ:AVGO) Q2 2024 Earnings Conference Call June
3
+ 12, 2024 5:00 PM ET
4
+ Company Participants
5
+ Ji Yoo - Head, IR
6
+ Hock Tan - President and CEO
7
+ Kirsten Spears - CFO
8
+ Charlie Kawwas - President, Semiconductor Solutions Group
9
+ Conference Call Participants
10
+ Vivek Arya - Bank of America Securities
11
+ Ross Seymore - Deutsche Bank
12
+ Stacy Rasgon - Bernstein
13
+ Harlan Sur - JPMorgan
14
+ Ben Reitzes - Melius Research
15
+ Toshiya Hari - Goldman Sachs
16
+ Blayne Curtis - Jefferies
17
+ Timothy Arcuri - UBS
18
+ Thomas O'Malley - Barclays
19
+ Karl Ackerman - BNP Paribas
20
+ CJ Muse - Cantor Fitzgerald
21
+ William Stein - Truist Securities
22
+ Operator
23
+ Welcome to Broadcom Inc. Second Quarter Fiscal Year 2024 Financial
24
+ Results Conference Call. At this time, for opening remarks and introductions, I
25
+ would like to turn the call over to Ji Yoo, Head of Investor Relations of
26
+ Broadcom Inc.
27
+ Ji Yoo
28
+ Thank you, Operator, and good afternoon, everyone. Joining me on today's
29
+ call are Hock Tan, President and CEO; Kirsten Spears, Chief Financial Officer
30
+ and Charlie Kawwas, President, Semiconductor Solutions Group.
31
+ Broadcom distributed a press release and financial tables after the market
32
+ closed, describing our financial performance for the second quarter of fiscal
33
+ year 2024. If you did not receive a copy, you may obtain the information from
34
+ the Investor section of Broadcom's website at Broadcom.com.
35
+ This conference call is being webcast live and an audio replay of the call can
36
+ be accessed for 1 year through the Investor section of Broadcom's website.
37
+ During the prepared comments, Hock and Kirsten will be providing details of
38
+ our second quarter fiscal year 2024 results, guidance for our fiscal year 2024,
39
+ as well as commentary regarding the business environment. We'll take
40
+ questions after the end of our prepared comments.
41
+ Please refer to our press release today and our recent filings with the SEC
42
+ for information on the specific risk factors that could cause our actual results
43
+ to differ materially from the forward-looking statements made on this call. In
44
+ addition to US GAAP reporting, Broadcom reports certain financial measures
45
+ on a non-GAAP basis. A reconciliation between GAAP and non-GAAP
46
+ measures is included in the tables attached to today's press release.
47
+ Comments made during today's call will primarily refer to our non-GAAP
48
+ financial results. I'll now turn the call over to Hock.
49
+ Hock Tan
50
+ Thank you, Ji. And thank you everyone for joining today. In our fiscal Q2 2024
51
+ results -- consolidated net revenue was $12.5 billion, up 43% year-on-year as
52
+ revenue included a full quarter of contribution from VMware. But if we exclude
53
+ VMware, consolidated revenue was up 12% year-on-year. And this 12%
54
+ organic growth in revenue was largely driven by AI revenue, which stepped up
55
+ 280% year-on-year to $3.1 billion, more than offsetting continued cyclical
56
+ weakness in semiconductor revenue from enterprises and telcos.
57
+ Let me now give you more color on our two reporting segments. Beginning
58
+ with software. In Q2 infrastructure software segment revenue of $5.3 billion
59
+ was up 175% year-on-year and included $2.7 billion in revenue contribution
60
+ from VMware, up from $2.1 billion in the prior quarter. The integration of
61
+ VMware is going very well. Since we acquired VMware, we have modernized
62
+ the product SKUs from over 8,000 disparate SKUs to four core product
63
+ offerings and simplified the go-to-market flow, eliminating a huge amount of
64
+ channel conflicts.
65
+ We are making good progress in transitioning all VMware products to a
66
+ subscription licensing model. And since closing the deal, we have actually
67
+ signed up close to 3,000 of our largest 10,000 customers to enable them to
68
+ build a self-service virtual private cloud on-prem. Each of these customers
69
+ typically sign up to a multi-year contract, which we normalize into an annual
70
+ measure known as Annualized Booking Value, or ABV. This metric, ABV for
71
+ VMware products, accelerated from $1.2 billion in Q1 to $1.9 billion in Q2.
72
+ For reference, for the consolidated Broadcom software portfolio, ABV grew
73
+ from $1.9 billion in Q1 to $2.8 billion over the same period in Q2. Meanwhile,
74
+ we have integrated SG&A across the entire platform and eliminated
75
+ redundant functions. Year-to-date, we have incurred about $2 billion of
76
+ restructuring and integration costs and drove our spending run rate at
77
+ VMware to $1.6 billion this quarter, from what used to be $2.3 billion per
78
+ quarter pre-acquisition.
79
+ We expect spending will continue to decline towards a $1.3 billion run rate
80
+ exiting Q4, better than our previous $1.4 billion plan, and will likely stabilize at
81
+ $1.2 billion post-integration. VMware revenue in Q1 was $2.1 billion, grew to
82
+ $2.7 billion in Q2, and will accelerate towards a $4 billion per quarter run rate.
83
+ We therefore expect operating margins for VMware to begin to converge
84
+ towards that of classic Broadcom software by fiscal 2025.
85
+ Turning to semiconductors, let me give you more color by end markets.
86
+ Networking. Q2 revenue of $3.8 billion grew 44% year-on-year, representing
87
+ 53% of semiconductor revenue. This was again driven by strong demand from
88
+ hyperscalers for both AI networking and custom accelerators. It's interesting
89
+ to note that as AI data center clusters continue to deploy, our revenue mix has
90
+ been shifting towards an increasing proportion of networking.
91
+ We doubled the number of switches we sold year-on-year, particularly the
92
+ PAM-5 and Jericho3, which we deployed successfully in close collaboration
93
+ with partners like Arista Networks, Dell, Juniper, and Supermicro. Additionally,
94
+ we also double our shipments of PCI Express switches and NICs in the AI
95
+ backend fabric. We're leading the rapid transition of optical interconnects in AI
96
+ data centers to 800 gigabit bandwidth, which is driving accelerated growth for
97
+ our DSPs, optical lasers, and PIN diodes. And we are not standing still.
98
+ Together with these same partners, we are developing the next generation
99
+ switches, DSP, and optics that will drive the ecosystem towards 1.6 terabit
100
+ connectivity to scale out larger AI accelerated clusters.
101
+ Talking of AI accelerators, you may know our hyperscale customers are
102
+ accelerating their investments to scale up the performance of these clusters.
103
+ And to that end, we have just been awarded the next generation custom AI
104
+ accelerators for these hyperscale customers of ours. Networking these AI
105
+ accelerators is very challenging, but the technology does exist today. In
106
+ Broadcom, with the deepest and broadest understanding of what it takes for
107
+ complex, large workloads to be scaled out in an AI fabric. Proof in
108
+ point, seven of the largest eight AI clusters in deployment today use
109
+ Broadcom Ethernet solutions.
110
+ Next year, we expect all mega-scale GPU deployments to be on Ethernet. We
111
+ expect the strength in AI to continue, and because of that, we now expect
112
+ networking revenue to grow 40% year-on-year compared to our prior
113
+ guidance of over 35% growth. Moving to wireless. Q2 wireless revenue of
114
+ $1.6 billion grew 2% year-on-year, was seasonally down 19% quarter-on-
115
+ quarter and represents 22% of semiconductor revenue.
116
+ And in fiscal '24, helped by content increases, we reiterate our previous
117
+ guidance for wireless revenue to be essentially flat year-on-year. This trend is
118
+ wholly consistent with our continued engagement with our North American
119
+ customer, which is deep, strategic, and multiyear and represents all of our
120
+ wireless business. Next, our Q2 server storage connectivity revenue was
121
+ $824 million or 11% of semiconductor revenue, down 27% year-on-year. We
122
+ believe though, Q2 was the bottom in server storage. And based on updated
123
+ demand forecast and bookings, we expect a modest recovery in the second
124
+ half of the year. And accordingly, we forecast fiscal '24 server storage
125
+ revenue to decline around the 20% range year-on-year.
126
+ Moving on to broadband. Q2 revenue declined 39% year-on-year to $730
127
+ million and represented 10% of semiconductor revenue. Broadband remains
128
+ weak on the continued pause in telco and service provider spending. We
129
+ expect Broadcom to bottom in the second half of the year with a recovery in
130
+ 2025. Accordingly, we are revising our outlook for fiscal '24 broadband
131
+ revenue to be down high 30s year-on-year from our prior guidance for a
132
+ decline of just over 30% year-on-year.
133
+ Finally, Q2 industrial rev -- resale of $234 million declined 10% year-on-year.
134
+ And for fiscal '24, we now expect industrial resale to be down double-digit
135
+ percentage year-on-year compared to our prior guidance for high single-digit
136
+ decline.
137
+ So to sum it all up, here's what we are seeing. For fiscal '24, we expect
138
+ revenue from AI to be much stronger at over $11 billion. Non-AI
139
+ semiconductor revenue has bottomed in Q2 and is likely to recover modestly
140
+ for the second half of fiscal '24.
141
+ On infrastructure software, we're making very strong progress in integrating
142
+ VMware and accelerating its growth. Pulling all these three key factors
143
+ together, we are raising our fiscal '24 revenue guidance to $51 billion. And
144
+ with that, let me turn the call over to Kirsten.
145
+ Kirsten Spears
146
+ Thank you, Hock. Let me now provide additional detail on our Q2 financial
147
+ performance, which included a full quarter of contribution from VMware.
148
+ Consolidated revenue was $12.5 billion for the quarter, up 43% from a year
149
+ ago. Excluding the contribution from VMware, Q2 revenue increased 12%
150
+ year-on-year. Gross margins were 76.2% of revenue in the quarter. Operating
151
+ expenses were $2.4 billion and R&D was $1.5 billion, both up year-on-year
152
+ primarily due to the consolidation of VMware.
153
+ Q2 operating income was $7.1 billion and was up 32% from a year ago with
154
+ operating margin at 57% of revenue. Excluding transition costs, operating
155
+ profit of $7.4 billion was up 36% from a year ago, with operating margin of
156
+ 59% of revenue. Adjusted EBITDA was $7.4 billion or 60% of revenue. This
157
+ figure excludes $149 million of depreciation. Now a review of the P&L for our
158
+ two segments, starting with semiconductors. Revenue for our semiconductor
159
+ solutions segment was $7.2 billion and represented 58% of total revenue in
160
+ the quarter. This was up 6% year-on-year.
161
+ Gross margins for our semiconductor solutions segment were approximately
162
+ 67%, down 370 basis points year-on-year, driven primarily by a higher mix of
163
+ custom AI accelerators. Operating expenses increased 4% year-on-year to
164
+ $868 million on increased investment in R&D, resulting in semiconductor
165
+ operating margins of 55%.
166
+ Now moving on to infrastructure software. Revenue for infrastructure software
167
+ was $5.3 billion, up 170% year-on-year, primarily due to the contribution of
168
+ VMware and represented 42% of revenue. Gross margin for infrastructure
169
+ software were 88% in the quarter, and operating expenses were $1.5 billion in
170
+ the quarter, resulting in infrastructure software operating margin of 60%.
171
+ Excluding transition costs, operating margin was 64%.
172
+ Now moving on to cash flow. Free cash flow in the quarter was $4.4 billion
173
+ and represented 36% of revenues. Excluding cash used for restructuring and
174
+ integration of $830 million, free cash flows of $5.3 billion were up 18% year-on
175
+ -year and represented 42% of revenue. Free cash flow as a percentage of
176
+ revenue has declined from 2023 due to higher cash interest expense from
177
+ debt related to the VMware acquisition and higher cash taxes due to a higher
178
+ mix of US income and the delay in the reenactment of Section 174.
179
+ We spent $132 million on capital expenditures. Days sales outstanding were
180
+ 40 days in the second quarter, consistent with 41 days in the first quarter. We
181
+ ended the second quarter with inventory of $1.8 billion down 4% sequentially.
182
+ We continue to remain disciplined on how we manage inventory across our
183
+ ecosystem. We ended the second quarter with $9.8 billion of cash and $74
184
+ billion of gross debt. The weighted average coupon rate and years to maturity
185
+ of our $48 billion in fixed rate debt is 3.5% and 8.2 years respectively.
186
+ The weighted average coupon rate and years to maturity of our $28 billion in
187
+ floating rate debt is 6.6% and 2.8 years, respectively. During the quarter, we
188
+ repaid $2 billion of our floating rate debt, and we intend to maintain this
189
+ quarterly repayment of debt throughout fiscal 2024. Turning to capital
190
+ allocation. In the quarter, we paid stockholders $2.4 billion of cash dividends
191
+ based on a quarterly common stock cash dividend of $5.25 per share.
192
+ In Q2, non-GAAP diluted share count was 492 million as the 54 million shares
193
+ issued for the VMware acquisition were fully weighted in the second quarter.
194
+ We paid $1.5 billion withholding taxes due on vesting of employee equity,
195
+ resulting in the elimination of 1.2 million AVGO shares. Today, we are
196
+ announcing a 10-for-1 forward stock split of Broadcom's common stock to
197
+ make ownership of Broadcom stock more accessible to investors and to
198
+ employees.
199
+ Our stockholders of record after the close of market on July 11, 2024, will
200
+ receive an additional nine shares of common stock after the close of market
201
+ on July 12, with trading on a split-adjusted basis expected to commence at
202
+ market open on July 15, 2024. In Q3, reflecting a post-split basis, we expect
203
+ share count to be approximately 4.92 billion shares.
204
+ Now on to guidance. We are raising our guidance for fiscal year 2024
205
+ consolidated revenue to $51 billion and adjusted EBITDA to 61%. For
206
+ modeling purposes, please keep in mind that GAAP net income and cash
207
+ flows in fiscal year 2024 are impacted by restructuring and integration-related
208
+ cash costs due to the VMware acquisition. That concludes my prepared
209
+ remarks. Operator, please open up the call for questions.
210
+ Question-and-Answer Session
211
+ Operator
212
+ Thank you. [Operator Instructions] And our first question will come from the
213
+ line of Vivek Arya with Bank of America. Your line is open.
214
+ Vivek Arya
215
+ Thanks for taking my question. Hock, I would appreciate your perspective on
216
+ the emerging competition between Broadcom and NVIDIA across both
217
+ Accelerators and Ethernet switching. So on the Accelerator side, they are
218
+ going to launch their Blackwell product that many of the same customers that
219
+ you have a very large position in the custom compute. So I'm curious how you
220
+ think customers are going to do that allocation decision, just broadly what the
221
+ visibility is.
222
+ And then I think Part B of that is as they launch their Spectrum-X Ethernet
223
+ switch, do you think that poses increasing competition for Broadcom and the
224
+ Ethernet switching side in AI for next year? Thank you.
225
+ Hock Tan
226
+ Very interesting question, Vivek. On AI accelerators, I think we are operating
227
+ on a different -- to start with scale, much as a different model. It is -- and on
228
+ the GPUs, which are the AI accelerator of choice on merchant -- in a
229
+ merchant environment is something that is extremely powerful as a model. It's
230
+ something that NVIDIA operates in, in a very, very effective manner.
231
+ We don't even think about competing against them in that space, not in the
232
+ least. That's where they're very good at and we know where we stand with
233
+ respect to that. Now what we do for very selected or selective hyperscalers is,
234
+ if there's a scale and the skills to try to create silicon solutions, which are AI
235
+ accelerators to do particular very complex AI workloads. We are happy to use
236
+ our IP portfolio to create those custom ASIC AI accelerator. So I do not see
237
+ them as truly competing against each other. And far for me to say I'm trying to
238
+ position myself to be a competitor on basically GPUs in this market. We're
239
+ not. We are not a competitor to them. We don't try to be, either.
240
+ Now on networking, maybe that's different. But again people may be
241
+ approaching and they may be approaching it from a different angle. We are as
242
+ I indicated all along, very deep in Ethernet as we've been doing Ethernet for
243
+ over 25 years, Ethernet networking. And we've gone through a lot of market
244
+ transitions, and we have captured a lot of market transitions from cloud-scale
245
+ networking to routing and now AI. So it is a natural extension for us to go into
246
+ AI. We also recognize that being the AI compute engine of choice in
247
+ merchants in the ecosystem, which is GPUs, that they are trying to create a
248
+ platform that is probably end-to-end very integrated.
249
+ We take the approach that we don't do those GPUs, but we enable the GPUs
250
+ to work very well. So if anything else, we supplement and hopefully
251
+ complement those GPUs with customers who are building bigger and bigger
252
+ GPU clusters.
253
+ Vivek Arya
254
+ Thank you.
255
+ Operator
256
+ One moment for our next question, and that will come from the line of Ross
257
+ Seymore with Deutsche Bank. Your line is open.
258
+ Ross Seymore
259
+ Hi guys. Thanks for taking my question. I want to stick on the AI theme, Hock.
260
+ The strong growth that you had in the quarter, the 280% year-over-year, could
261
+ you delineate a little bit between if that's the compute offload side versus the
262
+ connectivity side? And then as you think about the growth for the full year,
263
+ how are those split in that realm as well? Are they kind of going hand-in-
264
+ hand? Or is one side growing significantly faster than the other, especially
265
+ with the I guess, you said the next-generation accelerators are now going to
266
+ be Broadcom as well?
267
+ Hock Tan
268
+ Well, to answer your question on the mix, you are right. It's something we
269
+ don’t really predict very well, not understand completely except in hindsight.
270
+ Because it's like, to some extent, to the cadence of deployment of when they
271
+ put in the AI accelerators versus when they put in the infrastructure that puts it
272
+ together, the networking. And we don't really quite understand it 100%. All we
273
+ know, it used to be 80% accelerators, 20% networking. It's now running closer
274
+ to two-thirds accelerators, one-thirds networking and we'll probably head
275
+ towards 60%-40% by the close of the year.
276
+ Ross Seymore
277
+ Thank you.
278
+ Operator
279
+ Thank you. One moment for our next question. And that will come from the
280
+ line of Stacy Rasgon with Bernstein. Your line is open.
281
+ Stacy Rasgon
282
+ Hi, guys. Thanks for taking my question. I wanted to ask about the $11 billion
283
+ AI guide. You'd be at $11.6 billion even if you didn't grow AI from the current
284
+ level in the second half. And it feels to me like you're not suggesting that. It
285
+ feels to me like you think you could be [guided] (ph). So why wouldn't that AI
286
+ number be a lot more than $11.6 billion? It feels like it ought to be. Or am I
287
+ missing something?
288
+ Hock Tan
289
+ Because I guided just over $11 billion, Stacy could be what you think it is. It's
290
+ -- quarterly shipments get sometimes very lumpy. And it depends on rate of
291
+ deployment, depends on a lot of things. So you may be right. You may
292
+ estimate it better than I do, but the general trajectory is getting better.
293
+ Stacy Rasgon
294
+ Okay. So I guess again, how do I -- are you just suggesting that, that more
295
+ than $11 billion is sort of like the worst it could be because that would just be
296
+ flat at the current levels, but you're also suggesting that things are getting
297
+ better into the back half?
298
+ Hock Tan
299
+ Correct.
300
+ Stacy Rasgon
301
+ Okay. So I guess we just take that, that's a very -- if I'm reading it wrong,
302
+ that's just a very conservative number?
303
+ Hock Tan
304
+ That's the best forecast I have at this point, Stacy.
305
+ Stacy Rasgon
306
+ All right. Okay, Hock, thank you. I appreciate it.
307
+ Hock Tan
308
+ Thank you.
309
+ Operator
310
+ One moment for our next question, and that will come from the line of Harlan
311
+ Sur with JPMorgan. Your line is open.
312
+ Harlan Sur
313
+ Yeah, good afternoon. Thanks for taking my question. Hock, on cloud and AI
314
+ networking silicon, good to see that the networking mix is steadily increasing.
315
+ Like clockwork, the Broadcom team has been driving a consistent two year
316
+ cadence, right of new product introductions, Trident, Tomahawk, Jericho
317
+ family of switching and routing products for the past seven generations. You
318
+ layer on top of that your GPU -- TPU customers are accelerating their
319
+ cadence of new product introductions and deployments of their products.
320
+ So is this also driving faster adoption curve for your latest Tomahawk and
321
+ Jericho products? And then maybe just as importantly, like clockwork, it is
322
+ been two years since you've introduced Tomahawk 5 product introduction,
323
+ right which if I look back historically, means you have silicon and are getting
324
+ ready to introduce your next-generation three-nanometer Tomahawk 6
325
+ products, which would, I think, puts you two years to three years ahead of
326
+ your competitors. Can you just give us an update there?
327
+ Hock Tan
328
+ Harlan, you're pretty insightful there. Yes, we launched Tomahawk 5 in 2023.
329
+ So you're right, by late 2025, the time we should be coming out with
330
+ Tomahawk 6, which is the 100 terabit switch, yes.
331
+ Harlan Sur
332
+ And is the -- is this acceleration of cadence by your GPU and TPU partners, is
333
+ that also what's kind of driving the strong growth in the networking products?
334
+ Hock Tan
335
+ Well, you know what, sometimes you have to let things take its time. But it's
336
+ two-year cadence so we're right on. Late 2023, once when we shoot it out to a
337
+ Tomahawk 5 and adoption. You're correct with AI has been tremendous
338
+ because it ties in with the need for a very large bandwidth in the networking,
339
+ in the fabric for AI clusters, AI data centers. But regardless, we have always
340
+ targeted Tomahawk 6 to be out two years after that, which should put it into
341
+ late '25.
342
+ Harlan Sur
343
+ Okay, thank you Hock.
344
+ Operator
345
+ Thank you. One moment for our next question, and that will come from the
346
+ line of Ben Reitzes with Melius. Your line is open.
347
+ Ben Reitzes
348
+ Hi, thanks a lot. And congrats on the quarter and guide. Hock, I wanted to talk
349
+ a little bit more about VMware. Just wanted to clarify if it is indeed going better
350
+ than expectations. And how would you characterize the customer willingness
351
+ to move to subscription? And also just a little more color on Cloud Foundation.
352
+ You've cut the price there, and are you seeing that beat expectations? Thanks
353
+ a lot.
354
+ Hock Tan
355
+ Thanks, and thanks for your kind regards on the quarter. But it's -- as far as
356
+ VMware is concerned, we're making good progress. The journey is not over
357
+ by any means, but it is pretty much very much to expectation. Moving to
358
+ subscription, well, in VMware we are very slow compared to, I mean a lot of
359
+ other guys, Microsoft, Salesforce, Oracle, who have already been pretty much
360
+ in subscription. So VMware is late in that process. But we're trying to make up
361
+ for it by offering it and offering it in a very, very compelling manner because
362
+ subscription is the right things to do, right?
363
+ It's a situation where you put out your product offering, and you update it,
364
+ patch it, but update it feature-wise, everything as capabilities on a continual
365
+ basis, almost like getting your news on an ongoing basis, subscription online
366
+ versus getting it in a printed manner once a week. That's how I compare
367
+ perpetual to subscription. So it is very interesting for a lot of people to want to
368
+ can't get on. And so to no surprise, they are getting on very well. The big
369
+ selling point we have as I indicated, is the fact that we are not just trying to
370
+ keep customers kind of stuck on just server or compute virtualization.
371
+ That's a great product, great technology, but it's been out for 20 years. Based
372
+ on what we are offering now at a very compelling price point, compelling in a
373
+ very attractive price point, the whole stack, software stack to use vSphere and
374
+ its basic fundamental technology to virtualize networking, storage, operation
375
+ and management, the entire data center and create this self-service private
376
+ cloud.
377
+ And thanks for saying it, you're right, and we have priced it down to the point
378
+ where it is comparable with just compute virtualization. So yes, that is getting
379
+ a lot of interest, a lot of attention from the customers. We have signed up who
380
+ would like to deploy -- the ability to deploy private cloud, their own private
381
+ cloud on-prem. As a nice complement, maybe even alternative or hybrid to
382
+ public clouds, that's the selling point, and we are getting a lot of interest from
383
+ our customers in doing that.
384
+ Ben Reitzes
385
+ Great. And it's on track for $4 billion by the fourth quarter still, which is
386
+ reiterated?
387
+ Hock Tan
388
+ Well, I didn't give a specific time frame, did I? But it's on track as we see this
389
+ process growing towards a $4 billion quarter.
390
+ Ben Reitzes
391
+ Okay, thanks a lot Hock.
392
+ Hock Tan
393
+ Thanks.
394
+ Operator
395
+ Thank you. One moment for our next question, and that will come from the
396
+ line of Toshiya Hari with Goldman Sachs. Your line is open.
397
+ Toshiya Hari
398
+ Hi, thank you so much for taking my question. I guess kind of a follow-up to
399
+ the previous question on your software business. Hock, you seem to have
400
+ pretty good visibility into hitting that $4 billion run rate over the medium term,
401
+ perhaps. You also talked about your operating margin in that business
402
+ converging to classic Broadcom levels. I know the integration is not done and
403
+ you're still kind of in debt paydown mode. But how should we think about your
404
+ growth strategy beyond VMware? Do you think you have enough drivers, both
405
+ on the semiconductor side and the software side to continue to drive growth
406
+ or is M&A still an option beyond VMware? Thank you.
407
+ Hock Tan
408
+ Interesting question. And you're right. As I indicated in my remarks, even we
409
+ found the contribution from VMware this past quarter where we have AI
410
+ helping us, but we have non-AI semiconductor sort of bottoming out. We're
411
+ able to show 12% organic growth year-on-year. So almost have to say, so do
412
+ we need to rush to buy another company? Answer is no. But all options are
413
+ always open because we are trying to create the best value for our
414
+ shareholders who have entrusted us with the capital to do that.
415
+ So I would not discount that alternative because our strategy, our long-term
416
+ model has always been to grow through a combination of acquisition, but also
417
+ on the assets we acquire to really improve, invest, and operate them better to
418
+ show organic growth as well. But again, organic growth often enough is
419
+ determined very much by how fast your market would grow. So we do look
420
+ towards acquisitions now and then.
421
+ Toshiya Hari
422
+ All right. Thank you.
423
+ Operator
424
+ Thank you. One moment for our next question, and that will come from the
425
+ line of Blayne Curtis with Jefferies. Your line is open.
426
+ Blayne Curtis
427
+ Hi, thanks for taking my question. I wanted to ask you Hock, on the
428
+ networking business kind of ex AI. Obviously, I think there's an inventory
429
+ correction the whole industry is seeing. But just kind of curious, I don't think
430
+ you mentioned that it was at a bottom. So just the perspective, I think it's
431
+ down about [60%] (ph) year-over-year. Is that business finding a bottom? I
432
+ know you said overall whole semi business should -- non-AI should see a
433
+ recovery. Are you expecting any there any perspective on just customer
434
+ inventory levels in that segment?
435
+ Hock Tan
436
+ We see it behaving. I didn't particularly call it out, obviously because more
437
+ than anything else, I kind of link it very much to server storage, non-AI that is.
438
+ And we call server storage as at a bottom Q2, and we call it to recover
439
+ modestly second half of the year. We see the same thing in networking, which
440
+ is a combination of enterprise networking, as well as the hyperscalers who run
441
+ their traditional workloads on those, though it's hard to figure out sometimes.
442
+ But it is. So we see the same trajectory as we are calling out on server
443
+ storage.
444
+ Blayne Curtis
445
+ Okay, thank you.
446
+ Operator
447
+ Thank you. One moment for our next question, and that will come from the
448
+ line of Timothy Arcuri with UBS. Your line is open. Mr. Arcuri, your line is
449
+ open.
450
+ Timothy Arcuri
451
+ Hi, sorry. Thanks. Hock, is there a way to sort of map GPU demand back to
452
+ your AI networking opportunity? I think I've heard you say in the past that if
453
+ you spent $10 billion on GPU compute, you need to spend another $10 billion
454
+ on other [infrastructure] (ph), most of which is networking. So I'm just kind of
455
+ wondering if when you see these big GPU numbers, is there sort of a rule of
456
+ thumb that you use to map it back to what the opportunity will be for you?
457
+ Thanks.
458
+ Hock Tan
459
+ There is, but it's so complex, I stopped creating such a model, Tim. I've said it.
460
+ But there is because one would say that for every -- you almost say, for every
461
+ $1 billion you spend on GPU, you probably would spend probably on
462
+ networking, and if we include the optical interconnects as part of it, though we
463
+ are not totally in that market, except for the components like DSPs, lasers,
464
+ PIN diodes that go into those, high-bandwidth optical connect. But if you just
465
+ take optical connects in totality, switching, all the networking components, it
466
+ goes into -- attaches itself to clustering a bunch of GPUs, you probably would
467
+ say that about 25% of the value of the GPU goes to networking, the rest on
468
+ networking.
469
+ Now not entirely all of it is my available market. I don't do the optical connects,
470
+ but I do the few components I talked about in it. But roughly, the simple way to
471
+ look at it is probably about 25%, maybe 30% of all these infrastructure
472
+ components is kind of attached to the GPU value point itself. But having said
473
+ that, it's never – we are never that precise that deployment is the same way.
474
+ So you may see the deployment of GPU or purchase of GPU much earlier.
475
+ And the networking comes later or sometimes less the other way around,
476
+ which is why you're seeing the mix going on within my AI revenue mix. But
477
+ typically, you run towards that range over time.
478
+ Timothy Arcuri
479
+ Perfect Hock, thank you so much.
480
+ Operator
481
+ Thank you. One moment for our next question, and that will come from the
482
+ line of Thomas O'Malley with Barclays. Your line is open.
483
+ Thomas O’Malley
484
+ Hi, guys. Thanks for taking my question. And nice results. My question in
485
+ regards to the custom ASIC AI. Hock, you had a long run here of a very
486
+ successful business, particularly with one customer. If you look in the market
487
+ today, you have a new entrant who's playing with different customers. And I
488
+ know that you said historically, that's not really a direct customer to you. But
489
+ could you talk about what differentiates you from the new entrant in the
490
+ market as of late? And then there's been profitability questions around the
491
+ sustainability of gross margins longer term. Can you talk about if you see any
492
+ increased competition? And if there's really areas that you would deem more
493
+ or less defensible in your profile today? And if you would see kind of that
494
+ additional entrant maybe attack any of those in the future?
495
+ Hock Tan
496
+ Let me take the second part first, which is our AI -- custom AI accelerator
497
+ business. It is a very profitable business, and let me put to scale -- look
498
+ examine from a model point of view. I mean, each of these AI accelerators no
499
+ different from a GPU. The way these large language models get run
500
+ computing, get run on these accelerators, no one single accelerator, as you
501
+ know, can run these big large language models. You need multiple of it no
502
+ matter how powerful those accelerators are.
503
+ But also, and the way the models are run, there is a lot of memory access to
504
+ memory requirements. So each of this accelerator comes with a large amount
505
+ of cache memory, as you call it, what you guys probably now know as HBM,
506
+ high-bandwidth memory specialized for AI accelerators or GPUs. So we're
507
+ supplying both in our custom business.
508
+ And the logic side of it, where the compute function is on doing the chips, the
509
+ margin there are no different than the margin in any -- in most of any of a
510
+ semiconductor silicon chip business. But when you're attached to it, a huge
511
+ amount of memory, memory comes from a third-party. There are a few
512
+ memory makers who make this specialized thing. We don't do margin
513
+ stacking on that part. So by almost buying basic math will dilute the margin of
514
+ these AI accelerators when you sell them with memory, which we do. It does
515
+ push up revenue somewhat higher but it has diluted the margin.
516
+ But regardless, the spend, the R&D, the OpEx that goes to support this as a
517
+ percent of the revenue, which is higher revenue, so much less. So on an
518
+ operating margin level, this is easily as profitable, if not more profitable, given
519
+ the scale that each of those custom AI accelerators can go up to. It's even
520
+ better than our normal operating margin scale. So that's the return on
521
+ investment that attracts and keeps us going at this game. And this is more
522
+ than a game. It is a very difficult business. And to answer your first question,
523
+ there is only one Broadcom, period.
524
+ Thomas O'Malley
525
+ Thanks Hock.
526
+ Operator
527
+ Thank you. One moment for our next question, and that will come from the
528
+ line of Karl Ackerman with BNP. Your line is open.
529
+ Karl Ackerman
530
+ Hi, thank you. Good afternoon. Hock, your networking switch portfolio with
531
+ Tomahawk and Jericho chipsets allows hyperscalers to build AI clusters using
532
+ either a switch-scheduled or endpoint-scheduled network. And that, of course
533
+ is unique among competitors. But as hyperscalers seek to deploy their own
534
+ unique AI clusters, are you seeing a growing mix of white-box networking
535
+ switch deployments? I ask because while your custom sales and business
536
+ continues to broaden, it will be helpful to better understand the growing mix of
537
+ your $11 billion AI networking portfolio combined this year. Thank you.
538
+ Hock Tan
539
+ Let me have Charlie address this question. He's the expert.
540
+ Charlie Kawwas
541
+ Yes. Thank you, Hock. So two quick things on this. One is the – you are
542
+ exactly right that the portfolio we have is quite unique in providing that
543
+ flexibility. And by the way, this is exactly why Hock, in his statements earlier
544
+ on, mentioned that seven out of the top eight hyperscalers use our portfolio.
545
+ And they use it specifically because it provides that flexibility. So whether you
546
+ have an architecture that's based on an endpoint and you want to actually
547
+ build your platform that way or you want that switching to happen in the fabric
548
+ itself, that's why we have the full end-to-end portfolio. So that actually has
549
+ been a proven differentiator for us.
550
+ And then on top of that, we've been working, as you know, to provide a
551
+ complete network operating system that's open on top of that using SONiC
552
+ and Psi, which has been deployed in many of the hyperscalers. And so the
553
+ combination of the portfolio plus the stack really differentiates the solution that
554
+ we can offer to these hyperscalers. And if they decide to build their own NICs,
555
+ their own accelerators are custom or use standard products, whether it is from
556
+ Broadcom or other, that platform, that portfolio of infrastructure switching
557
+ gives you that full flexibility.
558
+ Karl Ackerman
559
+ Thank you.
560
+ Operator
561
+ Thank you. One moment for our next question, and that will come from the
562
+ line of C.J. Muse with Cantor Fitzgerald. Your line is open.
563
+ CJ Muse
564
+ Yeah. Good afternoon. Thank you for taking my question. I was hoping to ask
565
+ two part software question. So excluding VMware, your Brocade, CA, and
566
+ Symantec business now running $500 million higher for the last two quarters.
567
+ So curious, is that the new sustainable run rate or were there onetime events
568
+ in both January and April that we should be considering?
569
+ And then the second question is as you think about VMware Cloud
570
+ Foundation adoption, are you seeing any sort of crowding out of spending like
571
+ other software guys are seeing as they repurpose their budgets to IT? Or is
572
+ that business so less discretionary that it's just not an impact to you? Thanks
573
+ so much.
574
+ Hock Tan
575
+ Well, on the second one, I don't know about any crowding out, to be honest.
576
+ It's not. What we are offering, obviously, is not something that they would like
577
+ to use themselves, to be able to do themselves, which is they're already
578
+ spending on building their own on-prem data centers. And typical approach
579
+ people take, a lot of enterprises take historically continue today than most
580
+ people do a lot, people do is they have best of breed.
581
+ What I mean is they create a data center that is compute as a separate
582
+ category, best compute there is and they often enough use vSphere for
583
+ compute virtualization due to improved productivity, but best of breed there.
584
+ And best of breed on networking and best of breed on storage with a common
585
+ management operations layer, which very often is also VMware we realize.
586
+ And what we're trying to say is this mixed bag, and what they see -- is this
587
+ mixed bag best of [big] (ph) data center, very heterogenous, is not driving
588
+ that, is not a highly resilient data center.
589
+ I mean, you have a mixed bag. So it goes down. Where do you find quickly
590
+ root cause? Everybody is pointing fingers at the other. So you got a problem,
591
+ not very resilient and not necessary secure between bare metal in one side
592
+ and software on the other side.
593
+ So it's a natural thinking on the part of many CIOs we talk to, to say, hey, I
594
+ want to create one common platform as opposed to just [best-of-breed of age]
595
+ (ph). So that gets us into that. So it is a greenfield that’s not bad, they started
596
+ from scratch. If it's a brownfield, that means they have existing data centers
597
+ trying to upgrade. It's -- that sometimes that's more challenging for us to get
598
+ that adopted.
599
+ So I'm not sure there's a crowding out here. There's some competition,
600
+ obviously, on greenfield, where they can spend their budget on an entire
601
+ platform versus best-of-breed. But on the existing data center where you're
602
+ trying to upgrade, that's a trickier thing to do. And it cuts the other way as well
603
+ for us. So that's how I see it. So in that sense, best answer is I don't think
604
+ we're seeing a level of crowding out that is -- any and that very significant for
605
+ me to mention.
606
+ In terms of the revenue mix, no, Brocade is having a great, great field year so
607
+ far and still chugging along. But will that sustain? Hell no, you know that.
608
+ Brocade goes through cycles like most enterprise purchases. So we're
609
+ enjoying it while it lasts.
610
+ CJ Muse
611
+ Thank you.
612
+ Hock Tan
613
+ Thanks.
614
+ Operator
615
+ Thank you. And we do have time for one final question, and that will come
616
+ from the line of William Stein with Truist Securities. Your line is open.
617
+ William Stein
618
+ Great. Thanks for squeezing me in. Hock, congrats on the yet another great
619
+ quarter and a strong outlook in AI. I also want to ask about something you
620
+ mentioned with VMware. In your prepared remarks, you highlighted that
621
+ you've eliminated a tremendous amount of channel conflict. I'm hoping you
622
+ can linger on this a little bit and clarify maybe what you did. And specifically
623
+ also what you did in the heritage Broadcom software business, where I think
624
+ historically, you've shied away from the channel. And there was an idea that
625
+ perhaps you'd reintroduce those products to the channel through a more
626
+ unified approach using VMware's channel partners or resources. So any sort
627
+ of clarification here, I think, would be helpful.
628
+ Hock Tan
629
+ Yes, thank you. That's a great question. Yes, VMware taught me a few things.
630
+ They have 300,000 customers, 300,000. That's pretty amazing. And we look
631
+ at it. I know under CA, we took a position that let's pick an A-list strategic guy
632
+ and focus on it. I can't do that in VMware. I approached it differently. And I
633
+ start to learn the value of a very strong bunch of partners they have, which are
634
+ a network of distributors and something like 15,000 VARs, value-added
635
+ resales supported with these distributors.
636
+ So we have doubled down and invested in this resale network in a big way for
637
+ VMware. It's a great move, I think but six months into the game. But we are
638
+ seeing a lot more velocity out of it. Now these resellers, having said that, tend
639
+ to be very focused on a very long tail of their 300,000 customers. The largest
640
+ 10,000 customers of VMware are large enterprises who tend to -- they are
641
+ very large enterprises, the largest banks, the largest health care companies.
642
+ And their view is I want very bespoke service, support, engineering solutions
643
+ from us. So we've created a direct approach, supplemented with the VAR of
644
+ choice where they need to. But on the long tail of 300,000 customers, they get
645
+ a lot of services from the resellers, value-added resellers, and so in their way.
646
+ So we now strengthen that whole network of resellers so that they can go
647
+ direct, manage, supported financially with distributors.
648
+ And we don't try to challenge those guys unless the customers. On the end of
649
+ the day, the customer chose where they like to be supported. So we kind of
650
+ simplify this together with the number of SKUs they have. In the past, unlike
651
+ what we're trying to do here, everybody is a partner. I mean, you're talking a
652
+ full range of partners. And whoever makes the biggest deal gets the lowest.
653
+ The partner that makes the biggest deal gets the lower -- biggest discount,
654
+ lowest price. And they are out there basically kind of creating a lot of channel
655
+ chaos and conflict in the marketplace.
656
+ Here, we don't. The customers, I am aware. They can take it direct from
657
+ VMware to their direct sales force or they can easily move to the resellers to
658
+ get it that way. And as a third alternative which we offer, if they chose not --
659
+ they want to run their applications on VMware and they want to run it
660
+ efficiently on a full stack. They have a choice now of going to a hosted
661
+ environment managed by network of managed service providers, which we
662
+ set up globally, that will run the infrastructure, invest and operate the
663
+ infrastructure. And these enterprise customers just run their workloads in and
664
+ get it as a service, basically VMware as a service. That's a third alternative,
665
+ and we are clear to make it very distinct and differentiated for our end-use
666
+ customers. They're available to all three is how they choose to consume our
667
+ technology.
668
+ William Stein
669
+ Great. Thank you.
670
+ Operator
671
+ Thank you. I would now like to hand the call over to Ji Yoo, Head of Investor
672
+ Relations, for any closing remarks.
673
+ Ji Yoo
674
+ Thank you, Cherie. Broadcom currently plans to report its earnings for the
675
+ third quarter of fiscal '24 after close of market on Thursday, September 5,
676
+ 2024. A public webcast of Broadcom's earnings conference call will follow at
677
+ 2:00 p.m. Pacific Time. That will conclude our earnings call today. Thank you
678
+ all for joining. Operator, you may end the call.
679
+ Operator
680
+ Thank you all for participating. This concludes today's program. You may now
681
+ disconnect.
682
+ ",,,2024-07-25 21:01:29.337394