NVIDIA Corporation (NASDAQ:NVDA) Q4 2021 Earnings Conference Call February 24, 2021 5:00 PM ET
Company Participants
Jensen Huang – President and Chief Executive Officer
Simona Jankowski – Vice President of Investor Relations and Strategic Finance
Colette Kress – Executive Vice President and Chief Financial Officer
Conference Call Participants
C.J. Muse – Evercore ISI
Vivek Arya – BofA Securities
Stacy Rasgon – Bernstein Research
Timothy Arcuri – UBS
John Pitzer – Credit Suisse
Aaron Rakers – Wells Fargo
Mark Lipacis – Jefferies
Operator
Good afternoon. My name is Mariama, and I will be your conference operator today. At this time, I would like to welcome everyone to NVIDIA's Financial Results Conference Call. All lines have been placed on mute to prevent any background noise. After the speakers' remarks, there will be a question-and-answer session. [Operator Instructions] Thank you.
I will now turn the call over to Simona Jankowski, NVIDIA's Vice President of Investor Relations and Strategic Finance to begin the conference.
Simona Jankowski
Thank you. Good afternoon, everyone, and welcome to NVIDIA's conference call for the fourth quarter of fiscal 2021. With me on the call today from NVIDIA are Jensen Huang, President and Chief Executive Officer; and Colette Kress, Executive Vice President and Chief Financial Officer.
I'd like to remind you that our call is being webcast live on NVIDIA's Investor Relations website. The webcast will be available for replay until the conference call to discuss our financial results for the first quarter of fiscal 2022. The content of today's call is NVIDIA's property. It can't be reproduced or transcribed without our prior written consent.
During this call, we may make forward-looking statements based on current expectations. These are subject to a number of significant risks and uncertainties, and our actual results may differ materially. For a discussion of factors that could affect our future financial results and business, please refer to the disclosure in today's earnings release, our most recent forms 10-K and 10-Q and the reports that we may file on Form 8-K with the Securities and Exchange Commission.
All our statements are made as of today, February 24, 2021, based on information currently available to us. Except as required by law, we assume no obligation to update any such statements. During this call, we will discuss non-GAAP financial measures. You can find a reconciliation of these non-GAAP financial measures to GAAP financial measures in our CFO commentary, which is posted on our website.
With that, let me turn the call over to Colette.
Colette Kress
Thanks, Simona. Q4 was another record quarter with revenue exceeding $5 billion and year-on-year growth accelerating to 61%. Full-year revenue was also a record at $16.7 billion, up 53%. Our Gaming business have reached record revenue of $2.5 billion in Q4, up 10% sequentially, and up 67% from a year earlier. Full-year gaming revenue was a record at $7.8 billion, up 41%.
Demand is incredible for our new GeForce RTX 30 Series products based on the NVIDIA Ampere GPU architecture. In early December, we launched the GeForce RTX 3060 Ti, which joined the previously launched RTX 3090, 3080, and 3070. The entire 30 Series lineup has been hard to keep in stock and we exited Q4 with channel inventories even lower than when we started.
Although we are increasing supply, channel inventories will likely remain low throughout Q1. GeForce RTX 30 series graphics cards were a holiday sensation do not just do their amazing performance, but also to the rich features, including our second-generation RTX ray tracing technologies and DLSS, AI-powered performance accelerator, which massively boost frame rates in graphically demanding titles.
Three dozen games now all support RTX, including the top Battle Royale game, Fortnite; the top role-playing game, Cyberpunk 2077; the top massively multiplayer online game, World of Warcraft; and the best-selling game of all time, Minecraft. RTX has clearly set the new standard in gaming.
Building on this momentum at CES in January, we introduced a wave of Ampere architecture gaming products, including our biggest ever laptop launch powered by GeForce RTX 3060, 3070, and 3080 laptop GPUs and with our third-generation Max-Q technology. These new thin and lightweight gaming laptops increased performance and energy efficiency by up to 2x from the prior generation.
RTX 3060 laptops start $999 and are faster than the previous generation laptops, which sold for $2,500. The incredible performance, design and price points of these new laptops will delight the growing universe of gamers and creators as well as students and professionals.
The gaming laptop market has grown seven-fold in the past seven years and momentum is building. With top OEMs bringing to market a record 70-plus laptop models based on the GeForce RTX 30 series. GeForce laptops as a whole are the fastest-growing and one of the largest gaming platforms.
Also at CES, we announced the GeForce RTX 3060 GPU, priced at $329, extending the 30 series desktop liner further into the mainstream. We expect strong demand when it launches this Friday as 60-class GPUs have traditionally been our most popular products. Starting with the 3060, we're taking an important step to maximize the supply of GeForce GPUs for gamers. Users are constantly discovering new applications for our powerful programmable GPUs and cryptocurrency mining is one of them.
With rising Ethereum prices, there are indications that minors are behind GPUs. We would like GeForce GPUs to end up with gamers. So we have created a new special software drivers that will detect the Ethereum mining algorithm cutting in half the mining efficiency of the GeForce RTX 3060.
We suspect the significant increase in the Ethereum network hash rate observed over the past few months was driven by a combination of previously installed mining capacities that was reactivated as well as new sales of GPUs and ASICs. Since our GPUs are sold to graphics card manufacturers and then onto distribution, we don't have the ability to accurately track or quantify their end use.
Analyst estimate suggest that cryptomining contributed 100 million to 300 million to our Q4 revenue, a relatively small portion of our gaming revenue in Q4. Cryptocurrencies have recently started to be accepted by companies and financial institutions and show increased signs of staying power to address industrial Ethereum mining demand, last week we announced a new line of NVIDIA CMPs or cryptomining processors. Shipments will start in March. CMPs lack display outputs and have other optimizations that improve cryptomining power efficiency. CMP products will let us gain some visibility into the contribution of cryptomining to our overall revenue.
For Q1, we estimate that CMP will contribute approximately 15 million. We plan to sell these products to industrial miners. We will quantify their contribution each quarter for transparency. Over the past year, it has become clear that we've entered a new era in which gaming is an integral part of global culture. The number of concurrent users on Steam has more than doubled since 2018 and continues to hit new records.
In 2020 alone, more than 100 billion hours of gaming content was seen on YouTube and 0.5 billion people watched e-sports. Increasingly, we aren't just gaming. We're also watching sports, attending concerts, creating content and connecting with our friends in virtual environments.
Additionally, we are excited about the new experiences like VR. Significantly more content is now available, including arguably the first VR killer app, Beat Saber, and there is now almost 2 million VR users on Steam.
And with these powerful structural shifts, we expect our gaming business to remain on a robust growth trajectory. The GeForce RTX 30 Series GPUs have kicked off a powerful upgrade cycle and we estimate only around 15% of GeForce gamers on an RTX cost GPU, which is needed to experience the beautiful ray trace graphics of modern games. Moreover, the universe of gamers is rapidly expanding and the reach of GeForce has extended beyond gamers to some 45 million creators.
In addition, gaming revenue continues to benefit from a favorable mix shift as gamers and creators keep moving to higher-end GPUs, we expect another great year for GeForce.
Earlier this month, we celebrated the one year anniversary of the GeForce NOW cloud gaming platform, which is now over 6 million members strong. GeForce now offers 800 PCs from over 300 publishers, more than any other cloud gaming service including any of the most played free to play games, starting with support for windows PCs, Macs, and Android devices we added support in recent months to Chromebooks, iPhones and iPads. GFN has grown globally with more than 65 countries on our service and more – added regularly by our GeForce NOW alliance partners.
Moving to Pro Vis, Q4 revenue was $307 million, up 30% sequentially and down 10% year-on-year and ahead of our expectations full-year revenue was $1.1 billion was down 13%. Strong sequential growth was driven primarily by a recovery in desktop workstations as some customers returned to the office and enterprises resumed purchases that had been deferred by the pandemic. Notebook GPUs grew sequentially to a record as enterprises continue to support remote workforce initiatives. Looking ahead, the reopening of businesses will benefit desktop workstations, but longer-term workforce trends will likely shift our mix to notebook GPUs and cloud offerings. Healthcare was a standout vertical in the quarter with significant orders from GE Siemens and Oxford Nanopore technologies, public sector and automotive also so strength. Omniverse are real time 3D collaboration and simulation platform is now an open beta. Over 500 creators and professionals have tested Omniverse through our early access program. Omniverse is one of our most important and exciting platforms. We are divided by its initial acceptance and look forward to sharing more details on its long-term growth opportunity in the coming months.
Moving to Automotive; Q4 revenue was $145 million, up 16% sequentially and down 11% year-on-year. Full-year revenue of $536 million, declined 23%. Sequential growth was driven by continued recovering in the global automotive production volumes and growth in AI cockpit revenue. Year-on-year decline reflects the expected ramp down of legacy infotainment. Nvidia has emerged as the industry's leading end-to-end full stack technology provider for self-driving and AI enabled vehicles or in the SOC that drive self-driving platform is built on delivers an unrivaled 254 trillions of operations per second of performance on industry leading power efficiency, helping to revolutionize the transportation industry.
Our technology leadership has driven a robust rapidly growing set of opportunities. We have great momentum with an expanding list of electric vehicle OEMs, including NIO, SAIC, Li Auto and Xpeng which are all using the Nvidia drive platform to power their next generation of vehicles. We look forward to growing with them as they continue to scale.
Our software defined platforms is the only solution that spans from the data center for training deep neural net and running physically accurate simulations to a full stack in car solutions scaling for ADAS to Level 5 fully autonomous functionality. Autonomous vehicle companies are harnessing this technology. Zoox recently unveiled its Level 5 bi-directional robotaxi powered by Nvidia, and wide launched its next generation countless autonomous truck using NVIDIA Drive Orin.
And earlier this year, Mercedes announced a 56-inch wide MBUX Hyperscreen powered by Nvidia AI cockpit technology. This will builds on our momentum with Mercedes first-generation MBUX system, which is now in 1.8 million cars. We are in the early innings of a significant opportunity. We have built a multi-billion dollar design win pipeline for our self-driving the AI cockpit solutions, which will drive a material inflection in revenue over the next few years. Our transformational partnership with Mercedes announced last June demonstrates the power of our evolving business model as we expand our addressable market and layer in software revenue. We are exceptionally well positioned to capitalize on the significant opportunity that lies ahead.
Moving to data center; revenue was $1.9 billion, which exceeded our expectations was comparable to last quarter and up 97% from the year ago period, which did not include Mellanox. Data center compute revenue was up 45% year-on-year. Full-year data center revenue rose 125% to a record $6.7 billion, including almost 70% growth from data center compute. From a sequential perspective, the data center computes stronger than expected double-digit growth more than offset the anticipated decline in Mellanox revenue, which included a large non-reoccurring network sale to a single OEM in Q3. Compute growth was lead by vertical industries where OEM partners continued ramping up their 100 – day 100 base servers and our own DGX system sales were strong.
Vertical industries were well over 50% of data center revenue across compute and networking with particular strength in super computing, financial services, higher education and consumer internet verticals. Additionally, hyperscale customers continued to deploy the A100 driving both sequential growth and exceptionally strong year-on-year growth in data center compute. The A100 has been adopted by all major cloud customers globally and is being deployed by hyperscale customers for internal workloads. Still we are in the early stages of adoption and expect continued growth this year. The ramp of the A100 has been smoother and accomplished by better visibility than prior generation. It's universal AI training and inference capabilities as well as support for wider set of applications and outstanding performance are driving five customer utilization; a clear sign of the A100's value.
Turning to Mellanox. We are seeing continued strong traction and robust momentum across our customer set. Its revenue was up over 30% from Mellanox's Q4 revenue in calendar 2019 when it was still a standalone company. Year-on-year growth in the quarter was led by hyperscale and large consumer internet customers, which grew over 60% from last year with several contributing record revenues. Consistent with our outlook Mellanox had a sequential decline impacted by a non-reoccurring sales to a China OEM in Q3.
We expect to return to sequential growth in Q1 driven by strong demand for our high speed networking products, including the ramp of ConnectX adapters with CSPs and all major server OEMs in their upcoming refresh. We also see strong momentum in high-performance computing with HDR InfiniBand products. For example, we won six of the seven supercomputers awarded over the past few months by Euro HPC. Starting next quarter, we will continue to provide color on networking as part of the data center market platform, but we will no longer break out Mellanox revenue separately.
Looking forward, we are incredibly excited about the opportunities and data center. Accelerated computing is not only delivering super more loss, gains and performance but there's also an energy efficient and cost effective method of computing and virtually every industry is adopting technology with greater urgency as companies adapt to the new world of more distributed workers and customers. As industries embark on this journey, they are also increasingly focused on combating climate change. To that end, the A100 performed AI computations with one-twentieth the power consumption of CPU's. And powers are saline supercomputer, which is number one on the Green500 list of the world's most efficient supercomputers. Indeed Nvidia's powered machines recently captured 25 of the top 30 spots on the Green500 list. Accelerated computing is not only serving the exponential growth and demand for compute; it can also help bend the power consumption curve. With accelerated computing Nvidia is pioneering a [indiscernible] forward the computing industry.
Before I moved to the P&L and outlook, let me give you an update on our proposed acquisition of Arm. In September, we announced plans to acquire Arm from SoftBank Group and a transaction that will create the premier computing company for the age of AI. At that time, we said it would take approximately 18 months to secure regulatory approvals in the U.S., the UK, the EU, China and other jurisdictions. Thorough reviews are typical with the deal of this size. This process is moving forward as expected. We are in constructive dialogue with the relevant authorities and are confident that regulators will see the benefits to the entire tech ecosystem. As we have said, this combination will spur competition together Arm and Nvidia will provide greater choice to the data center ecosystem, a compelling alternative CPU architecture for the market and further enhance Arms' offering in mobile and embedded.
Our intention is to increase investment in Arm's existing roadmap, adding resources to stimulate growth in new markets. We love and intent to maintain Arm's open licensing model, a commitment guaranteed both by long-term legally binding contracts, as well as our own interest in ensuring this investment is a profitable one for us. We are on the cusp of new age in which AI fuels industries ranging from healthcare to scientific research to the environment. With this transaction, our vision is to boost Arm's potential, so it can thrive in this new era and grow into promising new markets.
Moving to the rest of the P&L. Q4 GAAP gross margins were 63.1% and non-GAAP gross margins were 65.5%. GAAP gross margins declined year-on-year due to amortization of developed technology acquired from Mellanox, partially offset by product mix. The sequential increase was due to higher margins for gaming GPU and lower IP related costs, partially offset by lower margin mix in our data center portfolio. Non-GAAP gross margins increased by 10 basis points year-on-year and was flat sequentially in-line with our expectations. Q4 GAAP EPS was $2.31, up 51% from a year earlier. Non-GAAP EPS was $3.10, up 64% from a year ago. Q4 cash from operations was a record $2.07 billion.
With that, let me turn to the outlook for the first quarter of fiscal 2022.
Revenue is expected to be $5.3 billion plus or minus 2% with most of the sequential growth driven by gaming. GAAP and non-GAAP gross margins are expected to be 63.8% and 66% respectively plus or minus 50 basis points. GAAP and non-GAAP operating expenses are expected to be approximately $1.67 billion, and $1.2 billion respectively. For the full-year we expect to grow non-GAAP OPEX in the mid-20% range. GAAP and non-GAAP, other income and expenses are both expected to be an extent of approximately $50 million. GAAP and non-GAAP tax rates are both expected to be 10% plus or minus 1%, excluding discrete items. Capital expenditures are expected to be approximately $300 million to $325 million. Further financial details are included in the CFO commentary and other information on our IR website.
In closing, let me highlight upcoming events for the financial community. We will be virtually attending to Raymond James Institutional Investors Conference on March 1st. The Morgan Stanley Technology, Media and Telecom Conference on March 3rd; and the Arete Virtual Semis Conference on March 3rd. In addition, we will be hosting a virtual Investor Day on Monday, April 12th following a live stream of Jensen's opening keynote at our GPU Technology Conference. Our earnings call to discuss our first quarter and full – our first quarter is scheduled for Wednesday, May 26.
We will now open the call for questions. Operator, would you please poll for questions. Thank you.
Question-and-Answer Session
Operator
[Operator Instructions] Your first question comes from the line of C.J. Muse with Evercore ISI. Your line is open.
C.J. Muse
Good afternoon. Thank you for taking the question. I guess, Jensen, higher-level question for you on the enterprise side. You're now a couple of quarters into the ramp of A100.And curious if you could speak to whether you've seen any surprises here, any areas of specific strength worth calling out? And any changes to how you're thinking about the size of this opportunity?
Jensen Huang
Yes. Thanks a lot. It's Jensen. As you know, A100 is a very different type of GPU. This is our first universal computing GPU. It's great at high-performance computing. It's great at data analytics. It's great at training. And also for our highest GPU, it’s also the first time that is incredible for imprint. It's some 20 times faster than previous generation.
We may introduce this some really exciting new computational formats like TF32, TensorFloat-32 for training. And with a multi-instance GPU, turning our GPU 1, GPU to a whole bunch of smaller GPUs, autonomous GPU to improve performance and reducing latency.
And so the capability is really quite exciting. We're seeing strength in hyperscalers as they continue to accelerate their adoption of AI. Some of the new applications we've spoken about a couple of times before, the transition to deep learning, the conversational AI, speech recognition to natural language understanding all the way to speech synthesis, which is now based on AI – based on deep learning.
The other area that's growing incredibly fast is the deep learning recommender models. Just about everything that you do on the Internet is based on recommenders. There are hundreds of different recommenders out there, whether you're shopping or recommending music or recommending news or recommending search, and so all the recommending ads. And so all of these different types of applications are driving that.
For the first time, we saw our industrial application - industrial data center growing to be larger than hyperscale. And we're seeing industrial applications across scientific computing where simulation-based approaches are now being fused with AI approaches for weather simulation, genomics, molecular dynamics simulation, quantum chemistry, even simulating quantum computing, which is one of the really exciting areas.
We're seeing AI being deployed for big data analytics, RAPIDS, which is NVIDIA’s created open source platform for data analytics of Spark 3.0, which NVIDIA really led and is now GPU accelerated. So now you could have big data in the cloud, while doing big data analytics in the cloud on all of the CSP platforms.
You could – we're seeing a lot of excitement around financial services and consumer Internet services are all really growing nicely. And so A100 adoption is just starting. I mean, we're going to see several – couple of years of continued growth ahead of us, while as AI gets adopted in clouds and industries.
Operator
Your next question comes from the line of Vivek Arya with BofA Securities. Your line is open.
Vivek Arya
Thanks for taking my question. Just a clarification and then a question for Jensen. On the clarification, Colette, I was hoping if you could give a little more color around Q1? Do you still expect data center to grow sequentially in Q1? I know you said that most of the growth will come from gaming, but any color on the data center would be useful?
And then, Jensen, the question for you is, in your press release you used the phrase AI driving the smartphone moment for every industry. Could you help us quantify what that means? And where I'm going with that is, is there a number in terms of what percentage of servers are shipping today with your accelerators? And where can that ratio go over time? Is that a fair way of looking at the adoption of your technology and AI?
Colette Kress
So thank you, Vivek. Your question regarding the guidance as we beat into Q1. We had indicated that, yes, a good percentage of our growth between Q4 and Q1 those come from gaming, but we also do expect data center to grow. Most of our sequential growth coming from gaming, but keep in mind, we also expect all of our market platforms will likely be able to grow quarter-over-quarter.
Jensen Huang
Because we are entering in the third phase of AI. The first phase of AI was when we invented the computing platforms, the new chips, the new systems, the new system software, the new middleware, the new way of working, the new way of developing software, which the industry, the world is now starting to call ML Ops.
The way that software is developed and the way that is deployed is completely different in the past. In fact, I heard a great term Software 2.0, and it makes a lot of sense. It's a computer that is writing software. The way that you develop software is completely different; the way you compute is different. And that was our first phase, and that that started in a journey that was some eight, nine years ago now.
The second phase was the adoption of using this in an industrial way for clouds. And we saw it revolutionize new services; whether it's speech-oriented services or search-oriented services, recommended services, the way you shop, the way you use the Internet is completely different today. And so that's really the second phase and those two phases are still continuing to grow, and you're still seeing the growth associated with that.
The third phase is the industrialization of AI. And some of the great examples when I say in terms of smartphone moment, I meant that it's a device with AI, it’s autonomous and it's connected to a cloud service, and it's continuously learning. So some of the exciting example that I saw, that I've seen and we're working with companies all over the world, we have some 7,000 AI startups that we're working with, and almost all of them are developing something like this. And large industrial companies whether it's John Deere or Walmart, they're all developing application kind of like this.
And basically it's an autonomous system, autonomous machine. In our case it's called Jetson. It's a robotics machine. That robotics machine is a car and it’s called DRIVE. And it's running an autonomous - an AI application on top and AI is still on top, and this could be – they could be moving device – moving things around, it can be picking and placing. It could be just watching a warehouse and monitoring traffic and keeping traffic flow going. It could be connected to a car. And whenever the car – whenever the fleet of cars needs to be retrained because of a new circumstance that was discovered, the cloud service would do the relearning and then we’d deploy into all of the autonomous devices.
And so in the future, we're seeing these industries, whether you're in retail or in logistics or transportation, or farming, ag tech to lawnmowers – consumer lawnmowers. They're not going to just be products that you buy and use from that point forward, but it likely be a connected device with an AI service that runs on top of it. And so these industries I'm so excited about because it gives an opportunity to change the way that they interact with their customers. Rather than selling something once, they sell something and provide service that's on top of it. And they can stay engaged with the customers. The customers could get a product that's improving all of the time, just like your smartphone. And that's kind of like – that's kind of the reason – that's the reason why I've been calling it a smartphone moment for all these industry.
And we saw what happened to the smartphone revolution. And then we saw what happened to the smart microphone, the smart speaker revolution, and you're going to see smart lawnmowers, smart tractors, smart air conditioners, smart elevators, smart building, smart warehouses, robotic retail storage and entire store – the entire retail store is like a robot. And they will all have autonomous capability that are being driven by AI. And so what’s new for the industry therefore. Is that all if the enterprises in the world use to have computers for IT to facilitate our employees and their supply chain. But in the future, all of these industries whether in medical imaging or lawnmowers, you're going to have data centers that are hosting your products just like the CSP.
And so that's a brand new industry and we have a platform that we call EGX, which is the 5G Edge AI systems. And we have the autonomous systems we call AGX, which is what goes into adjusting and drives. And between those two systems and the software stack that we have on top of it we're in a great position to help these industries one at a time transformed their business model from a object oriented business model, a thing based business model to a connected device.
Operator
Your next question comes from the line of Stacy Rasgon with Bernstein Research. Your line is open.
Stacy Rasgon
Hi guys. Thanks for taking my question. First, I don't want to be pedantic, I suppose, but I guess on the Q1 guide, you say the gaming is the majority of the growth. Was that an absolute statement or was that a percentage statement? Can you give us some idea of how you'd sort of rank the sequential percentage growth of say gaming versus data center versus other, especially since it sounds like you've got 50 OEMs in crypto specific stuff that will go into the other?
And then I guess just briefly, could you give us some indication of where your supply situation and lead times are on your ampere parts within data center? I think you'd said last quarter, they were many months on six months plus, are they still looking like that? And is that sort of a limiting factor at this point in terms of what you can actually ship on the compute side and data center?
Jensen Huang
Colette, you will take one and I'll take the one.
Colette Kress
Sure. Let me start off Stacy in terms of our guidance for Q1, as you know, we're still in the early innings of our Ampere Architecture. Our Ampere Architecture, as it relates to gaming, as well as what it relates to data center. As we articulated in our call we have been really seeing continued uplift of folks adoption of A100 and it's going quite smoothly than what we had seen in prior overall versions. So when we think about our guidance for Q1, there's many different types of conclusions that will happen at the end of the quarter in terms of what we should put all of our platforms can grow. But the majority of the majority of the growth from Q4 to Q1 will likely be gaming.
Jensen Huang
Thanks, Colette. You asked the question about normal time, our company – at the company level where supply constraint, our demand is greater than our supply. And however for data center, so long as the customers more closely with us, and we do a good job planning between my companies, there should be a supply issue for data center. We just had to do a good job planning and we have direct relationships with each one of the royalty of fees and we have direct relationships with all the young ones. And we could do excellent planning between us.
We shouldn't have a supplier. We shouldn't be supply constrained there. But at the company level, we're supply constrained. Demand is greater than supply. And yes, we have enough supply. Yes, we usually have enough supply to achieve that even the outlook. And we had that situation in Q4. We expect that situation in Q1. And we have enough supply to grow through the year. But supply against constraint and demand is really, really great.
And so we just have to do a really good job planning. And meanwhile, one of the things that, that really came through for us is we have the world's best operations team. Our company is really, really have had an amazing operations team. We build the most complex products in the world, the most complex chips, the most complex packages, the most complex systems. And during Q4, they improved our cycle time and during Q1 we expecting them to improve our cycle time again. And we really are blessed to have such an amazing operations team.
And so during these times it really comes in handy. But overall at the company level where we expected them to be greater than supply, we have enough supply to do better in the outlook. And we have enough supply to grow each quarter throughout the year.
Operator
Your next question comes from the line of Timothy Arcuri with UBS. Your line is open.
Timothy Arcuri
Hi, thanks. I had a question on crypto, I guess Jensen, I know that the CMP stuff and the software drivers stuff that you're doing for the 36. And that's going to help a lot, but I think there is like four or five the big currency was in a move or at least a moving or there on a faster move from a proof of work to proof of stake, which is going to be a lot less compute intensive. So I guess the question that I get a lot is how do you assess the degree to which that drives GPUs back into the secondary market? Is there any way to get treated, get kind of a handle on that? Thanks.
Jensen Huang
Yes. If you look at the recent cash rates first of all, the transitions is going to takes some time. It can't happen overnight. And people have to build trust in all of the new versions. And so they'll take a little bit of time, but I hope it does. I hope that people use our proof of stake over time. And a little bit of these questions don't have to be answered. However I don’t have the optimism either that it will be all proof of stake, I think that proof of work is a very legitimate way of securing the currency.
And in the beginning while any currency is building its reputation, then take something like a group of work to do so. And so I think proof of work is going to be around for a bit. We developed CMP for this very reason, just so that there are different versions. We have different versions of our products for gaming, for professional visualization, for high performance computing, for deep learning. It stands to reasoning that you have to do a different version for CMP and we can solve it directly the way that we go-to-market would be to go directly to the industrial miners to – and it’s a great benefit to them. So they don’t have to chasing around spot markets. It’s a great benefit to the gamers. And because they want a game, and the game demand is just probably is off a chart. And so this is visibility really beneficial to everybody. The recent hash rate growth was really a result of several dynamics. The first dynamic is the install base. Most people thought that the – once the mining – the GPUs come back into the aftermarket, a small part does that, some people do that, but the vast majority don't keep them. And the reason for that is because obviously they believe in Ethereum and so – they're industrial miners that's what they do. And so, they're keeping around for when the profitability returns and they could kick start the mining gear. We saw – that's what we saw in the latter part of last year. We saw the hash rates starting to grow.
Most of that was result of the installed miners reactivating their equipment. It wasn't until earlier this year that we started to see a demand in our own GPUs. And when that starts to happen, there are some different dynamics. There's – the primary source these days come from powerful ASICs. And then there's some that that comes from our GPU and other GPUs in the marketplace. And so, I think that this is going to be a part of our business. It won't grow extremely large, no matter what happens. And the reason for that is because when it starts to grow large more ASICs comes into market, which kind of needs it. And when the market becomes smaller, it's harder for ASICs to sustain the R&D.
And so the spot miners, industrial miners come back. And then we'll create CMPs. And so, we expect that to be kind of a – to be a small part of our business as we go forward. Now, one of the important things is to realize that in the near-term, the – because we're in the beginning parts of our Ampere ramp, only two quarters into a multi-year cycle. This is also the first time that we've completely changed computer graphics. RTX using ray tracing is completely different than virtualization. And so this is a fundamental change in the way we do computer graphics and the results have been spectacular. There are some 200 million install base in desktops and 50 million in laptop. And the vast majority of them, we've only upgraded approximately I think it's something like 50% of the install base has been upgraded to RTX. And so there is a giant install base and the install base is growing that we need to upgrade to the next generation of computer graphics.
Operator
Your next question comes from the line of John Pitzer with Credit Suisse. Your line is open.
John Pitzer
Yes, guys. Thanks for letting me to ask questions. I want to go back to data center. You've been very kind over the last couple of quarters to call out Mellanox both when it was a positive driver and when it was a headwind. I'm kind of curious as you do – when you look into the fiscal first quarter, is there anything of distinction to mention around Mellanox versus core data center? And I guess as a follow on, the key metric that a lot of investors were looking at is when does the core data center business year-over-year growth starts to reaccelerate? And some of that is just simple math where you're just comping very hard compares from last year. But Jensen how would you think about data center year-over-year growth in the context of a reopening trade or any sort of new applications out there? I mean, what happened – what helped last time around was the move to natural language AI. Is there another big sort of AI application we should be thinking about as we think about data center growth reaccelerating?
Jensen Huang
We're expecting – Mellanox was down this last quarter. And our compute business grew double digit and offset – more than offsets the decline in Mellanox. We expect Q1 to be a growth quarter to Mellanox and we expect this coming year to be quite an exciting year of growth for Mellanox. The business is growing and ethernet is growing for CSPs is growing in InfiniBand for high performance computing and the Switch – the Switches have grown. Switch business grew 50% year-over-year. And so, we're seeing really terrific growth there. One of the new initiatives and we're going to see success towards the second half because the number of adoptions, the number of engagements has grown as our BlueField DPUs. It's used for virtualization for hyperscalers.
It's also used for security. As you know quite well the future of computing and cloud and it's multi-tenant cloud and there's no VPN front door to the cloud. You've got millions of people who are using every aspect of computing. So you need to have distributed firewalls and you can't have it just in one place. The intense focus of security across all of the data centers around the world is really creating a great condition for BlueField, which is really perfect then. And so, I expect our Mellanox networking business to grow very nicely this year. And we expect Q1 to be a great growth quarter for compute as well as Mellanox.
The Killer, great driving application for AI are several – last year you're absolutely right that it with natural language understanding and the transformer model and what is the – what was the core of – and other versions like that really, really made it possible for us to enable all kinds of new applications. So you're going to see a natural language understanding do text completion, and it's going to be integrated – I think it was just announced today that it was going to be integrated into Microsoft Word, and we've been working with them on that for some time.
And so there are some really exciting applications out there, but the new ones that came – that emerged recently are deep learning based conversational AI, where the ASR, the speech recognition as well as the speech synthesis are now based on deep learning, it wasn't before. And they were based on models that ran on CPUs, but now with these deep learning models, the accuracy is much, much higher and it has the ability to also mimic your voice and be a lot more natural. And so, the ability – these models are much more complex and much larger.
The other big huge driver is the recommenders. This is something really worthwhile to take a look at is called deep learning recommender models, and recommenders have historically – whether it's for shopping and or personalizing websites are personalizing your store, recommending your basket, recommending your music. Historically, it's been – and use the traditional machine learning algorithms, but because of the accuracy and – just the extraordinary economic impact that comes from an incremental 1% in accuracy for most of the – mostly the world's large internet businesses people are moving very rapidly to deep learning based models. And these models are gigantic. They're utterly gigantic.
And this is an area that is really driving high-performance computers. And – so we – I expect us to see a lot of momentum there. And the last one is the one that I just spoken out, which has to do with industrial 5G and edge, IoT type of applications for all of the different industries whether it's retail or logistics or transportation, agriculture or warehouses to factories. And so, we're going to see AI and robotics in a very large number of applications in industries and we're just seeing so much excitement there.
Operator
Your next question comes from the line of Aaron Rakers with Wells Fargo. Your line is open.
Aaron Rakers
Yes. Thanks for taking the questions. I wanted to go back again on the data center business that you just mentioned, Jensen, that the BlueField-2 product poised to kind of ramp and materialize in the back half of the calendar year. How do you see that – is it an attach rate? I think there has been discussions in the past about – all servers could potentially over time incorporate this layer of acceleration. How quickly should we think about that ramp? And then the second question is, can you just at a high level talk about how CPU – how a CPU strategy you're thinking about that in the context of the broader data center market?
Jensen Huang
Sure. If I could just work backwards, I believe that every single data center node will be outfitted with a GPU someday. And that someday is probably call it five years from now. And the fundamental driver of it is going to be security. Every single application in the data center and every single node in the data center has to be individually secured, zero trust computing, zero or confidential computing or zero trust computing. These initiatives are going to cause every data center to have every single application and every single node would be secured, which means every one of those computers have to have a control plane that is isolated from the application plane. And all the applications cannot share the same resources because that application could be malware, that application could be an intruder.
No application could have access to the control plan. And yet today the software defined data centers, the software defined networking, software defined storage, all of the security agents are running in the same processors as the applications and that hasn't changed. You're seeing the cloud – the CSPs in the world moving in this direction. Every single data center will have to move in that direction. So every node will be a DPU process for the software, for the infrastructure. You're essentially going to see the data center infrastructure would be offloaded from the application point and it will be something like a Bluefield. So I think this is our next multi-billion dollar opportunity, CPS. We support every CPU in the world. And we're the only accelerated computing platform that accelerates every CPU.
Ironically, the only CPU that we don't accelerate for AI is ARM, but we want to change them. ARM has such an exciting future because the nature of their business model and the nature of their architecture is perfect for the future of hyperscalers and data centers. You wanted – you want the most energy efficiency and in every single data center, because every data center is power constrained. We are going to be power constrained in every aspect of computing going forward. And so we would love to build around the own processor and invest in building a great ecosystem around it. And so that all the world's peripherals and all the world's applications can work and – work on any one of the CPUs that we know today.
And I want to start – we're going to start with high-performance computing and start with – all the areas that we have a lot of expertise in to build out our platform. And so, you're starting to see one industry leader after another embraced on and I think that's terrific, but now we've got to energize it with all of the ecosystem support. It can't just be on vertical applications, but we want to create a broad general ARM ecosystem.
Operator
Your next question comes from the line of Mark Lipacis with Jefferies. Your line is open.
Mark Lipacis
Hi, thanks for taking my question. A question for Jensen, I think. Jensen, if you look at the past computing eras, typically it's one ecosystem that that captures 80% of the value of that computing era and mainframes as IBM and many computers with stack PCs, Wintel, cell phones, Nokia and then Apple. So, if you don't get the ecosystem right then you're splitting 20% of the market with a handful of players.
So in this next era of computing parallel processing or AI, I think you've articulated the most compelling architectural vision of the data center of the future with data center scale computing devices with CPUs, GPUs, DPUs integrating to the same box serving all workloads in machine virtualized environment. Can you help us understand where is the market in embracing that vision and where is NVIDIA in building out that the ecosystem for that data center scale competing vision. And then maybe as part of that to what extent is CUDA of the kernel for that ecosystem? Thank you.
Jensen Huang
Yes, we're – I think we've done a great job on building out the platforms for several ecosystems around the world. And the domain that we do incredibly well out on – the domains that I have to do with accelerated computing, we pioneered this approach. And we brought it to high-performance computing at first and we accelerated scientific computing and we democratized supercomputing for all researchers, anybody who wants to have a supercomputer now can. And computing, it will simply not be the obstacle that somebody's discovery. We did the same for artificial intelligence. We did the same for visualization. We brought – we expanded the nature of gaining tremendously. Our GeForce today is the largest gaming platform. It’s the largest single largest body of computers that are used for gaming.
And in each case, we expanded the market tremendously. We would like to do the same for data center scale computing, as it applies to virtualizing these applications, these applications are also in the process. They've historically required dedicated systems, but they're moving into a virtualized data center environment.
And we are best at doing that. They run on our platform today. We have the ability to virtualize it and put it into the data center and make it remotely available. And so these applications, these domains are some of the most important domains in the world. And so we're in the process of getting them. By doing so and making our architecture available to CSPs and OEMs, we could create this accelerated computing platform available to everybody.
And so that's we're seeing our journey doing them. First, creating an architecting this platform, and then putting it literally into every single data center in the world. But we would also like to the next step of our journey is there's the Phase 3 of AI and has to do about – it has to do with turning every end point into a data center, whether it's a 5G tower, a warehouse or retail store, a self-driving car, a self-driving truck, these are going to be – they're all going to be essentially autonomous data centers and/or they're going to run AI, but they're going to run a lot more. They're going to do security in real time. Its networking is going to be incredible. It's going to run software to 5G and GPO accelerating 5G, we call areal.
And so these platforms are going to become data centers. There'll be secure. The software is protected and we can't tamper with. It if you tamper with it, of course won't run. And so the capability of these clouds will move all the way out to the edge. And we're in the best position to be able to do that. So I did the – in this new world of post close to Moore's law post in arts gaming in this new world where AI and software that writes software in this new world, where data centers are going to be literally everywhere and they're unprotected. There's no giant building with a whole bunch of people that secured. And in this new world where a software is going to enable this autonomous feature, I think we are a perfectly positioned for it.
Operator
This is all the time we have for Q&A today. I will now turn the call back to CEO, Jensen Huang.
Jensen Huang
Thanks for joining us today. Q4 have the truly breakout year for Nvidia. The two biggest engines of our business gaming and data center posted powerful growth. Gaming has become the world's largest media and entertainment industry, and will grow to be much larger. And again gamers will create, will play, they'll learn, they'll connect. The medium of gaming can host any type of game and eventually evolve with the countless metaverses.
So I'm for placed on for work. Gaming at simultaneously a great technology and a great business driver for our company. This year, we also closed our Mellanox acquisition and successfully united the amazing talent of our companies. Combined we possess deep expertise in all aspects of computing and networking to drive the architecture of modern data centers. Crowd computing and hyperscalers have transformed the data center into the new unit of computing. Chips on servers are just elements of the data center scale computers now.
With our expertise in AI computing full stack accelerated computing, our deep network to computing expertise and cloud to edge platforms, Nvidia is helping to drive a great computer industry transformation. And our planned acquisition of Arm, the world's most popular and energy efficient CPU company will help position and video to lead in the age of AI. This year was extraordinary.
The pandemic will pass, but the world has been changed forever. Technology adoption is accelerating across every industry. Companies and products need to be more remote and autonomous. This will drive data centers, AI and robotics. This underlies the accelerated adoption of Nvidia’s technology. The urgency to digitize, automate and accelerate innovation has never been higher. We are ready. We look forward to updating you on our progress next quarter. Thanks a lot.
Operator
This concludes today's conference call. You may now disconnect.