Nvidia’s Strategic Double Moat: How It Stays Ahead in AI

16 7 月 2025

Phillip Nova高级投资分析师林知霖先生

 

Nvidia (+4.04%) jumped after announcing it will resume sales of its H20 AI chip to China, after CEO Jensen Huang met with Donald Trump last week.

  • Previously in April, the administration restricted Nvidia from selling H20 chips to China by tightening chip export licensing requirements to the country.
  • In May, CEO Jensen Huang said the H20 curb had a negative revenue impact of $10.5 billion in total across Nvidia’s April and July quarters.
  • With the latest development, 2 different GPUs—the H20 and the upcoming RTX PRO, this can potentially add ~$10 billion in revenue to Nvidia’s current FY.
  • CEO Huang has argued that limiting access to his company’s products forces companies to provide financial resources to China’s Huawei, which the rival will then use for R&D to compete with Nvidia.
  • At the same time Meta also said it will build several large data centers over the next few years which likely includes purchases of Nvidia chips. “We’re calling the first one Prometheus and it’s coming online in ’26,” Meta CEO Mark Zuckerberg said on Threads on Monday. “We’re building multiple more titan clusters as well,” Zuckerberg added.

As of Tuesday’s close, Nvidia is up 27% YTD after being down as much as 30% earlier in April. In this Market Trends article we will deep dive into Nvidia’s business model to learn more about what are some of the key features or quaities that give the AI giant a strategic advantage in the market.

 

Overview of Nvidia and the Market Landscape

The tariff landscape seems more favourable now, given the US-China trade truce and Trump’s scrapping of “AI diffusion” rules that would have limited purchases of US chips by other countries. As Huang put it — AI has “become more useful because its smarter, it can reason, it is more used” and amount of computation necessary to train those models and to inference those models has grown tremendously.


Key points to note:

  • Hopper vs Blackwell Shipments to the top 4 US Cloud Service Providers (CSPS): Amazon, Microsoft Azure, Alphabet, and Meta (excludes AI companies, startups, enterprises):
    • Peak year of Hopper in 2024 vs Blackwell: $1.3M vs $3.6M.
  • May see potential cost increases for non-semiconductor systems and rack components outsourced, like cables, connectors, power components etc.
  • Nevertheless, Nvidia’s market dominance and strong customer relationships may allow it to partially pass through higher costs.
  • Comments from White House AI adviser David Sacks appear to align with Jensen’s, as he has pushed for ensuring the world builds its AI tools and applications on an American “Tech stack” — a full complement of hardware and software based on US technology.
Image from Nvidia’s GTC 2025 Keynote

About New Blackwell Chips:

  • Blackwell is essentially 2 GPUs in 1 Blackwell superchip.
  • Blackwell is ~40x Hopper’s inference performance.
  • Hopper costs ~$30,000 vs Blackwell ~$40,000 (30-35% increase).
  • NVLink, NVIDIA’s high-speed, low-latency interconnect technology designed to facilitate rapid data transfer between GPUs and other processors in high-performance computing systems, enables quick communication between GPUs– turning them into one massive GPU.
  • Grace blackwell in full production. Available in coreweave, already in use by many CSPs.
  • Grace Blackwell NVL72 facilitates scale up (Make many GPU dies behave like one
    huge GPU) and scale out (Tie many of those giant racks together so the datacentre

 

Analyst Comments

  • Management believes Hopper and Blackwell demand will continue to outstrip supply well in FY2026.
  • Gradual reduction in H200 will allow more back-end capacity to be effectively used for Blackwell
  • “Reasoning AI” may spur greater demand for computing power, and consequently, Nvidia’s chips.
  • Trump has made American leadership in AI a priority and welcomed Silicon Valley figures into his inner circle. This suggests to us that the US is likely to continue supporting AI companies like Nvidia.
  • Recent Middle East deals and GTC 2025 signals an expansion for Nvidia beyond Big Tech & Silicon Valley for customers.
Photo: Hwa Cheng/Agency France Presse/Getty Images
Image from Nvidia’s GTC 2025 Keynote

Investment Thesis:

  • Nvidia made its name by selling graphics processors (GPUs) for gaming, rendering realistic images via parallel computing– where multiple processors operate simultaneously.
  • GPUs were found to be effective in deep learning, with CEO Jensen Huang “donating” DGX-1 to OpenAI in 2016.

Chart from Bloomberg.

 

Image from Nvidia’s GTC 2025 Keynote
Image from Nvidia’s GTC 2025 Keynote

NVIDIA’s Moat: The Software Stack

  • We think of NVIDIA like Apple in smartphones—not only does it sell the best hardware (GPUs), but it also owns the best software ecosystem (CUDA).
  • Hardware Moat (GPUs = iPhones) – just like Apple’s iPhones are premium and high-performance, NVIDIA’s AI chips (H100, A100) are the best in the industry.
  • Software Moat (CUDA = iOS) – AI developers have spent years building everything around CUDA, just like how app developers optimize for iPhones first. Bottom Line: Even if competitors make similar AI chips, most companies can’t easily switch because they are locked into NVIDIA’s ecosystem.


It is Hard to Replace NVIDIA:

  • Even though DeepSeek has its own AI models, it still relies on NVIDIA chips to train them (for now).
  • NVDA software stack refers to the collection of libraries and frameworks that make possible accelerated Computing, including CUDA, cuLitho, etc.
  • Libraries expand to multiple applications (science, physics, etc), accelerate those applications and also opens up new markets.
  • CUDA-X libraries sit on top of the core CUDA platform, giving developers ready-made “cookbooks”—so they can drop GPU acceleration into their code without writing low-level kernels.

 

CUDA: Nvidia’s “iOS for AI”

  • Rather than just a pure GPU seller, it would better to think of Nvidia’s main product being an entire ecosystem.
  • Nvidia has built an extensive library of code for using its GPU chips for AI purposes– called Compute Unified Device Architecture (CUDA).
  • Introduced in 06, CUDA unlocked the power of GPUs for general-purpose computing. Changing GPUs from graphics-only to general-purpose.
  • Most AI frameworks (TensorFlow, PyTorch) use CUDA libraries (cuDNN) to accelerate deep learning.
  • Nvidia then built a full software stack on top of CUDA– cuDNN for deep learning, TensorRT for inference optimization.
  • This results in an entire software ecosystem that enabled plug-and-play compute acceleration through CUDA-enabled GPUs.
Image from Nvidia’s GTC 2025 Keynote

CUDA as a Moat:

  • Deep Industry Lock-in – Companies and researchers have built their AI infrastructure around NVIDIA’s ecosystem, making switching extremely costly.
  • Cuda: Lets AI developers write programs that run on NVIDIA GPUs. Made NVIDIA GPUs usable for AI instead of just graphics.
  • If a AI developer is not on a Nvidia GPU, it falls back to CPU.
  • Most AI tools today was built with the assumption that Nvidia GPUs will be used. Similar to mobile apps and the Apple App Store.
  • Alibaba’s Qwen, Tencent’s Hunyuan, ByteDance’s Doubao are all trained via AI frameworks like PyTorch
  • Furthermore, CUDA is a proprietary platform, meaning AI developers who build on CUDA are tied to NVIDIA GPUs – high switching cost for developers.
  • Together with hardware, this software stack is typically bundled together to be sold as an entire computer system (e.g., DGX-1)
  • When tools like PyTorch are used to train AI models, it automatically tries to use Nvidia’s GPUs via Cuda if one is available, because it is built to work best with CUDA.
  • Because CUDA has >16 years of tooling, most AI code and research assume an NVIDIA GPU. Alternatives exist (ROCm, oneAPI, Core ML) but have smaller library support and developer mind-share.
  • Without a CUDA-like software ecosystem, developers switching away from NVIDIA is like switching from an iPhone to a brand-new OS High switching costs is a key moat for Nvidia.
 

Graphic from Nvidia

Nvidia Full Systems

  • Rather than individual GPUs, Nvidia’s products are typically sold in a full end-to-end system.
  • For example, a car engine (individual GPU) would be pointless to have by itself. Full systems are like selling an entire car to customers.
  • Data center products are typically sold in full integrated systems like DGX and HGX, rather than as individual GPUs.
  • These full systems have pre-integrated hardware and software, making them ready-to-use and deploy for AI applications right off the bat.

 

About the AI Landscape

  • As Jensen Huang mentioned, AI started with models that can understand patterns and recognize speech – “Perception AI”.
  • In the past 5 years, “Generative AI” took the torch, with AI going beyond recognition and being able to generate (e.g., text-to-text via ChatGPT).
  • Currently, the industry has shifted towards reasoning AI – breaking problems down step by step, simulate multiple options and weigh benefits. e.g. Chain of Thought, DeepSeek R1.
Image from Nvidia’s GTC 2025 Keynote

 The Next Phase:             

  • Once AI has successfully “mastered” perception and reasoning— it will then move to agentic AI. (e.g., give ai a goal, it reasons, weigh benefits, finds best solution, then executes)
  • Beyond this, we have Physical AI-– where AI aims to understand the world, understand physics – inertia, friction, cause and effect, etc
    • Nvidia Omniverse generates a virtual world to train self-driving cars in different scenarios

 

Tokens as the new unit of work

  • Training/inference measured in “tokens generated/sec”; reasoning models need 10-100× more tokens than prior.
  • Whenever you put a prompt into chatgpt, the output is actually those tokens being reformulated into words for the AI model to process.
  • A smarter AI generates more reasoning tokens – eg., coming up with a whole bunch of ideas so they can select the best of those ideas tokens.
  • But the time taken to generate and reason with tokens needs to be short – otherwise users will be waiting too long to answer a question.

 

3 Key Processes of AI:  

  1. Training– supervised pre-training that gives model its foundational [Builds the brain].
  2. Post-training: Continuous learning– learning on the fly without humans in the loop, reinforcement [Keep learning].
  3. Test time scaling (Inference-time reasoning): At run-time the model “thinks harder” by spending more compute to generate more reasoning tokens before answering [Reason].
  • According to Jensen Huang, the amount of tokens generated as a result of reasoning is substantially higher and could be “100x more”.

 

About the DeepSeek Developments

  • As demonstrated by Jensen Huang in the screencaps taken below, reasoning requires far more compute than traditional.
  • In the images below, a prompt asks to seat people around a wedding table while adhering to constraints like traditions, photogenic angles, and feuding family
  • Traditional LLm answers quickly with under 500 tokens. It makes mistakes in seating the guests, while the reasoning model thinks with over 8000 tokens to come up with the correct answer.
  • R1 reasons and tries all the different scenarios, tests its own answer and even asks itself whether it did it
  • Meanwhile, the last-gen language model does a “one shot” answer. So the one shot is 439 tokens. It was fast, it was effective, but it was wrong. So it was 439 wasted tokens.
  • Reasoning models like DeepSeek still require compute, which means Nvidia’s chips are still indispensable in AI.
  • Nvidia’s market value fell by almost $600 billion in a single day on DeepSeek-fuelled concerns over the pace of AI spending on computing
  • DeepSeek’s cost-efficient open-source approach calls into question the need for such enormous investments in data centers in the first place.
  • Nevertheless, cheaper AI implied by DeepSeek’s model could increase global appetite for AI services and the hardware behind.
  • Largest CAPEX spenders are still doubling down on AI – with hyperscalers and other tier 2 CSPs estimated to spend nearly $371b this year.
  • DeepSeek also reportedly still used older Nvidia chips. Inference still requires Nvidia’s software stack for scalability.
  • One of the few things DeepSeek’s leader has said publicly is that he would use more of these chips, if he could.
Image from Nvidia’s GTC 2025 Keynote


Trade Nvidia and other US stocks on Phillip Nova 2.0 now! Click here to open an account now!

 

在 Phillip Nova 交易股票、ETF、外汇和期货

Phillip Nova 的交易特点

  • 访问 20 多个全球交易所
    从 20 多个全球交易所的 200 多个全球期货中捕捉机会
  • 全球股票的交易机会
    新加坡、中国、香港、马来西亚和美国市场的 11,000 多只股票和 ETF。
  • 超过 90 项技术指标
    使用 Phillip Nova 平台上的 90 多种技术指标,查看实时图表并轻松进行交易
  • 在 Phillip Nova 上交易多种资产
    您可以在 Phillip Nova 的单一账本上交易股票、ETF、外汇和期货
交易所交易基金 (ETF) 是一种有价证券,可用于跟踪几乎所有内容,包括特定指数、行业、商品或越来越多的主题。它们最常用于跟踪一篮子股票,通常可以通过与常规股票相同的渠道访问。 ETF 通常分为被动管理的 ETF,它们仅反映它们所跟踪的证券(例如 STI),以及试图提供更高回报或特定投资目标的主动管理的 ETF,通常考虑到预先指定的主题(例如 ARK Invest 的创新 ETF)。

我为什么要交易 ETF 差价合约?

  • 多年来,ETF 越来越受欢迎。 2020 年是 ETF 最好的一年,全球股票 ETF 在 12 个月内的流入量超过 $1T。使用差价合约获得 ETF 的敞口可以提高资本效率,因为只需合约价值的一部分作为保证金即可建立头寸。
  • ETF 尤其受到寻求相对轻松的投资体验,同时希望接触一系列特定且相对易懂的证券的投资者的欢迎。交易 ETF 差价合约通过消除交易者持有多种货币以访问全球 ETF 的需要,带来了更大的便利。
  • 希望参与大流行后经济复苏的投资者可以在著名的 SPDR S&P 500 ETF (SPY) 中建仓,该指数跟踪标准普尔 500 指数的表现。另一位可能相信环境未来重要性的投资者,社会和治理问题 (ESG) 可能会发现,越来越多的 ESG 主题 ETF 选择跟踪一篮子 ESG 评级高的公司是一项不错的投资,而不是手动挑选单个股票。 ETF 差价合约可以作为一种强大的工具,交易者可以通过持有多头或空头头寸从市场的两个方向获利。

看看我们提供的两种 ETF 差价合约:

1) ARKK 被击沉了吗?

方舟创新ETF(ARKK) ARKK 是 ARK Invest 积极管理的 ETF,根据其创新和行业颠覆潜力投资于一系列公司。 ARKK 最大的持股是特斯拉、Square 和 Zoom 等公司。 ARKK 从 12 日的峰值下跌约 -33% 2 月,由于市场经历了资金的避险外流,今年迄今处于亏损状态。然而,超级明星基金经理凯西伍德一直在加倍押注,购买更多正在经历动荡时期的成长型股票,如 DraftKings、Peloton、Teladoc 和特斯拉。在她看来,ARKK 正在玩长期游戏,并且仍然坚信这些成长型股票在当前波动之后的长期前景。同样在流出方面,投资者仍然对 ARKK 押注很大,因为 ARK Invest 今年在其所有六只基金中仅损失了约 $1.2B 的资产,而同期则流入 $15.1B。最近,投资者一直紧张地关注 ARKK 的一篮子科技股,因为它们未来的盈利潜力仍然容易受到高通胀的侵蚀——这是最近几周市场的主要担忧。随着大宗商品——近期通胀担忧加剧的主要因素——从历史高位急剧下跌,投资者对恶性通货膨胀的担忧是否被夸大了?

2) 寻找亚洲股票的敞口?

iShares MSCI Asia ex Japan ETF (AAXJ) AAXJ 目前的交易价格为 -10.6%,偏离 2 月份创下的历史高点,在当时亚洲范围内的股票抛售中放弃了收益。鉴于该 ETF 持有的略高于 40% 的资产位于中国,中国股市的持续动荡目前在 AAXJ 中几乎完美延续,因为中国投资者在过去一年取得了惊人的收益后喘了口气。展望未来,亚洲——尤其是中国,正在加速其经济复苏。外界普遍预计,中国将成为今年表现最好的主要经济体之一,大大提振企业盈利前景。随着亚洲其他地区和世界逐渐开放自己的经济,在贸易前景增强的情况下,AAXJ 可能会再次受益于亚洲的强劲表现。

差价合约可用于在 Phillip MetaTrader 5 (MT5) 上进行交易。

交易差价合约的特点:

  • 在牛市和熊市中交易
    进入多头和/或空头头寸的能力使交易者能够利用上涨和下跌的市场。
  • 进入门槛更小
    灵活且较小的合约规模。这意味着交易者将能够以适度的资本签订合同。
  • 没有到期日期或交付风险
    与通常具有固定到期日的期货不同,差价合约允许交易者永久持有头寸。差价合约以现金结算,无需担心标的资产的交割。

 

使用飞利浦 MT5 的好处:

在提供低点差的动态平台上以零佣金进行交易。与 Autochartist 和 交易中心指标,并在移动、网络和桌面应用程序上可用,您将永远不会错过使用 Phillip MT5 的交易机会。

注册一个免费的 30 天 Phillip MetaTrader 5 模拟账户

更多市场趋势

China A50 Index ETFs and Futures — a powerful, simplified gateway into China’s growth story.

阅读更多 >

First JB-Singapore Rapid Transit System (RTS) Train Was Unveiled, But What Does It Mean For Traders?

阅读更多 >

First JB-Singapore Rapid Transit System (RTS) Train Was Unveiled, But What Does It Mean For Traders?

阅读更多 >