Data Center & Semiconductor Bottleneck Timeline: In-depth Analysis of CPU/GPU Bottlenecks (1)
- Source: https://x.com/tesla_teslaway/status/2022730281464271271?s=46
- Mirror: https://x.com/tesla_teslaway/status/2022730281464271271?s=46
- Published: 2026-02-14T17:50:54+00:00
- Saved: 2026-02-15
Content

Here is the English translation of the content, maintaining the original structure, tone, and emphasis.
Data Center/Semiconductor Bottleneck Timeline Summary
You must know this in the Great AI Era!!
You have to study this, everyone!!!!
Now, let’s analyze each bottleneck in depth.
First, let’s look back at the brief explanation I wrote before.
- CPU → GPU Transition (Computation Bottleneck)
Period: 2020~2022
Why did it become a bottleneck? Deep learning requires performing thousands of simple calculations (matrix multiplication) simultaneously.
CPUs are good at doing things "one by one in order" but weak at parallel processing.
GPUs have thousands of cores working simultaneously, making AI training 10~100 times faster.
Since AlexNet (2012), GPUs became essential, and with the advent of ChatGPT-class models, the transition became complete.
🔸 Easy Explanation
CPU: Starts in Seoul, then visits Daejeon, Gwangju, Daegu, Busan, and Gangwon-do one by one.
GPU: Departs from Seoul with multiple cars simultaneously heading to all destinations.
CPU: Like one genius tapping away at a calculator at a desk.
GPU: Like thousands of simple laborers doing calculations by hand.
Calculators are great for complex math, but for AI training, doing simple things simultaneously and quickly is best.
🔸 Result/Impact
Deep learning researchers used to train on CPUs, but after switching to GPUs (Pre-Hopper generation), tasks that took a day were reduced to a few hours.
NVIDIA captured the market by pushing this exclusively with their CUDA software.
🔸 Current Status (2026)
Already solved. Now GPUs are the standard; everyone trains on GPU clusters.
Single GPU Focus → Current: Bottleneck moved to entire system orchestration.
ㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡㅡ
First, let’s find out what CPU, GPU, and ASIC are and what their respective roles are.
AI workloads (especially in Data Centers) rely on heterogeneous computing as the standard.
The era of using purely CPUs or GPUs is over; the CPU + GPU (or ASIC) combination is the trend.
Role of CPU:
Orchestration & Data Management: Feeds data to the GPU (storage, sharding, indexing), scheduling, and power management. CPUs handle data preparation during AI pretraining/fine-tuning.
Sequential Processing & General Tasks: Handles non-AI server workloads (databases, networking).
Built-in AI Acceleration: AMD EPYC and Intel Xeon 6 have added AI features (e.g., AMX, built-in accelerators) → handling some inference tasks.
Strength: Stability, compatibility. Weakness: Parallel AI computation.
Role of GPU:
Core of Parallel Computing: Monopolies on AI training (massive matrix multiplication) and high-performance inference. NVIDIA Blackwell/Rubin, AMD MI300/MI400 are the main players.
Main AI Engine: GPUs handle 90%+ of the compute in Hyperscaler (Google, Meta, etc.) clusters.
Current Combination Methods:
Rack-scale Integration: NVIDIA Rubin (Late 2026): Vera CPU (36) + Rubin GPU (72) + NVLink connected like one "AI supercomputer." The CPU manages GPU data movement.
AMD Helios/MI400: EPYC Venice CPU + MI400 GPU, using unified memory for zero-copy data transfer → maximizing heterogeneous efficiency.
Intel: Xeon 6 host CPU + External GPU (mostly NVIDIA/AMD). Their own Gaudi is limited.
Custom ASIC Addition: Google TPU, Amazon Trainium replacing some GPUs in inference. Broadcom is leading the design of hyperscaler custom ASICs.
🤦♂️🤦♂️🤦♂️ How is it? You honestly have no idea what I'm talking about, right?
😁😁😁 Since we are investors, I don't think we need to know these gritty details.
Okay, let me explain it easily in my own way.
The most important parts here are CPU, GPU, and ASIC. Simply put:
CPU has become more important: Its role in data management and power regulation has grown, giving it a "resurgence" vibe.
GPU is still King: Nearly monopolizes model teaching (training).
Answering questions (Inference): Split between GPU + Custom Engines (ASIC) (to save costs).
Digging a little deeper:
The Role of the CPU in the Past (Team Manager Role)
The type that works smartly and sequentially: Good at calculating one by one (e.g., general programs, web servers, data management, power regulation, overall command).
Weak at AI parallel calculation (doing thousands of things at once) → Lost almost all AI parts when GPUs appeared.
Role in AI: Just giving data to the GPU and cleaning up. The GPU is the star; the CPU is the supporting actor.
Easy Explanation: The Manager just conducts; the Racer (GPU) does the actual driving.
The Role of the CPU Currently
Gained AI Muscles: Built-in small AI accelerators (AMX, etc.) → Can do simple AI calculations quickly.
Massive increase in Core Count (100+) + Smarter Memory/Interconnects.
Specialized in Data Movement/Management: Perfect "Traffic Control" when connecting multiple GPUs/ASICs.
Analogy: The Manager now acts as "Coach + Trainer." Trains the racers (GPU/ASIC), distributes fuel (data) efficiently, and plans the team strategy.
Why is the CPU suddenly rising again?
AI has changed from simply "Calculating like crazy with GPUs" to complex teamwork.
Past: Only teaching the AI model (training) was important → GPU solo run.
Present: Explosion of actual service usage (inference) + Mixing multiple chips (GPU + ASIC) → Data movement, power management, and chip coordination have become incredibly important.
The CPU is perfect for this. (Moving data fast, saving power, coordinating engines).
Example: In a 2026 rack-scale system, if there's no CPU, the GPUs/ASICs can't "talk to each other" and chaos ensues.
Also, as inference (answering questions) becomes much larger than training, CPU efficiency (low power/stability) shines.
Easy Explanation
Past Team: One Star Racer (GPU) could win alone → Manager (CPU) was less important.
Current Team: Multiple Racers (GPU + ASIC) + Complex Track → Without the Manager (CPU), the team falls apart. The Manager has "suddenly" become a hero.
Conclusion: The CPU was always essential, but as AI grew from "One Star" (GPU) to a "Big Team" (GPU+ASIC), the CPU's role expanded and shone brighter.
Therefore, the CPU could become a bottleneck again in the future.
My Thoughts
It is highly likely there won't be a major bottleneck.
CPU companies (AMD, Intel, even NVIDIA with Vera CPU) are developing frantically: Core counts exploding, memory connections getting smarter, built-in AI functions.
Even if inference grows, CPU demand is forecasted and being prepared for.
The overall bottlenecks are predicted to be Power, Memory (HBM), and Networking.
However, the probability is not 0%. If AI becomes more complex than expected (e.g., massive Agent AI), CPUs could briefly become a bottleneck. Companies that respond well, like AMD, will likely be the winners.
🙋♂️ Wait, you might have a question here: Then what is an ASIC? Is the GPU being replaced?
The biggest reason for using ASICs is Cost and Power Savings.
AI has two stages: "Teaching the model (Training)" and "Answering questions with the taught model (Inference)."
Inference stage repeats the same task every day (e.g., ChatGPT answering user questions). This is predictable and high volume, so a tool that does only that specific task perfectly (ASIC) is cheaper and uses less electricity.
Big companies (Hyperscalers like Google, Amazon, Meta) make chips tailored exactly to their AI work instead of buying GPUs → Reducing costs by 30~50% and saving on electricity bills.
Easy Explanation
ASIC is a "Electric Pot dedicated to boiling Ramen." It boils ramen incredibly fast and cheap (perfect for repetitive tasks like inference). So companies say, "We eat ramen (do inference) a lot, so let's just buy a dedicated pot!" instead of a GPU.
Why won't it replace GPUs completely? (Why GPUs still survive)
ASICs lack flexibility (can't do various tasks).
The Training stage involves changing models frequently and testing new ideas (e.g., testing thousands of times to make GPT-5). This is unpredictable, so you need a Kitchen Set (GPU) that can cook various dishes.
ASIC is specialized for "only one dish," so if you change the model slightly, you have to make a new chip → Costs huge amounts of time and money.
Software Ecosystem Strength: NVIDIA GPUs have CUDA, a "Recipe Book" that is so well-made that 90% of developers use it. ASICs are different for every company, so you have to rewrite code → Annoying and slow.
Development Cost & Time: Making one ASIC takes hundreds of millions of dollars + 2~3 years. You can just buy a GPU and use it immediately.
Easy Explanation: ASIC is a "Ramen Pot," so it's the best for ramen, but if a new menu (new model) comes out, you have to make a new pot. GPU is a "Gas Stove + Pot + Frying Pan Set," so you can cook anything immediately. That’s why GPU is still the best for high-experimentation, diverse AI training.
🙋♂️ NVIDIA is also developing new chips (H100→B100→B200, etc.), so doesn't it cost money just like ASICs?
But they overcome this with Economies of Scale + Software Ecosystem.
NVIDIA GPU: New chip design cost ≈ $1B~$2B, 1.5~2 year cycle. But with 90%+ market share, they sell billions of chips → Distributing development costs per chip.
Google TPU: Similar design cost, but used only internally (cloud service), so sales volume is limited → Slower cost recovery.
Conclusion
In the past, AI models evolved crazily: Transformer → BERT → GPT series → Multimodal. So, you had to develop ASIC chips frequently, or sometimes change them mid-development. So they couldn't be used much.
Present ~ Future: As model development speed stabilizes → TPU (ASIC) becomes much more profitable.
2026: The basic structure (Transformer) has become almost standard, so "major changes" are fewer. Models develop mainly by "increasing size (parameters)" or "fine-tuning."
Inference (Actual Service) volume explodes: Hundreds of millions of people use things like ChatGPT/Gemini daily → Massive repetitive work.
So big companies (Google, Amazon, Meta) invest in ASICs like TPU/Trainium: Once made well, they save huge amounts on cost/electricity in the long run (Billions of dollars in profit annually).
As model changes slow down, ASICs (TPUs) get stronger, and 2026 is exactly that turning point – this is why hyperscalers are pushing ASICs hard.
GPUs still hold the throne for "Experimentation/Learning," but ASICs are eating up the "Actual Money-Making Services (Inference)."
Final Conclusion
Ultimately, what we as investors need to think about is, as seen before:
Whether CPUs are keeping up well with GPUs (since a bottleneck might occur in CPUs).
Predicting how big the total pie of GPUs + ASICs will get, and within that, how much the share of ASICs will grow.(This will help you gauge who will race ahead: companies making GPUs vs. Big Tech pushing ASICs).
- GPU Related Companies
NVIDIA (NVDA): Dominates the AI and data center-centric GPU market, holding over 80% share in AI training GPUs. Strong in gaming and professional visualization, but facing increasing competition in the consumer sector.
AMD (AMD): Direct competitor in high-performance GPUs for data centers and gaming. Instinct series is expanding for AI tasks; recently data center GPU revenue surpassed CPU revenue. High growth potential amidst supply constraints.
Intel (INTC): Expanding beyond traditional integrated graphics to discrete GPUs, targeting data centers and AI with Arc and Max series. Aggressively pushing with recent hiring and investment, but market share is smaller than competitors.
ARM (ARM): Provides GPU IP designs like Mali and Immortalis, mainly for mobile and embedded devices. Supports AI acceleration but focuses on licensing rather than manufacturing; widely adopted in smartphones and IoT.
- CPU Related Companies
Intel (INTC): Traditional leader in the x86 architecture CPU market, targeting AI PCs and data centers with Core Ultra series (client) and Xeon (server). Maintains about 60% market share with strong server CPU demand in 2026, but under pressure from ARM competition.
AMD (AMD): x86 CPU specialist with Ryzen (consumer/client) and EPYC (server) lineups as core. Breaking 40% server market share by end of 2025 and growing rapidly. Launching 2nm-based Venice EPYC in 2026.
Apple (AAPL): Designs and produces custom ARM-based Apple Silicon (M series) CPUs, exclusive to its ecosystem (Mac/iPad/iPhone). M5 series strengthens AI performance; developing M6 for 2026.
Qualcomm (QCOM): Targeting Windows on ARM PC market with ARM-based Snapdragon X series (Oryon custom cores). Emphasizing multi-day battery and AI performance with X2 Plus, etc., in 2026.
SoftBank (9984.T): Leading CPU IP licensing business through ARM Holdings ownership. Does not manufacture directly but earns royalties on Arm architecture; revenue expected to increase in 2026 due to AI/Data Center growth.
- Custom Chip (ASIC) Related Companies
1) Masters of Chip Design (Big Tech/Hyperscalers)
Alphabet (GOOGL): Leader with TPU (Tensor Processing Unit). Reached 7th generation (Ironwood), equipped with massive 192GB HBM per chip, threatening NVIDIA in AI training/inference efficiency.
Amazon (AMZN): Developing Trainium for training and Inferentia for inference. Goal is to provide cheaper AI computing environments than NVIDIA to AWS customers by combining with their own Graviton CPU.
Meta (META): Designing MTIA (Meta Training and Inference Accelerator). Making chips optimized primarily for ad recommendation algorithms (Facebook/Instagram) and Large Language Model (Llama) inference.
Tesla (TSLA): Making FSD (Full Self-Driving) chips and Dojo chips for AI training supercomputers. Recently selected Samsung Electronics as a production partner for the Dojo D2 chip, accelerating chip internalization.
2) Design Partners
Broadcom (AVGO): Unrivaled #1 in the ASIC market. Key partner for Google TPU and also helps design chips for Meta and OpenAI. World's strongest in chip-to-chip connection technology (SerDes).
Marvell (MRVL): Broadcom's powerful rival, handling custom AI chip designs for Amazon and Microsoft. Specialized in networking and data center ASICs.
Alchip (3661.TW): Taiwan's #1 design house and TSMC's top partner. Leads Amazon's ASIC design and acts as the Taiwanese gateway for North American Big Tech volume.
GUC (3443.TW): Design house where TSMC is a major shareholder. Designs high-difficulty chips for large clients by applying TSMC's cutting-edge processes (3nm, 2nm) first.
3) Korean Design Solution Partners (DSP/VCA)
Korean companies act as a bridge for fabless companies wanting to use Samsung Electronics or TSMC processes.
Gaonchips (399720): Key partner (DSP) of Samsung Foundry. Recently proved technical capability by winning orders for Samsung's 2nm process-based AI chip design. The most specialized company for AI ASIC design within the Samsung ecosystem.
ASICLAND (445090): The only TSMC VCA (Value Chain Aggregator) in Korea. Handles designs for domestic and foreign fabless companies that need to use TSMC processes. Recent rapid earnings growth with large-scale mass production contracts in AI Edge chips and storage controllers.
ADTechnology (200710): One of Samsung's large DSPs, focusing on high-performance server-grade AI chip design utilizing past cooperation experience with ARM. Actively targeting overseas projects based on 2nm recently.
CoAsia (045970): Global partner of Samsung Foundry, with particular strength in Automotive ASIC design. Expanding orders for automotive AI chips in European and US markets.
My Thoughts
Companies experiencing bottlenecks should not be abandoned even if their stock prices have risen significantly.
These companies have a structure where bottlenecks occur cyclically, so they should be put on a long-term watchlist and continuously tracked/observed.
Looking at actual stock price trends, prices often do not crash after the bottleneck is resolved but tend to move sideways.
This is because, until the next bottleneck appears, although new growth momentum has vanished, existing demand is maintained to some extent, so revenue does not plummet.
From a corporate perspective:
Increasing supply blindly might increase profits again,
But demand is uncertain, and production capacity might have already reached its limit, so it's hard to expect massive growth.
Result: Revenue doesn't decrease easily, but it's also hard to expect high growth.
Therefore, an effective investment strategy is:
Buy low when stock prices adjust (drop or move sideways) due to bottleneck resolution.
Aim for the rise when the next bottleneck appears (Cyclical Trading).
Key Investment Strategy: Keep these companies on your watchlist and time your entry by continuously observing the bottleneck occurrence/resolution cycle.
Breaking this down into simple terms takes much more time and effort than expected...
Since I started it, I need to finish it, but allocating time is not easy. ㅠ
Anyway, I plan to summarize by sector, but the detailed analysis of individual companies I originally planned will likely have to move to the subscription service.
Link: http://x.com/i/article/2022729580700930048