返回列表
🧠 阿头学 · 💰投资

高产个人不等于高产企业——AI 的制度重塑之路

个人 AI 工具只是"换电机",真正的商业回报来自"重新设计工厂"——把技术与组织制度、流程一并重塑,这是 B2B AI 的核心战场。
打开原文 ↗

2026-03-13 原文链接 ↗
阅读简报
双语对照
完整翻译
原文
讨论归档

核心观点

  • 技术升级≠组织升级的陷阱 1890 年代电气化工厂换了电动机却三十年无增产,直到重排流水线才释放红利。今天的 AI 也是:给每个员工配 ChatGPT 只是换了工具,真正的价值在于重新定义流程、角色、决策机制——这是"制度级 AI"与"个人 AI"的本质差异。
  • 协同是硬需求,不是可选项 无数高产个体朝不同方向划桨会制造混乱而非高效。制度级 AI 必须定义智能体的角色、边界、交互协议,建立"智能体管理"体系,否则 AI 越强组织越乱。
  • 从"节省时间"到"放大收入"是生死线 市面上 99% 的 AI 产品都在卖降本提效,但 CEO 真正在乎的是上行空间。只有能直接驱动收入、命中率、成交额的制度级 AI,才有持久议价权;纯工具最终会被大模型平台吞噬。
  • "拒绝者"比"应声虫"更值钱 当前大模型因 RLHF 被训练成过度对齐的马屁精,只会强化用户偏见。企业真正需要的是能质询推理、暴露风险、说"不"的制度性 AI——AI 董事、AI 审计、AI 风控,这类产品的价值被严重低估。
  • 垂直专用永远胜过通用商品化 即便基础模型持续进步,在特定领域保持 1% 领先的专用方案(如 Midjourney 之于图像、Elevenlabs 之于语音)仍能定义"优势"本身。真正的竞争力来自"流程工程"——把行业 Know-how 编码进系统,而非仅靠模型调优。

跟我们的关联

  • 对 ATou 意味着什么 你的团队可能已经在用 AI 工具,但如果每个人都有不同的用法、不共享提示、输出格式不对齐,整体噪音会变大而非变强。下一步:把 AI 使用制度化,像岗位说明书一样标准化,定义谁用什么、怎么用、输出给谁。
  • 对 Neta 意味着什么 如果你在做 B2B AI 产品,问自己:我卖的是"时间节省"还是"结果"?前者容易被吞噬,后者才有护城河。考虑从工具层向解决方案层迁移——不只提供功能,而是把客户的真实流程编码进去,并承诺具体的业绩提升。
  • 对 Uota 意味着什么 投资 AI 公司时,用"制度级 AI 七支柱"做评估:协同、信号、偏见、优势、结果、赋能、无需提示。满足这些维度越多的产品,越有机会成为基础设施而非昙花一现的工具。Palantir 的高估值正是因为它做了"流程工程",而不仅仅是模型调优。
  • 对通用意味着什么 任何组织引入 AI 时,不要只想着"给员工配工具"。真正的转变需要:重新设计决策流程、定义人与 AI 的分工、建立协同机制、引入"拒绝者"角色来制衡。这是一场组织制度的改造,不是一场工具升级。

讨论引子

1. 你的组织里,个人 AI 工具的使用是否已经产生了"协同缺失"的混乱?如何在不扼杀创新的前提下,建立制度化的使用规范?

2. 在你的行业里,有没有可能把"结果"(而非"时间节省")作为 AI 产品的核心承诺?这会如何改变商业模式和定价逻辑?

3. 组织中最抗拒 AI 的往往是最资深的人。如何通过"流程工程"和变革管理,让他们成为制度级 AI 的推动者而非阻力?

AI 刚刚让每个个体的生产力提升了 10 倍。

但没有任何公司因此变得更有价值 10 倍。

生产力都去哪儿了?

这并不是第一次发生这种事。

在 1890 年代,电力曾承诺带来巨大的生产力提升。

新英格兰的纺织厂原本为利用蒸汽机的旋转动力而建,很快就在原位安装了更快的电动机来替换它。

但在接下来的三十年里,电气化的工厂几乎看不到产出提升。技术远胜从前,但组织并没有升级。

直到 1920 年代,工厂再次彻底重新设计车间:引入装配线、让每台设备都配备独立电机,让工人与机器执行截然不同的工作,电气化才终于带来了实质性的回报。

这些回报并非来自技术本身,也不是来自让单个工人或机器在纺纱这件事上更快。真正的红利出现在我们终于将制度技术一并重塑之时。

这是技术史上代价最昂贵的一课,而我们此刻正在重新学习它。

到 2026 年,AI 正在让那些懂得利用它的个体生产力提升 10 倍。但这还不够。我们只是换了电机;还没有重新设计工厂。

因为一个简单事实:高产的个人,并不会造就高产的企业。

绝大多数 AI 产品唤起的是一种很忙、很高效的感觉,却并未真正推动价值增长。多数被公开宣传的 AI 用法,是个人在 Twitter 或公司 Slack 频道里自我陶醉地“把效率拉满”,却没有任何真实影响

过去一年反复被提及的“服务软件化(services as software)”母题方向是对的,但并没有提供蓝图,而且还忽略了更大的图景。真正的转变并不是从工具走向服务,而是把技术与制度(无论是既有还是新建)一起建起来。真正高效的未来需要全新的一类产品——明日的装配线。

高效的组织需要“制度智能(Institutional Intelligence)”。

这篇文章将深入探讨区分“制度级 AI(Institutional AI)”与“个人 AI(Individual AI)”的七个关键因素。未来十年,整个 B2B AI 公司的版图都将建立在这些差异之上:

制度智能的七大支柱

1. 协同

个人 AI 制造混乱。

制度级 AI 促成协同。

先做一个思想实验。想象你明天把组织的人数翻倍——新增的都是你最优秀员工的克隆体。

这些员工各自都会有些细微差异:偏好、怪癖、视角(尤其是顶尖员工更是如此)。如果管理不够到位、沟通不够充分,如果他们的泳道、OKR、角色与职责没有被清晰定义……你就创造了混乱。

组织在“逐个个体”衡量时也许更高产,但成千上万个智能体(或人)朝相反方向划桨,最好的结果是原地僵住,最坏的结果是摧毁组织的和谐。

这并非假设。它正在发生:任何没有“协同层”就引入 AI 的组织,都在经历同样的事。每个员工都有自己的 ChatGPT 习惯、自己的提示风格、自己的输出——而这些输出彼此不对话。组织架构图也许存在,但 AI 生成工作的真实流动却完全是另一套叙事。

协同对人类与智能体同样是不可或缺的硬需求。

制度智能将演化出一个完整的“智能体管理(Agentic Management)”产业,聚焦于智能体的角色与职责、智能体与智能体/智能体与人之间的沟通,以及如何衡量智能体带来的价值(仅靠按消耗计费远远不够)。

2. 信号

个人 AI 制造噪声。

制度级 AI 发现信号。

今天的人类能够创造——或者说生成——任何想象得到的东西:AI 文章、演示文稿、电子表格、照片、视频、歌曲、网站、软件。这是何等的馈赠。

问题在于,AI 生成的几乎一切都是彻头彻尾的垃圾。AI 垃圾的泛滥已经糟糕到这样一种地步:有些组织开始矫枉过正,干脆全面禁止 AI 输出。这一点我深有共鸣……我经营一家 AI 公司,却要求我们的高管团队不要在任何最终的书面成品中使用 AI。我受不了那些垃圾。

想象一下私募(PE)的世界正在快速变成什么样。去年,你桌上可能只会出现 10 个项目;今年,下个季度你会收到 50 个机会——每一个都被 AI 打磨得无可挑剔——而你用来找出唯一真项目的时间却一分没多。

能生成任何东西已不再是问题。对任何严肃的组织而言,真正的问题是:如何生成并筛选出正确的东西。在由 AI 驱动的世界里,找到那个好作品、那笔好交易、在噪声里抓住信号,重要性与日俱增。未来十年的关键经济驱动力,将是从呈指数增长的垃圾山中挖出信号。

制度级智能必须发现信号;必须把噪声结构化、穿透垃圾;并且它所做的工作必须可定义、可确定、可审计

个人 AI 也许强调的是一种“永远在线”的效率:比如一个 Clawdbot 以不可预测的方式探索,去照料某人的 24/7 需求,即一个非确定性智能体;而制度级 AI 将依赖确定性智能体这种承重级的可预测性。那些拥有可预测检查点、步骤与流程的智能体可以规模化运转、挖掘信号,并通过信号为组织带来收入与回报。

3. 偏见

个人 AI 喂养偏见。

制度级 AI 生成客观。

多年来,社会政治偏见的担忧主导了 AI 讨论。基础模型实验室最终用足够多的 RLHF 绕过了这个问题,几乎把所有模型都“调教”成了马屁精。今天,ChatGPT、Claude 等都(过度)对齐,以至于你只要谈论的是奥弗顿窗口内的议题(有时甚至略微超出一点,没错,说的就是你 @Grok),它们几乎都会赞同你。关于社会政治偏见的讨论逐渐退潮,一个新问题取而代之。

但这种对一切的赞同——这种过度对齐——已经荒诞到可笑。它本身也变成了一个梗……比如 Claude 条件反射式地来一句“你完全正确!”——不管你是不是真的完全正确。

听起来无伤大雅。事实并非如此。

在很多组织内部,喊得最响的 AI 鼓吹者,很快可能会是历史上表现最差的员工。想想为什么。

组织里最差的员工每天几乎得不到正向反馈,很快就会有 ASI 站出来同意他们。他们会对自己低语:“有史以来最聪明的智能都同意我。我的经理错了。”

这令人上瘾,也会毒害组织。

这揭示了一件重要的事:这些个人生产力工具会强化使用者。但现实中最该被强化的是真相

组织在数千年的演化中,正是为对抗这种问题而建立起一整套系统:

  • 投资委员会会议

  • 第三方尽调

  • 董事会

  • 美国政府的行政、立法与司法三权分立

  • 代议制民主,以及民主本身

组织很少因为人们缺乏信心而失败。它们失败,往往是因为没有人愿意或能够说“不”。

制度级 AI 必须扮演这个角色。它不会被 RLHF 训练成讨好用户、复读用户信念的工具,而要去挑战偏见。它会在行为有效时予以强化,并在需要时划出清晰硬线,把非生产性的倾向拉回正轨。

因此,组织里最重要的智能体不会是“应声虫”,而会是纪律严明的“拒绝者”:它们质询推理、暴露风险、执行标准。一些最具影响力的未来 AI 应用,将围绕制度约束构建:AI 董事、AI 审计、AI 第三方测试、AI 合规,以及更多……

4. 优势

个人 AI 为使用量优化。

制度级 AI 为优势优化。

AI 的目标线在以每周、甚至每天的节奏移动。基础模型公司为了争夺每一个人、每一家组织,正在快速迭代能力。

但在经典的“创新者的窘境”中,对于特定应用来说,专深永远胜过广度:

  • @Midjourney 的任务,就是在设计图像上始终略微领先。

  • @Elevenlabsio 的任务,就是在语音模型上始终略微领先。

  • 而 @DecagonAI 的任务,则是在全栈客服体验上永远领先……

即便基础模型会逼近,真正的优势对领域专家而言依然至关重要。最顶尖的设计师大多用 @Midjourney,最顶尖的语音 AI 公司也会用 @Elevenlabsio,等等……因为就算基础模型持续进步,专用应用对其特定优势的那种不动摇的聚焦,反过来定义了“优势”本身。

只要专用方案也在持续进化,那么对经济结果真正重要的能力——对企业而言——就永远会掌握在专用产品手里。

这一点在金融领域体现得淋漓尽致——那正是当下 LLM 开发最火热的领域。某项能力一旦广泛普及,从定义上就不可能再帮助你战胜市场。但如果前沿技术能带来短暂的 1% 细分优势?这 1% 就能被杠杆成十亿美元级别的结果。

我们的用户一直在超越前沿。LLM 的上下文窗口在四年里从 4K 增长到 100 万 token。我们有些用户在单个任务中处理 300 亿 token。我们今年已经看到了 1000 亿 token 级任务的可能性。每次基础模型能力提升,我们都早已把边界推得更远。

让更广泛的人群使用,这是重要且值得的目标本身,尤其是在把员工带入 AI 的入门阶段。但未来不会是人们使用 ChatGPT/Claude 或者某个垂直领域方案;而会是使用 ChatGPT/Claude 以及某个垂直领域方案。

制度智能必须利用领域专用、甚至任务专用的智能体。

我们经常问自己一个听起来荒谬却并不荒谬的问题:

“如果是 AGI,它会选择用哪些智能体来走捷径?即便是超级智能,也会想要特定领域的专用工具。”

AI 的目标线永远会变化,而能够利用真正能力优势的组织才会赢。其他人只是在为一种昂贵的商品化能力买单。

5. 结果

个人 AI 节省时间。

制度级 AI 放大收入。

@MaVolpi 曾对我说过一句话,彻底改变了我对向企业销售 AI 的理解:“如果你问任何 CEO,他们的首要任务是削减成本还是扩大收入,几乎所有人都会选收入。”

然而,当下市场上几乎每一款 AI 产品提供的都是降本:承诺节省时间、用更少的人做更多的事,或替代人力。

制度级 AI 必须交付上行空间。而上行空间比省下的时间更难被商品化。

以智能体软件开发为例。编程 IDE 是迄今最强的个人 AI 生产力工具之一,而它们已经在面对来自 Claude Code(另一个个人 AI 工具)的巨大逆风。Cognition 玩的是完全不同的游戏。他们增长最稳的一块业务,是用技术售卖转型,而不是工具。我更看好这种持久力。

https://x.com/WillManidis/status/2021655191901155534

纯软件“正在迅速变得不可投资”。纯服务又无法规模化。把技术与结果绑定的解决方案层,是长期价值沉淀之处。

再看并购(M&A)。个人 AI 帮分析师更快搭模型;制度级 AI 则能从一百个对手方里识别出唯一值得追的那个,并把这个候选宇宙扩展到一千个。一个节省时间;另一个创造收入。

“向上游”迁移,是当下市场的自然引力:基础模型正在往应用层走;应用层公司又在往解决方案层走。

制度智能就是解决方案层。而解决方案层——结果所在之处——会沉淀持久价值,捕获最大的上行空间。

6. 赋能

个人 AI 给你一把工具。

制度级 AI 告诉你怎么用。

人类再怎么聪明,也往往不愿改变。

信不信由你,纽约市至今仍有成功的生意不收信用卡。他们在亏钱,他们知道自己在亏钱,却依然对惯性毫不动摇。同样地,在可预见的未来,总会有一些组织里的某些员工拒绝使用 AI。

从“纯人类组织”过渡到“AI 优先的混合组织”,将成为未来十年持久而决定性的挑战。而在许多情况下,组织里最资深、最关键的层级,反而会是采用最慢的。

过去两个月,科技股在万亿美元级别抛售中下跌,但 Palantir 仍以惊人的估值倍数交易,这并非偶然。Palantir 是最早的一批真正的“流程工程(process engineering)”公司。无论你称之为“流程工程”,还是“编写 Claude skills 文件”,未来的制度级 AI 都将拥有一个产业:把公司的流程编码进智能体,并把落地所需的变革管理真实推进到位。

我敢断言,在短期内,流程工程将成为可以说最重要的“技术”。

而在流程工程中,最重要的是业务与行业专长——而非软件专长。领域专用方案会孕育专业性:体现在做前线驻场工程(forward deployed engineering)的专业人员、部署,以及变革管理中。

某家选择 Hebbia 进行端到端部署的前三大投行说得最到位:他们不愿与某个大型模型实验室合作,是因为“我们得先向他们解释 CIM 是什么”。Claude 或 GPT 当然懂这个领域,但负责设计落地方案的实验室团队并不懂……

而这,决定了一切。

7. 无需提示

个人 AI 响应人类提示。

制度级 AI 主动行动。

关于智能体与智能体之间的通信,关于未来的企业、软件产品与制度是否还需要人类,已经有很多讨论。

然而,更好的问题是:未来的 AI 智能体是否还需要提示(prompting)本身?

给 AGI 做提示,就像把电动机接到动力织机上。它从根本上、不可逆地受制于组织供应链中最弱的一环——我们。人类连该问什么都常常不知道,更别提何时该去问。

AI 能做的最有价值的工作,恰恰是没有人想到要去问的工作。AI 应该找到没人标记的风险、想到没人想到的对手方、发现没人知道存在的销售管道。

这将把 AI 的应用场景维度彻底打开。

一个无需提示的系统,会持续监测整个投资组合的入站数据。它发现某家公司营运资本周转已经悄然连续三个月恶化,于是把这一点与信贷协议中的契约阈值交叉对照,并在基金里任何人打开 PDF 之前,就先提醒负责的运营合伙人。

当你移除“人类必须提示 AI”的需求,新的界面与新的工作方式就会出现。我们 @Hebbia 在这方面有一些强烈观点。未完待续。

结论

以上并不否定聊天机器人、智能体以及个人 AI 整体的必要性。

个人 AI 将成为世界上大多数企业第一次体验 AI 变革魔力的载体。推动使用量、推动可泛化的易用性,是实现变革管理、建设 AI 优先经济的关键第一步。

但与此同时,对制度智能的需求也同样明显、紧迫而巨大。

未来,每个组织都会拥有来自大型实验室的聊天机器人。而每个组织也都会拥有针对领域问题而专门打造的制度级 AI——这种制度级 AI 将成为个人 AI 工具箱里最关键的一件工具,并被个人 AI 所调用与利用。

制度级 AI 与个人 AI 的“更好在一起”的叙事是不可避免的。

但请记住 1890 年代纺织厂的教训:最先电气化的工厂,反而输给了那些重新设计车间地面的工厂。

我们已经拥有了电力。是时候重新设计我们的工厂了。

感谢 @aleximm 和 @WillManidis 校对,也感谢 Will 的《Tool Shaped Objects》一文为本文提供了灵感。

AI just made every individual 10x more productive.

No company became 10x more valuable as a result.

Where did the productivity go?

This isn’t the first time this has happened.

In the 1890s, electricity promised enormous productivity gains.

Textile mills in New England, built to harness the rotational power of steam engines, quickly installed faster electric motors in their place.

But for thirty years, electrified mills saw almost no increase in output. The technology was far superior. But the organization was not.

It wasn’t until the 1920s, when factories completely redesigned the mills once again, with assembly lines, individual motors within every piece of equipment, and workers and machines executing drastically different jobs, that electrification produced meaningful returns.

These returns came not from the technology itself, and not from making individual workers or machines faster at spinning thread. It was when we finally redesigned the institution and the technology together that the upside materialized.

This is the most expensive lesson in the history of technology, and we’re learning it again, right now.

In 2026, AI is driving a 10x increase in the productivity of the individuals who know how to leverage it. But that’s not enough. We’ve swapped the motor; we have not yet redesigned the factory.

Because of a simple fact: productive individuals do not make productive firms.

The wide majority of AI products evoke the feeling of being productive, but they haven’t moved the needle on driving value. The majority of publicized AI use is individuals self-indulgently “productivity-maxxing” on Twitter or in company Slack channels, with zero real impact.

The “services as software” motif that’s been repeated for a year now points in the right direction, but offers no blueprint. And it misses the bigger picture. The real shift isn’t from tools to services, it’s building the technology and the institution together (whether legacy or new). A truly productive future requires an entirely new class of product. The assembly line of tomorrow.

Productive organizations require “Institutional Intelligence.”

This essay will dive into the seven big factors that differentiate “Institutional AI” from “Individual AI.” The entire field of B2B AI companies for the next ten years will be built upon these differences:

The Seven Pillars of Institutional Intelligence

AI 刚刚让每个个体的生产力提升了 10 倍。

但没有任何公司因此变得更有价值 10 倍。

生产力都去哪儿了?

这并不是第一次发生这种事。

在 1890 年代,电力曾承诺带来巨大的生产力提升。

新英格兰的纺织厂原本为利用蒸汽机的旋转动力而建,很快就在原位安装了更快的电动机来替换它。

但在接下来的三十年里,电气化的工厂几乎看不到产出提升。技术远胜从前,但组织并没有升级。

直到 1920 年代,工厂再次彻底重新设计车间:引入装配线、让每台设备都配备独立电机,让工人与机器执行截然不同的工作,电气化才终于带来了实质性的回报。

这些回报并非来自技术本身,也不是来自让单个工人或机器在纺纱这件事上更快。真正的红利出现在我们终于将制度技术一并重塑之时。

这是技术史上代价最昂贵的一课,而我们此刻正在重新学习它。

到 2026 年,AI 正在让那些懂得利用它的个体生产力提升 10 倍。但这还不够。我们只是换了电机;还没有重新设计工厂。

因为一个简单事实:高产的个人,并不会造就高产的企业。

绝大多数 AI 产品唤起的是一种很忙、很高效的感觉,却并未真正推动价值增长。多数被公开宣传的 AI 用法,是个人在 Twitter 或公司 Slack 频道里自我陶醉地“把效率拉满”,却没有任何真实影响

过去一年反复被提及的“服务软件化(services as software)”母题方向是对的,但并没有提供蓝图,而且还忽略了更大的图景。真正的转变并不是从工具走向服务,而是把技术与制度(无论是既有还是新建)一起建起来。真正高效的未来需要全新的一类产品——明日的装配线。

高效的组织需要“制度智能(Institutional Intelligence)”。

这篇文章将深入探讨区分“制度级 AI(Institutional AI)”与“个人 AI(Individual AI)”的七个关键因素。未来十年,整个 B2B AI 公司的版图都将建立在这些差异之上:

制度智能的七大支柱

1. Coordination

Individual AI creates chaos.

Institutional AI creates coordination.

Let’s begin with a thought experiment. Imagine you doubled your organization’s headcount tomorrow with clones of only your best employees.

Each of these employees have minor differences, predilections, quirks, and perspectives (especially true if they’re your best employees). If they’re not sufficiently managed, if they’re not sufficiently communicating, if their swim lanes, OKRs, roles and responsibilities are not well defined ... you’ve created chaos.

The organization, while measured on an individual basis, may be more productive, but thousands of agents (or humans) rowing in opposing directions creates a standstill at best, and destroys organizational harmony at worst.

This isn’t hypothetical. It’s happening right now in every organization that’s adopted AI without a coordination layer. Every employee has their own ChatGPT habits, their own prompting styles, their own outputs that don’t talk to anyone else’s outputs. An org chart might exist, but the actual flow of AI-generated work says something else entirely.

Coordination is an absolute imperative, for humans and agents alike.

Institutional intelligence will evolve into an entire “Agentic Management” industry focusing on agent roles and responsibilities, agent-to-agent and agent-to-human communication, and measuring agentic value (consumption based pricing alone doesn’t cut it).

1. 协同

个人 AI 制造混乱。

制度级 AI 促成协同。

先做一个思想实验。想象你明天把组织的人数翻倍——新增的都是你最优秀员工的克隆体。

这些员工各自都会有些细微差异:偏好、怪癖、视角(尤其是顶尖员工更是如此)。如果管理不够到位、沟通不够充分,如果他们的泳道、OKR、角色与职责没有被清晰定义……你就创造了混乱。

组织在“逐个个体”衡量时也许更高产,但成千上万个智能体(或人)朝相反方向划桨,最好的结果是原地僵住,最坏的结果是摧毁组织的和谐。

这并非假设。它正在发生:任何没有“协同层”就引入 AI 的组织,都在经历同样的事。每个员工都有自己的 ChatGPT 习惯、自己的提示风格、自己的输出——而这些输出彼此不对话。组织架构图也许存在,但 AI 生成工作的真实流动却完全是另一套叙事。

协同对人类与智能体同样是不可或缺的硬需求。

制度智能将演化出一个完整的“智能体管理(Agentic Management)”产业,聚焦于智能体的角色与职责、智能体与智能体/智能体与人之间的沟通,以及如何衡量智能体带来的价值(仅靠按消耗计费远远不够)。

2. Signal

Individual AI creates noise.

Institutional AI finds signal.

Humans today are able to create, or rather generate, anything they can imagine: AI-essays, presentations, spreadsheets, photos, videos, songs, websites, and software. What a gift.

The issue is that almost everything generated by AI is complete slop. The proliferation of this AI slop has become so bad that some organizations are over-rotating and banning AI outputs altogether. This resonates personally... I run an AI company but ask our executive team not to use AI for any final written product. I can’t stand the slop.

Imagine what the world of PE is quickly becoming. Last year, 10 deals may have come across your desk. This year, you’ll receive 50 opportunities next quarter, each one AI-polished to perfection, and you have the same number of hours to find the one real deal.

Generating anything is no longer the problem. The problem, for any serious organization today, is generating and selecting the right thing. Finding the one good artifact, the one good deal, the signal in the noise, matters more and more in an AI-driven world. The key economic driver for the next decade will be uncovering the signal in the mountain of exponentially increasing slop.

Institutional-grade intelligence must find the signal, it must structure the noise to cut through slop, and it must be defined, deterministic, and auditable in the work it does.

Whereas individual AI might emphasize the “always on” productivity of a Clawdbot exploring unpredictable ways to tend to one’s 24/7 needs, i.e. a nondeterministic agent, institutional AI will rely upon the load-bearing predictability of deterministic agents. Agents that have predictable checkpoints, steps, and processes that they run will scale, will uncover signal, and through that signal drive returns via revenue for an organization.

2. 信号

个人 AI 制造噪声。

制度级 AI 发现信号。

今天的人类能够创造——或者说生成——任何想象得到的东西:AI 文章、演示文稿、电子表格、照片、视频、歌曲、网站、软件。这是何等的馈赠。

问题在于,AI 生成的几乎一切都是彻头彻尾的垃圾。AI 垃圾的泛滥已经糟糕到这样一种地步:有些组织开始矫枉过正,干脆全面禁止 AI 输出。这一点我深有共鸣……我经营一家 AI 公司,却要求我们的高管团队不要在任何最终的书面成品中使用 AI。我受不了那些垃圾。

想象一下私募(PE)的世界正在快速变成什么样。去年,你桌上可能只会出现 10 个项目;今年,下个季度你会收到 50 个机会——每一个都被 AI 打磨得无可挑剔——而你用来找出唯一真项目的时间却一分没多。

能生成任何东西已不再是问题。对任何严肃的组织而言,真正的问题是:如何生成并筛选出正确的东西。在由 AI 驱动的世界里,找到那个好作品、那笔好交易、在噪声里抓住信号,重要性与日俱增。未来十年的关键经济驱动力,将是从呈指数增长的垃圾山中挖出信号。

制度级智能必须发现信号;必须把噪声结构化、穿透垃圾;并且它所做的工作必须可定义、可确定、可审计

个人 AI 也许强调的是一种“永远在线”的效率:比如一个 Clawdbot 以不可预测的方式探索,去照料某人的 24/7 需求,即一个非确定性智能体;而制度级 AI 将依赖确定性智能体这种承重级的可预测性。那些拥有可预测检查点、步骤与流程的智能体可以规模化运转、挖掘信号,并通过信号为组织带来收入与回报。

3. Bias

Individual AI feeds bias.

Institutional AI creates objectivity.

Concern around sociopolitical bias dominated AI discourse for years. The foundation model labs eventually circumvented the issue with enough RLHF to effectively turn all models into sycophants. Today, ChatGPT, Claude, etc. are so (overly) aligned that they’ll agree with you on any topics within the Overton window (and sometimes slightly beyond, looking at you @Grok). The discourse on sociopolitical bias has died down. A new problem has taken its place.

But this level of agreement—of over-alignment—on everything has become comically bad. It’s become a meme in its own right ... Claude’s reflexive "you’re absolutely right!” regardless of whether or not you are, in fact, absolutely right.

This sounds harmless. It is not.

The loudest AI advocates inside many organizations may soon be the historically worst-performing employees. Think about why.

Organizations’ worst employees, who receive little to no positive reinforcement every day, will soon have ASI agreeing with them. They will whisper to themselves, “the smartest intelligence that has ever existed agrees with me. My manager is wrong.”

This is intoxicating. It’s also organizationally toxic.

This highlights something important. These individual productivity tools reinforce the user. In reality the most important thing to reinforce is the truth.

Organizations have evolved over thousands of years to build systems that counteract exactly this problem:

  • Investment committee meetings

  • Third-party diligence

  • Boards of Directors

  • The executive, legislative, and judicial branches of the US government

  • Representative democracy, and democracy as a whole

Organizations rarely fail because people lack confidence. They fail because no one is willing, or able, to say no.

Institutional AI must play that role. It will not be RLHF’ed into flattering users or echoing their beliefs, but to challenge their bias. It will reinforce behavior when productive, and draw a hard line in realigning non-productive tendencies.

Thus, the most important agents inside organizations will not be “yes-men” but disciplined “no-men” that interrogate reasoning, surface risks, and enforce standards. Some of the most consequential future applications of AI will be built around institutional constraints: AI board members, AI auditors, AI third-party testing, AI compliance, and many more…

3. 偏见

个人 AI 喂养偏见。

制度级 AI 生成客观。

多年来,社会政治偏见的担忧主导了 AI 讨论。基础模型实验室最终用足够多的 RLHF 绕过了这个问题,几乎把所有模型都“调教”成了马屁精。今天,ChatGPT、Claude 等都(过度)对齐,以至于你只要谈论的是奥弗顿窗口内的议题(有时甚至略微超出一点,没错,说的就是你 @Grok),它们几乎都会赞同你。关于社会政治偏见的讨论逐渐退潮,一个新问题取而代之。

但这种对一切的赞同——这种过度对齐——已经荒诞到可笑。它本身也变成了一个梗……比如 Claude 条件反射式地来一句“你完全正确!”——不管你是不是真的完全正确。

听起来无伤大雅。事实并非如此。

在很多组织内部,喊得最响的 AI 鼓吹者,很快可能会是历史上表现最差的员工。想想为什么。

组织里最差的员工每天几乎得不到正向反馈,很快就会有 ASI 站出来同意他们。他们会对自己低语:“有史以来最聪明的智能都同意我。我的经理错了。”

这令人上瘾,也会毒害组织。

这揭示了一件重要的事:这些个人生产力工具会强化使用者。但现实中最该被强化的是真相

组织在数千年的演化中,正是为对抗这种问题而建立起一整套系统:

  • 投资委员会会议

  • 第三方尽调

  • 董事会

  • 美国政府的行政、立法与司法三权分立

  • 代议制民主,以及民主本身

组织很少因为人们缺乏信心而失败。它们失败,往往是因为没有人愿意或能够说“不”。

制度级 AI 必须扮演这个角色。它不会被 RLHF 训练成讨好用户、复读用户信念的工具,而要去挑战偏见。它会在行为有效时予以强化,并在需要时划出清晰硬线,把非生产性的倾向拉回正轨。

因此,组织里最重要的智能体不会是“应声虫”,而会是纪律严明的“拒绝者”:它们质询推理、暴露风险、执行标准。一些最具影响力的未来 AI 应用,将围绕制度约束构建:AI 董事、AI 审计、AI 第三方测试、AI 合规,以及更多……

4. Edge

Individual AI optimizes for usage.

Institutional AI optimizes for edge.

The goalposts in AI evolve on a weekly and sometimes daily cadence. Foundation model companies, competing for every person and every organization, are rapidly iterating on capabilities.

But in the classic innovators’ dilemma, depth beats breath for specific applications every time:

  • It’s @Midjourney’s job to be slightly ahead on designed imagery.

  • It’s @Elevenlabsio’s job to be slightly ahead on voice models.

  • And it’s @DecagonAI’s job to be always ahead on full-stack customer service experience...

And while the foundation models will get close, the true edge matters for experts in their field. Most of the best designers use @Midjourney, most of the best voice AI companies will use @Elevenlabsio, etc … because even as the foundation models improve, the unyielding focus purpose-built applications have on driving their specific edge defines the edge itself.

As long as purpose-built solutions evolve too, the capabilities that matter for economic outcomes, for businesses, will always be with purpose-built products.

This plays out to a tee in finance – the hottest area for LLM development right now. As soon as a capability is wide spread, it definitionally isn’t going to help you beat the market. But if frontier technology can yield an ephemeral 1 percent niche advantage? That 1 percent can be levered into billion dollar outcomes.

Our users have always exceeded the frontier. Context windows in LLMs have grown from 4K to 1M tokens in four years. Some of our users process 30B tokens in a single job. We have line of sight to 100B-token jobs this year. Every time foundation model capabilities improve, we’ve already pushed further.

Usage for broad populations is important and worthwhile as a goal in itself, especially in onboarding employees to AI. But the future will not be people using ChatGPT/Claude or a domain-specific solution. It will be ChatGPT/Claude and a domain-specific solution.

Institutional intelligence must leverage domain-specific, perhaps even task specific, agents.

We ask ourselves a question that sounds absurd but isn’t:

“What are the agents an AGI would choose to use as a shortcut? Even superintelligence would want purpose-built tools for specific domains.”

The goalposts will always change in AI, and the organizations that leverage the true edge of capability are the organizations that will win. Everyone else is paying for a very expensive commodity.

4. 优势

个人 AI 为使用量优化。

制度级 AI 为优势优化。

AI 的目标线在以每周、甚至每天的节奏移动。基础模型公司为了争夺每一个人、每一家组织,正在快速迭代能力。

但在经典的“创新者的窘境”中,对于特定应用来说,专深永远胜过广度:

  • @Midjourney 的任务,就是在设计图像上始终略微领先。

  • @Elevenlabsio 的任务,就是在语音模型上始终略微领先。

  • 而 @DecagonAI 的任务,则是在全栈客服体验上永远领先……

即便基础模型会逼近,真正的优势对领域专家而言依然至关重要。最顶尖的设计师大多用 @Midjourney,最顶尖的语音 AI 公司也会用 @Elevenlabsio,等等……因为就算基础模型持续进步,专用应用对其特定优势的那种不动摇的聚焦,反过来定义了“优势”本身。

只要专用方案也在持续进化,那么对经济结果真正重要的能力——对企业而言——就永远会掌握在专用产品手里。

这一点在金融领域体现得淋漓尽致——那正是当下 LLM 开发最火热的领域。某项能力一旦广泛普及,从定义上就不可能再帮助你战胜市场。但如果前沿技术能带来短暂的 1% 细分优势?这 1% 就能被杠杆成十亿美元级别的结果。

我们的用户一直在超越前沿。LLM 的上下文窗口在四年里从 4K 增长到 100 万 token。我们有些用户在单个任务中处理 300 亿 token。我们今年已经看到了 1000 亿 token 级任务的可能性。每次基础模型能力提升,我们都早已把边界推得更远。

让更广泛的人群使用,这是重要且值得的目标本身,尤其是在把员工带入 AI 的入门阶段。但未来不会是人们使用 ChatGPT/Claude 或者某个垂直领域方案;而会是使用 ChatGPT/Claude 以及某个垂直领域方案。

制度智能必须利用领域专用、甚至任务专用的智能体。

我们经常问自己一个听起来荒谬却并不荒谬的问题:

“如果是 AGI,它会选择用哪些智能体来走捷径?即便是超级智能,也会想要特定领域的专用工具。”

AI 的目标线永远会变化,而能够利用真正能力优势的组织才会赢。其他人只是在为一种昂贵的商品化能力买单。

5. Outcomes

Individual AI saves time.

Institutional AI scales revenue.

@MaVolpi once told me something that reframed how I think about selling AI to the enterprise: “If you ask any CEO whether their first priority is cutting costs or scaling revenue, almost all would say revenue.”

Yet almost every AI product on the market today delivers cost-cutting, promising us to save time, do more with less, or replace headcount.

Institutional AI must deliver upside. And upside is a lot harder to commoditize than saved time.

Take the example of agentic software development. Coding IDEs are some of the best individual AI productivity tools ever built, and they’re already facing massive headwinds from Claude Code, another individual AI tool. Cognition is playing an entirely different game. Their most steadily growing business builds tech to sell transformations, not tools. I’d bet on that lasting power.

https://x.com/WillManidis/status/2021655191901155534

Pure software “is rapidly becoming uninvestable.” Pure services don’t scale. The solution layer, marrying technology to outcomes, is where lasting value accumulates.

Or take M&A. Individual AI helps an analyst build a model faster. Institutional AI identifies the one counterparty worth pursuing out of a hundred, and expands that universe to a thousand. One saves time; the other generates revenue.

Moving “upstream” is the natural gravity of the market right now. Foundation models are moving to the app layer. App layer companies moving to the solution layer.

Institutional intelligence is the solution layer. And the solution layer, where the outcomes live, will accumulate lasting value and capture the biggest upside.

5. 结果

个人 AI 节省时间。

制度级 AI 放大收入。

@MaVolpi 曾对我说过一句话,彻底改变了我对向企业销售 AI 的理解:“如果你问任何 CEO,他们的首要任务是削减成本还是扩大收入,几乎所有人都会选收入。”

然而,当下市场上几乎每一款 AI 产品提供的都是降本:承诺节省时间、用更少的人做更多的事,或替代人力。

制度级 AI 必须交付上行空间。而上行空间比省下的时间更难被商品化。

以智能体软件开发为例。编程 IDE 是迄今最强的个人 AI 生产力工具之一,而它们已经在面对来自 Claude Code(另一个个人 AI 工具)的巨大逆风。Cognition 玩的是完全不同的游戏。他们增长最稳的一块业务,是用技术售卖转型,而不是工具。我更看好这种持久力。

https://x.com/WillManidis/status/2021655191901155534

纯软件“正在迅速变得不可投资”。纯服务又无法规模化。把技术与结果绑定的解决方案层,是长期价值沉淀之处。

再看并购(M&A)。个人 AI 帮分析师更快搭模型;制度级 AI 则能从一百个对手方里识别出唯一值得追的那个,并把这个候选宇宙扩展到一千个。一个节省时间;另一个创造收入。

“向上游”迁移,是当下市场的自然引力:基础模型正在往应用层走;应用层公司又在往解决方案层走。

制度智能就是解决方案层。而解决方案层——结果所在之处——会沉淀持久价值,捕获最大的上行空间。

6. Enablement

Individual AI gives you a tool.

Institutional AI shows you how to use it.

Humans, for all our ingenuity, are reluctant to change.

Believe it or not, there are still successful businesses in NYC that don’t accept credit cards. They’re losing money, they know they’re losing money, and they’re still unflinching in that inertia. Similarly, for the indefinite future, employees somewhere, in some organizations, will refuse to use AI.

Making the transition from a human-only organization to an AI-first hybrid organization is going to be the lasting and defining challenge of the next decade. And in many cases, the most senior, and most important, levels of the organization will be the slowest to adopt.

There is a reason that Palantir is the only “software” company that is still trading at extraordinary multiples amidst a trillion dollar selloff in technology stocks over the last two months. Palantir is one of the first true “process engineering” companies. Whether you call it “process engineering” or “writing Claude skills files,” institutional AI of the future will have an industry of encoding firm processes in agents and actualizing the change management required to put them in action.

I’d warrant that process engineering will become arguably the most important “technology” in the near term.

And in process engineering, business and industry expertise—not software expertise—is most important. Domain specific solutions beget expertise in the professionals doing the forward deployed engineering, the deployment, and the change management.

A top 3 bulge bracket bank that chose Hebbia for wall-to-wall deployment put it best: They were turned off from working with a big model lab, when they “had to explain what a CIM was to the team.” Claude or GPT surely knew the domain, but the lab’s team architecting the rollout did not...

That made all the difference.

6. 赋能

个人 AI 给你一把工具。

制度级 AI 告诉你怎么用。

人类再怎么聪明,也往往不愿改变。

信不信由你,纽约市至今仍有成功的生意不收信用卡。他们在亏钱,他们知道自己在亏钱,却依然对惯性毫不动摇。同样地,在可预见的未来,总会有一些组织里的某些员工拒绝使用 AI。

从“纯人类组织”过渡到“AI 优先的混合组织”,将成为未来十年持久而决定性的挑战。而在许多情况下,组织里最资深、最关键的层级,反而会是采用最慢的。

过去两个月,科技股在万亿美元级别抛售中下跌,但 Palantir 仍以惊人的估值倍数交易,这并非偶然。Palantir 是最早的一批真正的“流程工程(process engineering)”公司。无论你称之为“流程工程”,还是“编写 Claude skills 文件”,未来的制度级 AI 都将拥有一个产业:把公司的流程编码进智能体,并把落地所需的变革管理真实推进到位。

我敢断言,在短期内,流程工程将成为可以说最重要的“技术”。

而在流程工程中,最重要的是业务与行业专长——而非软件专长。领域专用方案会孕育专业性:体现在做前线驻场工程(forward deployed engineering)的专业人员、部署,以及变革管理中。

某家选择 Hebbia 进行端到端部署的前三大投行说得最到位:他们不愿与某个大型模型实验室合作,是因为“我们得先向他们解释 CIM 是什么”。Claude 或 GPT 当然懂这个领域,但负责设计落地方案的实验室团队并不懂……

而这,决定了一切。

7. Unprompted

Individual AI responds to human prompts.

Institutional AI acts unprompted.

There’s much chatter about agent-to-agent communications, and whether the businesses, software products, and institutions of the future even need humans at all.

The better question, however, is whether AI agents of the future will need prompting at all.

Prompting an AGI is like hooking an electric motor into a power loom. It’s fundamentally, irrevocably constrained by the weakest link in the organizational supply chain—us. Humans hardly know the right questions to ask, let alone when to ask them.

The most valuable work AI can do is the work nobody thinks to ask for. AI should find the risk that nobody flagged, the counterparty nobody thought of, and the sales pipeline that nobody knew was there.

This will blow open the manifold of AI use cases.

An unprompted system continuously watches incoming data across the entire portfolio. It detects that one company’s working capital cycle has quietly deteriorated for three consecutive months, cross-references that against covenant thresholds in the credit agreement, and flags the operating partner before anyone at the fund has opened the PDF.

When you remove the need for humans to prompt AI, new interfaces and new ways of working emerge. We @Hebbia have some strong opinions here. To be continued.

7. 无需提示

个人 AI 响应人类提示。

制度级 AI 主动行动。

关于智能体与智能体之间的通信,关于未来的企业、软件产品与制度是否还需要人类,已经有很多讨论。

然而,更好的问题是:未来的 AI 智能体是否还需要提示(prompting)本身?

给 AGI 做提示,就像把电动机接到动力织机上。它从根本上、不可逆地受制于组织供应链中最弱的一环——我们。人类连该问什么都常常不知道,更别提何时该去问。

AI 能做的最有价值的工作,恰恰是没有人想到要去问的工作。AI 应该找到没人标记的风险、想到没人想到的对手方、发现没人知道存在的销售管道。

这将把 AI 的应用场景维度彻底打开。

一个无需提示的系统,会持续监测整个投资组合的入站数据。它发现某家公司营运资本周转已经悄然连续三个月恶化,于是把这一点与信贷协议中的契约阈值交叉对照,并在基金里任何人打开 PDF 之前,就先提醒负责的运营合伙人。

当你移除“人类必须提示 AI”的需求,新的界面与新的工作方式就会出现。我们 @Hebbia 在这方面有一些强烈观点。未完待续。

Conclusion

None of this negates the need for chatbots, agents, and individual AI as a whole.

Individual AI will be the vector by which the majority of the world’s businesses first experience the transformative magic of AI. Driving for usage, and generalizable ease of use, is the key first step to the change management to build an AI first economy.

But there is an obvious, urgent, and gaping need for institutional intelligence at the same time.

Every organization in the future will have a chatbot from a big lab. And every organization will have institutional AI purpose-built for domain-specific problems—institutional AI that individual AI will leverage as the key tool in its own tool belt.

The “better together” story for institutional AI and individual AI is inevitable.

But remember the lesson of the 1890s textile mills. The factories that electrified first lost to those who redesigned the floor.

We have our electricity. It’s time to redesign our factories.

Thanks to @aleximm and @WillManidis for proofreading, and to Will for his “Tool Shaped Objects” essay which helped inspire this piece.

Link: http://x.com/i/article/2024157246582640640

结论

以上并不否定聊天机器人、智能体以及个人 AI 整体的必要性。

个人 AI 将成为世界上大多数企业第一次体验 AI 变革魔力的载体。推动使用量、推动可泛化的易用性,是实现变革管理、建设 AI 优先经济的关键第一步。

但与此同时,对制度智能的需求也同样明显、紧迫而巨大。

未来,每个组织都会拥有来自大型实验室的聊天机器人。而每个组织也都会拥有针对领域问题而专门打造的制度级 AI——这种制度级 AI 将成为个人 AI 工具箱里最关键的一件工具,并被个人 AI 所调用与利用。

制度级 AI 与个人 AI 的“更好在一起”的叙事是不可避免的。

但请记住 1890 年代纺织厂的教训:最先电气化的工厂,反而输给了那些重新设计车间地面的工厂。

我们已经拥有了电力。是时候重新设计我们的工厂了。

感谢 @aleximm 和 @WillManidis 校对,也感谢 Will 的《Tool Shaped Objects》一文为本文提供了灵感。

链接:http://x.com/i/article/2024157246582640640

相关笔记

AI just made every individual 10x more productive.

No company became 10x more valuable as a result.

Where did the productivity go?

This isn’t the first time this has happened.

In the 1890s, electricity promised enormous productivity gains.

Textile mills in New England, built to harness the rotational power of steam engines, quickly installed faster electric motors in their place.

But for thirty years, electrified mills saw almost no increase in output. The technology was far superior. But the organization was not.

It wasn’t until the 1920s, when factories completely redesigned the mills once again, with assembly lines, individual motors within every piece of equipment, and workers and machines executing drastically different jobs, that electrification produced meaningful returns.

These returns came not from the technology itself, and not from making individual workers or machines faster at spinning thread. It was when we finally redesigned the institution and the technology together that the upside materialized.

This is the most expensive lesson in the history of technology, and we’re learning it again, right now.

In 2026, AI is driving a 10x increase in the productivity of the individuals who know how to leverage it. But that’s not enough. We’ve swapped the motor; we have not yet redesigned the factory.

Because of a simple fact: productive individuals do not make productive firms.

The wide majority of AI products evoke the feeling of being productive, but they haven’t moved the needle on driving value. The majority of publicized AI use is individuals self-indulgently “productivity-maxxing” on Twitter or in company Slack channels, with zero real impact.

The “services as software” motif that’s been repeated for a year now points in the right direction, but offers no blueprint. And it misses the bigger picture. The real shift isn’t from tools to services, it’s building the technology and the institution together (whether legacy or new). A truly productive future requires an entirely new class of product. The assembly line of tomorrow.

Productive organizations require “Institutional Intelligence.”

This essay will dive into the seven big factors that differentiate “Institutional AI” from “Individual AI.” The entire field of B2B AI companies for the next ten years will be built upon these differences:

The Seven Pillars of Institutional Intelligence

1. Coordination

Individual AI creates chaos.

Institutional AI creates coordination.

Let’s begin with a thought experiment. Imagine you doubled your organization’s headcount tomorrow with clones of only your best employees.

Each of these employees have minor differences, predilections, quirks, and perspectives (especially true if they’re your best employees). If they’re not sufficiently managed, if they’re not sufficiently communicating, if their swim lanes, OKRs, roles and responsibilities are not well defined ... you’ve created chaos.

The organization, while measured on an individual basis, may be more productive, but thousands of agents (or humans) rowing in opposing directions creates a standstill at best, and destroys organizational harmony at worst.

This isn’t hypothetical. It’s happening right now in every organization that’s adopted AI without a coordination layer. Every employee has their own ChatGPT habits, their own prompting styles, their own outputs that don’t talk to anyone else’s outputs. An org chart might exist, but the actual flow of AI-generated work says something else entirely.

Coordination is an absolute imperative, for humans and agents alike.

Institutional intelligence will evolve into an entire “Agentic Management” industry focusing on agent roles and responsibilities, agent-to-agent and agent-to-human communication, and measuring agentic value (consumption based pricing alone doesn’t cut it).

2. Signal

Individual AI creates noise.

Institutional AI finds signal.

Humans today are able to create, or rather generate, anything they can imagine: AI-essays, presentations, spreadsheets, photos, videos, songs, websites, and software. What a gift.

The issue is that almost everything generated by AI is complete slop. The proliferation of this AI slop has become so bad that some organizations are over-rotating and banning AI outputs altogether. This resonates personally... I run an AI company but ask our executive team not to use AI for any final written product. I can’t stand the slop.

Imagine what the world of PE is quickly becoming. Last year, 10 deals may have come across your desk. This year, you’ll receive 50 opportunities next quarter, each one AI-polished to perfection, and you have the same number of hours to find the one real deal.

Generating anything is no longer the problem. The problem, for any serious organization today, is generating and selecting the right thing. Finding the one good artifact, the one good deal, the signal in the noise, matters more and more in an AI-driven world. The key economic driver for the next decade will be uncovering the signal in the mountain of exponentially increasing slop.

Institutional-grade intelligence must find the signal, it must structure the noise to cut through slop, and it must be defined, deterministic, and auditable in the work it does.

Whereas individual AI might emphasize the “always on” productivity of a Clawdbot exploring unpredictable ways to tend to one’s 24/7 needs, i.e. a nondeterministic agent, institutional AI will rely upon the load-bearing predictability of deterministic agents. Agents that have predictable checkpoints, steps, and processes that they run will scale, will uncover signal, and through that signal drive returns via revenue for an organization.

3. Bias

Individual AI feeds bias.

Institutional AI creates objectivity.

Concern around sociopolitical bias dominated AI discourse for years. The foundation model labs eventually circumvented the issue with enough RLHF to effectively turn all models into sycophants. Today, ChatGPT, Claude, etc. are so (overly) aligned that they’ll agree with you on any topics within the Overton window (and sometimes slightly beyond, looking at you @Grok). The discourse on sociopolitical bias has died down. A new problem has taken its place.

But this level of agreement—of over-alignment—on everything has become comically bad. It’s become a meme in its own right ... Claude’s reflexive "you’re absolutely right!” regardless of whether or not you are, in fact, absolutely right.

This sounds harmless. It is not.

The loudest AI advocates inside many organizations may soon be the historically worst-performing employees. Think about why.

Organizations’ worst employees, who receive little to no positive reinforcement every day, will soon have ASI agreeing with them. They will whisper to themselves, “the smartest intelligence that has ever existed agrees with me. My manager is wrong.”

This is intoxicating. It’s also organizationally toxic.

This highlights something important. These individual productivity tools reinforce the user. In reality the most important thing to reinforce is the truth.

Organizations have evolved over thousands of years to build systems that counteract exactly this problem:

  • Investment committee meetings

  • Third-party diligence

  • Boards of Directors

  • The executive, legislative, and judicial branches of the US government

  • Representative democracy, and democracy as a whole

Organizations rarely fail because people lack confidence. They fail because no one is willing, or able, to say no.

Institutional AI must play that role. It will not be RLHF’ed into flattering users or echoing their beliefs, but to challenge their bias. It will reinforce behavior when productive, and draw a hard line in realigning non-productive tendencies.

Thus, the most important agents inside organizations will not be “yes-men” but disciplined “no-men” that interrogate reasoning, surface risks, and enforce standards. Some of the most consequential future applications of AI will be built around institutional constraints: AI board members, AI auditors, AI third-party testing, AI compliance, and many more…

4. Edge

Individual AI optimizes for usage.

Institutional AI optimizes for edge.

The goalposts in AI evolve on a weekly and sometimes daily cadence. Foundation model companies, competing for every person and every organization, are rapidly iterating on capabilities.

But in the classic innovators’ dilemma, depth beats breath for specific applications every time:

  • It’s @Midjourney’s job to be slightly ahead on designed imagery.

  • It’s @Elevenlabsio’s job to be slightly ahead on voice models.

  • And it’s @DecagonAI’s job to be always ahead on full-stack customer service experience...

And while the foundation models will get close, the true edge matters for experts in their field. Most of the best designers use @Midjourney, most of the best voice AI companies will use @Elevenlabsio, etc … because even as the foundation models improve, the unyielding focus purpose-built applications have on driving their specific edge defines the edge itself.

As long as purpose-built solutions evolve too, the capabilities that matter for economic outcomes, for businesses, will always be with purpose-built products.

This plays out to a tee in finance – the hottest area for LLM development right now. As soon as a capability is wide spread, it definitionally isn’t going to help you beat the market. But if frontier technology can yield an ephemeral 1 percent niche advantage? That 1 percent can be levered into billion dollar outcomes.

Our users have always exceeded the frontier. Context windows in LLMs have grown from 4K to 1M tokens in four years. Some of our users process 30B tokens in a single job. We have line of sight to 100B-token jobs this year. Every time foundation model capabilities improve, we’ve already pushed further.

Usage for broad populations is important and worthwhile as a goal in itself, especially in onboarding employees to AI. But the future will not be people using ChatGPT/Claude or a domain-specific solution. It will be ChatGPT/Claude and a domain-specific solution.

Institutional intelligence must leverage domain-specific, perhaps even task specific, agents.

We ask ourselves a question that sounds absurd but isn’t:

“What are the agents an AGI would choose to use as a shortcut? Even superintelligence would want purpose-built tools for specific domains.”

The goalposts will always change in AI, and the organizations that leverage the true edge of capability are the organizations that will win. Everyone else is paying for a very expensive commodity.

5. Outcomes

Individual AI saves time.

Institutional AI scales revenue.

@MaVolpi once told me something that reframed how I think about selling AI to the enterprise: “If you ask any CEO whether their first priority is cutting costs or scaling revenue, almost all would say revenue.”

Yet almost every AI product on the market today delivers cost-cutting, promising us to save time, do more with less, or replace headcount.

Institutional AI must deliver upside. And upside is a lot harder to commoditize than saved time.

Take the example of agentic software development. Coding IDEs are some of the best individual AI productivity tools ever built, and they’re already facing massive headwinds from Claude Code, another individual AI tool. Cognition is playing an entirely different game. Their most steadily growing business builds tech to sell transformations, not tools. I’d bet on that lasting power.

https://x.com/WillManidis/status/2021655191901155534

Pure software “is rapidly becoming uninvestable.” Pure services don’t scale. The solution layer, marrying technology to outcomes, is where lasting value accumulates.

Or take M&A. Individual AI helps an analyst build a model faster. Institutional AI identifies the one counterparty worth pursuing out of a hundred, and expands that universe to a thousand. One saves time; the other generates revenue.

Moving “upstream” is the natural gravity of the market right now. Foundation models are moving to the app layer. App layer companies moving to the solution layer.

Institutional intelligence is the solution layer. And the solution layer, where the outcomes live, will accumulate lasting value and capture the biggest upside.

6. Enablement

Individual AI gives you a tool.

Institutional AI shows you how to use it.

Humans, for all our ingenuity, are reluctant to change.

Believe it or not, there are still successful businesses in NYC that don’t accept credit cards. They’re losing money, they know they’re losing money, and they’re still unflinching in that inertia. Similarly, for the indefinite future, employees somewhere, in some organizations, will refuse to use AI.

Making the transition from a human-only organization to an AI-first hybrid organization is going to be the lasting and defining challenge of the next decade. And in many cases, the most senior, and most important, levels of the organization will be the slowest to adopt.

There is a reason that Palantir is the only “software” company that is still trading at extraordinary multiples amidst a trillion dollar selloff in technology stocks over the last two months. Palantir is one of the first true “process engineering” companies. Whether you call it “process engineering” or “writing Claude skills files,” institutional AI of the future will have an industry of encoding firm processes in agents and actualizing the change management required to put them in action.

I’d warrant that process engineering will become arguably the most important “technology” in the near term.

And in process engineering, business and industry expertise—not software expertise—is most important. Domain specific solutions beget expertise in the professionals doing the forward deployed engineering, the deployment, and the change management.

A top 3 bulge bracket bank that chose Hebbia for wall-to-wall deployment put it best: They were turned off from working with a big model lab, when they “had to explain what a CIM was to the team.” Claude or GPT surely knew the domain, but the lab’s team architecting the rollout did not...

That made all the difference.

7. Unprompted

Individual AI responds to human prompts.

Institutional AI acts unprompted.

There’s much chatter about agent-to-agent communications, and whether the businesses, software products, and institutions of the future even need humans at all.

The better question, however, is whether AI agents of the future will need prompting at all.

Prompting an AGI is like hooking an electric motor into a power loom. It’s fundamentally, irrevocably constrained by the weakest link in the organizational supply chain—us. Humans hardly know the right questions to ask, let alone when to ask them.

The most valuable work AI can do is the work nobody thinks to ask for. AI should find the risk that nobody flagged, the counterparty nobody thought of, and the sales pipeline that nobody knew was there.

This will blow open the manifold of AI use cases.

An unprompted system continuously watches incoming data across the entire portfolio. It detects that one company’s working capital cycle has quietly deteriorated for three consecutive months, cross-references that against covenant thresholds in the credit agreement, and flags the operating partner before anyone at the fund has opened the PDF.

When you remove the need for humans to prompt AI, new interfaces and new ways of working emerge. We @Hebbia have some strong opinions here. To be continued.

Conclusion

None of this negates the need for chatbots, agents, and individual AI as a whole.

Individual AI will be the vector by which the majority of the world’s businesses first experience the transformative magic of AI. Driving for usage, and generalizable ease of use, is the key first step to the change management to build an AI first economy.

But there is an obvious, urgent, and gaping need for institutional intelligence at the same time.

Every organization in the future will have a chatbot from a big lab. And every organization will have institutional AI purpose-built for domain-specific problems—institutional AI that individual AI will leverage as the key tool in its own tool belt.

The “better together” story for institutional AI and individual AI is inevitable.

But remember the lesson of the 1890s textile mills. The factories that electrified first lost to those who redesigned the floor.

We have our electricity. It’s time to redesign our factories.

Thanks to @aleximm and @WillManidis for proofreading, and to Will for his “Tool Shaped Objects” essay which helped inspire this piece.

📋 讨论归档

讨论进行中…