[by:whisper.cpp] [00:00.00]大家好,歡迎大家來到「Lit and Space Pockest」 [00:02.50]我是Alessio,會員,和CTO在職業的職業會員 [00:05.74]我是Makojo Swicks, founder of SmallAI [00:08.84]今天我們有David Luan, co-founder of ADEPT,在工作室,歡迎 [00:12.98]謝謝你 [00:14.10]一段時間在工作,我遇到你在VC的社交平台上 [00:17.98]你也說了,你很興奮,我們終於能夠做到這件事了 [00:21.88]對,很高興認識你 [00:23.88]我們想介紹你的職業,然後再說一下你剛才說了什麼,在你的連結,什麼人應該知道你 [00:32.02]你開始了一間公司,是第一次在實際視頻的視頻研究,例如DEXTRO,那是你的路,在你導致的AI,你開始了XON,然後你開始了30年,你開始了OpenAI? [00:47.06]對,30、35年,或是在那裡,或是在那裡,VP Avenge,兩年半,兩年半後, [00:53.48]我們在2022年開始了一個大型模式的創新 [00:57.08]然後在2022年開始了一個大型模式的創新 [01:00.32]所以那是一個短暫的CV [01:02.98]是否有其他東西? [01:03.98]對,是否有其他東西? [01:04.98]你覺得要做什麼? [01:05.98]或是人們應該知道更多? [01:07.98]我猜是一個比較大的故事 [01:09.48]是加入OpenAI比較早期的 [01:11.98]然後就做了兩、三個月的研究 [01:15.48]那是很有趣的 [01:16.48]第二或第三天的我的時間在OpenAI [01:18.98]Gregg and Ilya 找我住在房間,我們說要拿到我們的創新,我們會去… [01:23.98]我看過很多創新的工作 [01:25.98]所以那是很有趣的 [01:26.98]就在結合了一堆團隊 [01:28.98]有幾個早期的領導人已經有了 [01:30.98]公司的資料項目是很努力的 [01:32.98]然後再多次地在大型研究中放大型的圖案 [01:35.98]我們在做基本研究 [01:36.98]所以我花了很多時間在做這個 [01:37.98]然後我再加上Google的LM項目 [01:39.98]但也加上Google的Brain [01:41.98]是一個Brain的領導人,更多次地 [01:42.98]你知道,有幾個不同的領導人在AI的研究 [01:46.98]我們在2012 before prehistory [01:48.98]很多人很討厭我 [01:50.98]我跟你們三個最好的朋友 [01:51.98]寫了一個研究的文件 [01:53.98]從2012到2017 [01:56.98]我覺得遊戲的改善在2017 [01:58.98]然後很多學生都沒有發現 [01:59.98]但是我們在OpenAI上真的做了 [02:01.98]我想大部分的幫助是 [02:02.98]Ilya的 constant beating of the drum [02:04.98]讓世界被遮蓋在data centers [02:06.98]還有其他人需要… [02:07.98]對,我覺得我們有確定在那裡 [02:10.98]但沒有到我們開始看到 [02:11.98]結果的結果,那是我們要去的 [02:14.98]但也有一個部分 [02:15.98]是在OpenAI上 [02:16.98]我第一次加入 [02:17.98]我認為一件事我必須要做 [02:19.98]是如何告訴我們 [02:20.98]我們是否有不同的觀點 [02:22.98]比起我們是更小的GoogleBrain [02:25.98]或是我們在OpenAI上 [02:26.98]只要生活在SF [02:27.98]然後不想接受Mountain View [02:28.98]或不想要生活在London [02:29.98]那是不足夠的 [02:31.98]利用你的技術活動 [02:33.98]所以我們真的… [02:34.98]我花了很多時間在推廣這個 [02:36.98]就是我們要怎麼 [02:37.98]要專注在 [02:38.98]一個大學生的大學生 [02:41.98]你從最底下的研究 [02:44.98]變成了 [02:45.98]如何讓你放棄這個環境 [02:47.98]而讓你覺得 [02:48.98]什麼是大學生的大學生 [02:50.98]想要展現 [02:51.98]然後你把他們解決 [02:52.98]所有的財困 [02:53.98]不管是否要在創意 [02:54.98]創作什麼 [02:55.98]這就變成了 [02:56.98]大學生的大學生 [02:57.98]對嗎 [02:58.98]然後現在的改變 [02:59.98]是我認為 [03:00.98]第一次加入AiPrice [03:01.98]在下一幾年 [03:02.98]會是最深的 [03:03.98] co-design [03:04.98]和 co-evolution [03:05.98]產品和資料 [03:07.98]和實際技術 [03:08.98]而我認為 [03:09.98]每個技術的技術 [03:10.98]都會做得很好 [03:11.98]那是一大部分 [03:12.98]為何我開始深入 [03:13.98]你提及Dota [03:14.98]哪些記憶在想 [03:16.98]從RL 和 Transformers [03:18.98]在時間中 [03:19.98]然後我認為 [03:20.98]製造的工具 [03:21.98]更加在LM 上 [03:23.98]然後離開 [03:24.98]更多的Agent Simulation [03:25.98]工作 [03:26.98]像在移動的道路 [03:27.98]我覺得Agent [03:28.98]是一個 [03:29.98]完全正確的長途 [03:30.98]你只要去找 [03:31.98]AGI 是吧 [03:32.98]你會說 [03:33.98]首先 [03:34.98]我其實不喜歡AGI [03:35.98]用人的改變 [03:36.98]因為我真的不想 [03:37.98]這樣會發生 [03:38.98]我認為這個改變 [03:39.98]AGI 是一些 [03:40.98]人們表現的 [03:41.98]非常值得的技術 [03:43.98]是一個 [03:44.98]極端的看法 [03:45.98]和人的改變 [03:46.98]我認為 [03:47.98]我比較有興趣 [03:48.98]AGI 的改變 [03:49.98]就是 [03:50.98]一個模式 [03:51.98]可以做任何的 [03:52.98]人能做的 [03:53.98]如果你想到 [03:54.98]超級有趣 [03:55.98]Agent [03:56.98]是一種 [03:57.98]自然的 [03:58.98]改變 [03:59.98]所以 [04:00.98]所有的工作 [04:01.98]我們在RL [04:02.98]這些技術 [04:03.98]導致我們 [04:04.98]有很清楚的 [04:05.98]形容 [04:06.98]你需要增加 [04:07.98]你需要增加 [04:08.98]對 [04:09.98]而自然的LM [04:10.98]形容 [04:11.98]沒有出現 [04:12.98]我認為 [04:13.98]我們 [04:14.98]在這個場地 [04:15.98]有很多想法 [04:16.98]想想 [04:17.98]我們如何解決 [04:18.98]問題的問題 [04:19.98]然後 [04:20.98]我們忘記 [04:21.98]我們在RL [04:22.98]是一個 [04:23.98]很不容易的 [04:24.98]方式 [04:25.98]我們為何 [04:26.98]我們在世界 [04:27.98]找到所有的 [04:28.98]知識 [04:29.98]我們在一年 [04:30.98]和一位 [04:31.98]伯克里斯教授 [04:32.98]教授 [04:33.98]我們會拿到 [04:34.98]AGI [04:35.98]他的觀點 [04:36.98]對 [04:37.98]他的理想 [04:38.98]對 [04:39.98]所以 [04:40.98]我們都在 [04:41.98]記錄 [04:42.98]我們會 [04:43.98]解決 [04:44.98]我們已經解決 [04:45.98]LM [04:46.98]我們已經解決 [04:47.98]我們已經解決 [04:48.98]我們已經解決 [04:49.98]我們已經解決 [04:50.98]我們已經解決 [04:51.98]我們已經解決 [04:52.98]我們已經解決 [04:53.98]我們已經解決 [04:54.98]我們已經解決 [04:55.98]我們已經解決 [04:56.98]我們已經解決 [04:57.98]我們已經解決 [04:58.98]我們已經解決 [04:59.98]我們已經解決 [05:00.98]我們已經解決 [05:01.98]我們已經解決 [05:02.98]每一句 [05:03.98]文字 [05:04.98]然後所有的圖案都會學習到模式 [05:07.94]然後你能夠合作任何的組織 [05:10.14]例如寫進、聲音、畫面、其他畫面、影片等等 [05:14.42]這些都是圖案的圖案,可以學習到這類的動作 [05:18.50]所以我希望我們能夠解決這件事 [05:20.10]然後我們回到當時的歷史 [05:22.74]我們如何跟我們一起學習這些圖案的學習 [05:27.06]這就是我們要去進行的進步 [05:28.62]我還要向大家提醒你多多的明年開放的故事 [05:31.30]我們再回到大陸的故事 [05:32.90]在你的個人網站,我愛的,因為是一個很好的個人的故事 [05:37.38]故事的內容,像你的歷史 [05:39.38]我需要更新,因為太老了 [05:42.38]但是你提及GPC2,你忘記了GPC1嗎?我認為你忘記了,對吧? [05:46.18]我其實不太記得,我記得在那邊,我記得在那邊 [05:50.70]對,《Canonical Story》是阿力的故事,他很擔心傳播者和傳播者 [05:58.74]傳播者和傳播者和傳播者的訊息 [06:01.38]對,你帶我們去… 拿我們傳播者和傳播者和傳播者的訊息 [06:03.66]GPC的歷史,你也知道,對你來說 [06:07.46]對我來說,歷史和GPC的歷史是一個很好的問題 [06:10.02]所以我認為《Canonical Story》的故事,GPC的歷史是在谷歌上,對吧? [06:14.30]因為那是關於傳播者的故事 [06:17.30]而我認為最驚訝的一件事,是… [06:21.26]這是一個成績,例如在谷歌設立,你跟你的最好的朋友寫文章,對吧? [06:26.26]好,所以在調查,我認為我的工作,當我當了學校的學長,是一個領導的領導人,對吧? [06:33.02]所以我真的有很好的朋友,我的工作是把人們的小數目和好幾個好意義,然後向他們進行完結的工作 [06:41.10]我的工作不是在提供一百萬個意義,然後沒有任何股份的資料 [06:45.54]然後當我的想法開始合作,然後我開始工作,我的工作是向他們扭動資料,向他們做好工作 [06:52.50]然後開始將一些不正確的工作拆除,對吧? [06:56.06]那股股份並沒有存在在我的時間在谷歌上 [06:59.34]如果他們有做好工作,他們會說: [07:02.06]"喂,你真棒,你懂這些東西的效果嗎?" [07:05.98]"這裡是所有的我們的TPUs,然後我認為他們會殺掉我們" [07:09.94]他肯定是想要的,他在2017年也說了一百萬公升的計劃 [07:13.18]對,所以我認為這回合是在關於GPT的故事,對嗎? [07:15.98]就是我正在跳舞歷史,對嗎? [07:18.38]但在GPT2之後,我們都很期待GPT2,我可以告訴你更多的故事 [07:22.50]這是我最後的一篇文章,我甚至真的受到觸碍了,所以我變成了研究研究研究員 [07:27.70]每天每天我們進行GPT3,我會醒來,然後感到緊張 [07:32.38]我感到緊張,因為...你只要看看Fax,對嗎? [07:35.54]Google有所有的帖子,Google有所有的人 who invented all of these underlying technologies [07:40.74]有一個人叫Noam,他很聰明,他已經做了這個討論,他想要一百萬的計劃模式 [07:46.54]我認為我們可能只是在做一些複雜的研究,對嗎?他有這個扣子,只有轉換模式,他可能會在我們之前進行的 [07:54.66]我心想,拜託,讓這個模式結束,對嗎? [07:57.90]然後,整個時間都變成了他們沒有得到股票的資金 [08:01.62]所以,我年紀中,我帶了Google的LM的活動,我當時是一名手機的,我變得很清楚為什麼,對嗎? [08:06.98]那時候,有一個東西叫做"Brain Credit Marketplace" [08:11.06]你記得Brain Credit Marketplace嗎? [08:13.26]沒有,我沒聽過這說法 [08:14.30]其實,你會問任何Google,就像一件事,對嗎? [08:18.58]對,有限定資訊,你必須有一個市場的市場,對嗎? [08:23.06]你可能,有些時候是貧富,有些時候是政治欺負 [08:27.34]你可能,所以,基本上,每個人都要給錢,對嗎? [08:30.10]如果你有錢,你必須買N-CHIPS,按照貿易和責任的方式 [08:33.74]如果你想做一個大職業,你可能有19、20個朋友不願意去工作 [08:38.86]如果這就是它們的效果 [08:40.74]它們很難得獲得 [08:42.14]當中的肺炎 [08:43.86]去學習這些東西 [08:44.98]而 Google 的團隊 [08:45.86]正在打架 [08:47.02]但我們只能打擊它們 [08:48.22]因為我們拿了大大的肺炎 [08:50.62]然後我們注射 [08:51.42]然後我認為 [08:52.30]這就像是一部分的故事 [08:53.54]像是一部分的歷史 [08:54.34]像是一部分的歷史 [08:55.62]像是一部分的歷史 [08:57.58]像是一部分的歷史 [08:58.90]我認為同樣的 [09:00.22]我認為一部分的 [09:01.02]三部分會成為 [09:01.90]一部分的歷史 [09:03.22]因為是一部分的 [09:04.18]一部分的成績 [09:05.62]對 [09:06.30]我覺得這部分的內容是如何的 [09:07.70]和影片也有關的 [09:09.02]在前一天的情況下 [09:10.06]我認為可能 [09:11.10]我認為是Jensen [09:11.90]不確定是誰 [09:12.86]把最近的照片 [09:13.90]給大家看過的 [09:15.26]他在第一張DGX的照片中 [09:17.66]我覺得Jensen 已經是 [09:19.06]一個完美的 [09:21.22]技術 [09:21.94]和精神的一切 [09:24.10]我對NVIDIA的尊敬有多大關注 [09:26.22]是不實際的 [09:26.94]但我會打開 [09:27.74]我給他們的需要 [09:29.46]讓他們構思一下 [09:30.30]或者 [09:31.34]你只要用任何NVIDIA給他們的東西 [09:33.70]所以我們很接近他們的工作 [09:35.38]我不確定能分享所有的故事 [09:37.62]但例子是我找到的 [09:39.42]特別有趣的 [09:40.14]所以 Scott Gray 是很棒的 [09:41.54]我很喜歡他 [09:42.22]他在我的隊伍中 [09:43.30]是一名超級電腦隊伍 [09:45.62]就是Chris Burner 做的 [09:46.74]Chris Burner 還做了很多東西 [09:48.82]結果 [09:49.70]我們有很接近NVIDIA的 ties [09:52.62]其實我的 co-founder [09:53.70]在Adept Eric Elson [09:54.74]是一位以前的GPGPU人士 [09:56.78]所以他和Scott [09:57.82]和Brian Kanzaro [09:58.86]NVIDIA [09:59.66] and Jonah [10:00.26] and Ian at NVIDIA [10:01.14]我覺得我們全都很接近 [10:02.54]我們是一部分的組織 [10:03.70]我們如何推動這些股票的限度 [10:05.82]我覺得那種組織 [10:07.42]幫助了我們 [10:08.38]我想有趣的部分 [10:09.50]是 knowing the A100 generation [10:11.22]那個Quadsbar city [10:12.26]會是一件事 [10:12.98]是我們想找到的 [10:14.50]來解決 [10:15.22]這是我們可以利用的 [10:16.50]模特兒訓練 [10:17.14] really what it boils down to [10:18.50]是 [10:19.22]我認為更多人 [10:20.06]知道這件事 [10:21.26]6 年前 [10:22.34]甚至3 年前 [10:23.34]人們拒絕接受 [10:24.98]這個AI 是一件故事 [10:27.02]是一件故事 [10:27.62]如何讓你更能复入 [10:29.22]實際使用模特兒 [10:30.38]使用模特兒 [10:31.66]還有GPT 2 3 故事嗎 [10:35.78]你喜歡在外面 [10:37.78]我認為是 [10:38.78]很欣賞 [10:39.86]這個模特兒的作用 [10:41.66]有趣的GPT 2 故事 [10:43.66]我花了很長的時間 [10:45.86]幫Alex使用模特兒 [10:48.58]我記得 [10:49.82]最有趣的一刻 [10:52.22]是我們寫了模特兒 [10:54.70]我確定模特兒 [10:56.22]是一個最短的模特兒 [10:57.70]有任何ML [10:58.70]像是最理想的 [10:59.90]ML 模特兒 [11:01.42]是三個模特兒 [11:03.18]這是一種模特兒 [11:04.54]Vanilla 模特兒 [11:05.58]只有轉換的模特兒 [11:06.38]這些特別的東西 [11:07.34]我記得是在《ParaGraph》裡 [11:08.58]我記得是在《ParaGraph》裡 [11:09.42]我們都在看這件事 [11:11.02]我認為是很難看的模特兒 [11:11.82]OGs 在廣場上 [11:13.02]會很討厭這個模特兒 [11:14.02]他們會說沒有創意 [11:15.50]為什麼你們要做這個作用 [11:16.94]現在是很有趣的 [11:18.02]在後期的看法是 [11:19.54]一件很刺激的作用 [11:20.82]但我覺得是一件很早的事 [11:22.54]我們完全遲到 [11:24.42]我們都要關心的問題是 AI 和不關的 [11:27.58]是否有四種不同的想法 [11:29.34]是否有一個很簡單的想法 [11:30.34]是否有一個很簡單的想法 [11:31.34]是否有一個很簡單的想法 [11:32.34]是否有一個很簡單的想法 [11:33.34]是否有一個很簡單的想法 [11:34.34]是否有一個很簡單的想法 [11:35.34]是否有一個很簡單的想法 [11:36.34]是否有一個很簡單的想法 [11:37.34]是否有一個很簡單的想法 [11:38.34]是否有一個很簡單的想法 [11:39.34]是否有一個很簡單的想法 [11:40.34]是否有一個很簡單的想法 [11:41.34]是否有一個很簡單的想法 [11:42.34]是否有一個很簡單的想法 [11:43.34]是否有一個很簡單的想法 [11:44.34]是否有一個很簡單的想法 [11:45.34]是否有一個很簡單的想法 [11:46.34]是否有一個很簡單的想法 [11:47.34]是否有一個很簡單的想法 [11:48.34]是否有一個很簡單的想法 [11:49.34]是否有一個很簡單的想法 [11:50.34]是否有一個很簡單的想法 [11:51.34]是否有一個很簡單的想法 [11:52.34]是否有一個很簡單的想法 [11:53.34]是否有一個很簡單的想法 [11:54.34]是否有一個很簡單的想法 [11:55.34]是否有一個很簡單的想法 [11:56.34]是否有一個很簡單的想法 [11:57.34]是否有一個很簡單的想法 [11:58.34]是否有一個很簡單的想法 [11:59.34]是否有一個很簡單的想法 [12:00.34]是否有一個很簡單的想法 [12:01.34]是否有一個很簡單的想法 [12:02.34]是否有一個很簡單的想法 [12:03.34]是否有一個很簡單的想法 [12:04.34]是否有一個很簡單的想法 [12:05.34]是否有一個很簡單的想法 [12:06.34]是否有一個很簡單的想法 [12:07.34]是否有一個很簡單的想法 [12:08.34]是否有一個很簡單的想法 [12:09.34]之前 Microsoft invested in OpenAI [12:11.34]Sam Altman, myself, and our CFO [12:13.34] flew up to Seattle [12:14.34] to do the final pitch meeting [12:16.34] and I'd been a founder before [12:17.34] so I always had a tremendous amount of anxiety [12:19.34] about partner meetings [12:21.34] which this basis is what it was [12:22.34] it was like Kevin Scott [12:23.34] and Satya and Amy Hood [12:25.34] and it was my job to give the technical slides [12:27.34] about what's the path to AGI [12:29.34] what's our research portfolio [12:30.34] all of this stuff [12:31.34] but it was also my job to give the GPT-2 demo [12:34.34] we had a slightly bigger version of GPT-2 [12:36.34] that we had just cut [12:38.34] maybe a day or two before this flight up [12:40.34] and as we all know now [12:42.34]Model behaviors you find predictable [12:44.34] at one checkpoint [12:45.34] are not predictable in another checkpoint [12:46.34] and so like I spent all this time [12:48.34] trying to figure out how to keep this thing on rails [12:50.34] I had my canned demos [12:51.34] but I knew I had to go [12:52.34] turn it around over to Satya and Kevin [12:54.34] and let them type anything in [12:56.34] and that just that really kept me up all night [12:58.34]Nice, yeah [13:00.34]I mean that must have helped you [13:01.34] talking about partners meeting [13:03.34]You raised 420 million for ADAPT [13:06.34]The last round was a $350 million series B [13:09.34]So I'm sure you do great [13:10.34]Pitching and painting [13:12.34]Nice [13:13.34]No, that's a high compliment coming from a VC [13:15.34]Yeah, I mean you're doing great [13:17.34]Let's talk about ADAPT [13:19.34]and we were doing pre prep [13:21.34]and you mentioned that maybe a lot of people [13:22.34]don't understand what ADAPT is [13:23.34]So usually we try and introduce the product [13:26.34]and then have the founders fill in the blanks [13:27.34]but maybe let's do the reverse [13:28.34]Like what is ADAPT? [13:30.34]Yeah, so I think ADAPT [13:31.34]is the least understood company [13:34.34]in the broader space of foundation models [13:36.34]plus agents [13:37.34]So I'll give some color [13:39.34]and I'll explain what it is [13:40.34]and I'll explain also [13:41.34]why it's actually pretty different [13:43.34]from what people would have guessed [13:44.34]So the goal for ADAPT [13:46.34]is we basically want to build an AI agent [13:48.34]that can do [13:49.34]that can basically help humans [13:50.34]do anything a human does on a computer [13:51.34]and so what that really means is [13:53.34]we want this thing to be super good [13:55.34]at turning natural language [13:56.34]like goal specifications [13:58.34]right into the correct set of end steps [14:00.34]and then also have all the correct sensors [14:02.34]and actuators [14:03.34]to go get that thing done for you [14:04.34]across any software tool [14:05.34]that you already use [14:06.34]and so the end vision of this [14:07.34]is effectively like [14:08.34]I think in a couple years [14:09.34]everyone's going to have access [14:10.34]to an AI teammate [14:11.34]that they can delegate arbitrary tasks to [14:14.34]and then also be able to use it [14:16.34]to a sounding board [14:17.34]and just be way, way, way more productive [14:19.34]right and just changes the shape [14:21.34]of every job [14:22.34]from something where you're mostly [14:23.34]doing execution [14:24.34]to something where you're mostly [14:25.34]actually doing these core liberal arts skills [14:26.34]of what should I be doing and why [14:28.34]right and [14:29.34]I find this like really exciting [14:31.34]motivating because [14:32.34]I think it's actually [14:33.34]pretty different vision [14:34.34]for how AI will play out [14:36.34]I think systems like ADAPT [14:37.34]are the most likely systems [14:38.34]to be proto-AGI's [14:40.34]but I think the ways in which [14:41.34]we are really counterintuitive [14:42.34]to everybody [14:43.34]is that [14:44.34]we've actually been really quiet [14:45.34]because we are [14:46.34]not a developer company [14:47.34]we don't sell APIs [14:48.34]we don't sell open source models [14:50.34]we also don't sell bottom-up products [14:52.34]we're not a thing [14:53.34]that you go and click [14:54.34]and download the extension [14:55.34]and like we want more users [14:56.34]signing up for that thing [14:57.34]we're actually an enterprise company [14:58.34]so what we do is [14:59.34]we work with a range [15:00.34]of different companies [15:01.34]some like late-stage [15:02.34]multi-thousand people start-ups [15:04.34]some Fortune 500s etc [15:06.34]and what we do for them [15:07.34]is we basically give them [15:09.34]an out-of-the-box solution [15:11.34]where big complex workflows [15:12.34]that their employees [15:13.34]do every day [15:14.34]could be delegated to the model [15:15.34]and so we look a little [15:16.34]different from other companies [15:17.34]in that in order [15:18.34]to go build this [15:19.34]full agent thing [15:20.34]the most important thing [15:21.34]you gotta get right [15:22.34]is reliability [15:23.34]so initially zooming [15:24.34]way back when [15:25.34]one of the first things [15:26.34]debt did was we released [15:27.34]this demo called Act 1 [15:28.34]act 1 was like pretty cool [15:30.34]it's kind of become [15:31.34]a hello world thing [15:32.34]for people to show [15:33.34]agent demos [15:34.34]by going to redfin [15:35.34]and asking to buy a house [15:36.34]somewhere [15:37.34]because like we did that [15:38.34]in the original Act 1 demo [15:39.34]and like showed that [15:40.34]showed like Google Sheets [15:41.34]all this other stuff [15:42.34]over the last like year [15:44.34]since that has come out [15:45.34]there's been a lot [15:46.34]of really cool demos [15:47.34]and you go play with them [15:48.34]and you realize [15:49.34]they work 60% of the time [15:50.34]but since we've always [15:51.34]been focused on [15:52.34]how do we build [15:53.34]an amazing enterprise product [15:54.34]enterprises can't use [15:55.34]anything [15:56.34]the reliability [15:57.34]and so we've [15:58.34]actually had to go down [15:59.34]a slightly different [16:00.34]tech tree than what you [16:01.34]might find in the [16:02.34]prompt engineering [16:03.34]sort of plays in [16:04.34]the agent space [16:05.34]to get that reliability [16:06.34]and we've decided [16:07.34]to prioritize reliability [16:08.34]over all else [16:09.34]so like one of our use [16:10.34]cases is crazy enough [16:11.34]that it actually ends [16:12.34]with a physical truck [16:13.34]being sentto a place [16:15.34]as the result [16:16.34]of the agent workflow [16:17.34]and if you're like [16:18.34]if that works like 60% [16:19.34]of the time [16:20.34]you're just blowing money [16:21.34]and poor truck drivers [16:22.34]going places [16:23.34]interesting [16:24.34]one of the [16:25.34]common teams [16:26.34]has this idea of services [16:27.34]as software [16:28.34]I'm actually giving a talk [16:29.34]at nvidia gtc [16:30.34]about this [16:31.34]but basically [16:32.34]software as a service [16:33.34]you're wrapping [16:34.34]user productivity [16:35.34]in software [16:36.34]with agents [16:37.34]and services as software [16:38.34]is replacing things [16:39.34]that you know [16:40.34]you would ask somebody [16:41.34]to do [16:42.34]and the software [16:43.34]just does it for you [16:44.34]when you think [16:45.34]about these usecases [16:46.34]do the users [16:47.34]still go in [16:48.34]and look at the agent [16:49.34]kindof like [16:50.34]doing the things [16:51.34]and can intervene [16:52.34]or likeare they slowly [16:53.34]remove from them [16:54.34]are there people [16:55.34]in the middle [16:56.34]checking in [16:57.34]I think there's two current flaws [16:58.34]in the framing [16:59.34]for services [17:00.34]as software [17:01.34]or I think what you just said [17:02.34]I think that one of them [17:03.34]is likein our experience [17:04.34]as we've been rolling [17:05.34]out adept [17:06.34]the people who actually [17:07.34]do the jobs [17:08.34]are the most excited [17:09.34]about it [17:10.34]because they don't go from [17:11.34]I do this job [17:12.34]to I don't do this job [17:13.34]they go from [17:14.34]I do this job [17:15.34]for everything [17:16.34]including the shitty [17:17.34]wrote stuff [17:18.34]to I'm a supervisor [17:19.34]and I literally [17:20.34]likeit's pretty magical [17:21.34]when you watch the thing [17:22.34]being used [17:23.34]sequentially by hand [17:24.34]as a human [17:25.34]and you can just click [17:26.34]in any one of them [17:27.34]be like hey I want to watch [17:28.34]the trajectory [17:29.34]the agent went through [17:30.34]to go solve this [17:31.34]and the nice thing [17:32.34]about agent execution [17:33.34]as opposed to [17:34.34]like LLM generations [17:35.34]is that [17:36.34]a good chunk of the time [17:37.34]when the agent [17:38.34]fails to execute [17:39.34]it doesn't give you [17:40.34]the wrong result [17:41.34]it just fails to execute [17:42.34]and the whole trajectory [17:43.34]is just broken and dead [17:44.34]and the agent knows it [17:45.34]right so then [17:46.34]those are the ones [17:47.34]that the human [17:48.34]then goes and solves [17:49.34]and so then they become [17:50.34]a troubleshooter [17:51.34]they work on the more [17:52.34]present piece [17:53.34]of it [17:54.34]that we found [17:55.34]is our strategy [17:56.34]as a company [17:57.34]is to always be [17:58.34]an augmentation company [17:59.34]and I think [18:01.34]one out of principle [18:02.34]that's something [18:03.34]we really care about [18:04.34]but two [18:05.34]actually if you're [18:06.34]framing yourself [18:07.34]as an augmentation [18:08.34]company [18:09.34]you're always going to [18:10.34]live in the world [18:11.34]where you're solving [18:12.34]tasks that are a little [18:13.34]too hard for what [18:14.34]the model can do today [18:15.34]and still needs a human [18:16.34]to provide oversight [18:17.34]provide clarifications [18:18.34]provide human feedback [18:19.34]and that's how you [18:20.34]build a data flywheel [18:21.34]smart as humans [18:22.34]how to solve [18:23.34]things models [18:24.34]can't do today [18:25.34]and so I actually [18:26.34]think that [18:27.34]being an augmentation [18:28.34]company [18:29.34]forces you to go [18:30.34]develop your core [18:31.34]AI capabilities [18:32.34]faster than someone [18:33.34]who's saying [18:34.34]ah okay [18:35.34]my job's like [18:36.34]deliver you [18:37.34]a lights off [18:38.34]solution for X [18:39.34]it's interesting [18:40.34]because we've seen [18:41.34]two parts [18:42.34]of the market [18:43.34]one is [18:44.34]we have one company [18:45.34]that does [18:46.34]agents for [18:47.34]sock analysts [18:48.34]people just [18:49.34]don't have them [18:50.34]which is [18:51.34]the augmentation product [18:52.34]and then you have [18:53.34]sweep.dev [18:54.34]any of these products [18:55.34]which they just [18:56.34]do the whole thing [18:57.34]I'm really curious [18:58.34]to see how that evolves [18:59.34]I agree that today [19:00.34]the reliability is [19:01.34]so important [19:02.34]in the enterprise [19:03.34]that they just [19:04.34]don't use [19:05.34]most of them [19:06.34]that's cool [19:07.34]but it's great [19:08.34]to hear the story [19:09.34]because I think [19:10.34]from the outside [19:11.34]people are like [19:12.34]oh that [19:13.34]they do act one [19:14.34]they do person on [19:15.34]they do foo you [19:16.34]they do all these [19:17.34]it's just the public stuff [19:18.34]it's just the public stuff [19:19.34]我們想要更多的客人來領導 [19:22.20]所以我們想要更多的客人來領導 [19:26.08]但我們希望我們會更多的客人來領導 [19:29.32]我們想要更多的客人來領導 [19:31.48]我們想要更多的客人來領導 [19:33.68]所以這次我們想要更多的客人來領導 [19:36.70]為什麼你變得更多的客人? [19:38.78]如果整個推動... [19:40.12]你已經領導了你的公司 [19:41.82]但是你也會更加努力去領導更多的客人來領導 [19:46.20]我覺得我們剛剛領導過那一步 [19:48.14]因為我最近還沒有領導過那一步 [19:49.14]這是一個好問題 [19:50.14]我認為這兩件事其實是很重要的 [19:51.14]一件事我認為是... [19:53.14]坦白說,大部分是公共的歷史 [19:56.14]在公司中的公司中的歷史是最重要的 [19:58.14]我非常高興這件事發生 [20:00.14]因為當我們開始公司在2022年代 [20:03.14]大家都在社會中知道歷史的歷史 [20:06.14]但公司中的歷史沒有任何意義 [20:08.14]他們還會把所有的歷史都放在桌上 [20:11.14]所以我認為現在 [20:13.14]我真的要注意的是 [20:15.14]當人們認為歷史 [20:16.14]他們會認為是對的 [20:17.14]對,所有各種各樣的東西都會被引起 [20:19.14]會被引起的電話電話電話電話 [20:20.14]會被引起的東西都會被引起的東西 [20:21.14]或是被引起的電話電話電話 [20:22.14]我認為電話電話電話 [20:23.14]是一個可以給你一個目標 [20:25.14]再次進行的工作 [20:27.14]並且在最少數個步驟中 [20:28.14]所以這就是一個大部分的原因 [20:30.14]我認為其中一個部分 [20:31.14]是因為我認為更好讓人們 [20:33.14]更加 aware of the depth [20:34.14]他們想要做的事情 [20:35.14]他們的生意 [20:36.14]這塊地是在世界中 [20:38.14]在於在更多的利益 [20:40.14]我認為大量的利益 [20:43.14]會發生從 [20:44.14]你使用的研究模式 [20:46.14]作為大量學童的學童 [20:49.14]去解決這些事 [20:50.14]我認為那些人 [20:51.14]想要做的研究 [20:52.14]應該有所改善 [20:53.14]當你提到 [20:54.14]研究已經變成 [20:55.14]更多的一部分 [20:56.14]有什麼特別的東西 [20:57.14]你會問我嗎 [20:58.14]我會給你一個名字 [20:59.14] Bill Gates 在 his blog post [21:00.14]提及「Agent of the Future」 [21:02.14]我是那個人 who made OSs [21:04.14]我認為「Agent of the Next Thing」 [21:05.14]所以 Bill Gates [21:07.14]我會叫他出來 [21:08.14]然後 Sam Altman 也會說 [21:09.14]「Agent of the Future for Open AI」 [21:10.14]我認為之前 [21:11.14]我認為 [21:12.14]有些人在《紐約 Times》 [21:13.14]Kade Metz 也在《紐約 Times》 [21:15.14]對於現在 [21:16.14]在一些不同的 [21:17.14]我看過 AI 開始的 [21:18.14]使用的研究模式 [21:19.14]是 AI 公司 [21:20.14]現在的 AI 公司 [21:21.14]是 AI 公司 [21:22.14]只是我認為 [21:23.14]是一段時間 [21:24.14]從 VC 開始 [21:25.14]是有點混合 [21:26.14]是嗎 [21:27.14]我認為有很多 VC [21:28.14]會說我不會 [21:29.14]觸碰 any agent start-ups [21:30.14]因為 [21:31.14]為什麼 [21:32.14]你告訴我 [21:33.14]我認為有很多 VC [21:35.14]比較少技術 [21:37.14]不懂得 [21:38.14]限制的東西 [21:39.14]不不不 [21:40.14]你會這樣嗎 [21:41.14]不不 [21:42.14]我認為 [21:43.14]今天的可能性 [21:44.14]是否適用 [21:46.14]我認為 [21:47.14]人們會看你 [21:48.14]然後說 [21:49.14]這傢伙 [21:50.14]需要 400 億元 [21:51.14]去做 [21:52.14]所以有很多 VC [21:53.14]都會說 [21:54.14]我會再加上 [21:55.14]有些東西 [21:56.14]協助 AI [21:57.14]有些東西 [21:58.14]是比較容易 [21:59.14]進行 [22:00.14]進行的 [22:01.14]但我還驚訝 [22:02.14]有些 funders [22:03.14]不想做 agent [22:04.14]不只是 funding [22:05.14]有時候 [22:06.14]我們在看 [22:07.14]為什麼沒有人 [22:08.14]做 agent for acts [22:09.14]那是好 [22:10.14]其實 [22:11.14]我從沒知道 [22:12.14]我的觀點 [22:13.14]是 [22:14.14]有新的 agent company [22:16.14]在進行 [22:17.14]所以可能 [22:18.14]他們也有 [22:19.14]但我提供人員 [22:20.14]去取消 agent [22:21.14]他們的名字 [22:22.14]是因為 [22:23.14]他們的名字 [22:24.14]他們的名字 [22:25.14]所以 [22:26.14]他們不等待 [22:27.14]對 [22:28.14]那是好處 [22:29.14]你的 portfolio allocator [22:31.14]有些人 [22:32.14]知道 about persimmon [22:33.14]一些人知道 [22:34.14]for you and for you heavy [22:35.14]你覺得 [22:36.14]怎麼想 [22:37.14]那個 evolution of that [22:38.14]什麼人 [22:39.14]想想 [22:40.14]那是 [22:41.14]a depth [22:42.14]搜尋個案 [22:43.14] kind of take us [22:44.14]through the stuff [22:45.14]you should recently [22:46.14]and how people [22:47.14]should think about [22:48.14]the trajectory [22:49.14]what you're doing [22:50.14]the critical path [22:51.14]for adept [22:52.14]is we want to build [22:53.14]agents that can do [22:54.14]a higher and higher [22:55.14]level of abstraction [22:56.14]things over time [22:57.14]all while keeping [22:58.14]insanely [22:59.14]high reliability standard [23:00.14]because that's [23:01.14]what turns this from [23:02.14]research into something [23:03.14]that customers want [23:04.14]and if you build [23:05.14]agents with really [23:06.14]high reliability standard [23:07.14]your users [23:08.14]how to get that [23:09.14]next level of [23:10.14]straction faster [23:11.14]so that's how [23:12.14]you actually build [23:13.14]the data level [23:14.14]that's the critical path [23:15.14]for the company [23:16.14]everything we do [23:17.14]is in service of that [23:18.14]so you go zoom [23:19.14]way way back to [23:20.14]act one days right [23:21.14]like the core thing [23:22.14]behind act one [23:23.14]is can we teach [23:24.14]large model basically [23:25.14]how to even [23:26.14]actuate your computer [23:27.14]and I think we're [23:28.14]one of the first places [23:29.14]to have solved that [23:30.14]and shown it [23:31.14]and shown the generalization [23:32.14]that you get when you [23:33.14]give it various different [23:34.14]workflows and texts [23:35.14]but I think from [23:36.14]these models [23:37.14]to be able to [23:38.14]get a lot better [23:39.14]at having some [23:40.14]specificationof some [23:41.14]guardrails for what it [23:42.14]actually should be doing [23:43.14]and I think in conjunction [23:44.14]with that a giant thing [23:45.14]that was really [23:46.14]necessaryis really [23:47.14]fast multimodal models [23:48.14]that are really good [23:49.14]at understanding [23:50.14]knowledge work [23:51.14]and really good [23:52.14]at understanding screens [23:53.14]and that needs to [23:54.14]kind of be the base [23:55.14]for some of these [23:56.14]agentsback then [23:57.14]we had to do a ton [23:58.14]ofresearchbasically [23:59.14]on how do we [24:00.14]actually make that [24:01.14]possiblewell first off [24:02.14]back in [24:03.14]free at exact [24:04.14]one month of 23 [24:05.14]and then [24:06.14]we had to [24:07.14]get a lot better [24:08.14]at the first place [24:09.14]and then [24:10.14]we had to [24:11.14]get a lot better [24:12.14]at the first place [24:13.14]and then [24:14.14]we had to [24:15.14]get a lot better [24:16.14]at the first place [24:17.14]and then [24:18.14]we had to [24:19.14]get a lot better [24:20.14]at the first place [24:21.14]and then [24:22.14]we had to [24:23.14]get a lot better [24:24.14]at the first place [24:25.14]and then [24:26.14]we had to [24:27.14]get a lot better [24:28.14]at the first place [24:29.14]and then [24:30.14]we had to [24:31.14]get a lot better [24:32.14]at the first place [24:33.14]and then [24:34.14]we had to [24:35.14]get a lot better [24:36.14]at the first place [24:37.14]and then [24:38.14]we had to [24:39.14]get a lot better [24:40.14]at the first place [24:41.14]and then [24:42.14]we had to [24:43.14]get a lot better [24:44.14]at the first place [24:45.14]and then [24:46.14]we had to [24:47.14]get a lot better [24:48.14]at the first place [24:49.14]and then [24:50.14]we had to [24:51.14]get a lot better [24:52.14]at the first place [24:53.14]and then [24:54.14]we had to [24:55.14]get a lot better [24:56.14]at the first place [24:57.14]and then [24:58.14]we had to [24:59.14]get a lot better [25:00.14]at the first place [25:01.14]and then [25:02.14]we had to [25:03.14]get a lot better [25:04.12]at the first place [25:05.12]and then [25:06.12]we had to [25:07.12]get a lot better [25:08.12]at the first place [25:09.12]and then [25:10.12]we had to [25:11.12]get a lot better [25:12.12]at the first place [25:13.12]and then [25:14.12]we had to [25:15.12]get a lot better [25:16.12]at the first place [25:17.12]and then [25:18.12]we had to [25:19.12]get a lot better [25:20.12]at the first place [25:21.12]and then [25:22.12]we had to [25:23.12]get a lot better [25:24.12]at the first place [25:25.12]and then [25:26.12]we had to [25:27.12]get a lot better [25:28.12]at the first place [25:29.12]and then [25:30.12]we had to [25:31.12]get a lot better [25:32.12]at the first place [25:33.12]and then [25:34.12]we had to [25:35.12]get a lot better [25:36.12]at the first place [25:37.12]and then [25:38.12]we had to [25:39.12]get a lot better [25:40.12]at the first place [25:41.12]and then [25:42.12]we had to [25:43.12]get a lot better [25:44.12]at the first place [25:45.12]and then [25:46.12]we had to [25:47.12]get a lot better [25:48.12]at the first place [25:49.12]and then [25:50.12]we had to [25:51.12]get a lot better [25:52.12]at the first place [25:53.12]and then [25:54.12]we had to [25:55.12]get a lot better [25:56.12]at the first place [25:57.12]and then [25:58.12]we had to [25:59.12]get a lot better [26:00.12]at the first place [26:01.12]and then [26:02.12]we had to [26:03.12]get a lot better [26:04.12]at the first place [26:05.12]and then [26:06.12]we had to [26:07.12]get a lot better [26:08.12]at the first place [26:09.12]and then [26:10.12]we had to [26:11.12]get a lot better [26:12.12]at the first place [26:13.12]and then [26:14.12]we had to [26:15.12]get a lot better [26:16.12]at the first place [26:17.12]and then [26:18.12]we had to [26:19.12]get a lot better [26:20.12]at the first place [26:21.12]and then [26:22.12]we had to [26:23.12]get a lot better [26:24.12]at the first place [26:25.12]and then [26:26.12]we had to [26:27.12]get a lot better [26:28.12]at the first place [26:29.12]and then [26:30.12]we had to [26:31.12]get a lot better [26:32.12]at the browser level [26:33.12]I really want [26:34.12]at your papers [26:35.12]you have like a different representation [26:36.12]kind of like [26:37.12]you don't just take the dome [26:38.12]and act on it [26:39.12]you do a lot more stuff [26:40.12]how do you think about [26:41.12]the best way [26:42.12]the models will interact [26:43.12]with the software [26:44.12]and like how [26:45.12]the development of products [26:46.12]is going to change [26:47.12]with that in mind [26:48.12]as more and more [26:49.12]the work is done by agents [26:50.12]instead of people [26:51.12]this is [26:52.12]there's so much surface area here [26:53.12]and it's actually one of the things [26:54.12]I'm really excited about [26:55.12]and it's funny because [26:56.12]I've spent most of my time [26:57.12]doing research stuff [26:58.12]but this is like a whole [26:59.12]new ball game that I've been [27:00.12]doing about [27:01.12]and I find it [27:02.12]really cool [27:03.12]so I would say [27:04.12]the best analogy [27:05.12]I have to [27:06.12]why ADAPT [27:07.12]is pursuing a path [27:08.12]of being able to [27:09.12]use your computer [27:10.12]like a human [27:11.12]plus of course [27:12.12]being able to call [27:13.12]APIs [27:14.12]being able to call [27:15.12]APIs is the easy part [27:16.12]like being able to [27:17.12]use your gear like humans [27:18.12]is a hard part [27:19.12]it's in the same way [27:20.12]why people are excited [27:21.12]about humanoid robotics [27:22.12]right [27:23.12]in a world where [27:24.12]you had t=infinity [27:25.12]right you're probably [27:26.12]gonna have various [27:27.12]different form factors [27:28.12]that robots [27:29.12]do [27:30.12]without changing [27:31.12]everything along the way [27:32.12]it's the same thing [27:33.12]for software [27:34.12]right [27:35.12]if you go itemize out [27:36.12]the number of things [27:37.12]you wanna do on your computer [27:38.12]for which every step [27:39.12]has an api [27:40.12]those numbers [27:41.12]will workflows add up [27:42.12]pretty close to zero [27:43.12]and so then many [27:44.12]points along the way [27:45.12]you need the ability [27:46.12]to actually control [27:47.12]your computer like a human [27:48.12]it also lets you learn [27:49.12]from human usage [27:50.12]of computers [27:51.12]as a source of training [27:52.12]data that you don't get [27:53.12]if you have to somehow [27:54.12]figure out how every [27:55.12]particular step needs to be [27:56.12]some particular custom [27:57.12]private api thing [27:58.12]it's the most practical path [27:59.12]i think a lot of [28:00.12]success will come [28:01.12]from going down [28:02.12]this path [28:03.12]i kinda think about this [28:04.12]early days of the agent [28:05.12]interaction layer [28:06.12]level is a little bit [28:07.12]like do y'all remember [28:08.12]windows 3.1 [28:10.12]like those days [28:11.12]this might be [28:12.12]i might be too old [28:13.12]for you guys on this [28:14.12]but back in the day [28:15.12]windows 3.1 [28:16.12]we had this transition period [28:17.12]between pure command line [28:18.12]right [28:19.12]being the default [28:20.12]into this new world [28:21.12]with the gui is the default [28:22.12]and then you drop into the [28:23.12]command line for like [28:24.12]programmer things [28:25.12]the old way was [28:26.12]you booted your computer up [28:27.12]and then it would [28:28.12]give you the c colon [28:29.12]slash thing [28:30.12]and you typed windows [28:31.12]and you hit enter [28:32.12]and then you got [28:33.12]put into windows [28:34.12]and then the gui [28:35.12]kind of became a layer [28:36.12]above the command line [28:37.12]the same thing [28:38.12]is gonna happen [28:39.12]with agent interfaces [28:40.12]is like today [28:41.12]what we have in the gui [28:42.12]is like the base layer [28:44.12]and then the agent [28:45.12]just controls [28:46.12]the current gui [28:47.12]layer plus apis [28:48.12]and in the future [28:50.12]as more and more [28:51.12]trust is built towards [28:52.12]agents and more and more [28:53.12]things can be done by [28:54.12]agents and more UIs [28:55.12]for agents are actually [28:56.12]users [28:57.12]then that just becomes [28:58.12]a standard [28:59.12]interaction layer [29:00.12]and if that becomes [29:01.12]a standard [29:02.12]interaction layer [29:03.12]what changes for [29:04.12]software is that [29:05.12]a lot of software [29:06.12]is gonna be [29:07.12]either systems [29:08.12]or record [29:09.12]or like certain [29:10.12]customized [29:11.12]workflow [29:12.12]execution engines [29:13.12]and a lot of [29:14.12]how you actually [29:15.12]do stuff will be [29:16.12]controlled at the [29:17.12]agent layer [29:18.12]and you think the [29:19.12]rabbit interface [29:20.12]is more like [29:21.12]it would like [29:22.12]you're not actually [29:23.12]seeing the app [29:24.12]that the model [29:25.12]I can see that [29:26.12]being a model [29:27.12]I think [29:28.12]I don't know [29:29.12]enough about [29:30.12]what using [29:31.12]rabbit in real life [29:32.12]will actually be like [29:33.12]to comment on [29:34.12]that particular [29:35.12]thing but I think [29:36.12]the broader idea [29:37.12]that you know [29:38.12]you have a goal [29:39.12]the agent knows [29:40.12]how to break [29:41.12]your goal down into steps [29:42.12]the agent knows [29:43.12]how to use [29:44.12]the underlying [29:45.12]software [29:46.12]and systems [29:47.12]or record [29:48.12]to achieve [29:49.12]that goal for you [29:50.12]the agent may presents [29:51.12]you information [29:52.12]in a custom way [29:53.12]that's only [29:54.12]you're a power [29:55.12]user [29:56.12]for some niche thing [29:57.12]general question [29:58.12]so first of all [29:59.12]I think like [30:00.12]the sort of input [30:01.12]mode conversation [30:02.12]I wonder if you have [30:03.12]any analogies [30:04.12]that you like [30:05.12]with self-driving [30:06.12]because I do think [30:07.12]there's a little bit [30:08.12]of how the model [30:09.12]should perceive the world [30:10.12]and you know [30:11.12]the primary split [30:12.12]in self-driving [30:13.12]is LiDAR [30:14.12]versus camera [30:15.12]and I feel like [30:16.12]most agent companies [30:17.12]that I'm tracking [30:18.12]are all moving towards [30:19.12]camera approach [30:20.12]which is like [30:21.12]the multimodal approach [30:22.12]that we're doing [30:23.12]you're [30:24.12]focusing on that [30:25.12]including charts [30:26.12]and tables [30:27.12]and do you find [30:28.12]inspiration there [30:29.12]from the self-driving [30:30.12]world? [30:31.12]that's a good question [30:32.12]I think sometimes [30:33.12]the most useful [30:34.12]inspiration I've found [30:35.12]from self-driving [30:36.12]is the levels analogy [30:37.12]I think that's awesome [30:38.12]but I think that [30:39.12]our number one [30:40.12]goals for agents [30:41.12]not to look like [30:42.12]self-driving [30:43.12]we want to minimize [30:44.12]the chances [30:45.12]that agents are sort [30:46.12]of a thing [30:47.12]that you just [30:48.12]have to bang [30:49.12]your head at [30:50.12]for a long time [30:51.12]to get to like [30:52.12]completely [30:53.12]and that takes you [30:54.12]all the way [30:55.12]up to the top [30:56.12]but similarly [30:57.12]I mean [30:58.12]compared to self-driving [30:59.12]like two things [31:00.12]that people really [31:01.12]undervalue [31:02.12]that's like really [31:03.12]easy to driving [31:04.12]a car down [31:05.12]highway 101 [31:06.12]in a sunny day [31:07.12]demo [31:08.12]that actually [31:09.12]doesn't prove anything [31:10.12]anymore [31:11.12]and I think [31:12.12]the second thing [31:13.12]is that [31:14.12]as a non-self-driving [31:15.12]expert [31:16.12]I think one of the things [31:17.12]that we believe [31:18.12]really strongly [31:19.12]is that [31:20.12]everyone under [31:21.12]get a lot [31:22.12]of reliability [31:23.12]is a really [31:24.12]strong focus on [31:25.12]actually why [31:26.12]does the model [31:27.12]not do this thing [31:28.12]and the non-trivial amount [31:29.12]of time [31:30.12]the time the model [31:31.12]doesn't actually [31:32.12]do the thing [31:33.12]is because if [31:34.12]you're a wizard [31:35.12]of ozing it yourself [31:36.12]or if you have [31:37.12]unreliable actuators [31:38.12]you can't do the thing [31:39.12]and so we've [31:40.12]had to fix [31:41.12]a lot of those problems [31:42.12]I was slightly [31:43.12]surprised just because [31:44.12]I do generally [31:45.12]consider the way [31:46.12]most that we see [31:47.12]all around San Francisco [31:48.12]as the most [31:49.12]I guess real case [31:50.12]it's a big [31:51.12]job but it has taken [31:52.12]a long time [31:53.12]for self-driving [31:54.12]temperature from [31:55.12]when it entered [31:56.12]the consciousness [31:57.12]and the driving down [31:58.12]when it went on a sunny [31:59.12]day moment [32:00.12]happened to now. [32:01.12]so I want to see [32:02.12]the more compressed [32:03.12]cruise, you know, [32:04.12]R.I.P. [32:05.12]recently. [32:06.12]and then one more thing [32:07.12]on just like [32:08.12]just going back on [32:09.12]this reliability [32:10.12]thing, something [32:11.12]I have been holding [32:12.12]in my head [32:13.12]that I'm curious [32:14.12]to get your commentary on [32:15.12]is I think there's a [32:16.12]treatup between [32:17.12]reliability and generality [32:18.12]or I want to broaden [32:19.12]because you have [32:20.12]reliability also have [32:21.12]cost of speed [32:22.12]speed is a huge emphasis [32:23.12]for a debt [32:24.12]the tendency or the [32:25.12]attemptation is to reduce [32:26.12]generalityto improve [32:27.12]reliability [32:28.12]and to improve [32:29.12]cost improve speed [32:30.12]do you perceive a tradeoff [32:31.12]do you have any [32:32.12]insights that [32:33.12]solve those tradeoffs [32:34.12]for you guys [32:35.12]there's definitely a tradeoff [32:36.12]if you're at [32:37.12]the predo frontier [32:38.12]I think a lot of folks [32:39.12]aren't actually [32:40.12]at the predo frontier [32:41.12]I think the way you get [32:42.12]there is basically [32:43.12]how do you frame [32:44.12]the fundamental [32:45.12]agent problem in a way [32:46.12]that just continues [32:47.12]to benefit from data [32:48.12]I think one of [32:49.12]the main ways [32:50.12]of being able to solve [32:51.12]that particular tradeoff [32:52.12]is you basically [32:53.12]just want to formulate [32:54.12]the problem such that [32:55.12]every particular use [32:56.12]case just looks like [32:57.12]you collecting more [32:58.12]data to go make [32:59.12]that use case possible [33:00.12]I think that's how [33:01.12]you really solve it [33:02.12]then you get into the [33:03.12]other problems like [33:04.12]are you overfitting [33:05.12]on these end use cases [33:06.12]right but like you're [33:07.12]not doing a thing [33:08.12]where you're like [33:09.12]being super prescriptive [33:10.12]for the end steps [33:11.12]that the model can [33:12.12]only do for example [33:13.12]then the question becomes [33:14.12]kind of do you have [33:15.12]one sort of house model [33:16.12]they customize [33:17.12]the customer's [33:18.12]specific use case [33:19.12]we're not sharing [33:20.12]we're not sharing [33:21.12]it's tempting [33:22.12]but that doesn't [33:23.12]look like AGI to me [33:24.12]you know what I mean [33:25.12]that is just [33:26.12]you have a good [33:27.12]base model [33:28.12]and then [33:29.12]you fine tune it [33:30.12]for what it's worth [33:31.12]I think there's [33:32.12]two paths [33:33.12]to a lot more [33:34.12]capability coming out [33:35.12]of the models [33:36.12]that we [33:37.12]all are training [33:38.12]these days [33:39.12]one path [33:40.12]is you figure out [33:41.12]how to spend [33:42.12]compute and turn [33:43.12]into data [33:44.12]and so in that [33:45.12]path I consider [33:46.12]off play [33:47.12]all that stuff [33:48.12]the second path [33:49.12]is how do you [33:50.12]get super [33:52.12]competent [33:53.12]high intelligence [33:54.12]demonstrations [33:55.12]from humans [33:56.12]and I think [33:57.12]the right way [33:58.12]to move forward [33:59.12]is you kind of [34:00.12]want to combine the two [34:01.12]the first one [34:02.12]gives you maximum [34:03.12]sample efficiency [34:04.12]for the second [34:05.12]but I think [34:06.12]that is going to be [34:07.12]hard to be running [34:08.12]at max speed [34:09.12]towards AGI [34:10.12]without actually [34:11.12]solving a bit of both [34:12.12]you haven't talked [34:13.12]much about synthetic [34:14.12]data as far as I can [34:15.12]any insights [34:16.12]on using synthetic [34:17.12]data to augment [34:18.12]the expensive [34:19.12]human data [34:20.12]the best part [34:21.12]about framing AGI [34:22.12]is being able [34:23.12]to help people do [34:24.12]things on computers [34:25.12]is you have an environment [34:26.12]yes [34:27.12]so you can [34:28.12]simulate all of it [34:29.12]you can do a lot [34:30.12]of stuff [34:31.12]when you have an environment [34:32.12]we were having dinner [34:33.12]for our one year [34:34.12]anniversary [34:35.12]the other round [34:36.12]thank you [34:37.12]Raza from human [34:38.12]loop was there [34:39.12]and we mentioned [34:40.12]you were coming on [34:41.12]the pod [34:42.12]this is our first [34:43.12]so he submitted a question [34:44.12]now you had [34:45.12]gbd4 vision [34:46.12]and help you [34:47.12]building a lot [34:48.12]of those things [34:49.12]how do you think [34:50.12]about the things [34:51.12]that are unique to you [34:52.12]as a depth [34:53.12]and like going back [34:54.12]to like the maybe [34:55.12]research direction [34:56.12]that you want to take [34:57.12]the team and what you [34:58.12]want people to come [34:59.12]work on at a depth [35:00.12]versus what is maybe [35:01.12]not become commoditized [35:02.12]that you didn't expect [35:03.12]everybody would [35:04.12]have access to [35:05.12]yeah that's [35:06.12]a really good question [35:07.12]I think implicit [35:08.12]in that question [35:09.12]and I wish he were [35:10.12]tier two so he can [35:11.12]push back on my [35:12.12]assumption about his [35:13.12]questionbut I think [35:14.04]is calculus of where [35:16.04]does advantage a crew [35:18.04]in the overall [35:19.04]ML stack [35:20.04]and maybe part [35:21.04]of the assumption [35:22.04]is that advantage [35:23.04]a crew is solely [35:24.04]to base model scaling [35:25.04]but I actually [35:26.04]believe pretty strongly [35:27.04]that the way [35:28.04]that you really [35:29.04]win is that you [35:30.04]have to go build [35:31.04]an agent stack [35:32.04]that is much more [35:33.04]than that [35:34.04]of the base model itself [35:35.04]and so I think [35:36.04]like that is [35:37.04]always going to be [35:38.04]a giant advantage [35:39.04]of vertical integration [35:40.04]I think like [35:41.04]it lets us do things [35:42.04]like have a really [35:43.04]bad cat and dog [35:44.04]photo [35:45.04]it's pretty good [35:46.04]at cat and dog [35:47.04]photo [35:48.04]it's not like [35:49.04]soda at cat [35:50.04]and dogphoto [35:51.04]so like we're allocating [35:52.04]our capacity wisely [35:53.04]is like one thing [35:54.04]that you [35:55.04]really get to do [35:56.04]I also think that [35:57.04]the other thing [35:58.04]that is pretty [35:59.04]important now [36:00.04]in the broader [36:01.04]foundation modeling [36:02.04]space is [36:03.04]I feel despite any [36:04.04]potential concerns [36:05.04]about how good [36:06.04]is agents as [36:07.04]like a startup area [36:08.04]like we were talking [36:09.04]about earlier [36:10.04]I feel super good [36:11.04]that we're [36:12.04]cap just flowing [36:13.04]from can we make [36:14.04]a better agent [36:15.04]because right now [36:16.04]I think we all see [36:17.04]that you know [36:18.04]if you're training [36:19.04]on publicly available [36:20.04]web data [36:21.04]you put in the [36:22.04]flops and you do [36:23.04]reasonable things [36:24.04]then you get [36:25.04]decent results [36:26.04]and if you just [36:27.04]double the amount [36:28.04]of compute [36:29.04]then you get [36:30.04]predictably [36:31.04]better results [36:32.04]and so I think [36:33.04]pure play foundation [36:34.04]model companies [36:35.04]are just going to be [36:36.04]pinched by how [36:37.04]good the next couple [36:38.04]lamas are going to be [36:39.04]and the next [36:40.04]what good open source [36:41.04]on these base foundation [36:42.04]models I think it's [36:43.04]gonna commoditize a lot [36:44.04]of the regular llms [36:45.04]and soon regular [36:46.04]multimodal models [36:47.04]so I feel really good [36:48.04]that we're just focused [36:49.04]on agents so you [36:50.04]don't consider yourself [36:51.04]a pure play foundation [36:52.04]model company no [36:53.04]because if we were pure [36:54.04]play foundation model [36:55.04]company we would be [36:56.04]traininggeneral foundation [36:57.04]models that do [36:58.04]summarization and [36:59.04]all this dedicated [37:00.04]towards the agent [37:01.04]yeah and our business [37:02.04]is an agent business [37:03.04]we're not here to [37:04.04]sell you tokens right [37:05.04]and I think like [37:06.04]selling tokens unless [37:07.04]there's like yeah I [37:08.04]love it there's like [37:09.04]if you have a particular [37:10.04]area of specialty [37:11.04]right then you won't [37:13.04]get caught in the fact [37:14.04]that everyone's just [37:15.04]scaling to ridiculous [37:16.04]levels of compute [37:17.04]but if you don't have a [37:18.04]specialty I find that [37:19.04]I think it's gonna be [37:20.04]a little tougher [37:21.04]interesting are you [37:22.04]interested in robotics at [37:23.04]all just a personally [37:24.04]fascinated by robotics [37:25.04]have always loved robotics [37:26.04]embodied agents as a [37:27.04]business you know figure [37:28.04]is like a big also [37:29.04]so the open ai [37:30.04]affiliated company [37:31.04]that raises a lot of [37:32.04]money I think it's [37:33.04]cool I think I mean [37:34.04]I don't know exactly [37:35.04]what they're exactly [37:36.04]what they're doing but [37:37.04]robots yeah yeah [37:38.04]well I mean that's [37:39.04]well Christian [37:40.04]would you ask [37:41.04]like if we [37:42.04]had them on like [37:43.04]what would you ask them [37:44.04]oh I just want to [37:45.04]understand what their [37:46.04]overall strategy is [37:47.04]gonna be between now [37:48.04]and when there's reliable [37:49.04]stuff to be deployed [37:50.04]but honestly [37:51.04]I just don't know [37:52.04]enough about it [37:53.04]and if I told you [37:54.04]hey fire your entire [37:55.04]warehouse workforce [37:56.04]and you know [37:57.04]put robots in there [37:58.04]isn't that a strategy [37:59.04]oh yeah yeah sorry [38:00.04]I'm not questioning [38:01.04]whether [38:02.04]they're doing smart [38:03.04]things I genuinely [38:04.04]don't know what [38:05.04]they're doing as much [38:06.04]but I think there's [38:07.04]two things one [38:08.04]it's just [38:09.04]I think it's [38:10.04]just gonna work [38:11.04]like I will die [38:12.04]on this hill [38:13.04]like I mean [38:14.04]like again this whole [38:15.04]this whole time [38:16.04]like we've been on this [38:17.04]podcast it's just [38:18.04]gonna continually saying [38:19.04]these models [38:20.04]are basically behavioral [38:21.04]cloners right [38:22.04]so let's go behavioral [38:23.04]clone all this like [38:24.04]robot behavior right [38:25.04]and then [38:26.04]now you figure out [38:27.04]everything else [38:28.04]you have to do in order [38:29.04]to teach you how to [38:30.04]solve new problem [38:31.04]that's gonna work [38:32.04]I'm super stoked for that [38:33.04]I think unlike [38:34.04]what we're doing with [38:35.04]helping humans with [38:36.04]knowledge work [38:37.04]and I'm personally [38:38.04]less excited about that [38:39.04]we had a [38:40.04]canjun from imbu [38:41.04]on the podcast [38:42.04]we asked her [38:43.04]why people should [38:44.04]go work there [38:45.04]and not at adept [38:46.04]so I wanna [38:47.04]well she said [38:48.04]you know [38:49.04]there's space for everybody [38:50.04]in this market [38:51.04]we're all doing [38:52.04]interesting work [38:53.04]and she said [38:54.04]they're really excited [38:55.04]about building [38:56.04]an operating system [38:57.04]for agent [38:58.04]and for her [38:59.04]the biggest research [39:00.04]thing was like [39:01.04]getting models [39:02.04]better reasoning [39:03.04]and planning [39:04.04]for these agents [39:05.04]the reverse question [39:06.04]I'm excited to [39:07.04]come work at adept [39:08.04]instead of imbu [39:09.04]and maybe [39:10.04]what are like [39:11.04]the core research [39:12.04]questions [39:13.04]that people should [39:14.04]be passionate about [39:15.04]to have fun at adept [39:16.04]yeah first off [39:17.04]I think that [39:18.04]I'm sure you guys [39:19.04]believe this too [39:20.04]the AI space [39:21.04]to the center [39:22.04]there's an AI space [39:23.04]and the AI agent [39:24.04]space are both [39:25.04]exactly as [39:26.04]she likely said [39:27.04]I think colossal [39:28.04]opportunities [39:29.04]and people are just [39:30.04]going to end up [39:31.04]winning in different [39:32.04]areas and a lot [39:33.04]of companies are [39:34.04]going to do well [39:35.04]to be at [39:36.04]adept [39:37.04]I think there's [39:38.04]two huge reasons [39:39.04]to be at adept [39:40.04]I think one of them [39:41.04]is everything we do [39:42.04]is in the service [39:43.04]of like useful agents [39:44.04]we're not a [39:45.04]research lab [39:46.04]we do a lot of research [39:47.04]in service of that goal [39:48.04]but we don't [39:49.04]think about ourselves [39:50.04]as like a classic [39:51.04]research lab at all [39:52.04]and I think the second [39:53.04]reason at work at [39:54.04]adeptis [39:55.04]if you believe that [39:56.04]actually having customers [39:57.04]and a reward signal [39:58.04]from customers [39:59.04]lets you build [40:00.04]AGI faster [40:01.04]which we really believe [40:02.04]then you should come here [40:03.04]and I think the examples [40:04.04]are evaluations [40:05.04]they're not [40:06.04]academic evals [40:07.04]they're not simulator [40:08.04]evals [40:09.04]they're like [40:10.04]okay we have a [40:11.04]customer that [40:12.04]really needs us to do [40:13.04]these particular things [40:14.04]we can do some [40:15.04]of them [40:16.04]these other ones [40:17.04]they want us to [40:18.04]we can't do them at [40:19.04]all we've turned [40:20.04]those into evals [40:21.04]solve it [40:22.04]I think that's [40:23.04]really cool [40:24.04]like everybody knows [40:25.04]a lot of these evals [40:26.04]are like [40:27.04]pretty saturated [40:28.04]and the new ones [40:29.04]that even are [40:30.04]not saturated you look [40:31.04]at someone and you're [40:32.04]like is this actually [40:33.04]and all of this stuff [40:34.04]but they're very grounded [40:35.04]and actual needs [40:36.04]right now [40:37.04]which is really cool [40:38.04]yeah this has been [40:39.04]wonderful dive [40:40.04]I wish we had more time [40:41.04]but I'll just leave it [40:42.04]kind of open to you [40:43.04]I think you have broad thoughts [40:44.04]you know just about [40:45.04]the agent space [40:46.04]but also just general AI [40:47.04]space any sort of rants [40:48.04]or things that [40:49.04]they're just helping [40:50.04]might for you right now [40:51.04]any rants [40:52.04]minding you [40:53.04]for just general [40:54.04]wow okay [40:55.04]so Amelia's already [40:56.04]made the rant better [40:57.04]than I have [40:58.04]but not just [40:59.04]not just chatbots [41:00.04]is like kind of rant one [41:01.04]but the rant two [41:02.04]is AI's really been [41:03.04]the story of compute [41:04.04]and compute plus data [41:06.04]and ways in which [41:07.04]you could change one [41:08.04]for the other [41:09.04]and I think as much as [41:10.04]our research community [41:11.04]is really smart [41:12.04]we have made many [41:13.04]many advancements [41:14.04]and that's going to [41:15.04]continue to be important [41:16.04]but now I think [41:17.04]the game is [41:18.04]increasingly changing [41:19.04]and the rapid [41:20.04]industrialization [41:21.04]error has begun [41:22.04]and I think [41:23.04]we unfortunately [41:24.04]have to embrace it [41:25.04]excellent awesome David [41:26.04]thank you so much [41:27.04]for your time [41:28.04]cool yeah thanks guys [41:29.04]this was fun [41:30.04]thank you [41:31.04]thank you [41:32.04]thank you [41:32.04]thank you [41:33.04]thank you [41:34.04]thank you [41:35.04]thank you [41:36.04]thank you [41:37.04]thank you [41:38.04]thank you [41:39.04]thank you [41:40.04]thank you [41:41.04]thank you [41:42.04]thank you [41:43.04]thank you [41:44.04]thank you [41:45.04]thank you [41:46.04]thank you [41:47.04]字幕by索兰娅 [41:49.04]字幕:J Chong [41:50.04]请不吝点赞 订阅 转发 打赏 打赏