Kazuki
@kazuki
Cofounder of Glasp. I collect ideas and stories worth sharing 📚
San Francisco, CA
Joined Oct 9, 2020
1068
Following
5595
Followers
1.44k
13.37k
164.98k
www.vox.com/even-better/23744304/how-much-social-interaction-do-you-need-loneliness-burnout
Jun 19, 2023
143
jamesclear.com/checklist-solutions
Jun 17, 2023
41
jamesclear.com/creative-thinking
Jun 16, 2023
183
jayacunzo.com/blog/best-quote-on-creativity-ira-glass-gap
Jun 14, 2023
63
magazine.wharton.upenn.edu/issues/spring-summer-2022/the-power-of-the-underdog/
Jun 13, 2023
3
openai.com/blog/function-calling-and-other-api-updates
Jun 13, 2023
5
collabfund.com/blog/paying-attention/
Jun 12, 2023
82
www.forbes.com/sites/katevitasek/2022/05/17/knowledge-is-powerand-why-you-should-share-it/?sh=7eb29585c7c6
Jun 12, 2023
4
pmarca.substack.com/p/why-ai-will-save-the-world
Jun 7, 2023
213
every.to/chain-of-thought/we-re-building-ai-into-our-media-business
Jun 6, 2023
132
technomancers.ai/japan-goes-all-in-copyright-doesnt-apply-to-ai-training/
Jun 1, 2023
51
kylepoyar.substack.com/p/typeforms-viral-growth-and-its-disruption
Jun 1, 2023
162
future.com/cohort-based-courses/
Jun 1, 2023
251
a16z.com/2023/05/25/ai-canon/
May 26, 2023
1
jamesaltucher.com/blog/how-to-make-millions-with-idea-sex/
May 25, 2023
31
thegeneralist.substack.com/p/where-do-great-ideas-come-from
May 24, 2023
133
www.goldmansachs.com/intelligence/pages/the-creator-economy-could-approach-half-a-trillion-dollars-by-2027.html
May 17, 2023
41
greatergood.berkeley.edu/article/item/how_to_make_sure_you_keep_growing_and_learning
May 15, 2023
8
www.implications.com/p/the-personalization-wave-a-surge
May 9, 2023
184
every.to/chain-of-thought/gpt-4-is-a-reasoning-engine
May 6, 2023
52
www.semianalysis.com/p/google-we-have-no-moat-and-neither
May 5, 2023
132
medium.com/crv-insights/powertotheconsumer-insights-from-30-leading-consumer-ai-founders-operators-and-thinkers-c3c56e6db04e
May 5, 2023
4
www.mattprd.com/p/the-complete-beginners-guide-to-autonomous-agents
May 2, 2023
1
www.pinecone.io/learn/vector-database/
May 2, 2023
9
waitbutwhy.com/2018/04/picking-career.html
Apr 27, 2023
91
www.sahilbloom.com/newsletter/intellectual-sparring-partners
Apr 25, 2023
5
bigthink.com/neuropsych/reading-fiction-empathy-better-person/
Apr 19, 2023
9
hbr.org/2022/09/emotions-arent-the-enemy-of-good-decision-making
Apr 18, 2023
91
www.sahilbloom.com/newsletter/the-four-idols-money-power-pleasure-fame
Apr 17, 2023
61
www.generalist.com/briefing/socials-next-wave
Apr 13, 2023
162
www.slideshare.net/ahrefs/how-search-works-256157502
Apr 12, 2023
2
fs.blog/why-write/
Apr 11, 2023
62
on.substack.com/p/introducing-notes
Apr 6, 2023
41
medium.datadriveninvestor.com/7-powerful-ai-tools-for-creators-and-entrepreneurs-to-speed-up-your-success-1141c8a1f37e
Apr 6, 2023
31
www.aboutamazon.com/news/company-news/2016-letter-to-shareholders
Apr 4, 2023
81
marginalrevolution.com/marginalrevolution/2023/03/existential-risk-and-the-turn-in-human-history.html
Apr 4, 2023
3
acecreamu.substack.com/p/is-y-combinator-worth-the-money
Mar 31, 2023
123
www.youtube.com/watch?v=L_Guz73e6fw
Mar 29, 2023
335
www.darcycoolican.com/blog/ideamaze
Mar 27, 2023
9
Open-source models are faster, more customizable, more private, and pound-for-pound more capable.
People will not pay for a restricted model when free, unrestricted alternatives are comparable in quality.
Giant models are slowing us down. In the long run, the best models are the ones
which can be iterated upon quickly. We should make small variants more than an afterthought, now that we know what is possible in the <20B parameter regime.
Most importantly, they have solved the scaling problem to the extent that anyone can tinker. Many of the new ideas are from ordinary people. The barrier to entry for training and experimentation has dropped from the total output of a major research organization to one person, an evening, and a beefy laptop.
LoRA works by representing model updates as low-rank factorizations, which reduces the size of the update matrices by a factor of up to several thousand. This allows model fine-tuning at a fraction of the cost and time. Being able to personalize a language model in a few hours on consumer hardware is a big deal, particularly for aspirations that involve incorporating new and diverse knowledge in near real-time.
If we really do have major architectural improvements that preclude directly reusing model weights, then we should invest in more aggressive forms of distillation that allow us to retain as much of the previous generation’s capabilities as possible.
LoRA updates are very cheap to produce (~$100) for the most popular model sizes. This means that almost anyone with an idea can generate one and distribute it. Training times under a day are the norm. At that pace, it doesn’t take long before the cumulative effect of all of these fine-tunings overcomes starting off at a size disadvantage.
the best are already largely indistinguishable from ChatGPT. Focusing on maintaining some of the largest models on the planet actually puts us at a disadvantage.
Many of these projects are saving time by training on small, highly curated datasets. This suggests there is some flexibility in data scaling laws. The existence of such datasets follows from the line of thinking in Data Doesn't Do What You Think, and they are rapidly becoming the standard way to do training outside Google.
holding on to a competitive advantage in technology becomes even harder now that cutting edge research in LLMs is affordable. Research institutions all over the world are building on each other’s work, exploring the solution space in a breadth-first way that far outstrips our own capacity.
the one clear winner in all of this is Meta. Because the leaked model was theirs, they have effectively garnered an entire planet's worth of free labor. Since most open source innovation is happening on top of their architecture, there is nothing stopping them from directly incorporating it into their products.
The value of owning the ecosystem cannot be overstated. Google itself has successfully used this paradigm in its open source offerings, like Chrome and Android. By owning the platform where innovation happens, Google cements itself as a thought leader and direction-setter, earning the ability to shape the narrative on ideas that are larger than itself.
in the end, OpenAI doesn’t matter. They are making the same mistakes we are in their posture relative to open source, and their ability to maintain an edge is necessarily in question. Open source alternatives can and will eventually eclipse them unless they change their stance.