Kazuki
@kazuki
Cofounder of Glasp. I collect ideas and stories worth sharing 📚
San Francisco, CA
Joined Oct 9, 2020
1069
Following
5759
Followers
1.46k
13.54k
169.50k
markmanson.net/the-backwards-law
Jan 11, 2023
133
curationmonetized.substack.com/p/personal-viewpoint-cm-2
Jan 10, 2023
1
bakadesuyo.com/2014/02/samurai/
Jan 9, 2023
81
every.to/superorganizers/the-end-of-organizing
Jan 8, 2023
102
www.nirandfar.com/labeling-yourself
Jan 6, 2023
63
robotic.substack.com/p/ml-moats
Dec 30, 2022
9
every.to/superorganizers/understanding-the-science-of-creativity
Dec 26, 2022
162
www.marktechpost.com/2022/12/22/this-artificial-intelligence-ai-application-does-youtube-summary-with-chatgpt/
Dec 23, 2022
5
www.surgehq.ai/blog/googles-existential-threat-chatgpt-matches-googles-performance-on-informational-search-queries-and-smashes-it-on-coding
Dec 21, 2022
6
www.bensbites.co/p/youtube-summaries
Dec 19, 2022
11
www.nfx.com/post/generative-ai-tech-5-layers
Dec 19, 2022
192
nesslabs.com/everything-is-aiming
Dec 17, 2022
61
openai.com/blog/introducing-text-and-code-embeddings/
Dec 17, 2022
51
www.ben-evans.com/benedictevans/2022/12/14/ChatGPT-imagenet
Dec 17, 2022
4
deskoflawyer.com/glasp-web-clipper-web-highlighter/
Dec 12, 2022
8
leighmariebraswell.substack.com/p/overview-and-applications-of-large
Dec 12, 2022
71
ourworldindata.org/brief-history-of-ai
Dec 10, 2022
61
lspace.swyx.io/p/everything-we-know-about-chatgpt
Dec 6, 2022
62
thealgorithmicbridge.substack.com/p/chatgpt-is-the-worlds-best-chatbot
Dec 5, 2022
31
openai.com/blog/chatgpt/
Nov 30, 2022
72
bensbites.beehiiv.com/p/openai-update
Nov 29, 2022
11
writingcooperative.com/malcolm-gladwells-top-13-writing-tips-c833432c005f
Nov 29, 2022
164
bensbites.beehiiv.com/p/first-ai-laugh-lex-fridman-interviews-richard-feynman
Nov 25, 2022
11
nesslabs.com/prestige-psychology
Nov 24, 2022
91
every.to/napkin-math/6-new-theories-about-ai
Nov 24, 2022
142
gltr.io/
Nov 23, 2022
61
www.nfx.com/post/generative-tech
Nov 22, 2022
81
medium.com/taking-notes/make-highlights-and-annotations-social-with-glasp-the-social-highlighter-6028ff0358c2
Nov 22, 2022
4
davidspinks.substack.com/p/building-for-believers
Nov 18, 2022
143
a16z.com/2022/09/16/the-new-learning-economy-its-time-to-build-in-education/
Nov 18, 2022
101
whilstinarrakis.wordpress.com/2022/11/11/the-next-google-search-engine-will-be-generative-ai/
Nov 16, 2022
61
jmcdonnell.substack.com/p/the-near-future-of-ai-is-action-driven
Nov 16, 2022
61
themindcollection.com/the-feynman-technique/
Nov 15, 2022
141
every.to/napkin-math/how-pinterest-can-win
Nov 15, 2022
6
www.youtube.com/watch?v=TUD6iN_EuXc
Nov 11, 2022
5
www.youtube.com/watch?v=9uOMectkCCs
Nov 10, 2022
111
talia.gold/2022/10/31/a-trillion-dollar-opportunity/
Nov 9, 2022
85
hardfork.substack.com/p/easy-come-easy-go-understanding-the
Nov 9, 2022
61
www.generalist.com/briefing/what-to-watch-in-ai
Nov 9, 2022
51
The dialogue format makes it possible for ChatGPT to answer followup questions, admit its mistakes, challenge incorrect premises, and reject inappropriate requests.
We trained this model using Reinforcement Learning from Human Feedback (RLHF), using the same methods as InstructGPT, but with slight differences in the data collection setup.
We trained an initial model using supervised fine-tuning: human AI trainers provided conversations in which they played both sides—the user and an AI assistant.
we took conversations that AI trainers had with the chatbot. We randomly selected a model-written message, sampled several alternative completions, and had AI trainers rank them. Using these reward models, we can fine-tune the model using Proximal Policy Optimization.
ChatGPT is fine-tuned from a model in the GPT-3.5 series, which finished training in early 2022. You can learn more about the 3.5 series here. ChatGPT and GPT 3.5 were trained on an Azure AI supercomputing infrastructure.
ChatGPT sometimes writes plausible-sounding but incorrect or nonsensical answers. Fixing this issue is challenging
Ideally, the model would ask clarifying questions when the user provided an ambiguous query. Instead, our current models usually guess what the user intended.