Chip Huyen

I help companies deploy machine learning into production. I write about AI applications, tooling, and best practices.
https://huyenchip.com/blog/ (RSS)
visit blog
Building A Generative AI Platform
25 Jul 2024 | original ↗

After studying how companies deploy generative AI applications, I noticed many similarities in their platforms. This post outlines the common components of a generative AI platform, what they do, and how they are implemented. I try my best to keep the architecture general, but certain applications might deviate. This is what the overall...

Measuring personal growth
17 Apr 2024 | original ↗

My founder friends constantly think about growth. They think about how to measure their business growth and how to get to the next order of magnitude scale. If they’re making $1M ARR today, they think about how to get to $10M ARR. If they have 1,000 users today, they think about how to get to 10,000 users. This made me wonder if/how people are...

What I learned from looking at 900 most popular open source AI tools
14 Mar 2024 | original ↗

[Hacker News discussion, LinkedIn discussion, Twitter thread] Four years ago, I did an analysis of the open source ML ecosystem. Since then, the landscape has changed, so I revisited the topic. This time, I focused exclusively on the stack around foundation models. The full list of open source AI repos is hosted at llama-police. The list is...

Predictive Human Preference: From Model Ranking to Model Routing
28 Feb 2024 | original ↗

A challenge of building AI applications is choosing which model to use. What if we don’t have to? What if we can predict the best model for any prompt? Predictive human preference aims to predict which model users might prefer for a specific query. Table of contents Ranking Models Using Human Preference …. How Preferential Ranking Works …....

Sampling for Text Generation
16 Jan 2024 | original ↗

ML models are probabilistic. Imagine that you want to know what’s the best cuisine in the world. If you ask someone this question twice, a minute apart, their answers both times should be the same. If you ask a model the same question twice, its answer can change. If the model thinks that Vietnamese cuisine has a 70% chance of being the best...

Multimodality and Large Multimodal Models (LMMs)
10 Oct 2023 | original ↗

For a long time, each ML model operated in one data mode – text (translation, language modeling), image (object detection, image classification), or audio (speech recognition). However, natural intelligence is not limited to just a single modality. Humans can read, talk, and see. We listen to music to relax and watch out for strange noises to...

Open challenges in LLM research
16 Aug 2023 | original ↗

[LinkedIn discussion, Twitter thread] Never before in my life had I seen so many smart people working on the same goal: making LLMs better. After talking to many people working in both industry and academia, I noticed the 10 major research directions that emerged. The first two directions, hallucinations and context learning, are probably the...

Generative AI Strategy
7 Jun 2023 | original ↗

I had a lot of fun preparing the talk: “Leadership needs us to do generative AI. What do we do?” for Fully Connected. The idea for the talk came from many conversations I’ve had recently with friends who need to figure out their generative AI strategy, but aren’t sure what exactly to do. This talk is a simple framework to explore what to do with...

RLHF: Reinforcement Learning from Human Feedback
2 May 2023 | original ↗

[LinkedIn discussion, Twitter thread] In literature discussing why ChatGPT is able to capture so much of our imagination, I often come across two narratives: Scale: throwing more data and compute at it. UX: moving from a prompt interface to a more natural chat interface. One narrative that is often glossed over is the incredible technical...

Building LLM applications for production
11 Apr 2023 | original ↗

[Hacker News discussion, LinkedIn discussion, Twitter thread] Update: My upcoming book, AI Engineering (late 2024/early 2025) will cover building aplications with foundation models in depth. A question that I’ve been asked a lot recently is how large language models (LLMs) will change machine learning workflows. After working with several...

↑ these items are from RSS. Visit the blog itself at https://huyenchip.com/blog/ to find other articles and to appreciate the author's digital home.