10 Comments

Really interesting article and very complete. Thanks for sharing!

Let me add a remark/question about some companies that would be in between the two categories: those that offer a consumable AI service via API to be integrated into software or automation, while relying themselves on more "low-level" third-party services. However, they offer additional processing and provide an "intelligence" that justifies their use.

Concretely, this concerns for example document parsing services (contracts, invoices, resumes, etc.) that rely on Google or AWS for OCR but have trained their own NLP engines. The same goes for services based on speech recognition. Where would you categorize them?

We are working at Eden AI (www.edenai.co) to aggregate and harmonize foundation models to make them easier to use for the people who use them and to make it very easy for them to swap them according to the performance achieved for their specific data or the price evolution (which is a very important point). The borderline can however be quite thin between models that are 100% provided by one vendor and those that are partially based on someone else.

Expand full comment

Great read, thank you :) I was wondering how you predict defensibility for Level 1 companies? Will this become an enclosed circle of a few powerful companies or do you see challengers gaining traction?

Expand full comment
Jan 10, 2023Liked by Viet Le

Please share the prompt used to generate the title image

Expand full comment

So what are your conclusions then about which companies are most likely to challenge foundational model Generative A.I. darlings? Consolidation seems to occur faster here. Stability.AI already siding with AWS.

Expand full comment

Awesome article, thanks for sharing!

Expand full comment

Really cool article, Viet! :)

Contains lots of insights into technical topics that founders of AI companies have to look into (e.g. short-term cost of ML) while still having to deal with building a product people want.

Just as you mentioned, my "fear" is open-source falling behind proprietary solutions in AI because of the reasons you named. But (from my understanding), this _strict_ platform-dependency is mainly for generative AI use cases, isn't it? In discriminative use cases, LLMs like GPT-X can be used to create the proprietary data itself (at least partially). For instance, we currently like using GPT-3 to create labeled datasets and then train super simple models on top of it, e.g. using an open-source encoder (distilbert or something from HF) for the embeddings and a simple old-school logistic regression :-) Maybe there is something like partial platform dependency? (e.g. for building, but not for inference). But just my thoughts.

Let's see. Really exciting times :)

Expand full comment