šŸ”„Tencent Unveils DeepSeek Rival

PLUS: Diffusion-Based LLM Breakthrough

Reading time: 5 minutes

Today we will discuss:

Product for Engineers: Flex your product muscles

Product for Engineers is PostHogā€™s newsletter dedicated to helping engineers and founders improve their product skills. In it, they provide curated advice on building great products, lessons (and mistakes) theyā€™ve learned building PostHog, and deep dives on the culture and strategies of top startups. Subscribe for free.

GoogleImages

Key Points 

  • Tencent says its new AI model, Turbo S, can generate responses in under a second, outperforming DeepSeek R1.

  • The company claims Turbo S matches DeepSeek-V3 in knowledge, math, and reasoning.

ā™ØļøNews - Tencent has unveiled Hunyuan Turbo S, a new AI model it claims can respond faster than DeepSeekā€™s R1, a model that has gained significant attention in the AI space.

šŸ§Does it measure up? According to Tencent, Turbo S can generate responses in under a second, setting it apart from models like DeepSeek R1 and Hunyuan T1, which take longer to process before answering. 

The company also says that in areas like knowledge, math, and reasoning, Turbo S performs at the same level as DeepSeek-V3, the model behind DeepSeekā€™s AI chatbot, which recently surpassed OpenAIā€™s ChatGPT in app store downloads.

Tencent is also emphasizing cost efficiency, saying Turbo S is significantly cheaper to use than its previous versions. The move reflects a growing trend in Chinaā€™s AI industry, where DeepSeekā€™s open-source approach and low pricing have forced major players to rethink their pricing strategies.

GoogleImages

Key Points 

  • Unlike LLMs, which generate words sequentially, Inceptionā€™s model refines entire text blocks in parallel for greater efficiency.

  • The startup claims its AI models run 10x faster than competitors while reducing GPU costs for enterprise users.

ā˜•News - A new AI startup, Inception, is making bold claims about a faster, more efficient approach to language models. Founded by Stanford professor Stefano Ermon, the Palo Alto-based company has introduced a ā€œdiffusion-based large language modelā€ (DLM), applying diffusion technologyā€”commonly used for images and videoā€”to text generation.

šŸ‘Øā€šŸ’»For context - Most generative AI models today fall into two categories: large language models (LLMs), which generate text, and diffusion models, which create images, video, and audio. While LLMs like ChatGPT process words sequentially, diffusion models work differently, refining an entire output in parallel.

Inceptionā€™s DLM is designed to offer the same core capabilities as LLMsā€”code generation, answering questions, and moreā€”but with significantly faster performance and lower computational costs.

āœØThe unique approach - Ermonā€™s research focused on overcoming a key limitation of LLMs: their step-by-step generation process. ā€œYou cannot generate the second word until youā€™ve generated the first one,ā€ he explained. Diffusion models, however, start with a rough estimate and refine everything simultaneously.

After years of research, Ermon and a student developed a way to apply this approach to text. Their breakthrough, published last year, demonstrated that generating and modifying large blocks of text in parallel was possibleā€”potentially reshaping how language models are built.

šŸ¤“What it's delivering - According to Ermon, this method allows models to use GPUs far more efficiently, making AI systems both faster and cheaper. Inception claims its DLMs run up to 10 times faster than traditional LLMs while cutting costs by the same factor.

The company is already offering an API, on-premises and edge deployment, and model fine-tuning options. A spokesperson compared Inceptionā€™s smallest coding model to OpenAIā€™s GPT-4o mini, saying it delivers similar results but at more than 10 times the speed.

šŸ™†šŸ»ā€ā™€ļøWhat else is happening?

We just launched a new edition of Workflow Wednesday for AI-minded professionals like you. Actionable AI-workflows delivered directly to your inbox.  This week's Topic? AI-Monetization

šŸ‘©šŸ¼ā€šŸš’Discover mind-blowing AI tools

  1. Learn How to Use AI - Starting January 8, 2025, weā€™re launching Workflow Wednesday, a series where we teach you how to use AI effectively. Lock in early bird pricing now and secure your spot. Check it out here

  2. OpenTools AI Tools Expert  - Find the perfect AI Tool to solve supercharge your workflow. This GPT is connected to our database, so you can ask in depth questions on any AI tool directly in ChatGPT (free)

  3. Gling - Automatically detects and cut silences and disfluencies from videos

  4. Juno - Makes data science 10x better by writing, editing, and automatically debugging your code with you as you go

  5. Klenty - A sales engagement platform that helps sales teams streamline their outreach efforts

  6. Wave.video - Live streaming studio, video editor, thumbnail maker, video hosting, video recording, and stock library combined in one platform

  7. Recraft - A tool that lets users create and edit digital illustrations, vector art, icons, and 3D graphics in a uniform brand style

  8. SaveDay - A Telegram bot that allows users to save, search and converse with various types of content, such as web pages, articles, PDFs, YouTube videos, and podcasts

  9. Westidol - An AI-powered tool that allows users to do professional photoshoots using just a single selfie

How likely is it that you would recommend the OpenTools' newsletter to a friend or colleague?

Login or Subscribe to participate in polls.

Interested in featuring your services with us? Email us at [email protected]