Liquid AI Debuts New LFM-Based AI Models, Surpassing Traditional LLMs in  Performance

| Updated on October 4, 2024

An AI startup and MIT spinoff, Liquid AI, co-founded by former researchers from the Massachusetts Institute of Technology (MIT)’s Computer Science and Artificial Intelligence Laboratory (CSAIL), announced its first set of generative AI models called “Liquid Foundation Models,” or LFMs.

The new models are built on a fundamentally new architecture and are built from “first principles…the same way engineers built engines, cars, and airplanes.” 

They are said to deliver impressive performance, surpassing even the traditional and some of the best large language models, such as Meta’s Llama 3.1-8B and Microsoft’s Phi-3.5 3.8B.

Liquid AI is releasing three models at launch:

  • LFM-1B is a dense model with 1.3 billion parameters, which is designed for resource-constrained environments.
  • LFM-3B is slightly powerful, has 3.1 billion parameters, and is aimed at edge deployments, such as mobile applications, robots, and drones.
  • LFM-40B is the most powerful “mixture of experts” model with 40.3 billion parameters, designed to be used in cloud servers to handle the complex use cases.

The startup believes new models have shown exceptional results, showing a smaller memory footprint for long-context processing.

LFM 3B model interference memory footprint statistic

These models are expected to become formidable competitors to generative AI models such as ChatGPT.

However, the new models are not open source, and thus they are available in early access via platforms, such as Liquid’s inference playground, Lambda Chat, or Perplexity AI.

What are LFMs?

According to Liquid, LFMs represent a new wave of AI models that are designed to deliver both efficiency and performance while using minimal system memory and delivering exceptional computing power. It is said they are ideal for handling sequential data, including text, audio, images, video, and signals.

LNNs are based on the concept of artificial neurons, and thus they combine those neurons with innovative mathematical formulations to transform data with the same performance but fewer neurons.

Akriti Rana

Tech Journalist