All Articles  (X)

Clear
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.

🚨

NEW

Company

Our launch & what's next

Last week, we launched Modular to the world after more than 16 months in stealth. We started Modular with a deep conviction — after 6+ years of building and scaling AI infrastructure to billions of users and 20+ years of building foundational compute infrastructure — it was clear the world needed a better path forward. Everyone wants less complexity, better access to compute and hardware, and the ability to develop and deploy AI faster.

May 11, 2023

/

Tim Davis

Read

🚨

NEW

Product

A unified, extensible platform to superpower your AI

We’re excited to finally share what we’ve been building at Modular. This announcement begins Modular’s journey to radically change the nature of AI programmability, usability, scalability, and compute.

May 2, 2023

/

Chris Lattner

Tim Davis

Eric Johnson

Read

🚨

NEW

Engineering

The world's fastest unified matrix multiplication

In this post, we describe Modular’s approach to solving this problem and its game-changing benefits, including a new standard in state-of-the-art (SOTA) performance on CPU as compared to existing solutions.

April 20, 2023

/

Abdul Dakkak

Chad Jarvis

Eric Johnson

Hengjie Wang

Ian Tramble

Read

🚨

NEW

Engineering

AI’s compute fragmentation: what matrix multiplication teaches us

AI is powered by a virtuous circle of data, algorithms (“models”), and compute. Growth in one pushes needs in the others and can grossly affect the developer experience on aspects like usability and performance. Today, we have more data and more AI model research than ever before, but compute isn’t scaling at the same speed due to … well, physics.

March 23, 2023

/

Eric Johnson

Abdul Dakkak

Chad Jarvis

Read

🚨

NEW

Company

We want to hear from you

At Modular, we are rebuilding AI infrastructure for the world. Our goal is to move past AI tools that are themselves research projects and into a future where AI development and deployment are orders of magnitude more efficient for everyone. You should be able to do this without trading off performance or having to rewrite your entire code base.

December 15, 2022

/

Eric Johnson

Read

🚨

NEW

Engineering

If AI serving tech can’t solve today’s problems, how do we scale into the future?

The technological progress that has been made in AI over the last ten years is breathtaking — from AlexNet in 2012 to the recent release of ChatGPT, which has taken large foundational models and conversational AI to another level.

December 8, 2022

/

Eric Johnson

Tim Davis

Read

🚨

NEW

Engineering

Part 2: Increasing development velocity of giant AI models

The first four requirements address one fundamental problem with how we've been using MLIR: weights are constant data, but shouldn't be managed like other MLIR attributes. Until now, we've been trying to place a square peg into a round hole, creating a lot of wasted space that's costing us development velocity (and, therefore, money for users of the tools).

November 10, 2022

/

Abdul Dakkak

Eric Johnson

Read

🚨

NEW

Company

Modular is rebuilding AI in the face of a new economy

Here in November 2022, we see a continuing onslaught of bad news: significant layoffs of incredible people as companies tighten their belts; companies that raised too much money, too fast, without core fundamentals are dying; and a changing climate where over-tightening rather than under-tightening is seemingly the new normal.

November 8, 2022

/

Chris Lattner

Tim Davis

Read

🚨

NEW

Case Study

Modular's Brand Story

What makes a brand? When you see an apple with a bite out of it, you immediately associate it with the famous technology company — but what about the logo elicits the feelings it does? 

August 18, 2022

/

Tim Davis

Eric Johnson

Matt Ellis

Read

🚨

NEW

Engineering

Increasing development velocity of giant AI models

Machine learning models are getting larger and larger — some might even say, humongous. The world’s most advanced technology companies have been in an arms race to see who can train the largest model (MUM, OPT, GPT-3, Megatron), while other companies focused on production systems have scaled their existing models to great effect. Through all the excitement, what’s gone unsaid is the myriad of practical challenges larger models present for existing AI infrastructure and developer workflows.

August 12, 2022

/

Eric Johnson

Read

🤔

No results for this query