Three years ago, we began reimagining AI development by rebuilding its infrastructure to be more performant, programmable, and portable. Just a few weeks ago, we introduced MAX 24.6, featuring MAX GPU–a preview of the first vertically integrated generative AI serving stack that eliminates dependency on vendor-specific libraries like NVIDIA CUDA.
MAX GPU is built on two groundbreaking technologies:
- MAX Engine: A high-performance AI model compiler and runtime supporting vendor-agnostic Mojo GPU kernels for NVIDIA GPUs.
- MAX Serve: A Python-native serving layer engineered for LLMs, handling complex request batching and scheduling for reliable performance under heavy workloads.
We appreciate the community's feedback and questions about this exciting release. As we continue to refine MAX GPU, we encourage you to continue sharing your thoughts, ideas, and experiences in the forum and Discord. Your input plays a key role in shaping the future of this technology.
Blogs, Tutorials, and Videos
- Community Meeting #11 was a special event called Modular milestones: GPUs, 2024 reflections, and the road ahead. Chris Lattner and team shared updates on MAX 24.6 and MAX GPU, our Mojo open source approach, the standard library contribution policy, and more.
- In Community Meeting #12, Max Hutchinson and Tyler Kenney covered MAX GPU benchmarking, Brad Larson shared his new MAX-powered image processing framework, MAX-CV, and the team answered questions from the community.
- MAX 24.6 is here, featuring MAX GPU, the first vertically integrated gen AI serving stack, with SoTA performance on NVIDIA A100 and support for deployment across all major clouds. Check out the announcement.
- MAX GPU preview delivers SOTA LLM throughput performance on NVIDIA A100s. Dig into our benchmarking methodology, results, and takeaways with our blog post.
- Want to see MAX 24.6 in action? Build your own GPU-accelerated chat app with MAX Serve and Llama 3.1 by following our blog post.
- Mojo 24.6 brings important changes to argument conventions and lifetimes management, making Mojo’s memory and ownership model more intuitive while maintaining strong safety guarantees. Deep dive into the exciting Mojo 24.6 updates with our blog post.
Awesome MAX + Mojo
Want to learn more about these projects? Chat with their creators in the #community-showcase channel of our Discord server or the Community Showcase category of our forum.
- Bridge, EmberJson, Endia, ExtraMojo, Hue, Mojmelo, and NuMojo are now available for installation via Magic! Just add the Modular community channel (https://repo.prefix.dev/modular-community) to your
mojoproject.toml
orpixi.toml
and runmagic add <package name>
. Learn more here. - McCoy Becker has started building a compiler middleware project called juju to bring JAX to the world of MAX Graphs.
- Talha Tahir implemented moondream, a small vision language model, in Mojo and Python, using the MAX Graph API.
- benz0li shared a JupyterLab MAX and Mojo Docker stack, including a GPU-accelerated option, MAX and Mojo data science dev containers, and a dev container for Mojo standard library development.
- Zbornak created a terminal game in Mojo based on Morabaraba, a traditional 2-player strategy game from South Africa.
- Sarctiann and other community members shared Neovim configs for Mojo.
- PhoToN 旭輝 released HEPJo v0.1.0, a high-performance library for numerical computations in particle physics, written in Mojo.
- Cdjones wrote an implementation of min heap in Mojo.
- duck_tape created a benchmark using ExtraMojo to compare Mojo to Python and Rust.
- samufi shared a performance-oriented archetype-based Entity Component System for Mojo, based on Go’s Arche.
- mlange-42 released Modo, a documentation generator for Mojo.
- Leandro Campos built MPFR for Mojo, a proof of concept that demonstrates how the GNU MPFR library can be used a gold standard to test the correctness of mathematical functions implemented in Mojo.
- Robert created PyMo, a framework in Mojo for AI, ML, and DL applications.
Open-Source Contributions
If you’ve recently had your first PR merged, message Caroline Frasca (@Caroline_Frasca) on Discord or the forum to claim your epic Mojo swag!
Check out the recently merged contributions from our valued community members:
- martinvuyk [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15]
- owenhilyard [1] [2] [3]
- PunkPun [1]
- rd4com [1]
- msaelices [1] [2]
- jjvraw [1] [2]
- avitkauskas [1]
- cnhz95 [1]
- guzmanri [1]
- soraros [1] [2] [3]
- bmerkle [1]
- el2e10 [1]
- miguelcsx [1]
- kraudy [1]
- bgreni [1] [2]
- mahiro21h [1]
Coming Up
Democratize Intelligence Summit
Chris Lattner will speak at the second Democratize Intelligence Summit this Friday, January 24th, at 11 AM PT in San Francisco. Chris will also participate in a panel discussion at 12:35 PM PT.
Modular Community Meeting
Our next community meeting will take place on February 3rd at 10 AM PT. RSVP in Discord.