4M is a training framework for “any-to-any” vision foundation models that uses tokenization and masking to scale across many modalities and tasks. The same model family can classify, segment, detect, caption, and even generate images, with a single interface for both discriminative and generative use. The repository releases code and models for multiple variants (e.g., 4M-7 and 4M-21), emphasizing transfer to unseen tasks and modalities. Training/inference configs and issues discuss things like depth tokenizers, input masks for generation, and CUDA build questions, signaling active research iteration. The design leans into flexibility and steerability, so prompts and masks can shape behavior without bespoke heads per task. In short, 4M provides a unified recipe to pretrain large multimodal models that generalize broadly while remaining practical to fine-tune.

Features

  • Any-to-any modeling across diverse vision tasks
  • Masked modeling with unified tokenization for multiple modalities
  • Released model families (e.g., 4M-7, 4M-21) with training/eval code
  • Promptable and steerable behavior without task-specific heads
  • Transfer to unseen tasks and modalities from a single backbone
  • Research-grade configs and examples for reproduction

Project Samples

Project Activity

See All Activity >

Categories

AI Models

License

Apache License V2.0

Follow 4M

4M Web Site

You Might Also Like
MongoDB Atlas runs apps anywhere Icon
MongoDB Atlas runs apps anywhere

Deploy in 115+ regions with the modern database for every enterprise.

MongoDB Atlas gives you the freedom to build and run modern applications anywhere—across AWS, Azure, and Google Cloud. With global availability in over 115 regions, Atlas lets you deploy close to your users, meet compliance needs, and scale with confidence across any geography.
Start Free
Rate This Project
Login To Rate This Project

User Reviews

Be the first to post a review of 4M!

Additional Project Details

Programming Language

Python

Related Categories

Python AI Models

Registered

2025-10-08