BryanW commited on
Commit
beb7c63
·
verified ·
1 Parent(s): 6a2acd5

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -4
README.md CHANGED
@@ -7,12 +7,24 @@ sdk: static
7
  pinned: false
8
  ---
9
 
10
- MeissonFlow Research is a non-commercial research group sharing the interest in artistic creation techniques. The main products of MeissonFlow Research involves models in image synthesis.
11
 
12
- Our goal is to design algorithms and build tools to make it easier for artists and designers to create cool things.
 
13
 
 
 
14
 
 
15
 
16
- Official homepage for presenting demos: https://sites.google.com/view/meissonic/home
 
17
 
18
- Inference Code of Meissonic: https://github.com/viiika/Meissonic
 
 
 
 
 
 
 
 
7
  pinned: false
8
  ---
9
 
10
+ # MeissonFlow Research
11
 
12
+ **MeissonFlow Research** is a non-commercial research group dedicated to advancing generative modeling techniques for structured visual and multimodal content creation.
13
+ We aim to design models and algorithms that help creators produce high-quality content with greater efficiency and control.
14
 
15
+ Our journey began with [**MaskGIT**](https://arxiv.org/abs/2202.04200), a pioneering work by [**Huiwen Chang**](https://scholar.google.com/citations?hl=en&user=eZQNcvcAAAAJ), which introduced a bidirectional transformer decoder for image synthesis—outperforming traditional raster-scan autoregressive (AR) generation.
16
+ This paradigm was later extended to text-to-image synthesis in [**MUSE**](https://arxiv.org/abs/2301.00704).
17
 
18
+ Building upon these foundations, we scaled masked generative modeling with the latest architectural designs and sampling strategies—culminating in [**Monetico** and **Meissonic**](https://github.com/viiika/Meissonic) from scratch, which on par with leading diffusion models such as SDXL, while maintaining greater efficiency.
19
 
20
+ Having verified the effectiveness of this approach, we began to ask a deeper question — one that reaches beyond performance benchmarks: **what foundations are required for general-purpose generative intelligence**?
21
+ Through discussions with researchers at Safe Superintelligence (SSI) Club, University of Illinois Urbana-Champaign (UIUC) and Riot Video Games, we converged on the vision of a **visual-centric world model** — a generative and interactive system capable of simulating, interacting with, and reasoning about multimodal environments.
22
 
23
+ > We believe that **masking** is a fundamental abstraction for building such controllable, efficient, and generalizable intelligence.
24
+
25
+ To pursue this vision, we introduced [**Muddit** and **Muddit Plus**](https://github.com/M-E-AGI-Lab/Muddit), unified generative models built upon visual priors (Meissonic), and capable of unified generation across text and image within a single architecture and paradigm.
26
+
27
+ We look forward to releasing more models and algorithms in this direction.
28
+ We thank our amazing teammates — and you, the reader — for your interest in our work.
29
+
30
+ Special thanks to [**Style2Paints Research**](https://lllyasviel.github.io/Style2PaintsResearch/), which helped shape our taste and research direction in the early days.