The worthwhile problems are the ones you can really solve or help solve, the ones you can really contribute something to. A problem is grand in science if it lies before us unsolved and we see some way for us to make some headway into it.
No problem is too small or too trivial if we can really do something about it.
dynamic computation seems like a big deal right now, with MOD and Quiet-STaR we can have dynamic routing on tokens with an arbitrary (?) amount of computation spent on each token. sounds like we can craft system that can choose important tokens and reason more on some. combined with some latent representation we could achieve some sort of abstract reasoning?
similar idea could be applied to image generation to avoid spending computation on un-important parts of the image and focus on more important, or to filter un-relevant tokens in text conditioning. the current vae used in stable diffusion is a bit of a glorified downscaler, but with a token representation the same rationale could be applied.
gecko embeddings for t2i as it looks they can carry more information than “normal” CLIP and T5 embeddings.