Let’s make a generation of amazing image-generation models
The best image generation models are trained on human preference datasets, where annotators have selected the best image from a choice of two. Unfortunately, many of these datasets are closed source so the community cannot train open models on them. Let’s change that!
The community can contribute image preferences for an open-source dataset that could be used for building AI models that convert text to image, like the flux or stable diffusion families. The dataset will be open source so everyone can use it to train models that we can all use.
I have been seeing a specific type of AI hype more and more, I call it, releasing research expecting that no one will ever reproduce your methods, then overhyping your results. I test the methodology of maybe 4-5 research papers per day. That is how I find a lot of my research. Usually, 3-4 of those experiments end up not being reproduceable for some reason. I am starting to think it is not accidental.
So, I am launching a new series where I specifically showcase a research paper by reproducing their methodology and highlighting the blatant flaws that show up when you actually do this. Here is Episode 1!
Good folks from @Microsoft Research have just released bitnet.cpp, a game-changing inference framework that achieves remarkable performance gains.
Key Technical Highlights: - Achieves speedups of up to 6.17x on x86 CPUs and 5.07x on ARM CPUs - Reduces energy consumption by 55.4–82.2% - Enables running 100B parameter models at human reading speed (5–7 tokens/second) on a single CPU
Features Three Optimized Kernels: 1. I2_S: Uses 2-bit weight representation 2. TL1: Implements 4-bit index lookup tables for every two weights 3. TL2: Employs 5-bit compression for every three weights
Performance Metrics: - Lossless inference with 100% accuracy compared to full-precision models - Tested across model sizes from 125M to 100B parameters - Evaluated on both Apple M2 Ultra and Intel i7-13700H processors
This breakthrough makes running large language models locally more accessible than ever, opening new possibilities for edge computing and resource-constrained environments.
💾🧠Want to know how much VRAM you will need for training your model? 💾🧠 Now you can use this app in which you can input a torch/tensorflow summary or the parameters count and get an estimate of the required memory! Use it in: howmuchvram.com
⚡ My PhD thesis, “Scalable Nested Optimization for Deep Learning,” is now on arXiv! ⚡
tl;dr: We develop various optimization tools with highlights, including: · Making the momentum coefficient complex for adversarial games like GANs. · Optimizing millions of hyperparameters using implicit differentiation. · Tuning hyperparameters using hypernetworks. · Differentiably finding bifurcations in optimization for diverse solutions.