AI News Meta and Stanford Researchers Propose Fast Byte Latent Transformer That Reduces Inference Memory Bandwidth by Over 50% Without Tokenization May 11, 2026 0 7 FacebookXPinterestWhatsAppLinkedinReddItEmailPrintTumblrTelegramMix