Quantizable Transformers: Removing Outliers by Helping Attention Heads Do Nothing | Read Paper on Bytez