DuoGPT: Training-free Dual Sparsity through Activation-aware Pruning in LLMs | Read Paper on Bytez