nm-testing/SparseLlama-3.1-8B-gsm8k-pruned.2of4-FP8-Dynamic-Channel-BitMaskCompressed Updated Dec 6, 2024 • 3
nm-testing/Meta-Llama-3-8B-Instruct-FP8-Dynamic-IA-Per-Tensor-Weight-testing Updated Dec 6, 2024 • 35
nm-testing/TinyLlama-1.1B-Chat-v1.0-INT8-Dynamic-IA-Per-Channel-Weight-testing Updated Dec 8, 2024 • 36
nm-testing/TinyLlama-1.1B-Chat-v1.0-INT8-Dynamic-IA-Per-Tensor-Weight-testing Updated Dec 8, 2024 • 35