| | Unsloth – Dynamic 4-bit Quantization (unsloth.ai) |
| 3 points by amrrs 11 months ago | past |
|
| | Dynamic 4bit Quantization (unsloth.ai) |
| 3 points by danielhanchen 11 months ago | past | 5 comments |
|
| | Unsloth: Easily finetune and train LLMs Get faster with unsloth (unsloth.ai) |
| 3 points by handfuloflight 12 months ago | past |
|
| | Unsloth creators fix universal error with gradient accumulation (unsloth.ai) |
| 4 points by ZQ-Dev8 on Oct 16, 2024 | past | 2 comments |
|
| | Bugs in LLM Training – Gradient Accumulation Fix (unsloth.ai) |
| 81 points by apsec112 on Oct 16, 2024 | past | 16 comments |
|
| | Bugs in LLM Training – Gradient Accumulation Fix (unsloth.ai) |
| 3 points by jasondavies on Oct 15, 2024 | past |
|
| | 2x faster Gemma 2 finetuning and 63% less VRAM (unsloth.ai) |
| 3 points by ricopags on July 4, 2024 | past | 1 comment |
|
| | Continued LLM Pretraining with Unsloth (unsloth.ai) |
| 1 point by swyx on June 5, 2024 | past |
|
| | Fixing Gemma Bugs (unsloth.ai) |
| 166 points by danielhanchen on March 11, 2024 | past | 63 comments |
|
| | Fixing All Gemma Bugs (unsloth.ai) |
| 2 points by xyzzyrz on March 7, 2024 | past |
|
| | CodeLlama-34B 13x faster finetuning (unsloth.ai) |
| 2 points by danielhanchen on Dec 16, 2023 | past |
|
| | Reducing FLOPs for transformers (unsloth.ai) |
| 1 point by danielhanchen on Dec 14, 2023 | past | 1 comment |
|
| | Unsloth: 30x faster AI training (unsloth.ai) |
| 3 points by Tomte on Dec 1, 2023 | past | 1 comment |
|
| | Finetune language models 30x faster (unsloth.ai) |
| 2 points by danielhanchen on Nov 30, 2023 | past |
|