Llama 3.3 (70B) Finetuning - now with 90K context length and fits on <41GB VRAM.
You can now run DeepSeek-R1 on your own local device!
1.58bit DeepSeek R1 - 131GB Dynamic GGUF
I was able to 1-shot prompt the unlsoth "python flappy bird game" test with Deepseek R1 distilled 70B. The distilled models deserve more credit.
The new Mistral Small model is disappointing
DeepSeek AI blocked by Italian authorities
Running Deepseek R1 IQ2XXS (200GB) from SSD actually works
Fine Tuning On Completions only using Unsloth
Unsloth made dynamic R1 quants - can be run on as little as 80gb of RAM
I have a 12GB 3060, is it possible to fine-tune ANY model?
[R] Replicating DeepSeek-R3-Zero RL recipe on 3B LLM for <30$, the model develops self-verification and search abilities all on its own
I fixed 4 bugs in Microsoft's open-source Phi-4 model
Deepseek-R1 GGUFs + All distilled 2 to 16bit GGUFs + 2bit MoE GGUFs
Deepseek just uploaded 6 distilled verions of R1 + R1 "full" now available on their website.
let’s goo, DeppSeek-R1 685 billion parameters!
DeepSeek R1 has been officially released!
[P] How I found 8 bugs in Google's Gemma 6T token model
[P] How I found & fixed 4 bugs in Microsoft's Phi-4 model
I accidentally built an open alternative to Google AI Studio
Phi-4 Llamafied + 4 Bug Fixes + GGUFs, Dynamic 4bit Quants
Phi-4 has been released
Phi-4 Finetuning - now with >128K context length + Bug Fix Details
DeepSeek-V3 imatrix quants by team mradermacher
Now that Phi-4 has been out for a while what do you think?
Resources for AI