Readit News logoReadit News
galeos commented on Everything we announced at our first LlamaCon   ai.meta.com/blog/llamacon... · Posted by u/meetpateltech
philipkglass · 8 months ago
I recently needed to classify thousands of documents according to some custom criteria. I wanted to use LLM classification from these thousands of documents to train a faster, smaller BERT (well, ModernBERT) classifier to use across millions of documents.

For my task, Llama 3.3 was still the best local model I could run. I tried newer ones (Phi4, Gemma3, Mistral Small) but they produced much worse results. Some larger local models are probably better if you have the hardware for them, but I only have a single 4090 GPU and 128 GB of system RAM.

galeos · 8 months ago
How did you find ModernBERT performance Vs prior BERT models?
galeos commented on BitNet b1.58 2B4T Technical Report   arxiv.org/abs/2504.12285... · Posted by u/galeos
galeos · 8 months ago
You can try out the model in a demo they have setup: https://bitnet-demo.azurewebsites.net/

u/galeos

KarmaCake day817December 4, 2014View Original