Remote: yes
Willing to relocate: no
Technologies: python/django SQL, metabase, C#
resume/cv: https://drive.google.com/file/d/1px-oOulk-JZM9Ir6uh24w6pyNBL...
email: joshuamkogut@gmail.com
Deleted Comment
Remote: yes
Willing to relocate: no
Technologies: python/django SQL, metabase, C#
resume/cv: https://drive.google.com/file/d/1px-oOulk-JZM9Ir6uh24w6pyNBL...
email: joshuamkogut@gmail.com
I get that this still sucks for any individual with a Telescope though.
See: Hubble, JWST, and more to come
edit: JWST is in a different point from LEO but still counts
[1] https://github.com/Hellisotherpeople/DebateSum [2] https://github.com/EleutherAI/the-pile/issues/56
Stella was waiting for you to submit your dataset. Did you? She closed the ticket many months later.
I expected to like lenses and favoring/blocking specific domains. What I didn’t expect was how much their “Quick Answer” would change how I search.
I’ve been “AI hesitant”, in general the chance that an LLM will hallucinate makes these kinds of tools more trouble than they’re worth for me personally. In Kagi’s case, though, the individual facts it states in the quick answers have citations linking to the site it drew that information from.
Here’s what I’ve found:
- it’s been accurate most of the time, but not 100% (as expected)
- citations are pretty accurate most of the time
- every so often the citation links to a page that seemingly doesn’t back the claim in the quick answer
Unsurprisingly, I don’t trust the AI generated quick answer in isolation, what it does do is let me scan a few paragraphs, find the one that answers my question most specifically, and visit the sites it links to as citations for that piece of the answer. This saves me the time of clicking through the top $N results and scanning each page to find the one that seems to answer my query most directly. It’s like a layer on top of the page rank.
I remember using Google the first time and being impressed how the top answers were so much more relevant than Yahoo, it was a huge time saver. Now I find myself wondering if the “quick answer” citations will prove to be a similar jump in accelerating my ability to find the right web page.
It also makes me wonder if their own page rank algorithm could incorporate the quick answer output as an input to a site’s rank? That would be an interesting experiment!
With DeepSeek we can now run on non-GPU servers with a lot of RAM. But surely quite a lot of the 671 GB or whatever is knowledge that is usually irrelevant?
I guess what I sort of am thinking of is something like a model that comes with its own built in vector db and search as part of every inference cycle or something.
But I know that there is something about the larger models that is required for really intelligent responses. Or at least that is what it seems because smaller models are just not as smart.
If we could figure out how to change it so that you would rarely need to update the background knowledge during inference and most of that could live on disk, that would make this dramatically more economical.
Maybe a model could have retrieval built in, and trained on reducing the number of retrievals the longer the context is. Or something.
I tried the same thing on the 7B and 32B model today, neither are as effective as codellama.