Readit News logoReadit News

Deleted Comment

Deleted Comment

jncraton commented on Run Python in the Browser Effortlessly   kai.bi/post/run-python-pr... · Posted by u/ccbikai
david_draco · 8 months ago
Wake me up when it supports numpy&scipy.
jncraton · 8 months ago
jncraton commented on Go library for in-process vector search and embeddings with llama.cpp   github.com/kelindar/searc... · Posted by u/kelindar
ausbah · 10 months ago
could anyone recommend a similar library for python?
jncraton · 10 months ago
The languagemodels[1] package that I maintain might meet your needs.

My primary use case is education, as myself and others use this for short student projects[2] related to LLMs, but there's nothing preventing this package from being used in other ways. It includes a basic in-process vector store[3].

[1] https://github.com/jncraton/languagemodels

[2] https://www.merlot.org/merlot/viewMaterial.htm?id=773418755

[3] https://github.com/jncraton/languagemodels?tab=readme-ov-fil...

jncraton commented on Phind-405B and faster, high quality AI answers for everyone   phind.com/blog/introducin... · Posted by u/rushingcreek
jncraton · a year ago
It would be nice to see the Phind Instant weights released under a permissive license. It looks like it could be a useful tool in the local-only code model toolbox.

Deleted Comment

jncraton commented on Lossless Acceleration of LLM via Adaptive N-Gram Parallel Decoding   arxiv.org/abs/2404.08698... · Posted by u/PaulHoule
MyFirstSass · a year ago
If this is "plug and play" can it be added to say llama.cpp and give ~3.67 speedup to existing models or is there some complication?
jncraton · a year ago
The speedup would not be that high in practice for folks already using speculative decoding[1]. ANPD is similar but uses a simpler and faster drafting approach. These two enhancements can't be meaningfully stacked. Here's how the paper describes it:

> ANPD dynamically generates draft outputs via an adaptive N-gram module using real-time statistics, after which the drafts are verified by the LLM. This characteristic is exactly the difference between ANPD and the previous speculative decoding methods.

ANPD does provide a more general-purpose solution to drafting that does not require training, loading, and running draft LLMs.

[1] https://github.com/ggerganov/llama.cpp/pull/2926

jncraton commented on Launch HN: Greptile (YC W24) - RAG on codebases that actually works    · Posted by u/dakshgupta
jaffee · 2 years ago
> embedding vectors you've calculated from the code? If so, those are likely quite easily reversible

I don't think embeddings are generally reversible... you're usually projecting onto a lower dimensional space, and therefore losing information.

jncraton · 2 years ago
You might be interested in "Text Embeddings Reveal (Almost) As Much As Text":

> We train our model to decode text embeddings from two state-of-the-art embedding models, and also show that our model can recover important personal information (full names) from a dataset of clinical notes.

https://arxiv.org/pdf/2310.06816.pdf

There's certainly information loss, but there is also a lot of information still present.

u/jncraton

KarmaCake day1381June 1, 2008
About
I'm always interested in new ideas. Feel free to get in touch.

https://joncraton.com

my username [at] g m a i l

View Original