Hacker News new | past | comments | ask | show | jobs | submit login

I am constantly searching for this. To point a locally hosting LLM to a locally hosted code repo - and then start improving it from there would be awesome.



Depending on how serious you are about this (and how well you can slap together a few different python packages), it is very doable today.

Get one of the better llama versions fine-tuned on code (e.g. WizardCoder), take your entire code base and create embeddings from it, put those into a vector database.

Now, every time you ask your LLM a question about your code base, you first turn that prompt into an embedding and perform a search on your vector database. The results of that search are appended as context to the actual prompt before passing it to the LLM itself.

There's tons of packages that help with all of that, Langchain and Faiss are probably the most popular right now.


interesting, i'd love to do this too. but it sounds like there aren't any full-featured, opensource packages/projects that do this all-together? I'd love to hack the parts together, but i don't have the time/energy these days to do it.

thanks for the helpful keywords though, it helps point me in the right direction.




Join us for AI Startup School this June 16-17 in San Francisco!

Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: