this post was submitted on 21 Nov 2023
1 points (100.0% liked)

LocalLLaMA

3 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 1 year ago
MODERATORS
 

I tried to run locally various networks for code completion, but it turns out most of the plugins are just directly feed data to neural network and that’s why they all suck, so I built a new that has required preprocessing and uses ollama instead of something weird.

you are viewing a single comment's thread
view the rest of the comments
[–] geepytee@alien.top 1 points 11 months ago

I built a new that has required preprocessing

Nice! How are you preprocessing the code?