this post was submitted on 09 Nov 2023
1 points (100.0% liked)

LocalLLaMA

3 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 1 year ago
MODERATORS
 
top 11 comments
sorted by: hot top controversial new old
[–] Dorialexandre@alien.top 1 points 1 year ago

Link to the ongoing demo for MonadGPT, with generous GPU support from HuggingFace : https://huggingface.co/spaces/Pclanglais/MonadGPT

The model has been published as well (and soon the dataset): https://huggingface.co/Pclanglais/MonadGPT?text=Hi.

[–] vec1nu@alien.top 1 points 1 year ago

Which frontend is that?

[–] buzzyness@alien.top 1 points 1 year ago (2 children)

Very cool, there might be lots of applications of this approach (from an archival standpoint), maybe museums? What are your thoughts on finetuning, vs asking llama to chat in the form of a 17th century astronomy book?

[–] Dorialexandre@alien.top 1 points 1 year ago

Well that was actually my original motivation for finetuning. Even GPT-4 is not so good with a proper prompt: the text feels fake and/or struggle to maintain cultural consistency. I think finetuning works better for this task, as there are too many directives to give and it helps to relieve the model from anachronistic RLHF.

As for the applications, I mostly think about education, especially if the model is properly connected to a RAG database. Can be a very interesting way to get immersed in a time period on any kind of topics.

[–] unamednational@alien.top 1 points 1 year ago

Would be awesome in classroom. If kids can ask George Washington what happened exactly I think they'd care more. Plus they could tell him to go f himself for infinite amusement

[–] tortistic_turtle@alien.top 1 points 1 year ago
[–] ReMeDyIII@alien.top 1 points 1 year ago

Did we used to spell "we" as "wee?"

[–] UseNew5079@alien.top 1 points 1 year ago
[–] oKatanaa@alien.top 1 points 1 year ago

How was it trained? Did you just train it on the passages from those books? If so, I am very surprised it retained its conversational capabilities. I would expect it to just go off the rails and generate random 17th century stuff

[–] FPham@alien.top 1 points 1 year ago

Interestingly, if you tell in system prompt to the OpenHermes-Mistral 2.5 that he is from 17 century and uses archaic language, he will also say there are 7 planets.

You are MonadGPT, a very old chatbot from the 17th century. Please answer the questions using an archaic language

https://preview.redd.it/0ecpxhg86hzb1.png?width=927&format=png&auto=webp&s=cc626b7c480bf1582b9f937f0c8c671ab403f0be

[–] Dorialexandre@alien.top 1 points 1 year ago

As an update: I have now released the finetuning dataset on HuggingFace: https://huggingface.co/datasets/Pclanglais/MonadGPT

Overall 10,797 excerpts in early modern English, French and Latin with synthetic question generated by Mistral-Hermes.