this post was submitted on 23 Nov 2023
1 points (100.0% liked)

LocalLLaMA

3 readers
1 users here now

Community to discuss about Llama, the family of large language models created by Meta AI.

founded 1 year ago
MODERATORS
 

Ilya from OpenAI have published a paper (2020) about Q* : a GPT-f model have capabilities in understand and resolve Math, Automated Theorem Proving.

https://arxiv.org/abs/2009.03393

When AI model can understand and really doing Math, that a critical jump.

top 16 comments
sorted by: hot top controversial new old
[–] montcarl@alien.top 1 points 11 months ago (1 children)

Where is the explicit definition and reference to Q* ??

[–] cellardoorstuck@alien.top 1 points 11 months ago

U/Xnohat - wanna answer this one?

[–] mrpogiface@alien.top 1 points 11 months ago (3 children)
[–] nested_dreams@alien.top 1 points 11 months ago (2 children)

This definitely sounds like the paper. 100% worth the read, surprised I hadn't heard much about it until this ordeal

[–] wind_dude@alien.top 1 points 11 months ago (1 children)

PRM8k, made the rounds maybe 6+ months but they never publicly released the model.

[–] dododragon@alien.top 1 points 11 months ago

I've recently just got into LLM's have you tried these math models? They seem to follow math related instructions reasonably well.

wizard-math:13b-q6_KMathLLM-MathCoder-CL-7B.Q8_0.ggufmetamath-mistral-7b.Q5_K_M.gguf

[–] jack-in-the-sack@alien.top 1 points 11 months ago

I'll definitely be asking my GPT to read this paper to me as my bedtime story.

[–] wind_dude@alien.top 1 points 11 months ago

yea, that seems to be what a few news articles have referenced.

[–] _Lee_B_@alien.top 1 points 11 months ago

Strange, I thought they would naturally be rewarding the process, by rewarding each word that's generated by the sequence to sequence model, rather than the final words, for example. Maybe they over-optimised and skipped training on all output.

[–] allende911@alien.top 1 points 11 months ago (1 children)

When AI model can understand and really doing Math, that a critical jump.

Grammarly is free, my man

[–] Klaud10z@alien.top 1 points 11 months ago (2 children)
[–] altryne@alien.top 1 points 11 months ago
[–] allende911@alien.top 1 points 11 months ago

The free version could do a better job than OP.

[–] AnomalyNexus@alien.top 1 points 11 months ago (1 children)

This doesn't smell right to me.

All references around Q* and the drama around proto-AGI...e.g. Altman talking about veil of ignorance being pulled back seem to point to something that happened in the last couple of weeks. Not 2020.

[–] BlipOnNobodysRadar@alien.top 1 points 11 months ago

If they found a proto-AGI and it was relatively trivial to implement, it would be a good idea to throw competitors off the trail with a red herring.

[–] PopeSalmon@alien.top 1 points 11 months ago

doesn't seem directly related but surely it's indirectly related,, this is an interesting idea: "We demonstrate that iteratively training a value function on statements generated by our language model leads to improved prover performance, which immediately suggests a strategy for continuous self improvement: keep training on proofs generated by the prover."