xadiant

joined 10 months ago
[–] xadiant@alien.top 1 points 9 months ago

As I understand LLMs basically write the average pattern of a billion books, so when you add gpt-4 and 3.5 data into the mix, which averages the average, things get boring very fast. For model suggestion, Yi-34b based ones look fine for literary purposes.

I think being very specific and editing (co-writing with the model) could help. Some LoRA training on specific books could be helpful to mimic a certain style.

High temperature and repetition penalty could help too.

[–] xadiant@alien.top 1 points 9 months ago

High repetition penalty? One model I merged suddenly started speaking Spanish in one summarisation task lol

[–] xadiant@alien.top 1 points 9 months ago

Really cool, will check the video out. Since we found an actually qualified person though, let me ask a few layman questions, hope you have time to answer them!

sampling methods. Most of them look simple, but we still don't really know how to tune them. Do you think novel sampling methods or specific combinations could improve output quality by a lot?

For instance, beam search. Does beam search provide a linear improvement in quality as you go up or not?

Do you think ideal numbers for temperature, top_k and top_p are context or model based, or both?

 

I want to see if some presets and custom modifications work well in benchmarks, but running HellaSwag or MMLU looks too complicated for me, and it takes 10+ hours to upload 20GBs of data.

I assume there isn't a convenient webui for chumps to run benchmarks with (apart from ooba perplexity, which I assume isn't the same thing?). Any advise?

[–] xadiant@alien.top 1 points 9 months ago

Exactly what I was thinking. I just fail miserably each time I merge the layers.

[–] xadiant@alien.top 1 points 9 months ago (4 children)

Any tips/attempts on frankensteining 2 yi-34b models together to make a ~51B model?

[–] xadiant@alien.top 1 points 10 months ago (1 children)

That's acceptable. Did you full train or fine-tune though? And how much data?

[–] xadiant@alien.top 1 points 10 months ago (4 children)

Goddammit, I just fine-tuned Tortoise with custom voice. Can't wait for webui's for the StyleTTS. Hope it's easy to fine-tune

[–] xadiant@alien.top 1 points 10 months ago
[–] xadiant@alien.top 1 points 10 months ago (1 children)

Sure, it's just going to generate 5 tokens per week

[–] xadiant@alien.top 1 points 10 months ago (5 children)

No fucking way. GPT-3 has 175B params. In no shape or form they could have discovered the "secret sauce" to make an ultra smart 20B model. TruthfulQA paper suggests that bigger models are more likely to score worse, and ChatGPT's TQA score is impressively bad. I think the papers responsible for impressive open-source models are max 12-20 months old. Turbo version is probably quantized, that's all.