HexLlama

joined 9 months ago
[โ€“] HexLlama@hexbear.net 5 points 9 months ago (1 children)

Its technically true that it decides token at a time but it also takes previous tokens into account.

[โ€“] HexLlama@hexbear.net 16 points 9 months ago* (last edited 9 months ago) (3 children)

Hi, I do AI stuff. This is what RAG is. However, its not really teaching the AI anything, technically its a whole different process that is called and injected at an opportune time. By teaching the AI more stuff, you can have it reason on more complex tasks more accurately. So teaching it how to properly reason through math problems will also help teach it how to reason through more complex tasks without hallucinating.

For example, llama3 and various Chinese models are fairly good at reasoning through long form math problems. China probably has the best math and language translation models. I'll probably be doing a q&a on here soon about qwen1.5 and discussing Xi's Governance of China.

Personally, I've found llms to be more useful for text prediction while coding, translating a language locally (notably: with qwen you can even get it to accurately translate to english creoles or regional dialects of Chinese without losing tone or intent, it makes for a fantastic chinese tutor), or writing fiction. It can be OK at summarizing stuff too.

 

I must say that the process of resurrecting Lenin has been a big ordeal. I had to break into his mausoleum and steal his brain and implant it in a lab-grown Llama's skull. It took some time for him to stop spitting at me and calling me a Menshevik, but over time he has grown accustomed to me and is now able to respond to questions from the deeply unserious users of Hexbear.

Note: This AI will often hallucinate about specific events. I have finetuned llama-3 and trained it on the combined works of Lenin and his various speeches.

edit: Off to do other stuff. If anyone responds I'll do another batch later in the day.