AUTHOREA
Log in Sign Up Browse Preprints
LOG IN SIGN UP
Katheryne Laurent
Katheryne Laurent

Public Documents 1
Optimizing Large Language Models Through Highly Dense Reward Structures and Recursive...
Katheryne Laurent
Owen Blanchard

Katheryne Laurent

and 2 more

September 17, 2024
Optimizing language models to achieve both high performance and efficiency remains a complex challenge due to the vast parameter spaces and the intricacies of text generation. Introducing a novel approach that leverages highly dense reward structures combined with a recursive thought process through Monte Carlo Tree Search (MCTS) offers significant improvements in model training and generative capabilities. The enhanced LLaMA model benefits from granular feedback provided through dense rewards, allowing it to better navigate the high-dimensional space and converge more rapidly to optimal solutions. The recursive evaluation mechanism of MCTS further enables the model to make informed decisions through an iterative exploration of potential text sequences, resulting in more contextually coherent outputs. Experiments conducted on diverse datasets demonstrate the model's superior performance in terms of BLEU, ROUGE, and METEOR scores, as well as its adaptability to unseen linguistic contexts. The modified LLaMA model exhibits faster convergence rates and more robust generalization compared to baseline models, indicating the effectiveness of the proposed optimization techniques in enhancing language model performance.

| Powered by Authorea.com

  • Home