Hacker News new | past | comments | ask | show | jobs | submit login

Anyone know of research or tools for using an existing text generating LLM with diffusion like techniques with no new pre-training, or at most, a bit of fine-tuning, such that it works with a small GPT / Phi 3 / Gwen model, for example? I know about Tree of Thoughts with MCTS etc, that are somewhat similar (though often with a different reward learned goal) but I'm interested in something closer to token level generation. Is this possible?





Guidelines | FAQ | Lists | API | Security | Legal | Apply to YC | Contact

Search: