Can we use chain-of-thought to make an LLM that considers every sentence carefully before stating it? For instance, simply require some thinking tokens between every sentence. How would text from such a model be qualitatively different?
If you are inspired by this idea, you can reach out to the authors for collaboration or cite it:
@misc{holtzman-an-llm-thats-2026,
author = {Holtzman, Ari},
title = {An LLM That's Careful With Its Words},
year = {2026},
url = {https://hypogenic.ai/ideahub/idea/QAbWf2FN525tKq4HLs0I}
}Please sign in to comment on this idea.
No comments yet. Be the first to share your thoughts!