Karpathy Endorses HTML Output for Large Language Models, Predicts Interactive Neural Video as Ultimate Form

According to Andrej Karpathy, OpenAI founding member and “vibe coding” concept creator, today he endorsed the Claude Code team’s approach of using HTML instead of Markdown for large language model outputs. Karpathy outlined an evolution roadmap for AI interaction interfaces: from plain text to Markdown to HTML, followed by multiple intermediate forms, ultimately reaching the final stage of interactive neural video generated directly by diffusion models.

Karpathy attributed this evolution to human brain bandwidth, noting that approximately one-third of the human brain processes visual signals in parallel—a “ten-lane highway” for information input. He argued the optimal human-AI interaction combines efficient voice for human input and high-bandwidth visual output (images, animations, or video) from AI. He recommended users immediately add “structure replies as HTML” to prompts as a near-term improvement.

Disclaimer: The information on this page may come from third parties and does not represent the views or opinions of Gate. The content displayed on this page is for reference only and does not constitute any financial, investment, or legal advice. Gate does not guarantee the accuracy or completeness of the information and shall not be liable for any losses arising from the use of this information. Virtual asset investments carry high risks and are subject to significant price volatility. You may lose all of your invested principal. Please fully understand the relevant risks and make prudent decisions based on your own financial situation and risk tolerance. For details, please refer to Disclaimer.
Comment
0/400
No comments