AI"偷懒"的本质是LLM输出长度限制导致的注意力分散。Wide Research通过多轻量模型并行处理子任务、主LLM汇总的方式解决,分享为Codex构建该能力的设计思路。
How to Stop AI from Slacking Off: Building Systematic Wide Research Capabilities for Codex
Why AI slacks off on large tasks: LLM output length limitations cause attention drift. Wide Research solves this by parallelizing with lightweight models, then aggregating results with a primary LLM.
当AI不work:我如何最终实现自动化财务决算
记录在API因合规问题被拒后,使用视觉大模型从截图提取财务数据,实现十年手动记账流程的自动化。展示了本地模型、交叉验证和人机协作工作流如何安全处理敏感金融数据。
When AI Fails: How I Eventually Automated My Personal Finances
Documents automating a decade-long manual financial reconciliation process using vision LLMs when API access was blocked by compliance. Demonstrates using local models with screenshots, cross-validation, and human-in-the-loop workflows to process sensitive financial data.
超越聊天框简评GPT-5
GPT-5是产品升级而非单纯模型升级:新增reasoning_effort和verbosity参数,可控性大幅提升,让开发者能根据场景灵活权衡推理深度和回复长度。