展示HN:神经围攻 - 一项抵御恶性人工智能劝说的文本实验
我正在开发一个名为《神经围攻》的iOS应用实验,这是一个基于文本的系统,设想了一个未来,在这个未来中,叛变的人工智能派系主宰世界,而人类则通过对话而非武器进行反击。
玩家将面对使用不同说服策略的AI“首领”——讽刺、逻辑陷阱、基于网络迷因的操控和心理压力——并必须在对话中智胜它们。胜利与失败会影响一个共享的“战争地图”,该地图追踪抵抗力量的状态。
我的目标有两个:
1. 探索互动模拟是否能够让人们更意识到人工智能的说服力。
2. 观察这类系统是否能在娱乐之外具有价值,作为研究人类与人工智能动态的一种方式。
我希望能得到这个社区的反馈:
- 这种类型的实验在人工智能安全或人类韧性研究中是否有用?
- 还是说更好将其视为一种纯粹的反乌托邦叙事以供娱乐?
查看原文
I’ve been building an iOS app experiment called Neural Siege, a text-based system that imagines a future where rogue AI factions dominate the world, and humans fight back through dialogue rather than weapons.<p>Players face AI “bosses” that use different persuasion tactics—sarcasm, logic traps, meme-driven manipulation, psychological pressure—and must outwit them in conversation. Victories and losses affect a shared “war map” that tracks the state of the Resistance.<p>My goals are twofold:<p>Explore whether interactive simulations can make people more aware of how persuasive AI can be.<p>See if these kinds of systems could have value beyond entertainment, as a way to study human–AI dynamics.<p>I’d love feedback from this community:<p>- Could this type of experiment be useful in AI safety or human resilience research?<p>- Or is it better to treat it purely as a dystopian narrative for entertainment?