Teaching National Security Policy with AI

48 enescakir 21 6/10/2025, 1:54:25 PM steveblank.com ↗

Comments (21)

suddenlybananas · 1d ago
>Policy students have to read reams of documents weekly. Our hypotheses was that our student teams could use AI to ingest and summarize content, identify key themes and concepts across the content, provide an in-depth analysis of critical content sections, and then synthesize and structure their key insights and apply their key insights to solve their specific policy problem.

Yeah who cares about actually reading and properly understanding anything at all. Given the policy world is filled with so much BS, no wonder they like a BS machine.

alephnerd · 1d ago
Enhanced information retrieval is a good tool to have - at some point close reading does become difficult to scale out.

Building experience on how to use tools to automated expected drudgery like making PPT slides or wordsmithing an NSC memo is a good skill to build.

There is a lot of low hanging fruit in professional tooling that can and should be automated where possible, and some class similar to the "Missing Semester" at MIT except oriented towards productivity tools would be helpful.

FuriouslyAdrift · 1d ago
Synthesis and summarization is literally the main job of an analyst. Frequently the real information is hidden in the tone, tenor, and syntax, not necessarily in the broader content (aka reading between the lines).
neilv · 1d ago
Reading between the lines is also a skill in general human communication.

Which is why, when someone sends me an AI-generated message that previously would've been written by them, it's like they're jamming one of my skills.

Not only are they not giving me some information I had before (e.g., that the person thought of this aspect to mention, that they expressed it this way, that they invested this effort into this message to me, etc.), but, (if I don't know it's AI-generated) the message is giving me wrong information about all those things I read into it.

(I'm reasonably OK at reading between the lines, for someone with only basic schooling. Though sometimes I'm reminded that some of my humanities major friends are obviously much better at interpreting and expressing. Maybe they're going to be taking the AI-slop apocalypse even worse than I do.)

vouaobrasil · 1d ago
I don't think it is when the students are random variables, because enhanced information retrieval will increase the proportion of the lazy in the class.
radioactivist · 1d ago
At one point this states:

> Claude was also able to create a list of leaders with the Department of Energy Title17 credit programs, Exim DFC, and other federal credit programs that the team should interview. In addition, it created a list of leaders within Congressional Budget Office and the Office of Management and Budget that would be able to provide insights. See the demo here:

and then there is a video of them "doing" this. But the video basically has Claude just responding saying "I'm sorry I can't do that, please look at their website/etc".

Am I missing something here?

radioactivist · 1d ago
It happens again in the next video. It says:

> The team came up with a use case the teaching team hadn’t thought of – using AI to critique the team’s own hypotheses. The AI not only gave them criticism but supported it with links from published scholars. See the demo here:

But the video just shows Claude giving some criticism but then just says go look at some journals and talk to experts (doesn't give any references or specifics).

kenjackson · 1d ago
That was really weird. I did do this with ChatGPT 4o and it seems to do a good job of creating this list. But I don't know anything about this field, so I don't know how accurate it is.
cptroot · 1d ago
This article says "the students did X", without providing any metrics to compare the result on. It's frustrating to again and again get articles saying "AI is great and speeds learning" without actually evaluating that learning process.
troelsSteegin · 1d ago
What's missing from this is the "before and after" - how this quarter's class experience was different from previous quarters without the AI tool emphasis.
mapt · 1d ago
The very first thing you have to learn about original research is the basis of the experimental scientific method, of the idea of empiricism and improvement through reason, observation, and iterative comparative testing. It is a little bit shocking when you encounter the broad swath of the population that has not internalized this.
bjelkeman-again · 1d ago
It feels like the tools are used as a shortcut to not read documents, and then have the tools produce output from the shortcut taken. What did they accentually learn that they will retain afterwards?
OWaz · 1d ago
I find it perplexing how people are so open to just dumping personal effort onto these tools and believing the tools work accurately.
sureokbutyeah · 1d ago
Work accurately? Relative to what? Old humans? Make up something about psychology? Physics? Economics? History? Academics have been doing that for years and we all blindly agreed their work was accurate, lauded them, then found out decades later it was garbage.

Seems typical for humans; centuries of false belief religion was accurate, now contemporary nation state politics, economics, and the engineered things they sell for profit.

So long as enough stuff is available on shelves to keep people sedate, they'll believe whatever. Our biology couples us to knowing when we need food, water; keep those normal and no one cares about anything else. Riots only occur when biology is threatened. Everything else about humanity is 100% made up false belief, appeals to empty trust in what we say.

Physics makes it pretty clear its all just skins suits pulling illusions out their ass all the way down. We can never change the immutable forces of physics, there's too much other stuff in universe rushing in to correct. This is it for humans; idle about on Earth hallucinating.

jay_kyburz · 22h ago
I can agree on Psychology, Economics, and History, but most of Physics is reproducible science.

I think, now more than ever, we need to clearly distinguish reproducible science from untested hypothesise. Reality vs Opinion.

update: opinion is not quite the right word here. Perhaps somebody else can think of a better word.

sarchertech · 1d ago
So if you’re reading a summary of a bullshit document written by an old human, created by a machine trained on billions of bullshit documents written by old humans, what do you get out of that?
bgwalter · 1d ago
Just read Mearsheimer and the think tank policy papers if you want to know what is actually going on. Go to the Stanford Hoover Institute if you want to sell what is actually going on to the American public.

Why would LLMs help, unless trained on classified information for which you could also use an internal search engine? In the end it comes down to how much military, economic and propaganda power you have and how much you are willing to deploy it.

The whole interaction with LLMs, which focuses on clicking, wrestling with a stupid and recalcitrant dialogue partner distracts from thinking. Better read original information yourself and take a long walk to organize it in your own mind.

formerphotoj · 1d ago
And after the walk, talk with an intelligent human dialog partner to exchange ideas and concepts that illuminate the schemas. Heck, walk and talk together! :)
psunavy03 · 23h ago
Someone apparently is taking the old war college joke about "it's only a lot of reading if you do it" a little too seriously . . .
einpoklum · 1d ago
National Security and AI -

Two domains which are rife with hype, and self-serving self-nominated experts, and are both put to use for manipulating the public for questionable purposes.

sidewndr46 · 1d ago
A most perfect union?