Article
8 minutes
read

The Selectric Trap: Why AI’s Polish Keeps Your Team Stuck

AI for L&D
Performance Support
Upskilling and Reskilling

When I was young, I learned to type on a manual typewriter. Every mistake meant reaching for liquid white-out or correction tape. I could see exactly where I'd messed up thanks to the little bumps and shadows on the page.

Then I got an IBM Selectric with its correction feature, and something fascinating happened: When I made a mistake and corrected it, the page looked perfect…and I stopped reading my work critically.

The absence of visible errors made my brain treat everything as correct. I'd miss logical flaws, awkward phrasing, and even factual mistakes—things I would have caught immediately before. The polish created a false sense of quality; now, years later, I’m watching the same phenomenon happen with AI.

Why AI’s Polish Creates a Productivity Trap

The problem isn’t what you think. When leaders tell me their teams aren't getting value from AI, they think it's a skills problem. They frame it as “We need better prompt engineering training," or "People don't understand the tools." 

But I see something else. Their people are stuck in two mutually reinforcing AI productivity traps:

  1. They use AI transactionally—ask a question, get an answer, copy-paste, done.
  2. They trust AI output too much because it looks perfectly polished, including formatting, punctuation, transitions, and fluid language that never manages to say much at all.

When you use AI transactionally, you never develop the judgment you need to evaluate its output. And when you trust the polish too much, you never push back enough to have the human-AI conversation that makes the AI tool worth using. 

The result? Teams generate confident-sounding content that doesn't actually innovate, advance their work, or add value.

The Problem with “Perfect” AI Drafts

With AI, there’s no such thing as a rough draft. It gives you complete sentences, proper formatting, confident

assertions, and professional-sounding language. There are no typos, crossed-out words, or insertions. It looks right. And that shuts down our critical thinking. 

Think about how you'd respond if a colleague sent you a messy draft with handwritten notes in the margins and "I'm not sure about this part" scrawled at the top. You'd read it critically, following the thought process that unfolds on the page. In doing so, you'd build on the ideas. You’d question the logic. You'd push back. In short, you’d have a conversation. 

Now imagine the same colleague sent you a polished document with perfect formatting, confident language, and a tidy presentation. The polish signals "this is done" rather than "this needs more thought." Our brains have a cognitive bias toward viewing the polished version as final, even when the thinking is shallow or absent.

Moving Beyond Transactional AI

Here's the kind of AI use I see all too often from well-meaning L&D folks: “Give me three learning objectives for a course on delegation.”

In response, AI dutifully delivers three objectives that sound professional and are nicely formatted. Transaction complete. Next task. 

But those objectives are probably generic. They probably don’t align with actual business outcomes. They probably address delegation in theory while missing the specific challenges this team faces.

The unwitting user has no way to know because they never pushed further. They asked for three

objectives and got three objectives. They never questioned whether these were the right objectives, and they missed the opportunity to explore alternatives, provide context, iterate, refine, or test whether the AI output actually solves the problem. 

What Actually Works: The AI Thought Partner Approach

The teams getting real value from AI work with it completely differently. They treat it like an AI thought partner—someone smart but in need of context, constraints, and a push to do their best work. 

Here’s what AI thought partnership looks like for these teams: 

  • They start with context, not just requests.
  • They treat the first response as a rough draft, because it rarely captures the full picture.
  • They iterate based on what they know that AI doesn't.
  • They use AI to explore alternatives, not just execute tasks.
  • They evaluate AI outputs critically, asking whether it addresses the need or instead constitutes well-formatted, generic “workslop.”

Using AI as a thought partner isn’t more difficult than transactional use; it’s just a different strategy. It requires you to think while you work with AI rather than outsourcing the thinking entirely.

Breaking the Pattern: Strategies for L&D Teams

To avoid the Selectric effect, you have to make it visible. 

  1. Show examples of AI content that looks perfect but is shallow or generic. Train people to see beyond the formatting.
  2. Model iterative use. Let them see that the first AI response is a starting point, not a final product. Show how much output improves when you treat the interaction with AI as a conversation.
  3. Create space for exploration. If your organizational culture measures productivity primarily by output volume, you'll get polished mediocrity. Legitimize the time spent iterating and refining.

Teams who are able to break through the Selectric effect aren’t simply AI power users—they’re better thinkers. 

By questioning AI outputs and thinking critically about whether they address the need, they become better at questioning their own thinking. In the process of providing clear context for their AI thought partners, they improve their ability to articulate their own implicit knowledge and ideas. 

AI is a true thought partner not because it thinks like a fellow human, but because working with it well requires us to think more clearly ourselves. A truly effective human-AI pairing is founded in what we call the unpromptable: inimitable human skills like insight, judgment, and vision. 

Got a vision of your own for adding AI literacy and thought partnership to your people’s human skills? We’d love to talk shop! Reach out to explore the possibilities.

Contributors
Clare Dygert

Subscribe to our newsletter

Connect with our industry experts on the subjects that matter most to you.

Clare Dygert

SweetRush Newsletter

Keep pace with your peers— get the latest L&D innovations and insigths!

By subscribing you agree to with our Privacy Policy.

Thanks for sharing with us!
We appreciate your interest in SweetRush.
Soon, your inbox will receive the wisdom of the ages—the modern, culture, and disruptive ages, that is.

Oops! Something went wrong while submitting the form.