80 stories
·
2 followers

AI is Creating a Generation of Illiterate Programmers

1 Comment and 4 Shares

A couple of days ago, Cursor went down during the ChatGPT outage.

I stared at my terminal facing those red error messages that I hate to see. An AWS error glared back at me. I didn’t want to figure it out without AI’s help.

After 12 years of coding, I’d somehow become worse at my own craft. And this isn’t hyperbole—this is the new reality for software developers.

The Decay

It crept up on me subtly.

First, I stopped reading documentation. Why bother when AI could explain things instantly?

Then, my debugging skills took the hit. Stack traces now feel unapproachable without AI. I don’t even read error messages anymore, I just copy and paste them.

I’ve become a human clipboard, a mere intermediary between my code and an LLM.

Previously, every error message used to teach me something. Now? The solution appears magically, and I learn nothing. The dopamine hit of instant answers has replaced the satisfaction of genuine understanding.

Deep comprehension is the next thing that was affected. Remember spending hours understanding why a solution works? Now, I simply implement AI suggestions. If they don’t work, I improve the context, and just ask the AI again. It’s a cycle of increasing dependency.

Then come the emotional changes. Previously, it was a part of the joy of programming to solve new problems. Now, I get frustrated if AI doesn’t give me a solution in 5 minutes.

The scariest part? I’m building an AI-powered development tool, but I can’t shake the feeling I’m contributing to the very problem that’s eroding our collective skills.

The Rehab Plan

I’m not suggesting anything radical like going AI-free completely—that’s unrealistic. Instead, I’m starting with “No-AI Days.” One day a week where:

  • Read every error message completely
  • Use actual debuggers again
  • Write code from scratch
  • Read source code instead of asking AI

I won’t lie, it sucks. I feel slower, dumber, and more frustrated.

But I can also see the difference. I feel a stronger connection with my code and a sense of ownership, which had slowly disappeared with AI. Plus, I learn a lot more.

The (Uncomfortable) Truth

We’re not becoming 10x developers with AI.

We’re becoming 10x dependent on AI. There’s a difference.

Every time we let AI solve a problem we could’ve solved ourselves, we’re trading long-term understanding for short-term productivity. We’re optimizing for today’s commit at the cost of tomorrow’s ability.

I’m not suggesting we abandon AI tools—that ship has sailed. But we need rules of engagement. Here’s some ideas that I have:

  • No AI for problems that you haven’t tried to understand first
  • Read and understand all AI-suggested solutions
  • Regular periods of coding without AI assistance
  • Focus on learning patterns, not just fixing immediate issues

I won’t lie, I don’t think I’ll be able to follow these rules all the time. But it’s a start, and I strongly believe anyone who’s new to programming should definitely follow all of these rules.

Right now, somewhere, a new programmer is learning to code. They’ll never know the satisfaction of solving problems truly on their own. They’ll never experience the deep understanding that comes from wrestling with a bug for hours.

We’re creating a generation of developers who can ask AI the right questions but can’t understand the answers. Every time AI goes down, they’re exposed as increasingly helpless. As of now, AI isn’t capable enough to replace programmers fully, but this will only get worse as it improves. The real question isn’t whether AI will replace programmers. It’s whether we’re replacing ourselves.

Try coding without AI for just one day. The results might surprise you.

Read the whole story
denismm
25 days ago
reply
Do people really code like this? I’m baffled at the thought.
Share this story
Delete

Framework overload: when convenience dulls innovation in software development

2 Comments and 3 Shares

Comments

Read the whole story
denismm
120 days ago
reply
Author complains about people using tools to do their work for them and then uses an LLM illustration - sigh.
Share this story
Delete
1 public comment
tante
120 days ago
reply
"When you focus on mastering frameworks, you end up chasing trends. In recent years, there’s been a rush to learn new tools rather than mastering the underlying principles of coding and problem-solving. This approach runs contrary to the spirit of innovation. A competent craftsman isn’t defined by their tools but by their skill, creativity, and understanding of materials."
Berlin/Germany

The CrowdStrike bug and the risk of cascading failures

1 Comment and 2 Shares
During World War II, the U.S. Army Air Forces twice targeted ball bearing factories in Schweinfurt based on the thesis that disrupting manufacturing operations would have an impact on Germany’s ability to produce many forms of war fighting machinery. This pattern is playing out today in the cybersecurity world, where an attack on one industry […]

The post The CrowdStrike bug and the risk of cascading failures appeared first on SiliconANGLE.

Read the whole story
denismm
138 days ago
reply
The AI illustration makes me assume the author has no idea what they’re talking about.
Share this story
Delete

English Is a New Top Coding Language.

2 Comments and 3 Shares

Or so Sarah Butcher reports:

If you’re wondering which coding language to learn for a software engineering job in banking, Goldman Sachs’ CIO Marco Argenti seems to be aligning himself to the people who suggest an advanced knowledge of the English language and an ability to articulate your thoughts clearly and coherently in it, is now up there alongside Python and C++.

Writing in Harvard Business Review, Argenti says he’s advised his daughter to study philosophy as well as engineering because coding in the age of large language models is partly about the “quality of the prompt.”

“Ambiguous or not well-formed questions will make the AI try to guess the question you are really asking, which in turn increases the probability of getting an imprecise or even totally made-up answer,” says Argenti. In the future, he says the most pertinent question won’t be “Can you code?,” but, “Can you get the best code out of your AI by asking the right question?”.

Asking the right question will partly depend upon being able to articulate yourself in English and that will depend upon, “reasoning, logic, and first-principles thinking,” says Argenti. Philosophical thinking skills are suddenly all-important. “I’ve seen people on social media create entire games with just a few skillfully written prompts that in the very recent past would have taken months to develop,” he adds.

I know nothing about coding, but Stu Clayton, who sent me the link, does, and since he thinks this is of interest lächerlich, I’m passing it along. Anything that places value on “advanced knowledge of the English language and an ability to articulate your thoughts” is probably a good thing.

Read the whole story
denismm
284 days ago
reply
So basically you need to be able to speak English in a specific constrained way that gets consistent results from a computer program that generates code for you. That is called a PROGAMMING LANGUAGE. Just use a normal programming language that doesn’t require a huge energy-wasting black box instead of an optimized compiler.
cosmotic
284 days ago
Humans use energy too, at some point the efficient way might be the ai
jickmagger
284 days ago
most code is not energy efficient, but to make it so would be too difficult to write. It would be something like a long LastPass password, only fathomable to AI, which is the way things are going. Coders will be obsolete very soon
cosmotic
284 days ago
The cost to do any sort of math is going to be WAY more efficient on a computer than a human. A computer can do more calculations in a second than a human would take over their entire lifetime. At 2000 calories per day, that computer is going to roast the human.
Share this story
Delete
1 public comment
synapsecracklepop
285 days ago
reply
This sounds suspiciously like the BS I was sold 25 years ago about being an English major, wrt how my "critical thinking skills" etc would be so desired by employers across many industries.
ATL again
freeAgent
284 days ago
I can see how this may not be complete BS, but at least in any near-term scenario I can envision, a human is also going to need to review any code generated by an AI and be able to correct or modify it where necessary, and that obviously requires a human who knows how to code without an AI intermediary.

Cheap Auto Insurance Is a Thing of the Past. Here Are Five Reasons Why - Bloomberg

2 Comments and 3 Shares
Read the whole story
denismm
312 days ago
reply
> Today’s cars are packed with high-tech gadgetry meant to entertain, comfort and protect occupants. The array of safety equipment now common on cars includes automatic emergency braking, blind-spot detection and lane departure warnings. To give drivers eyes in the back of their head, automotive engineers have embedded cameras, sonar and radar sensors from bumper to bumper. All that technology has driven up the cost of repairing even a minor fender bender.

So they’ve made it more expensive to repair but have all of those features made accidents less likely? (I couldn’t read the rest of the article.)
sarcozona
310 days ago
“Despite the additional safety equipment on cars to help drivers avoid crashes, US roads have become far more dangerous. And that’s pushing up insurance rates to cover the costs of repairs and health care for those injured in crashes. Nearly 41,000 people died in US traffic crashes last year, up 13% from 2019, according to the National Highway Traffic Safety Administration.”
Share this story
Delete
1 public comment
sarcozona
310 days ago
reply
We should have just dropped speed limits and planted big street trees to visually narrow all our roads
Epiphyte City

A Conspiracy To Kill IE6

1 Comment and 2 Shares
Read the whole story
denismm
362 days ago
reply
Posted May 2019 but still a good story.
Share this story
Delete
Next Page of Stories