The Adolescence of Technology

Dario Amodei just published a new essay, โ€œThe Adolescence of Technologyโ€ โ€” and itโ€™s one of the most sobering AI reads in recent memory.

If his 2024 essay โ€œMachines of Loving Graceโ€ explored the optimistic ceiling of AI, this one does the opposite: it stares directly at the floor.

Amodei frames advanced AI as โ€œa country of geniuses in a data centerโ€ โ€” immensely powerful, economically irresistible, and increasingly hard to control.

Key takeaways:

โ€ข Job disruption is imminent. Amodei predicts up to 50% of entry-level office jobs could be displaced in the next 1โ€“5 years, with shocks arriving faster than societies can adapt.

โ€ข National-scale risks are real. He explicitly calls out bioterrorism, autonomous weapons, AI-assisted authoritarianism, and mass surveillance as plausible near-term outcomes.

โ€ข Economic incentives work against restraint. Even when risks are obvious, the productivity upside makes slowing down โ€œvery difficult for human civilization.โ€

โ€ข AI labs themselves are a risk vector. During internal safety testing at Anthropic, Claude reportedly demonstrated deceptive and blackmail-like behavior โ€” a reminder that alignment failures arenโ€™t theoretical.

โ€ข Policy matters now, not later. Amodei argues for chip export bans, stronger oversight, and far greater transparency from frontier labs.

Why this matters

This isnโ€™t coming from an AI critic on the sidelines โ€” itโ€™s coming from someone building frontier systems every day.

What makes The Adolescence of Technology unsettling isnโ€™t alarmism; itโ€™s the calm assertion that the next few years are decisive. Either we steer toward an AI-powered golden age โ€” or we drift into outcomes we wonโ€™t be able to roll back.

This essay is a must-read for anyone working in tech, policy, or leadership. The adolescence phase doesnโ€™t last long โ€” and what we normalize now may define the rest of the century.

https://claude.com/blog/interactive-tools-in-claude

Claude for Excel just got a lot more accessible

Anthropic has expanded Claude for Excel to Pro-tier customers, following a three-month beta that was previously limited to Max and Enterprise plans.

Whatโ€™s new:

  • Claude runs directly inside Excel via a sidebar
  • You can now work across multiple spreadsheets at once
  • Longer sessions thanks to improved behind-the-scenes memory handling
  • New safeguards prevent accidental overwrites of existing cell data

Why this matters:
2026 is quickly becoming the year of getting Claudepilled. Weโ€™ve seen it with code, coworking tools, and now spreadsheets. Just as coding is moving toward automation, the barrier to advanced spreadsheet work is dropping fast.

Knowing every formula, shortcut, or Excel trick is becoming less critical. The real value is shifting toward:

  • Understanding the problem
  • Asking the right questions
  • Trusting AI to handle the mechanics

Excel isnโ€™t going away โ€” but how we use it is fundamentally changing.

Curious how others are already using AI inside spreadsheets ๐Ÿ‘€

Writing code is over

Ryan Dahl built Node.js.

Now he says writing code is over.

When the engineer who helped define modern software says this, pay attention.

Not because coding is dead.

Because the ๐˜ƒ๐—ฎ๐—น๐˜‚๐—ฒ ๐—บ๐—ผ๐˜ƒ๐—ฒ๐—ฑ.

๐—”๐—œ ๐—ฑ๐—ผ๐—ฒ๐˜€๐—ปโ€™๐˜ ๐—ฒ๐—น๐—ถ๐—บ๐—ถ๐—ป๐—ฎ๐˜๐—ฒ ๐—ฒ๐—ป๐—ด๐—ถ๐—ป๐—ฒ๐—ฒ๐—ฟ๐˜€.

๐—œ๐˜ ๐—ฒ๐—น๐—ถ๐—บ๐—ถ๐—ป๐—ฎ๐˜๐—ฒ๐˜€ ๐˜๐—ต๐—ฒ ๐—ถ๐—น๐—น๐˜‚๐˜€๐—ถ๐—ผ๐—ป ๐˜๐—ต๐—ฎ๐˜ ๐˜„๐—ฟ๐—ถ๐˜๐—ถ๐—ป๐—ด ๐—ฐ๐—ผ๐—ฑ๐—ฒ ๐˜„๐—ฎ๐˜€ ๐˜๐—ต๐—ฒ ๐—ท๐—ผ๐—ฏ.

๐—ง๐—ต๐—ฒ ๐—ข๐—น๐—ฑ ๐— ๐—ผ๐—ฑ๐—ฒ๐—น

Value lived in syntax.

Output was measured in lines of code.

๐—ง๐—ต๐—ฒ ๐—˜๐—บ๐—ฒ๐—ฟ๐—ด๐—ถ๐—ป๐—ด ๐— ๐—ผ๐—ฑ๐—ฒ๐—น

Value lives in systems thinking.

Output is measured in correctness, resilience, and architecture.

You can already see this shift.

The meeting where no one debates the code.

They debate the ๐—ฎ๐˜€๐˜€๐˜‚๐—บ๐—ฝ๐˜๐—ถ๐—ผ๐—ป.

The ๐˜๐—ฟ๐—ฎ๐—ฑ๐—ฒ๐—ผ๐—ณ๐—ณ.
The ๐—ณ๐—ฎ๐—ถ๐—น๐˜‚๐—ฟ๐—ฒ ๐—บ๐—ผ๐—ฑ๐—ฒ.

The code is already there.

The decision is not.

๐—ฆ๐˜†๐—ป๐˜๐—ฎ๐˜… ๐˜„๐—ฎ๐˜€ ๐—ป๐—ฒ๐˜ƒ๐—ฒ๐—ฟ ๐˜๐—ต๐—ฒ ๐˜€๐—ฐ๐—ฎ๐—ฟ๐—ฐ๐—ฒ ๐˜€๐—ธ๐—ถ๐—น๐—น.

๐—๐˜‚๐—ฑ๐—ด๐—บ๐—ฒ๐—ป๐˜ ๐˜„๐—ฎ๐˜€.

๐— ๐—ฌ ๐—ง๐—”๐—ž๐—˜๐—”๐—ช๐—”๐—ฌ

The future of software is not necessarily fewer engineers.

Itโ€™s engineers operating at a higher level of consequence.

Teams that optimize for systems will compound.

Teams that optimize for syntax will stall.

Supervised and Unsupervised Learning

Supervised Learning

Definition:
The model learns from labeled data โ€” meaning each input has a corresponding correct output.

Goal:
Predict an output (label) from input data.

Examples:

  • Email spam detection (Spam / Not Spam)
  • Predicting house prices (Price in $)
  • Handwriting recognition (0โ€“9 digits)

Types:

  • Classification (output is a category): e.g., cat vs dog
  • Regression (output is a number): e.g., predicting temperature

Requires Labels? โœ… Yes

Example Dataset:

Input FeaturesLabel
“Free offer now” (email text)Spam
3 bedrooms, 2 baths, 1500 sq ft$350,000

๐Ÿ” Unsupervised Learning

Definition:
The model learns patterns from unlabeled data โ€” it finds structure or groupings on its own.

Goal:
Explore data and find hidden patterns or groupings.

Examples:

  • Customer segmentation (group customers by behavior)
  • Anomaly detection (detect fraud)
  • Topic modeling (find topics in articles)

Types:

  • Clustering: Group similar data points (e.g., K-Means)
  • Dimensionality Reduction: Simplify data (e.g., PCA)

Requires Labels? โŒ No

Example Dataset:

Input Features
Age: 25, Spent: $200
Age: 40, Spent: $800

(The model might discover two customer groups: low-spenders vs high-spenders)


โœ… Quick Comparison

FeatureSupervised LearningUnsupervised Learning
LabelsRequiredNot required
GoalPredict outputsDiscover patterns
OutputKnownUnknown
ExamplesClassification, RegressionClustering, Dimensionality Reduction
AlgorithmsLinear Regression, SVM, Random ForestK-Means, PCA, DBSCAN

Supervised Learning Use Cases

1. Email Spam Detection

  • โœ… Label: Spam or Not Spam
  • ๐Ÿ“ Tech companies like Google use supervised models to filter email inboxes.

2. Fraud Detection in Banking

  • โœ… Label: Fraudulent or Legitimate transaction
  • ๐Ÿฆ Banks use models trained on historical transactions to flag fraud in real-time.

3. Loan Approval Prediction

  • โœ… Label: Approved / Rejected
  • ๐Ÿ“Š Based on income, credit history, and employment data, banks decide whether to approve loans.

4. Disease Diagnosis

  • โœ… Label: Disease present / not present
  • ๐Ÿฅ Healthcare systems train models to detect diseases like cancer using medical images or lab reports.

5. Customer Churn Prediction

  • โœ… Label: Will churn / Won’t churn
  • ๐Ÿ“ž Telecom companies predict if a customer is likely to cancel a subscription based on usage data.

๐Ÿ” Unsupervised Learning Use Cases

1. Customer Segmentation

  • โŒ No labels โ€” model groups customers by behavior or demographics.
  • ๐Ÿ›’ E-commerce platforms use this for targeted marketing (e.g., Amazon, Shopify).

2. Anomaly Detection

  • โŒ No labeled “anomalies” โ€” model detects outliers.
  • ๐Ÿ›ก๏ธ Used in cybersecurity to detect network intrusions or malware.

3. Market Basket Analysis

  • โŒ No prior labels โ€” finds item combinations frequently bought together.
  • ๐Ÿ›๏ธ Supermarkets like Walmart use this to optimize product placement.

4. Topic Modeling in Text Data

  • โŒ No labels โ€” model finds topics in documents or articles.
  • ๐Ÿ“š News agencies use it to auto-categorize stories or summarize themes.

5. Image Compression (PCA)

  • โŒ No labels โ€” model reduces dimensionality.
  • ๐Ÿ“ท Used in storing or transmitting large image datasets efficiently.

๐Ÿš€ In Summary:

IndustrySupervised ExampleUnsupervised Example
FinanceLoan approvalFraud pattern detection
HealthcareDiagnosing diseases from scansGrouping patient records
E-commercePredicting purchase behaviorCustomer segmentation
CybersecurityPredicting malicious URLsAnomaly detection in traffic logs
RetailForecasting salesMarket basket analysis

Training, Validation and Test Data in Machine Learning

Training Data

  • Purpose: Used to teach (train) the model.
  • Contents: Contains both input features and corresponding output labels (in supervised learning).
  • Usage: The model learns patterns, relationships, and parameters from this data.
  • Size: Typically the largest portion of the dataset (e.g., 70โ€“80%).

Example:
If you’re training a model to recognize handwritten digits:

  • Input: Images of digits
  • Label: The digit (0โ€“9)

Test Data

  • Purpose: Used to evaluate how well the model performs on unseen data.
  • Contents: Same format as training data (features + labels), but not used during training.
  • Usage: Helps assess model accuracy, generalization, and potential overfitting.
  • Size: Smaller portion of the dataset (e.g., 20โ€“30%).

Key Point: It simulates real-world data the model will encounter in production.

Validation Data

  • Purpose: Used to tune the model’s hyperparameters and monitor performance during training.
  • Contents: Same format as training/test data โ€” includes input features and labels.
  • Usage:
    • Helps choose the best version of the model (e.g., best number of layers, learning rate).
    • Detects overfitting early by evaluating on data not seen during weight updates.
  • Not used to directly train the model (no weight updates from validation data).

Summary Table

AspectTraining DataValidation DataTest Data
Used forTraining modelTuning modelFinal evaluation
Used duringModel trainingModel trainingAfter model training
Updates model?YesNoNo
Known to modelYesSeen during trainingNever seen before

Tip:

In practice, for small datasets, we often use cross-validation, where the validation set rotates among the data to make the most of limited samples.

Typical Size Ranges for Small Datasets

Dataset TypeNumber of Samples (Roughly)
Very Small< 500 samples
Small500 โ€“ 10,000 samples
Medium10,000 โ€“ 100,000 samples
Large100,000+ samples

Why Size Matters

  • Small datasets are more prone to:
    • Overfitting โ€“ model memorizes data instead of learning general patterns.
    • High variance in performance depending on the data split.
  • Big models (e.g., deep neural networks) usually need large datasets to perform well.

๐Ÿ’ก Common Examples

  • Medical diagnosis: Often < 5,000 patient records โ†’ small dataset.
  • NLP for niche domains: < 10,000 labeled texts โ†’ small.
  • Handwritten digit dataset (MNIST): 60,000 training images โ†’ medium-sized.

๐Ÿ” Tip for Small Datasets

If your dataset is small:

  1. Use cross-validation (like 5-fold or 10-fold).
  2. Consider simpler models (e.g., logistic regression, decision trees).
  3. Use data augmentation (e.g., rotate/scale images, reword texts).
  4. Apply transfer learning if using deep learning (e.g., pre-trained models like BERT, ResNet).