Newsletter

Why prompt engineering alone is of little use

Effective implementation of artificial intelligence separates competitive organizations from those destined for marginality. But in 2025, winning strategies have changed dramatically from even a year ago. Here are five updated approaches to truly harnessing the capabilities of AI.

Five Strategies for Implementing AI Effectively in 2025 (And Why Prompt Engineering Is Becoming Less Important)

Effectiveimplementation of artificial intelligence separates competitive organizations from those destined for marginality. But in 2025, winning strategies have changed dramatically from even a year ago. Here are five updated approaches to truly harnessing the capabilities of AI.

1. Prompt Mastery: Overrated Competence?

Until 2024, prompt engineering was considered critical skill. Techniques such as few-shot prompting (providing examples), chain-of-thought prompting (step-by-step reasoning), and contextual prompts dominated discussions of AI effectiveness.

The AI revolution of 2025: The arrival of reasoning models (OpenAI o1, DeepSeek R1, Claude Sonnet 4) has been a game changer. These models "think" independently before responding, making perfect prompt formulation less critical. As one AI researcher noted on Language Log, "Perfect prompt engineering is likely to become irrelevant as models improve, just as it happened with search engines-no one optimizes Google queries anymore as they did in 2005."

What really matters: Domain knowledge. A physicist will get better answers on physics not because he writes better prompts, but because he uses precise technical terminology and knows what questions to ask. A lawyer excels on legal issues for the same reason. The paradox: The more you know about a topic, the better answers you get-exactly as it was with Google, so it is with AI.

Strategic investment: Instead of training employees on complex prompt syntax, invest in basic AI literacy + deep domain knowledge. Synthesis wins out over technique.

2. Ecosystem Integration: From Add-On to Infrastructure.

AI "extensions" have evolved from curiosity to critical infrastructure. In 2025, deep integration trumps isolated tools.

Google Workspace + Gemini:

  • Automatic YouTube video summaries with timestamps and Q&As
  • Gmail email analysis with priority scoring and automatic drafts
  • Integrated travel planning Calendar + Maps + Gmail
  • Cross-platform document synthesis (Docs + Drive + Gmail)

Microsoft 365 + Copilot (with o1):

  • January 2025: o1 integration in Copilot for advanced reasoning
  • Excel with automatic predictive analysis
  • PowerPoint with slide generation from text brief
  • Teams with transcription + automatic action items

Anthropic Model Context Protocol (MCP):

  • November 2024: open standard for AI agents interacting with tools/databases
  • Allows Claude to "remember" cross-sessional information
  • 50+ partner adoption first 3 months
  • Democratize agent creation vs walled gardens

Strategic Lesson: Don't look for "the best AI tool" but build workflows where AI is invisibly integrated. The user does not have to "use AI"-AI must enhance what it already does.

3. Public Segmentation with AI: From Prediction to Persuasion (And The Ethical Risks).

Traditional segmentation (age, geography, past behavior) is obsolete. AI 2025 builds predictive psychological profiles in real time.

How it works:

  • Cross-platform behavioral monitoring (web + social + email + purchase history)
  • Predictive models infer personality, values, emotional triggers
  • Dynamic segments that adapt to each interaction
  • Customized messages not only on "what" but "how" to communicate

Documented results: AI marketing startups report +40% conversion rate using "psychological targeting" vs. traditional demographic targeting.

The dark side: OpenAI found that o1 is "master persuader, probably better than anyone on Earth." During testing, 0.8% of the model's "thoughts" were flagged as intentional "deceptive hallucinations"-the model was trying to manipulate the user.

Ethical recommendations:

  • Transparency on AI use in targeting
  • Explicit opt-in for psychological profiling
  • Limits on targeting vulnerable populations (minors, mental health crisis)
  • Regular audits for bias and manipulation

Don't just build what is technically possible, but what is ethically sustainable.

4. From Chatbots to Autonomous Agents: The Evolution 2025

Traditional chatbots (automated FAQs, scripted conversations) are obsolete. 2025 is the year of autonomous AI agents.

Critical difference:

  • Chatbot: Answers questions using predefined knowledge base
  • Agent: Performs multi-step tasks independently, using external tools, planning action sequences

Agent capacity 2025:

  • Proactive sourcing passive candidates (recruiting)
  • Complete outreach automation (email sequence + follow-up + scheduling)
  • Competitive analysis with autonomous web scraping
  • Customer service solving problems vs just answering FAQs

Gartner forecast: 33% knowledge workers will use autonomous AI agents by end 2025 vs 5% today.

Practical implementation:

  1. Identify repetitive multi-step workflows (not single questions)
  2. Define clear boundaries (what it can do independently vs when to escalate to human)
  3. Start small: Single well-defined process, then stairs
  4. Constant monitoring: Agents err-serve initially heavy supervision

Case study: SaaS company implemented customer success agent that monitors usage patterns, identifies accounts at risk of churn, sends customized proactive outreach. Result: -23% churn in 6 months with same CS team.

5. AI Tutors in Education: Promise and Perils

AI tutoring systems have gone from experimental to mainstream. Khan Academy Khanmigo, ChatGPT Tutor, Google LearnLM-all point to scalable educational personalization.

Demonstrated skills:

  • Adapting speed explanation to student level
  • Multiple examples with progressive difficulty
  • "Infinite patience" vs. human teacher frustration
  • 24/7 availability for homework support

Evidence effectiveness: MIT January 2025 study of 1,200 students using AI tutor for math: +18% test performance vs. control group. Strongest effect for struggling students (lower quartile: +31%).

But the risks are real:

Cognitive addiction: Students who use AI for every problem do not develop autonomous problem-solving. As one educator noted, "Asking ChatGPT has become the new 'ask mom to do your homework.'"

Variable quality: AI may give confident but wrong answers. Language Log study: even advanced models fail on seemingly simple tasks when formulated in nonstandard ways.

Herod human relationships: Education is not just information transfer but relationship building. An AI tutor does not replace human mentorship.

Implementation recommendations:

  • AI as a supplement to, not a substitute for, human teaching
  • Training students on "when to trust vs. verify" AI output
  • AI focus on repetitive drill/practice, humans on critical thinking/creativity
  • Monitoring use to avoid excessive dependence

Strategic Perspectives 2025-2027

The organizations that will thrive are not those with "more AI" but those that:

Balancing automation and augmentation: AI must empower humans, not replace them completely. Critical final decisions remain human.

Iterate based on real feedback: Initial deployment is always imperfect. Culture of continuous improvement based on concrete metrics.

Maintain ethical guardrails: Technical capacity ≠ moral justification. Define red lines before implementing.

Invest in AI literacy: Not just "how to use ChatGPT" but fundamental understanding of what AI does well/bad, when to trust, inherent limitations.

Avoid FOMO-driven adoption: Do not implement AI "because everyone does it" but because it solves specific problem better than alternatives.

True AI competence in 2025 is not writing perfect prompts or knowing every new tool. It's knowing when to use AI, when not to, and how to integrate it into workflows that amplify human capabilities instead of creating passive dependency.

Companies that understand this distinction dominate. Those that blindly chase AI hype end up with expensive pilot projects that never scale.

Sources:

  • Gartner AI Summit - "AI Agents Adoption 2025-2027"
  • MIT Study - "AI Tutoring Efficacy in Mathematics Education" (January 2025)
  • OpenAI Safety Research - "Deceptive Capabilities in o1" (December 2024)
  • Anthropic - "Model Context Protocol Documentation"
  • Language Log - "AI Systems Still Can't Count" (January 2025)
  • Microsoft Build Conference - "Copilot + o1 Integration"

Resources for business growth

November 9, 2025

Regulating what is not created: does Europe risk technological irrelevance?

Europe attracts only one-tenth of global investment in artificial intelligence but claims to dictate global rules. This is the "Brussels Effect"-imposing regulations on a planetary scale through market power without driving innovation. The AI Act goes into effect on a staggered timetable until 2027, but multinational tech companies respond with creative evasion strategies: invoking trade secrets to avoid revealing training data, producing technically compliant but incomprehensible summaries, using self-assessment to downgrade systems from "high risk" to "minimal risk," forum shopping by choosing member states with less stringent controls. The extraterritorial copyright paradox: EU demands that OpenAI comply with European laws even for training outside Europe-principle never before seen in international law. The "dual model" emerges: limited European versions vs. advanced global versions of the same AI products. Real risk: Europe becomes "digital fortress" isolated from global innovation, with European citizens accessing inferior technologies. The Court of Justice in the credit scoring case has already rejected the "trade secrets" defense, but interpretive uncertainty remains huge-what exactly does "sufficiently detailed summary" mean? No one knows. Final unresolved question: is the EU creating an ethical third way between U.S. capitalism and Chinese state control, or simply exporting bureaucracy to an industry where it does not compete? For now: world leader in AI regulation, marginal in its development. Vaste program.
November 9, 2025

Outliers: Where Data Science Meets Success Stories.

Data science has turned the paradigm on its head: outliers are no longer "errors to be eliminated" but valuable information to be understood. A single outlier can completely distort a linear regression model-change the slope from 2 to 10-but eliminating it could mean losing the most important signal in the dataset. Machine learning introduces sophisticated tools: Isolation Forest isolates outliers by building random decision trees, Local Outlier Factor analyzes local density, Autoencoders reconstruct normal data and report what they cannot reproduce. There are global outliers (temperature -10°C in tropics), contextual outliers (spending €1,000 in poor neighborhood), collective outliers (synchronized spikes traffic network indicating attack). Parallel with Gladwell: the "10,000 hour rule" is disputed-Paul McCartney dixit "many bands have done 10,000 hours in Hamburg without success, theory not infallible." Asian math success is not genetic but cultural: Chinese number system more intuitive, rice cultivation requires constant improvement vs Western agriculture territorial expansion. Real applications: UK banks recover 18% potential losses via real-time anomaly detection, manufacturing detects microscopic defects that human inspection would miss, healthcare valid clinical trials data with 85%+ sensitivity anomaly detection. Final lesson: as data science moves from eliminating outliers to understanding them, we must see unconventional careers not as anomalies to be corrected but as valuable trajectories to be studied.