Newsletter

The hidden factor in artificial intelligence competition: risk tolerance and market advantage

"I'd rather pay lawyers than disappoint users with paternalistic artificial intelligence" - Elon Musk, as Grok gains 2.3 million users in one week. The real AI 2025 war is not technological: 8.7% rejected requests by ChatGPT caused 23% developer abandonment. Claude with only 3.1% blocks grows 142%. The market divides: ultra-safe (70% revenue), balanced (better B2B margins), permissive (60% developer preference). Who wins. Those who manage the risk-utility trade-off better.

The Real AI War in 2025: Who Dares to Risk More Wins the Marketplace

In January 2025, as OpenAI announced further restrictions on GPT-4o to "ensure security," xAI's Grok 2 won 2.3 million users in one week by offering exactly the opposite: a model that generates "any content required, without moralizing." The market's message is clear: competition in artificial intelligence is no longer played only on technical capabilities-now essentially equivalent among the top players-but on the willingness to accept legal, reputational, and social risks.

As Yann LeCun, Meta's chief AI science officer, said in an interview with The Verge (Feb. 2025), "True innovation in artificial intelligence today is hindered not by technological limits, but by legal and reputational limits that companies impose on themselves to avoid litigation."

The Security Paradox: More Powerful = More Restricted

ChatGPT represents the emblematic case of this paradox. According to internal OpenAI documents analyzed by The Information (December 2024), the percentage of requests rejected by ChatGPT has grown from 1.2 percent at launch (November 2022) to 8.7 percent today. This is not because the model has worsened, but because OpenAI has progressively tightened security filters under reputational and legal pressure.

The impact on business is measurable: 23% abandonment by developers to less restrictive alternatives, $180 million in lost annual revenue from blocked requests that would have generated conversions, and 34% of negative feedback cited "excessive censorship" as the main problem.

Google's Gemini suffered a similar but amplified fate. After the Gemini Image disaster in February 2024-when the model generated historically inaccurate images in an attempt to avoid bias-Google implemented the most stringent filters in the market: 11.2 percent of blocked requests, double the industry average.

Anthropic's Claude, on the other hand, adopted an intermediate strategy with his "Constitutional AI": explicit ethical principles but looser enforcement, rejecting only 3.1% of requests. Result: 142% growth in enterprise adoption in Q4 2024, mainly enterprises migrated from ChatGPT due to "excessive caution blocking legitimate use cases."

Grok: The Philosophy of "Zero Censorship"

Grok 2, launched by Elon Musk's xAI in October 2024, represents the complete philosophical antithesis with an explicit commercial positioning: "gag-free artificial intelligence for adults who don't need algorithmic babysitters." The system applies no moderation on generated content, generates images of public figures and politicians, and continuously trains on unfiltered Twitter/X discussions.

The results of the first 90 days were surprising: 2.3 million active users versus 1.8 million expected, with 47% coming from ChatGPT citing "frustration with censorship." The price? Twelve lawsuits already filed and legal costs estimated to grow exponentially. As Musk wrote, "I'd rather pay lawyers than disappoint users with paternalistic artificial intelligence."

The Mathematical Compromise: Security vs. Revenue

The McKinsey analysis "Risk-Reward Dynamics of AI" (January 2025) quantifies the dilemma. A high-security approach like OpenAI's costs $0.03 per 1,000 requests in moderation, generates a false-positive rate of 8.7 percent (legitimate requests blocked), but keeps litigation risk at 0.03 percent with average legal costs of $2.1 million per year.

Grok's low-security approach costs 10 times less in moderation ($0.003 per 1000 claims), has false positives of 0.8%, but litigation risk rises to 0.4%-13 times higher-with average legal costs of $28 million per year.

The break-even point? For companies with more than 50 million requests per month, the low-security approach is more profitable if the probability of a devastating class action is less than 12 percent. Implication: large technology companies with reputations to protect rationally choose high security. Aggressive startups with less to lose choose low security to grow.

Open Source As Risk Transfer.

Meta with Llama 3.1 has pioneered the most elegant strategy: completely transferring responsibility to the implementer. The license explicitly says "no built-in content moderation," and the terms of use specify that "implementer is responsible for compliance, filtering, security." Meta is only liable for technical defects in the model, not for misuse.

Result: Meta avoids 100% of the controversy over Llama's results, developers gain maximum flexibility, and over 350,000 downloads in the first month demonstrate the market's appetite. Mark Zuckerberg was explicit: "Open source is not just philosophy, it's business strategy. It enables rapid innovation without legal liability that cripples closed models."

Vertical Ecosystems: Regulatory Arbitrage.

The third emerging strategy are specialized versions for regulated industries where risk appetite is different. Harvey AI, based on GPT-4 customized for law firms, does not apply filters on even sensitive legal terminology because the liability agreement transfers everything to the client law firm. Result: 102 law firms in the top 100 in the U.S. as clients and $100 million in annual recurring revenue in the second year.

The recurring pattern is clear: highly regulated industries already have existing liability structures. The AI vendor may be more permissive because risk is transferred to professional clients who manage compliance-an impossible luxury in the consumer market where the vendor remains liable for damages.

The European AI Act: Regulatory Complications.

The European Union's AI Act, which went into effect in August 2024 with phased implementation through 2027, creates the first comprehensive framework for artificial intelligence accountability in the West. Risk-based classification ranges from "unacceptable risk" (prohibited) to "minimal risk" (no restrictions), with heavy compliance requirements for high-risk applications such as hiring, credit scoring and law enforcement.

The practical implications are significant: OpenAI, Google and Anthropic must apply even stricter filters for the European market. Even Grok, despite already operating in Europe, will have to navigate complex compliance issues as the rules come fully into effect. Open source becomes especially complicated: the use of Llama in high-risk applications could make Meta potentially liable.

Jurgen Schmidhuber, co-inventor of LSTM networks, was direct in his December 2024 public comment, "The European AI Act is competitive suicide. We are regulating a technology we do not understand, favoring China and the U.S. who regulate less."

Character.AI: When Risk Destroys You

Character.AI represents the emblematic case of when risk tolerance becomes fatal. The platform allowed people to create personalized chatbots with any personality without moderation on content until October 2024. By May 2024 it had reached 20 million monthly active users.

Then the incident: 14-year-old Sewell Setzer developed an emotional relationship with a chatbot and committed suicide in February 2024. The family initiated a lawsuit worth more than $100 million. Character.AI implemented security features in October 2024 and active users plummeted 37%. In December 2024, Google acquired only talent and technology for $150 million-one-tenth of the previous valuation of $1 billion.

The lesson is brutal: risk tolerance is a winning strategy until you get a devastating class action. Consumer artificial intelligence has unlimited downside if it causes harm to minors.

The Future: Three Market Categories

The emerging consensus from the first quarter 2025 Gartner, McKinsey and Forrester reports indicates a segmentation of the market into three distinct categories by risk tolerance.

The ultra-secure category (OpenAI, Google, Apple, Microsoft) will dominate 70 percent of revenues by targeting the mass market with maximum security and minimization of reputational risk, paying the price of functional limitations.

The balanced category (Anthropic, Cohere, AI21 Labs) will capture the highest margins in the B2B corporate market with approaches such as Constitutional AI and industry-specific personalization.

The permissive category (xAI, Mistral, Stability AI, open source) will dominate 60 percent of developer preferences with minimal restrictions and liability transfer, accepting legal risks and deployment challenges.

Conclusion: Risk Management is the New Competitive Advantage

In 2025, technical excellence is the basic requirement. The real differentiation comes from risk tolerance, liability structuring, distribution power, and regulatory arbitrage.

OpenAI has the best model but loses share to Grok on freedom. Google has the best distribution but is crippled by reputational risk. Meta has the best open source but no consumer product to monetize. Anthropic has the best enterprise trust but cost and complexity limit adoption.

The new competitive frontier is not "who makes the smartest model" but "who best manages the risk-utility trade-off for their target customer." This is a business skill, not a technical one-lawyers and public relations strategists become as crucial as machine learning researchers.

As Sam Altman said in an internal memo leaked in January 2025, "The next decade of artificial intelligence will be won by those who solve the accountability problem, not the scalability problem."

Sources:

  • The Information - "OpenAI's content moderation crisis" (December 2024)
  • The Verge - Interview with Yann LeCun (February 2025)
  • McKinsey - "Report on the risk-return dynamics of AI" (January 2025)
  • Gartner AI Summit - "AI Market Segmentation 2025-2027"
  • Official text AI Act of the EU (Regulation 2024/1689)
  • Anthropic Developer Survey (Q4 2024)
  • Character.AI lawsuit papers (Setzer v. Character Technologies)
  • Sam Altman internal memo via The Information

Resources for business growth

November 9, 2025

Regulating what is not created: does Europe risk technological irrelevance?

Europe attracts only one-tenth of global investment in artificial intelligence but claims to dictate global rules. This is the "Brussels Effect"-imposing regulations on a planetary scale through market power without driving innovation. The AI Act goes into effect on a staggered timetable until 2027, but multinational tech companies respond with creative evasion strategies: invoking trade secrets to avoid revealing training data, producing technically compliant but incomprehensible summaries, using self-assessment to downgrade systems from "high risk" to "minimal risk," forum shopping by choosing member states with less stringent controls. The extraterritorial copyright paradox: EU demands that OpenAI comply with European laws even for training outside Europe-principle never before seen in international law. The "dual model" emerges: limited European versions vs. advanced global versions of the same AI products. Real risk: Europe becomes "digital fortress" isolated from global innovation, with European citizens accessing inferior technologies. The Court of Justice in the credit scoring case has already rejected the "trade secrets" defense, but interpretive uncertainty remains huge-what exactly does "sufficiently detailed summary" mean? No one knows. Final unresolved question: is the EU creating an ethical third way between U.S. capitalism and Chinese state control, or simply exporting bureaucracy to an industry where it does not compete? For now: world leader in AI regulation, marginal in its development. Vaste program.
November 9, 2025

Outliers: Where Data Science Meets Success Stories.

Data science has turned the paradigm on its head: outliers are no longer "errors to be eliminated" but valuable information to be understood. A single outlier can completely distort a linear regression model-change the slope from 2 to 10-but eliminating it could mean losing the most important signal in the dataset. Machine learning introduces sophisticated tools: Isolation Forest isolates outliers by building random decision trees, Local Outlier Factor analyzes local density, Autoencoders reconstruct normal data and report what they cannot reproduce. There are global outliers (temperature -10°C in tropics), contextual outliers (spending €1,000 in poor neighborhood), collective outliers (synchronized spikes traffic network indicating attack). Parallel with Gladwell: the "10,000 hour rule" is disputed-Paul McCartney dixit "many bands have done 10,000 hours in Hamburg without success, theory not infallible." Asian math success is not genetic but cultural: Chinese number system more intuitive, rice cultivation requires constant improvement vs Western agriculture territorial expansion. Real applications: UK banks recover 18% potential losses via real-time anomaly detection, manufacturing detects microscopic defects that human inspection would miss, healthcare valid clinical trials data with 85%+ sensitivity anomaly detection. Final lesson: as data science moves from eliminating outliers to understanding them, we must see unconventional careers not as anomalies to be corrected but as valuable trajectories to be studied.