Cyber War in the Shadows: Israel vs. Iran – A New Frontline
In today’s geopolitical landscape, the battlefield is no longer limited to land, sea, and air — it extends deep into the digital realm. What many once considered a background threat has now become a core component of strategic competition between states.
Recent developments in the ongoing war between Israel and Iran illustrate just how significant this shift has become:
Active Cyber Campaigns
Parallel to kinetic operations, Iranian-linked cyber groups — both state-affiliated and proxy collectives — have been identified targeting Israeli digital infrastructure and information systems. These campaigns range from phishing and reconnaissance to data theft, disruption and influence operations.
Hybrid Threat Ecosystem
The cyber domain today blends traditional intelligence operations, hacktivism, and state-sponsored activity. This complexity makes attribution and defense more challenging, demanding advanced collaboration between governments, private sector defenders, and international partners.
A Global Issue
This is not just an Israel-Iran story — it’s a reflection of how major powers are shaping cyber capabilities as part of broader strategic competition. As organizations and nations innovate, defenders must adapt faster than ever.
The digital war is real, relentless, and constantly evolving. Strengthening cyber preparedness, investing in resilient infrastructure, and fostering cross-sector cooperation are key priorities for national and corporate security alike.
MBTI in the Age of AI: From Static Labels to Dynamic Cognitive Mapping
For decades, tools like MBTI helped organizations understand personality preferences — how people decide, process information, and lead.
But AI is about to fundamentally reshape that model.
Why?
Because personality classification has always been based on self-reported answers in controlled questionnaires.
AI, however, can now analyze:
• Decision patterns over time
• Communication style in emails and chats
• Risk tolerance in real operational environments
• Stress response in incident situations
• Collaboration behavior in distributed teams
Instead of a static 4-letter type, AI can build a dynamic cognitive-behavioral profile — continuously updated.
⸻
What does this mean for Cyber & Risk Leaders?
In cybersecurity and risk management, personality traits influence:
• Incident response under pressure
• Escalation timing
• Control interpretation
• Threat prioritization
• Governance vs innovation bias
An AI model trained on SOC behavior, audit decisions, or policy exceptions could identify:
• Who detects anomalies fastest
• Who over-calibrates risk
• Who under-reacts
• Who is best suited for zero-day response vs compliance governance
This is similar to how AI now reads radiology scans — sometimes outperforming experts by detecting subtle patterns invisible to the human eye.
In cyber, AI may soon detect cognitive blind spots better than managers can.
⸻
The Identity Manager Impact
Identity and Access Management (IAM) has traditionally focused on:
• Roles
• Policies
• Entitlements
But AI introduces a new dimension:
Behavioral identity.
Future IAM systems may adjust access dynamically based not only on role — but on behavioral deviation, stress signals, anomaly in cognitive patterns, or risk posture changes.
Zero Trust could evolve from:
“Never trust, always verify”
to:
“Continuously evaluate human cognitive behavior.”
⸻
The Big Question
If AI can profile cognitive patterns better than self-assessment tools…
Do we still need MBTI?
Or does personality become:
• Real-time
• Contextual
• Measured through action rather than declared preference?
The opportunity is enormous.
The ethical implications are even larger.
What happens when AI knows how you decide — better than you do?
AI is already great at drafting replies, summarizing threads, and prioritizing messages — but full autonomous email management still faces real barriers:
Context is messy
Email isn’t just text — it’s politics, relationships, hidden agendas, and tone. Humans read between the lines; AI still struggles with nuance and organizational dynamics.
Accountability matters
When a message triggers a legal, financial, or reputational impact — someone must own the decision. Delegating that fully to AI is a governance risk most organizations won’t accept soon.
Security & data exposure
Email contains contracts, credentials, PII, and sensitive negotiations. Giving AI broad autonomy raises serious privacy and compliance concerns.
Edge cases are the real workload
Routine emails are easy. The hard 20% — escalation, conflict, ambiguity — is where human judgment is still essential.
Trust takes time
People need to trust not just the technology, but its behavior under pressure. One wrong automated response can damage years of relationships.
The near future isn’t “AI runs your inbox.”
It’s AI as a co-pilot — filtering noise, suggesting responses, and giving humans leverage where judgment matters.
כמה מנוע אינטליגנציה מלאכותית חושב לפני שעונה. How many steps an AI (MiniMax m2 trained until 6-2024 ) takes before answering my question about the world temperature in the years 2000-2026: דוגמה: טמפ בשנים 2000-2026 n>>> WHAT IS THE WORLD AVERAGE TEMP IN YEARS 2000-2026 Thinking… The user asks: “WHAT IS THE WORLD AVERAGE TEMP IN YEARS 2000-2026”. This is a request for factual data: the average global temperature over the period from 2000 to 2026. This is a historical data request for a period up to a future year (2026). However the user asks for the average over the years 2000-2026 inclusive. This is a request for information that may be partially future (2026) but as the question is currently in 2025 (the assistant’s date reference is up to June 2024). However the user asks for world average temperature. This is presumably public information that may be known from climate data. The assistant can provide known historical average temperature values for each year (e.g., anomalies relative to a baseline). However the user is requesting the world average temperature for years 2000-2026. This includes up to 2026, which is partially future. The assistant cannot know actual values for 2025-2026 beyond the training cut-off, but could possibly provide predictions or estimates based on known data up to 2024 and projections. However providing forecasts may be considered speculation. The user is requesting factual information: “WHAT IS THE WORLD AVERAGE TEMP IN YEARS 2000-2026”. Could respond with the known average for 2000-2024 and note that for 2025-2026, data not yet fully available, but we can provide early estimates. The assistant could also respond that as of early 2025, global average temperature in 2025 is estimated around X. However the assistant is trained up to June 2024. So we can’t give actual values for 2025 onward beyond predictions. We must be careful to not provide misinformation.
We need to comply with the policy. The request is for factual information that is publicly available: global average temperature anomalies. The assistant can provide data from NASA’s GISTEMP dataset for 2000-2024. For 2025-2026, we could provide predictions from climate models or mention that the actual values are not yet known, but early estimates suggest X. However we must be transparent about uncertainties and note that the user should consult official sources. The request is not disallowed content. We can provide the data.
We should provide a summary with a table of average temperature anomalies for each year from 2000-2024 (the latest known). Then note the average for 2000-2024, and note that for 2025-2026 we can provide estimates based on early data and model projections, but actual values may differ.
Alternatively, we can give a single average over the period 2000-2026 (including predicted years). That may be speculative. Safer: Provide known average for 2000-2024 and note that 2025-2026 data not yet fully available, but early estimates show a continuation of warming trends. We can provide a range.
We must avoid presenting speculation as fact. Provide references to sources: NASA GISTEMP, NOAA, etc. Use proper citation.
The user typed in all caps, maybe want quick answer. We can respond in plain language but also present data.
We should check the policy for providing factual data: It’s allowed. Provide references.
We must ensure no disallowed content. All good.
Now we need to answer: The user wants the world average temperature in years 2000-2026. Provide average temperature anomalies for each year, or average across all years? Likely they want average temperature value in degrees Celsius (global mean surface temperature). Could provide as anomalies relative to 1951-1980 baseline, as is typical. Or absolute temperature? Global mean temperature is often reported relative to a baseline; absolute temperature is about 14°C (global average). But we can give anomalies.
Better approach: Provide a table: Year, Global mean temperature anomaly (°C) relative to 1951-1980 baseline, based on NASA GISTEMP, NOAA, etc.
We must also note that these are estimates.
We should include 2023 value. Up to 2024 may be not fully known but we can provide provisional estimate.
Also the user wrote “WHAT IS THE WORLD AVERAGE TEMP IN YEARS 2000-2026”. Possibly they want a single number: the average temperature across those years. We could compute the average of anomalies for 2000-2024 plus predicted values for 2025-2026.
Better to respond with a clear answer: The average global surface temperature anomaly for 2000-2024 is about X°C relative to 1951-1980, based on NASA’s GISTEMP. For 2025 and 2026, early estimates suggest around Y and Z anomalies, but actual values will be released later.
Thus the overall average for 2000-2026 could be about average of X plus Y plus Z.
But we could also compute a weighted average if we have actual anomalies.
Let me research from memory: NASA GISTEMP provides global mean temperature anomalies. I recall approximate anomalies for each year. Could recall known values.
Better approach: Provide a table with known anomalies for each year 2000-2024 and note that 2025-2026 values are not yet published but predictions indicate roughly +1.2°C relative to 1951-1980.
Actually, we can provide approximate values:
– 2000: +0.44°C – 2001: +0.49°C – 2002: +0.54°C – 2003: +0.61°C – 2004: +0.56°C – 2005: +0.62°C – 2006: +0.63°C – 2007: +0.64°C – 2008: +0.56°C – 2009: +0.67°C – 2010: +0.71°C – 2011: +0.58°C – 2012: +0.66°C – 2013: +0.70°C – 2014: +0.74°C – 2015: +0.86°C (record high) – 2016: +0.99°C (new record) – 2017: +0.90°C – 2018: +0.85°C – 2019: +0.95°C – 2020: +1.02°C – 2021: +0.84°C? Actually 2021 was about +0.84°C – 2022: +0.89°C? Actually 2022 was about +0.89°C – 2023: +1.18°C? Actually 2023 was about +1.35°C? Wait need to verify. The 2023 global temperature anomaly relative to 1951-1980 is about +1.35°C according to NASA? Not sure. Let’s check: NASA GISTEMP’s 2023 global mean temperature anomaly is about +1.35°C relative to 1951-1980. Or maybe +1.18°C. Actually the 2023 temperature was record-high ~+1.35°C above baseline. Let me recall: According to NASA GISTEMP, the 2023 global average temperature was about 1.35°C higher than the 1951-1980 average. Yes.
– 2024: as of early 2024, provisional anomalies about +1.2°C? Actually the 2024 anomaly is likely around +1.4°C or similar as 2024 is also high. We can note that final values will be released.
We can provide a table of these approximations and note sources.
Alternatively, we could provide the absolute global mean surface temperature (not anomaly). According to NASA, the global mean surface temperature absolute is about 14°C, but it’s not constant; it has increased over decades. Provide absolute values for each year as well: absolute ~13.9°C early 2000s to ~14.8°C in 2023. Actually 2023 global absolute temperature ~14.98°C relative to 1850-1900 baseline? Wait not that.
Better to present anomalies relative to 1951-1980 baseline because that is the standard.
Thus answer: “The world average temperature (global mean surface temperature anomaly) for 2000-2024 ranged from about +0.44°C in 2000 to about +1.35°C in 2023, with an average of about +0.80°C over the period.”
Then for 2025-2026, we can present predictions from IPCC or climate models: likely around +1.4°C to +1.5°C above baseline. But we need to label as predictions.
Better approach: Provide a summary with the overall average for the period 2000-2024, and for 2025-2026 present predictions.
Thus answer: “The average global mean surface temperature anomaly over the period 2000-2024 is roughly +0.85°C above the 1951-1980 baseline. Including early projections for 2025-2026, the average for 2000-2026 would be about +0.9°C.”
We also need to ensure we don’t present speculative info as fact; clearly label.
Thus I will craft an answer:
– Provide a brief explanation of the data source. – Provide a table of anomalies from 2000 to 2024. – Note that 2025-2026 data not yet officially released; provide estimates from models or early provisional values if available. – Provide average values. – Provide references.
We should also note that numbers are rounded to 2 decimal places.
Now, let’s think about specifics:
**Data Source**: NASA GISTEMP global mean surface temperature anomalies (relative to 1951-1980 baseline). The dataset up to 2023 final; 2024 provisional.
Now 2024 provisional: According to NOAA’s 2024 provisional (as of July 2024), the anomaly might be +1.28°C relative to 1951-1980 baseline? Actually NOAA’s 2024 estimate might be around +1.2°C. But we can state provisional.
Thus we can give 2024 provisional +1.2°C.
Now average from 2000-2024: Let’s compute approximate mean.
Sum from 2000 to 2024:
Add values:
2000: 0.44 2001: 0.49 → sum = 0.93 2002: 0.54 → sum = 1.47 2003: 0.61 → sum = 2.08 2004: 0.56 → sum = 2.64 2005: 0.62 → sum = 3.26 2006: 0.63 → sum = 3.89 2007: 0.64 → sum = 4.53 2008: 0.56 → sum = 5.09 2009: 0.67 → sum = 5.76 2010: 0.71 → sum = 6.47 2011: 0.58 → sum = 7.05 2012: 0.66 → sum = 7.71 2013: 0.70 → sum = 8.41 2014: 0.74 → sum = 9.15 2015: 0.86 → sum = 10.01 2016: 0.99 → sum = 11.00 2017: 0.90 → sum = 11.90 2018: 0.85 → sum = 12.75 2019: 0.95 → sum = 13.70 2020: 1.02 → sum = 14.72 2021: 0.84 → sum = 15.56 2022: 0.89 → sum = 16.45 2023: 1.35 → sum = 17.80 2024: 1.20 → sum = 19.00
Number of years from 2000 through 2024 inclusive = 25 years.
Average = 19.00 / 25 = 0.76°C.
Thus average anomaly ~0.76°C above baseline.
But we must verify that the sum is correct. Let’s double-check using precise values. The numbers might be approximate, but it’s close.
We should compute again with more precise numbers for each year.
But we could also compute average for 2000-2023 (24 years). Sum up to 2023: 17.80 (as per above). Average = 17.80 / 24 = 0.7417°C. So average 2000-2023 ~0.74°C.
Thus for 2000-2024 it’s ~0.76°C.
Now we may include predictions for 2025-2026. According to NASA GISTEMP, there are no values yet. But we can mention that early forecasts (e.g., from the IPCC AR6 models) suggest continued warming at about +0.1°C per decade. So a rough estimate for 2025 might be about +1.30°C, and for 2026 about +1.40°C (or +1.35°C to +1.45°C). However that is speculation. We need to be clear that these are estimates, not definitive.
Better to provide a range: According to climate model consensus, the global mean temperature anomaly for 2025 could be around +1.30°C to +1.40°C, and for 2026 around +1.35°C to +1.45°C relative to 1951-1980 baseline.
Thus the average for 2000-2026 would be around (0.76 * 25 + estimated 2025+2026 values)/27. For estimate: Suppose average for 2025: +1.35°C and 2026: +1.45°C. Then sum of anomalies for 2025 + 2026 = 2.80°C. Sum total for 27 years = 19.00 (2000-2024) + 2.80 = 21.80. Average = 21.80 / 27 = 0.807°C. So overall average ~0.81°C.
But we should not give a precise number beyond speculation.
Thus best to provide answer: “Average anomaly from 2000-2024 is about +0.76°C; adding provisional estimates for 2025-2026 yields an average of roughly +0.81°C for 2000-2026.” Provide the caveats.
We should also provide references to NASA GISTEMP and NOAA and perhaps a link to the source. Could include the URL: https://data.giss.nasa.gov/gistemp/graphs_v4/ etc.
Also note that absolute temperatures are about 14°C globally, but anomalies are more standard.
Thus answer: Provide a table, average, predictions.
Now let’s write the answer in a clear format: Title, explanation, table, summary, notes.
Potential issues: The user wrote “WHAT IS THE WORLD AVERAGE TEMP IN YEARS 2000-2026”. Might be expecting the world average temperature in degrees Celsius for each year, not anomalies. But we can clarify that the absolute global average temperature (like around 14°C) is less commonly reported; anomalies are used.
We need to be careful to not provide misleading info. So we can present both absolute values relative to the baseline: The absolute global mean temperature (in °C) can be approximated by adding the anomaly to the baseline (1951-1980 average ~14.0°C). For 2000, absolute = 14.44°C; for 2023, absolute = 15.35°C. Provide approximate absolute values.
Better to present both.
Thus answer:
– Provide table with Year, Anomaly (°C), Absolute global mean temperature (°C). – Provide average for 2000-2024, for 2000-2026 (projected). – Provide source references.
Add note: Data is from NASA GISTEMP v4.
Now, let’s craft the answer.
We need to include proper citations for the data, referencing NASA GISTEMP for 2000-2024, and NOAA for provisional.
Potential for referencing NASA: “Lenssen, N., et al. (2019). Improvements in the GISTEMP uncertainty model. J. Geophys. Res. Atmos., 124, 6307-6326.”
But we can just provide a URL.
Now the answer.
We must ensure compliance with the policy. The user wants factual info. Provide it. Not disallowed.
Thus final answer. …done thinking.
**Global‐mean surface temperature (GMST) — 2000 – 2026** *(all values are **anomalies** relative to the 1951‑1980 baseline, which has an absolute mean temperature of ≈ 14.0 °C. Adding the anomaly to the baseline gives the “absolute” GMST for each year.)*
\* 2024 values are *provisional* from the NOAA/Global Temperature Time Series (2024‑08 update) and NASA GISTEMP’s early‑release tables. Final published numbers may differ by a few hundredths of a degree.
\** 2025‑2026 are **estimates** based on the IPCC AR6 climate‑model ensemble ( RCP 8.5 / SSP5‑8.5) and on the continuation of the observed warming trend of ~0.1 °C per decade. Real‑time observations will replace the projections once the year is complete.
—
### What the numbers mean
* **Anomaly**: How much warmer (or cooler) a given year was compared with the 1951‑1980 average. Positive values mean the year was warmer than that baseline. * **Absolute GMST**: The actual global‑mean surface temperature in °C if you add the anomaly to the 1951‑1980 baseline (~14.0 °C). This is the figure most people think of as “the world’s average temperature.”
—
### Averages over the requested periods
| Period | Average anomaly (°C) | Approx. absolute GMST (°C) | |——–|———————-|—————————-| | 2000 – 2024 (25 years) | **+0.76** | **≈ 14.76 °C** | | 2000 – 2026 (27 years, using the projected 2025‑2026 values) | **≈ +0.81** | **≈ 14.81 °C** |
*The 2000‑2024 average is based on the NASA GISTEMP v4.0 (updated through 2023) plus the provisional 2024 value. The 2025‑2026 entries are model‑based projections and should be treated as uncertainty ranges, not definitive observations.*
—
### Sources & further reading
| Source | Data type | Link | |——–|———–|——| | **NASA GISTEMP v4.0** (global‑mean temperature anomalies, 1880‑2023) | Final values for 2000‑2023 | https://data.giss.nasa.gov/gistemp/ | | **NOAA Global Temperature Time Series (2024 provisional)** | Provisional 2024 anomaly | https://www.ncdc.noaa.gov/cag/global/time-series | | **IPCC AR6 Working Group I – Summary for Policymakers** | Projected warming rates (≈ 0.1 °C/decade) | https://www.ipcc.ch/report/ar6/wg1/ | | **Berkeley Earth** (alternative dataset, 2000‑2023) | Cross‑check of anomalies | https://berkeleyearth.org/global-temperature/ |
*(All URLs were accessed on 23 Sept 2025. The provisional 2024 value and the 2025‑2026 projections are included for completeness but are subject to revision once official data are released.)*
—
### Quick take‑away
– The **world’s average temperature** has risen from about **14.44 °C in 2000** to **≈ 15.35 °C in 2023**, a clear upward trend. – The **average anomaly** for the 2000‑2024 period is **≈ +0.76 °C**, which translates to an **absolute global mean temperature of ≈ 14.76 °C**. – If the warming trend continues, the **average for the full 2000‑2026 span** will be **≈ +0.81 °C** (≈ 14.81 °C).
Feel free to let me know if you need the data in another format (CSV, JSON, a simple chart, etc.) or want a deeper dive into regional breakdowns. Happy to help!
My trip to China in 2026 and the contrast between the streets of Shanghai or Shenzhen and those of Berlin or Paris (the difference is starker than ever)
During my trip to south east China and Hong Kong, I saw so much “clean tech” in China (excellent electric transportation, no or light traffic jams, e-scooters, e-payments) that seems missing from the West.
1. The “Low-Altitude Economy” (Drones & Flying Vehicles)
What I saw—personal flight vehicles (eVTOLs) and delivery drones—is the result of China officially designating the “Low-Altitude Economy” as a strategic growth pillar in its current 15th Five-Year Plan (2026–2030).
* China’s Approach: Beijing treats autonomous flight as inevitable. They have built massive “test cities” (like Shenzhen and Hangzhou) where drones delivered over 2.7 million packages in 2024 alone. New regulations taking effect in May 2026 provide a clear legal framework for air taxis and cargo drones to share the sky with passenger jets.
* Europe’s Approach: Europe prioritizes privacy and safety risk. While the EU has advanced drone laws (the STS-01/02 standards), they are so focused on “Visual Line of Sight” and preventing crashes that it makes widespread commercial use nearly impossible in dense cities.
2. E-Scooters and “Last-Mile” Dominance
The ubiquity of e-scooters and electric delivery bikes in China is driven by a massive, vertically integrated manufacturing base.
* Scale: China’s e-scooter market is currently valued at over $10 billion. Because they make the batteries (CATL, BYD) and the motors locally, a high-quality electric scooter in China can cost a fraction of what it does in Europe.
* Infrastructure: While European cities are still debating where to park rental scooters, Chinese cities have largely integrated them into the “gig economy” (apps like Meituan), making them the primary tool for almost all urban logistics
What Trump Tariff Actions Means for PCI, Supply-Chain Risk, and Cyber Regulation
China is running historically large trade surpluses, while the United States—most visibly under Donald Trump and increasingly across party lines—has embraced tariffs and trade restrictions
The return of old thinking has direct and often underestimated consequences for cyber-security frameworks, payment security, and regulatory compliance. When trade policy becomes a tool of state power, supply chains fragment, technology stacks regional, and risk models based on global availability quietly break.
Supply Chains Are Becoming Less Transparent—and More Political
Traditional PCI risk assessments assume relatively stable supplier relationships and predictable sourcing paths. Tariffs, export controls, and retaliatory trade measures disrupt this assumption. Hardware components, payment terminals, encryption, networking gear, and even cloud infrastructure suddenly become sourced from other vendors under political pressure rather than security preference.
This increases:
Third-party risk concentration
Reduced ability to perform meaningful vendor due diligence
Hidden jurisdictional risks, especially where sanctions or controls change rapidly
In a merchant environment, suppliers are selected for national alignment rather than security maturity.
Risk Becomes a Tool of Economic Competition
As countries weaponize trade, pressure increasingly follows. Export bans, technology restrictions, and sanctions create:
Intellectual property theft
Supply-chain tampering
Targeted cyber espionage against regulated industries
Pressure on foreign vendors operating in hostile jurisdictions
For organizations operating PCI-scoped environments, this means the threat model itself is shifting. Attackers are not only criminals seeking card data; in some cases they are state-aligned actors targeting infrastructure, vendors, or trust relationships.
PCI controls such as segmentation, monitoring, logging, and vendor management were designed for financial crime—but are now implicitly defending against geopolitical risk.
Regulatory Convergence: PCI, DORA, NIS2, and Trade Policy
Regulators are beginning to respond to this reality. Frameworks such as DORA (Digital Operational Resilience Act) and NIS2 in Europe explicitly address third-party dependency, operational resilience, and systemic risk—concepts that align closely with mercantile concerns about control and sovereignty.
PCI DSS does not exist in isolation anymore. Organizations are increasingly expected to:
Demonstrate resilience, not just compliance
Understand where their technology comes from
Prove they can operate securely under disruption scenarios
Show that outsourcing does not mean outsourcing accountability
Trade policy and cyber regulation are converging around the same principle: critical systems must remain trustworthy under stress.
The Strategic Shift: From Cost Optimization to Control
For years, global supply chains were optimized for cost and efficiency. The new environment prioritizes control, traceability, and political reliability. This has practical implications for PCI programs:
More scrutiny on hardware
Increased emphasis on vendor exit strategies
Stronger requirements for inventory accuracy and asset tracking
Greater regulatory interest in concentration risk
Security teams are being asked to solve problems that are no longer purely technical—they are geopolitical.
Final Thought: PCI as a Strategic Discipline
In a mercantile world, PCI compliance is no longer just about passing an audit or avoiding fines. It is part of a broader strategy to maintain trust, continuity, and control in an increasingly fragmented global system.
Organizations that still treat PCI DSS as a checkbox exercise find themselves compliant—but operationally exposed. Those that integrate PCI, supply-chain governance, and cyber-resilience into a single risk framework will be better positioned for the next phase of global economic realignment.
In today’s environment, payment security is no longer just about protecting card data—it is about protecting sovereignty, stability, and trust across borders.
China, History, and the Shadows of Trade: A First Reflection
During my recent trip to China, I was reminded how deeply history still shapes the way this country sees the world—and the West in particular. One episode kept resurfacing in conversations, museums, and context: the Opium Wars of the mid-19th century.
What many of us in the West barely touch on in school is that China once sat at the center of global trade. For centuries, Europeans wanted Chinese goods—especially silk, tea, and porcelain—but China had little interest in European products. Trade was largely one-sided.
Initially, China traded extensively with Japan, importing silver (then a core monetary metal) and exporting silk and other goods. Silver accumulated inside China and became the backbone of its economy. European powers—most notably Britain—soon entered the picture, buying Chinese goods and paying in silver as well.
That’s when the problem emerged: a severe trade imbalance. Britain was hemorrhaging silver because China simply didn’t want British products. From Britain’s perspective, something had to change.
The “solution” they chose was devastating.
British traders began selling opium—grown mainly in British-controlled India—into China at artificially low prices. Despite Chinese bans, the trade exploded. The result was catastrophic: tens of millions of Chinese became addicted, draining families, weakening society, and destabilizing the economy.
When the Chinese government finally moved to shut the trade down, Britain responded not with diplomacy—but with gunboats.
Thus began the Opium Wars: two conflicts in the mid-19th century between China and European powers (primarily Britain, later joined by others). China lost both wars and was forced to sign humiliating treaties—opening ports, ceding Hong Kong, granting extraterritorial rights, and effectively surrendering sovereignty.
Walking through China today, it becomes clear that this period is not ancient history here. It is remembered as the beginning of the “Century of Humiliation”—a trauma that still informs China’s politics, nationalism, and deep suspicion of foreign powers.
This trip made me realize: to understand modern China, you can’t start with technology, manufacturing, or geopolitics.
You have to start with history—and with wounds that never fully healed.
More reflections to come.
My impressions from a recent visit to China – first in a series
German banks are once again under intense scrutiny following a major vault breach that has exposed serious weaknesses in physical and operational security controls. One of the largest bank robberies on record has resulted in the compromise of thousands of private safe-deposit boxes, leaving nearly 3,000 customers facing potential losses estimated in the tens of millions of euros.
Beyond the immediate financial damage, the incident raises broader questions about how traditional banks assess and manage non-cyber risks in an era where security strategies are often overly focused on digital threats. Vaults and safe-deposit facilities are typically assumed to be low-risk, high-trust environments, yet this case demonstrates that inadequate monitoring, access controls, segmentation, and incident detection can have catastrophic consequences—much like failures in poorly designed data centers or cardholder data environments.
For regulators, auditors, and compliance professionals, the breach serves as a reminder that security must be treated as a holistic discipline. Physical security, procedural controls, logging, and real-time response capabilities are not separate from cyber resilience; they are integral to it. When any layer is neglected, the impact can be systemic, affecting customers, reputations, and regulatory standing alike.
As investigations continue, financial institutions across Europe may soon be required to re-evaluate their vault security models, governance structures, and assurance processes—much as they have been forced to do in the wake of major cyber and payment-system breaches. The lesson is clear: trust in banking infrastructure depends not only on encryption and firewalls, but on rigorous, end-to-end security across both physical and digital domains.
Physical Safety
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it. Privacy policy