Μετά το χτύπημα της πόλης DNIPRO
από ρωσικό πύραυλο τη νύχτα 21/22 Νοε 2024 για τον οποίο οι ΗΠΑ δήλωσαν ότι ήταν πειραματικό όπλο (όχι σε παραγωγή δηλαδή), πρόχειρη έρευνα σε δυτικές
πηγές, δείχνει ότιεπρόκειτο για βαλλιστικό
πύραυλο μέσου βεληνεκούς (IRBM=Intermediate-Range
Ballistic
Missile) εναντίον συγκεκριμένων στόχων στην Ουκρανία, σε απόσταση 1500χλμ.
Οι ΗΠΑ πρέπει λογικά (σύμφωνα με την
START- StrategicArmsReductionTreaty/
Ομπάμα-Πούτιν /2009) να είχαν ειδοποιηθεί από τους Ρώσσους για την εκτόξευση
Διηπειρωτικού πυραύλου, προς αποφυγή «παρεξηγήσεων». Αν δεν υπήρχε προειδοποίηση,
ίσως ο κόσμος να ήταν αλλιώς σήμερα. Το άλλο ενδεχόμενο, είναι ότι απλά δεν
επρόκειτο για διηπειρωτικό πύραυλο. ΠΡΟΣΘΗΚΗ ΤΗΣ 24 Νοε 2024:Η επακριβής
δήλωση της Αναπληρωτού Εκπροσώπου Τύπου του Πενταγώνου (ΗΠΑ) ήταν:
"Μπορώ να επιβεβαιώσω ότι η Ρωσσία εξετόξευσε ένα πειραματικό βαλλιστικό
πύραυλο μέσου βεληνεκούς. Αυτός ο IRBM ήταν βασισμένος στο Ρωσσικό
μοντέλο ICBM, to RS-26 Rubezh". Επίσης δήλωσε ότι η Αμερικανική
κυβέρνηση είχε ενημερωθεί από τη Ρωσσία έγκαιρα για την εκτόξευση,
"through Nuclear Risk Reduction channels".
Η Ρωσία στόχευσε νοτιοδυτικά του Dnipro, το Youzhmash, το εργοστάσιο όπου
παράγεται στρατιωτικός και πολιτικός αεροδιαστημικός εξοπλισμός από την
Ουκρανία. Αυτό το συγκρότημα, μαζί με το γειτονικό γραφείο σχεδιασμού
(Yozhnoye) ήταν το κέντρο ανάπτυξης των αεροδιαστημικών τεχνολογιών της ΕΣΣΔ.
Αυτό το χτύπημα είναι η απάντηση της
Ρωσίας στην απόφαση των ΗΠΑ, του Ηνωμένου Βασιλείου και της Γαλλίας να άρουν
τους περιορισμούς στη χρήση των ATACMS, Storm Shadow/
SCAPL-EG σε ρωσικό έδαφος και στην απόφαση της Ουκρανίας να τα χρησιμοποιήσει
για αυτόν ακριβώς τον σκοπό.
Ο Πούτιν λέει ότι χρησιμοποίησε ένα νέο πύραυλο που ονομάζεται
Oreshnik και που «είναι αδύνατο να αναχαιτιστεί και να καταρριφθεί από τα
δυτικά αμυντικά συστήματα». Είπε επίσης ότι απαίτησε ο νέος πύραυλος να μπει σε
μαζική παραγωγή (άρα δεν είναι τώρα σε παραγωγή), προειδοποιώντας ότι από εδώ και στο εξής θα χρησιμοποιείται στην
Ουκρανία.
Σύμφωνα με τον Fabian Hoffmann,
αμυντικό εμπειρογνώμονα, πρόκειται για μια παραλλαγή του
βαλλιστικού πυραύλου μέσου βεληνεκούς (IRBM) που δεν πρέπει να συγχέεται με τον
διηπειρωτικό βαλλιστικό πύραυλο, (ICBM) και στην επίθεση χρησιμοποιήθηκε ο γνωστός
πύραυλος RS-26 Rubezh.
Το RS-26 Rubezh είναι ένα IRBM βάρους 40-50 τ. με εμβέλεια έως και 5500 km. Αναπτύχθηκε
τη δεκαετία του 2000 από τη Ρωσία ως απάντηση στα αμερικανικά συστήματα THAAD
και Aegis, ειδικά μετά την αποκάλυψη των αμερικανικών σχεδίων για ανάπτυξη
τέτοιων συστημάτων στην Ανατολική Ευρώπη (Ρουμανία, Deveselu). Το RS-26 Rubezh
είναι ένα πραγματικά τρομακτικό σύστημα από τεχνικής άποψης, ικανό να μεταφέρει
πολλαπλές κεφαλές MIRV (Multiple Independently targetable Reentry
Vehicle/Πολλαπλές
Κεφαλές Ανεξάρτητης Στόχευσης), περιλαμβανομένης της κεφαλής Avangard.
Κάποιοι (IlyaKramnikστην
Izvestiya)
λένε ότι πρόκειται για νέα γενιά ρωσικών πυραύλων στο υψηλότερο όριο επιδόσεων των IRBM, ενώ άλλοι (DmitryKornev στην ίδια εφημερίδα) ότι το Oreshnikείναι απλά μια βελτιωμένη
έκδοση του Iskader που ήδη χρησιμοποιείται ευρέως στην Ουκρανία, με βελτιωμένο
κινητήρα.
Την περασμένη άνοιξη δοκιμάσθηκε ένας τέτοιος κινητήρας στο KapustinYarστη νότια Ρωσσία.
Από την ίδια ακριβώς περιοχή εκτοξεύθηκε ο υπόψη πύραυλος, ο οποίος θα μπορούσε
επίσης να είναι μια σμίκρυνση του ICBMσυστήματος
Yars-M.
Οι Ουκρανοί λένε ότι πρόκειται για
νέου τύπου ICBM, το Kedr (Κέδρος) και ότι η περιοχή εκτόξευσης
ήδη βομβαρδίσθηκε από αυτούς (τους Ουκρανούς) την επομένη της εκτόξευσης.
Το σύστημα αντιμετωπίζεται
αποτελεσματικά στη φάση «ανόδου» δηλαδή πριν το διαχωρισμό των κεφαλών, αφού μετά το διαχωρισμό, οι απειλές πολλαπλασιάζονται.
Οι ονομαστικές δυνατότητες της κύριας κεφαλής (Avangard, βάρους 2 τ.) επιτρέπουν
στον Πούτιν να λέει ότι «είναι αδύνατο να αναχαιτιστεί», επειδή πράγματι είναι πολύ
δύσκολο έως σχεδόν αδύνατο να αναχαιτιστεί εάν ισχύουν όσα είναι γνωστά για αυτή
την κεφαλή, δηλαδή:
- Είναι υπερηχητική, άνω των 10 mach (μέχρι και 27 machδιαδίδεται). 10 mach είναι 12.250 χαω.
- Ίπταται σε μεταβλητή τροχιά (μπορεί να αλλάξει το ύψος και να κάνει ελιγμούς
αποφυγής αριστερά-δεξιά, με προεπιλογές που ενεργοποιούνται από τους αισθητήρες
του).
- Το σύστημα που χρησιμοποιήθηκε έφερε 6 κεφαλές, με 6 υποπυρομαχικά κάθε μία.
- Ο διαχωρισμός των κεφαλών γίνεται στην ενδιάμεση φάση, σε χαμηλά διαστημικά
υψόμετρα (περίπου εκεί που βρίσκεται σε τροχιά ο ISS – δηλαδή στα 400 χλμ.),
άρα είναι σχεδόν αδύνατο να εντοπισθεί από ραντάρ και αισθητήρες εδάφους.
- Είναι αυτόνομη κεφαλή και δεν χρειάζεται γεωγραφική καθοδήγηση πχ. GPS (Αμερικανικό), GLONASS
(Ρωσσικό) ή Galileo (Ευρωπαϊκό), άρα πολύ δύσκολο να παρεμβληθεί με γνωστά
μέσα Ηλεκτρονικού Πολέμου.
Τέτοιες τεχνικές λεπτομέρειες κάνουν
αυτά τα βλήματα επικίνδυνα για περιοχές που καλύπτονται μόνο από τα συνηθισμένα αντιπυραυλικά
συστήματα όπως Aegis Ashore, THAAD, BMD και Patriot PAC-3. Θεωρητικά, η Avangard μπορεί να διαπεράσει όλα αυτά
τα συστήματα και να χτυπήσει τους στόχους της.
Σύμφωνα με τις δυτικές πηγές
φαίνεται ότι ο πύραυλος Oreshnik δεν υπάρχει στην πραγματικότητα, δηλαδή είναι
concept και δεν έχουν παραχθεί πάνω από 2 τέτοιοι πύραυλοι, που ναι, μπορούν να
φέρουν και πυρηνικές κεφαλές. Ο πύραυλος RS-26 Rubezh είναι στα ίδια χνάρια, αφού η
ανάπτυξή του σταμάτησε το 2015 λόγω έλλειψης κεφαλαίων και δεν υπάρχουν
πληροφορίες ότι παράγεται αυτή τη στιγμή.
Απ’ όσα κυκλοφορούν, ένας τέτοιος
βαλλιστικός πύραυλος δεν μπορεί να παραχθεί σε λιγότερο από 6 μήνες, και αφού
έχουν γίνει όλες οι δοκιμές. Το κόστος είναι μάλλον αδιάφορο για τη Ρωσία, αφού
βρίσκεται σε πόλεμο και μπορεί να επενδύσει πολλά για να ολοκληρώσει την
ανάπτυξη τέτοιων όπλων.
Το τρωτό σημείο του συστήματος,
είναι τα κέντρα σχεδίασης, ανάπτυξης, παραγωγής και συναρμολόγησής τους. Ενώ οι
κεφαλές Avangard , σύμφωνα με ανοικτές πηγές, είναι σχεδόν
αδύνατο να αναχαιτιστούν, οι πύραυλοι όμως μπορούν να καταστραφούν από την «κούνια» τους
εάν αυτά τα κέντρα (όχι πολλά) χτυπηθούν από την Ουκρανία, που έχει δείξει ότι
έχει εξαιρετικές δυνατότητες πληροφοριών και κρούσης πολύ
μεγάλης εμβέλειας (Ουκρανικά drones έχουν χτυπήσει κατ’ επανάληψη μέχρι 1500
χλμ εντός ρωσικού εδάφους).
Το πιο σημαντικό είναι ότι παρ’ όλες
τις απειλές τα τελευταία δυόμισι χρόνια για χρήση πυρηνικών όπλων «αν η
τελευταία κόκκινη γραμμή ξεπεραστεί», αυτό δεν συνέβη ποτέ. Ο Πούτιν υποτίθεται ότι είναι αυτό που στη
θεωρία παιγνίων ονομάζεται «ορθολογικός παίκτης». Η ρητορική του εξυπηρετεί τη
μεγιστοποίηση των κερδών του στην Ουκρανία, με βάση το φόβο που ευελπιστεί να
προκαλέσει στη Δύση ώστε να μη δοθούν στην Ουκρανία έγκαιρα και αρκετά όπλα για
να αμυνθεί.
Η 22/11/2024 καταγράφεται στην
ιστορία ως η μέρα που για πρώτη φορά μια κατοικημένη πόλη δέχτηκε επίθεση με
βαλλιστικό πύραυλο μέσου βεληνεκούς με δυνατότητα να φέρει πολλαπλές πυρηνικές
κεφαλές.
Για την Ελλάδα, εκτός από την
(απίθανη) περίπτωση της στοχοποίησής της ως ΝΑΤΟϊκής χώρας σε περίπτωση συμμαχικής
ευθείας εμπλοκής με τη Ρωσσία (κατά το άρθρο 5 του Βορειοατλαντικού Συμφώνου) πιθανές
νέες επιπτώσεις θα μπορούσαν να είναι η συνέχιση της ενεργειακής κρίσης και οι καθυστερήσεις
στην υλοποίηση εξοπλιστικών προγραμμάτων λόγω αλλαγής αναγκών των προτεραιοτήτων των
προμηθευτών, υπέρ των αναγκών σε άλλα μέτωπα.
Με σταχυολόγηση δημοσιευμάτων, Κων. Β. Κωνσταντάρας
From
the recalibration of military strategy to the reconstitution of
diplomacy, artificial intelligence will become a key determinant of
order in the world. Immune to fear and favor, AI introduces a new
possibility of objectivity in strategic decision-making. But that
objectivity, harnessed by both the warfighter and the peacemaker, should
preserve human subjectivity, which is essential for the responsible
exercise of force. AI in war will illuminate the best and worst
expressions of humanity. It will serve as the means both to wage war and
to end it.
Humanity’s long-standing struggle to constitute itself
in ever-more complex arrangements, so that no state gains absolute
mastery over others, has achieved the status of a continuous,
uninterrupted law of nature. In a world where the major actors are still
human—even if equipped with AI to inform, consult, and advise
them—countries should still enjoy a degree of stability based on shared
norms of conduct, subject to the tunings and adjustments of time.
But
if AI emerges as a practically independent political, diplomatic, and
military set of entities, that would force the exchange of the age-old
balance of power for a new, uncharted disequilibrium. The international
concert of nation-states—a tenuous and shifting equilibrium achieved in
the last few centuries—has held in part because of the inherent equality
of the players. A world of severe asymmetry—for instance, if some
states adopted AI at the highest level more readily than others—would be
far less predictable. In cases where some humans might face off
militarily or diplomatically against a highly AI-enabled state, or
against AI itself, humans could struggle to survive, much less compete.
Such an intermediate order could witness an internal implosion of
societies and an uncontrollable explosion of external conflicts.
Other
possibilities abound. Beyond seeking security, humans have long fought
wars in pursuit of triumph or in defense of honor. Machines—for now—lack
any conception of either triumph or honor. They may never go to war,
choosing instead, for instance, immediate, carefully divided transfers
of territory based on complex calculations. Or they might—prizing an
outcome and deprioritizing individual lives—take actions that spiral
into bloody wars of human attrition. In one scenario, our species could
emerge so transformed as to avoid entirely the brutality of human
conduct. In another, we would become so subjugated by the technology
that it would drive us back to a barbaric past.
THE AI SECURITY DILEMMA
Many
countries are fixated on how to “win the AI race.” In part, that drive
is understandable. Culture, history, communication, and perception have
conspired to create among today’s major powers a diplomatic situation
that fosters insecurity and suspicion on all sides. Leaders believe that
an incremental tactical advantage could be decisive in any future
conflict, and that AI could offer just that advantage.
If
each country wished to maximize its position, then the conditions would
be set for a psychological contest among rival military forces and
intelligence agencies the likes of which humanity has never faced
before. An existential security dilemma awaits. The logical first wish
for any human actor coming into possession of superintelligent AI—that
is, a hypothetical AI more intelligent than a human—might be to attempt
to guarantee that nobody else gains this powerful version of the
technology. Any such actor might also reasonably assume by default that
its rival, dogged by the same uncertainties and facing the same stakes,
would be pondering a similar move.
Short of war, a
superintelligent AI could subvert, undermine, and block a competing
program. For instance, AI promises both to strengthen conventional
computer viruses with unprecedented potency and to disguise them
thoroughly. Like the computer worm Stuxnet—the cyberweapon uncovered in
2010 that was thought to have ruined a fifth of Iran’s uranium
centrifuges—an AI agent could sabotage a rival’s progress in ways that
obfuscate its presence, thereby forcing enemy scientists to chase
shadows. With its unique capacity for manipulation of weaknesses in
human psychology, an AI could also hijack a rival nation’s media,
producing a deluge of synthetic disinformation so alarming as to inspire
mass opposition against further progress in that country’s AI
capacities.
It will be hard for countries to get a clear sense of
where they stand relative to others in the AI race. Already the largest
AI models are being trained on secure networks disconnected from the
rest of the Internet. Some executives believe that AI development will
itself sooner or later migrate to impenetrable bunkers whose
supercomputers will be powered with nuclear reactors. Data centers are
even now being built on the bottom of the ocean floor. Soon they could
be sequestered in orbits around Earth. Corporations or countries might
increasingly “go dark,” ceasing to publish AI research so as not only to
avoid enabling malicious actors but also to obscure their own pace of
development. To distort the true picture of their progress, others might
even try deliberately publishing misleading research, with AI assisting
in the creation of convincing fabrications.
AI in war will illuminate the best and worst expressions of humanity.
There
is a precedent for such scientific subterfuge. In 1942, the Soviet
physicist Georgy Flyorov correctly inferred that the United States was
building a nuclear bomb after he noticed that the Americans and the
British had suddenly stopped publishing scientific papers on atomic
fission. Today, such a contest would be made all the more unpredictable
given the complexity and ambiguity of measuring progress toward
something so abstract as intelligence. Although some see advantage as
commensurate with the size of the AI models in their possession, a
larger model is not necessarily superior across all contexts and may not
always prevail over smaller models deployed at scale. Smaller and more
specialized AI machines might operate like a swarm of drones against an
aircraft carrier—unable to destroy it, but sufficient to neutralize it.
An
actor might be perceived to have an overall advantage were it to
demonstrate achievement in a particular capability. The problem with
this line of thinking, however, is that AI refers merely to a process of
machine learning that is embedded not just in a single technology but
also in a broad spectrum of technologies. Capability in any one area may
thus be driven by factors entirely different from capability in
another. In these senses, any “advantage” as ordinarily calculated may
be illusory.
Moreover, as demonstrated by the exponential and
unforeseen explosion of AI capability in recent years, the trajectory of
progress is neither linear nor predictable. Even if one actor could be
said to “lead” another by an approximate number of years or months, a
sudden technical or theoretical breakthrough in a key area at a critical
moment could invert the positions of all players.
In
such a world, where no leaders could trust their most solid
intelligence, their most primal instincts, or even the basis of reality
itself, governments could not be blamed for acting from a position of
maximum paranoia and suspicion. Leaders are no doubt already making
decisions under the assumption that their endeavors are under
surveillance or harbor distortions created by malign influence.
Defaulting to worst-case scenarios, the strategic calculus of any actor
at the frontier would be to prioritize speed and secrecy over safety.
Human leaders could be gripped by the fear that there is no such thing
as second place. Under pressure, they might prematurely accelerate the
deployment of AI as deterrence against external disruption.
A NEW PARADIGM OF WAR
For
almost all of human history, war has been fought in a defined space in
which one could know with reasonable certainty the capability and
position of hostile enemy forces. The combination of these two
attributes offered each side a sense of psychological security and
common consensus, allowing for the informed restraint of lethality. Only
when enlightened leaders were unified in their basic understanding of
how a war might be fought could opposing forces determine whether a war
should be fought.
Speed and mobility have been among the most
predictable factors underpinning the capability of any given piece of
military equipment. An early illustration is the development of the
cannon. For a millennium after their construction, the Theodosian Walls
protected the great city of Constantinople from outside invaders. Then,
in 1452, a Hungarian artillery engineer proposed to Emperor Constantine
XI the construction of a giant cannon that, firing from behind the
defensive walls, would pulverize attackers. But the complacent emperor,
possessing neither the material means nor the foresight to recognize the
technology’s significance, dismissed the proposal.
Unfortunately
for him, the Hungarian engineer turned out to be a mercenary. Switching
tactics (and sides), he updated his design to be more
mobile—transportable by no fewer than 60 oxen and 400 men—and approached
the emperor’s rival, the Ottoman Sultan Mehmed II, who was preparing to
besiege the impermeable fortress. Winning the young sultan’s interest
with his claim that this gun could “shatter the walls of Babylon
itself,” the entrepreneurial Hungarian helped the Turkish forces to
breach the ancient walls in only 55 days.
The contours of this
fifteenth-century drama can be seen again and again throughout history.
In the nineteenth century, speed and mobility transformed the fortunes
first of France, as Napoleon’s army overwhelmed Europe, and then of
Prussia, under the direction of Helmuth von Moltke (the Elder) and
Albrecht von Roon, who capitalized on the newly developed railways to
enable faster and more flexible maneuvering. Similarly, blitzkrieg—an
evolution of the same German military principles—would be used against
the Allies in World War II to great and terrible effect.
“Lightning
war” has taken on new meaning—and ubiquity—in the era of digital
warfare. Speeds are instantaneous. Attackers need not sacrifice
lethality to sustain mobility, as geography is no longer a constraint.
Although that combination has largely favored the offense in digital
attacks, an AI era could see the increase of the velocity of response
and allow cyberdefenses to match cyberoffenses.
In kinetic
warfare, AI will provoke another leap forward. Drones, for instance,
will be extremely quick and unimaginably mobile. Once AI is deployed not
only to guide one drone but to direct fleets of them, clouds of drones
will form and fly in sync as a single cohesive collective, perfect in
their synchronicity. Future drone swarms will dissolve and reconstitute
themselves effortlessly in units of every size, much as elite
special-operations forces are built from scalable detachments, each of
which is capable of sovereign command.
In addition, AI will
provide similarly speedy and flexible defenses. Drone fleets are
impractical if not impossible to shoot down with conventional
projectiles. But AI-enabled guns firing rounds of photons and electrons
(instead of ammunition) could re-create the same lethal disabling
capacities as a solar storm that can fry the circuitry of exposed
satellites.
AI-enabled
weapons will be unprecedentedly exact. Limits to the knowledge of an
antagonist’s geography have long constrained the capabilities and
intentions of any warring party. But the alliance between science and
war has come to ensure increasing accuracy in instruments, and AI can be
expected to make more breakthroughs. AI will thus shrink the gap
between original intent and ultimate outcome, including in the
application of lethal force. Whether land-based drone swarms, machine
corps deployed in the sea, or possibly interstellar fleets, machines
will possess highly precise capabilities of killing humans with little
degree of uncertainty and with limitless impact. The bounds of the
potential destruction will hinge only on the will, and the restraint, of
both human and machine.
In kinetic warfare, AI will provoke a huge leap forward.
That
being so, the AI age of warfare will be reduced primarily to an
assessment not of an adversary’s capabilities but rather of its
intentions and their strategic applications. In the nuclear age, we have
already entered such a phase—but its dynamics and significance will
come into much sharper focus as AI proves its worth as a weapon of war.
With
such valuable technology involved, humans may not even be the primary
targets of AI-enabled war. AI could in fact remove humans as a proxy in
warfare entirely, making war less deadly but potentially no less
decisive. Similarly, territory alone seems unlikely to provoke AI
aggression—but data centers and other critical digital infrastructure
certainly could.
Surrender, then, will come not when the
opponent’s numbers are diminished and its armory empty but when the
survivors’ shield of silicon is rendered incapable of saving its
technological assets—and finally its human deputies. War could evolve
into a game of purely mechanical fatalities, the deciding factor being
the psychological strength of the human (or AI) who must contest to
risk, or forfeit to prevent, a breakthrough moment of total destruction.
Even
the motives governing the new battlefield would be alien, to some
extent. The English writer G. K. Chesterton once quipped that “the true
soldier fights not because he hates what is in front of him, but because
he loves what is behind him.” An AI war is unlikely to involve love or
hate, let alone a concept of soldierly bravery. On the other hand, it
may still incorporate ego, identity, and loyalty—although the nature of
those identities and loyalties may not be consistent with those of
today.
The calculation in warfare has always been relatively
straightforward: whichever side first finds intolerable the pain of
battle will likely be conquered. The consciousness of one’s own
shortcomings has in the past produced restraint. Without such awareness,
and with no sense for (and thus a great tolerance of) pain, one cannot
but wonder what, if anything, would prompt restraint in an AI that has
been introduced into warfare, and what would conclude the conflicts it
wages. A chess-playing AI, if it had never been informed of the rules
dictating the end of the game, could play to the very last pawn.
GEOPOLITICAL RESTRUCTURING
In
every age of humanity, almost as if in obedience to some natural law,
there has emerged, as one of us (Kissinger) once put it, a unit “with
the power, the will, and the intellectual and moral impetus to shape the
entire international system in accordance with its own values.” The
most familiar arrangement of human civilizations is that of the
Westphalian system as conventionally understood. The idea of the
sovereign nation-state, however, is only a few centuries old, having
emerged from treaties that are collectively known as the Peace of
Westphalia in the mid-seventeenth century. It is not the preordained
unit of social organization, and it may not be suited for the age of AI.
Indeed, as mass disinformation and automated discrimination trigger a
loss of faith in that arrangement, AI may pose an inherent challenge to
the power of national governments. Alternatively, AI may well reset the
relative positions of competitors within today’s system. If its powers
are harnessed primarily by nation-states themselves, humanity could be
forced toward a hegemonic stasis, or else toward a new equilibrium of
AI-empowered nation-states. But the technology could also be the
catalyst of an even more fundamental transition—a shift to an entirely
new system, in which state governments would in turn be forced to
abandon their central role in the global political infrastructure.
One
possibility is that the companies that own and develop AI will accrue
totalizing social, economic, military, and political power. Today’s
governments are forced to contend with their difficult position both as
cheerleaders for private corporations—lending their military power,
diplomatic capital, and economic heft to promote these homegrown
firms—and as supporters of the average citizen suspicious of
monopolistic greed and secrecy. That may prove an untenable
contradiction.
Meanwhile,
corporations could form alliances to consolidate their already
considerable strength. Those alliances might be built on complementary
advantages and the profit of amalgamation or, alternatively, on a shared
philosophy of development and deployment of AI systems. These corporate
alliances might take on traditional nation-state functions, though
rather than seeking to define and expand bounded territories, they would
cultivate diffuse digital networks as their domains.
AI may pose an inherent challenge to the power of national governments.
And
there is still another alternative. Uncontrolled, open-source diffusion
could give rise to smaller gangs or tribes with substandard but
substantial AI capacities, sufficient to administer to, provide for, and
defend themselves within some limited scope. Among human groups that
reject established authority in favor of decentralized finance,
communication, and governance, such technology-enabled proto-anarchy
could win out. Or such groupings might incorporate a religious
dimension. After all, in terms of reach, Christianity, Islam, and
Hinduism have all been larger and longer-lasting than any state in
history. In the age to come, religious denomination, more than national
citizenship, might conceivably prove the more relevant framework for
identity and loyalty.
In either future, whether dominated by
corporate alliances or diffused into loose religious groupings, the new
“territory” that each group would claim—and over which they would
fight—would not be inches of land but a digital landscape, seeking the
loyalties of individual users. Linkages between these users and any
administration would subvert the traditional notion of citizenship, and
agreements between the entities would be unlike ordinary alliances.
Historically,
alliances have been forged by individual leaders and have served to
augment a nation’s strength in case of war. By contrast, the prospect of
citizenships and alliances—and perhaps conquests or crusades—structured
around the opinions, beliefs, and subjective identities of ordinary
people in times of peace would require a new (or very old) conception of
empire. It would also force a reassessment of the obligations entailed
in pledging allegiance and the cost of exit options, if indeed any were
to exist in the AI-entangled future.
PEACE AND POWER
The
foreign policies of nation-states have been built and then adjusted by
balancing idealism and realism. The temporary balances struck by our
leaders are seen in retrospect not as end-states but as only ephemeral
(if necessary) strategies for their time. With each new age, this
tension has produced a different expression of what constitutes
political order. The dichotomy between the pursuit of interests and the
pursuit of values—or between a particular nation-state’s advantage and
the global good—has been part of this unending evolution. In the conduct
of their diplomacy, leaders of smaller states historically have
responded straightforwardly, prioritizing the necessities of their own
survival. By contrast, those responsible for global empires, with the
means to realize additional goals, have faced a more agonizing
predicament.
Since the beginning of civilization, as human units
of organization have grown, they have simultaneously achieved new levels
of cooperation. But today, perhaps because of the scale of planetary
challenges as well as to the material inequalities evident among and
within states, a backlash against this trend has surfaced. AI could
prove commensurate to the demands of this still-grander scale of human
governance, capable of seeing with granularity and fidelity not merely
the imperatives of the country but also the interplay of the globe.
We
harbor a hope that AI, deployed for political ends at home and abroad,
might do more than just illuminate balanced tradeoffs. Ideally, it could
provide new, globally optimal solutions, acting on a longer time
horizon and with greater precision than humans are capable of, and thus
bringing competing human interests into alignment. In the coming world,
machine intelligences navigating conflict and negotiating peace might
help clarify, or even surmount, traditional dilemmas.
However, if
AI were indeed to fix problems that we should have hoped to solve
ourselves, we could face a crisis of confidence—of both overconfidence
and the lack of confidence. To the former, once we understand the limits
of our own ability for self-correction, it may be difficult to admit
that we have come to cede too much power to machines in handling
existential issues of human conduct. To the latter, the realization that
simply removing human agency from the handling of our affairs has been
enough to solve our most intractable problems might reveal too
explicitly the shortcomings of human design. If peace has always been
but a simple voluntary choice, the price of human imperfection has been
paid in the coin of perpetual war. To know that a solution has always
existed but has never been conceived by us would be crushing to human
pride.
In
the case of security, unlike that of the displacement of people in
scientific or other academic endeavors, we may more readily accept the
impartiality of a mechanical third party as necessarily superior to the
self-interestedness of a human—just as humans easily recognize the need
for a mediator in a contentious divorce. Some of our worst traits will
enable us to exhibit some of our best: that the human instinct toward
self-interest, even at the expense of others, may prepare us for
accepting AI’s transcendence of the self.
HENRY A. KISSINGER served as U.S. Secretary of State from 1973 to 1977 and as U.S. National Security Adviser from 1969 to 1975.
ERIC SCHMIDT is Chair of the Special Competitive Studies Project and former CEO and Chair of Google.
CRAIG MUNDIE is the Co-Founder of Alliant Computing Systems and the former Senior Adviser to the CEO at Microsoft.
A Turkish Air Force F-16 receives a
mid air refuel from a NATO allied aircraft on Oct 23, 2018 during
exercise Trident Juncture 18. Trident Juncture is a multinational NATO
exercise that enhances professional relationships and improves overall
coordination with Allied and partner nations. (photo by Nebil; Turkish
Air Force)
A video circulating on Telegram and other social media suggests North Korean Special Forces dispatched to fight Ukraine on Russia’s behalf have seen their first combat. The video purports
to interview the single North Korean survivor from a unit of 40
compatriots who encountered Ukrainian forces near the Ukraine-occupied
Russian town of Kursk. While unclear if the video is authentic—some suggest it is psychological warfare—it is believable. While Ukrainian President Volodymyr Zelensky urges
the world to intercede before more North Korean units can reach the
battlefield, military analysts wait to see how North Koreans do in
battle. After all, despite their fearsome displays and bellicose rhetoric, it has been decades since the North Korean Army engaged in open combat. North Koreans are increasinglyshorter and lighter than their South Korean neighbors are.
North Korea is not the only country that coasts on reputation. As much as the United States fears China’s rise,
the fighting ability of the People’s Liberation Army is an unknown. It
is perhaps the only army in the world entirely comprised of only
children. The last time the People’s Republic of China fought an open
war—a month-long conflict with Vietnam in 1979—China lost. Since then,
China has only engaged small and unarmed or only lightly armed
opponents—Filipino coast guard speedboats, Vietnamese fishing boats, or
small squads of Indian soldiers high up in the Himalayas. China can
bluster about conquering Taiwan.
The People’s Liberation Army can cause incredible devastation with
missiles and drones, but their ability to occupy the country is a
different matter. The second the People’s Liberation Army engages,
Beijing knows, their carefully crafted image of invincibility might
crater.
Russia has been another paper tiger. As Russian forces massed on the Ukrainian border in February 2022, President Joe Biden
and National Security Advisor Jake Sullivan urged Zelensky to surrender
preemptively and flee the country. The intelligence briefings Biden and
Sullivan received from the U.S. intelligence community grossly
exaggerated the capabilities of the Russian army. Rather than march
triumphantly through Kyiv, the Russian Army
today loses every two months more than the United States lost in the
entire Vietnam War. Russia might still win, but not as it envisioned.
Rather, it will simply seek to outlast its Ukrainian opponents in a new
Stalingrad. The Kremlin might have sold an image of itself as a first
world military capable of shock and awe, but what it showed the world
was not much different from Russian forces in World War I.
The Saddam-era Iraqi Army was also a paper tiger. In the lead up to
the 1991 Operation Desert Storm and the liberation of Kuwait, Iraq had
the fifth largest army in the world. Colin Powell, then the chairman of
the Joint Chiefs of Staff, confused size with competence and sought a
diplomatic compromise with Iraq, a proposal President George H.W. Bush
wisely rejected. The subsequent decimation of Iraq’s army showed what a
paper tiger they had been.
While North Korea, China, Russia, and Saddam-era Iraq are or were all U.S. adversaries, the same dynamics may also apply to NATO.
The Turkish military forms the second largest force component within
NATO, after the United States. Diplomats, analysts, and Turkey’s
lobbyists on K Street and in Washington think tanks conflate Turkey’s military power and strategic importance, but seldom consider if Turkey’s military power is real.
Turkish President Recep Tayyip Erdogan took a formidable military and
sought to remake it in his own image. To expedite its transformation,
he cited various conspiracies as fact, culminating in the “Reichstag Fire coup.” The result of Erdogan’s purges has been the prioritization of politics over competence. One-in-five Turkish F-16 pilots, for example, ended up in prison; their replacements had a fraction of their imprisoned colleagues’ experience.
The same is true with Turkey’s Ground Forces. While Turkish forces
have pushed into Syria into some Kurdish districts of Syria, they only
do so with proxies or against lightly armed opponents. While Turkey has
waged war against Syrian Kurds’ civilian infrastructure such as oil
pipelines and electrical substations or Yezidi farms across the border
in Iraq, the Turkish military has failed to engage the Islamic State.
There are two possible explanations for this: Either Turkey as a whole
or certain commanders do not consider the Islamic State to be an enemy
or Turkish commanders fear directly engaging the Islamic State would
expose the weakness of Turkey’s post-Erdogan ground forces.
Here, Libya provides some clues where the Wagner Group effectively
has held NATO’s second largest military to a draw. The state-controlled
Turkish media will cite its air support for Azerbaijan’s assault on
Armenians or provision of drones to Ukraine, but neither of these
supposed successes involved the deployment of Turkish troops with the
exception perhaps of some Special Forces to Nagorno-Karabakh.
Too often, the United States self-deters in the face of paper tigers
in a way that empowers them. In effect, the United States might hold a
full house militarily, but it folds before a pair of twos. Turkey
today presents another problem, however, since it is nominally an ally
rather than adversary. While Erdogan seeks benefit from an illusion of
strength, it is time to question whether the size of Turkey’s military
mattes if it has effectively become a third world force, little
different than Iraq 1991 or Russia 2022. If so, then perhaps the next
administration must recalculate the deference to which Turkey is due and
even such basic questions about whether the Syrian Kurds, if properly
armed, can contribute more to regional security than Turkish troops
whose illusion of power will dissipate the moment they leave their
barracks.
Michael Rubin is a senior fellow at the American Enterprise Institute and director of policy analysis at the Middle East Forum.
A former Pentagon official, Dr. Rubin has lived in post-revolution
Iran, Yemen, and pre-and postwar Iraq. He also spent time with the
Taliban before 9/11. For over a decade, he taught classes at sea about
the Horn of Africa and Middle East conflicts, culture, and terrorism to
deployed US Navy and Marine units. Dr. Rubin is the author, coauthor,
and coeditor of several books exploring diplomacy, Iranian history, Arab
culture, Kurdish studies, and Shi’ite politics. The opinions and views
expressed are his own.
Michael Rubin is a senior fellow at the American Enterprise
Institute and director of policy analysis at the Middle East Forum. A
former Pentagon official, Dr. Rubin has lived in post-revolution Iran,
Yemen, and both pre- and postwar Iraq. He also spent time with the
Taliban before 9/11. For more than a decade, he taught classes at sea
about the Horn of Africa and Middle East conflicts, culture, and
terrorism, to deployed US Navy and Marine units. Dr. Rubin is the
author, coauthor, and coeditor of several books exploring diplomacy,
Iranian history, Arab culture, Kurdish studies, and Shi’ite politics.