The battlefield is changing as artificial intelligence (AI) becomes a part of military operations. AI technologies are making military strategies better by processing data and running autonomous systems faster. This allows for quicker and more effective decisions during combat. Amir Husain and other experts believe it’s crucial to look into the dangers of AI in warfare. They highlight concerns about ethics and how these technologies are used in battle. Countries like China and Saudi Arabia are heavily investing in AI and autonomous weapons, aiming to lead in this area.
Experts predict that by 2035, 70% of the US Air Force could be remotely piloted aircraft. This shift is huge and comes with many challenges. It’s sparking debates about the risks of artificial intelligence. There’s a need for detailed rules to handle these issues. We will explore how AI is changing warfare, the risks involved, and the moral questions it raises.
Key Takeaways
- AI is transforming military strategies and operations, with significant implications for decision-making.
- Investment in AI technologies is rapidly increasing among global military powers.
- The rise of autonomous weapons systems raises ethical and operational challenges.
- Understanding AI risks is essential for ensuring effective and responsible military engagement.
- The future of warfare will increasingly intertwine AI with traditional combat methods.
Introduction to AI in Modern Warfare
The way wars are fought is drastically changing with the Introduction to AI. This term refers to technology that lets machines think like humans. In the military, AI is used for analyzing data and controlling combat systems autonomously.
AI-augmented weapons are a key example. They include the autonomous rifle used in the 2020 assassination of an Iranian scientist. This rifle could shoot six hundred rounds a minute. It shows how Modern Warfare and Impact of AI on Warfare are evolving.
AI is being used by countries like Russia and China to boost their military strength. This change is altering global power dynamics. The United States stresses the careful use of AI in the military. It has set ethical guidelines for AI in 2020 and updated them for autonomous systems in 2023.
AI can improve military operations and quick decision-making. Yet, there are big challenges. Adversaries can attack AI systems in ways that trick or bypass them. This shows the need for strong defense against new dangers.
The US Department of Defense is working hard on this. They have a plan called the “Responsible Artificial Intelligence Strategy and Implementation Pathway.” It aims to train people in AI and manage its risks in the military. Combining AI with current military technology promises to make data-based decisions better. This can change how efficient and effective modern wars are fought.
The Rise of Autonomous Weapons Systems
The emergence of Autonomous Weapons Systems (AWS) is changing military strategy and operations. It’s key to understand the Definition of AWS to fully grasp their impact. AWS are weapons that decide and attack targets without humans, raising big questions about who is in charge and responsible in AI in Armed Conflict.
Definition of Autonomous Weapons
Autonomous weapons work with complex algorithms and machine learning, enabling them to fight without human control. They excel in tasks that are dull, dirty, or dangerous. Examples include disposing of bombs or keeping watch for long hours. According to the Department of Defense’s Unmanned Systems Roadmap, they could make soldiering less risky and change how enemies are fought. However, using machine learning comes with challenges like bias and unpredictability.
Global Perspectives on Automated Weapons
Different countries have varied views on autonomous weapons. Some are enhancing AWS, while others want rules to keep human decision-making in the loop. The International Committee of the Red Cross (ICRC) argues that these systems shouldn’t act unpredictably or target people directly. More than 3,000 AI and robotics experts have raised alarms about moving too fast with these technologies. They worry about an AI arms race and the ethical dilemmas of using such systems without strict rules.
There’s a lot of talk about how autonomous weapons are being used now, like in Libya where they’re already in play. As nations step up their game in this area because of global tensions, creating new international laws for AWS is becoming more important.
Understanding Hyperwar: The New Landscape of Conflict
The idea of Hyperwar marks a big change in the way conflicts are seen. It’s all thanks to tech and AI improvements. Nations are spending a lot to strengthen their defense systems with new tech. This push into the future is changing how military strategies are made, especially with AI-fueled warfare.
AI is changing how wars are fought by making decisions faster than people can. It lets armies work better in different places, like social media and spy satellites. Whoever leads in AI tech might become a world power. So, using this new tech right can help in fights and keep peace.
Even with new tech, old military ways are still important. Nuclear weapons and cyber attacks play big roles in power. The competition to be the strongest shows why armies need to keep getting better.
With Hyperwar coming, quick thinking is key in military plans. Leaders need to make fast, smart choices, using AI to help. But this focus on speed brings up tough questions about ethics and how we talk about AI in war.
AI Risks in Warfare: What to Consider
The use of AI in warfare brings big challenges. It can change battle outcomes in unexpected ways. It is important to think carefully about the Potential Hazards of AI in combat. AI, especially when it learns by itself, can make choices without human input. This raises Ethical AI Concerns. Using these technologies wisely means keeping a close watch and setting strict rules.
Potential Hazards of AI in Battle
AI is changing fast and this adds risks in the military world. Key Potential Hazards of AI include:
- Mistakes in decisions by AI might cause unintentional harm.
- Conflicts might grow quickly as machines make moves too fast for people to keep up.
- Learning from flawed data could lead to biased or unfair actions in military missions.
- A challenge is keeping clear who is responsible because AI algorithms are complex.
Ethical AI Concerns Surrounding Automated Decision Making
Talking about right and wrong in war matters more when Ethical AI Concerns come up. Main concerns are:
- It’s a big moral concern to let machines decide on life and death.
- It’s crucial that people can step in during critical moments.
- Worries about worldwide safety from an AI arms race are similar to fear from the nuclear arms races before.
- We need strict safety rules to avoid taking shortcuts in making AI.
The Role of Machine Learning in Military Strategy
Machine Learning is changing how militaries operate by boosting their ability to perceive and decide. It lets them quickly examine tons of data to make smarter strategic decisions. With generative AI, this growth is even faster, giving leaders new ways to beef up old methods.
Enhancements in Perception and Decision Making
Machine Learning greatly improves AI’s ability to understand things. It digs through various data sources like satellite pictures, social media, and intercepted messages. This helps military experts get a full view of the battlefield and better understand the situation. They can also spot trends in enemy tactics, helping to plan better and even guess where foes will move next.
Examples of Machine Learning Applications in Warfare
Here are some ways machine learning helps in military efforts:
- Autonomous Drones: Drones that can think for themselves spot and follow enemy activities on their own, covering more ground efficiently.
- Data Processing: AI helps sift through huge amounts of data quickly to find important trends, influencing military strategies.
- Combat Simulations: Better simulation tech makes training feel more real, getting soldiers ready for actual combat.
- Threat Monitoring: AI systems keep an eye on different dangers, helping to stop threats before they happen.
- Cybersecurity: Machine learning boosts defenses against digital attacks, protecting vital military systems.
Machine learning is becoming a key part of military strategy upgrades. By adopting these techs, armies can operate more smoothly and tackle the complex issues of modern warfare effectively.
Technology Threats: How AI Changes the Game
AI technology is changing warfare in big ways, bringing new threats. As countries use more AI in battle, they face new weaknesses. These can be used against them. One big worry is cybersecurity. Enemies could hack into military networks. This could lead to spying or breaking crucial communication links.
AI changes how decisions are made in war. It can look at lots of data quickly. This worries people about how enemies might use it. They could uncover secrets or launch hidden attacks. AI’s quick data analysis changes military strategy. Countries need to rethink their defense plans.
Sometimes AI can be biased because of bad data. This can lead to dangerous mistakes in military choices. Biased AI concerns people because it might not be fair or dependable. This questions the ethics of AI in military decisions. It’s a big issue for safety and fairness.
Military leaders must adjust to AI in warfare. They need to see both the good and bad sides. To deal with these advanced threats, they should improve cybersecurity. They must also make sure AI is used ethically. Understanding and managing these new challenges is key to modern warfare success.
AI Bias Issues Affecting Military Operations
Adding AI to Military Operations has its challenges, especially with AI Bias Issues. Biases can sneak in at many stages like data gathering, designing, developing, and deploying. These biases can mess up decision-making. They can even cause legal and ethical problems. This is really important when decisions could affect people’s lives.
Implications of Bias in AI Decision Making
Biases in military AI systems are a big deal. Biases from humans can sneak into AI, leading to automation bias. This means people might trust AI too much. This trust can make mistakes like wrong identifications or conclusions based on AI’s advice. Plus, the data AI uses might not be diverse enough. This can make AI less accurate in the military than in civilian life.
Biases can spread throughout the machine learning cycle from start to end. Studies show that fixing these biases is tricky. It’s not only about how AI is used. It requires a deep dive into the technical fixes that haven’t been fully figured out yet. Another issue is the lack of diversity among the tech folks working on AI. This can make AI biases worse. That’s why it’s key to focus on making AI decisions fair and ethical. This will help avoid problems during crucial military operations.
Data Privacy Risks in AI Warfare
The use of AI in war raises big Data Privacy Risks. These risks come from the need for large datasets in machine learning. When the military gathers a lot of data, the risk of leaks increases.
AI mishaps have shown how easily private data can get exposed. This situation makes it clear we need strong privacy protections.
To handle AI in Warfare, steps must be taken against cybersecurity threats. Measures like encryption and secure access are key. They help keep sensitive data safe from hackers.
Collecting too much data can also threaten privacy. So, it’s important to collect only what’s needed and keep it anonymous. Checking these practices regularly can reduce privacy risks.
To govern AI well, we need advice from many fields. This lets us check data for errors or biases continually. Using fake data can help make AI models fairer. Doing tests called red team exercises can also improve security against complex attacks.
New laws like the European Union’s GDPR and U.S. orders aim to protect data privacy. They offer guidelines and support AI’s ethical use. As AI’s role in the military grows, keeping data safe becomes even more crucial to avoid risks.
The OODA Loop and AI Integration
The OODA Loop was created by US Air Force Colonel John Boyd. It stands for Observe, Orient, Decide, and Act. This framework is key in military decisions, helping forces react quickly during battles. With AI, every stage of the OODA Loop improves, leading to faster, more informed decisions in combat.
How AI Facilitates Faster Decision Making
AI works with military systems to make decisions faster. It uses sensors and surveillance to gather and analyze lots of data quickly. This gives clear, up-to-date information about the situation. By looking at past data, AI helps predict what opponents might do next.
AI also works on the front lines to make systems more flexible and quick. It looks at different actions and assesses risks, helping commanders choose the best path. This is crucial in quick, uncertain situations.
Potential Challenges with AI in the OODA Loop
However, using AI in the OODA Loop has its problems. There are risks like accidental escalation and AI making decisions we can’t explain. Also, if the AI training data is not good, it can lead to biased decisions. These issues make trusting AI difficult in tense situations.
As we mix AI with military plans, we must find the right speed and quality of decisions. It’s important to ensure AI helps, not replaces, human judgment. This balance is essential for effective military strategies.
The Future of AI and Cybersecurity Risks
Technology is always evolving, and so is the link between AI and Cybersecurity. AI could greatly improve how we protect our nation. It does this through advanced predictive analytics. These tools help find potential threats before they even happen.
This means we can manage risks better and react to cyber threats faster. It’s a step forward in keeping ourselves safe in the digital age.
AI’s Role in Strengthening National Security
AI helps make our nation’s security stronger. It does this by making systems that can do tasks on their own, find harmful software, and predict future risks. Companies like Trend Micro use AI to find and deal with risks faster.
Other companies, like Normalyze, work on keeping data safe in different places. Proofpoint fights against email threats. Then, there’s Torq’s “HyperSOC,” which shows how using a lot of automation makes defense better and faster against tough threats.
Weaknesses in AI Security that Could Be Exploited
But AI isn’t perfect. It has its weaknesses. Designing, setting up, and teaching AI systems can be very pricey. Smaller and bigger businesses might see their costs go up when they start using AI.
Also, we don’t have enough AI experts for all the jobs out there. This means there’s a gap in skills for handling these advanced systems. AI’s weaknesses change as threats get more complex. We need to keep checking and improving our security to stay ahead.
International Regulations and AI Use in Warfare
The world of warfare is changing. We need strong International Regulations for AI use. As countries and groups include AI in military, it’s important to have rules. The United Nations is working on AI Warfare Guidelines to control this tech.
The Role of the UN in Establishing Guidelines
The UN sees the need to tackle AI in the military. They started the “Political Declaration on Responsible Military Use of Artificial Intelligence and Autonomy” with 47 states joining. These efforts aim to make sure AI in warfare is used ethically and responsibly. The UN’s “AI for Good” meetings also focus on using AI right, to avoid making global conflicts worse.
Current Gaps in Legal Frameworks for Automated Weapons
There are still big gaps in rules for automated weapons. The EU’s AI Act tried to set regulations but left out military uses. This omission is worrying as technology moves fast. Tech companies becoming part of military projects, like in Ukraine’s conflict, makes regulation harder.
AI in the military isn’t just about robots that can fight. It’s also key in cybersecurity, making big decisions, and other operations. As AI becomes more common, how to keep humans in charge is debated. This shows how hard it is to make rules that can keep up with tech changes.
Case Studies: AI in Recent Conflicts
Understanding AI in modern warfare is key. It shows us how wars are changing. Many militaries now use AI to get ahead. This has changed how battles are fought. The case studies tell us how different countries use new tech to be stronger.
Ukraine’s Use of AI-Enhanced Drones
Ukraine has changed its defense with AI drones. These drones help find targets and gather info. They make attacks precise and safer for soldiers. Ukraine’s use of drones shows tech can change war and make forces more effective.
The Impact of AI in the Israel-Hamas Conflict
The Israel-Hamas fight is seen as the first “AI war.” This marks a big change in how wars are fought. AI helped make decisions faster, especially about where and when to attack. This use of AI asks big questions about how to keep civilians safe in war. It shows we need rules to use this powerful tech responsibly.
The Arms Race in AI Development
The world is seeing a race in AI development. Countries want to boost their military strength with new tech. They focus on AI and autonomous weapons. This race is changing the world power structure. It shows how crucial it is to stay in front in AI.
Comparative Spending on AI and Autonomous Weapons
The U.S. Department of Defense spent about $5 billion on AI in 2020. This was just a small part of its huge $700 billion budget. Countries find it hard to say how much they spend on military AI. This is because AI can be used in many ways. Although China is pushing forward, it’s still behind the U.S. in creating very advanced AI or AGI.
- China’s big AI projects are 1-3 years behind the U.S.
- There are worries about rushing unsafe AI into use because of competition.
- Leading AI experts call for careful development, putting safety first.
Geopolitical Implications of AI in Military Strategy
The race to spend on AI development has big global effects. Countries want to lead in military AI to stay strong. This situation fuels fears of falling behind. It makes it tough to regulate AI use worldwide. Although we’re not seeing a classic arms race in AI investment, it still poses a security challenge. Countries enhance their defenses, often without knowing the full impact.
Even if military AI spending isn’t skyrocketing, the stakes in global politics are high. Countries have to deal with a tough competition. They must ensure AI advances without risking safety or peace.
Future Trends in AI and Warfare
The way we fight wars is changing fast, with new tech playing a big role. Military plans are now taking AI into account. This change is crucial for keeping countries safe. Modern tech like drones and robots are leading to more automation in battle. This forces us to think over our military plans.
Emerging Technologies and their Potential Impact
Deputy Defense Secretary Kathleen Hicks started the Replicator initiative. It plans to bring lots of AI to the battlefield soon. The goal is to have thousands of smart weapons systems ready to use. This shows how serious the U.S. is about AI in combat.
The war in Ukraine is known as the first big drone war. It shows us how important quick adaptation is in military strategy. Using drones in fights is becoming more common. This change is shaping the future of how wars are fought.
There’s a global talk about making rules for AI weapons that act on their own. U.N. Secretary-General António Guterres wants new laws in three years. These laws would limit these kinds of weapons. The Red Cross and many countries agree. They say we need clear rules to make sure these weapons are used right.
In the future, tech could help make fairer decisions on the battlefield. Armed forces are looking into this tech. They hope it will make them better at hitting targets while obeying the law. This path to a tech-driven battlefield requires ongoing strategy updates. It’s key to making the most of these new tools.
Ethical Considerations in the Age of AI Warfare
AI in warfare brings up major ethical issues that need careful review. As AI takes on roles once done by people, we must consider the ethics of letting machines make combat decisions. The use of autonomous weapons brings up questions about who is responsible and the chances of them being misused.
The Moral Implications of Autonomous Decision Making
AI-driven warfare moves critical choices to algorithms. This leads to dilemmas about missing human judgment in vital moments. When AI systems decide, who is accountable becomes a significant question. AI’s quick and complicated decision-making might lower our moral standards. This makes it crucial to talk deeply about the ethics of AI in warfare.
Regulatory Challenges within Warfare Ethics
It’s important to create effective rules for AI use in the military. Today’s laws can’t always keep up with how fast AI is advancing. Experts argue we need regulations focused on responsibility, clarity in operation, and ethical responsibility. Ongoing discussions try to set rules that keep these technologies within human rights and ethical standards.
Conclusion
The study of AI in warfare shows a world of great chances and big challenges. Technology keeps getting better, offering new ways for the military to make decisions and work more efficiently. Yet, we must be careful. It’s vital to have strong ethical and rule-based systems for AI in the military. These systems help us deal with issues like losing jobs and increasing biases.
Groups and governments need to talk more about making AI in warfare safer and more responsible. Working together is key to lessen the dangers and make the most of what AI offers. By making laws that tackle these problems and guarantee fair access to technology, we protect everyone involved.
To wrap up, it’s really important to know the risks and benefits of AI in war. The technology could change how we fight completely. But, we must pay close attention to its effects on jobs, fairness, and moral choices. As we bring more AI into warfare, making sure we don’t lose sight of ethical values and people’s well-being is crucial.
FAQ
Q: What are the primary AI risks associated with warfare?
Q: How are autonomous weapons systems defined?
Q: What is the concept of “Hyperwar” in military strategy?
Q: How does machine learning enhance military operations?
Q: What are the ethical concerns surrounding AI in warfare?
Q: What data privacy risks are associated with AI in warfare?
Q: How does the OODA loop integrate AI technology?
Q: Why is there concern over AI bias in military operations?
Q: How can AI be used to improve national security?
Q: What are some existing gaps in international regulations regarding AI warfare?
Source Links
- AI & The Future of Conflict | GJIA – https://gjia.georgetown.edu/2024/07/12/war-artificial-intelligence-and-the-future-of-conflict/
- AI is Shaping the Future of War – https://ndupress.ndu.edu/Media/News/News-Article-View/Article/2846375/ai-is-shaping-the-future-of-war/
- Artificial Intelligence and Future Warfare – https://www.hudson.org/defense-strategy/artificial-intelligence-future-warfare
- Artificial Intelligence, Real Risks: Understanding—and Mitigating—Vulnerabilities in the Military Use of AI – Modern War Institute – https://mwi.westpoint.edu/artificial-intelligence-real-risks-understanding-and-mitigating-vulnerabilities-in-the-military-use-of-ai/
- A.I. Joe: The Dangers of Artificial Intelligence and the Military – Public Citizen – https://www.citizen.org/article/ai-joe-report/
- What you need to know about autonomous weapons – https://www.icrc.org/en/document/what-you-need-know-about-autonomous-weapons
- Pros and Cons of Autonomous Weapons Systems – https://www.armyupress.army.mil/Journals/Military-Review/English-Edition-Archives/May-June-2017/Pros-and-Cons-of-Autonomous-Weapons-Systems/
- Problems with autonomous weapons – https://www.stopkillerrobots.org/stop-killer-robots/facts-about-autonomous-weapons/
- PDF – https://researchcentre.trtworld.com/wp-content/uploads/2023/11/Hyperwar.pdf
- SparkCognition Releases "Hyperwar: Conflict and Competition in the AI Century" at Time Machine 2018 – https://www.prnewswire.com/news-releases/sparkcognition-releases-hyperwar-conflict-and-competition-in-the-ai-century-at-time-machine-2018-300744381.html
- Op-ed: Hyperwar is coming. America needs to bring AI into the fight to win — with caution – https://www.cnbc.com/2020/07/12/why-america-needs-to-bring-ai-into-the-upcoming-hyperwar-to-win.html
- Militarization of AI Has Severe Implications for Global Security and Warfare – https://unu.edu/article/militarization-ai-has-severe-implications-global-security-and-warfare
- AI Risks that Could Lead to Catastrophe | CAIS – https://www.safe.ai/ai-risk
- AI and International Stability: Risks and Confidence-Building Measures – https://www.cnas.org/publications/reports/ai-and-international-stability-risks-and-confidence-building-measures
- Military Applications of AI in 2024- Sentient Digital, Inc. – https://sdi.ai/blog/the-most-useful-military-applications-of-ai/
- The Role of AI and Machine Learning in Modern Military Intelligence – https://www.linkedin.com/pulse/role-ai-machine-learning-modern-military-intelligence-monser-kernosh-i1pte
- AI—The good, the bad, and the scary – https://eng.vt.edu/magazine/stories/fall-2023/ai.html
- How Will AI Affect the Gaming Industry? – University of Silicon Valley – https://usv.edu/blog/how-will-ai-affect-the-gaming-industry/
- The problem of algorithmic bias and military applications of AI. – https://blogs.icrc.org/law-and-policy/2024/03/14/falling-under-the-radar-the-problem-of-algorithmic-bias-and-military-applications-of-ai/
- The Machine Got it Wrong? Uncertainties, Assumptions, and Biases in Military AI – https://www.justsecurity.org/95630/biases-in-military-ai/
- The Dark Side of AI Data Privacy – https://coalfire.com/the-coalfire-blog/the-dark-side-of-ai-data-privacy
- Protecting Data Privacy as a Baseline for Responsible AI – https://www.csis.org/analysis/protecting-data-privacy-baseline-responsible-ai
- Risks of AI & Cybersecurity | Risks of Artificial Intelligence – https://www.malwarebytes.com/cybersecurity/basics/risks-of-ai-in-cyber-security
- AI and the OODA loop: How AI enhances strategic decisions for today’s warfighters – Military Embedded Systems – https://militaryembedded.com/ai/big-data/ai-and-the-ooda-loop-how-ai-enhances-strategic-decisions-for-todays-warfighters
- Speeding Up the OODA Loop with AI – Joint Air Power Competence Centre – https://www.japcc.org/essays/speeding-up-the-ooda-loop-with-ai/
- Stuck between real and ideal: the OODA loop and military AI – Luftled – https://luftled.info/stuck-between-real-and-ideal-the-ooda-loop-and-military-ai/
- The Impact of AI on Cybersecurity: Predictions for the Future | UpGuard – https://www.upguard.com/blog/the-impact-of-ai-on-cybersecurity
- AI Is The Past, Present And Future Of Cybersecurity – https://www.forbes.com/sites/tonybradley/2024/05/17/ai-is-the-past-present-and-future-of-cybersecurity/
- Governing Military AI Amid a Geopolitical Minefield – https://carnegieendowment.org/research/2024/07/governing-military-ai-amid-a-geopolitical-minefield
- U.S. Endorses Responsible AI Measures for Global Militaries – https://www.defense.gov/News/News-Stories/Article/article/3597093/
- International Community Must Urgently Confront New Reality of Generative, Artificial Intelligence, Speakers Stress as Security Council Debates Risks, Rewards – https://press.un.org/en/2023/sc15359.doc.htm
- The risks and inefficacies of AI systems in military targeting support – https://blogs.icrc.org/law-and-policy/2024/09/04/the-risks-and-inefficacies-of-ai-systems-in-military-targeting-support/
- AI Risks: Exploring the Critical Challenges of Artificial Intelligence | Lakera – Protecting AI teams that disrupt the world. – https://www.lakera.ai/blog/risks-of-ai
- A Race to Extinction: How Great Power Competition Is Making Artificial Intelligence Existentially Dangerous – https://hir.harvard.edu/a-race-to-extinction-how-great-power-competition-is-making-artificial-intelligence-existentially-dangerous/
- Debunking the AI Arms Race Theory – Texas National Security Review – https://tnsr.org/2021/06/debunking-the-ai-arms-race-theory/
- AI and the Future of Drone Warfare: Risks and Recommendations – https://www.justsecurity.org/89033/ai-and-the-future-of-drone-warfare-risks-and-recommendations/
- Will AI fundamentally alter how wars are initiated, fought and concluded? – https://blogs.icrc.org/law-and-policy/2024/09/26/will-ai-fundamentally-alter-how-wars-are-initiated-fought-and-concluded/
- The Ethical Considerations of Artificial Intelligence | Capitol Technology University – https://www.captechu.edu/blog/ethical-considerations-of-artificial-intelligence
- Ethical Issues of AI – https://www.ncbi.nlm.nih.gov/pmc/articles/PMC7968615/
- PDF – https://www.europarl.europa.eu/RegData/etudes/STUD/2020/634452/EPRS_STU(2020)634452_EN.pdf
- 14 Dangers of Artificial Intelligence (AI) | Built In – https://builtin.com/artificial-intelligence/risks-of-artificial-intelligence
- The 15 Biggest Risks Of Artificial Intelligence – https://www.forbes.com/sites/bernardmarr/2023/06/02/the-15-biggest-risks-of-artificial-intelligence/