Taxonomía de la inteligencia artificial en el entorno cognitivo de los conflictos

  1. Flores Vivar, Jesús Miguel 1
  2. Gómez-de-Ágreda, Ángel
  3. Gómez-López, Jacinto 1
  1. 1 Universidad Complutense de Madrid
    info

    Universidad Complutense de Madrid

    Madrid, España

    ROR 02p0gd045

Revista:
Disertaciones: Anuario electrónico de estudios en Comunicación Social

ISSN: 1856-9536

Any de publicació: 2023

Títol de l'exemplar: Public Communication and Armed Conflicts: The Role of Cognition and the Algorithm in Defense and National Security Environments

Volum: 16

Número: 2

Tipus: Article

DOI: 10.12804/REVISTAS.UROSARIO.EDU.CO/DISERTACIONES/A.12804 DIALNET GOOGLE SCHOLAR lock_openDialnet editor

Altres publicacions en: Disertaciones: Anuario electrónico de estudios en Comunicación Social

Resum

The article focuses on the state of the art and analysis of the artificial intelligence (ai) paradigm as applied to warfare. It shows how contemporary wars are largely fought outside the battlefields, even in non-physical environments, including cybernetic and cognitive ones. In these two environments, there is intense use of ai to alter the will of adversaries, mainly through influence and disinformation tools. The methodology used is based on a literature review and analysis of documents prepared by experts and reports from institutions dedicated to treating cybersecurity, defense, and warfare in developed countries. Here, we propose a taxonomy of warfare and malicious applications of ai that dispenses with the technological characteristics of ai to focus on its operational application. This does not result in a closed set of options but branches into numerous alternatives that can be developed in other works. The primary conclusion is that countries should consider the potential of ai to develop tactical and strategic actions and prepare for unconventional scenarios.

Referències bibliogràfiques

  • Aguado, J. M. (2020). Mediaciones ubicuas. Gedisa.
  • Aguado, J. M., Feijóo, C., & Martínez Martínez, I. J. (2013). La comunicación móvil hacia un nuevo ecosistema digital. Gedisa.
  • Allen, G., & Chan, T. (2017). Artificial Intelligence and National Security: A study on behalf of Dr. Jason Matheny, director of the U.S. Intelligence Advanced Research Projects Activity (iarpa). Belfer Center for Science and International Affairs-Harvard Kennedy School.
  • Allcott, H., & Gentzkow, M. (2017). Social media and fake news in the 2016 election. Journal of Economic Perspectives, 31(2), 211-236. https://doi.org/10.1257/jep.31.2.211
  • Aslama-Horowitz, M. (2019). Disinformation as warfare in the digital age: dimensions, dilemmas, and solutions dimensions, dilemmas, and solutions. Journal of Vincentian Social Action, 4(2). https://scholar.stjohns.edu/jovsa/vol4/iss2/5
  • Ayerbe, A. (2020). La ciberseguridad y su relación con la inteligencia artificial. ari (128). https://media.realinstitutoelcano.org/wp-content/uploads/2021/10/ari128-2020-ayerbe-ciberseguridad-y-su-relacion-con-inteligencia-artificial.pdf]
  • Baker-Brunnbauer, J. (2023). Introduction. En Trustworthy artificial intelligence implementation: Business guides on the go. Springer. https://doi.org/10.1007/978-3-031-18275-4_1
  • Barnes, C., & Barraclough, T. (2019). Perception Iinception: Preparing for deepfakes and the synthetic media of tomorrow. The Law Foundation.
  • Bonfanti, M. E. (2020). The weaponisation of synthetic media: What threat does this pose to national security? ari, 93. http://www.realinstitutoelcano.org/wps/portal/rielcano_en/contenido?WCM_GLOBAL_CONTEXT=/elcano/elcano_in/zonas_in/ari93-2020-bonfanti-weaponisation-of-synthetic-media-whatthreat-does-this-pose-to-national-security
  • Boulanin, V. V. V. (2018). The impact of artificial intelligence on strategic stability and nuclear risk. Sipri, I(May). https://doi.org/10.2991/icsshe-18.2018.203
  • Boyer, P. (2018). Minds make societies: How cognition explains the world humans create. Yale University Press.
  • Brundage, M., Avin, S., Clark, J., Toner, H., Eckersley, P., Garfinkel, B., Dafoe, A., Scharre, P., Zeitzoff, T., Filar, B., Anderson, H., Roff, H., Allen, G. C., Steinhardt, J., Flynn, C., HÉigeartaigh, S. Ó., Beard, S., Belfield, H., Farquhar, S., … Amodei, D. (2018, febrero). The malicious use of artificial intelligence: forecasting, prevention, and mitigation. https://arxiv.org/pdf/1802.07228
  • Bughin, J., Hazan, E., Ramaswamy, S., Chui, M., Allas, T., Dahlström, P., Henke, N., & Trench, M. (2017). Artificial intelligence: The next digital frontier? [discussion paper]. McKinsey Global Institute (mgi).
  • Calvo-Albero, J. L. (2020). Implicaciones en el ámbito cognitivo en las operaciones militares. Centro Superior de Estudios de la Defensa Nacional. https://emad.defensa.gob.es/Galerias/CCDC/files/IMPLICACIONES_DEL_AMBITO_COGNITIVO_EN_LAS_OPERACIONES_MILITARES.pdf
  • ccw. (2018). Chair’s summary of the discussion. 6(April), 1-13.
  • Chesney, R., & Citron, D. (2018, 11 de diciembre). Deepfakes and the new disinformation war. Foreign Affairs. https://www.foreignaffairs.com/articles/world/2018-12-11/deepfakes-and-new-disinformation-war
  • Debord, G. (2005). La sociedad del espectáculo. Pre-Textos.
  • Deptula, D. A. (2022). A new battle command architecture for Joint all-domain operations. æther: A Journal of Strategic Airpower & Spacepower, 1(1), 51-56. https://www.airuniversity.af.edu/Portals/10/AEtherJournal/Journals/Volume-1_Issue-1/08-Deptula.pdf?source=GovD
  • Flores-Vivar, J., & García-Peñalvo, F. (2023). Reflections on the ethics, potential, and challenges of artificial intelligence in the framework of quality education (SDG4). Comunicar, 74, 37-47. https://doi.org/10.3916/C74-2023-03
  • Future of Life Institute. (2020). Asilomar AI principles. Future of Life Institute, 1-49. https://futureoflife.org/ai-principles/?submitted=1#confirmation
  • Gartner. (2018, 24 de julio). Hype CYCLe for artificial intelligence, 2018. https://www.gartner.com/en/documents/3883863
  • Gómez-de-Ágreda, Á. (2019). Mundo Orwell: Manual de supervivencia para un mundo hiperconectado. Ariel.
  • Gómez-de-Ágreda, Á. (2020). Ethics of autonomous weapons systems and its applicability to any AI systems. Telecommunications Policy, 44(6), 1-15. https://doi.org/10.1016/j.telpol.2020.101953
  • Gómez-de-Ágreda, Á., & Feijoo, C. (2021). Hacia una ética del ecosistema híbrido del espacio físico y el ciberespacio. Diecisiete (4), 47-58. https://doi.org/10.36852/2695-4427_2021_04.02
  • Gómez-de-Ágreda, Á., Feijóo, C., & Salazar, I. (2021). Una nueva taxonomía del uso de la imagen en la conformación interesada del relato digital: Deep fakes e inteligencia artificial. El Profesional de la Información, 30(2). https://doi.org/10.3145/epi.2021.mar.16
  • Gómez-de-Ágreda, Á., Martínez, J. M., Mohíno Herranz, I., Barragán Montes, R., Marín Gutierrez, F. A., Cubeiro Cabello, E., & Aznar Lahoz, J. L. (2019). Usos militares de la inteligencia artificial, la automatización y la robótica (iaa&r). Centro Superior de Estudios de la Defensa Nacional. http://www.ieee.es/contenido/noticias/2019/11/DIEEET04-2019InteligenciaRobotica.html
  • Gómez-López, J. (2019). Tecnologías de la información y los mensajes en los nuevos espectros del conflicto. Revista de Ciencias de la Comunicación e Información, (24), 45-56.
  • Han, B. C. (2021). No-cosas: Quiebras del mundo de hoy. Taurus.
  • Helmus, T. C. (2022, julio). Artificial intelligence, deepfakes, and disinformation: A primer. RAND Corporation. https://doi.org/10.7249/pea1043-1
  • International Committee of the Red Cross. (2018, 3 de abril.). Ethics and autonomous weapon systems: An ethical basis for human control? https://www.icrc.org/en/document/ethics-and-autonomous-weaponsystems-ethical-basis-human-control
  • Johnson, J. (2020). Artificial intelligence, drone swarming and escalation risks in future warfare. rusi Journal, 165(2), 26-36. https://doi.org/10.1080/03071847.2020.1752026
  • Lassalle, J. M. (2022, 22 de octubre). Capitalismo digital de sustitución. La Vanguardia. https://www.lavanguardia.com/opinion/20221022/8576754/capitalismo-digital-sustitucion.html
  • Leonardo-Oviedo, G. (2004). La definición del concepto de percepción en psicología con base en la teoría Gestalt. Revista de Estudios Sociales, 18, 89-96.
  • Marina, J. A. (2022). El deseo interminable. Ariel.
  • Mir, R., & Rodríguez, K. (2020). If privacy dies in vr, it dies in real life. https://www.eff.org/deeplinks/2020/08/if-privacy-dies-vr-it-dies-real-life
  • Parasuraman, R., Sheridan, T. B., & Wickens, C. D. (2000). A model for types and levels of human interaction with automation. ieee Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans, 30(3), 286-297. https://doi.org/10.1109/3468.844354
  • Paris, B., & Donovan, J. (2019). Deepfakes and cheap fakes. Data & Society, 47. https://site.ieee.org/sagroups-7011/
  • Postman, N. (2005). Amusing ourselves to death. Penguin Books.
  • Pulido, G. (2021). Guerra multidominio y mosaico. Catarata. https://www.catarata.org/libro/guerramultidominio-y-mosaico_133474/
  • Scharre, B. P., & Fish, L. (2018, 7 de noviembre). Human performance enhancement. Center for a New American Security. https://www.cnas.org/publications/reports/human-performance-enhancement-1
  • Scharre, P. (2019). Killer apps: The real dangers of an ai arms race. Foreign Affairs, 98(3), 135-144. https://www.foreignaffairs.com/articles/2019-04-16/killer-apps
  • Schick, N. (2020, 28 de diciembre). Los “deepfake” no han roto la democracia en 2020; los “cheapfake”, sí. MIT Technology Review. https://www.technologyreview.es/s/13044/los-deepfake-no-han-rotola-democracia-en-2020-los-cheapfake-si
  • Shahriari, K., & Shahriari, M. (2017). ieee standard review. Ethically aligned design: A vision for prioritizing human wellbeing with artificial intelligence and autonomous systems. 2017 IEEE Canada International Humanitarian Technology Conference (ihtc), Toronto, on, Canada, 197-201. https://doi.org/10.1109/IHTC.2017.8058187
  • Simon, H. A. (1971). Designing organizations for an information-rich world. En M. Greenberger (Ed.), Computers, communications, and the public interest. John Hopkins University Press.
  • Stop Killer Robots. (2018). Campaign to stop killer robots. https://www.stopkillerrobots.org/
  • Templeton, E. M., Chang, L. J., Reynolds, E. A., LeBeaumont, M. D. C., & Wheatley, T. (2022). Fast response times signal social connection in conversation. Proceedings of the National Academy of Sciences of the United States of America, 119(4). https://doi.org/10.1073/pnas.2116915119
  • Torres Soriano, M. R. (2022, 28 de junio). Operaciones de influencia vs. desinformación: Diferencias y puntos de conexión. Instituto Español de Asuntos Estratégicos, (64). https://www.ieee.es/Galerias/fichero/docs_opinion/2022/DIEEEO64_2022_MANTOR_Operaciones.pdf
  • Tschopp, M. (2019). Good AI, bad AI - psychological aspects of a dual-use technology. https://www.academia.edu/40718200/Good_AI_Bad_AI_Psychological_Aspects_of_a_Dual_Use_Technology
  • Veliz, C. (2020). Privacy is power: Why and how you should take back control of your data. Transworld.
  • Von Clausewitz, K. (2002). De la guerra. https://www.psicosocial.net/historico/index.php?option=com_docman&view=download&alias=870-de-la-guerra&category_slug=psicologia-y-violencia-politica&Itemid=100225
  • Watts, C. (2019). Advanced persistent manipulators, part one: The threat to the social media industry - alliance for securing democracy. https://securingdemocracy.gmfus.org/advanced-persistent-manipulators-part-one-the-threat-to-the-social-media-industry/
  • Woolley, S., & Joseff, K. (2020). Demand for deceit: How the way we think drives disinformation. https://www.ned.org/wp-content/uploads/2020/01/Demand-for-Deceit.pdf
  • Zuboff, S. (2020). La era del capitalismo de la vigilancia: La lucha por un futuro humano frente a las nuevas fronteras del poder. Paidós.