Academic Handbook Course Descriptors and Programme Specifications
LDSCI7230 Advanced Topics in Responsible Artificial Intelligence Course Descriptor
Course code | cd | Discipline | Computer Science |
UK Credit | 30 | US Credit | NA |
FHEQ level | 7 | Date approved | June 2023 |
Core attributes | N/A | ||
Pre-requisites | None | ||
Co-requisites | None |
Course Summary
This course explores the social and ethical dimensions of responsible Artificial intelligence. Artificial intelligence systems and their development can pose risks to individuals and societies. Responsible AI encompasses the design, development, implementation, and regulation of AI systems in order to ensure that they are ethical, safe, and trustworthy. Responsible AI approaches are naturally interdisciplinary; they include investigations into ethics, societal impact, regulation, design, and responsibility for AI systems, with a particular focus on AI that will have significant impact on our lives. There is a particular focus on sustainability in terms of societal and environmental impact.
Students will cover a range of advanced topics in Responsible AI with particular attention to recent, current, and near-future issues in AI. Students will gain an understanding of how philosophy can interact with other disciplines in discussions of artificial intelligence, and how these approaches can inform the development and management of new technologies. Each topic is examined through a philosophical lens, and at least one other disciplinary approach that has a particular relevance to that topic. Additional disciplinary approaches may include (but are not limited to) business studies, digital humanities, law, politics, sociology, psychology, or computer science. Topics will be research-led, informed by the expertise of Northeastern University London.
Course Aims
The aims of the course are to:
- To provide a foundation of philosophical concepts and values relevant to current and emerging issues in responsible AI
- To develop the critical skills needed to apply concepts and approaches from a variety of disciplines to ethical and societal questions in AI.
- To develop an understanding of how philosophical thinking can contribute to the beneficial development of artificial intelligence.
- To enable students to form, elaborate, and defend their own views in this field, to appreciate the relevance and value of diverse viewpoints, and contribute constructively to work aimed at addressing concrete problems.
- To develop students’ capacity to engage constructively with peers, process arguments in a charitable manner, and engage in productive class debate.
Learning Outcomes
On successful completion of the course, students will be able to:
Knowledge and Understanding
K1d | Demonstrate sound knowledge and sophisticated understanding of key questions and debates in the field of Responsible AI from an interdisciplinary perspective. |
K3d | Evaluate societal dimensions of AI and data practices and demonstrate a comprehensive understanding and critical awareness of key philosophical issues (ethical, cultural, privacy, or policy) surrounding data use, data processing, and AI. |
K4d | Critically review and analyse key issues in the development and application of artificial intelligence in society, identify issues and limitations, and propose directions for the responsible development, management, and future innovation of AI systems. |
Subject Specific Skills
S1d | Critically assess the design, development, and implementation of artificial intelligence systems in society, and propose ways to improve this. |
S3d | Produce original ideas for the future development, implementation, and management of responsible AI systems. |
S4d | Understand the importance of embedding ethical considerations into the development of data applications and AI systems. |
Transferable and Professional Skills
T2d | Consistently display an excellent level of technical proficiency in written English and command of scholarly terminology, so as to be able to deal with complex issues in a sophisticated and systematic way. |
T4d | Communicate effectively with rigorous arguments appropriately for both technical and non-technical audiences in relation to the development and application of responsible AI systems and the contemporary philosophical questions that surround it, through oral presentations and written reports. |
T3d | Demonstrate initiative in working independently, effectively, and to deadlines. |
Teaching and Learning
This course has a dedicated Virtual Learning Environment (VLE) page with a syllabus and range of additional resources (e.g. readings, question prompts, tasks, assignment briefs, discussion boards) to orientate and engage you in your studies.
The scheduled teaching and learning activities for this course are:
Lectures/Labs: Contact hours are typically a mix of weekly lectures and lab sessions:
- Version 1: All sessions in the same sized group, or
- Version 2: most of the sessions in larger groups; some of the sessions in smaller groups
Faculty hold regular ‘office hours’, which are opportunities for students to drop in or sign up to explore ideas, raise questions, or seek targeted guidance or feedback, individually or in small groups.
Students are to attend and participate in all the scheduled teaching and learning activities for this course and to manage their directed learning and independent study.
Indicative total learning hours for this course: 300
Assessment
Both formative and summative assessment are used as part of this course, with formative opportunities typically embedded within interactive teaching activities delivered via the VLE.
Summative
AE: | Assessment Activity | Weighting (%) | Duration | Length |
1 | Presentation | 20% | 10 Minutes | N/A |
2 | Written Assignment | 80% | N/A | 4,000 Words |
Further information about the assessments can be found in the Course Syllabus.
Feedback
Students will receive formative and summative feedback in a variety of ways, written (e.g. marked up on assignments or via the VLE) or oral (e.g. as part of interactive teaching sessions or in office hours).
Indicative Reading
Note: Comprehensive and current reading lists for courses are produced annually in the Course Syllabus or other documentation provided to students; the indicative reading list provided below is used as part of the approval/modification process only.
Books
- Voeneky, S., Kellmeyer, P., Mueller, O., & Burgard, W. (Eds.) (2022). The Cambridge Handbook of Responsible Artificial Intelligence: Interdisciplinary Perspectives. Cambridge University Press.
- Dubber, M.D., Pasquale, F. and Das, S. (Eds.) (2020). The Oxford handbook of ethics of AI. Oxford University Press.
- Dignum, V. (2019) Responsible Artificial Intelligence: How to develop and use AI in a responsible way. Springer.
- von Braun, S. Archer, M., Reichberg, G. M., & Sánchez Sorondo, M. (2021). Robotics, AI, and Humanity: Science, Ethics, and Policy. Springer Nature.
- Noble, S. U. (2018). Algorithms of oppression. In Algorithms of oppression. New York University Press
- Sandler. (2014). Ethics and Emerging Technologies. Palgrave Macmillan UK.
Journals
- Autor, D. H. (2015), ‘Why Are There Still So Many Jobs? The History and Future of Workplace Automation’, The Journal of Economic Perspectives, 29, pp. 3–30
- Bender, E. M., Gebru, T., McMillan-Major, A., & Shmitchell, S. (2021, March). On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency (pp. 610-623).
- Spencer, D. A. (2022) ‘Automation and Well-Being: Bridging the Gap between Economics and Business Ethics.’ Journal of Business Ethics.
- Hristov, K. (2016). ‘Artificial intelligence and the copyright dilemma’. Idea, 57, p.431.
- van Wynsberghe, A. (2021) ‘Sustainable AI: AI for sustainability and the sustainability of AI’. AI Ethics, 1, 213–218 (2021). https://doi.org/10.1007/s43681-021-00043-6
- Hongladarom, S. and Bandasak, J. (2023) ‘Non-western AI ethics guidelines: implications for intercultural ethics of technology’. AI & Society. pp.1-14.
- La Fors, K. (2022) ‘Toward children-centric AI: a case for a growth model in children-AI interactions’. AI & Society. pp.1-13.
Reports
- Taylor, Steve, Pickering, Brian, Boniface, Michael, Anderson, Michael, Danks, David,
- Følstad, Asbjørn, Leese, Matthias, Müller, Vincent, Sorell, Tom, Winfield, Alan, & Woollard,
- Fiona. (2018). Responsible AI – Key Themes, Concerns & Recommendations for
- European Research and Innovation (1.0). Zenodo. https://doi.org/10.5281/zenodo.1303253
Electronic Resources
University of Oxford. Ethics in AI [Podcast] Available at: https://podcasts.ox.ac.uk/series/ethics-ai
Ethical OS Toolkit https://ethicalos.org/
Indicative Topics
Note: Comprehensive and current topics for courses are produced annually in the Course Syllabus or other documentation provided to students. The indicative topics provided below are used as a general guide and part of the approval/modification process only.
- The ethics of automation
- Authorship, AI, and copyright
- AI in the classroom
- Human-computer interaction
- Human-centred AI
- Sustainable AI
- AI in healthcare
- Global governance for AI
Version History
Title: LDSCI7230 Advanced Topics in Responsible AI
Approved by: Academic Board Location: Academic Handbook/Programme specifications and Handbooks/ Postgraduate Programme Specifications/MSc Computer Science Programme Specification/Course Descriptors |
|||||
Version number | Date approved | Date published | Owner | Proposed next review date | Modification (As per AQF4) & category number |
1.0 | June 2023 | June 2023 | Alexandros Koliousis | April 2028 |