Technology is rapidly changing the world around us. It offers new ways for human rights defenders to use new tools to their advantage. Machine learning is one of them.
Machine learning is a powerful tool that offers tremendous opportunities in the field of human rights. For instance, it can help us detect patterns of corruption to support advocacy and predict poverty to drive policy change. Additionally, it can analyze evidence of human rights violations to support transitional justice efforts. However, despite these promising opportunities, the same technology also raises significant human rights concerns. In particular, algorithmic biases have the potential to drastically alter individuals’ lives, as well as reinforce and even accelerate existing social and economic inequalities. Furthermore, flawed facial recognition systems, the misclassification of videos documenting war crimes as terrorist propaganda, and the creation of racist chatbots illustrate the risks posed by these technologies.
Given these challenges, our goal as human rights defenders is to distinguish between beneficial machine learning systems and harmful automated decision-making processes. By doing so, we can minimize the risks and maximize the positive impact of new technologies on human rights work. A few good practices to help achieve this include ensuring fair and transparent machine learning algorithms and fostering close collaboration with experts from various fields to promote open and informed conversations.
Thank you to our featured resource practitioners who led the conversation:
- Enrique Piracés, Carnegie Mellon University
- Natalie Widmann, HURIDOCS
- Micaela Mantegna, Center for Technology and Society, San Andreas University
- Nani Jansen Reventlow, Digital Freedom Fund
- Bill Doran, Ushahidi
- Santiago Borrajo, CELS Centro de Estudios Legales y Sociales
- Adam Harvey, VFRAME
- Vivian Ng, University of Essex
Where to start: Machine learning 101 for human rights defenders
Machine learning (ML) is a subfield of artificial intelligence, with a goal of enabling computers to learn on their own. Through the computer’s algorithm, the computer can identify patterns, build models that explain the world, and make predictions without having pre-programmed rules and models governing the predictions. Arthur Samuel described machine learning as giving “computers the ability to learn without being explicitly programmed”. Before ML, patterns for classifying data would have to be manually defined, which was a long-winded process. ML can lead to a drastic reduction in the labor needed, that can be redirected towards other human rights focused endeavors. There are several tools and resources that provide information about ML. Distill focuses on clarity and transparency when working within the field of ML. Google has provided several learning tools and games that people can undertake to understand the concepts and methods within ML.
Further, there are online tools providing an introduction to ML in more than 10 languages, such as R2D3 that can be very helpful for non-English speaking practitioners. With the development of resources, it is suggested that practitioners will have a better chance to enter into the ML space as there are relatively low-cost resources available, such as services, products, libraries and hardware. Such resources include Tensorflow, DeepLens, AWS Machine Learning, and Google’s Cloud AI, which make ML increasingly accessible.
Promise and perils: Machine learning applied to human rights practice
ML is still in the developing phases, but has already been implemented into some human rights work. ML programs can aid detection of human rights abuses, improve existing systems and prevent dangerous situations. When working in human rights, practitioners are often faced with reports, evidence and other data that needs to be categorized. ML tools decrease the amount of time needed to accomplish this. For example by implementing a tool that classifies sentences and is adaptable to the specific research questions human rights defenders are interested in. Other tools currently developed are those using video analysis that can detect objects, sound, speech, text and event types. This allows users to run semantic queries within video collections to discern what is happening. They can also document human rights violations, predict judicial hearings, and be used as open source computer vision tools in large video datasets. Currently, video analysis is being used in Syria in an attempt at providing verified videos that can be used as evidence of war crimes.
Challenges and Ethical Concerns in the Use of Machine Learning for Human Rights
However, there are still challenges to ML that can provide obstacles and limitations. The Danish Institute has gathered a large database of information based on Universal Periodic Reviews (UPR). Currently, they are exploring the usage of ML in making predictions, but the availability of the database itself was only made possible by manually categorizing reports and recommendations that have been made searchable through their website. In addition, the growth of ML also leads to concerns that it will be used in ways contrary to human rights standards. A common concern is the misuse of facial recognition, which could potentially put human rights practitioners and vulnerable populations at risk. Other ML projects have been developed to answer this risk, like Harvard Law School’s EqualAIs, which slightly alters an image in a way that is undetectable to the human eye but prevents it from being identified through other ML technology.
Impact of machine learning on society
ML technology provides exciting new opportunities for human rights defenders as it can lead to a significant decrease in the amount of time practitioners use to categorize and classify data. However, as mentioned above, ML also brings about a set of new obstacles that must be addressed. The Toronto Declaration of 2018 serves as an example of public policy attempting to protect against discrimination and human rights abuses in ML technology. Not only are public policy officials working on mitigating the potential harms of ML, but companies who are leaders in ML are taking their own steps to ensure that it is used ethically. One example of this is Google’s principles on AI. Some have raised concerns about ML systems used by governments, like the Netherland’s System Risk Indication program (SyRI), which creates risk profiles of its citizens to detect fraud. Many have argued that systems such as these reinforce their own findings and have a disproportionate impact on vulnerable members of society.
Further, some have highlighted the potential ethical question of whether or not ML should be used to assist or replace judicial decision making, pointing to how legal systems are created by humans to ensure social order. It is suggested that perhaps ML can be used to make processes more efficient. That for decisions revolving around critical aspects of personal and social life, humans should make the last call. To mitigate these practical and philosophical issues, special attention has been paid to ensuring that fairness, transparency and diversity are present in ML programs. Some solutions include applying the current human rights framework to the use of ML technology. Others point to the EU’s current General Data Protection Regulation (GDPR) as another source of guidelines for ML.
Struggles of machine learning practitioners in the human rights field
Despite the many useful applications of ML in the human rights field. There is a gap in the understanding of ML and its potential by human rights defenders, while ML practitioners struggle to understand human rights practice. To bridge this gap suggestions include open and diverse dialogue between the two groups, as well as more long term projects where the different actors work closely together. In addition, the Fairness, Accountability and Transparency in Machine Learning, provides resources that can resonate with both human rights defenders and ML practitioners. It plays a fundamental role within the human rights framework. It connects to conversations about design and development processes of ML systems.
Ushahidi is a non-profit tech company focused on helping marginalized people. Their decision to start utilizing ML has brought them to the intersection of human rights and ML professionals. However, their work has also presented challenges, particularly in data rarity and sparsity. The data is often very specific and limited, which prevents it from transferring well to other instances. This difficulty makes it challenging to train ML algorithms in new domains.
Open Discussion
While ML is still new in the human rights field, discussions on the strengths and challenges related to its use in the field are becoming more and more common. The
Huridocs Collaboratory facilitates discussions between technologists and human rights defenders on the relationship between ML and human rights. Other panels, like “Beyond Explainability: Regulating Machine Learning In Practice” at the Strata NY 2018 also touch on similar issues of ML and human rights.
Tactic Examples
Resources
Resources
More Resources:
- FaceForensics: A large-scale video dataset used to detect face manipulations.
- Google Cloud AI: A modern learning service from Google Cloud.
- “How to Upgrade Judges with Machine Learning”: Published by MIT Technology Review portraying a positive view on the use of software that helps judges decide whether to jail a defendant awaiting trial.
- “How will the GDPR impact machine learning?”: Article focused on how to maintain GDRP-compliant data science.
- HURIDOCS: Provides technological and methodological support to the documentation of human rights violations.
- HURIDOCS Collaboratory: A forum for human rights practitioners and technologists to share information and experience.
- Machine Learning for Humans: An explanation of machine learning using math, code and real-world examples.
- Machine learning to uncover mass graves in Mexico: Published by Quartz. Machine learning is being used to uncover mass graves in Mexico.
- Mozilla Fellowship: A fellowship for researchers in open science and data sharing.
- NLTK: Natural Language Toolkit for building Python programs to work with human language data.
- R2D3: A visual introduction to machine learning.
- Scikit-learn: A tool for data mining and data analysis.
- Security in-a-box: A digital security tool developed for human rights defenders.
- Submission to the House of Lords on AI: Analysis of the challenges and opportunities presented by big data and technologies from a human rights perspective.
Final Resources:
- System Risk Indication (SyRI): A tool by the government in the Netherlands that allows them to use data collected through public institutions to combat abuse of social security provisions and prevent tax fraud.
- TensorFlow: An open-source machine learning library for research and production.
- “The Challenge from AI: is ”human” always better?”: Blogpost by Sherif Elsayed-Ali discussing if human decisions are always better.
- The Danish Institute for Human Rights: Project attempting to develop and train an algorithm for automatic classification of recommendations from UN human rights monitoring bodies.
- The Human Rights, Big Data and Technology Project: Analyze the challenges and opportunities presented by the use of big data, AI and associated technologies from a human rights perspective.
- Toronto Declaration of 2018: Protects the rights equality and non-discrimination in machine learning systems.
- Tracking Twitter Abuse against Women: Amnesty International used twitter to track abuse against women.
- UNHR Innovation Service: Projects that are sponsored by the UNHR to inform refugee systems.
- United Nations Global Pulse: Collaborative research, prototypes and experiments that support humanitarian action and development.
- UPR Database: Database of Recommendations from the UPR (Universal Periodic Review).
- Ushahidi: Article describing Ushahidi’s attempt at implementing ML.
- VFRAME: A collection of open-source computer vision software tools designed specifically for human rights investigations.
- Weka 3: A machine learning software in Java.