Timnit Gebru facts for kids
Quick facts for kids
Timnit Gebru
|
|
---|---|
Gebru in 2018
|
|
Born | 1982/1983 (age 40–41) Addis Ababa, Ethiopia
|
Alma mater | Stanford University |
Known for |
|
Scientific career | |
Fields | Computer science |
Institutions | |
Doctoral advisor | Fei-Fei Li |
Timnit Gebru (Amharic and Tigrinya: ትምኒት ገብሩ; 1982/1983) is an Eritrean Ethiopian-born computer scientist who works in the fields of artificial intelligence (AI), algorithmic bias and data mining. She is an advocate for diversity in technology and co-founder of Black in AI, a community of Black researchers working in AI. She is the founder of the Distributed Artificial Intelligence Research Institute (DAIR).
In December 2020, public controversy erupted over the nature of Gebru's departure from Google, where she was technical co-lead of the Ethical Artificial Intelligence Team. Together with five co-authors, four of whom were also from Google, Gebru had authored a paper on the risks of large language models (LLMs) and submitted it for publication. Google management requested that Gebru either withdraw the paper or remove the names of all the authors employed by Google, claiming that the paper ignored recent research. Gebru requested an explanation and stated that if Google refused she would talk to her manager about "a last date". Google terminated her employment immediately, stating that they were accepting her resignation. Gebru maintained that she had not formally offered to resign, and only threatened to.
Gebru has been recognized widely for her expertise in the ethics of artificial intelligence. She was named one of the World's 50 Greatest Leaders by Fortune and one of Nature's ten people who shaped science in 2021, and in 2022, one of Time's most influential people.
Contents
Early life and education
Gebru was born and raised in Addis Ababa, Ethiopia. Her father, an electrical engineer with a Doctor of Philosophy (PhD), died when she was five years old, and she was raised by her mother, an economist. Both her parents are from Eritrea. When Gebru was 15, during the Eritrean–Ethiopian War, she fled Ethiopia after some of family was deported to Eritrea and compelled to fight in the war. She was initially denied a US visa and briefly lived in Ireland, but she eventually received political asylum in the US, an experience she said was "miserable". Gebru settled in Somerville, Massachusetts to attend high school, where she says she immediately started to experience racially-based discrimination, with some teachers refusing to allow her to take certain Advanced Placement courses, despite being a high-achiever.
After completing high school, an encounter with the police set Gebru on a course toward a focus on ethics in technology. A friend of hers, a Black woman, was assaulted in a bar, and Gebru called the police to report it. She says that instead of filing the assault report, her friend was arrested and remanded in a cell. Gebru called it a pivotal moment and a "blatant example of systemic racism".
In 2001, Gebru was accepted at Stanford University. There she earned her Bachelor of Science and Master of Science degrees in electrical engineering, and her PhD in computer vision in 2017. Gebru was advised during her PhD program by Fei-Fei Li.
During the 2008 United States presidential election, Gebru canvassed in support of Barack Obama.
Gebru presented her doctoral research at the 2017 LDV Capital Vision Summit competition, where computer vision scientists present their work to members of industry and venture capitalists. Gebru won the competition, starting a series of collaborations with other entrepreneurs and investors.
Both during her PhD program in 2016 and in 2018, Gebru returned to Ethiopia with Jelani Nelson's programming campaign AddisCoder.
While working on her PhD, Gebru authored a paper that was never published about her concern over the future of AI. She wrote of the dangers of the lack of diversity in the field, centered on her experiences with the police, and a ProPublica investigation into predictive policing, which revealed a projection of human biases in machine learning. .....
Career
2004–2013: Software development at Apple
Gebru joined Apple as an intern while at Stanford, working in their hardware division making circuitry for audio components, and was offered a full-time position the following year. Of her work as an audio engineer, her manager told Wired she was "fearless," and well-liked by her colleagues. During her tenure at Apple, Gebru became more interested in building software, namely computer vision that could detect human figures. She went on to develop signal processing algorithms for the first iPad. At the time, she said she did not consider the potential use for surveillance, saying "I just found it technically interesting."
Long after leaving the company, during the #AppleToo movement in the summer of 2021, which was led by Apple engineer Cher Scarlett, who consulted with Gebru, Gebru revealed she experienced "so many egregious things" and "always wondered how they manage[d] to get out of the spotlight." She said that accountability at Apple was long overdue, and warned they could not continue to fly under the radar for much longer. Gebru also criticized the way the media covers Apple and other tech giants, saying that the press helps shield such companies from public scrutiny.
2013–2017: Research at Stanford and Microsoft
In 2013, Gebru joined Fei-Fei Li's lab at Stanford. She used data mining of publicly available images. She was interested in the amount of money spent by governmental and non-governmental organisations trying to collect information about communities. To investigate alternatives, Gebru combined deep learning with Google Street View to estimate the demographics of United States neighbourhoods, showing that socioeconomic attributes such as voting patterns, income, race, and education can be inferred from observations of cars. If the number of pickup trucks outnumbers the number of sedans, the community is more likely to vote for the Republican party. They analysed over 15 million images from the 200 most populated U.S. cities. The work was extensively covered in the media, being picked up by BBC News, Newsweek, The Economist, and The New York Times.
In 2015, Gebru attended the field's top conference, Neural Information Processing Systems (NIPS), in Montreal, Canada. Out of 3,700 attendees, she noted she was one of only a few Black researchers. When she attended again the following year, she kept a tally and noted that there were only five Black men and that she was the only Black woman out of 8,500 delegates. Together with her colleague Rediet Abebe, Gebru founded Black in AI, a community of Black researchers working in artificial intelligence.
In the summer of 2017, Gebru joined Microsoft as a postdoctoral researcher in the Fairness, Accountability, Transparency, and Ethics in AI (FATE) lab. In 2017, Gebru spoke at the Fairness and Transparency conference, where MIT Technology Review interviewed her about biases that exist in AI systems and how adding diversity in AI teams can fix that issue. In her interview with Jackie Snow, Snow asked Gebru, "How does the lack of diversity distort artificial intelligence and specifically computer vision?" and Gebru pointed out that there are biases that exist in the software developers. While at Microsoft, Gebru co-authored a research paper called Gender Shades, which became the namesake of a project of a broader Massachusetts Institute of Technology project led by co-author Joy Buolamwini. The pair investigated facial recognition software, finding that in one particular implementation Black women were 35% less likely to be recognized than White men.
2018–2020: Artificial intelligence ethics at Google
Gebru joined Google in 2018, where she co-led a team on the ethics of artificial intelligence with Margaret Mitchell. She studied the implications of artificial intelligence, looking to improve the ability of technology to do social good.
In 2019, Gebru and other artificial intelligence researchers "signed a letter calling on Amazon to stop selling its facial-recognition technology to law enforcement agencies because it is biased against women and people of color," citing a study that was conducted by MIT researchers showing that Amazon's facial recognition system had more trouble identifying darker-skinned females than any other technology company's facial recognition software. In a New York Times interview, Gebru has further expressed that she believes facial recognition is too dangerous to be used for law enforcement and security purposes at present.
Exit from Google
In 2020 Gebru and five co-authors wrote a paper titled "On the Dangers of Stochastic Parrots: Can Language Models Be Too Big? 🦜". The paper examined risks of very large language models, including their environmental footprint, financial costs, the inscrutability of large models, the potential for LLMs to display prejudice against certain groups, the inability of LLMs to understand the language they process, and the use of LLMs to spread disinformation.
In December 2020, her employment with Google ended after Google management asked her to either withdraw the paper before publication, or remove the names of all the Google employees from the paper. Of the six authors, only Emily M. Bender was not at the time employed at Google. In a six-page mail sent to an internal collaboration list, Gebru describes how she was summoned to a meeting at short notice where she was asked to withdraw the paper and she requested to know the names and reasons of everyone who made that decision, along with advice for how to revise it to Google's liking. She said she would work with Google on an employment end date after an appropriate amount of time if not provided with that information. Google did not meet her request and terminated her employment immediately, declaring that they accepted her resignation. Jeff Dean, Google's head of AI research, replied with an email saying that they made the decision because the paper ignored too much relevant recent research on ways to mitigate some of the problems described in it, about environmental impact, and bias of these models.
Dean went on to publish his internal email regarding Gebru's departure and his thoughts on the matter, defending Google's research paper process to "tackle ambitious problems, but to do so responsibly." Gebru and others blame this initial publication, and Dean's subsequent silence on the matter, to have both catalyzed and enabled harassment that followed close behind his response. Gebru was serially harassed by a number of sock puppet accounts and internet trolls on Twitter, making racist and obscene comments. Gebru and her supporters alleged some of the harassment to be from machine learning researcher Pedro Domingos and businessman Michael Lissack, who had said that her work was "advocacy disguised as science." Lissack also allegedly harassed Mitchell and Bender, along with other colleagues on Gebru's former team. Twitter temporarily suspended Lissack's account access on 1 February 2021.
Gebru has repeatedly maintained that she was fired, and close to 2,700 Google employees and more than 4,300 academics and civil society supporters signed a letter condemning Gebru's alleged firing. Nine members of Congress sent a letter to Google asking it to clarify the circumstances around Timnit Gebru's exit. Gebru's former team demanded that Vice President Megan Kacholia be removed from the team's management chain. Kacholia had allegedly fired Gebru without notifying Gebru's direct manager Samy Bengio first, and Gebru's former team demanded Kacholia and Dean apologize for how Gebru was treated. Mitchell took to Twitter to criticize Google's treatment of employees working to eliminate bias and toxicity in AI, including its alleged dismissal of Gebru. Mitchell was later terminated for alleged exfiltration of confidential business-sensitive documents and private data of other employees.
Following the negative publicity over the circumstances of her exit, Sundar Pichai, CEO of Alphabet, Google's parent company, publicly apologized on Twitter without clarifying whether Gebru was terminated or resigned and initiated a months-long investigation into the incident. Upon conclusion of the review, Dean announced Google would be changing its "approach for handling how certain employees leave the company," but still did not clarify whether or not Gebru's leaving Google was voluntary. Additionally, Dean said there would be changes to how research papers with "sensitive" topics would be reviewed, and diversity, equity, and inclusion goals would be reported to Alphabet's board of directors quarterly. Gebru wrote on Twitter that she "expected nothing more" from Google and pointed out that the changes were due to the requests she was allegedly terminated for but that no one was held accountable for it. In the aftermath, two Google employees resigned from their positions at the company.
Following her departure, Google held a forum to discuss experiences with racism at the company, and employees reported to NBC News that half of it was spent discrediting Gebru, which they took as the company making an example of her for speaking out. The town hall was followed up with a group psychotherapy session for Google's Black employees with a licensed therapist, which the employees said was dismissive over the harm they felt Gebru's alleged termination had caused.
Since 2021: Independent research
In November 2021, the Nathan Cummings Foundation, partnered with Open MIC and endorsed by Color of Change, filed a shareholder proposal calling for a "racial equity audit" to analyze Alphabet's "adverse impact" on "Black, Indigenous and People of Color (BIPOC) communities." The proposal also requests investigation into whether or not Google retaliated against minority employees who raised concerns of discrimination, citing Gebru's firing, her previous urge for Google to hire more BIPOC, and her research into racially based biases in Google's technology. The proposal followed a less formal request from a group of Senate Democratic Caucus members led by Cory Booker from earlier that year, also citing Gebru's separation from the company and her work.
In December 2021, Reuters reported that Google was under investigation by California Department of Fair Employment and Housing (DFEH) for its treatment of Black women, after numerous formal complaints of discrimination and harassment by current and former workers. The probe comes after Gebru, and other BIPOC employees, reported that when they brought up their experiences with racism and sexism to Human Resources, they were advised to take medical leave and therapy through the company's Employee Assistance Program (EAP). Gebru, and others, believe that her alleged dismissal was retaliatory and evidence that Google is institutionally racist. Google said that it "continue[s] to focus on this important work and thoroughly investigate[s] any concerns, to make sure [Google] is representative and equitable."
In June 2021, Gebru announced that she was raising money to "launch an independent research institute modeled on her work on Google's Ethical AI team and her experience in Black in AI."
On 2 December 2021 she launched the Distributed Artificial Intelligence Research Institute (DAIR), which is expected to document the effect of artificial intelligence on marginalized groups, with a focus on Africa and African immigrants in the United States. One of the organization's initial projects plans to analyze satellite imagery of townships in South Africa with AI to better understand legacies of apartheid.
Gebru and Émile P. Torres have used the acronym TESCREAL to criticize what they see as a group of overlapping futurist philosophies: transhumanism, extropianism, singularitarianism, cosmism, rationalism, effective altruism, and longtermism. Gebru considers these to be a right-leaning influence in Big Tech and compares proponents to "the eugenicists of the 20th century" in their production of harmful projects they portray as "benefiting humanity".
Gebru has criticized research into Artificial General Intelligence (AGI) as being rooted in eugenics. Gebru states that focus should be shifted away from AGI and that trying to build AGI is an inherently unsafe practice.
Awards and recognition
Gebru, Buolamwini, and Inioluwa Deborah Raji won VentureBeat's 2019 AI Innovations Award in the category AI for Good for their research highlighting the significant problem of algorithmic bias in facial recognition. Gebru was named one of the world's 50 greatest leaders by Fortune in 2021. Gebru was included in a list of ten scientists who had had important roles in scientific developments in 2021 compiled by the scientific journal Nature.
Gebru was named one of Time's most influential people of 2022.
In 2023, Gebru was named by Carnegie Corporation of New York as an honoree of the Great Immigrants Awards.
In November 2023, she was named to the BBC's 100 Women list.
See also
In Spanish: Timnit Gebru para niños
- Coded Bias
- Claire Stapleton
- Meredith Whittaker
- Sophie Zhang