Pope Francis urges followers to pray that AI and robots ‘always serve mankind’

Pope Francis has asked believers around the world to pray that robots and artificial intelligence “always serve mankind.”

The message is one of the pope’s monthly prayer intentions — regular missives shared on YouTube that are intended to help Catholics “deepen their daily prayer” by focusing on particular topics or events. In August, the pope urged prayer for “the maritime world”; in April, the topic was “freedom for addiction.” Now, in November, it’s AI and robots.

Although the message sounds similar to warnings issued by tech notables like Elon Musk (the Tesla CEO famously compared work on artificial intelligence to “summoning the demon”), the pope’s focus is more prosaic. He doesn’t seem to be worrying about the sort of exotic doomsday scenario where a superintelligent AI turns the world into paperclips, but more about how the tech could exacerbate existing inequalities here and now.

(We should note also that the call to prayer came out earlier this month, but we only saw it recently via the Import AI newsletter because of the… events that have taken up so much of everyone’s time, energy, and general mental acuity in recent weeks.)

In his message, the pope said AI was “at the heart of the epochal change we are experiencing” and that robotics had the power to change the world for the better. But this would only be the case if these forces are harnessed correctly, he said. “Indeed, if technological progress increases inequalities, it is not true progress. Future advances should be orientated towards respecting the dignity of the person.”

Perhaps surprisingly, this isn’t new territory for the pope. Earlier this year, the Vatican, along with Microsoft and IBM, endorsed the “Rome Call for AI Ethics” — a policy document containing six general principles that guide the deployment of artificial intelligence. These include transparency, inclusion, impartiality, and reliability, all sensible attributes when it comes to deploying algorithms.

Although the pope didn’t touch on any particular examples in his video, it’s easy to think of ways that AI is entrenching or increasing divisions in society. Examples include biased facial recognition systems that lead to false arrests and algorithmically allotted exam results that replicate existing inequalities between students. In other words: regardless of whether you think prayer is the appropriate course of action, the pope certainly has a point.

Repost: Original Source and Author Link


Fortnite Fortography contest urges players to get close to dinosaurs

Epic has revealed the next Fortnite Fortography challenge, and this time you’ll have to get up close and personal with dinosaurs. The raptors have just hatched from their eggs and, well, they don’t have the friendliest disposition. That’ll make this one of the most fun photography challenges.

The Fortnite Fortography contest is a bragging-rights challenge that seeks in-game screenshots featuring a certain theme. Epic announces the theme every couple of weeks and they usually revolve around a new skin or item added to the game. This week’s challenge is no exception.

With the latest update comes a new order: Get screenshots that involve your best moments with the nearly added raptors. This can include, for example, defeating them in a battle or simply running away in a phenomenal display of terror — or both, perhaps, if you’re lucky.

As with previous Fortography challenges, you’ll need to share the screenshots on social media using the #Fortography hashtag to bring them to Epic’s attention. The company will review the images submitted on social media and will highlight the ones it likes the most on its Fortnite blog.

Getting screenshots while actively playing the game is very difficult, of course, and there’s where playback footage comes in. You can pull up your recorded matches after the game, reposition the camera, and get the perfect screenshot after the action has taken place.

Repost: Original Source and Author Link


Google employee group urges Congress to strengthen whistleblower protections for AI researchers

Join Transform 2021 for the most important themes in enterprise AI & Data. Learn more.

Google’s decision to fire its AI ethics leaders is a matter of “urgent public concern” that merits strengthening laws to protect AI researchers and tech workers who want to act as whistleblowers. That’s according to a letter published by Google employees today in support of the Ethical AI team at Google and former co-leads Margaret Mitchell and Timnit Gebru, who Google fired two weeks ago and in December 2020, respectively.

Firing Gebru, one of the best known Black female AI researchers in the world and one of few Black women at Google, drew public opposition from thousands of Google employees. It also led critics to claim the incident may have shattered Google’s Black talent pipeline and signaled the collapse of AI ethics research in corporate environments.

“We must stand up together now, or the precedent we set for the field — for the integrity of our own research and for our ability to check the power of big tech — bodes a grim future for us all,” reads the letter published by the group Google Walkout for Change. “Researchers and other tech workers need protections which allow them to call out harmful technology when they see it, and whistleblower protection can be a powerful tool for guarding against the worst abuses of the private entities which create these technologies.”

Google Walkout for Change was created in 2018 by Google employees organizing to force change at Google. According to organizers, the global walkout that year involved 20,000 Googlers in 50 cities around the world.

The letter also urges academic conferences to refuse to review papers subjected to editing by corporate lawyers and to begin declining sponsorship from businesses that retaliate against ethics researchers. “Too many institutions of higher learning are inextricably tied to Google funding (along with other Big Tech companies), with many faculty having joint appointments with Google,” the letter reads.

The letter addressed to state and national lawmakers cites a VentureBeat article published two weeks after Google fired Gebru about potential policy outcomes that could include changes to whistleblower protection laws and unionization. That analysis — which drew on conversations with ethics, legal, and policy experts — cites UC Berkeley Center for Law and Technology co-director Sonia Katyal, who analyzed whistleblower protection laws in 2019 in the context of AI. In an interview with VentureBeat late last year, Katyal called them “totally insufficient.”

“What we should be concerned about is a world where all of the most talented researchers like [Gebru] get hired at these places and then effectively muzzled from speaking. And when that happens, whistleblower protections become essential,” Katyal told VentureBeat.

VentureBeat spoke with two sources familiar with Google AI ethics and policy matters who said they want to see stronger whistleblower protection for AI researchers. One person familiar with the matter said that at Google and other tech companies, people sometimes know something is broken but won’t fix it because they either don’t want to or don’t know how to.

“They’re stuck in this weird place between making money and making the world more equitable, and sometimes that inherent tension is very difficult to resolve,” the person, who spoke on condition of anonymity, told VentureBeat. “But I believe that they should resolve it because if you want to be a company that touches billions of people, then you should be responsible and held accountable for how you touch those billions of people.”

After Gebru was fired, that source described a sense among people from underrepresented groups at Google that if they push the envelope too far now they might be perceived as hostile and people will start filing complaints to push them out. She said this creates a feeling of “genuine unsafety” in the workplace and a “deep sense of fear.”

She also told VentureBeat that when we’re looking at technology with the power to shape human lives, we need to have people throughout the design process with the authority to overturn potentially harmful decisions and make sure models learn from mistakes.

“Without that, we run the risk of … allowing algorithms that we don’t understand to literally shape our ability to be human, and that inherently isn’t fair,” she said.

The letter also criticizes Google leadership for “harassing and intimidating” not only Gebru and Mitchell, but other Ethical AI team members as well. Ethical AI team members were reportedly told to remove their names from a paper under review at the time Gebru was fired. The final copy of that paper, titled “On the Dangers of Stochastic Parrots: Can Language Models Be Too Big?” was published this week at AI ethics conference Fairness, Accountability, and Transparency (FAccT) and lists no authors from Google. But a copy of the paper VentureBeat obtained lists Mitchell as a coauthor of the paper, as well as three other members of the Ethical AI team, each with an extensive background in examining bias in language models or human speech. Google AI chief Jeff Dean questioned the veracity of the research represented in that paper in an email to Google Research. Last week, FAccT organizers told VentureBeat the organization has suspended sponsorship from Google.

The letter published today calls on academics and policymakers to take action and follows changes to company diversity policy and reorganization of 10 teams within Google Research. These include Ethical AI, now under Google VP Marian Croak, who will report directly to AI chief Jeff Dean. As part of the change, Google will double staff devoted to employee retention and enact policy to engage HR specialists when certain employee exits are deemed sensitive. While Google CEO Sundar Pichai mentioned better de-escalation strategies as part of the solution in a companywide memo, in an interview with VentureBeat, Gebru called his memo “dehumanizing” and an attempt to characterize her as an angry Black woman.

A Google spokesperson told VentureBeat in an email following the reorganization last month that diversity policy changes were undertaken based on the needs of the organization, not in response to any particular team at Google Research.

In the past year or so, Google’s Ethical AI team has explored a range of topics, including the need for a culture change in machine learning and an internal algorithm auditing framework, algorithmic fairness issues specific to India, the application of critical race theory and sociology, and the perils of scale.

The past weeks and months have seen a rash of reporting about the poor experiences of Black people and women at Google, as well as reporting that raises concerns about corporate influence over AI ethics research. Reuters reported in December 2020 that AI researchers at Google were told to strike a positive tone when referring to “sensitive” topics. Last week, Reuters reported that Google will reform its approach to research review and additional instances of interference in AI research. According to an email obtained by Reuters, the coauthor of another paper about large language models referred to edits made by Google’s legal department as “deeply insidious.”

In recent days, the Washington Post has detailed how Google treats candidates from historically Black colleges and universities in a separate and unequal fashion, and NBC News reported that Google employees who experienced racism or sexism were told by HR to “assume good intent” and encouraged to take mental health leave instead of addressing the underlying issues.

Instances of gender discrimination and toxic work environments for women and people of color have been reported at other major tech companies, including Amazon, Dropbox, Facebook, Microsoft, and Pinterest. Last month, VentureBeat reported that dozens of current and former Dropbox employees, particularly women of color, reported witnessing or experiencing gender discrimination at their company. Former Pinterest employee Ifeoma Ozoma, who previously spoke with VentureBeat about whistleblower protections, helped draft the proposed Silenced No More Act in California last month. If passed, that law will allow employees to report discrimination even if they have signed a non-disclosure agreement.

The letter published by Google employees today follows other correspondence sent to Google company leadership since Gebru was fired in December 2020. Thousands of Google employees signed a Google Walkout letter protesting the way Gebru was treated and “unprecedented research censorship.” That letter also called for a public inquiry into Gebru’s termination for the sake of Google users and employees. Members of Congress with records of proposing regulation like the Algorithmic Accountability Act, including Rep. Yvette Clark (D-NY) and Senator Cory Booker (D-NJ), also sent Google CEO Sundar Pichai an email questioning the way Gebru was fired, Google’s research integrity, and steps the company takes to mitigate bias in large language models.

About a week after Gebru was fired, members of the Ethical AI team sent their own letter to company leadership. According to a copy obtained by VentureBeat, Ethical AI team members demanded Gebru be reinstated and Samy Bengio remain the direct report manager for the Ethical AI team. They also state that reorganization is sometimes used for “shunting workers who’ve engaged in advocacy and organizing into new roles and managerial relationships.” The letter described Gebru’s termination as having a demoralizing effect on the Ethical AI team and outlined a number of steps needed to re-establish trust. That letter cosigns letters of support for Gebru from Google’s Black Researchers group and the DEI Working Group. A Google spokesperson told VentureBeat an investigation was carried out by outside counsel but declined to share details. The Ethical AI letter also demands Google maintain and strengthen their team, guarantee the integrity of independent research, and clarify its sensitive review process by the end of Q1 2021. And it calls for a public statement that guarantees research integrity at Google, including in areas tied to the company’s business interests, such as large language models or datasets like JFT-300, a dataset with over a billion labeled images.

A Google spokesperson said Croak will oversee the work of about 100 AI researchers going forward. A source familiar with the matter told VentureBeat a reorganization that brings Google’s numerous AI fairness efforts under a single leader makes sense and had been discussed before Gebru was fired. The question, they said, is whether Google will fund fairness testing and analysis.

“Knowing what these communities need consistently becomes hard when these populations aren’t necessarily going to make the company a bunch of money,” a person familiar with the matter told VentureBeat. “So yeah, you can put us all under the same team, but where’s the money at? Are you going to give a bunch of headcount and jobs so that people can actually go do this work inside of products? Because these teams are already overtaxed — like these teams are really, really small in comparison to the products.”

Before Gebru and Mitchell, Google walkout organizers Meredith Whittaker and Claire Stapleton claimed they were the targets of retaliation before leaving the company, as did employees who attempted to unionize, many of whom identify as queer. Shortly before Gebru was fired, the National Labor Review Board filed a complaint against Google that accuses the company of retaliating against and illegally spying on employees.

The AI Index, an annual accounting of performance advances and AI’s impact on startups, business, and government policy, was released last week and found that the United States differs from other countries in its quantity of industry-backed research. The index report also called for more fairness benchmarks, found that Congress is talking about AI more than ever, and cites research finding only 3% of AI Ph.D. graduates in the U.S. are Black and 18% are women. The report notes that AI ethics incidents — including Google firing Gebru —  were among the most popular AI news stories of 2020.

VentureBeat requested an interview with Google VP Marian Croak, but a Google spokesperson declined on her behalf.

In a related matter, VentureBeat analysis about the “fight for the soul of machine learning” was cited in a paper published this week at FAccT about power, exclusion, and AI ethics education.


VentureBeat’s mission is to be a digital town square for technical decision-makers to gain knowledge about transformative technology and transact.

Our site delivers essential information on data technologies and strategies to guide you as you lead your organizations. We invite you to become a member of our community, to access:

  • up-to-date information on the subjects of interest to you
  • our newsletters
  • gated thought-leader content and discounted access to our prized events, such as Transform 2021: Learn More
  • networking features, and more

Become a member

Repost: Original Source and Author Link