Google CEO Still Insists AI Revolution Bigger Than Invention of Fire — from gizmodo.com by Matt Novak
Pichai suggests the internet and electricity are also small potatoes compared to AI.

Excerpt:

The artificial intelligence revolution is poised to be more “profound” than the invention of electricity, the internet, and even fire, according to Google CEO Sundar Pichai, who made the comments to BBC media editor Amol Rajan in a podcast interview that first went live on Sunday.

“The progress in artificial intelligence, we are still in very early stages, but I viewed it as the most profound technology that humanity will ever develop and work on, and we have to make sure we do it in a way that we can harness it to society’s benefit,” Pichai said.

“But I expect it to play a foundational role pretty much across every aspect of our lives. You know, be it health care, be it education, be it how we manufacture things and how we consume information. 

 

The Future of Social Media: Re-Humanisation and Regulation — by Gerd Leonhard

How could social media become ‘human’ again? How can we stop the disinformation, dehumanisation and dataism that has resulted from social media’s algorithmic obsessions? I foresee that the EXTERNALTIES i.e. the consequences of unmitigated growth of exponential digital technologies will become just as big as the consequences of climate change. In fact, today, the social media industry already has quite a few parallels to the oil, gas and coal business: while private make huge profits from extracting the ‘oil’ (i.e. user data), the external damage is left to society and governments to fix. This needs to change! In this keynote I make some precise suggestions as to how that could happen.

Some snapshots/excerpts:

The future of social media -- a video by Gerd Leonhard in the summer of 2021

 

 

 

 


From DSC:
Gerd brings up some solid points here. His presentation and perspectives are not only worth checking out, but they’re worth some time for us to seriously reflect on what he’s saying.

What kind of future do we want?

And for you professors, teachers, instructional designers, trainers, and presenters out there, check out *how* he delivers the content. It’s well done and very engaging.


 

21 jobs of the future: A guide to getting — and staying — employed over the next 10 years — from cognizant.com and  the Center for The Future of Work

Excerpt:

WHAT THE NEXT 10 YEARS WILL BRING: NEW JOBS
In this report, we propose 21 new jobs that will emerge over the next 10 years and will become cornerstones of the future of work. In producing this report, we imagined hundreds of jobs that could emerge within the major macroeconomic, political, demographic, societal, cultural, business and technology trends observable today, e.g., growing populations, aging populations, populism, environmentalism, migration, automation, arbitrage, quantum physics, AI, biotechnology, space exploration, cybersecurity, virtual reality.

Among the jobs we considered, some seemed further out on the horizon and are not covered here: carbon farmers, 3-D printing engineers, avatar designers, cryptocurrency arbitrageurs, drone jockeys, human organ developers, teachers of English as a foreign language for robots, robot spa owners, algae farmers, autonomous fleet valets, Snapchat addiction therapists, urban vertical farmers and Hyperloop construction managers. These are jobs that younger generations may do in the further off future.

21 jobs on a chart where tech-centricity is on the vertical axis and the time horizon is on the horizontal axis. 21 jobs are represented in this graphic and report.

Also see:

Here are the top 10 jobs of the future — from bigthink.com by Robert Brown
Say hello to your new colleague, the Workplace Environment Architect.

Excerpt:

6. Algorithm Bias Auditor – “All online, all the time” lifestyles for work and leisure accelerated the competitive advantage derived from algorithms by digital firms everywhere. But from Brussels to Washington, given the increasing statutory scrutiny on data, it’s a near certainty that when it comes to how they’re built, verification through audits will help ensure the future workforce is also the fair workforce.

 

Let’s Teach Computer Science Majors to Be Good Citizens. The Whole World Depends on It. — from edsurge.com by Anne-Marie Núñez, Matthew J. Mayhew, Musbah Shaheen and Laura S. Dahl

Excerpt:

To mitigate the perpetuation of these and related inequities, observers have called for increased diversification of the technology workforce. However, as books like “Brotopia” by Emily Chang and “Race after Technology” by Ruha Benjamin indicate, the culture of tech companies can be misogynistic and racist and therefore unwelcoming to many people. Google’s firing of a well-regarded Black scientist for her research on algorithmic bias in December 2020 suggests that there may be limited capacity within the industry to challenge this culture.

Change may need to start earlier in the workforce development pipeline. Undergraduate education offers a key opportunity for recruiting students from historically underrepresented racial and ethnic, gender, and disability groups into computing. Yet even broadened participation in college computer science courses may not shift the tech workforce and block bias from seeping into tech tools if students aren’t taught that diversity and ethics are essential to their field of study and future careers.

Also mentioned/see:

  • Teaching Responsible Computing Playbook
    The ultimate goal of Teaching Responsible Computing is to educate a new wave of students who bring holistic thinking to the design of technology products. To do this, it is critical for departments to work together across computing, humanistic studies, and more, and collaborate across institutions. This Playbook offers the lessons learned from the process of adapting and enhancing curricula to include responsible computing in a broad set of institutions and help others get started doing the same in their curricula. 
 

This is an abstract picture of a person's head made of connections peering sideways -- it links to Artificial intelligence and the future of national security from ASU

Artificial intelligence and the future of national security — from news.asu.edu

Excerpt:

Artificial intelligence is a “world-altering” technology that represents “the most powerful tools in generations for expanding knowledge, increasing prosperity and enriching the human experience” and will be a source of enormous power for the companies and countries that harness them, according to the recently released Final Report of the National Security Commission on Artificial Intelligence.

This is not hyperbole or a fantastical version of AI’s potential impact. This is the assessment of a group of leading technologists and national security professionals charged with offering recommendations to Congress on how to ensure American leadership in AI for national security and defense. Concerningly, the group concluded that the U.S. is not currently prepared to defend American interests or compete in the era of AI.

Also see:

EU Set to Ban Surveillance, Start Fines Under New AI Rules — from bloomberg.com by Natalia Drozdiak

Excerpt:

The European Union is poised to ban artificial intelligence systems used for mass surveillance or for ranking social behavior, while companies developing AI could face fines as high as 4% of global revenue if they fail to comply with new rules governing the software applications.

Also see:

Wrongfully arrested man sues Detroit police over false facial recognition match — from washingtonpost.com by Drew Harwell
The case could fuel criticism of police investigators’ use of a controversial technology that has been shown to perform worse on people of color

Excerpts:

A Michigan man has sued Detroit police after he was wrongfully arrested and falsely identified as a shoplifting suspect by the department’s facial recognition software in one of the first lawsuits of its kind to call into question the controversial technology’s risk of throwing innocent people in jail.

Robert Williams, a 43-year-old father in the Detroit suburb of Farmington Hills, was arrested last year on charges he’d taken watches from a Shinola store after police investigators used a facial recognition search of the store’s surveillance-camera footage that identified him as the thief.

Prosecutors dropped the case less than two weeks later, arguing that officers had relied on insufficient evidence. Police Chief James Craig later apologized for what he called “shoddy” investigative work. Williams, who said he had been driving home from work when the 2018 theft had occurred, was interrogated by detectives and held in custody for 30 hours before his release.

Williams’s attorneys did not make him available for comment Tuesday. But Williams wrote in The Washington Post last year that the episode had left him deeply shaken, in part because his young daughters had watched him get handcuffed in his driveway and put into a police car after returning home from work.

“How does one explain to two little girls that a computer got it wrong, but the police listened to it anyway?” he wrote. “As any other black man would be, I had to consider what could happen if I asked too many questions or displayed my anger openly — even though I knew I had done nothing wrong.”

Addendum on 4/20/21:

 

How a Discriminatory Algorithm Wrongly Accused Thousands of Families of Fraud — from vice.com by Gabriel Geiger; with thanks to Sam DeBrule for this resource
Dutch tax authorities used algorithms to automate an austere and punitive war on low-level fraud—the results were catastrophic.

Excerpt:

Last month, Prime Minister of the Netherlands Mark Rutte—along with his entire cabinet—resigned after a year and a half of investigations revealed that since 2013, 26,000 innocent families were wrongly accused of social benefits fraud partially due to a discriminatory algorithm.

Forced to pay back money they didn’t owe, many families were driven to financial ruin, and some were torn apart. Others were left with lasting mental health issues; people of color were disproportionately the victims.

On a more positive note, Sam DeBrule (in his Machine Learnings e-newsletter) also notes the following article:

Can artificial intelligence combat wildfires? Sonoma County tests new technology — from latimes.com by Alex Wigglesworth

 

From DSC:
The items below are from Sam DeBrule’s Machine Learnings e-Newsletter.


By clicking this image, you will go to Sam DeBrule's Machine Learning e-Newsletter -- which deals with all topics regarding Artificial Intelligence

#Awesome

“Sonoma County is adding artificial intelligence to its wildfire-fighting arsenal. The county has entered into an agreement with the South Korean firm Alchera to outfit its network of fire-spotting cameras with software that detects wildfire activity and then alerts authorities. The technology sifts through past and current images of terrain and searches for certain changes, such as flames burning in darkness, or a smoky haze obscuring a tree-lined hillside, according to Chris Godley, the county’s director of emergency management…The software will use feedback from humans to refine its algorithm and will eventually be able to detect fires on its own — or at least that’s what county officials hope.” – Alex Wigglesworth Learn More from Los Angeles Times >

#Not Awesome

Hacked Surveillance Camera Firm Shows Staggering Scale of Facial Recognition — from
A hacked customer list shows that facial recognition company Verkada is deployed in tens of thousands of schools, bars, stores, jails, and other businesses around the country.

Excerpt:

Hackers have broken into Verkada, a popular surveillance and facial recognition camera company, and managed to access live feeds of thousands of cameras across the world, as well as siphon a Verkada customer list. The breach shows the astonishing reach of facial recognition-enabled cameras in ordinary workplaces, bars, parking lots, schools, stores, and more.

The staggering list includes K-12 schools, seemingly private residences marked as “condos,” shopping malls, credit unions, multiple universities across America and Canada, pharmaceutical companies, marketing agencies, pubs and bars, breweries, a Salvation Army center, churches, the Professional Golfers Association, museums, a newspaper’s office, airports, and more.

 

The Chegg situation is worse than you think — from eliterate.us by Michael Feldstein

Excerpts:

Forbes just ran with an article entitled “This $12 Billion Company Is Getting Rich Off Students Cheating Their Way Through Covid“.

Ouch.

Chegg -- This $12 Billion Company Is Getting Rich Off Students Cheating Their Way Through Covid

[Per Michael] To sum up:

  • Publishers, after selling expensive textbooks to students, sold the answers to the homework questions in those expensive books to Chegg.
  • Chegg sells the answers to the questions to the students, who often use them to cheat.
  • To combat this problem, universities pay for proctoring software, which is apparently more effective at preventing students from going to the bathroom than it is at preventing cheating.
  • To add insult to all of this injury, “to chegg” is now apparently a verb in the English language. We will all have to live with that linguistic violence.

Addendum on 2/9/21:

 

Could 2021 Be The Year Of Civil Justice Reform? — from law360.com by Cara Bayles

Excerpt:

Now, any tenant in Boulder, regardless of means, can get an attorney to represent them in housing court for free. While a handful of other cities like New York and San Francisco offer similar programs, there is no universal right to an attorney in eviction cases.

Justice advocates are hoping that could change. COVID-19, they say, has drawn attention to access to justice issues that have plagued civil proceedings for years. They hope the tragedies of 2020 can fuel reform in 2021.

Six months into the pandemic, millions of Americans had fallen behind on rent. One statistic that civil justice advocates have known for years became apparent — that while 90% of landlords have legal counsel in eviction proceedings, only 10% of tenants do.

 

 

Thank you LORD for Dr. Martin Luther King Jr.!!!

Martin Luther King Jr. Day

Excerpt:

[Dr.] King was the chief spokesperson for nonviolent activism in the Civil Rights Movement, which successfully protested racial discrimination in federal and state law. The campaign for a federal holiday in King’s honor began soon after his assassination in 1968. President Ronald Reagan signed the holiday into law in 1983, and it was first observed three years later. At first, some states resisted observing the holiday as such, giving it alternative names or combining it with other holidays. It was officially observed in all 50 states for the first time in 2000.

 

He has shown you, O mortal, what is good.
    And what does the Lord require of you?
To act justly and to love mercy
    and to walk humbly[a] with your God.

 

From DSC:
An interesting, more positive use of AI here:

Deepdub uses AI to dub movies in the voice of famous actors — from protocol.com by Janko Roettgers
Fresh out of stealth, the startup is using artificial intelligence to automate the localization process for global streaming.

Excerpt:

Tel Aviv-based startup Deepdub wants to help streaming services accelerate this kind of international rollout by using artificial intelligence for their localization needs. Deepdub, which came out of stealth on Wednesday, has built technology that can translate a voice track to a different language, all while staying true to the voice of the talent. This makes it possible to have someone like Morgan Freeman narrate a movie in French, Italian or Russian without losing what makes Freeman’s voice special and recognizable.

From DSC:
A much more negative use of AI here:

A much more negative use of AI here...

 

 

The Digital Divide for Tribal College Students — COVID, CARES Act, and Critical Next Steps — from diverseeducation.com

Excerpt:

In this episode staff writer Sara Weissman shares a story that focuses on the digital divide for Native Americans by bringing in voices of tribal college leaders and their students during the COVID 19 pandemic.

Many don’t know but Native American colleges and universities have long struggled with the worst internet connectivity in the nation while ironically paying the highest rates for service. Hear first-hand how students from Diné College and other institutions are currently affected. Carrie Billie (Big Water Clan), President & CEO of the American Indian Higher Education Consortium (AIHEC) and Dr. Cynthia Lindquist (Star Horse Woman), President of Cankdeska Cikana Community College in North Dakota, break down the data and lay out critical next steps necessary to address the digital divide.

Many don’t know but Native American colleges and universities have long struggled with the worst internet connectivity in the nation while ironically paying the highest rates for service.

From DSC:
When will there be justice!? Let’s join in and make amends and provide the funding, concrete assistance, products, and services to Native American colleges, universities, and communities. Some potential ideas:

  • For the short term, could there be Loon balloons deployed immediately to provide free and stronger access to the Internet?

Could Project Loon assist Native American colleges, universities, and communities?

  • Could our Federal Government make amends and do the right thing here? (e-rate program, put Internet access in, make policy changes, offer more grants, other?)
  • Could Silicon Valley assist with hardware and software? For example:
    • Can Apple, HP, Microsoft, and others donate hardware and software?
    • Can Zoom, Adobe, Cisco Webex, Microsoft Teams, and others donate whatever these communities need to provide videoconferencing licenses?
  • Could telecom providers provide free internet access?
  • Could MOOCs offer more free courses?
  • Could furniture makers such as Steelcase, Herman Miller, and others donate furniture and help establish connected learning spaces?
  • How might faculty members and staff within higher education contribute?
  • How could churches, synagogues, and such get involved?
  • Could the rest of us locate and donate to charities that aim to provide concrete assistance to Native American schools, colleges, universities, and communities?

We need to do the right thing here. This is another area* where our nation can do much better.

* Here’s another example/area where we can do much better and make amends/changes.

 


Addendum on 12/7/20:

 

Pushback is growing against automated proctoring services. But so is their use — from edsurge.com by Jeffrey R. Young

Excerpt:

Many students have pushed back, arguing that remote proctoring tools result in a serious invasion of privacy and create stress that can hinder academic performance. More than 60,000 students across the U.S. have signed petitions calling on their colleges to stop using automated proctoring tools, meaning that the technology has become arguably the most controversial tool of the pandemic at colleges.

From DSC:
We have an issue oftentimes within higher education — including graduate schools/professional schools as well — where the student and the professor aren’t always on the same team (or at least that’s the percaption). To me, the professors need to be saying (and living out the message that), “We ARE on your team. We are working to help make you successful in the future that you have chosen for yourself. We’re here to help you…not monitor you.”

It’s like I feel when I walk into so many public places these days (and even out on the roadways as well). When I walk into a store, it’s like the security cameras are whispering to me and to others…”We don’t trust you. Some of you have stolen in the past. so we’re going to carefully watch every single one of you. And we aren’t just going to watch you, we’re going to record you as well.”

The message? We don’t trust you.

This severely hampers the relationships involved.

And I’m sure that cheating is going on. But then, that makes me think that perhaps it’s time to change the way we assess students — and to help them see assessments as opportunities to learn, not to cheat. 

Lower the stakes. Offer tests more frequently. Provide more opportunities to practice recall. And be on their team.

 

“Especially given that these systems replicate and amplify the harms of structural racism and historical discrimination, which fall predominantly on Black, brown, and poor communities.”

From DSC:
Some serious fodder for thought in this article. I’d like to see
#computerscience students and faculty members debate and/or weigh in on this type of topic.

 
© 2021 | Daniel Christian