Artificial Intelligence: Should You Teach It To Your Employees?— from forbes.com by Tom Taulli

Excerpt:

“If more people are AI literate and can start to participate and contribute to the process, more problems–both big and small–across the organization can be tackled,” said David Sweenor, who is the Senior Director of Product Marketing at Alteryx. “We call this the ‘Democratization of AI and Analytics.’ A team of 100, 1,000, or 5,000 working on different problems in their areas of expertise certainly will have a bigger impact than if left in the hands of a few.”

New Artificial Intelligence Tool Accelerates Discovery of Truly New Materials — from scitechdaily.com
The new artificial intelligence tool has already led to the discovery of four new materials.

Excerpt:

Researchers at the University of Liverpool have created a collaborative artificial intelligence tool that reduces the time and effort required to discover truly new materials.

AI development must be guided by ethics, human wellbeing and responsible innovation — from healthcareitnews.com by Bill Siwicki
An expert in emerging technology from the IEEE Standards Association describes the elements that must be considered as artificial intelligence proliferates across healthcare.

 

In the US, the AI Industry Risks Becoming Winner-Take-Most — from wired.com by Khari Johnson
A new study illustrates just how geographically concentrated AI activity has become.

Excerpt:

A NEW STUDY warns that the American AI industry is highly concentrated in the San Francisco Bay Area and that this could prove to be a weakness in the long run. The Bay leads all other regions of the country in AI research and investment activity, accounting for about one-quarter of AI conference papers, patents, and companies in the US. Bay Area metro areas see levels of AI activity four times higher than other top cities for AI development.

“When you have a high percentage of all AI activity in Bay Area metros, you may be overconcentrating, losing diversity, and getting groupthink in the algorithmic economy. It locks in a winner-take-most dimension to this sector, and that’s where we hope that federal policy will begin to invest in new and different AI clusters in new and different places to provide a balance or counter,” Mark Muro, policy director at the Brookings Institution and the study’s coauthor, told WIRED.

Also relevant/see:

 

“Algorithms are opinions embedded in code.”

 

The Fight to Define When AI Is ‘High Risk’ — from wired.com by Khari Johnson
Everyone from tech companies to churches wants a say in how the EU regulates AI that could harm people.

Excerpt:

The AI Act is one of the first major policy initiatives worldwide focused on protecting people from harmful AI. If enacted, it will classify AI systems according to risk, more strictly regulate AI that’s deemed high risk to humans, and ban some forms of AI entirely, including real-time facial recognition in some instances. In the meantime, corporations and interest groups are publicly lobbying lawmakers to amend the proposal according to their interests.

 

Many Americans aren’t aware they’re being tracked with facial recognition while shopping  — from techradar.com by Anthony Spadafora
You’re not just on camera, you’re also being tracked

Excerpt:

Despite consumer opposition to facial recognition, the technology is currently being used in retail stores throughout the US according to new research from Piplsay.

While San Francisco banned the police from using facial recognition back in 2019 and the EU called for a five year ban on the technology last year, several major retailers in the US including Lowe’s, Albertsons and Macy’s have been using it for both fraud and theft detection.

From DSC:
I’m not sure how prevalent this practice is…and that’s precisely the point. We don’t know what all of those cameras are actually doing in our stores, gas stations, supermarkets, etc. I put this in the categories of policy, law schools, legal, government, and others as the legislative and legal realm need to scramble to catch up to this Wild Wild West.

Along these lines, I was watching a portion of 60 minutes last night where they were doing a piece on autonomous trucks (reportedly to hit the roads without a person sometime later this year). When asked about oversight, there was some…but not much.

Readers of this blog will know that I have often wondered…”How does society weigh in on these things?”

Along these same lines, also see:

  • The NYPD Had a Secret Fund for Surveillance Tools — from wired.com by Sidney Fussell
    Documents reveal that police bought facial-recognition software, vans equipped with x-ray machines, and “stingray” cell site simulators—with no public oversight.
 

The Future of Social Media: Re-Humanisation and Regulation — by Gerd Leonhard

How could social media become ‘human’ again? How can we stop the disinformation, dehumanisation and dataism that has resulted from social media’s algorithmic obsessions? I foresee that the EXTERNALTIES i.e. the consequences of unmitigated growth of exponential digital technologies will become just as big as the consequences of climate change. In fact, today, the social media industry already has quite a few parallels to the oil, gas and coal business: while private make huge profits from extracting the ‘oil’ (i.e. user data), the external damage is left to society and governments to fix. This needs to change! In this keynote I make some precise suggestions as to how that could happen.

Some snapshots/excerpts:

The future of social media -- a video by Gerd Leonhard in the summer of 2021

 

 

 

 


From DSC:
Gerd brings up some solid points here. His presentation and perspectives are not only worth checking out, but they’re worth some time for us to seriously reflect on what he’s saying.

What kind of future do we want?

And for you professors, teachers, instructional designers, trainers, and presenters out there, check out *how* he delivers the content. It’s well done and very engaging.


 

Let’s Teach Computer Science Majors to Be Good Citizens. The Whole World Depends on It. — from edsurge.com by Anne-Marie Núñez, Matthew J. Mayhew, Musbah Shaheen and Laura S. Dahl

Excerpt:

To mitigate the perpetuation of these and related inequities, observers have called for increased diversification of the technology workforce. However, as books like “Brotopia” by Emily Chang and “Race after Technology” by Ruha Benjamin indicate, the culture of tech companies can be misogynistic and racist and therefore unwelcoming to many people. Google’s firing of a well-regarded Black scientist for her research on algorithmic bias in December 2020 suggests that there may be limited capacity within the industry to challenge this culture.

Change may need to start earlier in the workforce development pipeline. Undergraduate education offers a key opportunity for recruiting students from historically underrepresented racial and ethnic, gender, and disability groups into computing. Yet even broadened participation in college computer science courses may not shift the tech workforce and block bias from seeping into tech tools if students aren’t taught that diversity and ethics are essential to their field of study and future careers.

Also mentioned/see:

  • Teaching Responsible Computing Playbook
    The ultimate goal of Teaching Responsible Computing is to educate a new wave of students who bring holistic thinking to the design of technology products. To do this, it is critical for departments to work together across computing, humanistic studies, and more, and collaborate across institutions. This Playbook offers the lessons learned from the process of adapting and enhancing curricula to include responsible computing in a broad set of institutions and help others get started doing the same in their curricula. 
 

This is an abstract picture of a person's head made of connections peering sideways -- it links to Artificial intelligence and the future of national security from ASU

Artificial intelligence and the future of national security — from news.asu.edu

Excerpt:

Artificial intelligence is a “world-altering” technology that represents “the most powerful tools in generations for expanding knowledge, increasing prosperity and enriching the human experience” and will be a source of enormous power for the companies and countries that harness them, according to the recently released Final Report of the National Security Commission on Artificial Intelligence.

This is not hyperbole or a fantastical version of AI’s potential impact. This is the assessment of a group of leading technologists and national security professionals charged with offering recommendations to Congress on how to ensure American leadership in AI for national security and defense. Concerningly, the group concluded that the U.S. is not currently prepared to defend American interests or compete in the era of AI.

Also see:

EU Set to Ban Surveillance, Start Fines Under New AI Rules — from bloomberg.com by Natalia Drozdiak

Excerpt:

The European Union is poised to ban artificial intelligence systems used for mass surveillance or for ranking social behavior, while companies developing AI could face fines as high as 4% of global revenue if they fail to comply with new rules governing the software applications.

Also see:

Wrongfully arrested man sues Detroit police over false facial recognition match — from washingtonpost.com by Drew Harwell
The case could fuel criticism of police investigators’ use of a controversial technology that has been shown to perform worse on people of color

Excerpts:

A Michigan man has sued Detroit police after he was wrongfully arrested and falsely identified as a shoplifting suspect by the department’s facial recognition software in one of the first lawsuits of its kind to call into question the controversial technology’s risk of throwing innocent people in jail.

Robert Williams, a 43-year-old father in the Detroit suburb of Farmington Hills, was arrested last year on charges he’d taken watches from a Shinola store after police investigators used a facial recognition search of the store’s surveillance-camera footage that identified him as the thief.

Prosecutors dropped the case less than two weeks later, arguing that officers had relied on insufficient evidence. Police Chief James Craig later apologized for what he called “shoddy” investigative work. Williams, who said he had been driving home from work when the 2018 theft had occurred, was interrogated by detectives and held in custody for 30 hours before his release.

Williams’s attorneys did not make him available for comment Tuesday. But Williams wrote in The Washington Post last year that the episode had left him deeply shaken, in part because his young daughters had watched him get handcuffed in his driveway and put into a police car after returning home from work.

“How does one explain to two little girls that a computer got it wrong, but the police listened to it anyway?” he wrote. “As any other black man would be, I had to consider what could happen if I asked too many questions or displayed my anger openly — even though I knew I had done nothing wrong.”

Addendum on 4/20/21:

 

How a Discriminatory Algorithm Wrongly Accused Thousands of Families of Fraud — from vice.com by Gabriel Geiger; with thanks to Sam DeBrule for this resource
Dutch tax authorities used algorithms to automate an austere and punitive war on low-level fraud—the results were catastrophic.

Excerpt:

Last month, Prime Minister of the Netherlands Mark Rutte—along with his entire cabinet—resigned after a year and a half of investigations revealed that since 2013, 26,000 innocent families were wrongly accused of social benefits fraud partially due to a discriminatory algorithm.

Forced to pay back money they didn’t owe, many families were driven to financial ruin, and some were torn apart. Others were left with lasting mental health issues; people of color were disproportionately the victims.

On a more positive note, Sam DeBrule (in his Machine Learnings e-newsletter) also notes the following article:

Can artificial intelligence combat wildfires? Sonoma County tests new technology — from latimes.com by Alex Wigglesworth

 

From DSC:
The items below are from Sam DeBrule’s Machine Learnings e-Newsletter.


By clicking this image, you will go to Sam DeBrule's Machine Learning e-Newsletter -- which deals with all topics regarding Artificial Intelligence

#Awesome

“Sonoma County is adding artificial intelligence to its wildfire-fighting arsenal. The county has entered into an agreement with the South Korean firm Alchera to outfit its network of fire-spotting cameras with software that detects wildfire activity and then alerts authorities. The technology sifts through past and current images of terrain and searches for certain changes, such as flames burning in darkness, or a smoky haze obscuring a tree-lined hillside, according to Chris Godley, the county’s director of emergency management…The software will use feedback from humans to refine its algorithm and will eventually be able to detect fires on its own — or at least that’s what county officials hope.” – Alex Wigglesworth Learn More from Los Angeles Times >

#Not Awesome

Hacked Surveillance Camera Firm Shows Staggering Scale of Facial Recognition — from
A hacked customer list shows that facial recognition company Verkada is deployed in tens of thousands of schools, bars, stores, jails, and other businesses around the country.

Excerpt:

Hackers have broken into Verkada, a popular surveillance and facial recognition camera company, and managed to access live feeds of thousands of cameras across the world, as well as siphon a Verkada customer list. The breach shows the astonishing reach of facial recognition-enabled cameras in ordinary workplaces, bars, parking lots, schools, stores, and more.

The staggering list includes K-12 schools, seemingly private residences marked as “condos,” shopping malls, credit unions, multiple universities across America and Canada, pharmaceutical companies, marketing agencies, pubs and bars, breweries, a Salvation Army center, churches, the Professional Golfers Association, museums, a newspaper’s office, airports, and more.

 

The Chegg situation is worse than you think — from eliterate.us by Michael Feldstein

Excerpts:

Forbes just ran with an article entitled “This $12 Billion Company Is Getting Rich Off Students Cheating Their Way Through Covid“.

Ouch.

Chegg -- This $12 Billion Company Is Getting Rich Off Students Cheating Their Way Through Covid

[Per Michael] To sum up:

  • Publishers, after selling expensive textbooks to students, sold the answers to the homework questions in those expensive books to Chegg.
  • Chegg sells the answers to the questions to the students, who often use them to cheat.
  • To combat this problem, universities pay for proctoring software, which is apparently more effective at preventing students from going to the bathroom than it is at preventing cheating.
  • To add insult to all of this injury, “to chegg” is now apparently a verb in the English language. We will all have to live with that linguistic violence.

Addendum on 2/9/21:

 

Could 2021 Be The Year Of Civil Justice Reform? — from law360.com by Cara Bayles

Excerpt:

Now, any tenant in Boulder, regardless of means, can get an attorney to represent them in housing court for free. While a handful of other cities like New York and San Francisco offer similar programs, there is no universal right to an attorney in eviction cases.

Justice advocates are hoping that could change. COVID-19, they say, has drawn attention to access to justice issues that have plagued civil proceedings for years. They hope the tragedies of 2020 can fuel reform in 2021.

Six months into the pandemic, millions of Americans had fallen behind on rent. One statistic that civil justice advocates have known for years became apparent — that while 90% of landlords have legal counsel in eviction proceedings, only 10% of tenants do.

 

 

From DSC:
An interesting, more positive use of AI here:

Deepdub uses AI to dub movies in the voice of famous actors — from protocol.com by Janko Roettgers
Fresh out of stealth, the startup is using artificial intelligence to automate the localization process for global streaming.

Excerpt:

Tel Aviv-based startup Deepdub wants to help streaming services accelerate this kind of international rollout by using artificial intelligence for their localization needs. Deepdub, which came out of stealth on Wednesday, has built technology that can translate a voice track to a different language, all while staying true to the voice of the talent. This makes it possible to have someone like Morgan Freeman narrate a movie in French, Italian or Russian without losing what makes Freeman’s voice special and recognizable.

From DSC:
A much more negative use of AI here:

A much more negative use of AI here...

 

 

DC: This is not right! When will there be justice?! We need to help Native American colleges, universities, & communities out!

The Digital Divide for Tribal College Students — COVID, CARES Act, and Critical Next Steps — from diverseeducation.com

Excerpt:

In this episode staff writer Sara Weissman shares a story that focuses on the digital divide for Native Americans by bringing in voices of tribal college leaders and their students during the COVID 19 pandemic.

Many don’t know but Native American colleges and universities have long struggled with the worst internet connectivity in the nation while ironically paying the highest rates for service. Hear first-hand how students from Diné College and other institutions are currently affected. Carrie Billie (Big Water Clan), President & CEO of the American Indian Higher Education Consortium (AIHEC) and Dr. Cynthia Lindquist (Star Horse Woman), President of Cankdeska Cikana Community College in North Dakota, break down the data and lay out critical next steps necessary to address the digital divide.

Many don’t know but Native American colleges and universities have long struggled with the worst internet connectivity in the nation while ironically paying the highest rates for service.

From DSC:
When will there be justice!? Let’s join in and make amends and provide the funding, concrete assistance, products, and services to Native American colleges, universities, and communities. Some potential ideas:

  • For the short term, could there be Loon balloons deployed immediately to provide free and stronger access to the Internet?

Could Project Loon assist Native American colleges, universities, and communities?

  • Could our Federal Government make amends and do the right thing here? (e-rate program, put Internet access in, make policy changes, offer more grants, other?)
  • Could Silicon Valley assist with hardware and software? For example:
    • Can Apple, HP, Microsoft, and others donate hardware and software?
    • Can Zoom, Adobe, Cisco Webex, Microsoft Teams, and others donate whatever these communities need to provide videoconferencing licenses?
  • Could telecom providers provide free internet access?
  • Could MOOCs offer more free courses?
  • Could furniture makers such as Steelcase, Herman Miller, and others donate furniture and help establish connected learning spaces?
  • How might faculty members and staff within higher education contribute?
  • How could churches, synagogues, and such get involved?
  • Could the rest of us locate and donate to charities that aim to provide concrete assistance to Native American schools, colleges, universities, and communities?

We need to do the right thing here. This is another area* where our nation can do much better.

* Here’s another example/area where we can do much better and make amends/changes.

 


Addendum on 12/7/20:

 
© 2024 | Daniel Christian