Daniel Christian: My slides for the Educational Technology Organization of Michigan’s Spring 2024 Retreat

From DSC:
Last Thursday, I presented at the Educational Technology Organization of Michigan’s Spring 2024 Retreat. I wanted to pass along my slides to you all, in case they are helpful to you.

Topics/agenda:

  • Topics & resources re: Artificial Intelligence (AI)
    • Top multimodal players
    • Resources for learning about AI
    • Applications of AI
    • My predictions re: AI
  • The powerful impact of pursuing a vision
  • A potential, future next-gen learning platform
  • Share some lessons from my past with pertinent questions for you all now
  • The significant impact of an organization’s culture
  • Bonus material: Some people to follow re: learning science and edtech

 

Education Technology Organization of Michigan -- ETOM -- Spring 2024 Retreat on June 6-7

PowerPoint slides of Daniel Christian's presentation at ETOM

Slides of the presentation (.PPTX)
Slides of the presentation (.PDF)

 


Plus several more slides re: this vision.

 

 

Apple Intelligence: every new AI feature coming to the iPhone and Mac — from theverge.com by Wes Davis

Apple announced “Apple Intelligence” at WWDC 2024, its name for a new suite of AI features for the iPhone, Mac, and more. Starting later this year, Apple is rolling out what it says is a more conversational Siri, custom, AI-generated “Genmoji,” and GPT-4o access that lets Siri turn to OpenAI’s chatbot when it can’t handle what you ask it for.

Apple jumps into the AI arms race with OpenAI deal — from washingtonpost.com by Gerrit De Vynck
The iPhone maker has mostly stayed on the sidelines as the tech industry goes wild for AI. Not anymore.

SAN FRANCISCO — Apple officially launched itself into the artificial intelligence arms race, announcing a deal with ChatGPT maker OpenAI to use the company’s technology in its products and showing off a slew of its own new AI features.

The announcements, made at the tech giant’s annual Worldwide Developers Conference on Monday in Cupertino, Calif., are aimed at helping the tech giant keep up with competitors such as Google and Microsoft, which have boasted in recent months that AI makes their phones, laptops and software better than Apple’s. In addition to Apple’s own homegrown AI tech, the company’s phones, computers and iPads will also have ChatGPT built in “later this year,” a huge validation of the importance of the highflying start-up’s tech.

Apple Intelligence: AI for the rest of us. — from apple.com

  • Built into your iPhone, iPad, and Mac to help you write, express yourself, and get things done effortlessly.
  • Draws on your personal context while setting a brand-new standard for privacy in AI.

Introducing Apple Intelligence, the personal intelligence system that puts powerful generative models at the core of iPhone, iPad, and Mac — from apple.com
Setting a new standard for privacy in AI, Apple Intelligence understands personal context to deliver intelligence that is helpful and relevant

Apple doubles down on artificial intelligence, announcing partnership with OpenAI — from npr.org by Lola Murti and Dara Kerr

The highly anticipated AI partnership is the first of its kind for Apple, which has been regarded by analysts as slower to adopt artificial intelligence than other technology companies such as Microsoft and Google.

The deal allows Apple’s millions of users to access technology from OpenAI, one of the highest-profile artificial intelligence companies of recent years. OpenAI has already established partnerships with a variety of technology and publishing companies, including a multibillion-dollar deal with Microsoft.

 

The real deal here is that Apple is literally putting AI into the hands of >1B people, most of whom will probably be using AI for the 1st time. And it’s delivering AI that’s actually useful (forget those Genmojis, we’re talking about implanting ChatGPT-4o’s brain into Apple devices).

Noah Edelman (source)

Here’s everything Apple announced at the WWDC 2024 keynote, including Apple Intelligence, Siri makeover — from techcrunch.com by Christine Hall

It’s WWDC 2024 keynote time! Each year Apple kicks off its Worldwide Developers Conference with a few hours of just straight announcements, like the long-awaited Apple Intelligence and a makeover for smart AI assistant, Siri. We expected much of them to revolve around the company’s artificial intelligence ambitions (and here), and Apple didn’t disappoint. We also bring you news about Vision Pro and lots of feature refreshes.

Here’s how to watch the archive of WWDC 2024.


Why Gamma is great for presentations — from Jeremy Caplan

Gamma has become one of my favorite new creativity tools. You can use it like Powerpoint or Google Slides, adding text and images to make impactful presentations. It lets you create vertical, square or horizontal slides. You can embed online content to make your deck stand out with videos, data or graphics. You can even use it to make quick websites.

Its best feature, though, is an easy-to-use application of AI. The AI will learn from any document you import, or you can use a text prompt to create a strong deck or site instantly.
.


107 Up-to-Date ChatGPT Statistics & User Numbers [April 2024] — from nerdynav.com

Top ChatGPT Statistics

  • ChatGPT has 180.5 million users out of which 100 million users are active weekly.
  • In January 2024, ChatGPT got 2.3 billion website visits and 2 million developers are using its API.
  • The highest percentage of ChatGPT users belong to USA (46.75%), followed by India (5.47%). ChatGPT is banned in 7 countries including Russia and China.
  • OpenAI’s projected revenue from ChatGPT is $2billion in 2024.
  • Running ChatGPT costs OpenAI around $700,000 daily.
  • Sam Altman is seeking $7 trillion for a global AI chip project while Open AI is also listed as a major shareholder in Reddit.
  • ChatGPT offers a free version with GPT-3.5 and a Plus version with GPT-4, which is 40% more accurate and 82% safer costing $20 per month.
  • ChatGPT is being used for automation, education, coding, data-analysis, writing, etc.
  • 43% of college students and 80% of the Fortune 500 companies are using ChatGPT.
  • A 2023 study found 25% of US companies surveyed saved $50K-$70K using ChatGPT, while 11% saved over $100K.
 
 
 

Introducing Copilot+ PCs — from blogs.microsoft.com

[On May 20th], at a special event on our new Microsoft campus, we introduced the world to a new category of Windows PCs designed for AI, Copilot+ PCs.

Copilot+ PCs are the fastest, most intelligent Windows PCs ever built. With powerful new silicon capable of an incredible 40+ TOPS (trillion operations per second), all–day battery life and access to the most advanced AI models, Copilot+ PCs will enable you to do things you can’t on any other PC. Easily find and remember what you have seen in your PC with Recall, generate and refine AI images in near real-time directly on the device using Cocreator, and bridge language barriers with Live Captions, translating audio from 40+ languages into English.

From DSC:
As a first off-the-hip look, Recall could be fraught with possible security/privacy-related issues. But what do I know? The Neuron states “Microsoft assures that everything Recall sees remains private.” Ok…


From The Rundown AI concerning the above announcements:

The details:

  • A new system enables Copilot+ PCs to run AI workloads up to 20x faster and 100x more efficiently than traditional PCs.
    Windows 11 has been rearchitected specifically for AI, integrating the Copilot assistant directly into the OS.
  • New AI experiences include a new feature called Recall, which allows users to search for anything they’ve seen on their screen with natural language.
  • Copilot’s new screen-sharing feature allows AI to watch, hear, and understand what a user is doing on their computer and answer questions in real-time.
  • Copilot+ PCs will start at $999, and ship with OpenAI’s latest GPT-4o models.

Why it matters: Tony Stark’s all-powerful JARVIS AI assistant is getting closer to reality every day. Once Copilot, ChatGPT, Project Astra, or anyone else can not only respond but start executing tasks autonomously, things will start getting really exciting — and likely initiate a whole new era of tech work.


 

AI’s New Conversation Skills Eyed for Education — from insidehighered.com by Lauren Coffey
The latest ChatGPT’s more human-like verbal communication has professors pondering personalized learning, on-demand tutoring and more classroom applications.

ChatGPT’s newest version, GPT-4o ( the “o” standing for “omni,” meaning “all”), has a more realistic voice and quicker verbal response time, both aiming to sound more human. The version, which should be available to free ChatGPT users in coming weeks—a change also hailed by educators—allows people to interrupt it while it speaks, simulates more emotions with its voice and translates languages in real time. It also can understand instructions in text and images and has improved video capabilities.

Ajjan said she immediately thought the new vocal and video capabilities could allow GPT to serve as a personalized tutor. Personalized learning has been a focus for educators grappling with the looming enrollment cliff and for those pushing for student success.

There’s also the potential for role playing, according to Ajjan. She pointed to mock interviews students could do to prepare for job interviews, or, for example, using GPT to play the role of a buyer to help prepare students in an economics course.

 

 

Hello GPT-4o — from openai.com
We’re announcing GPT-4o, our new flagship model that can reason across audio, vision, and text in real time.

GPT-4o (“o” for “omni”) is a step towards much more natural human-computer interaction—it accepts as input any combination of text, audio, image, and video and generates any combination of text, audio, and image outputs. It can respond to audio inputs in as little as 232 milliseconds, with an average of 320 milliseconds, which is similar to human response time in a conversation. It matches GPT-4 Turbo performance on text in English and code, with significant improvement on text in non-English languages, while also being much faster and 50% cheaper in the API. GPT-4o is especially better at vision and audio understanding compared to existing models.

Example topics covered here:

  • Two GPT-4os interacting and singing
  • Languages/translation
  • Personalized math tutor
  • Meeting AI
  • Harmonizing and creating music
  • Providing inflection, emotions, and a human-like voice
  • Understanding what the camera is looking at and integrating it into the AI’s responses
  • Providing customer service

With GPT-4o, we trained a single new model end-to-end across text, vision, and audio, meaning that all inputs and outputs are processed by the same neural network. Because GPT-4o is our first model combining all of these modalities, we are still just scratching the surface of exploring what the model can do and its limitations.





From DSC:
I like the assistive tech angle here:





 

 

Description:

I recently created an AI version of myself—REID AI—and recorded a Q&A to see how this digital twin might challenge me in new ways. The video avatar is generated by Hour One, its voice was created by Eleven Labs, and its persona—the way that REID AI formulates responses—is generated from a custom chatbot built on GPT-4 that was trained on my books, speeches, podcasts and other content that I’ve produced over the last few decades. I decided to interview it to test its capability and how closely its responses match—and test—my thinking. Then, REID AI asked me some questions on AI and technology. I thought I would hate this, but I’ve actually ended up finding the whole experience interesting and thought-provoking.


From DSC:
This ability to ask questions of a digital twin is very interesting when you think about it in terms of “interviewing” a historical figure. I believe character.ai provides this kind of thing, but I haven’t used it much.


 

Smart(er) Glasses: Introducing New Ray-Ban | Meta Styles + Expanding Access to Meta AI with Vision — from meta.com

  • Share Your View on a Video Call
  • Meta AI Makes Your Smart Glasses Smarter
  • All In On AI-Powered Hardware

New Ray-Ban | Meta Smart Glasses Styles and Meta AI Updates — from about.fb.com

Takeaways

  • We’re expanding the Ray-Ban Meta smart glasses collection with new styles.
  • We’re adding video calling with WhatsApp and Messenger to share your view on a video call.
  • We’re rolling out Meta AI with Vision, so you can ask your glasses about what you’re seeing and get helpful information — completely hands-free.

 

12 Books for Instructional Designers to Read This Year — from theelearningcoach.com by Connie Malamed

Over the past year, many excellent and resourceful books have crossed my desk or Kindle. I’m rounding them up here so you can find a few to expand your horizons. The list below is in alphabetical order by title.

Each book is unique, yet as a collection, they reflect some common themes and trends in Learning and Development: a focus on empathy and emotion, adopting best practices from other fields, using data for greater impact, aligning projects with organizational goals, and developing consultative skills. The authors listed here are optimistic and forward-thinking—they believe change is possible. I hope you enjoy the books.

 

Below are some items for those creatives who might be interested in telling stories, designing games, crafting audio-based experiences, composing music, developing new worlds using 3D graphics, and more. 


CREATING THE SOUNDS OF LIGHTFALL — from bungie.net; via Mr. Robert Bender

The sounds of any game can make or break the experience for its players. Many of our favorite adventures come roaring back into our minds when we hear a familiar melody, or maybe it’s a special sound effect that reminds us of our time performing a particularly heroic feat… or the time we just caused some havoc with friends. With Lightfall sending Guardians to explore the new destination of Neomuna, there’s an entire universe hidden away within the sounds—both orchestral and diegetic—for Guardians to uncover and immerse themselves in. We recently assembled some of Destiny’s finest sound designers and composers to dive a little bit deeper into the stunning depths of Neomuna’s auditory experience.

Before diving into the interview with our incredible team, we wanted to make sure you have seen the Lightfall music documentary that went out shortly after the expansion’s release. This short video is a great introduction to how our team worked to create the music of Lightfall and is a must-see for audiophiles and Destiny fans alike.

.

.


Game Dev Diaries: The Hidden World of Audio — from lianaruppert.medium.com by Liana Ruppert, via Mr. Robert Bender

Every game has a story to tell, a journey to take players through that — if done well — can inspire wonderful memories that last a lifetime. Unlike other storytelling mediums, the art of video games is an intricate interweaving of experiences, including psychological cues that are designed to entrance players and make them feel like they’re a part of the story. One way this is achieved is through the art of audio. And no, we aren’t just talking about the many incredible soundtracks out there, we’re talking about the oftentimes overlooked universe of audio design.

What does an audio designer do?
“Number one? We don’t work on music. That’s a thing almost everyone thinks every audio designer does,” jokes Nyte when opening up about beginning her quest into the audio world. “That, or for a game like Destiny, people just assume we only work on weapon sounds and nothing else. Which, [Juan] Uribe does, but a lot of us don’t. There is this entire gamut of other sounds that are in-game that people don’t really notice. Some do, and that’s always cool, but audio is about all sounds coming together for a ‘whole’ audio experience.”


Also relevant/see:

The New Stack of Entertainment, Tensions of the AI Age, & Navigating Cambrian Explosions — from implications.com by Scott Belsky
Let’s explore some fun albeit heretical Hollywood possibilities, face key tensions, and talk about how to stay grounded with customer needs.

On the Transformation of Entertainment
What company will be the Pixar of the AI era? What talent agency will be the CAA of the AI era? How fast can the entertainment industry evolve to natively leverage AI, and what parts will be disrupted by the industry’s own ambivalence? Or are all of these questions myopic…and should we anticipate a wave of entirely new categories of entertainment?

We are starting to see material adoption of AI tools across many industries, including media and entertainment. No doubt, these tools will transform the processes behind generating content. But what entirely new genres of content might emerge? The platform shift to AI-based workflows might give rise to entirely new types of companies that transform entertainment as we know it – from actor representation, Hollywood economics, consumption devices and experiences, to the actual mediums of entertainment themselves. Let’s explore just a few of the more edgy implications:

 

What is executive function?

What is executive function? — from understood.org by Gail Belsky

Executive function is a set of mental skills that include working memory, flexible thinking, and self-control. We use these skills every day to learn, work, and manage daily life. Trouble with executive function can make it hard to focus, follow directions, and handle emotions, among other things.

Snapshot: What executive function is
Some people describe executive function as “the management system of the brain.” That’s because the skills involved let us set goals, plan, and get things done. When people struggle with executive function, it impacts them at home, in school, and in life.

There are three main areas of executive function. They are…

 
 

From DSC:
I recently ran into the following item:


UK university opens VR classroom — from inavateonthenet.net

Students at the University of Nottingham will be learning through a dedicated VR classroom, enabling remote viewing and teaching for students and lecturers.

Based in the university’s Engineering Science and Learning Centre (ELSC), this classroom, believed to be the first in the UK to use a dedicated VR classroom, using 40 VR headsets, 35 of which are tethered overhead to individual PCs, with five available as traditional, desk-based systems with display screens.


I admit that I was excited to see this article and I congratulate the University of Nottingham on their vision here. I hope that they can introduce more use cases and applications to provide evidence of VR’s headway.

As I look at virtual reality…

  • On the plus side, I’ve spoken with people who love to use their VR-based headsets for fun workouts/exercises. I’ve witnessed the sweat, so I know that’s true. And I believe there is value in having the ability to walk through museums that one can’t afford to get to. And I’m sure that the gamers have found some incredibly entertaining competitions out there. The experience of being immersed can be highly engaging. So there are some niche use cases for sure.
  • But on the negative side, the technologies surrounding VR haven’t progressed as much as I thought they would have by now. For example, I’m disappointed Apple’s taken so long to put a product out there, and I don’t want to invest $3500 in their new product. From the reviews and items on social media that I’ve seen, the reception is lukewarm. At the most basic level, I’m not sure people want to wear a headset for more than a few minutes.

So overall, I’d like to see more use cases and less nausea.


Addendum on 2/27/24:

Leyard ‘wall of wonder’ wows visitors at Molecular Biology Lab — from inavateonthenet.net

 

Enter the New Era of Mobile AI With Samsung Galaxy S24 Series — from news.samsung.com

Galaxy AI introduces meaningful intelligence aimed at enhancing every part of life, especially the phone’s most fundamental role: communication. When you need to defy language barriers, Galaxy S24 makes it easier than ever. Chat with another student or colleague from abroad. Book a reservation while on vacation in another country. It’s all possible with Live Translate,2 two-way, real-time voice and text translations of phone calls within the native app. No third-party apps are required, and on-device AI keeps conversations completely private.

With Interpreter, live conversations can be instantly translated on a split-screen view so people standing opposite each other can read a text transcription of what the other person has said. It even works without cellular data or Wi-Fi.


Galaxy S24 — from theneurondaily.com by Noah Edelman & Pete Huang

Samsung just announced the first truly AI-powered smartphone: the Galaxy S24.


For us AI power users, the features aren’t exactly new, but it’s the first time we’ve seen them packaged up into a smartphone (Siri doesn’t count, sorry).


Samsung’s Galaxy S24 line arrives with camera improvements and generative AI tricks — from techcrunch.com by Brian Heater
Starting at $800, the new flagships offer brighter screens and a slew of new photo-editing tools

 
© 2024 | Daniel Christian