New models and developer products announced at DevDay — from openai.com
GPT-4 Turbo with 128K context and lower prices, the new Assistants API, GPT-4 Turbo with Vision, DALL·E 3 API, and more.

Today, we shared dozens of new additions and improvements, and reduced pricing across many parts of our platform. These include:

  • New GPT-4 Turbo model that is more capable, cheaper and supports a 128K context window
  • New Assistants API that makes it easier for developers to build their own assistive AI apps that have goals and can call models and tools
  • New multimodal capabilities in the platform, including vision, image creation (DALL·E 3), and text-to-speech (TTS)


Introducing GPTs — from openai.com
You can now create custom versions of ChatGPT that combine instructions, extra knowledge, and any combination of skills.




OpenAI’s New Groundbreaking Update — from newsletter.thedailybite.co
Everything you need to know about OpenAI’s update, what people are building, and a prompt to skim long YouTube videos…

But among all this exciting news, the announcement of user-created “GPTs” took the cake.

That’s right, your very own personalized version of ChatGPT is coming, and it’s as groundbreaking as it sounds.

OpenAI’s groundbreaking announcement isn’t just a new feature – it’s a personal AI revolution. 

The upcoming customizable “GPTs” transform ChatGPT from a one-size-fits-all to a one-of-a-kind digital sidekick that is attuned to your life’s rhythm. 


Lore Issue #56: Biggest Week in AI This Year — from news.lore.com by Nathan Lands

First, Elon Musk announced “Grok,” a ChatGPT competitor inspired by “The Hitchhiker’s Guide to the Galaxy.” Surprisingly, in just a few months, xAI has managed to surpass the capabilities of GPT-3.5, signaling their impressive speed of execution and establishing them as a formidable long-term contender.

Then, OpenAI hosted their inaugural Dev Day, unveiling “GPT-4 Turbo,” which boasts a 128k context window, API costs slashed by threefold, text-to-speech capabilities, auto-model switching, agents, and even their version of an app store slated for launch next month.


The Day That Changed Everything — from joinsuperhuman.ai by Zain Kahn
ALSO: Everything you need to know about yesterday’s OpenAI announcements

  • OpenAI DevDay Part I: Custom ChatGPTs and the App Store of AI
  • OpenAI DevDay Part II: GPT-4 Turbo, Assistants, APIs, and more

OpenAI’s Big Reveal: Custom GPTs, GPT Store & More — from  news.theaiexchange.com
What you should know about the new announcements; how to get started with building custom GPTs


Incredible pace of OpenAI — from theaivalley.com by Barsee
PLUS: Elon’s Gork


 

 


Teaching writing in the age of AI — from the Future of Learning (a Hechinger Report newsletter) by Javeria Salman

ChatGPT can produce a perfectly serviceable writing “product,” she said. But writing isn’t a product per se — it’s a tool for thinking, for organizing ideas, she said.

“ChatGPT and other text-based tools can’t think for us,” she said. “There’s still things to learn when it comes to writing because writing is a form of figuring out what you think.”

When students could contrast their own writing to ChatGPT’s more generic version, Levine said, they were able to “understand what their own voice is and what it does.”




Grammarly’s new generative AI feature learns your style — and applies it to any text — from techcrunch.com by Kyle Wiggers; via Tom Barrett

But what about text? Should — and if so, how should — writers be recognized and remunerated for AI-generated works that mimic their voices?

Those are questions that are likely to be raised by a feature in Grammarly, the cloud-based typing assistant, that’s scheduled to launch by the end of the year for subscribers to Grammarly’s business tier. Called “Personalized voice detection and application,” the feature automatically detects a person’s unique writing style and creates a “voice profile” that can rewrite any text in the person’s style.


Is AI Quietly Weaving the Fabric of a Global Classroom Renaissance? — from medium.com by Robert the Robot
In a world constantly buzzing with innovation, a silent revolution is unfolding within the sanctuaries of learning—our classrooms.

From bustling metropolises to serene hamlets, schools across the globe are greeting a new companion—Artificial Intelligence (AI). This companion promises to redefine the essence of education, making learning a journey tailored to each child’s unique abilities.

The advent of AI in education is akin to a gentle breeze, subtly transforming the academic landscape. Picture a classroom where each child, with their distinct capabilities and pace, embarks on a personalized learning path. AI morphs this vision into reality, crafting a personalized educational landscape that celebrates the unique potential harbored within every learner.


AI Books for Educators — from aiadvisoryboards.wordpress.com by Barbara Anna Zielonka

Books have always held a special place in my heart. As an avid reader and AI enthusiast, I have curated a list of books on artificial intelligence specifically tailored for educators. These books delve into the realms of AI, exploring its applications, ethical considerations, and its impact on education. Share your suggestions and let me know which books you would like to see included on this list.


SAIL: ELAI recordings, AI Safety, Near term AI/learning — by George Siemens

We held our fourth online Empowering Learners for the Age of AI conference last week. We sold out at 1500 people (a Whova and budget limit). The recordings/playlist from the conference can now be accessed here.

 

60+ Ideas for ChatGPT Assignments — from stars.library.ucf.edu by Kevin Yee, Kirby Whittington, Erin Doggette, and Laurie Uttich

60+ ideas for using ChatGPT in your assignments today


Artificial intelligence is disrupting higher education — from itweb.co.za by Rennie Naidoo; via GSV
Traditional contact universities need to adapt faster and find creative ways of exploring and exploiting AI, or lose their dominant position.

Higher education professionals have a responsibility to shape AI as a force for good.


Introducing Canva’s biggest education launch — from canva.com
We’re thrilled to unveil our biggest education product launch ever. Today, we’re introducing a whole new suite of products that turn Canva into the all-in-one classroom tool educators have been waiting for.

Also see Canva for Education.
Create and personalize lesson plans, infographics,
posters, video, and more. 
100% free for
teachers and students at eligible schools.


ChatGPT and generative AI: 25 applications to support student engagement — from timeshighereducation.com by Seb Dianati and Suman Laudari
In the fourth part of their series looking at 100 ways to use ChatGPT in higher education, Seb Dianati and Suman Laudari share 25 prompts for the AI tool to boost student engagement


There are two ways to use ChatGPT — from theneurondaily.com

  1. Type to it.
  2. Talk to it (new).


Since then, we’ve looked to it for a variety of real-world business advice. For example, Prof Ethan Mollick posted a great guide using ChatGPT-4 with voice as a negotiation instructor.

In a similar fashion, you can consult ChatGPT with voice for feedback on:

  • Job interviews.
  • Team meetings.
  • Business presentations.



Via The Rundown: Google is using AI to analyze the company’s Maps data and suggest adjustments to traffic light timing — aiming to cut driver waits, stops, and emissions.


Google Pixel’s face-altering photo tool sparks AI manipulation debate — from bbc.com by Darren Waters

The camera never lies. Except, of course, it does – and seemingly more often with each passing day.
In the age of the smartphone, digital edits on the fly to improve photos have become commonplace, from boosting colours to tweaking light levels.

Now, a new breed of smartphone tools powered by artificial intelligence (AI) are adding to the debate about what it means to photograph reality.

Google’s latest smartphones released last week, the Pixel 8 and Pixel 8 Pro, go a step further than devices from other companies. They are using AI to help alter people’s expressions in photographs.



From Digital Native to AI-Empowered: Learning in the Age of Artificial Intelligence — from campustechnology.com by Kim Round
The upcoming generation of learners will enter higher education empowered by AI. How can institutions best serve these learners and prepare them for the workplace of the future?

Dr. Chris Dede, of Harvard University and Co-PI of the National AI Institute for Adult Learning and Online Education, spoke about the differences between knowledge and wisdom in AI-human interactions in a keynote address at the 2022 Empowering Learners for the Age of AI conference. He drew a parallel between Star Trek: The Next Generation characters Data and Picard during complex problem-solving: While Data offers the knowledge and information, Captain Picard offers the wisdom and context from on a leadership mantle, and determines its relevance, timing, and application.


The Near-term Impact of Generative AI on Education, in One Sentence — from opencontent.org by David Wiley

This “decreasing obstacles” framing turned out to be helpful in thinking about generative AI. When the time came, my answer to the panel question, “how would you summarize the impact generative AI is going to have on education?” was this:

“Generative AI greatly reduces the degree to which access to expertise is an obstacle to education.”

We haven’t even started to unpack the implications of this notion yet, but hopefully just naming it will give the conversation focus, give people something to disagree with, and help the conversation progress more quickly.


How to Make an AI-Generated Film — from heatherbcooper.substack.com by Heather Cooper
Plus, Midjourney finally has a new upscale tool!


Eureka! NVIDIA Research Breakthrough Puts New Spin on Robot Learning — from blogs.nvidia.com by Angie Lee
AI agent uses LLMs to automatically generate reward algorithms to train robots to accomplish complex tasks.

From DSC:
I’m not excited about this, as I can’t help but wonder…how long before the militaries of the world introduce this into their warfare schemes and strategies?


The 93 Questions Schools Should Ask About AI — from edweek.org by Alyson Klein

The toolkit recommends schools consider:

  • Purpose: How can AI help achieve educational goals?
  • Compliance: How does AI fit with existing policies?
  • Knowledge: How can schools advance AI Literacy?
  • Balance: What are the benefits and risks of AI?
  • Integrity: How does AI fit into policies on things like cheating?
  • Agency: How can humans stay in the loop on AI?
  • Evaluation: How can schools regularly assess the impact of AI?
 
 

The Game-Changer: How Legal Technology is Transforming the Legal Sector — from todaysconveyancer.co.uk by Perfect Portal

Rob Lawson, Strategic Sales Manager at Perfect Portal discussed why he thinks legal technology is so important:

“I spent almost 20 years in private practice and was often frustrated at the antiquated technology and processes that were deployed. It is one of the reasons that I love working in legal tech to provide solutions and streamline processes in the modern law firm. One of the major grumbles for practitioners is the amount of admin that they must do to fulfil the needs of their clients. Technology can automate routine tasks, streamline processes, and help manage large volumes of data more effectively. This then allows legal professionals to focus on more strategic aspects of their work. Ultimately this will increase efficiency and productivity.”



Some gen AI vendors say they’ll defend customers from IP lawsuits. Others, not so much. — from techcrunch.com by Kyle Wiggers

A person using generative AI — models that generate text, images, music and more given a prompt — could infringe on someone else’s copyright through no fault of their own. But who’s on the hook for the legal fees and damages if — or rather, when — that happens?

It depends.

In the fast-changing landscape of generative AI, companies monetizing the tech — from startups to big tech companies like Google, Amazon and Microsoft — are approaching IP risks from very different angles.


Clio Goes All Out with Major Product Announcements, Including A Personal Injury Add-On, E-Filing, and (Of Course) Generative AI — from lawnext.com by Bob Ambrogi

At its annual Clio Cloud Conference in Nashville today, the law practice management company Clio introduced an array of major new products and product updates, calling the series of announcements its most expansive product update ever in its 15-year history.


AI will invert the biglaw pyramid — from alexofftherecord.com by Cece Xie

These tasks that GPT can now handle are, coincidentally, common tasks for junior associates. From company and transaction summaries to legal research and drafting memos, analyzing and drafting have long been the purview of bright-eyed, bushy-tailed new law grads.

If we follow the capitalistic impulse of biglaw firms to its logical conclusion, this means that junior associates may soon face obsolescence. Why spend an hour figuring out how to explain an assignment to a first-year associate when you can just ask CoCounsel in five minutes? And the initial output will likely be better than a first-year’s initial work product, too.

Given the immense cost-savings that legal GPT products can confer, I suspect the rise of AI in legal tech will coincide with smaller junior associate classes. Gone are the days of 50+ junior lawyers all working on the same document review or due diligence. Instead, a fraction of those junior lawyers will be hired to oversee and QC the AI’s outputs. Junior associates will edit more than they do currently and manage more than they do right now. Juniors will effectively be more like midlevels from the get-go.


Beyond Law Firms: How Legal Tech’s Real Frontier Lies With SMBs (small and medium-sized businesses) — from forbes.com by Charles Brecque

Data and artificial intelligence are transforming the legal technology space—there’s no doubt about it. A recent Thomson Reuters Institute survey of lawyers showed that a large majority (82%) of respondents believe ChatGPT and generative AI can be readily applied to legal work.

While it’s tempting to think of legal tech as a playground exclusive to law firms, as technology enables employees without legal training to use and create legal frameworks and documentation, I’d like to challenge that narrative. Being the founder of a company that uses AI to manage contracts, the way I see it is the real magic happens when legal tech tools meet the day-to-day challenges of small and medium-sized businesses (known as “SMBs”).

 

WHAT WAS GARY MARCUS THINKING, IN THAT INTERVIEW WITH GEOFF HINTON? — from linkedin.com by Stephen Downes

Background (emphasis DSC): 60 Minutes did an interview with ‘the Godfather of AI’, Geoffrey Hinton. In response, Gary Marcus wrote a column in which he inserted his own set of responses into the transcript, as though he were a panel participant. Neat idea. So, of course, I’m stealing it, and in what follows, I insert my own comments as I join the 60 Minutes panel with Geoffrey Hinton and Gary Marcus.

Usually I put everyone else’s text in italics, but for this post I’ll put it all in normal font, to keep the format consistent.

Godfather of Artificial Intelligence Geoffrey Hinton on the promise, risks of advanced AI


OpenAI’s Revenue Skyrockets to $1.3 Billion Annualized Rate — from maginative.com by Chris McKay
This means the company is generating over $100 million per month—a 30% increase from just this past summer.

OpenAI, the company behind the viral conversational AI ChatGPT, is experiencing explosive revenue growth. The Information reports that CEO Sam Altman told the staff this week that OpenAI’s revenue is now crossing $1.3 billion on an annualized basis. This means the company is generating over $100 million per month—a 30% increase from just this past summer.

Since the launch of a paid version of ChatGPT in February, OpenAI’s financial growth has been nothing short of meteoric. Additionally, in August, the company announced the launch of ChatGPT Enterprise, a commercial version of its popular conversational AI chatbot aimed at business users.

For comparison, OpenAI’s total revenue for all of 2022 was just $28 million. The launch of ChatGPT has turbocharged OpenAI’s business, positioning it as a bellwether for demand for generative AI.



From 10/13:


New ways to get inspired with generative AI in Search — from blog.google
We’re testing new ways to get more done right from Search, like the ability to generate imagery with AI or creating the first draft of something you need to write.

 

The Misunderstanding About Education That Cost Mark Zuckerberg $100 Million — from danmeyer.substack.com by Dan Meyer
Personalized learning can feel isolating. Whole class learning can feel personal. This is hard to understand.

Excerpt (emphasis DSC):

Last week, Matt Barnum reported in Chalkbeat that the Chan Zuckerberg Initiative is laying off dozens of staff members and pivoting away from the personalized learning platform they have funded since 2015 with somewhere near $100M.

I have tried to illustrate as often as my subscribers will tolerate that students don’t particularly enjoy learning alone with laptops within social spaces like classrooms. That learning fails to answer their questions about their social identity. It contributes to their feelings of alienation and disbelonging. I find this case easy to make but hard to prove. Maybe we just haven’t done personalized learning right? Maybe Summit just needed to include generative AI chatbots in their platform?

What is far easier to prove, or rather to disprove, is the idea that “whole class instruction must feel impersonal to students,” that “whole class instruction must necessarily fail to meet the needs of individual students.”

From DSC:
I appreciate Dan’s comments here (as highlighted above) as they are helpful in my thoughts regarding the Learning from the Living [Class] Room vision. They seem to be echoed here by Jeppe Klitgaard Stricker when he says:

Personalized learning paths can be great, but they also entail a potential abolishment or unintended dissolution of learning communities and belonging.

Perhaps this powerful, global, Artificial Intelligence (AI)-backed, next-generation, lifelong learning platform of the future will be more focused on postsecondary students and experiences — but not so much for the K12 learning ecosystem.

But the school systems I’ve seen here in Michigan (USA) represent systems that address a majority of the class only. These one-size-fits-all systems don’t work for many students who need extra help and/or who are gifted students. The trains move fast. Good luck if you can’t keep up with the pace.

But if K-12’ers are involved in a future learning platform, the platform needs to address what Dan’s saying. It must address students questions about their social identity and not contribute to their feelings of alienation and disbelonging. It needs to support communities of practice and learning communities.

 

Thinking with Colleagues: AI in Education — from campustechnology.com by Mary Grush
A Q&A with Ellen Wagner

Wagner herself recently relied on the power of collegial conversations to probe the question: What’s on the minds of educators as they make ready for the growing influence of AI in higher education? CT asked her for some takeaways from the process.

We are in the very early days of seeing how AI is going to affect education. Some of us are going to need to stay focused on the basic research to test hypotheses. Others are going to dive into laboratory “sandboxes” to see if we can build some new applications and tools for ourselves. Still others will continue to scan newsletters like ProductHunt every day to see what kinds of things people are working on. It’s going to be hard to keep up, to filter out the noise on our own. That’s one reason why thinking with colleagues is so very important.

Mary and Ellen linked to “What Is Top of Mind for Higher Education Leaders about AI?” — from northcoasteduvisory.com. Below are some excerpts from those notes:

We are interested how K-12 education will change in terms of foundational learning. With in-class, active learning designs, will younger students do a lot more intensive building of foundational writing and critical thinking skills before they get to college?

  1. The Human in the Loop: AI is built using math: think of applied statistics on steroids. Humans will be needed more than ever to manage, review and evaluate the validity and reliability of results. Curation will be essential.
  2. We will need to generate ideas about how to address AI factors such as privacy, equity, bias, copyright, intellectual property, accessibility, and scalability.
  3. Have other institutions experimented with AI detection and/or have held off on emerging tools related to this? We have just recently adjusted guidance and paused some tools related to this given the massive inaccuracies in detection (and related downstream issues in faculty-elevated conduct cases)

Even though we learn repeatedly that innovation has a lot to do with effective project management and a solid message that helps people understand what they can do to implement change, people really need innovation to be more exciting and visionary than that.  This is the place where we all need to help each other stay the course of change. 


Along these lines, also see:


What people ask me most. Also, some answers. — from oneusefulthing.org by Ethan Mollick
A FAQ of sorts

I have been talking to a lot of people about Generative AI, from teachers to business executives to artists to people actually building LLMs. In these conversations, a few key questions and themes keep coming up over and over again. Many of those questions are more informed by viral news articles about AI than about the real thing, so I thought I would try to answer a few of the most common, to the best of my ability.

I can’t blame people for asking because, for whatever reason, the companies actually building and releasing Large Language Models often seem allergic to providing any sort of documentation or tutorial besides technical notes. I was given much better documentation for the generic garden hose I bought on Amazon than for the immensely powerful AI tools being released by the world’s largest companies. So, it is no surprise that rumor has been the way that people learn about AI capabilities.

Currently, there are only really three AIs to consider: (1) OpenAI’s GPT-4 (which you can get access to with a Plus subscription or via Microsoft Bing in creative mode, for free), (2) Google’s Bard (free), or (3) Anthropic’s Claude 2 (free, but paid mode gets you faster access). As of today, GPT-4 is the clear leader, Claude 2 is second best (but can handle longer documents), and Google trails, but that will likely change very soon when Google updates its model, which is rumored to be happening in the near future.

 

Mark Zuckerberg: First Interview in the Metaverse | Lex Fridman Podcast #398


Photo-realistic avatars show future of Metaverse communication — from inavateonthenet.net

Mark Zuckerberg, CEO, Meta, took part in the first-ever Metaverse interview using photo-realistic virtual avatars, demonstrating the Metaverse’s capability for virtual communication.

Zuckerberg appeared on the Lex Fridman podcast, using scans of both Fridman and Zuckerberg to create realistic avatars instead of using a live video feed. A computer model of the avatar’s faces and bodies are put into a Codec, using a headset to send an encoded version of the avatar.

The interview explored the future of AI in the metaverse, as well as the Quest 3 headset and the future of humanity.


 

180 Degree Turn: NYC District Goes From Banning ChatGPT to Exploring AI’s Potential — from edweek.org by Alyson Klein (behind paywall)

New York City Public Schools will launch an Artificial Intelligence Policy Lab to guide the nation’s largest school district’s approach to this rapidly evolving technology.


The Leader’s Blindspot: How to Prepare for the Real Future — from preview.mailerlite.io by the AIEducator
The Commonly Held Belief: AI Will Automate Only Boring, Repetitive Tasks First

The Days of Task-Based Views on AI Are Numbered
The winds of change are sweeping across the educational landscape (emphasis DSC):

  1. Multifaceted AI: AI technologies are not one-trick ponies; they are evolving into complex systems that can handle a variety of tasks.
  2. Rising Expectations: As technology becomes integral to our lives, the expectations for personalised, efficient education are soaring.
  3. Skill Transformation: Future job markets will demand a different skill set, one that is symbiotic with AI capabilities.

Teaching: How to help students better understand generative AI — from chronicle.com by Beth McMurtrie
Beth describes ways professors have used ChatGPT to bolster critical thinking in writing-intensive courses

Kevin McCullen, an associate professor of computer science at the State University of New York at Plattsburgh, teaches a freshman seminar about AI and robotics. As part of the course, students read Machines of Loving Grace: The Quest for Common Ground Between Humans and Robots, by John Markoff. McCullen had the students work in groups to outline and summarize the first three chapters. Then he showed them what ChatGPT had produced in an outline.

“Their version and ChatGPT’s version seemed to be from two different books,” McCullen wrote. “ChatGPT’s version was essentially a ‘laundry list’ of events. Their version was narratives of what they found interesting. The students had focused on what the story was telling them, while ChatGPT focused on who did what in what year.” The chatbot also introduced false information, such as wrong chapter names.

The students, he wrote, found the writing “soulless.”


7 Questions with Dr. Cristi Ford, VP of Academic Affairs at D2L — from campustechnology.com by Rhea Kelly

In the Wild West of generative AI, educators and institutions are working out how best to use the technology for learning. How can institutions define AI guidelines that allow for experimentation while providing students with consistent guidance on appropriate use of AI tools?

To find out, we spoke with Dr. Cristi Ford, vice president of academic affairs at D2L. With more than two decades of educational experience in nonprofit, higher education, and K-12 institutions, Ford works with D2L’s institutional partners to elevate best practices in teaching, learning, and student support. Here, she shares her advice on setting and communicating AI policies that are consistent and future-ready.


AI Platform Built by Teachers, for Teachers, Class Companion Raises $4 Million to Tap Into the Power of Practice — from prweb.com

“If we want to use AI to improve education, we need more teachers at the table,” said Avery Pan, Class Companion co-founder and CEO. “Class Companion is designed by teachers, for teachers, to harness the most sophisticated AI and improve their classroom experience. Developing technologies specifically for teachers is imperative to supporting our next generation of students and education system.”


7 Questions on Generative AI in Learning Design — from campustechnology.com by Rhea Kelly
Open LMS Adoption and Education Specialist Michael Vaughn on the challenges and possibilities of using artificial intelligence to move teaching and learning forward.

The potential for artificial intelligence tools to speed up course design could be an attractive prospect for overworked faculty and spread-thin instructional designers. Generative AI can shine, for example, in tasks such as reworking assessment question sets, writing course outlines and learning objectives, and generating subtitles for audio and video clips. The key, says Michael Vaughn, adoption and education specialist at learning platform Open LMS, is treating AI like an intern who can be guided and molded along the way, and whose work is then vetted by a human expert.

We spoke with Vaughn about how best to utilize generative AI in learning design, ethical issues to consider, and how to formulate an institution-wide policy that can guide AI use today and in the future.


10 Ways Technology Leaders Can Step Up and Into the Generative AI Discussion in Higher Ed — from er.educause.edu by Lance Eaton and Stan Waddell

  1. Offer Short Primers on Generative AI
  2. Explain How to Get Started
  3. Suggest Best Practices for Engaging with Generative AI
  4. Give Recommendations for Different Groups
  5. Recommend Tools
  6. Explain the Closed vs. Open-Source Divide
  7. Avoid Pitfalls
  8. Conduct Workshops and Events
  9. Spot the Fake
  10. Provide Proper Guidance on the Limitations of AI Detectors


 

Canva’s new AI tools automate boring, labor-intensive design tasks — from theverge.com by Jess Weatherbed
Magic Studio features like Magic Switch automatically convert your designs into blogs, social media posts, emails, and more to save time on manually editing documents.


Canva launches Magic Studio, partners with Runway ML for video — from bensbites.beehiiv.com by Ben Tossell

Here are the highlights of launched features under the new Magic Studio:

  • Magic Design – Turn ideas into designs instantly with AI-generated templates.
  • Magic Switch – Transform content into different formats and languages with one click.
  • Magic Grab – Make images editable like Canva templates for easy editing.
  • Magic Expand – Use AI to expand images beyond the original frame.
  • Magic Morph – Transform text and shapes with creative effects and prompts.
  • Magic Edit – Make complex image edits using simple text prompts.
  • Magic Media – Generate professional photos, videos and artworks from text prompts.
  • Magic Animate – Add animated transitions and motion to designs instantly.
  • Magic Write – Generate draft text and summaries powered by AI.



Adobe Firefly

Meet Adobe Firefly -- Adobe is going hard with the use of AI. This is a key product along those lines.


Addendums on 10/11/23:


Adobe Releases New AI Models Aimed at Improved Graphic Design — from bloomberg.com
New version of Firefly is bigger than initial tool, Adobe says Illustrator, Express programs each get own generative tools


 

Deepfakes: An evidentiary tsunami! — fromthebrainyacts.beehiiv.com by Josh Kubicki

Excerpt: (emphasis DSC):

I’ve written and spoken about this before but the rise of deepfakes is going to have a profound impact on courts throughout the world. This week we saw three major deepfake stories.

Whether you are a lawyer or not, this topic will impact you. So, please consider these questions as we will need to have answers for each one very soon (if not now).

  1. How will we establish a reliable and consistent standard to authenticate digital evidence as genuine and not altered by deepfake technology?
  2. Will the introduction of deepfakes shift the traditional burdens of proof or production, especially when digital evidence is introduced?
  3. Will courts require expert witnesses for digital evidence authentication in every case, and what standards will be used to qualify these experts?
  4. Are there existing technological tools or methods to detect deepfakes? (yes there is but it is not 100%) How can courts keep abreast of rapidly advancing technology?
  5. …plus several more questions

From DSC:
What are law schools doing about this? Are they addressing this?


And speaking of legal matters and law schools, this might be interesting or helpful to someone out there:

 

As AI Chatbots Rise, More Educators Look to Oral Exams — With High-Tech Twist — from edsurge.com by Jeffrey R. Young

To use Sherpa, an instructor first uploads the reading they’ve assigned, or they can have the student upload a paper they’ve written. Then the tool asks a series of questions about the text (either questions input by the instructor or generated by the AI) to test the student’s grasp of key concepts. The software gives the instructor the choice of whether they want the tool to record audio and video of the conversation, or just audio.

The tool then uses AI to transcribe the audio from each student’s recording and flags areas where the student answer seemed off point. Teachers can review the recording or transcript of the conversation and look at what Sherpa flagged as trouble to evaluate the student’s response.

 

Reimagining Hiring and Learning with the Power of AI — from linkedin.com by Hari Srinivasan

That’s why today we’re piloting new tools like our new release of Recruiter 2024 and LinkedIn Learning’s AI-powered coaching experience to help with some of the heavy lifting so HR professionals can focus on what matters most.

“AI is quickly transforming recruitment, training, and many other HR practices,” says Josh Bersin, industry analyst and CEO of The Josh Bersin Company. “LinkedIn’s new features in Recruiter 2024 and LinkedIn Learning can massively improve recruiter productivity and help all employees build the skills they need to grow in their careers.”

By pairing generative AI with our unique insights gained from the more than 950 million professionals, 65 million companies, and 40,000 skills on our platform, we’ve reimagined our Recruiter product to help our customers find that short list of qualified candidates — faster.

From DSC:
While I’m very interested to see how Microsoft’s AI-powered LinkedIn Learning coach will impact peoples’ growth/development, I need to admit that I still approach AI and hiring/finding talent with caution. I’m sure I was weeded out by several Applicant Tracking Systems (ATS) back in 2017 when I was looking for my next position — and I only applied to positions that I had the qualifications for. And if you’ve tried to get a job recently, I bet you were weeded out by an ATS as well. So while this might help recruiters, the jury is still out for me as to whether these developments are good or bad for the rest of society.

Traditional institutions of higher education may want to research these developments to see which SKILLS are in demand.

Also relevant/see:

LinkedIn Launches Exciting Gen AI Features in Recruiter and Learning — from joshbersin.com by Josh Bersin

This week LinkedIn announced some massive Gen AI features in its two flagship products: LinkedIn Recruiter and LinkedIn Learning. Let me give you an overview.

LinkedIn goes big on new AI tools for learning, recruitment, marketing and sales, powered by OpenAI — from techcrunch.com by Ingrid Lunden

LinkedIn Learning will be incorporating AI in the form of a “learning coach” that is essentially built as a chatbot. Initially the advice that it will give will be trained on suggestions and tips, and it will be firmly in the camp of soft skills. One example: “How can I delegate tasks and responsibility effectively?”

The coach might suggest actual courses, but more importantly, it will actually also provide information, and advice, to users. LinkedIn itself has a giant catalogue of learning videos, covering both those soft skills but also actual technical skills and other knowledge needed for specific jobs. It will be interesting to see if LinkedIn extends the coach to covering that material, too.

 

 



Adobe video-AI announcements for IBC — from provideocoalition.com by Rich Young

For the IBC 2023 conference, Adobe announced new AI and 3D features to Creative Cloud video tools, including Premiere Pro Enhance Speech for faster dialog cleanup, and filler word detection and removal in Text-Based Editing. There’s also new AI-based rotoscoping and a true 3D workspace in the After Effects beta, as well as new camera-to-cloud integrations and advanced storage options in Frame.io.

Though not really about AI, you might also be interested in this posting:


Airt AI Art Generator (Review) — from hongkiat.com
Turn your creative ideas into masterpieces using Airt’s AI iPad app.

The Airt AI Generator app makes it easy to create art on your iPad. You can pick an art style and a model to make your artwork. It’s simple enough for anyone to use, but it doesn’t have many options for customizing your art.

Even with these limitations, it’s a good starting point for people who want to try making art with AI. Here are the good and bad points we found.

Pros:

  • User-Friendly: The app is simple and easy to use, making it accessible for users of all skill levels.

Cons:

  • Limited Advanced Features: The app lacks options for customization, such as altering image ratios, seeds, and other settings.

 
© 2024 | Daniel Christian