But a few applications of machine learning stood out as genuinely helpful or surprising — here are a few examples of AI that might actually do some good.
The whole idea that AI might not be a total red flag occurred to me when I chatted with Whispp at a press event. This small team is working on voicing the voiceless, meaning people who have trouble speaking normally due to a condition or illness.
The Emerging Role of Smart TVs Boothe and Comiskey claim that our TVs will become even smarter and better connected to the web and the internet. Our TVs will become an intelligent center for a variety of applications powered through our smartphone. TVs will be able to direct things like appliances and security cameras. Perhaps even more importantly, our TVs can become e-commerce centers, allowing us to speak with them and conduct business.
…
This increased TV capability means that the TV could become a more dominant mode of working and computing for lawyers. As TVs become more integrated with the internet and capable of functioning as communication hubs, they could potentially replace traditional computing devices in legal settings. With features like voice control and pattern recognition, TVs could serve as efficient tools for such things as document preparation and client meetings.
From DSC: Now imagine the power of voice-enabled chatbots and the like. We could be videoconferencing (or holograming) with clients, and be able to access information at the same time. Language translation — like that in the Timekettle product — will be built in.
As we step into 2024, the legal industry continues to be reshaped by technological advancements. This year promises to bring new developments that could revolutionize how legal professionals work and interact with clients. Here are key predictions for legal tech in 2024:
Last Friday was Legaltech Week’s year-end show, in which our panel of journalists and bloggers picked the year’s top stories in legal tech and innovation.
So what were the top stories? Well, if you missed it, no worries. Here’s the video:
During lockdown, we were forced to start producing videos for Moocs remotely. This was quite successful, and we continue to use these techniques whenever a contributor can’t come to the studio. The same principles can be used by anyone who doesn’t have access to a media production team. Here are our tips on producing educational videos on a budget.
Programmable medicines. AI tools for kids. We asked
over 40 partners across a16z to preview one big idea
they believe will drive innovation in 2024.
Narrowly Tailored, Purpose-Built AI In 2024, I predict we’ll see narrower AI solutions. While ChatGPT may be a great general AI assistant, it’s unlikely to “win” for every task. I expect we’ll see an AI platform purpose-built for researchers, a writing generation tool targeted for journalists, and a rendering platform specifically for designers, to give just a few examples.
Over the longer term, I think the products people use on an everyday basis will be tailored to their use cases — whether this is a proprietary underlying model or a special workflow built around it. These companies will have the chance to “own” the data and workflow for a new era of technology; they’ll do this by nailing one category, then expanding. For the initial product, the narrower the better.
— via Olivia Moore, who focuses on marketplace startups
So asks a recent study by two academics from Stanford Law School, David Freeman Engstrom and Nora Freeman Engstrom, on the potential impact of AI on the civil litigation landscape in the US.
It is against this landscape, the study observes, that champions of legal tech have suggested that there is an opportunity for legal tech to “democratise” litigation and put litigation’s “haves” and “have nots” on a more equal footing, by arming smaller firms and sole practitioners with the tools necessary to do battle against their better resourced opponents, and cutting the cost of legal services, putting lawyers within reach of a wider swathe of people.
But is this a real opportunity, and will AI be key to its realisation?
…
However, while AI may reduce the justice gap between the “haves” and “have-nots” of litigation, it could also exacerbate existing inequalities.
From DSC: While this article approaches things from the lawyer’s viewpoint, I’d like to see this question and the use of AI from the common man’s/woman’s viewpoint. Why? In order to provide FAR GREATER access to justice (#A2J) for those who can’t afford a lawyer as they head into the civil law courtrooms.
Should I take my case to court? Do I have a chance to win this case? If so, how?
What forms do I need to complete if I’m going to go to court?
When and how do I address the judge?
What does my landlord have to do?
How do I prevent myself from falling into a debt-collection mess and/or what options do I have to get out of this mess?
Are there any lawyers in my area who would take my case on a pro bono basis?
…and judges and lawyers — as well as former litigants — could add many more questions (and answers) to this list
Bottom line: It is my hope that technology can help increase access to justice.
A number of products are already under development, or have been launched. One example is a project that Norton Rose Fulbright is working on, together with not-for-profit legal service Justice Connect. The scope is to develop an automated natural language processing AI model that seeks to interpret the ‘everyday’ language used by clients in order to identify the client’s legal issues and correctly diagnose their legal problem. This tool is aimed at addressing the struggles that individuals often face in deciphering legal jargon and understanding the nature of their legal issue and the type of lawyer, or legal support, they need to resolve that problem.
From DSC: As I’ve long stated on the Learning from the Living [Class]Room vision, we are heading toward a new AI-empowered learning platform — where humans play a critically important role in making this new learning ecosystem work.
Along these lines, I ran into this site out on X/Twitter. We’ll see how this unfolds, but it will be an interesting space to watch.
From DSC: This future learning platform will also focus on developing skills and competencies. Along those lines, see:
Scale for Skills-First— from the-job.beehiiv.com by Paul Fain An ed-tech giant’s ambitious moves into digital credentialing and learner records.
A Digital Canvas for Skills
Instructure was a player in the skills and credentials space before its recent acquisition of Parchment, a digital transcript company. But that $800M move made many observers wonder if Instructure can develop digital records of skills that learners, colleges, and employers might actually use broadly.
…
Ultimately, he says, the CLR approach will allow students to bring these various learning types into a coherent format for employers.
Instructure seeks a leadership role in working with other organizations to establish common standards for credentials and learner records, to help create consistency. The company collaborates closely with 1EdTech. And last month it helped launch the 1EdTech TrustEd Microcredential Coalition, which aims to increase quality and trust in digital credentials.
It’s an era many instructors would like to put behind them: black boxes on Zoom screens, muffled discussions behind masks, students struggling to stay engaged. But how much more challenging would teaching during the pandemic have been if colleges did not have experts on staff to help with the transition? On many campuses, teaching-center directors, instructional designers, educational technologists, and others worked alongside professors to explore learning-management systems, master video technology, and rethink what and how they teach.
A new book out this month, Higher Education Beyond Covid: New Teaching Paradigms and Promise, explores this period through the stories of campus teaching and learning centers. Their experiences reflect successes and failures, and what higher education could learn as it plans for the future.
As usual, our readers were full of suggestions. Kathryn Schild, the lead instructional designer in faculty development and instructional support at the University of Alaska at Anchorage, shared a guide she’s compiled on holding asynchronous discussions, which includes a section on difficult topics.
In an email, Schild also pulled out a few ideas she thought were particularly relevant to Le’s question, including:
Set the ground rules as a class. One way to do this is to share your draft rules in a collaborative document and ask students to annotate it and add suggestions.
Plan to hold fewer difficult discussions than in a face-to-face class, and work on quality over quantity. This could include multiweek discussions, where you spiral through the same issue with fresh perspectives as the class learns new approaches.
Start with relationship-building interactions in the first few weeks, such as introductions, low-stakes group assignments, or peer feedback, etc.
Forty-one states and the District of Columbia are suing Meta, alleging that the tech giant harms children by building addictive features into Instagram and Facebook. Tuesday’s legal actions represent the most significant effort by state enforcers to rein in the impact of social media on children’s mental health.
Mark Zuckerberg, CEO, Meta, took part in the first-ever Metaverse interview using photo-realistic virtual avatars, demonstrating the Metaverse’s capability for virtual communication.
Zuckerberg appeared on the Lex Fridman podcast, using scans of both Fridman and Zuckerberg to create realistic avatars instead of using a live video feed. A computer model of the avatar’s faces and bodies are put into a Codec, using a headset to send an encoded version of the avatar.
The interview explored the future of AI in the metaverse, as well as the Quest 3 headset and the future of humanity.
Canva’s new AI tools automate boring, labor-intensive design tasks — from theverge.com by Jess Weatherbed Magic Studio features like Magic Switch automatically convert your designs into blogs, social media posts, emails, and more to save time on manually editing documents.
Google DeepMind cofounder Mustafa Suleyman thinks that generative is a passing phase, and that interactive AI is the next big thing in AI. Suleyman called the transformation “a profound moment” in the history of technology.
Suleyman divided AI’s evolution into 3 waves:
Classification: Training computers to classify various types of data like images and text.
Generative: The current wave, which takes input data to generate new data. ChatGPT is the best example of this.
Interactive: The next wave, where an AI will be capable of communicating and operating autonomously.
“Think of it as autonomous software that can talk to other apps to get things done.”
From DSC: Though I find this a generally positive thing, the above sentence makes me exclaim, “No, nothing could possibly go wrong there.”
During this special keynote presentation, Western Michigan University (WMU) professor Sue Ellen Christian speaks about the importance of media literacy for all ages and how we can help educate our friends and families about media literacy principles. Hosted by the Grand Rapids Public Library and GRTV, a program of the Grand Rapids Community Media Center. Special thanks to the Grand Rapids Public Library Foundation for their support of this program.
Inspired by my recent Review: Shure MV7 dynamic hybrid studio microphone – near, far and beyond, Beaker Films of Fairfield, Connecticut, US has developed and deployed a first batch of 10 kits to capture remote conversations from different locations worldwide. Beaker Films is frequently contracted to record remote interviews or testimonials from medical professionals. For this project, Beaker Films’ clients wanted consistent, high quality audio and video, but with 3 additional challenges: they preferred to have no visible microphone in the shot, they needed a teleprompter function and the whole kit needed to be as simple as possible for non-technical guests.
West Suffolk College in the UK has opened its Extended Reality Lab (XR Lab), the facilities comprise of four distinct areas: an Immersion Lab, a Collaboration Theatre, a Green Room, and a Conference Room. The project was designed by architects WindsorPatania for Eastern Colleges Group.
Systems integrator CJP Broadcast Service Solutions, has won a tender to build a virtual production environment for Solent University in the UK.
The new facilities, converted from an existing studio space, will provide students on the film production courses with outstanding opportunities to develop their creative output.
— Daniel Christian (he/him/his) (@dchristian5) June 23, 2023
.
On giving AI eyes and ears— from oneusefulthing.org by Ethan Mollick AI can listen and see, with bigger implications than we might realize.
Excerpt:
But even this is just the beginning, and new modes of using AI are appearing, which further increases their capabilities. I want to show you some examples of this emerging world, which I think will soon introduce a new wave of AI use cases, and accompanying disruption.
We need to recognize that these capabilities will continue to grow, and AI will be able to play a more active role in the real world by observing and listening. The implications are likely to be profound, and we should start thinking through both the huge benefits and major concerns today.
Even though generative AI is a new thing, it doesn’t change why students cheat. They’ve always cheated for the same reason: They don’t find the work meaningful, and they don’t think they can achieve it to their satisfaction. So we need to design assessments that students find meaning in.
Tricia Bertram Gallant
Caught off guard by AI— from chonicle.com by Beth McMurtrie and Beckie Supiano Professor scrambled to react to ChatGPT this spring — and started planning for the fall
Excerpt:
Is it cheating to use AI to brainstorm, or should that distinction be reserved for writing that you pretend is yours? Should AI be banned from the classroom, or is that irresponsible, given how quickly it is seeping into everyday life? Should a student caught cheating with AI be punished because they passed work off as their own, or given a second chance, especially if different professors have different rules and students aren’t always sure what use is appropriate?
…OpenAI built tool use right into the GPT API with an update called function calling. It’s a little like a child’s ability to ask their parents to help them with a task that they know they can’t do on their own. Except in this case, instead of parents, GPT can call out to external code, databases, or other APIs when it needs to.
Each function in function calling represents a tool that a GPT model can use when necessary, and GPT gets to decide which ones it wants to use and when. This instantly upgrades GPT capabilities—not because it can now do every task perfectly—but because it now knows how to ask for what it wants and get it. .
.
How ChatGPT can help disrupt assessment overload— from timeshighereducation.com by David Carless Advances in AI are not necessarily the enemy – in fact, they should prompt long overdue consideration of assessment types and frequency, says David Carless
Excerpt:
Reducing the assessment burden could support trust in students as individuals wanting to produce worthwhile, original work. Indeed, students can be co-opted as partners in designing their own assessment tasks, so they can produce something meaningful to them.
A strategic reduction in quantity of assessment would also facilitate a refocusing of assessment priorities on deep understanding more than just performance and carries potential to enhance feedback processes.
If we were to tackle assessment overload in these ways, it opens up various possibilities. Most significantly there is potential to revitalise feedback so that it becomes a core part of a learning cycle rather than an adjunct at its end. End-of-semester, product-oriented feedback, which comes after grades have already been awarded, fails to encourage the iterative loops and spirals typical of productive learning. .
Since AI in education has been moving at the speed of light, we built this AI Tools in Education database to keep track of the most recent AI tools in education and the changes that are happening every day.This database is intended to be a community resource for educators, researchers, students, and other edtech specialists looking to stay up to date. This is a living document, so be sure to come back for regular updates.
These claims conjure up the rosiest of images: human resource departments and their robot buddies solving discrimination in workplace hiring. It seems plausible, in theory, that AI could root out unconscious bias, but a growing body of research shows the opposite may be more likely.
…
Companies’ use of AI didn’t come out of nowhere: For example, automated applicant tracking systems have been used in hiring for decades. That means if you’ve applied for a job, your resume and cover letter were likely scanned by an automated system. You probably heard from a chatbot at some point in the process. Your interview might have been automatically scheduled and later even assessed by AI.
From DSC:
Here was my reflection on this:
DC: Along these lines, I wonder if Applicant Tracking Systems cause us to become like typecast actors and actresses — only thought of for certain roles. Pigeonholed.
— Daniel Christian (he/him/his) (@dchristian5) June 23, 2023
In June, ResumeBuilder.com surveyed more than 1,000 employees who are involved in hiring processes at their workplaces to find out about their companies’ use of AI interviews.
The results:
43% of companies already have or plan to adopt AI interviews by 2024
Two-thirds of this group believe AI interviews will increase hiring efficiency
15% say that AI will be used to make decisions on candidates without any human input
More than half believe AI will eventually replace human hiring managers
Watch OpenAI CEO Sam Altman on the Future of AI — from bloomberg.com Sam Altman, CEO & Co-Founder, OpenAI discusses the explosive rise of OpenAI and its products and what an AI-laced future can look like with Bloomberg’s Emily Chang at the Bloomberg Technology Summit.
The implementation of generative AI within these products will dramatically improve educators’ ability to deliver personalized learning to students at scale by enabling the application of personalized assessments and learning pathways based on individual student needs and learning goals. K-12 educators will also benefit from access to OpenAI technology…