San Diego’s Nanome Inc. releases collaborative VR-STEM software for free — from vrscout.com by Becca Loux

Excerpt:

The first collaborative VR molecular modeling application was released August 29 to encourage hands-on chemistry experimentation.

The open-source tool is free for download now on Oculus and Steam.

Nanome Inc., the San Diego-based start-up that built the intuitive application, comprises UCSD professors and researchers, web developers and top-level pharmaceutical executives.

 

“With our tool, anyone can reach out and experience science at the nanoscale as if it is right in front of them. At Nanome, we are bringing the craftsmanship and natural intuition from interacting with these nanoscale structures at room scale to everyone,” McCloskey said.

 

San Diego’s Nanome Inc. Releases Collaborative VR-STEM Software For Free

 

 

10 ways VR will change life in the near future — from forbes.com

Excerpts:

  1. Virtual shops
  2. Real estate
  3. Dangerous jobs
  4. Health care industry
  5. Training to create VR content
  6. Education
  7. Emergency response
  8. Distraction simulation
  9. New hire training
  10. Exercise

 

From DSC:
While VR will have its place — especially for timeswhen you need to completely immerse yourself into another environment — I think AR and MR will be much larger and have a greater variety of applications. For example, I could see where instructions on how to put something together in the future could use AR and/or MR to assist with that process. The system could highlight the next part that I’m looking for and then highlight the corresponding parts where it goes — and, if requested, can show me a clip on how it fits into what I’m trying to put together.

 

How MR turns firstline workers into change agents — from virtualrealitypop.com by Charlie Finkand
Mixed Reality, a new dimension of work — from Microsoft and Harvard Business Review

Excerpts:

Workers with mixed-reality solutions that enable remote assistance, spatial planning, environmentally contextual data, and much more,” Bardeen told me. With the HoloLens Firstline Workers workers conduct their usual, day-to-day activities with the added benefit of a heads-up, hands-free, display that gives them immediate access to valuable, contextual information. Microsoft says speech services like Cortana will be critical to control along with gesture, according to the unique needs of each situation.

 

Expect new worker roles. What constitutes an “information worker” could change because mixed reality will allow everyone to be involved in the collection and use of information. Many more types of information will become available to any worker in a compelling, easy-to-understand way. 

 

 

Let’s Speak: VR language meetups — from account.altvr.com

 

 

 

 

Google’s VR Labs provide STEM students with hands-on experience — from vrscout.com by Kyle Melnick

Excerpt:

STEM students engaged in scientific disciplines, such as biochemistry and neuroscience, are often required by their respective degrees to spend a certain amount of time engaged in an official laboratory environment. Unfortunately, crowded universities and the rise of online education have made it difficult for these innovators-in-training to access properly equipped labs and log their necessary hours.

Cue Google VR Labs, a series of comprehensive virtual lab experiences available on the Google Daydream platform. Developed as part of partnership between Google and simulation education company Labster, the in-depth program boasts 30 interactive lab experiences in which biology students can engage in a series of hands-on scientific activities in a realistic environment.

These actions can include everything from the use of practical tools, such as DNA sequencers and microscopes, to reality-bending experiences only capable in a virtual environment, like traveling to the surface of the newly discovered Astakos IV exoplanet or examining and altering DNA on a molecular level.

 

Google’s VR Labs Provide STEM Students With Hands-On Experience

 

Also see:

 

 

 

Adobe Announces the 2019 Release of Adobe Captivate, Introducing Virtual Reality for eLearning Design — from theblog.adobe.com

Excerpt:

  • Immersive learning with VR experiences: Design learning scenarios that your learners can experience in Virtual Reality using VR headsets. Import 360° media assets and add hotspots, quizzes and other interactive elements to engage your learners with near real-life scenarios
  • Interactive videos: Liven up demos and training videos by making them interactive with the new Adobe Captivate. Create your own or bring in existing YouTube videos, add questions at specific points and conduct knowledge checks to aid learner remediation
  • Fluid Boxes 2.0: Explore the building blocks of Smart eLearning design with intelligent containers that use white space optimally. Objects placed in Fluid Boxes get aligned automatically so that learners always get fully responsive experience regardless of their device or browser.
  • 360° learning experiences: Augment the learning landscape with 360° images and videos and convert them into interactive eLearning material with customizable overlay items such as information blurbs, audio content & quizzes.

 

 

Blippar unveils indoor visual positioning system to anchor AR — from martechtoday.com by Barry Levine
Employing machine vision to recognize mapped objects, the company says it can determine which way a user is looking and can calculate positioning down to a centimeter.

A Blippar visualization of AR using its new indoor visual positioning system

 

The Storyteller’s Guide to the Virtual Reality Audience — from medium.com by Katy Newton

Excerpt:

To even scratch the surface of these questions, we need to better understand the audience’s experience in VR — not just their experience of the technology, but the way that they understand story and their role within it.

 

 

Hospital introducing HoloLens augmented reality into the operating room — from medgadget.com

Excerpt:

HoloLens technology is being paired with Microsoft’s Surface Hub, a kind of digital whiteboard. The idea is that the surgical team can gather together around a Surface Hub to review patient information, discuss the details of a procedure, and select what information should be readily accessible during surgery. During the procedure, a surgeon wearing a HoloLens would be able to review a CT or MRI scan, access other data in the electronic medical records, and to be able to manipulate these so as to get a clear picture of what is being worked on and what needs to be done.

 

 

Raleigh Fire Department invests in virtual reality to enrich training — from vrfocus.com by Nikholai Koolon
New system allows department personnel to learn new skills through immersive experiences.

Excerpt:

The VR solution allows emergency medical services (EMS) personnel to dive into a rich and detailed environment which allows them to pinpoint portions of the body to dissect. This then allows them then see each part of the body in great detail along with viewing it from any angle. The goal is to allow for users to gain the experience to diagnose injuries from a variety of vantage points all where working within an virtual environment capable of displaying countless scenarios.

 

 

For another emerging technology, see:

Someday this tiny spider bot could perform surgery inside your body — from fastcompany.com by Jesus Diaz
The experimental robots could also fix airplane engines and find disaster victims.

Excerpt:

A team of Harvard University researchers recently achieved a major breakthrough in robotics, engineering a tiny spider robot using tech that could one day work inside your body to repair tissues or destroy tumors. Their work could not only change medicine–by eliminating invasive surgeries–but could also have an impact on everything from how industrial machines are maintained to how disaster victims are rescued.

Until now, most advanced, small-scale robots followed a certain model: They tend to be built at the centimeter scale and have only one degree of freedom, which means they can only perform one movement. Not so with this new ‘bot, developed by scientists at Harvard’s Wyss Institute for Biologically Inspired Engineering, the John A. Paulson School of Engineering and Applied Sciences, and Boston University. It’s built at the millimeter scale, and because it’s made of flexible materials–easily moved by pneumatic and hydraulic power–the critter has an unprecedented 18 degrees of freedom.

 


Plus some items from a few weeks ago


 

After almost a decade and billions in outside investment, Magic Leap’s first product is finally on sale for $2,295. Here’s what it’s like. — from

Excerpts (emphasis DSC):

I liked that it gave a new perspective to the video clip I’d watched: It threw the actual game up on the wall alongside the kind of information a basketball fan would want, including 3-D renderings and stats. Today, you might turn to your phone for that information. With Magic Leap, you wouldn’t have to.

Abovitz also said that intelligent assistants will play a big role in Magic Leap’s future. I didn’t get to test one, but Abovitz says he’s working with a team in Los Angeles that’s developing high-definition people that will appear to Magic Leap users and assist with tasks. Think Siri, Alexa or Google Assistant, but instead of speaking to your phone, you’d be speaking to a realistic-looking human through Magic Leap. Or you might be speaking to an avatar of someone real.

“You might need a doctor who can come to you,” Abovitz said. “AI that appears in front of you can give you eye contact and empathy.”

 

And I loved the idea of being able to place a digital TV screen anywhere I wanted.

 

 

Magic Leap One Available For Purchase, Starting At $2,295 — from vrscout.com by Kyle Melnick

Excerpt:

December of last year U.S. startup Magic Leap unveiled its long-awaited mixed reality headset, a secretive device five years and $2.44B USD in the making.

This morning that same headset, now referred to as the Magic Leap One Creator Edition, became available for purchase in the U.S. On sale to creators at a hefty starting price of $2,275, the computer spatial device utilizes synthetic lightfields to capture natural lightwaves and superimpose interactive, 3D content over the real-world.

 

 

 

Magic Leap One First Hands-On Impressions for HoloLens Developers — from magic-leap.reality.news

Excerpt:

After spending about an hour with the headset running through set up and poking around its UI and a couple of the launch day apps, I thought it would be helpful to share a quick list of some of my first impressions as someone who’s spent a lot of time with a HoloLens over the past couple years and try to start answering many of the burning questions I’ve had about the device.

 

 

World Campus researches effectiveness of VR headsets and video in online classes — from news.psu.edu

Excerpt:

UNIVERSITY PARK, Pa. — Penn State instructional designers are researching whether using virtual reality and 360-degree video can help students in online classes learn more effectively.

Designers worked with professors in the College of Nursing to incorporate 360-degree video into Nursing 352, a class on Advanced Health Assessment. Students in the class, offered online through Penn State World Campus, were offered free VR headsets to use with their smartphones to create a more immersive experience while watching the video, which shows safety and health hazards in a patient’s home.

Bill Egan, the lead designer for the Penn State World Campus RN to BSN nursing program, said students in the class were surveyed as part of a study approved by the Institutional Review Board and overwhelmingly said that they enjoyed the videos and thought they provided educational value. Eighty percent of the students said they would like to see more immersive content such as 360-degree videos in their online courses, he said.

 

 

7 Practical Problems with VR for eLearning — from learnupon.com

Excerpt:

In this post, we run through some practical stumbling blocks that prevent VR training from being feasible for most.

There are quite a number of practical considerations which prevent VR from totally overhauling the corporate training world. Some are obvious, whilst others only become apparent after using the technology a number of times. It’s important to be made aware of these limitations so that a large investment isn’t made in tech that isn’t really practical for corporate training.

 

Augmented reality – the next big thing for HR? — from hrdconnect.com
Augmented reality (AR) could have a huge impact on HR, transforming long-established processes into engaging and exciting something. What will this look like? How can we shape this into our everyday working lives?

Excerpt (emphasis DSC):

AR also has the potential to revolutionise our work lives, changing the way we think about office spaces and equipment forever.

Most of us still commute to an office every day, which can be a time-consuming and stressful experience. AR has the potential to turn any space into your own customisable workspace, complete with digital notes, folders and files – even a digital photo of your loved ones. This would give you access to all the information and tools that you would typically find in an office, but wherever and whenever you need them.

And instead of working on a flat, stationary, two-dimensional screen, your workspace would be a customisable three-dimensional space, where objects and information are manipulated with gestures rather than hardware. All you would need is an AR headset.

AR could also transform the way we advertise brands and share information. Imagine if your organisation had an AR stand at a conference – how engaging would that be for potential customers? How much more interesting and fun would meetings be if we used AR to present information instead of slides on a projector?

AR could transform the on-boarding experience into something fun and interactive – imagine taking an AR tour of your office, where information about key places, company history or your new colleagues pops into view as you go from place to place. 

 

 

RETINA Are Bringing Augmented Reality To Air Traffic Control Towers — from vrfocus.com by Nikholai Koolonavi

Excerpt:

A new project is aiming to make it easier for staff in airport control towers to visualize information to help make their job easier by leveraging augmented reality (AR) technology. The project, dubbed RETINA, is looking to modernise Europe’s air traffic management for safer, smarter and even smoother air travel.

 

 

 

25 skills LinkedIn says are most likely to get you hired in 2018 — and the online courses to get them — from businessinsider.com by Mara Leighton

Excerpt:

With the introduction of far-reaching and robust technology, the job market has experienced its own exponential growth, adaptation, and semi-metamorphosis. So much so that it can be difficult to guess what skills employer’s are looking for and what makes your résumé — and not another — stand out to recruiters.

Thankfully, LinkedIn created a 2018 “roadmap”— a list of hard and soft skills that companies need the most.

LinkedIn used data from their 500+ million members to identify the skills companies are currently working the hardest to fill. They grouped the skills members add to their profiles into several dozen categories (for example, “Android” and “iOS” into the “Mobile Development” category). Then, the company looked at all of the hiring and recruiting activity that happened on LinkedIn between January 1 and September 1 (billions of data points) and extrapolated the skill categories that belonged to members who were “more likely to start a new role within a company and receive interest from companies.”

LinkedIn then coupled those specific skills with related jobs and their average US salaries — all of which you can find below, alongside courses you can take (for free or for much less than the cost of a degree) to support claims of aptitude and stay ahead of the curve.

The online-learning options we included — LinkedIn Learning, Udemy, Coursera, and edX— are among the most popular and inexpensive.

 

 

Also see:

 

 

 

Despite increase in Instructional Designers, there is no ‘universal profile’ for the role — from campustechnology.com by Dian Schaffhauser

Excerpt:

In higher education, one 2016 estimate pegged the “minimum” number of instructional designers employed by colleges and universities at some 13,000. But, as a new report from the Online Learning Consortium pointed out, “There is still a certain mystery surrounding who instructional designers are.”

Instructional Design in Higher Education: Defining an Evolving Field seeks to clear up the mystery. The report is the second in a series to examine aspects of the digital learning landscape by doing a literature survey of existing research. The first report looked at accessibility. The series is being produced by OLC’s Research Center for Digital Learning & Leadership.

 

 

 

Three AI and machine learning predictions for 2019 — from forbes.com by Daniel Newman

Excerpt:

What could we potentially see next year? New and innovative uses for machine learning? Further evolution of human and machine interaction? The rise of AI assistants? Let’s dig deeper into AI and machine learning predictions for the coming months.

 

2019 will be a year of development for the AI assistant, showing us just how powerful and useful these tools are. It will be in more places than your home and your pocket too. Companies such as Kia and Hyundai are planning to include AI assistants in their vehicles starting in 2019. Sign me up for a new car! I’m sure that Google, Apple, and Amazon will continue to make advancements to their AI assistants making our lives even easier.

 

 

DeepMind AI matches health experts at spotting eye diseases — from endgadget.com by Nick Summers

Excerpt:

DeepMind has successfully developed a system that can analyze retinal scans and spot symptoms of sight-threatening eye diseases. Today, the AI division — owned by Google’s parent company Alphabet — published “early results” of a research project with the UK’s Moorfields Eye Hospital. They show that the company’s algorithms can quickly examine optical coherence tomography (OCT) scans and make diagnoses with the same accuracy as human clinicians. In addition, the system can show its workings, allowing eye care professionals to scrutinize the final assessment.

 

 

Microsoft and Amazon launch Alexa-Cortana public preview for Echo speakers and Windows 10 PCs — from venturebeat.com by Khari Johnson

Excerpt:

Microsoft and Amazon will bring Alexa and Cortana to all Echo speakers and Windows 10 users in the U.S. [on 8/15/18]. As part of a partnership between the Seattle-area tech giants, you can say “Hey Cortana, open Alexa” to Windows 10 PCs and “Alexa, open Cortana” to a range of Echo smart speakers.

The public preview bringing the most popular AI assistant on PCs together with the smart speaker with the largest U.S. market share will be available to most people today but will be rolled out to all users in the country over the course of the next week, a Microsoft spokesperson told VentureBeat in an email.

Each of the assistants brings unique features to the table. Cortana, for example, can schedule a meeting with Outlook, create location-based reminders, or draw on LinkedIn to tell you about people in your next meeting. And Alexa has more than 40,000 voice apps or skills made to tackle a broad range of use cases.

 

 

What Alexa can and cannot do on a PC — from venturebeat.com by Khari Johnson

Excerpt:

Whatever happened to the days of Alexa just being known as a black cylindrical speaker? Since the introduction of the first Echo in fall 2014, Amazon’s AI assistant has been embedded in a number of places, including car infotainment systems, Alexa smartphone apps, wireless headphones, Echo Show and Fire tablets, Fire TV Cube for TV control, the Echo Look with an AI-powered fashion assistant, and, in recent weeks, personal computers.

Select computers from HP, Acer, and others now make Alexa available to work seamlessly alongside Microsoft’s Cortana well ahead of the Alexa-Cortana partnership for Echo speakers and Windows 10 devices, a project that still has no launch date.

 

 

Vanguard Projects: Expanding Teaching and Learning Horizons
Authors — from er.educause.edu by Malcolm Brown
Members of the higher education community provided a long list of projects that exemplify the six developments in educational technology from this year’s Horizon Report. A few of the projects are highlighted to convey a sense of the range and direction of current innovations in teaching and learning in higher education.

Excerpt:

This list summarizes the distribution across the six developments (as laid out in the preview):

  • Mixed reality: 31.4%
  • Makerspaces: 28.6%
  • Adaptive learning technology: 16.2%
  • Analytics technologies: 15.2%
  • Artificial intelligence: 4.8%
  • Robotics: 3.8%

 

 

 

A more strategic approach to arranging students into groups — from facultyfocus.com by Maryellen Weimer

Excerpt:

What’s the best way to put students into groups? It’s the first task that confronts teachers who want students to work together. And the best reply is one of those “it depends” answers. Here are the questions on which it depends.

 

If the group work is a project that requires extended collaboration and will benefit from a variety of opinions and perspectives, letting students form the groups may not be the best approach. On the other hand, for short, ad-hoc group work and for students who may be shy and not used to working with peers, knowing others in the group makes the experience less intimidating.

 

If one of the goals of the group work is getting students acquainted with others in the course or providing the experience of learning to work with peers they don’t know (which frequently occurs in professional contexts), then teachers should consider forming the groups.

 

What criteria should teachers use when forming groups? There’s a range of options. Here’s some of the more common criteria.

  • No criteria
  • Ability
  • Personality traits
  • Skills and experiences

 

 

 

Can we design online learning platforms that feel more intimate than massive? — from edsurge.com by Amy Ahearn

Excerpt:

This presents a challenge and an opportunity: How can we design online learning environments that achieve scale and intimacy? How do we make digital platforms feel as inviting as well-designed physical classrooms?

The answer may be that we need to balance massiveness with miniaturization. If the first wave of MOOCs was about granting unprecedented numbers of students access to high-quality teaching and learning materials, Wave 2 needs to focus on creating a sense of intimacy within that massiveness.

We need to be building platforms that look less like a cavernous stadium and more like a honeycomb. This means giving people small chambers of engagement where they can interact with a smaller, more manageable and yet still diverse groups. We can’t meaningfully listen to the deafening roar of the internet. But we can learn from a collection of people with perspectives different than ours.

 

 

What will it take to get MOOC platforms to begin to offer learning spaces that feel more inviting and intimate? Perhaps there’s a new role that needs to emerge in the online learning ecosystem: a “learning architect” who sits between the engineers and the instructional designers.

 

 

 

 

 

 

Instructional Design in Higher Education: Defining an Evolving Field
From OLC Outlook: An Environmental Scan of the Digital Learning Landscape
Elaine Beirne (Dublin City University) and Matthew Romanoski (The University of Arizona)
July 2018

Description:
This white paper provides an overview of the growing field of instructional design in higher education, from why the field is growing to how designers are functioning in their role. It explores why there is a growing demand for designers, who is filling these roles, what the responsibilities of designers are, and how instructional designers are addressing the challenges they face.

From onlinelearningconsortium.org

 

Report: Accessibility in Digital Learning Increasingly Complex — from campustechnology.com by Dian Schaffhauser

Excerpt:

The Online Learning Consortium (OLC)has introduced a series of original reports to keep people in education up-to-date on the latest developments in the field of digital learning. The first report covers accessibility and addresses both K-12 and higher education. The series is being produced by OLC’s Research Center for Digital Learning & Leadership.

The initial report addresses four broad areas tied to accessibility:

  • The national laws governing disability and access and how they apply to online courses;
  • What legal cases exist to guide online course design and delivery in various educational settings;
  • The issues that emerge regarding online course access that might be unique to higher ed or to K-12, and which ones might be shared; and
  • What support online course designers need to generate accessible courses for learners across the education life span (from K-12 to higher education).

 

 

The title of this article being linked to is: Augmented and virtual reality mean business: Everything you need to know

 

Augmented and virtual reality mean business: Everything you need to know — from zdnet by Greg Nichols
An executive guide to the technology and market drivers behind the hype in AR, VR, and MR.

Excerpt:

Overhyped by some, drastically underestimated by others, few emerging technologies have generated the digital ink like virtual reality (VR), augmented reality (AR), and mixed reality (MR).  Still lumbering through the novelty phase and roller coaster-like hype cycles, the technologies are only just beginning to show signs of real world usefulness with a new generation of hardware and software applications aimed at the enterprise and at end users like you. On the line is what could grow to be a $108 billion AR/VR industry as soon as 2021. Here’s what you need to know.

 

The reason is that VR environments by nature demand a user’s full attention, which make the technology poorly suited to real-life social interaction outside a digital world. AR, on the other hand, has the potential to act as an on-call co-pilot to everyday life, seamlessly integrating into daily real-world interactions. This will become increasingly true with the development of the AR Cloud.

The AR Cloud
Described by some as the world’s digital twin, the AR Cloud is essentially a digital copy of the real world that can be accessed by any user at any time.

For example, it won’t be long before whatever device I have on me at a given time (a smartphone or wearable, for example) will be equipped to tell me all I need to know about a building just by training a camera at it (GPS is operating as a poor-man’s AR Cloud at the moment).

What the internet is for textual information, the AR Cloud will be for the visible world. Whether it will be open source or controlled by a company like Google is a hotly contested issue.

 

Augmented reality will have a bigger impact on the market and our daily lives than virtual reality — and by a long shot. That’s the consensus of just about every informed commentator on the subject.

 

 

 

Mixed reality will transform learning (and Magic Leap joins act one) — from edsurge.com by Maya Georgieva

Excerpt:

Despite all the hype in recent years about the potential for virtual reality in education, an emerging technology known as mixed reality has far greater promise in and beyond the classroom.

Unlike experiences in virtual reality, mixed reality interacts with the real world that surrounds us. Digital objects become part of the real world. They’re not just digital overlays, but interact with us and the surrounding environment.

If all that sounds like science fiction, a much-hyped device promises some of those features later this year. The device is by a company called Magic Leap, and it uses a pair of goggles to project what the company calls a “lightfield” in front of the user’s face to make it look like digital elements are part of the real world. The expectation is that Magic Leap will bring digital objects in a much more vivid, dynamic and fluid way compared to other mixed-reality devices such as Microsoft’s Hololens.

 

The title of the article being linked to here is Mixed reality will transform learning (and Magic Leap joins act one)

 

Now think about all the other things you wished you had learned this way and imagine a dynamic digital display that transforms your environment and even your living room or classroom into an immersive learning lab. It is learning within a highly dynamic and visual context infused with spatial audio cues reacting to your gaze, gestures, gait, voice and even your heartbeat, all referenced with your geo-location in the world. Unlike what happens with VR, where our brain is tricked into believing the world and the objects in it are real, MR recognizes and builds a map of your actual environment.

 

 

 

Also see:

virtualiteach.com
Exploring The Potential for the Vive Focus in Education

 

virtualiteach.com

 

 

 

Digital Twins Doing Real World Work — from stambol.com

Excerpt:

On the big screen it’s become commonplace to see a 3D rendering or holographic projection of an industrial floor plan or a mechanical schematic. Casual viewers might take for granted that the technology is science fiction and many years away from reality. But today we’re going to outline where these sophisticated virtual replicas – Digital Twins – are found in the real world, here and now. Essentially, we’re talking about a responsive simulated duplicate of a physical object or system. When we first wrote about Digital Twin technology, we mainly covered industrial applications and urban infrastructure like transit and sewers. However, the full scope of their presence is much broader, so now we’re going to break it up into categories.

 

The title of the article being linked to here is Digital twins doing real world work

 

Digital twin — from Wikipedia

Digital twin refers to a digital replica of physical assets (physical twin), processes and systems that can be used for various purposes.[1] The digital representation provides both the elements and the dynamics of how an Internet of Things device operates and lives throughout its life cycle.[2]

Digital twins integrate artificial intelligence, machine learning and software analytics with data to create living digital simulation models that update and change as their physical counterparts change. A digital twin continuously learns and updates itself from multiple sources to represent its near real-time status, working condition or position. This learning system, learns from itself, using sensor data that conveys various aspects of its operating condition; from human experts, such as engineers with deep and relevant industry domain knowledge; from other similar machines; from other similar fleets of machines; and from the larger systems and environment in which it may be a part of. A digital twin also integrates historical data from past machine usage to factor into its digital model.

In various industrial sectors, twins are being used to optimize the operation and maintenance of physical assets, systems and manufacturing processes.[3] They are a formative technology for the Industrial Internet of Things, where physical objects can live and interact with other machines and people virtually.[4]

 

 

Disney to debut its first VR short next month — from techcrunch.com by Sarah Wells

Excerpt:

Walt Disney Animation Studio is set to debut its first VR short film, Cycles, this August in Vancouver, the Association for Computing Machinery announced today. The plan is for it to be a headliner at the ACM’s computer graphics conference (SIGGRAPH), joining other forms of VR, AR and MR entertainment in the conference’s designated Immersive Pavilion.

This film is a first for both Disney and its director, Jeff Gipson, who joined the animation team in 2013 to work as a lighting artist on films like Frozen, Zootopia and Moana. The objective of this film, Gipson said in the statement released by ACM, is to inspire a deep emotional connection with the story.

“We hope more and more people begin to see the emotional weight of VR films, and with Cycles in particular, we hope they will feel the emotions we aimed to convey with our story,” said Gipson.

 

 

 

 

 

Guiding faculty into immersive environments — from campustechnology.com by David Raths
What’s the best way to get faculty to engage with emerging technologies and incorporate new learning spaces into their teaching? Five institutions share their experiences.

Guiding faculty into immersive environments -- by David Raths

Excerpt:

One of the biggest hurdles for universities has been the high cost of VR-enabled computers and headsets, and some executives say prices must continue to drop before we’ll see more widespread usage. But John Bowditch, director of the Game Research and Immersive Design Lab at Ohio University’s Scripps College of Communication, is already seeing promising developments on that front as he prepares to open a new 20-seat VR classroom. “Probably the best thing about VR in 2018 is that it is a lot more affordable now and that democratizes it,” he said. “We purchased a VR helmet 13 years ago, and it was $12,000 just for the headset. The machine that ran it cost about $20,000. That would be a nonstarter beyond purchasing just one or two. Today, you can get a VR-enabled laptop and headset for under $2,000. That makes it much easier to think about integrating it into classes.”

 

 

Colleges and universities face several hurdles in getting faculty to incorporate virtual reality or immersive experiences in their courses. For one, instructional designers, instructional technologists and directors of teaching and learning centers may not have access to these tools yet, and the budgets aren’t always there to get the labs off the ground, noted Daniel Christian, instructional services director at Western Michigan University‘s Cooley Law School. “Many faculty members’ job plates are already jam-packed — allowing little time to even look at emerging technologies,” he said. “Even if they wanted to experiment with such technologies and potential learning experiences, they don’t have the time to do so. Tight budgets are impacting this situation even further.”

 

 

 

 

Computing in the Camera — from blog.torch3d.com by Paul Reynolds
Mobile AR, with its ubiquitous camera, is set to transform what and how human experience designers create.

One of the points Allison [Woods, CEO, Camera IQ] made repeatedly on that call (and in this wonderful blog post of the same time period) was that the camera is going to be at the center of computing going forward, an indispensable element. Spatial computing could not exist without it. Simple, obvious, straightforward, but not earth shaking. We all heard what she had to say, but I don’t think any of us really understood just how profound or prophetic that statement turned out to be.

 

“[T]he camera will bring the internet and the real world into a single time and space.”

— Allison Woods, CEO, Camera IQ

 

 

The Camera As Platform — from shift.newco.co by Allison Wood
When the operating system moves to the viewfinder, the world will literally change

“Every day two billion people carry around an optical data input device — the smartphone Camera — connected to supercomputers and informed by massive amounts of data that can have nearly limitless context, position, recognition and direction to accomplish tasks.”

– Jacob Mullins, Shasta Ventures

 

 

 

The State Of The ARt At AWE 18 — from forbes.com by Charlie Fink

Excerpt:

The bigger story, however, is how fast the enterprise segment is growing as applications as straightforward as schematics on a head-mounted monocular microdisplay are transforming manufacturing, assembly, and warehousing. Use cases abounded.

After traveling the country and most recently to Europe, I’ve now experienced almost every major VR/AR/MR/XR related conference out there. AWE’s exhibit area was by far the largest display of VR and AR companies to date (with the exception of CES).

 

AR is being used to identify features and parts within cars

 

 

 

 

Student Learning and Virtual Reality: The Embodied Experience — from er.educause.edu by Jaime Hannans, Jill Leafstedt and Talya Drescher

Excerpts:

Specifically, we explored the potential for how virtual reality can help create a more empathetic nurse, which, we hypothesize, will lead to increased development of nursing students’ knowledge, skills, and attitudes. We aim to integrate these virtual experiences into early program coursework, with the intent of changing nursing behavior by providing a deeper understanding of the patient’s perspective during clinical interactions.

In addition to these compelling student reflections and the nearly immediate change in reporting practice, survey findings show that students unanimously felt that this type of patient-perspective VR experience should be integrated and become a staple of the nursing curriculum. Seeing, hearing, and feeling these moments results in significant and memorable learning experiences compared to traditional classroom learning alone. The potential that this type of immersive experience can have in the field of nursing and beyond is only limited by the imagination and creation of other virtual experiences to explore. We look forward to continued exploration of the impact of VR on student learning and to establishing ongoing partnerships with developers.

 

Also see:

 

 

 

 

AR & VR — Education’s marvelous revolution — from verizoninternet.com

Excerpt:

Virtual Reality (VR) and Augmented Reality (AR), often used in video games and mobile apps, are transforming the world—and with that, the way we learn. These technologies have the capability to change students’ outlook on the world and the way they engage with it. After all, why would you learn about outer space from a classroom when you could learn about it from the International Space Station?

As AR and VR technology become more widely available and user-friendly, interest and market value have spiked across the world. In 2017, interest in VR hardware such as PlayStation VR, HTC Vive, Oculus Rift, and Samsung Gear VR spiked around the globe.2 In China in particular, AR and VR are booming.

With AR and VR, geographical distances are no longer an obstacle. Interactive experiences, tutorial videos, and learning apps work just as well, whether the teacher and student are in the same room, or across the world from each other.

From their site, here are some additional resources:

 

 

Study shows VR increases learning — from Donald Clark

 Excerpt:

I have argued that several conditions for good learning are likely to be enhanced by VR. First there’s increased attention, where the learner is literally held fast within the created environment and cannot be distracted by external stimuli. Second is experiential learning, where one has to ‘do’ something where that active component leads to higher retention. Third is emotion, the affective component in learning, which is better achieved where the power to induce empathy, excitement, calm and so on is easier. Fourth is context, where providing an albeit simulated context aids retention and recall. Fifth is transfer, where all of these conditions lead to greater transfer of knowledge and skills to the real world.

 

 

Example Use Cases of How to Use Virtual Reality (VR) for Training — from instavr.co

Some of the topics covered include:

  • Employee Onboarding (and Cross-Training)
  • Preparing for Rare or Unexpected Events
  • Employee Testing
  • Customer/Client Interaction Practice

 

 

8 of the Wildest Augmented Reality Glasses You Haven’t Seen Yet — from next.reality.news by Adario Strange

 

 

 

 
© 2025 | Daniel Christian