Toggle light / dark theme

Varjo’s Unique Masking Tool Lets You Bring the Real World into VR

Varjo’s XR-3 headset has perhaps the best passthrough view of any MR headset on the market thanks to color cameras that offer a fairly high resolution and a wide field-of-view. But rather than just using the passthrough view for AR (bringing virtual objects into the real world) Varjo has developed a new tool to do the reverse (bringing real objects into the virtual world).

At AWE 2021 this week I got my first glimpse at ‘Varjo Lab Tools’, a soon-to-be released software suite that will work with the company’s XR-3 mixed reality headset. The tool allows users to trace arbitrary shapes that then become windows into the real world, while the rest of the view remains virtual.

Nvidia’s Omniverse adds AR/VR viewing, AI training, and AI avatar creation

Nvidia’s Omniverse, billed as a “metaverse for engineers,” has grown to more than 700 companies and 70,000 individual creators that are working on projects to simulate digital twins that replicate real-world environments in a virtual space.

The Omniverse is Nvidia’s simulation and collaboration platform delivering the foundation of the metaverse, the universe of virtual worlds that are all interconnected, like in novels such as Snow Crash and Ready Player One. Omniverse is now moving from beta to general availability, and it has been extended to software ecosystems that put it within reach of 40 million 3D designers.

And today during Nvidia CEO Jensen Huang’s keynote at the Nvidia GTC online conference, Nvidia said it has added features such as Omniverse Replicator, which makes it easier to train AI deep learning neural networks, and Omniverse avatar, which makes it simple to create virtual characters that can be used in the Omniverse or other worlds.

Lynx R-1 MR Headset Kickstarter Comes to an End with over $800,000

French startup Lynx launched a Kickstarter campaign for Lynx R-1 in October, a standalone MR headset which is capable of both VR and passthrough AR. Starting at €530 (or $500 if you’re not subject to European sales tax), the MR headset attracted a strong response from backers as it passed its initial funding goal in under 15 hours, going on to garner over $800,000 throughout the month-long campaign.

Update (November 10th, 2021): Lynx R-1 Kickstarter is now over, and it’s attracted €725,281 (~$835,000) from 1,216 backers. In the final hours the campaign managed to pass its first stretch goal at $700,000—a free facial interface pad.

If you missed out, the company is now offering direct preorders for both its Standard Edition for $600 and Enterprise Edition for $1,100. It’s also selling a few accessories including compatible 6DOF controllers, facial interfaces, and a travel case.

SEVEN CONVERGING TECHNOLOGICAL FUNCTIONS WILL PERMANENTLY CHANGE CITIZEN ASTRONOMY

By Jeremy Batterson 11-09-2021

The equivalent of cheap 100-inch binoculars will soon be possible. This memo is a quick update on seven rapidly converging technologies that augur well for astronomy enthusiasts of the near future. All these technologies already exist in either fully developed or nascent form, and all are being rapidly improved due to the gigantic global cell phone market and the retinal projection market that will soon replace it. Listed here are the multiple technologies, after which they are brought together into a single system.

1) Tracking.
2) Single-photon image sensing.
3) Large effective exit pupils via large sensors.
4) Long exposure non-photographic function.
5) Flat optics (metamaterials)
6) Off-axis function of flat optics.
7) Retinal projection.

1) TRACKING: this is already being widely used in so-called “go-to” telescopes, where the instrument will find any object and track it, so Earth’s rotation does not take the object viewed out of the field of vision. The viewer doesn’t have to find the object and doesn’t have to set up the clock drive to track it. Tracking is also partly used in image stabilization software for cameras and smart phones, to prevent motion blurring of images.

2) SINGLE-PHOTON IMAGE SENSORS, whether of the single-photon avalanching diode type, or the type developed by Dr. Fossum, will allow passive imaging in nearly totally dark environments, without the use of IR or other illumination. This new type of image sensor will replace the monochromatic analogue “night-vision” devices, allowing color imaging at higher resolution than they can produce. Unlike these current devices, such sensors will not be destroyed by being exposed to normal or high lighting. Effectively, these sensors increase the effective light-gathering power of a telescope by at least an order of magnitude, allowing small telescopes to see what observatory telescopes see now.

3) EXIT PUPIL: The pupil of the dark-adapted human eye is around 7mm, which means light exiting a telescope must not have a wider-cross axis than this, or a percent of the light captured by the objective lens or mirror will be lost. If the magnification of a system is lowered, to give brighter images, this is limited by this roadblock. This is a well-known problem for visual astronomers. Astro-photographers get around this by two tricks. The first is to use a photographic sensor wider than 7mm, allowing a larger exit pupil and thus brighter images. A 1-inch sensor or photographic plate, for example, already allows an image thirteen times brighter than what a 7mm human pupil can see.

4) LONG EXPOSURE: The other trick astro-photographers use is to keep the shutter of their cameras open for longer periods, thus capturing more light, and allowing a bright image of a faint object to build up over time. As a telescope tracks the stars–so that they appear motionless in the telescopic view–this can be done for hours. The Hubble Space Telescope took a 100 hour long-exposure photograph leading to the famous “deep field” of ultra-faint distant galaxies. An example of a visual use of the same principle is the Sionyx Pro camera, which keeps the shutter open for a fraction of a second. If the exposures are short enough, a video can be produced which appears brighter than what the unaided eye sees. Sionyx adds to this with its black-silicon sensors, which are better at retaining all light that hits them. For astronomy, where stellar objects do not move and do not cause blurring if they are tracked, longer exposures can be created, with the image rapidly brightening as the viewer watches. Unistellar’s eVscope and Vaonis’s Stellina telescope, already use this function, but without an eyepiece. Instead, their images are projected onto people’s cell phones or other viewing devices. However, most astronomers want to be able to see something directly with their eyes, which is a limiting point on such types of telescopes.

Qualcomm is Preparing Developers for the Coming Wave of Smartphone-tethered AR Glasses

Today at AWE 2,021 Qualcomm announced Snapdragon Spaces XR Developer Platform, a head-worn AR software suite the company is using to kickstart a broader move towards smartphone-tethered AR glasses.

Qualcomm says its Snapdragon Spaces XR Developer Platform offers a host of machine perception functions that are ideal for smartphone-tethered AR glasses. The software tool kit focuses on performance and low power, and provides the sort of environmental and human interaction stuff it hopes will give AR developers a good starting point.

Mansoor Hanif — Executive Director, Emerging Technologies, NEOM — An Accelerator Of Human Progress

A US$500 billion accelerator of human progress — mansoor hanif, executive director, emerging technologies, NEOM.


Mansoor Hanif is the Executive Director of Emerging Technologies at NEOM (https://www.neom.com/en-us), a fascinating $500 billion planned cognitive city” & tourist destination, located in north west Saudi Arabia, where he is responsible for all R&D activities for the Technology & Digital sector, including space technologies, advanced robotics, human-machine interfaces, sustainable infrastructure, digital master plans, digital experience platforms and mixed reality. He also leads NEOM’s collaborative research activities with local and global universities and research institutions, as well as manages the team developing world-leading Regulations for Communications and Connectivity.

Prior to this role, Mr Hanif served as Executive Director, Technology & Digital Infrastructure, where he oversaw the design and implementation of NEOM’s fixed, mobile, satellite and sub-sea networks.

An industry leader, Mr Hanif has over 25 years of experience in planning, building, optimizing and operating mobile networks around the world. He is patron of the Institute of Telecommunications Professionals (ITP), a member of the Steering Board of the UK5G Innovation Network, and on the Advisory Boards of the Satellite Applications Catapult and University College London (UCL) Electrical and Electronic Engineering Dept.

Prior to joining NEOM, Mr Hanif was Chief Technology Officer of Ofcom, the UK telecoms and media regulator, where he oversaw the security and resilience of the nation’s networks.

10 EMERGING Technologies That Will Change The World

Today, we’re bringing you 10 EMERGING Technologies That Will Change The World. Better stick around for #1 to find out how Elon Musk may have plans to turn us all into human robots some day.
What’s up tech-heads and welcome to another episode of TechJoint! It really feels like we’re already living in the future every day. From 5G connectivity to self-driving cars being even more accessible, innovation is everywhere we look! It can sometimes be hard to imagine a world with even more innovation—a future world. What would it look like? What everyday problems would be solved? It’s a pretty good bet some technologies like artificial intelligence will be in our lives, playing important roles in the future of humankind. Other technologies may seem far fetched, unnecessary and frankly, unattainable. So buckle up, and let’s take a look at these 10 futuristic technologies that are going to change the world as we know it.

► Subscribe For More! https://goo.gl/FQqpG8

10 EMERGING technologies that will change the world.

10. 10 Voice Assistants.
9. Gene Splicing.
8. Mixed Reality.
7. Regenerative Medicine.
6. Fully Autonomous Vehicles.
5. Digital Wallets.
4. Artificial Intelligence.
3. Automation.
2. ‘Alive’ Building Materials.
1. Internet For Everyone.

► Disclaimer: This video description contains affiliate links. Meaning, if you click on one of the product links and make a purchase, we receive a small commission. This helps us keep making more videos. Thank you for your support!

For Copyright Issues, Please Contact [email protected]

Microsoft Teams enters the metaverse race with 3D avatars and immersive meetings

Microsoft is entering the race to build a metaverse inside Teams, just days after Facebook rebranded to Meta in a push to build virtual spaces for both consumers and businesses. Microsoft is bringing Mesh, a collaborative platform for virtual experiences, directly into Microsoft Teams next year. It’s part of a big effort to combine the company’s mixed reality and HoloLens work with meetings and video calls that anyone can participate in thanks to animated avatars.

With today’s announcement, Microsoft and Meta seem to be on a collision course to compete heavily in the metaverse, particularly for the future of work.

Microsoft Mesh always felt like the future of Microsoft Teams meetings, and now it’s starting to come to life in the first half of 2022. Microsoft is building on efforts like Together Mode and other experiments for making meetings more interactive, after months of people working from home and adjusting to hybrid work.

Apple Patent Outlines Direct Retinal Projection To Beam AR Images Onto Your Eyeballs

Apple is looking into how it may change how you view AR (augmented reality) altogether…literally. Instead of projecting an image onto a lens, which is viewed by someone wearing an AR headset or glasses, Apple envisions beaming the image directly onto the user’s eyeball itself.

Apple recently unveiled its upcoming lineup of new products. What it did not showcase, however, was revealed in a recent patent—Apple is shown researching how it can change how we see AR and the future of its “Apple Glass” product, if one comes to exist. The patent reveals how Apple intends to move away from the traditional way of projecting an image onto a lens, to projecting the image directly onto the retina of the wearer. This will be achieved through the use of micro projectors.

The issue that Apple is trying to avoid is the nausea and headaches some people experience while viewing AR and VR (virtual reality). The patent states the issue as “accommodation-convergence mismatch,” which causes eyestrain for some. Apple hopes that by using its “Direct Retinal Projector” it can alleviate those symptoms and make the AR and VR realm accessible for more users.

Samsung invests in DigiLens XR glasses firm at valuation over $500M

DigiLens has raised funding from Samsung Electronics in a round that values the augmented reality smart glasses makers at more than $500 million.

Sunnyvale, California-based DigiLens did not say the exact amount it raised for the development of its extended reality glasses (XR), which will offer AR features, such as overlaying digital images on what you see.

DigiLens CEO Chris Pickett said in a previous interview with VentureBeat that the latest smart glasses are more advanced than models the company showed in 2019.

/* */