Toggle light / dark theme

SEVEN CONVERGING TECHNOLOGICAL FUNCTIONS WILL PERMANENTLY CHANGE CITIZEN ASTRONOMY

By Jeremy Batterson 11-09-2021

The equivalent of cheap 100-inch binoculars will soon be possible. This memo is a quick update on seven rapidly converging technologies that augur well for astronomy enthusiasts of the near future. All these technologies already exist in either fully developed or nascent form, and all are being rapidly improved due to the gigantic global cell phone market and the retinal projection market that will soon replace it. Listed here are the multiple technologies, after which they are brought together into a single system.

1) Tracking.
2) Single-photon image sensing.
3) Large effective exit pupils via large sensors.
4) Long exposure non-photographic function.
5) Flat optics (metamaterials)
6) Off-axis function of flat optics.
7) Retinal projection.

1) TRACKING: this is already being widely used in so-called “go-to” telescopes, where the instrument will find any object and track it, so Earth’s rotation does not take the object viewed out of the field of vision. The viewer doesn’t have to find the object and doesn’t have to set up the clock drive to track it. Tracking is also partly used in image stabilization software for cameras and smart phones, to prevent motion blurring of images.

2) SINGLE-PHOTON IMAGE SENSORS, whether of the single-photon avalanching diode type, or the type developed by Dr. Fossum, will allow passive imaging in nearly totally dark environments, without the use of IR or other illumination. This new type of image sensor will replace the monochromatic analogue “night-vision” devices, allowing color imaging at higher resolution than they can produce. Unlike these current devices, such sensors will not be destroyed by being exposed to normal or high lighting. Effectively, these sensors increase the effective light-gathering power of a telescope by at least an order of magnitude, allowing small telescopes to see what observatory telescopes see now.

3) EXIT PUPIL: The pupil of the dark-adapted human eye is around 7mm, which means light exiting a telescope must not have a wider-cross axis than this, or a percent of the light captured by the objective lens or mirror will be lost. If the magnification of a system is lowered, to give brighter images, this is limited by this roadblock. This is a well-known problem for visual astronomers. Astro-photographers get around this by two tricks. The first is to use a photographic sensor wider than 7mm, allowing a larger exit pupil and thus brighter images. A 1-inch sensor or photographic plate, for example, already allows an image thirteen times brighter than what a 7mm human pupil can see.

4) LONG EXPOSURE: The other trick astro-photographers use is to keep the shutter of their cameras open for longer periods, thus capturing more light, and allowing a bright image of a faint object to build up over time. As a telescope tracks the stars–so that they appear motionless in the telescopic view–this can be done for hours. The Hubble Space Telescope took a 100 hour long-exposure photograph leading to the famous “deep field” of ultra-faint distant galaxies. An example of a visual use of the same principle is the Sionyx Pro camera, which keeps the shutter open for a fraction of a second. If the exposures are short enough, a video can be produced which appears brighter than what the unaided eye sees. Sionyx adds to this with its black-silicon sensors, which are better at retaining all light that hits them. For astronomy, where stellar objects do not move and do not cause blurring if they are tracked, longer exposures can be created, with the image rapidly brightening as the viewer watches. Unistellar’s eVscope and Vaonis’s Stellina telescope, already use this function, but without an eyepiece. Instead, their images are projected onto people’s cell phones or other viewing devices. However, most astronomers want to be able to see something directly with their eyes, which is a limiting point on such types of telescopes.

Qualcomm is Preparing Developers for the Coming Wave of Smartphone-tethered AR Glasses

Today at AWE 2,021 Qualcomm announced Snapdragon Spaces XR Developer Platform, a head-worn AR software suite the company is using to kickstart a broader move towards smartphone-tethered AR glasses.

Qualcomm says its Snapdragon Spaces XR Developer Platform offers a host of machine perception functions that are ideal for smartphone-tethered AR glasses. The software tool kit focuses on performance and low power, and provides the sort of environmental and human interaction stuff it hopes will give AR developers a good starting point.

Mansoor Hanif — Executive Director, Emerging Technologies, NEOM — An Accelerator Of Human Progress

A US$500 billion accelerator of human progress — mansoor hanif, executive director, emerging technologies, NEOM.


Mansoor Hanif is the Executive Director of Emerging Technologies at NEOM (https://www.neom.com/en-us), a fascinating $500 billion planned cognitive city” & tourist destination, located in north west Saudi Arabia, where he is responsible for all R&D activities for the Technology & Digital sector, including space technologies, advanced robotics, human-machine interfaces, sustainable infrastructure, digital master plans, digital experience platforms and mixed reality. He also leads NEOM’s collaborative research activities with local and global universities and research institutions, as well as manages the team developing world-leading Regulations for Communications and Connectivity.

Prior to this role, Mr Hanif served as Executive Director, Technology & Digital Infrastructure, where he oversaw the design and implementation of NEOM’s fixed, mobile, satellite and sub-sea networks.

An industry leader, Mr Hanif has over 25 years of experience in planning, building, optimizing and operating mobile networks around the world. He is patron of the Institute of Telecommunications Professionals (ITP), a member of the Steering Board of the UK5G Innovation Network, and on the Advisory Boards of the Satellite Applications Catapult and University College London (UCL) Electrical and Electronic Engineering Dept.

Prior to joining NEOM, Mr Hanif was Chief Technology Officer of Ofcom, the UK telecoms and media regulator, where he oversaw the security and resilience of the nation’s networks.

10 EMERGING Technologies That Will Change The World

Today, we’re bringing you 10 EMERGING Technologies That Will Change The World. Better stick around for #1 to find out how Elon Musk may have plans to turn us all into human robots some day.
What’s up tech-heads and welcome to another episode of TechJoint! It really feels like we’re already living in the future every day. From 5G connectivity to self-driving cars being even more accessible, innovation is everywhere we look! It can sometimes be hard to imagine a world with even more innovation—a future world. What would it look like? What everyday problems would be solved? It’s a pretty good bet some technologies like artificial intelligence will be in our lives, playing important roles in the future of humankind. Other technologies may seem far fetched, unnecessary and frankly, unattainable. So buckle up, and let’s take a look at these 10 futuristic technologies that are going to change the world as we know it.

► Subscribe For More! https://goo.gl/FQqpG8

10 EMERGING technologies that will change the world.

10. 10 Voice Assistants.
9. Gene Splicing.
8. Mixed Reality.
7. Regenerative Medicine.
6. Fully Autonomous Vehicles.
5. Digital Wallets.
4. Artificial Intelligence.
3. Automation.
2. ‘Alive’ Building Materials.
1. Internet For Everyone.

► Disclaimer: This video description contains affiliate links. Meaning, if you click on one of the product links and make a purchase, we receive a small commission. This helps us keep making more videos. Thank you for your support!

For Copyright Issues, Please Contact [email protected]

Microsoft Teams enters the metaverse race with 3D avatars and immersive meetings

Microsoft is entering the race to build a metaverse inside Teams, just days after Facebook rebranded to Meta in a push to build virtual spaces for both consumers and businesses. Microsoft is bringing Mesh, a collaborative platform for virtual experiences, directly into Microsoft Teams next year. It’s part of a big effort to combine the company’s mixed reality and HoloLens work with meetings and video calls that anyone can participate in thanks to animated avatars.

With today’s announcement, Microsoft and Meta seem to be on a collision course to compete heavily in the metaverse, particularly for the future of work.

Microsoft Mesh always felt like the future of Microsoft Teams meetings, and now it’s starting to come to life in the first half of 2022. Microsoft is building on efforts like Together Mode and other experiments for making meetings more interactive, after months of people working from home and adjusting to hybrid work.

Apple Patent Outlines Direct Retinal Projection To Beam AR Images Onto Your Eyeballs

Apple is looking into how it may change how you view AR (augmented reality) altogether…literally. Instead of projecting an image onto a lens, which is viewed by someone wearing an AR headset or glasses, Apple envisions beaming the image directly onto the user’s eyeball itself.

Apple recently unveiled its upcoming lineup of new products. What it did not showcase, however, was revealed in a recent patent—Apple is shown researching how it can change how we see AR and the future of its “Apple Glass” product, if one comes to exist. The patent reveals how Apple intends to move away from the traditional way of projecting an image onto a lens, to projecting the image directly onto the retina of the wearer. This will be achieved through the use of micro projectors.

The issue that Apple is trying to avoid is the nausea and headaches some people experience while viewing AR and VR (virtual reality). The patent states the issue as “accommodation-convergence mismatch,” which causes eyestrain for some. Apple hopes that by using its “Direct Retinal Projector” it can alleviate those symptoms and make the AR and VR realm accessible for more users.

Samsung invests in DigiLens XR glasses firm at valuation over $500M

DigiLens has raised funding from Samsung Electronics in a round that values the augmented reality smart glasses makers at more than $500 million.

Sunnyvale, California-based DigiLens did not say the exact amount it raised for the development of its extended reality glasses (XR), which will offer AR features, such as overlaying digital images on what you see.

DigiLens CEO Chris Pickett said in a previous interview with VentureBeat that the latest smart glasses are more advanced than models the company showed in 2019.

A system to control robotic arms based on augmented reality and a brain-computer interface

For people with motor impairments or physical disabilities, completing daily tasks and house chores can be incredibly challenging. Recent advancements in robotics, such as brain-controlled robotic limbs, have the potential to significantly improve their quality of life.

Researchers at Hebei University of Technology and other institutes in China have developed an innovative system for controlling robotic arms that is based on augmented reality (AR) and a . This system, presented in a paper published in the Journal of Neural Engineering, could enable the development of bionic or prosthetic arms that are easier for users to control.

“In recent years, with the development of robotic arms, brain science and information decoding technology, brain-controlled robotic arms have attained increasing achievements,” Zhiguo Luo, one of the researchers who carried out the study, told TechXplore. “However, disadvantages like poor flexibility restrict their widespread application. We aim to promote the lightweight and practicality of brain-controlled robotic arms.”

Facebook: Creating the Metaverse (Zuckerberg’s Master Plan)

Oculus Quest 2: https://amzn.to/3w0j6dO (Affiliate link)
Ray Ban Stories: https://amzn.to/3mv62tT (Affiliate link)

Timestamps:
00:00 Intro.
01:04: Brief History of Facebook.
04:47: VR & AR Today.
14:21: Mark Zuckerberg’s Master Plan.
23:19: Support Perhaps?

Links (In order of appearance):

October 28 2021: Mark Zuckerberg talking at Facebook Connect: [https://fb.watch/8X18pssy2q/]

October 23 2021: [https://youtu.be/99BnZ8js1_k?t=26] Reuters: What is the Metaverse?

July 23 2021: CNBC Television: [https://youtu.be/d5KzLgbb5Xo?t=260]

New Algorithms Give Digital Images More Realistic Color

In Optica, The Optical Society’s (OSA) journal for high impact research, Qiu and colleagues describe a new approach for digitizing color. It can be applied to cameras and displays — including ones used for computers, televisions and mobile devices — and used to fine-tune the color of LED lighting.

“Our new approach can improve today’s commercially available displays or enhance the sense of reality for new technologies such as near-eye-displays for virtual reality and augmented reality glasses,” said Jiyong Wang, a member of the PAINT research team. “It can also be used to produce LED lighting for hospitals, tunnels, submarines and airplanes that precisely mimics natural sunlight. This can help regulate circadian rhythm in people who are lacking sun exposure, for example.”