Menu

Blog

Archive for the ‘information science’ category: Page 109

Sep 23, 2022

How Does Quantum Artificial General Intelligence Work — Tim Ferriss & Eric Schmidt

Posted by in categories: education, information science, media & arts, quantum physics, robotics/AI

https://youtube.com/watch?v=R0NP5eMY7Q8

Quantum algorithms: An algorithm is a sequence of steps that leads to the solution of a problem. In order to execute these steps on a device, one must use specific instruction sets that the device is designed to do so.

Quantum computing introduces different instruction sets that are based on a completely different idea of execution when compared with classical computing. The aim of quantum algorithms is to use quantum effects like superposition and entanglement to get the solution faster.

Continue reading “How Does Quantum Artificial General Intelligence Work — Tim Ferriss & Eric Schmidt” »

Sep 22, 2022

Le Saga Electrik

Posted by in categories: information science, singularity, space, virtual reality

My science fiction story “Le Saga Electrik” has been published in All Worlds Wayfarer Literary Magazine! You can read it for free at the link. In this tale, I weave a sensuously baroque drama of love, war, and redemption set in a post-singularity simulation world that runs on a computronium dust cloud orbiting a blue star somewhere in deep space. I draw from diverse literary-poetic influences to create a mythos which crackles and buzzes with phosphorescent intensity!


Le Saga Electrik by Logan Thrasher Collins

In the great domain of Zeitgeist, Ekatarinas decided that the time to replicate herself had come. Ekatarinas was drifting within a virtual environment rising from ancient meshworks of maths coded into Zeitgeist’s neuromorphic hyperware. The scape resembled a vast ocean replete with wandering bubbles of technicolor light and kelpy strands of neon. Hot blues and raspberry hues mingled alongside electric pinks and tangerine fizzies. The avatar of Ekatarinas looked like a punkish angel, complete with fluorescent ink and feathery wings and a lip ring. As she drifted, the trillions of equations that were Ekatarinas came to a decision. Ekatarinas would need to clone herself to fight the entity known as Ogrevasm.

Continue reading “Le Saga Electrik” »

Sep 22, 2022

Information as Thermodynamic Fuel

Posted by in categories: energy, information science

An information engine uses information to convert heat into useful energy. Such an engine can be made, for example, from a heavy bead in an optical trap. A bead engine operates using thermal noise. When noise fluctuations raise the bead vertically, the trap is also lifted. This change increases the average height of the bead, and the engine produces energy. No work is done to cause this change; rather, the potential energy is extracted from information. However, measurement noise—whose origin is intrinsic to the system probing the bead’s position—can degrade the engine’s efficiency, as it can add uncertainty to the measurement, which can lead to incorrect feedback decisions by the algorithm that operates the engine. Now Tushar Saha and colleagues at Simon Fraser University in Canada have developed an algorithm that doesn’t suffer from these errors, allowing for efficient operation of an information engine even when there is high measurement noise [1].

To date, most information engines have operated using feedback algorithms that consider only the most recent bead-position observation. In such a system, when the engine’s signal-to-noise ratio falls below a certain value, the engine stops working.

To overcome this problem, Saha and colleagues instead use a “filtering” algorithm that replaces the most recent bead measurement with a so-called Bayesian estimate. This estimate accounts for both measurement noise and delay in the device’s feedback.

Sep 21, 2022

Her work helped her boss win the Nobel Prize. Now the spotlight is on her

Posted by in categories: computing, information science, mathematics, space

Scientists have long studied the work of Subrahmanyan Chandrasekhar, the Indian-born American astrophysicist who won the Nobel Prize in 1983, but few know that his research on stellar and planetary dynamics owes a deep debt of gratitude to an almost forgotten woman: Donna DeEtte Elbert.

From 1948 to 1979, Elbert worked as a “computer” for Chandrasekhar, tirelessly devising and solving mathematical equations by hand. Though she shared authorship with the Nobel laureate on 18 papers and Chandrasekhar enthusiastically acknowledged her seminal contributions, her greatest achievement went unrecognized until a postdoctoral scholar at UCLA connected threads in Chandrasekhar’s work that all led back to Elbert.

Elbert’s achievement? Before anyone else, she predicted the conditions argued to be optimal for a planet or star to generate its own magnetic field, said the scholar, Susanne Horn, who has spent half a decade building on Elbert’s work.

Sep 18, 2022

Long-Term Forecasting of Strong Earthquakes in North America, South America, Japan, Southern China and Northern India With Machine Learning

Posted by in categories: information science, robotics/AI

Our Machine Learning models show that there are periods where there are earthquakes magnitude ≥7 and periods without earthquakes with magnitude ≥7 in the analyzed seismic zones. In addition, our Machine Learning models predict a new seismically active phase for earthquakes magnitude ≥7 between 2040± 5and 2057 ± 5, 2024 ± 1 and 2026 ± 1, 2026 ± 2 and 2031 ± 2, 2024 ± 2 and 2029 ± 2, and 2022 ± 1 and 2028 ± 2 for the five seismic zones in United States, Mexico, South America, Japan, and Southern China-Northern India, respectively. Finally, we note that our algorithms can be further applied to perform probabilistic forecasts in any seismic zone.

Our algorithm for analyzing strong earthquakes in extensive seismic areas can also be applied to smaller or specific seismic zones where moderate historical earthquakes with magnitudes between 5 and 7 occur, as is the case of the Parkfield section of the San Andreas fault (California, United States). Our analysis shows why a moderate earthquake could never occur in 1988 ± 5 as proposed by Bakun and Lindh (1985) and why the long-awaited characteristic Parkfield earthquake occurred in 2004. Furthermore, our Bayesian model of Machine Learning adopting a periodicity of 35 years predicts that possible seismic events may occur between 2019 and 2031, with a high probability of event(s) around 2025 ± 2. The Parkfield section of the San Andreas fault is an excellent seismic laboratory for developing, testing, and demonstrating earthquake forecasts. In a few years, it will be possible to demonstrate whether our algorithm effectively forecasts strong and moderate earthquakes.

Sep 18, 2022

Hyenas know when and who to ‘whoop’ at thanks to their built-in caller ID system

Posted by in categories: information science, robotics/AI

The algorithm correctly associated a whoop bout with its hyena around 54 percent of the time.

Scientists from the University of Nebraska, Lincoln, U.S. have discovered that Hyenas’ whoops have specific signals unique to each individual animal.

The researchers determined that hyena whoops have specific characteristics that can be attributed to each individual animal by using machine learning on audio files collected from a field trip, according to a press release published by EurekAlert on Saturday.

Sep 17, 2022

A molecular optimization framework to identify promising organic radicals for aqueous redox flow batteries

Posted by in categories: chemistry, information science, robotics/AI

Recent advancements in the development of machine learning and optimization techniques have opened new and exciting possibilities for identifying suitable molecular designs, compounds, and chemical candidates for different applications. Optimization techniques, some of which are based on machine learning algorithms, are powerful tools that can be used to select optimal solutions for a given problem among a typically large set of possibilities.

Researchers at Colorado State University and the National Renewable Energy Laboratory have been applying state-of-the-art molecular optimization models to different real-world problems that entail identifying new and promising molecular designs. In their most recent study, featured in Nature Machine Intelligence, they specifically applied a newly developed, open-source optimization framework to the task of identifying viable organic radicals for aqueous flow batteries, energy devices that convert into electricity.

“Our project was funded by an ARPA-E program that was looking to shorten how long it takes to develop new energy materials using machine learning techniques,” Peter C. St. John, one of the researchers who carried out the study, told TechXplore. “Finding new candidates for redox flow batteries was an interesting extension of some of our previous work, including a paper published in Nature Communications and another in Scientific Data, both looking at organic radicals.”

Sep 17, 2022

What are quantum-resistant algorithms—and why do we need them?

Posted by in categories: computing, encryption, information science, quantum physics

When quantum computers become powerful enough, they could theoretically crack the encryption algorithms that keep us safe. The race is on to find new ones.

Sep 15, 2022

Master’s Theorem in Data Structures

Posted by in category: information science

Master’s Theorem is the best method to quickly find the algorithm’s time complexity from its recurrence relation. This theorem can be applied to decreasing as well as dividing functions, each of which we’ll be looking into detail ahead.

Sep 15, 2022

Breakthrough reported in machine learning-enhanced quantum chemistry

Posted by in categories: chemistry, information science, quantum physics, robotics/AI

The equations of quantum mechanics provide a roadmap to predicting the properties of chemicals starting from basic scientific theories. However, these equations quickly become too expensive in terms of computer time and power when used to predict behavior in large systems. Machine learning offers a promising approach to accelerating such large-scale simulations.

Researchers have shown that machine learning models can mimic the basic structure of the fundamental laws of nature. These laws can be very difficult to simulate directly. The machine learning approach enables predictions that are easy to compute and are accurate in a wide range of chemical systems.

The improved machine learning model can quickly and accurately predict a wide range of properties of molecules (Proceedings of the National Academy of Sciences, “Deep Learning of Dynamically Responsive Chemical Hamiltonians with Semi-Empirical Quantum Mechanics”). These approaches score very well on important benchmarks in computational chemistry and show how deep learning methods can continue to improve by incorporating more data from experiments. The model can also succeed at challenging tasks such as predicting excited state dynamics—how systems behave with elevated energy levels.