Menu

Blog

Archive for the ‘information science’ category: Page 30

Apr 4, 2024

Holographic Breakthrough: Scientists Create Full-Color 3D Holographic Displays with Ordinary Smartphone Screen

Posted by in categories: computing, holograms, information science, military, mobile phones

In science fiction, holograms are used for anything from basic communications to advanced military weaponry. In the real world, 3D holographic displays have yet to break through to everyday products and devices. That’s because creating holograms that look real and have significant fidelity requires laser emitters or other advanced pieces of optical equipment. This situation has stymied commercial development, as these components are complex and expensive.

More recently, research scientists were able to create realistic 3D holographic images without lasers by using a white chip-on-board light-emitting diode. Unfortunately, that method required two spatial light modulators to control the wave fronts of the emitted light, adding a prohibitive amount of complexity and cost.

Now, those same scientists say they have created a simpler, more cost-effective way to create realistic-looking 3D holographic displays using only one spatial light modulator and new software algorithms. The result is a simpler and cheaper method for creating holograms that an everyday technology like a smartphone screen can emit.

Apr 3, 2024

Neural feedback loops algorithms and consciousness

Posted by in categories: information science, neuroscience

Shared with Dropbox.

Apr 3, 2024

Classical optical neural network exhibits ‘quantum speedup’

Posted by in categories: information science, quantum physics, robotics/AI

In recent years, artificial intelligence technologies, especially machine learning algorithms, have made great strides. These technologies have enabled unprecedented efficiency in tasks such as image recognition, natural language generation and processing, and object detection, but such outstanding functionality requires substantial computational power as a foundation.

Mar 30, 2024

Novel quantum algorithm proposed for high-quality solutions to combinatorial optimization problems

Posted by in categories: information science, quantum physics, robotics/AI

Combinatorial optimization problems (COPs) have applications in many different fields such as logistics, supply chain management, machine learning, material design and drug discovery, among others, for finding the optimal solution to complex problems. These problems are usually very computationally intensive using classical computers and thus solving COPs using quantum computers has attracted significant attention from both academia and industry.

Mar 27, 2024

Transcerebral information coordination in directional hippocampus-prefrontal cortex network during working memory based on bimodal neural electrical signals

Posted by in categories: information science, neuroscience

Working memory (WM) is a kind of advanced cognitive function, which requires the participation and cooperation of multiple brain regions. Hippocampus and prefrontal cortex are the main responsible brain regions for WM. Exploring information coordination between hippocampus and prefrontal cortex during WM is a frontier problem in cognitive neuroscience. In this paper, an advanced information theory analysis based on bimodal neural electrical signals (local field potentials, LFPs and spikes) was employed to characterize the transcerebral information coordination across the two brain regions. Firstly, LFPs and spikes were recorded simultaneously from rat hippocampus and prefrontal cortex during the WM task by using multi-channel in vivo recording technique. Then, from the perspective of information theory, directional hippocampus-prefrontal cortex networks were constructed by using transfer entropy algorithm based on spectral coherence between LFPs and spikes. Finally, transcerebral coordination of bimodal information at the brain-network level was investigated during acquisition and performance of the WM task. The results show that the transfer entropy in directional hippocampus-prefrontal cortex networks is related to the acquisition and performance of WM. During the acquisition of WM, the information flow, local information transmission ability and information transmission efficiency of the directional hippocampus-prefrontal networks increase over learning days. During the performance of WM, the transfer entropy from the hippocampus to prefrontal cortex plays a leading role for bimodal information coordination across brain regions and hippocampus has a driving effect on prefrontal cortex. Furthermore, bimodal information coordination in the hippocampus → prefrontal cortex network could predict WM during the successful performance of WM.

Keywords: Bimodal neural electrical signals; Graph theory; Transcerebral information coordination; Transfer entropy; Working memory.

© The Author(s), under exclusive licence to Springer Nature B.V. 2022.

Mar 27, 2024

Machines Are on the Verge of Tackling Fermat’s Last Theorem—a Proof That Once Defied Them

Posted by in category: information science

Advanced algorithms are now deciphering what once was the domain of pure human intellect.

Mar 27, 2024

AI’s Learning Path: Surprising Uniformity Across Neural Networks

Posted by in categories: information science, robotics/AI

Summary: Neural networks, regardless of their complexity or training method, follow a surprisingly uniform path from ignorance to expertise in image classification tasks. Researchers found that neural networks classify images by identifying the same low-dimensional features, such as ears or eyes, debunking the assumption that network learning methods are vastly different.

This finding could pave the way for developing more efficient AI training algorithms, potentially reducing the significant computational resources currently required. The research, grounded in information geometry, hints at a more streamlined future for AI development, where understanding the common learning path of neural networks could lead to cheaper and faster training methods.

Mar 24, 2024

God’s Number Revealed: 20 Moves Proven Enough to Solve Any Rubik’s Cube Position

Posted by in categories: alien life, computing, information science, mathematics

Year 2010 😗😁


The world has waited with bated breath for three decades, and now finally a group of academics, engineers, and math geeks has discovered the number that explains life, the universe, and everything. That number is 20, and it’s the maximum number of moves it takes to solve a Rubik’s Cube.

Known as God’s Number, the magic number required about 35 CPU-years and a good deal of man-hours to solve. Why? Because there’s-1 possible positions of the cube, and the computer algorithm that finally cracked God’s Algorithm had to solve them all. (The terms God’s Number/Algorithm are derived from the fact that if God was solving a Cube, he/she/it would do it in the most efficient way possible. The Creator did not endorse this study, and could not be reached for comment.)

Continue reading “God’s Number Revealed: 20 Moves Proven Enough to Solve Any Rubik’s Cube Position” »

Mar 24, 2024

Bayesian neural networks using magnetic tunnel junction-based probabilistic in-memory computing

Posted by in categories: information science, particle physics, robotics/AI

Bayesian neural networks (BNNs) combine the generalizability of deep neural networks (DNNs) with a rigorous quantification of predictive uncertainty, which mitigates overfitting and makes them valuable for high-reliability or safety-critical applications. However, the probabilistic nature of BNNs makes them more computationally intensive on digital hardware and so far, less directly amenable to acceleration by analog in-memory computing as compared to DNNs. This work exploits a novel spintronic bit cell that efficiently and compactly implements Gaussian-distributed BNN values. Specifically, the bit cell combines a tunable stochastic magnetic tunnel junction (MTJ) encoding the trained standard deviation and a multi-bit domain-wall MTJ device independently encoding the trained mean. The two devices can be integrated within the same array, enabling highly efficient, fully analog, probabilistic matrix-vector multiplications. We use micromagnetics simulations as the basis of a system-level model of the spintronic BNN accelerator, demonstrating that our design yields accurate, well-calibrated uncertainty estimates for both classification and regression problems and matches software BNN performance. This result paves the way to spintronic in-memory computing systems implementing trusted neural networks at a modest energy budget.

The powerful ability of deep neural networks (DNNs) to generalize has driven their wide proliferation in the last decade to many applications. However, particularly in applications where the cost of a wrong prediction is high, there is a strong desire for algorithms that can reliably quantify the confidence in their predictions (Jiang et al., 2018). Bayesian neural networks (BNNs) can provide the generalizability of DNNs, while also enabling rigorous uncertainty estimates by encoding their parameters as probability distributions learned through Bayes’ theorem such that predictions sample trained distributions (MacKay, 1992). Probabilistic weights can also be viewed as an efficient form of model ensembling, reducing overfitting (Jospin et al., 2022). In spite of this, the probabilistic nature of BNNs makes them slower and more power-intensive to deploy in conventional hardware, due to the large number of random number generation operations required (Cai et al., 2018a).

Mar 24, 2024

Probabilistic Neural Computing with Stochastic Devices

Posted by in categories: information science, robotics/AI

The brain has effectively proven a powerful inspiration for the development of computing architectures in which processing is tightly integrated with memory, communication is event-driven, and analog computation can be performed at scale. These neuromorphic systems increasingly show an ability to improve the efficiency and speed of scientific computing and artificial intelligence applications. Herein, it is proposed that the brain’s ubiquitous stochasticity represents an additional source of inspiration for expanding the reach of neuromorphic computing to probabilistic applications. To date, many efforts exploring probabilistic computing have focused primarily on one scale of the microelectronics stack, such as implementing probabilistic algorithms on deterministic hardware or developing probabilistic devices and circuits with the expectation that they will be leveraged by eventual probabilistic architectures. A co-design vision is described by which large numbers of devices, such as magnetic tunnel junctions and tunnel diodes, can be operated in a stochastic regime and incorporated into a scalable neuromorphic architecture that can impact a number of probabilistic computing applications, such as Monte Carlo simulations and Bayesian neural networks. Finally, a framework is presented to categorize increasingly advanced hardware-based probabilistic computing technologies.

Keywords: magnetic tunnel junctions; neuromorphic computing; probabilistic computing; stochastic computing; tunnel diodes.

© 2022 The Authors. Advanced Materials published by Wiley-VCH GmbH.

Page 30 of 322First2728293031323334Last