Archive for the ‘computing’ category: Page 825
Nov 19, 2015
Laser ‘freeze ray’ could change cooling biology labs, computer processors
Posted by Shailesh Prasad in category: computing
Laser cooling isn’t a new idea, but this is the first time it’s actually worked in real-world conditions.
Nov 19, 2015
Quantum Computers One Step Closer After Australian Breakthrough
Posted by Julius Garcia in categories: computing, quantum physics
Engineers from the University of New South Wales, Australia, have made an important breakthrough that brings quantum computers one step closer to reality.
The team created a quantum version of a standard computer code within a silicon chip. The discovery shows that it is possible to construct realistic and reliable quantum computers.
Quantum computers have the potential to solve problems much more quickly than any computer that exists today, as they combine the rules of informatics to phenomena of quantum mechanics that are not observed in everyday life. Namely, the principle of superposition, popularized by Schrödinger’s cat being both alive and dead, and entanglement.
Nov 19, 2015
Los Alamos National Laboratory Orders a 1000+ Qubit D-Wave 2X Quantum Computer
Posted by Andreas Matt in categories: computing, quantum physics, security
D-Wave Systems Inc., the world’s first quantum computing company, announced that Los Alamos National Laboratory will acquire and install the latest D-Wave quantum computer, the 1000+ qubit D-Wave 2X™ system. Los Alamos, a multidisciplinary research institution engaged in strategic science on behalf of national security, will lead a collaboration within the Department of Energy and with select university partners to explore the capabilities and applications of quantum annealing technology, consistent with the goals of the government-wide National Strategic Computing Initiative. The National Strategic Computing Initiative, created by executive order of President Barack Obama in late July, is intended “to maximize [the] benefits of high-performance computing (HPC) research, development, and deployment.”
“Eventually Moore’s Law (that predicted that the number of transistors on an integrated circuit would double every two years) will come to an end,” said John Sarrao, associate director for Theory, Simulation, and Computation at Los Alamos. “Dennard Scaling (that predicted that performance per watt of computing would grow exponentially at roughly the same rate) already has. Beyond these two observations lies the end of the current ‘conventional’ computing era, so new technologies and ideas are needed.”
“As conventional computers reach their limits in terms of scaling and performance per watt, we need to investigate new technologies to support our mission,” said Mark Anderson of the Laboratory’s Weapons Physics Directorate. “Researching and evaluating quantum annealing as the basis for new approaches to address intractable problems is an essential and powerful step, and will enable a new generation of forward thinkers to influence its evolution in a direction most beneficial to the nation.”
Nov 18, 2015
Nvidia unveils Pascal specifics — up to 16GB of VRAM, 1TB of bandwidth
Posted by Shailesh Prasad in category: computing
Nvidia shared some more details on its upcoming Pascal architecture for 2016 — the new GPU will offer 1TB/s of memory bandwidth and up to 16GB of VRAM.
Nov 17, 2015
Quantum Super Computer Detects and Self-corrects its Errors
Posted by Shailesh Prasad in categories: computing, quantum physics
A team of quantum physicists in Martinis Lab have come a step closer in creating the circuitry that would allow them to process super computing done by quantum computers. The revolution is promised by the new quantum bits (qubits) compared to the previously done classical computing. Qubits infuse the system with high levels of reliability and speed, thus building foundations for large scale superconducting quantum computers.
Till now computing has been done by classical methods in which the bits were either in states 0 or 1, but qubits exist at all the positions simultaneously, in different dimensions. This special property of being omnipresent is called ‘superpositioning’. However, one of the difficulties is keeping the qubits stable to reproduce same result each time. This superpositioning characteristic makes qubits prone to ‘flipping’, therefore making it difficult to work with.
Julian Kelly, graduate student researcher and co-lead author of a research paper that was published in the journal Nature said:
Nov 16, 2015
A network of artificial neurons learns to use human language
Posted by Shailesh Prasad in categories: computing, neuroscience
A computer simulation of a cognitive model entirely made up of artificial neurons learns to communicate through dialog starting from a state of tabula rasa —
A group of researchers from the University of Sassari (Italy) and the University of Plymouth (UK) has developed a cognitive model, made up of two million interconnected artificial neurons, able to learn to communicate using human language starting from a state of ‘tabula rasa’, only through communication with a human interlocutor. The model is called ANNABELL (Artificial Neural Network with Adaptive Behavior Exploited for Language Learning) and it is described in an article published in PLOS ONE. This research sheds light on the neural processes that underlie the development of language.
How does our brain develop the ability to perform complex cognitive functions, such as those needed for language and reasoning? This is a question that certainly we are all asking ourselves, to which the researchers are not yet able to give a complete answer. We know that in the human brain there are about one hundred billion neurons that communicate by means of electrical signals. We learned a lot about the mechanisms of production and transmission of electrical signals among neurons. There are also experimental techniques, such as functional magnetic resonance imaging, which allow us to understand which parts of the brain are most active when we are involved in different cognitive activities. But a detailed knowledge of how a single neuron works and what are the functions of the various parts of the brain is not enough to give an answer to the initial question.
Nov 16, 2015
Lessons from the PC video game industry — By Chris Dixon | Medium
Posted by Odette Bohr Dienel in categories: business, computing, economics, futurism, internet, media & arts
“The subtitle to this post is a variation of William Gibson’s famous remark: “The future is already here — it’s just not very evenly distributed.” An obvious follow up question is: if the future is already here, where can I find it?”
Nov 16, 2015
Graphene Could Give Us Sleek Night Vision Sensors
Posted by Shailesh Prasad in categories: computing, electronics, materials, mobile phones
Because of its unique chemical and physical properties, graphene has helped scientists design new gadgets from tiny computer chips to salt water filters. Now a team of researchers from MIT has found a new use for the 2D wonder material: in infrared sensors that could replace bulky night-vision goggles, or even add night vision capabilities to high-tech windshields or smartphone cameras. The study was published last week in Nano Letters.
Night vision technology picks up on infrared wavelengths, energy usually emitted in the form of heat that humans can’t see with the naked eye. Researchers have known for years that because of how it conducts electricity, graphene is an excellent infrared detector, and they wanted to see if they could create something less bulky than current night-vision goggles. These goggles rely on cryogenic cooling to reduce the amount of excess heat that might muddle the image. To create the sensor, the researchers integrated graphene with tiny silicon-based devices called MEMS. Then, they suspended this chip over an air pocket so that it picks up on incoming heat and eliminates the need for the cooling mechanisms found in other infrared-sensing devices. That signal is then transmitted to another part of the device that creates a visible image. When the researchers tested their sensor, they found that it clearly and successfully picked up the image of a human hand.
Nov 16, 2015
Researchers suggest that the universe could be a computer simulation: Is it possible?
Posted by Andreas Matt in categories: alien life, computing, robotics/AI, virtual reality
It‘s older, but interesting!
The year is 2050 and super-intelligent robots have emerged as the masters of Earth. Unfortunately, you have no idea of that fact because we are immersed in a computer simulation set decades ago. Everything you see and touch has now been created and programmed by machines that use mankind for their own benefit. This radical theory, demonstrated in numerous books and science fiction films, has been, and is currently regarded by science as possible; Moreover, scientists are taking this theory to a cosmic level and even believe that if only one extraterrestrial civilization in the universe go the technological level to “emulate” an entire “multiverse,” then even our probes and space telescopes, which are out there exploring the universe, belong to that “creepy simulation.”