Menu

Blog

Archive for the ‘information science’ category: Page 22

Jun 4, 2024

New model suggests partner anti-universe could explain accelerated expansion without the need for dark energy

Posted by in categories: cosmology, information science, quantum physics

The accelerated expansion of the present universe, believed to be driven by a mysterious dark energy, is one of the greatest puzzles in our understanding of the cosmos. The standard model of cosmology called Lambda-CDM, explains this expansion as a cosmological constant in Einstein’s field equations. However, the cosmological constant itself lacks a complete theoretical understanding, particularly regarding its very small positive value.

Jun 2, 2024

Memristor-based adaptive neuromorphic perception in unstructured environments

Posted by in categories: information science, robotics/AI, transportation

Differential neuromorphic computing, as a memristor-assisted perception method, holds the potential to enhance subsequent decision-making and control processes. Compared with conventional technologies, both the PID control approach and the proposed differential neuromorphic computing share a fundamental principle of smartly adjusting outputs in response to feedback, they diverge significantly in the data manipulation process (Supplementary Discussion 12 and Fig. S26); our method leverages the nonlinear characteristics of the memristor and a dynamic selection scheme to execute more complex data manipulation than linear coefficient-based error correction in PID. Additionally, the intrinsic memory function of memristors in our system enables real-time adaptation to changing environments. This represents a significant advantage compared to the static parameter configuration of PID systems. To perform similar adaptive control functions in tactile experiments, the von Neumann architecture follows a multi-step process involving several data movements: 1. Input data about the piezoresistive film state is transferred to the system memory via an I/O interface. 2. This sensory data is then moved from the memory to the cache. 3. Subsequently, it is forwarded to the Arithmetic Logic Unit (ALU) and waits for processing.4. Historical tactile information is also transferred from the memory to the cache unless it is already present. 5. This historical data is forwarded to the ALU. 6. ALU calculates the current sensory and historical data and returns the updated historical data to the cache. In contrast, our memristor-based approach simplifies this process, reducing it to three primary steps: 1. ADC reads data from the piezoresistive film. 2. ADC reads the current state of the memristor, which represents the historical tactile stimuli. 3. DAC, controlled by FPGA logic, updates the memristor state based on the inputs. This process reduces the costs of operation and enhances data processing efficiency.

In real-world settings, robotic tactile systems are required to elaborate large amounts of tactile data and respond as quickly as possible, taking less than 100 ms, similar to human tactile systems58,59. The current state-of-the-art robotics tactile technologies are capable of elaborating sudden changes in force, such as slip detection, at millisecond levels (from 500 μs to 50 ms)59,60,61,62, and the response time of our tactile system has also reached this detection level. For the visual processing, suppose a vehicle travels 40 km per hour in an urban area and wants control effective for every 1 m. In that case, the requirement translates a maximum allowable response time of 90 ms for the entire processing pipeline, which includes sensors, operating systems, middleware, and applications such as object detection, prediction, and vehicle control63,64. When incorporating our proposed memristor-assisted method with conventional camera systems, the additional time delay includes the delay from filter circuits (less than 1 ms) and the switching time for the memristor device, which ranges from nanoseconds (ns) to even picoseconds (ps)21,65,66,67. Compared to the required overall response time of the pipeline, these additions are negligible, demonstrating the potential of our method application in real-world driving scenarios68. Although our memristor-based perception method meets the response time requirement for described scenarios, our approach faces several challenges that need to be addressed for real-world applications. Apart from the common issues such as variability in device performance and the nonlinear dynamics of memristive responses, our approach needs to overcome the following challenges:

Currently, the modulation voltage applied to memristors is preset based on the external sensory feature, and the control algorithm is based on hard threshold comparison. This setting lacks the flexibility required for diverse real-world environments where sensory inputs and required responses can vary significantly. Therefore, it is crucial to develop a more automatic memristive modulation method along with a control algorithm that can dynamically adjust based on varying application scenarios.

Jun 2, 2024

A 3D ray traced biological neural network learning model

Posted by in categories: biological, information science, robotics/AI

In artificial neural networks, many models are trained for a narrow task using a specific dataset. They face difficulties in solving problems that include dynamic input/output data types and changing objective functions. Whenever the input/output tensor dimension or the data type is modified, the machine learning models need to be rebuilt and subsequently retrained from scratch. Furthermore, many machine learning algorithms that are trained for a specific objective, such as classification, may perform poorly at other tasks, such as reinforcement learning or quantification.

Even if the input/output dimensions and the objective functions remain constant, the algorithms do not generalize well across different datasets. For example, a neural network trained on classifying cats and dogs does not perform well on classifying humans and horses despite both of the datasets having the exact same image input1. Moreover, neural networks are highly susceptible to adversarial attacks2. A small deviation from the training dataset, such as changing one pixel, could cause the neural network to have significantly worse performance. This problem is known as the generalization problem3, and the field of transfer learning can help to solve it.

Transfer learning4,5,6,7,8,9,10 solves the problems presented above by allowing knowledge transfer from one neural network to another. A common way to use supervised transfer learning is obtaining a large pre-trained neural network and retraining it for a different but closely related problem. This significantly reduces training time and allows the model to be trained on a less powerful computer. Many researchers used pre-trained neural networks such as ResNet-5011 and retrained them to classify malicious software12,13,14,15. Another application of transfer learning is tackling the generalization problem, where the testing dataset is completely different from the training dataset. For example, every human has unique electroencephalography (EEG) signals due to them having distinctive brain structures. Transfer learning solves the generalization problem by pretraining on a general population EEG dataset and retraining the model for a specific patient16,17,18,19,20. As a result, the neural network is dynamically tailored for a specific person and can interpret their specific EEG signals properly. Labeling large datasets by hand is tedious and time-consuming. In semi-supervised transfer learning21,22,23,24, either the source dataset or the target dataset is unlabeled. That way, the neural networks can self-learn which pieces of information to extract and process without many labels.

May 31, 2024

New Machine Learning Algorithm Promises Advances in Computing

Posted by in categories: information science, robotics/AI

Digital twin models may enhance future autonomous systems.

Systems controlled by next-generation computing algorithms could give rise to better and more efficient machine learning products, a new study suggests.

Using machine learning tools to create a digital twin, or a virtual copy, of an electronic circuit that exhibits chaotic behavior, researchers found that they were successful at predicting how it would behave and using that information to control it.

May 31, 2024

Dan Dennett: Sir Roger Penrose Is WRONG About Human Consciousness!

Posted by in categories: information science, neuroscience, quantum physics

Join my mailing list https://briankeating.com/list to win a real 4 billion year old meteorite! All.edu emails in the USA 🇺🇸 will WIN!

Previous guest and friend of the show, Sir Roger Penrose, argues that human consciousness is not algorithmic and, therefore, cannot be modeled by Turing machines. In fact, he believes in a quantum mechanical understanding of human consciousness. However, as with any issue related to human consciousness, many disagree with him. One of his opponents is Daniel Dennett, with whom I recently had the pleasure of talking. Tune in to find out why Dennett thinks Penrose is wrong!

Continue reading “Dan Dennett: Sir Roger Penrose Is WRONG About Human Consciousness!” »

May 30, 2024

Secrets from the Algorithm: Google Search’s Internal Engineering Documentation Has Leaked

Posted by in categories: engineering, information science

Learn what you always wish you knew about Google’s algorithms.

May 25, 2024

Generative AI: The New Lifeline To Overwhelmed Healthcare Systems

Posted by in categories: biotech/medical, health, information science, robotics/AI

As the world’s population continues to grow and age, the healthcare system in different geographies is inching closer to the brink of collapse. According to the World Health Organization, the current number of health workers, including physicians, radiologists, and other professionals, is not sufficient to handle the rising caseload. On top of it, the increased stress and burnout stemming from the surge in cases is pushing many to exit the field, further reducing the number of practicing workers. Becker Health estimates show that nearly 72,000 American physicians left the workforce between 2021 and 2022, and some 30,000 who will join the workforce will not be enough to meet the growing demand.

At the core, both these challenges – the rising caseload and dwindling workforce – are leaving one major impact: diminished quality of patient care. This is where the much talked about generative AI can come in, saving healthcare staffers valuable time and resources and enabling them to focus on enhancing clinical outcomes.

First off, it’s important to understand AI is not new in healthcare. Organizations have been experimenting with predictive and computer vision algorithms for a while now, most notably to forecast the success of treatments and diagnose dangerous diseases earlier than humans. However, when it comes to generative AI, things are still pretty fresh, given the technology came to the forefront just a couple of years ago with the launch of ChatGPT. Gen AI models use neural networks to identify patterns and structures in existing data and generate new content such as text and images. They are applicable across sectors, including healthcare – where organizations cumulatively generate about 300 petabytes of data every single day.

May 25, 2024

A Warp Drive Breakthrough Inches a Tiny Bit Closer to ‘Star Trek’

Posted by in categories: information science, physics, space travel

While tantalizing, Alcubierre’s design has a fatal flaw. To provide the necessary distortions of spacetime, the spacecraft must contain some form of exotic matter, typically regarded as matter with negative mass. Negative mass has some conceptual problems that seem to defy our understanding of physics, like the possibility that if you kick a ball that weighs negative 5 kilograms, it will go flying backwards, violating conservation of momentum. Plus, nobody has ever seen any object with negative mass existing in the real universe, ever.

These problems with negative mass have led physicists to propose various versions of “energy conditions” as supplements to general relativity. These aren’t baked into relativity itself, but add-ons needed because general relativity allows things like negative mass that don’t appear to exist in our universe—these energy conditions keep them out of relativity’s equations. They’re scientists’ response to the unsettling fact that vanilla GR allows for things like superluminal motion, but the rest of the universe doesn’t seem to agree.

The energy conditions aren’t experimentally or observationally proven, but they are statements that concord with all observations of the universe, so most physicists take them rather seriously. And until recently, physicists have viewed those energy conditions as making it absolutely 100 percent clear that you can’t build a warp drive, even if you really wanted to.

May 25, 2024

Researchers identify best algorithms to optimize performance of functionally graded materials

Posted by in categories: computing, engineering, information science

A study from Japan published in the International Journal of Computer Aided Engineering and Technology reveals a way to optimize the composition of functionally graded materials (FGMs). FGMs are advanced composite materials with a gradual variation in composition and properties across their volume, designed to optimize performance under specific loading conditions.

May 25, 2024

Can ChatGPT Mimic Theory of Mind? Psychology Is Probing AI’s Inner Workings

Posted by in categories: information science, robotics/AI

Scientists pitted OpenAI and Meta chatbots against over 1,900 humans. The algorithms bested participants in areas like the detection of irony and faux pas.

Page 22 of 322First1920212223242526Last