БЛОГ

Archive for the ‘information science’ category: Page 27

Jun 5, 2024

A Safer Future for AI with Stronger Algorithms

Posted by in categories: cybercrime/malcode, information science, robotics/AI

This post is also available in: עברית (Hebrew)

AI technology is spreading quickly throughout many different industries, and its integration depends on users’ trust and safety concerns. This matter becomes complicated when the algorithms powering AI-based tools are vulnerable to cyberattacks that could have detrimental results.

Dr. David P. Woodruff from Carnegie Mellon University and Dr. Samson Zhou from Texas A&M University are working to strengthen the algorithms used by big data AI models against attacks.

Jun 4, 2024

CMSP series of lectures on “Topology and dynamics of higher-order networks”: lecture 3

Posted by in categories: computing, information science, mathematics, quantum physics

ICTP lectures “Topology and dynamics of higher-order networks”

- Network topology: 1 https://youtube.com/watch?v=mbmsv9RS3Pc

Continue reading “CMSP series of lectures on ‘Topology and dynamics of higher-order networks’: lecture 3” »

Jun 4, 2024

Scientists spot 60 stars appearing to show signs of alien power plants

Posted by in categories: alien life, information science, robotics/AI

I don’t know if this true but it definitely could be as most civilizations are probably more advanced than the earth.


A survey of five million distant solar systems, aided by ‘neural network’ algorithms, has discovered 60 stars that appear to be surrounded by giant alien power plants.

Seven of the stars — so-called M-dwarf stars that range between 60 percent and 8 percent the size of our sun — were recorded giving off unexpectedly high infrared ‘heat signatures,’ according to the astronomers.

Continue reading “Scientists spot 60 stars appearing to show signs of alien power plants” »

Jun 4, 2024

New model suggests partner anti-universe could explain accelerated expansion without the need for dark energy

Posted by in categories: cosmology, information science, quantum physics

The accelerated expansion of the present universe, believed to be driven by a mysterious dark energy, is one of the greatest puzzles in our understanding of the cosmos. The standard model of cosmology called Lambda-CDM, explains this expansion as a cosmological constant in Einstein’s field equations. However, the cosmological constant itself lacks a complete theoretical understanding, particularly regarding its very small positive value.

Jun 2, 2024

Memristor-based adaptive neuromorphic perception in unstructured environments

Posted by in categories: information science, robotics/AI, transportation

Differential neuromorphic computing, as a memristor-assisted perception method, holds the potential to enhance subsequent decision-making and control processes. Compared with conventional technologies, both the PID control approach and the proposed differential neuromorphic computing share a fundamental principle of smartly adjusting outputs in response to feedback, they diverge significantly in the data manipulation process (Supplementary Discussion 12 and Fig. S26); our method leverages the nonlinear characteristics of the memristor and a dynamic selection scheme to execute more complex data manipulation than linear coefficient-based error correction in PID. Additionally, the intrinsic memory function of memristors in our system enables real-time adaptation to changing environments. This represents a significant advantage compared to the static parameter configuration of PID systems. To perform similar adaptive control functions in tactile experiments, the von Neumann architecture follows a multi-step process involving several data movements: 1. Input data about the piezoresistive film state is transferred to the system memory via an I/O interface. 2. This sensory data is then moved from the memory to the cache. 3. Subsequently, it is forwarded to the Arithmetic Logic Unit (ALU) and waits for processing.4. Historical tactile information is also transferred from the memory to the cache unless it is already present. 5. This historical data is forwarded to the ALU. 6. ALU calculates the current sensory and historical data and returns the updated historical data to the cache. In contrast, our memristor-based approach simplifies this process, reducing it to three primary steps: 1. ADC reads data from the piezoresistive film. 2. ADC reads the current state of the memristor, which represents the historical tactile stimuli. 3. DAC, controlled by FPGA logic, updates the memristor state based on the inputs. This process reduces the costs of operation and enhances data processing efficiency.

In real-world settings, robotic tactile systems are required to elaborate large amounts of tactile data and respond as quickly as possible, taking less than 100 ms, similar to human tactile systems58,59. The current state-of-the-art robotics tactile technologies are capable of elaborating sudden changes in force, such as slip detection, at millisecond levels (from 500 μs to 50 ms)59,60,61,62, and the response time of our tactile system has also reached this detection level. For the visual processing, suppose a vehicle travels 40 km per hour in an urban area and wants control effective for every 1 m. In that case, the requirement translates a maximum allowable response time of 90 ms for the entire processing pipeline, which includes sensors, operating systems, middleware, and applications such as object detection, prediction, and vehicle control63,64. When incorporating our proposed memristor-assisted method with conventional camera systems, the additional time delay includes the delay from filter circuits (less than 1 ms) and the switching time for the memristor device, which ranges from nanoseconds (ns) to even picoseconds (ps)21,65,66,67. Compared to the required overall response time of the pipeline, these additions are negligible, demonstrating the potential of our method application in real-world driving scenarios68. Although our memristor-based perception method meets the response time requirement for described scenarios, our approach faces several challenges that need to be addressed for real-world applications. Apart from the common issues such as variability in device performance and the nonlinear dynamics of memristive responses, our approach needs to overcome the following challenges:

Currently, the modulation voltage applied to memristors is preset based on the external sensory feature, and the control algorithm is based on hard threshold comparison. This setting lacks the flexibility required for diverse real-world environments where sensory inputs and required responses can vary significantly. Therefore, it is crucial to develop a more automatic memristive modulation method along with a control algorithm that can dynamically adjust based on varying application scenarios.

Jun 2, 2024

A 3D ray traced biological neural network learning model

Posted by in categories: biological, information science, robotics/AI

In artificial neural networks, many models are trained for a narrow task using a specific dataset. They face difficulties in solving problems that include dynamic input/output data types and changing objective functions. Whenever the input/output tensor dimension or the data type is modified, the machine learning models need to be rebuilt and subsequently retrained from scratch. Furthermore, many machine learning algorithms that are trained for a specific objective, such as classification, may perform poorly at other tasks, such as reinforcement learning or quantification.

Even if the input/output dimensions and the objective functions remain constant, the algorithms do not generalize well across different datasets. For example, a neural network trained on classifying cats and dogs does not perform well on classifying humans and horses despite both of the datasets having the exact same image input1. Moreover, neural networks are highly susceptible to adversarial attacks2. A small deviation from the training dataset, such as changing one pixel, could cause the neural network to have significantly worse performance. This problem is known as the generalization problem3, and the field of transfer learning can help to solve it.

Transfer learning4,5,6,7,8,9,10 solves the problems presented above by allowing knowledge transfer from one neural network to another. A common way to use supervised transfer learning is obtaining a large pre-trained neural network and retraining it for a different but closely related problem. This significantly reduces training time and allows the model to be trained on a less powerful computer. Many researchers used pre-trained neural networks such as ResNet-5011 and retrained them to classify malicious software12,13,14,15. Another application of transfer learning is tackling the generalization problem, where the testing dataset is completely different from the training dataset. For example, every human has unique electroencephalography (EEG) signals due to them having distinctive brain structures. Transfer learning solves the generalization problem by pretraining on a general population EEG dataset and retraining the model for a specific patient16,17,18,19,20. As a result, the neural network is dynamically tailored for a specific person and can interpret their specific EEG signals properly. Labeling large datasets by hand is tedious and time-consuming. In semi-supervised transfer learning21,22,23,24, either the source dataset or the target dataset is unlabeled. That way, the neural networks can self-learn which pieces of information to extract and process without many labels.

May 31, 2024

New Machine Learning Algorithm Promises Advances in Computing

Posted by in categories: information science, robotics/AI

Digital twin models may enhance future autonomous systems.

Systems controlled by next-generation computing algorithms could give rise to better and more efficient machine learning products, a new study suggests.

Using machine learning tools to create a digital twin, or a virtual copy, of an electronic circuit that exhibits chaotic behavior, researchers found that they were successful at predicting how it would behave and using that information to control it.

May 31, 2024

Dan Dennett: Sir Roger Penrose Is WRONG About Human Consciousness!

Posted by in categories: information science, neuroscience, quantum physics

Join my mailing list https://briankeating.com/list to win a real 4 billion year old meteorite! All.edu emails in the USA 🇺🇸 will WIN!

Previous guest and friend of the show, Sir Roger Penrose, argues that human consciousness is not algorithmic and, therefore, cannot be modeled by Turing machines. In fact, he believes in a quantum mechanical understanding of human consciousness. However, as with any issue related to human consciousness, many disagree with him. One of his opponents is Daniel Dennett, with whom I recently had the pleasure of talking. Tune in to find out why Dennett thinks Penrose is wrong!

Continue reading “Dan Dennett: Sir Roger Penrose Is WRONG About Human Consciousness!” »

May 30, 2024

Secrets from the Algorithm: Google Search’s Internal Engineering Documentation Has Leaked

Posted by in categories: engineering, information science

Learn what you always wish you knew about Google’s algorithms.

May 25, 2024

Generative AI: The New Lifeline To Overwhelmed Healthcare Systems

Posted by in categories: biotech/medical, health, information science, robotics/AI

As the world’s population continues to grow and age, the healthcare system in different geographies is inching closer to the brink of collapse. According to the World Health Organization, the current number of health workers, including physicians, radiologists, and other professionals, is not sufficient to handle the rising caseload. On top of it, the increased stress and burnout stemming from the surge in cases is pushing many to exit the field, further reducing the number of practicing workers. Becker Health estimates show that nearly 72,000 American physicians left the workforce between 2021 and 2022, and some 30,000 who will join the workforce will not be enough to meet the growing demand.

At the core, both these challenges – the rising caseload and dwindling workforce – are leaving one major impact: diminished quality of patient care. This is where the much talked about generative AI can come in, saving healthcare staffers valuable time and resources and enabling them to focus on enhancing clinical outcomes.

First off, it’s important to understand AI is not new in healthcare. Organizations have been experimenting with predictive and computer vision algorithms for a while now, most notably to forecast the success of treatments and diagnose dangerous diseases earlier than humans. However, when it comes to generative AI, things are still pretty fresh, given the technology came to the forefront just a couple of years ago with the launch of ChatGPT. Gen AI models use neural networks to identify patterns and structures in existing data and generate new content such as text and images. They are applicable across sectors, including healthcare – where organizations cumulatively generate about 300 petabytes of data every single day.

Page 27 of 328First2425262728293031Last