Toggle light / dark theme

Scientists say they have developed a new AI-assisted model of a digital twin with the ability to adapt and control the physical machine in real time.

The discovery, reported in the journal IEEE Access, adds a new dimension to the digital copies of real-world machines, like robots, drones, or even autonomous cars, according to the authors.

Digital twins are exact replicas of things in the physical world. They are likened to video game versions of real machines with which they digitally twin, and are constantly updated with real-time data.

Imagine a large city recovering from a devastating hurricane. Roads are flooded, the power is down, and local authorities are overwhelmed. Emergency responders are doing their best, but the chaos is massive.

AI-controlled drones survey the damage from above, while process and data from sensors on the ground and air to identify which neighborhoods are most vulnerable.

Meanwhile, AI-equipped robots are deployed to deliver food, water and into areas that human responders can’t reach. Emergency teams, guided and coordinated by AI and the insights it produces, are able to prioritize their efforts, sending rescue squads where they’re needed most.

In today’s AI news, Investor interest in AI coding assistants is exploding. Anysphere, the developer of AI-powered coding assistant Cursor, is in talks with venture capitalists to raise capital at a valuation of nearly $10 billion, Bloomberg reported. The round, if it transpires, would come about three months after Anysphere completed its previous fundraise of $100 million at a pre-money valuation of $2.5 billion.

And, there’s a new voice model in town, and it’s called Sesame. As he so often does, John Werner got a lot of information on this new technology from Nathaniel Whittemore at AI Daily Brief, where he covered interest in this conversational AI. Quoting Deedy Das of Menlo Ventures calling Sesame “the GPT-3 moment for voice,” Whittemore talked about what he called an “incredible explosion” of voice-based models happening now.

In other advancements, along with the new M4 MacBook Pro series Apple is releasing, the company is also quite proud of the new Mac mini. The Mac mini is arguably the more radical of the two. Apple’s diminutive computer has now received its first major design overhaul in 13 years. And this new tiny computer is the perfect machine for experimenting with and learning AI.

S biggest defense tech startups by valuation, raising $240 million at a $5.3 billion valuation in its latest round. Shield AI, the San Diego defense tech startup that builds drones and other AI-powered military systems, has raised a $240 million round at a $5.3 billion valuation, it announced today.” + In videos, while he hardly needs an introduction, few leaders have shaped the future of technology quite like Satya Nadella. He stepped into Microsoft’s top job at a catalytic moment—making bold bets on the cloud, embedding AI into the fabric of computing, all while staying true to Microsoft’s vision of becoming a “software factory.”

T just think, it delivers results. Manus excels at various tasks in work and life, getting everything done while you rest. + Then, join Boris Starkov and Anton Pidkuiko, the developers behind GibberLink, for a fireside chat with Luke Harries from ElevenLabs. On February 24, Georgi Gerganov, the creator of the GGwave protocol, showcased their demo at the ElevenLabs London hackathon on X, garnering attention from around the world—including Forbes, TechCrunch, and the entire developer community.

We close out with, Sam Witteveen looking at the latest release from Mistral AI, which is their Mistral OCR model. He looks at how it works and how it compares to other models, as well as how you can get started using it with code.

Thats all for today, but AI is moving fast — subscribe and follow for more Neural News.

Satellite-based optical remote sensing from missions such as ESA’s Sentinel-2 (S2) have emerged as valuable tools for continuously monitoring the Earth’s surface, thus making them particularly useful for quantifying key cropland traits in the context of sustainable agriculture [1]. Upcoming operational imaging spectroscopy satellite missions will have an improved capability to routinely acquire spectral data over vast cultivated regions, thereby providing an entire suite of products for agricultural system management [2]. The Copernicus Hyperspectral Imaging Mission for the Environment (CHIME) [3] will complement the multispectral Copernicus S2 mission, thus providing enhanced services for sustainable agriculture [4, 5]. To use satellite spectral data for quantifying vegetation traits, it is crucial to mitigate the absorption and scattering effects caused by molecules and aerosols in the atmosphere from the measured satellite data. This data processing step, known as atmospheric correction, converts top-of-atmosphere (TOA) radiance data into bottom-of-atmosphere (BOA) reflectance, and it is one of the most challenging satellite data processing steps e.g., [6, 7, 8]. Atmospheric correction relies on the inversion of an atmospheric radiative transfer model (RTM) leading to the obtaining of surface reflectance, typically through the interpolation of large precomputed lookup tables (LUTs) [9, 10]. The LUT interpolation errors, the intrinsic uncertainties from the atmospheric RTMs, and the ill posedness of the inversion of atmospheric characteristics generate uncertainties in atmospheric correction [11]. Also, usually topographic, adjacency, and bidirectional surface reflectance corrections are applied sequentially in processing chains, which can potentially accumulate errors in the BOA reflectance data [6]. Thus, despite its importance, the inversion of surface reflectance data unavoidably introduces uncertainties that can affect downstream analyses and impact the accuracy and reliability of subsequent products and algorithms, such as vegetation trait retrieval [12]. To put it another way, owing to the critical role of atmospheric correction in remote sensing, the accuracy of vegetation trait retrievals is prone to uncertainty when atmospheric correction is not properly performed [13].

Although advanced atmospheric correction schemes became an integral part of the operational processing of satellite missions e.g., [9,14,15], standardised exhaustive atmospheric correction schemes in drone, airborne, or scientific satellite missions remain less prevalent e.g., [16,17]. The complexity of atmospheric correction further increases when moving from multispectral to hyperspectral data, where rigorous atmospheric correction needs to be applied to hundreds of narrow contiguous spectral bands e.g., [6,8,18]. For this reason, and to bypass these challenges, several studies have instead proposed to infer vegetation traits directly from radiance data at the top of the atmosphere [12,19,20,21,22,23,24,25,26].

DARPA lifts the veil on concealed bio-weapons and astonishing drone technology 🤖🦾 To try everything Brilliant has to offer—free—for a full 30 days, visit http://brilliant.org/BeeyondIdeas/ The first 200 of you will get 20% off Brilliant’s annual premium subscription. 🪐

Beeyond Ideas follows the viewpoint of Harry, a human-AI synthesis from the 22nd century. Someday in 2123, he found a way to access the secret old database of information or the “2023 Internet” as we know it.

Follow Harry’s adventure by subscribing to this channel Want to support our production? Feel free to join our membership at https://youtube.com/watch?v=wMeOlJjEvSc&si=YQODBYXZ1-dq4Leh #AI #Robotics #ArtificialIntelligence #darpa.

Want to support our production? Feel free to join our membership at https://youtube.com/watch?v=wMeOlJjEvSc&si=YQODBYXZ1-dq4Leh.

#AI #Robotics #ArtificialIntelligence #darpa

Transparent aluminum oxide (TAlOx), a real material despite its sci-fi name, is incredibly hard and resistant to scratches, making it perfect for protective coatings on electronics, optical sensors, and solar panels. On the sci-fi show Star Trek, it is even used for starship windows and spacefaring aquariums.

Current methods of making TAlOx are expensive and complicated, requiring high-powered lasers, vacuum chambers, or large vats of dangerous acids. That may change thanks to research co-authored by Filipino scientists from the Ateneo de Manila University.

Instead of immersing entire sheets of metal into acidic solutions, the researchers applied microdroplets of acidic solution onto small aluminum surfaces and applied an . Just two volts of electricity—barely more than what’s found in a single AA household flashlight battery—was all that was needed to transform the metal into glass-like TAlOx.

In this video, we delve into The Future of Electronic Warfare, exploring how advancements in AI, drone swarms, and cyber integration are reshaping military strategies. Historically, electronic warfare (EW) began with basic communication interception in World War I and evolved through World War II with techniques like radar jamming. Today, we stand at the brink of a new era where technology significantly enhances operational capabilities.

The Evolution of Drone Swarms.

Recent developments have seen the emergence of AI-powered drone swarms, which offer unprecedented adaptability and efficiency on the battlefield. For instance, Thales’s COHESION demonstrator showcases how these swarms can operate autonomously, reducing the cognitive load on human operators while maintaining control during critical mission phases. Unlike traditional systems that require one operator per drone, these advanced systems leverage AI to allow multiple drones to work collaboratively, enhancing surveillance and attack capabilities across vast terrains.

Key features of ai-powered drone swarms.

Wide-Area Surveillance: Swarms can cover extensive areas, providing comprehensive monitoring and real-time situational awareness, ensuring no part of the terrain goes unmonitored.

Decentralized Coordination: Each drone operates autonomously while contributing to a collective intelligence network, allowing for effective mission execution even if individual drones are lost.