Toggle light / dark theme

Tesla AI5 & AI6 Chips “Compressing Reality”?! What Did Elon See?!

Elon Musk has revealed Tesla’s new AI chips, AI5 and AI6, which will drive the company’s shift towards AI-powered services, enabling significant advancements in Full Self-Driving capabilities and potentially revolutionizing the self-driving car industry and beyond.

## Questions to inspire discussion.

Tesla’s AI Chip Advancements.

🚀 Q: What are the key features of Tesla’s AI5 and AI6 chips? A: Tesla’s AI5 and AI6 chips are inference-first, designed for high-throughput and efficient processing of AI models on devices like autos, Optimus, and Grok voice agents, being 40x faster than previous models.

💻 Q: How do Tesla’s AI5 and AI6 chips compare to previous models? A: Tesla’s AI5 chip is a 40x improvement over AI4, with 500 TOPS expanding to 5,000 TOPS, enabling excellent performance in full self-driving and Optimus humanoid robots.

🧠 Q: What is the significance of softmax in Tesla’s AI5 chip? A: AI5 is designed to run softmax natively in a few steps, unlike AI4 which relies on CPU and runs softmax in 40 steps in emulation mode.

AI to integrate bulk multi-omics data for precision oncology

“Cancer and other complex diseases arise from the interplay of various biological factors, for example, at the DNA, RNA, and protein levels,” explains the author. Characteristic changes at these levels — such as the amount of HER2 protein produced in breast or stomach cancer — are often recorded, but typically not yet analyzed in conjunction with all other therapy-relevant factors.

This is where Flexynesis comes in. “Comparable tools so far have often been either difficult to use, or only useful for answering certain questions,” says the author. “Flexynesis, by contrast, can answer various medical questions at the same time: for example, what type of cancer is involved, what drugs are particularly effective in this case, and how these will affect the patient’s chances of survival.” The tool also helps identify suitable biomarkers for diagnosis and prognosis, or — if metastases of unknown origin are discovered — to identify the primary tumor. “This makes it easier to develop comprehensive and personalized treatment strategies for all kinds of cancer patients,” says the author.


Nearly 50 new cancer therapies are approved every year. This is good news. “But for patients and their treating physicians, it is becoming increasingly difficult to keep track and to select the treatment methods from which the people affected — each with their very individual tumor characteristics — will benefit the most,” says the senior author. The researcher has been working for some time on developing tools that use artificial intelligence to make more precise diagnoses and that also determine the best form of therapy tailored to individual patients.

The team has now developed a toolkit called Flexynesis, which does not rely solely on classical machine learning but also uses deep learning to evaluate very different types of data simultaneously — for example, multi-omics data as well as specially processed texts and images, such as CT or MRI scans. “In this way, it enables doctors to make better diagnoses, prognoses, and develop more precise treatment strategies for their patients,” says the author. Flexynesis is described in detail in a paper published in “Nature Communications.”

“We are running multiple translational projects with medical doctors who want to identify biomarkers from multi-omics data that align with disease outcomes,” says the first and co-corresponding author of the publication. “Although many deep-learning based methods have been published for this purpose, most have turned out to be inflexible, tied to specific modeling tasks, or difficult to install and reuse. That gap motivated us to build Flexynesis as a proper toolkit, which is flexible for different modeling tasks and packaged on PyPI, Guix, Docker, Bioconda, and Galaxy, so others can readily apply it in their own pipelines.”

Machine learning for materials discovery and optimization

This Collection supports and amplifies research related to SDG 9 — Industry, Innovation & Infrastructure.

Discovering new materials with customizable and optimized properties, driven either by specific application needs or by fundamental scientific interest, is a primary goal of materials science. Conventionally, the search for new materials is a lengthy and expensive manual process, frequently based on trial and error, requiring the synthesis and characterization of many compositions before a desired material can be found. In recent years this process has been greatly improved by a combination of artificial intelligence and high-throughput approaches. Advances in machine learning for materials science, data-driven materials prediction, autonomous synthesis and characterization, and data-guided high-throughput exploration, can now significantly accelerate materials discovery.

This Collection brings together the latest computational and experimental advances in artificial intelligence, machine learning and data-driven approaches to accelerate high-throughput prediction, synthesis, characterization, optimization, discovery, and understanding of new materials.

Defeating Nondeterminism in LLM Inference

Reproducibility is a bedrock of scientific progress. However, it’s remarkably difficult to get reproducible results out of large language models.

For example, you might observe that asking ChatGPT the same question multiple times provides different results. This by itself is not surprising, since getting a result from a language model involves “sampling”, a process that converts the language model’s output into a probability distribution and probabilistically selects a token.

What might be more surprising is that even when we adjust the temperature down to 0This means that the LLM always chooses the highest probability token, which is called greedy sampling. (thus making the sampling theoretically deterministic), LLM APIs are still not deterministic in practice (see past discussions here, here, or here). Even when running inference on your own hardware with an OSS inference library like vLLM or SGLang, sampling still isn’t deterministic (see here or here).

Mo Gawdat on AI, ethics & machine mastery: How Artificial Intelligence will rule the world

Mo Gawdat warns that AI will soon surpass human intelligence, fundamentally changing society, but also believes that with collective action, ethical development, and altruistic leadership, humans can ensure a beneficial future and potentially avoid losing control to AI

## Questions to inspire discussion.

AI’s Impact on Humanity.

🤖 Q: How soon will AI surpass human intelligence? A: According to Mo Gawdat, AI will reach AGI by 2026, with intelligence measured in thousands compared to humans, making human intelligence irrelevant within 3 years.

🌍 Q: What potential benefits could AI bring to global issues? A: 12% of world military spending redirected to AI could solve world hunger, provide universal healthcare, and end extreme poverty, creating a potential utopia.

Preparing for an AI-Driven Future.

US Energy Secretary’s INSANE Bet Against Elon Musk

Questions to inspire discussion.

Energy for AI and Infrastructure.

🤖 Q: How does AI development impact energy demands? A: AI development will drive massive demand for electricity, with solar and batteries being the only energy source with an unbounded upper limit to scale and meet these demands.

⛽ Q: Can solar energy support existing infrastructure? A: Solar energy can produce synthetic biofuels and oil and gas through chemical processes, enabling it to power existing infrastructure that runs on traditional fuels.

Expert Predictions.

🚗 Q: What does Elon Musk predict about future energy sources? A: Elon Musk predicts that solar and batteries will dominate the future energy landscape, citing China’s massive investment as a key factor in this prediction.

AI system leverages standard security cameras to detect fires in seconds

Fire kills nearly 3,700 Americans annually and destroys $23 billion in property, with many deaths occurring because traditional smoke detectors fail to alert occupants in time.

Now, the NYU Fire Research Group at NYU Tandon School of Engineering has developed an artificial intelligence system that could significantly improve by detecting fires and smoke in using ordinary security cameras already installed in many buildings.

Published in the IEEE Internet of Things, the research demonstrates a system that can analyze and identify fires within 0.016 seconds per frame—faster than the blink of an eye—potentially providing crucial extra minutes for evacuation and . Unlike conventional smoke detectors that require significant smoke buildup and proximity to activate, this AI system can spot fires in their earliest stages from video alone.

Elon Musk: Robotaxis Will Replace Personal Cars, Not Just Uber

Questions to inspire discussion.

🧠 Q: How does Tesla’s upcoming AI chip compare to the current one? A: Tesla’s AI5 chip will be 40 times better than the current AI4 chip, which is already capable of achieving self-driving safety at least 2–3 times that of a human.

💰 Q: What is the expected pricing for Tesla’s robotaxi service? A: Tesla’s robotaxi service is projected to cost $2 per mile at launch, which is cheaper than Uber rides in high-cost areas like Seattle.

Impact on Transportation.

🚘 Q: How will robotaxis affect car ownership? A: Robotaxis are expected to become a viable alternative to car ownership, especially when prices reach $1 per mile, making them cheaper than options like airport parking.

💼 Q: How does Tesla’s robotaxi cost compare to competitors? A: Tesla’s robotaxi can be built and deployed for half the cost of competitors like Whim, potentially offering more competitive pricing.

/* */