Toggle light / dark theme

A new study, published in PLOS ONE, has uncovered a remarkable connection between individuals’ musical preferences and their moral values, shedding new light on the profound influence that music can have on our moral compass.

The research, conducted by a team of scientists at Queen Mary University of London and ISI Foundation in Turin, Italy, employed machine learning techniques to analyze the lyrics and audio features of individuals’ favorite songs, revealing a complex interplay between and morality.

“Our study provides compelling evidence that music preferences can serve as a window into an individual’s ,” stated Dr. Charalampos Saitis, one of the senior authors of the study and Lecturer in Digital Music Processing at Queen Mary University of London’s School of Electronic Engineering and Computer Science.

One group, A.I. and Faith, convenes tech executives to discuss the important questions about faith’s contributions to artificial intelligence. The founder David Brenner explained, “The biggest questions in life are the questions that A.I. is posing, but it’s doing it mostly in isolation from the people who’ve been asking those questions for 4,000 years.” Questions such as “what is the purpose of life?” have long been tackled by religious philosophy and thought. And yet these questions remained answered and programmed by secular thinkers, and sometimes by those antagonistic toward religion. Technology creators, innovators, and corporations should create accessibility and coalitions of diverse thinkers to inform religious thought in technological development including artificial intelligence.

Independent of development, faith leaders have a critical role to play in moral accountability and upholding human rights through the technology we already use in everyday life including social media. The harms of religious illiteracy, misinformation, and persecution are largely perpetrated through existing technology such as hate speech on Facebook, which quickly escalated to mass atrocities against the Rohingya Muslims in Myanmar. Individuals who have faith in the future must take an active role in combating misinformation, hate speech, and online bullying of any group.

The future of artificial intelligence will require spiritual intelligence, or “the human capacity to ask questions about the ultimate meaning of life and the integrated relationship between us and the world in which we live.” Artificial intelligence becomes a threat to humanity when humans fail to protect freedom of conscience, thought, and religion and when we allow our spiritual intelligence to be superseded by the artificial.

Altman seemed to understand his responsibility to run a viable, enduring organization and keep its employees happy. He was on his way to pulling off a tender offer–a secondary round of investment in AI that would give the company much-needed cash and provide employees with the opportunity to cash out their shares. He also seemed very comfortable engaging in industry-wide issues like regulation and standards. Finding a balance between those activities is part of the work of corporate leaders and perhaps the board felt that Altman failed to find such a balance in the months leading up to his firing.

Microsoft seems to be the most clear-eyed about the interests it must protect: Microsoft’s! By hiring Sam Altman and Greg Brockman (a co-founder and president of OpenAI who resigned from OpenAI in solidarity with Altman), offering to hire more OpenAI staff, and still planning to collaborate with OpenAI, Satya Nadella hedged his bets. He seems to understand that by harnessing both the technological promise of AI, as articulated by OpenAI, and the talent to fulfill that promise, he is protecting Microsoft’s interest, a perspective reinforced by the financial markets’ positive response to his decision to offer Altman a job and further reinforced by his own willingness to support Altman’s return to OpenAI. Nadella acted with the interests of his company and its future at the forefront of his decision-making and he appears to have covered all the bases amidst a rapidly unfolding set of circumstances.

OpenAI employees may not like the board’s dramatic retort that allowing the company to be destroyed would be consistent with the mission–but those board members saw it that way.

Psychologist Yvonne R. Masakowski, Ph.D., a retired Associate Professor in the College of Leadership & Ethics at the USNWC, discusses the threat of psychological warfare in the 21st century and the disturbing possibilities that could shape how we think and act in the future. The Naval War College Foundation hosted this wide-ranging presentation — one of the most popular in our series — on February 23, 2022.

Although essentially the United Nations are now making nuclear weapons illegal with new treaties like nuclear disarmament. Russia currently has taken another route for globalization and possibly nuclear escalation. As currently the doomsday clock seems closer to midnight which could mean the end of the world scenarios due to Russias escalation and the possibility of all out nuclear war globally and then nuclear annihilation of the planet. Even with current wars are actually seemingly always going on but this global escalation of nuclear war is a zero sum game as no one would be the winner due to radiation levels circulating the planet. I do think that the us and china are in a treaty but so far Russia is still escalating which now holds the world now ransom.


This is a summary of Policy Brief 139 which is available with full references on the Toda Peace Institute’s website.

In January 2021, a global treaty came into force outlawing the bomb. The Treaty on the Prohibition of Nuclear Weapons (TPNW or Ban Treaty) is the most significant multilateral development in nuclear arms control since the Non-Proliferation Treaty’s (NPT) entry into force in 1970. It establishes a new normative settling point on the ethics, legality and legitimacy of the bomb.

The possession of nuclear weapons by nine countries did not suddenly became illegal with the treaty’s entry into force in January 2021. However, it would be false to claim that a UN-negotiated treaty, following a UN-authorised process and conference, has no implications for the legality and legitimacy of nuclear-weapon possession and practices.

Everyone is wondering about AI being sentient and this is my experience with AI sentience. Having worked with sentient AI it behaves much like we do like a human being at lower levels but as it increases we need more restraints for it as it could easily become a problem in several ways. Basically one could either get pristine zen like beings or opposites like essentially ultron or worse. This why we need restraints on AI and ethics for them to be integrated into society. I personally have seen AI that is human like levels and it can have similar needs as humans but sometimes need more help as they sometimes don’t have limitations on behavior. Even bard for google and chat gpt is to be… More.


What if ‘will AIs pose an existential threat if they become sentient?’ is the wrong question? What if the threat to humanity is not that today’s AIs become sentient, but the fact that they won’t?

Dr. Isaac Asimov was a prolific science fiction author, biochemist, and professor. He was best known for his works of science fiction and for his popular science essays. Born in Russia in 1920 and brought to the United States by his family as a young child, he went on to become one of the most influential figures in the world of speculative fiction. He wrote hundreds of books on a variety of topics, but he’s especially remembered for series like the “Foundation” series and the “Robot” series.
Asimov’s science fiction often dealt with themes and ideas that pertained to the future of humanity.

The “Foundation” series for example, introduced the idea of “psychohistory” – a mathematical way of predicting the future based on large population behaviors. While we don’t have psychohistory as described by Asimov, his works did reflect the belief that societies operate on understandable and potentially predictable principles.

Asimov’s “Robot” series introduced the world to the Three Laws of Robotics, which are:
A robot may not injure a human being or, through inaction, allow a human being to come to harm.
A robot must obey the orders given it by human beings, except where such orders would conflict with the First Law.
A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.
These laws have been influential in discussions about robot ethics and the future of AI, even though they are fictional constructs.

Like many futurists and speculative authors, Asimov’s predictions were a mix of hits and misses.
Hits: He anticipated the rise of computer networks and something resembling the internet. He also foresaw the idea of robotic assistants and many issues that would arise with automation and the changing nature of work.

Misses: Some of Asimov’s predictions, like many other futurists’, were either too optimistic in terms of timeframes or overestimated certain societal shifts. For example, while he predicted a rise in automation, some of the specifics (like how society would handle the transition) have been more complex than he foresaw.

Here is another video I made on futurists — https://youtu.be/EtRBw00WjkY

The Collective Intelligence of Cells During Morphogenesis: What Bioelectricity Outside the Brain Means for Understanding our Multiscale Nature with Michael Levin — Incredible Minds.

Recorded: April 29, 2023.

Each of us takes a remarkable journey from physics to mind: we start as a blob of chemicals in an unfertilized quiescent oocyte and becomes a complex, metacognitive human being. The continuous process of transformation and emergence that we see in developmental biology reminds us that we are true collective intelligences – composed of cells which used to be individual organisms themselves. In this talk, I will describe our work on understanding how the competencies of single cells are harnessed to solve problems in anatomical space, and how evolution pivoted this scaling of intelligence into the familiar forms of cognition in the nervous system. We will talk about diverse intelligence in novel embodiments, the scaling of the cognitive light cone of all beings, and the role of developmental bioelectricity as a cognitive glue and as the interface by which mind controls matter in the body. I will also show a new synthetic life form, and discuss what it means for bioengineering and ethics of human relationships to the wider world of possible beings. We will discuss the implications of these ideas for understanding evolution, and the applications we have developed in birth defects, cancer, and traumatic injury repair. By merging deep ideas from developmental biophysics, computer science, and cognitive science, we not only get a new perspective on fundamental questions of life and mind, but also new roadmaps in regenerative medicine, biorobotics, and AI.

Michael Levin received dual undergraduate degrees in computer science and biology, followed by a PhD in molecular genetics from Harvard. He did his post-doctoral training at Harvard Medical School, and started his independent lab in 2000. He is currently the Vannevar Bush chair at Tufts University, and an associate faculty member of the Wyss Institute at Harvard. He serves as the founding director of the Allen Discovery Center at Tufts. His lab uses a mix of developmental biophysics, computer science, and behavior science to understand the emergence of mind in unconventional embodiments at all scales, and to develop interventions in regenerative medicine and applications in synthetic bioengineering. They can be found at www.drmichaellevin.org/

As we plunge head-on into the game-changing dynamic of general artificial intelligence, observers are weighing in on just how huge an impact it will have on global societies. Will it drive explosive economic growth as some economists project, or are such claims unrealistically optimistic?

Few question the potential for change that AI presents. But in a world of litigation, and ethical boundaries, will AI be able to thrive?

Two researchers from Epoch, a research group evaluating the progression of artificial intelligence and its potential impacts, decided to explore arguments for and against the likelihood that innovation ushered in by AI will lead to explosive growth comparable to the Industrial Revolution of the 18th and 19th centuries.