From nanoelectrode arrays that peer inside heart cells to single-layer spintronics that manipulate quantum states, researchers are advancing how we observe and control matter at multiple scales. This roundup examines breakthroughs in non-invasive cardiac monitoring and AI-powered genomic modeling, alongside developments in energy-efficient computing through anomalous Hall torques. Also in this week in AI and semiconductors, Mayo Clinic’s partnerships with Microsoft Research and Cerebras Systems aim to build foundation models for medical imaging and genomic analysis, while Penn Medicine’s MISO platform processes up to 30,000 data points per pixel for enhanced cancer detection.
AI in healthcare

[Adobe Stock]
OpenAI’s new frontier is longevity science
OpenAI has developed a new AI model aimed at stem cell engineering, according to MIT’s Technology Review. Known as “GPT-4b micro,” the firm trained the model on large-scale biological data to suggest ways to improve reprogramming proteins — often called Yamanaka factors — making cells more youthful and versatile. Early lab tests, conducted in partnership with Retro Biosciences, show a more than 50-fold enhancement in key protein performance markers. “Just across the board, the proteins seem better than what the scientists were able to produce by themselves,” John Hallman, an OpenAI researcher, told Technology Review. This marks OpenAI’s first venture into biological research.
AI-driven method offers noninvasive peek inside heart cells
Researchers at the UC San Diego and Stanford have developed a technique that reconstructs the electrical activity inside heart cells without the need for invasive probes. In the research, described in a study in Nature Communications, stem-cell-derived heart cells were grown on nanoelectrode arrays. Next, the scientists relied on thousands of paired intracellular and extracellular signals to allow a deep learning model to infer subcellular events from surface data alone. “We discovered that extracellular signals hold the information we need to unlock the intracellular features that we’re interested in,” said Zeinab Jahed, a UC San Diego professor, in a press release.

Comparing methods for recording signals within cells. Existing approaches (left) are expensive, highly invasive, and limited in throughput. A new technique (right) uses AI to reconstruct intracellular signals from less invasive extracellular recordings. [Images by Keivan Rahmani.]
This approach could reduce the reliance on invasive methods for testing new heart drugs, offering a more efficient route to assess drug toxicity. The researchers demonstrated the technique’s potential for high-throughput drug screening using stem cell-derived cardiomyocytes and aim to expand this technology for use with neurons and other cell types.
Mayo Clinic, Microsoft Research and Cerebras to partner on foundation models for personalized medicine

[Adobe]
Meanwhile, Mayo Clinic’s partnership with Cerebras focuses on creating a genomic foundation model using both patient exome data and public genome references. Early studies indicate this approach may reveal disease patterns and therapy responses more quickly than standard methods, proving especially valuable for complex conditions like rheumatoid arthritis. The alliance was unveiled at the J.P. Morgan Healthcare Conference.
AI tool ‘MISO’ mines thousands of data points per pixel to improve cancer detection

[From https://www.pennmedicine.org/]
By integrating spatial omics data—such as transcriptomics and proteomics—MISO constructs a map that links the physical layout of tumor tissue with its molecular makeup. Future plans involve scaling the platform to handle larger datasets and additional data types, including epigenetic markers.
AI model from Case Western Reserve aims to predict heart failure risk from CT scans

Using AI, physicians could better predict cardiovascular risk by analyzing CT scans and integrating patient biometric data. This image, courtesy of a Case Western Reserve University press release, illustrates how AI distinguishes different cardiac segments, such as the left and right atria, left and right ventricles, and cardiac arteries.
Investigators at Case Western Reserve University, University Hospitals, and Houston Methodist received a $4 million National Institutes of Health grant to create an AI tool for predicting cardiovascular events. The tool will use routine CT scans, specifically calcium-scoring CT, to analyze existing scan data and patient metrics. This analysis aims to identify risk factors by revealing coronary artery plaque accumulation, providing crucial insights into cardiovascular health.
By integrating large-scale imaging datasets with comprehensive clinical risk factors (including age, body composition, and demographic data), the research team aims to enhance physicians’ ability to identify and manage patients at elevated risk for cardiovascular disease. The technology makes use of existing screening protocols from both Houston Methodist and University Hospitals. It offers a non-invasive approach that could enable healthcare providers to implement more targeted preventative treatments and optimize patient outcomes.
The research team will develop AI-driven predictive models that analyze coronary calcium levels, heart shape, bone density, and visceral fat measurements to provide comprehensive cardiovascular risk assessments.
ALAFIA introduces on-prem HPC to power clinical AI workflows

[ALAFIA]
This HPC-based approach arose from hospitals’ reliance on older computing infrastructure to drive modern imaging equipment costing millions of dollars. Cloud-based GPU rentals, while flexible, can quickly generate high monthly fees when labs process thousands of slides daily. By keeping data on-prem and tapping significant on-board compute resources, ALAFIA systems enable local analytics with reduced latency and lower recurring costs. Hospitals also benefit from a long-term software maintenance guarantee (12 years through Canonical), helping ensure that intensive AI workflows—ranging from cell segmentation to volumetric imaging—remain feasible on-site for both clinical care and research.
Emerging tech spotlight
Researchers unveil ‘brand new physics’ for next-generation spintronics
Physicists from the University of Utah and the University of California, Irvine, have reported a new spin–orbit torque phenomenon known as the anomalous Hall torque, which may pave the way for ultra-fast, energy-efficient computing devices. Traditional spintronic approaches typically rely on multiple magnetic layers to flip electron spins, but this discovery demonstrates how a single layer of ferromagnetic material can transfer spin currents, potentially removing the need for multi-layered structures. By tuning symmetry and leveraging Hall-like effects, the team can precisely control electron spins, offering a promising path for high-speed data storage and neuromorphic computing. “This is brand new physics, which on its own is interesting, but there’s also a lot of potential new applications that go along with it,” said Eric Montoya, assistant professor of physics and astronomy at the University of Utah, in a press release. “These self-generated spin-torques are uniquely qualified for new types of computing like neuromorphic computing, an emerging system that mimics human brain networks.” The research was published in Nature Nanotechnology.
The study adds to the “Hall triad” of spin-orbit torques—encompassing the spin Hall, planar Hall, and now anomalous Hall torques—which appears to be broadly applicable across all conductive spintronic materials. Having already constructed a prototype spin-torque oscillator to demonstrate the effect, the researchers plan to link multiple nanoscale oscillators into larger networks for tasks such as image recognition. This approach could lead to a new generation of computing devices that operate at higher speeds and with fewer energy demands than current hardware.
Tuning Magnetism with Voltage Opens a New Path to Neuromorphic Circuits

Applying a voltage to lanthanum strontium manganite (LSMO) causes it to separate into regions with dramatically different magnetic properties, represented by the red and blue areas in the figure. [New York University and the University of California San Diego/US DOE]
The discovery complements existing work on voltage-tunable resistance in LSMO, suggesting that future spintronic neuromorphic devices could exploit both resistance and magnetism changes in the same material. By enabling robust yet efficient control over electron spins, scientists hope to build networks of “spin oscillators” for next-generation AI, sensing, and computing applications. This research was conducted by the Quantum Materials for Energy Efficient Neuromorphic Computing (Q-MEEN-C) Center, funded by the U.S. Department of Energy’s Office of Science, Basic Energy Sciences. Read more about it in DOE PAGES.
Patent update
Samsung and TSMC led USPTO granted patents in 2024
Samsung Electronics and TSMC have obtained the most granted USPTO patents of 2024, securing 6,377 and 3,989 grants respectively in 2024, according to data from IFI Claims. Samsung’s continued dominance, marking its third consecutive year at the top, represents a 3.44% increase from 2023, while TSMC’s second-place finish showcases 8.19% growth in patent grants. Asian companies now holding 56% of all U.S. patents. A recent article on R&D World also dug into the news, noting, for instance, that there is a growing concentration in electrical digital data processing CPC codes such as (G06F) and digital information transmission (H04L). Emerging technologies like battery technology (H01M) and organic electric solid-state devices (H10K) are also showing significant growth.