TxGemma - New Open Model for Drug Development

One of the biggest challenges in drug development is finding lead compounds beyond the clinical phase, as 90% of candidates fail in the initial trial phase. In this context, TxGemma—an open model collection developed on top of Google's DeepMind Gemma and its family of modern, lightweight open models—represents a breakthrough. TxGemma aims to harness the power of large language models to improve the efficiency of therapeutic discovery, from identifying promising targets to predicting clinical trial outcomes.

TxGemma is the Successor to Tx-LLM

Launched last October, Tx-LLM was trained to perform a range of therapeutic tasks in the drug development process. The model generated significant interest, so the developers quickly fine-tuned it based on user feedback, resulting in TxGemma. The model is available in three different sizes—2B, 9B, and 27B—each with a "predict" version optimized specifically for narrow therapeutic tasks, such as predicting the toxicity of a molecule or its ability to cross the blood-brain barrier.

TxGemma is based on millions of practical examples that enable the model to excel in various tasks—classification, regression, and generation. The largest predict version, at 27B, outperforms or at least keeps pace with the previous Tx-LLM model in almost all tasks tested, and even outperforms many models optimized for specific tasks. Based on detailed performance data, TxGemma produced similar or better results in 64 out of 66 tasks and performed better than the previous model in 45 tasks.

Chat Capabilities and Further Fine-Tuning Options

The developers focused not only on raw predictive capabilities but also integrated chat features into the models. This allows the models to answer complex questions, justify their decisions, and provide feedback in multi-step conversations. For example, a researcher can ask why a particular molecule was classified as toxic, and the model can justify its answer by referring to the molecule’s structure.

The release of TxGemma offers not only an end product but also a customizable platform for developers and researchers. With the included Colab notebook, it is easy to fine-tune the model based on your own therapeutic data and tasks—such as predicting adverse events in clinical trials. Additionally, TxGemma can be integrated with the Agentic-Tx system, which includes 18 advanced tools, such as PubMed and Wikipedia search, as well as molecular, gene, and protein tools. This solution helps combine everyday research workflows with the multi-step inference capabilities provided by agent systems.

Availabilty

TxGemma is available in the Vertex AI Model Garden and on the Hugging Face platform, so anyone interested can explore how the system works, try out its inference and fine-tuning features, and experiment with the complex workflows offered by Agentic-Tx. As an open model, TxGemma also offers the possibility for further development, since researchers can tailor it to their specific therapeutic development needs using their own data.

The advent of TxGemma could open a new chapter in drug development, significantly shortening the process from the laboratory to the patient’s bedside and reducing development costs. 

Share this post
Google Introduces the Agent2Agent (A2A) Open Source Protocol
In a recent speech, Jensen Huang (CEO of NVIDIA) divided the evolution of artificial intelligence into several phases and called the current phase the era of Agentic AI. Although he mainly focused on the next phase of the physical AI era, we should not forget that the Agentic AI era also started only this year, so its fully developed form has not yet been seen. The recent announcement by Google of the open source Agent2Agent protocol gives us a hint of what this more advanced form might look like. The protocol is designed to bridge the gap between AI agents created on different platforms, frameworks, and by various vendors, enabling smooth communication and collaboration.
Apple in Trouble with Artificial Intelligence Developments?
With Trump's tariffs, Apple appears to be facing increasing problems. One reason is that, besides the tariffs—which have hit Apple's shares hard—there are internal conflicts, especially in the division responsible for AI integration. Tripp Mickle, a journalist for The New York Times, reports that Apple has not been able to produce any new innovations lately. Although this may not be entirely true—since, after much debate, the company finally managed to launch Apple Intelligence—there is no doubt that it is lagging behind its competitors in the field of artificial intelligence.
New Collaboration Between Netflix and OpenAI
Netflix recently began testing a new artificial intelligence-based search feature that uses OpenAI’s technology to improve content search. This feature is a significant departure from traditional search methods because it allows users to find movies and TV shows using specific terms, such as their mood or preferences, rather than only using titles, genres, or actor names.
Strong Turbulence Around Meta Llama Models
Less than a week after its market debut, Llama 4 has already received harsh criticism from users. As mentioned before, one of Llama 4’s new features is its architecture built from different modules. This design lets the model have a much larger effective parameter set than the one it uses at run time, so in theory, it should perform much better. However, several independent user tests show that it does not meet the expected results, especially for mathematical tasks and coding. Some users claim that Meta heavily manipulated benchmarks to achieve better scores, while others believe that an internal version of the model was tested while a more modest version was released to the public.