A strong open-model and data-centric page because his work sits close to the infrastructure that made OLMo and Dolma useful to the broader research community rather than just another benchmark-driven model release.
Researcher Profile
Editor reviewedXiangru Tang
RWKV and efficient sequence modeling
Research scientist at Google working on agents after PhD research at Yale on agentic AI for biomedical discovery
Worth keeping because it connects an early RWKV byline to a much more visible later research program in agentic AI, biomedical discovery, and code-focused evaluation, which makes the page far more useful than a one-paper ghost profile.
Organizations
About This Page
This profile is meant to help you get oriented quickly: why this researcher matters, what to read first, and where to explore next.
Last reviewed
March 18, 2026
Known For
The ideas, systems, and research directions that make this person worth knowing.
01
Original RWKV authorship
02
Agentic AI for biomedical discovery
03
Code and biomedical benchmark work
04
RWKV and efficient sequence modeling
05
RWKV: Reinventing RNNs for the Transformer Era
06
RWKV (project)
Start Here
Canonical papers, project pages, or repositories that anchor this profile.
Signature Works
Additional papers, projects, or repositories that help flesh out the profile.
Supporting Sources
Additional links that help verify and flesh out this profile.
Related Researchers
People worth exploring next because they share topics, labs, or source material with this profile.
A useful page because it turns an otherwise stray RWKV byline into a visible builder profile: his public work is less about academic publishing and more about making efficient models, AI agents, and production RWKV systems usable in practice.
Worth keeping because he is one of the original RWKV coauthors who clearly did not stop there: his public work moves into production AI for crisis intelligence, security-aware infrastructure tooling, and later open-LLM experimentation.
A useful long-tail page because he is present on both the original RWKV paper and Eagle/Finch, then shows up again on multilingual embedding evaluation work, which makes him more than a one-paper launch contributor.
A useful page because his public trail moved well beyond the original RWKV paper into reasoning, in-context learning, and decoder-only model design, which makes him one of the stronger later-follow-up names from that original author list.
Worth tracking if you care about alternatives to the standard transformer playbook, especially the line of work trying to keep strong language-model performance while making inference and memory use much cheaper.