Philipp Moritz, Tyler Griggs, and the SkyRL Team

🗓️ Posted: December 8, 2025

<aside>

We are happy to announce SkyRL tx v0.2!

SkyRL tx is a unified training and inference engine that implements the Tinker API and allows people to set up a Tinker-like service running on their own hardware.

In this release, we add several performance improvements and add support for an external inference engine.

</aside>

<aside> 📢

We gave a talk on SkyRL tx: A unified training and inference engine at this year’s Ray Summit, check out the recording and slides.

</aside>

Updates

There are a number of PRs that are currently in-flight and will be part of the next release, including supporting more sampling parameters (#742, #680), supporting the training_runs endpoint (#720), custom loss functions (#698), FSDP support (#674), Llama3 support (#657), enabling full parameter optimization (#611) and supporting migrations for the database (#580). Thanks Sriran, Ago, Thejas, Benji, Taro, Atem and Lukas for the contributions!

As always, we welcome more contributions!

We are currently falling a little behind in extending the documentation. If you are excited about contributing to it, that would be particularly welcome. Similarly welcome is implementing more functionality of the Tinker API, performance optimizations, and any of the currently open tasks here or really anything you would like to see implemented. One bigger item we plan to work on next is multi-node support, if you are interested in collaborating let us know!

Comparing the performance with vLLM

We have recently implemented a number of performance improvements, especially for the sampling code path. In this section, we show some performance comparisons of SkyRL tx’s native inference support against using vLLM as an external inference engine.