Ray

Latest version: v2.44.1

Safety actively analyzes 723650 Python packages for vulnerabilities to keep your Python projects secure.

Scan your dependencies

Page 9 of 18

1.12.1

Not secure
Patch release with the following fixes:

- **Ray now works on Google Colab again!** The bug with memory limit fetching when running Ray in a container is now fixed (https://github.com/ray-project/ray/pull/23922).
- `ray-ml` Docker images for CPU will start being built again after they were stopped in Ray 1.9 (https://github.com/ray-project/ray/pull/24266).
- [Train/Tune] Start MLflow run under the correct experiment for Ray Train and Ray Tune integrations (https://github.com/ray-project/ray/pull/23662).
- [RLlib] Fix for APPO in eager mode (https://github.com/ray-project/ray/pull/24268).
- [RLlib] Fix Alphastar for TF2 and tracing enabled (https://github.com/ray-project/ray/commit/c5502b2aa57376b26408bb297ff68696c16f48f1).
- [Serve] Fix replica leak in anonymous namespaces (https://github.com/ray-project/ray/pull/24311).

1.12.0

Not secure
Highlights
- Ray AI Runtime (AIR), an open-source toolkit for building end-to-end ML applications on Ray, is now in Alpha. AIR is an effort to unify the experience of using different Ray libraries (Ray Data, Train, Tune, Serve, RLlib). You can find more information on the [docs](https://docs.ray.io/en/master/ray-air/getting-started.html) or on the [public RFC](https://github.com/ray-project/ray/issues/22488).
- Getting involved with Ray AIR. We’ll be holding office hours, development sprints, and other activities as we get closer to the Ray AIR Beta/GA release. Want to join us? Fill out this [short form](https://forms.gle/wCCdbaQDtgErYycT6)!
- Ray [usage data collection](https://github.com/ray-project/ray/issues/20857) is now off by default. If you have any questions or concerns, please comment [on the RFC](https://github.com/ray-project/ray/issues/20857).
- New algorithms are added to RLlib: SlateQ & Bandits (for recommender systems use cases) and AlphaStar (multi-agent, multi-GPU w/ league-based self-play)
- Ray Datasets: new lazy execution model with automatic task fusion and memory-optimizing move semantics; first-class support for Pandas DataFrame blocks; efficient random access datasets.


Ray Autoscaler

🎉 New Features
- Support cache_stopped_nodes on Azure (21747)
- AWS Cloudwatch support (21523)

💫 Enhancements
- Improved documentation and standards around built in autoscaler node providers. (22236, 22237)
- Improved KubeRay support (22987, 22847, 22348, 22188)
- Remove redis requirement (22083)

🔨 Fixes
- No longer print infeasible warnings for internal placement group resources. Placement groups which cannot be satisfied by the autoscaler still trigger warnings. (22235)
- Default ami’s per AWS region are updated/fixed. (22506)
- GCP node termination updated (23101)
- Retry legacy k8s operator on monitor failure (22792)
- Cap min and max workers for manually managed on-prem clusters (21710)
- Fix initialization artifacts (22570)
- Ensure initial scaleup with high upscaling_speed isn't limited. (21953)

Ray Client

🎉 New Features:
- ray.init has consistent return value in client mode and driver mode 21355

💫Enhancements:
- Gets and puts are streamed to support arbitrary object sizes 22100, 22327

🔨 Fixes:
- Fix ray client object ref releasing in wrong context 22025


Ray Core

🎉 New Features
- RuntimeEnv:
- Support setting timeout for runtime_env setup. (23082)
- Support setting pip_check and pip_version for runtime_env. (22826, 23306)
- env_vars will take effect when the pip install command is executed. (temporarily ineffective in conda) (22730)
- Support strongly-typed API ray.runtime.RuntimeEnv to define runtime env. (22522)
- Introduce [virtualenv](https://github.com/pypa/virtualenv) to isolate the pip type runtime env. (#21801,22309)
- Raylet shares fate with the dashboard agent. And the dashboard agent will stay alive when it catches the port conflicts. (22382,23024)
- Enable dashboard in the minimal ray installation (21896)
- Add task and object reconstruction status to ray memory cli tools(22317)

🔨 Fixes
- Report only memory usage of pinned object copies to improve scaledown. (22020)
- Scheduler:
- No spreading if a node is selected for lease request due to locality. (22015)
- Placement group scheduling: Non-STRICT_PACK PGs should be sorted by resource priority, size (22762)
- Round robin during spread scheduling (21303)
- Object store:
- Increment ref count when creating an ObjectRef to prevent object from going out of scope (22120)
- Cleanup handling for nondeterministic object size during transfer (22639)
- Fix bug in fusion for spilled objects (22571)
- Handle IO worker failures correctly (20752)
- Improve ray stop behavior (22159)
- Avoid warning when receiving too much logs from a different job (22102)
- Gcs resource manager bug fix and clean up. (22462, 22459)
- Release GIL when running `parallel_memcopy()` / `memcpy()` during serializations. (22492)
- Fix registering serializer before initializing Ray. (23031)

🏗 Architecture refactoring
- Ray distributed scheduler refactoring: (21927, 21992, 22160, 22359, 22722, 22817, 22880, 22893, 22885, 22597, 22857, 23124)
- Removed support for bootstrapping with Redis.


Ray Data Processing
🎉 New Features
- Big Performance and Stability Improvements:
- Add lazy execution mode with automatic stage fusion and optimized memory reclamation via block move semantics (22233, 22374, 22373, 22476)
- Support for random access datasets, providing efficient random access to rows via binary search (22749)
- Add automatic round-robin load balancing for reading and shuffle reduce tasks, obviating the need for the `_spread_resource_prefix` hack (21303)
- More Efficient Tabular Data Wrangling:
- Add first-class support for Pandas blocks, removing expensive Arrow <-> Pandas conversion costs (21894)
- Expose `TableRow` API + minimize copies/type-conversions on row-based ops (22305)
- Groupby + Aggregations Improvements:
- Support mapping over groupby groups (22715)
- Support ignoring nulls in aggregations (20787)
- Improved Dataset Windowing:
- Support windowing a dataset by bytes instead of number of blocks (22577)
- Batch across windows in `DatasetPipeline`s (22830)
- Better Text I/O:
- Support streaming snappy compression for text files (22486)
- Allow for custom decoding error handling in `read_text()` (21967)
- Add option for dropping empty lines in `read_text()` (22298)
- New Operations:
- Add `add_column()` utility for adding derived columns (21967)
- Support for metadata provider callback for read APIs (22896)
- Support configuring autoscaling actor pool size (22574)

🔨 Fixes
- Force lazy datasource materialization in order to respect `DatasetPipeline` stage boundaries (21970)
- Simplify lifetime of designated block owner actor, and don’t create it if dynamic block splitting is disabled (22007)
- Respect 0 CPU resource request when using manual resource-based load balancing (22017)
- Remove batch format ambiguity by always converting Arrow batches to Pandas when `batch_format=”native”` is given (21566)
- Fix leaked stats actor handle due to closure capture reference counting bug (22156)
- Fix boolean tensor column representation and slicing (22323)
- Fix unhandled empty block edge case in shuffle (22367)
- Fix unserializable Arrow Partitioning spec (22477)
- Fix incorrect `iter_epochs()` batch format (22550)
- Fix infinite `iter_epochs()` loop on unconsumed epochs (22572)
- Fix infinite hang on `split()` when `num_shards < num_rows` (22559)
- Patch Parquet file fragment serialization to prevent metadata fetching (22665)
- Don’t reuse task workers for actors or GPU tasks (22482)
- Pin pipeline executor actors to driver node to allow for lineage-based fault tolerance for pipelines (​​22715)
- Always use non-empty blocks to determine schema (22834)
- API fix bash (22886)
- Make label_column optional for `to_tf()` so it can be used for inference (22916)
- Fix `schema()` for `DatasetPipeline`s (23032)
- Fix equalized split when `num_splits == num_blocks` (23191)

💫 Enhancements
- Optimize Parquet metadata serialization via batching (21963)
- Optimize metadata read/write for Ray Client (21939)
- Add sanity checks for memory utilization (22642)

🏗 Architecture refactoring
- Use threadpool to submit `DatasetPipeline` stages (22912)

RLlib

🎉 New Features
- New “AlphaStar” algorithm: A parallelized, multi-agent/multi-GPU learning algorithm, implementing league-based self-play. (21356, 21649)
- SlateQ algorithm has been re-tested, upgraded (multi-GPU capable, TensorFlow version), and bug-fixed (added to weekly learning tests). (22389, 23276, 22544, 22543, 23168, 21827, 22738)
- Bandit algorithms: Moved into `agents` folder as first-class citizens, TensorFlow-Version, unified w/ other agents’ APIs. (22821, 22028, 22427, 22465, 21949, 21773, 21932, 22421)
- ReplayBuffer API (in progress): Allow users to customize and configure their own replay buffers and use these inside custom or built-in algorithms. (22114, 22390, 21808)
- Datasets support for RLlib: Dataset Reader/Writer and documentation. (21808, 22239, 21948)

🔨 Fixes
- Fixed memory leak in SimpleReplayBuffer. (22678)
- Fixed Unity3D built-in examples: Action bounds from -inf/inf to -1.0/1.0. (22247)
- Various bug fixes. (22350, 22245, 22171, 21697, 21855, 22076, 22590, 22587, 22657, 22428, 23063, 22619, 22731, 22534, 22074, 22078, 22641, 22684, 22398, 21685)

🏗 Architecture refactoring
- A3C: Moved into new `training_iteration` API (from `exeution_plan` API). Lead to a ~2.7x performance increase on a Atari + CNN + LSTM benchmark. (22126, 22316)
- Make `multiagent->policies_to_train` more flexible via callable option (alternative to providing a list of policy IDs). (20735)

💫Enhancements:
- Env pre-checking module now active by default. (22191)
- Callbacks: Added `on_sub_environment_created` and `on_trainer_init` callback options. (21893, 22493)
- RecSim environment wrappers: Ability to use google’s RecSim for recommender systems more easily w/ RLlib algorithms (3 RLlib-ready example environments). (22028, 21773, 22211)
- MARWIL loss function enhancement (exploratory term for stddev). (21493)

📖Documentation:
- Docs enhancements: Setup-dev instructions; Ray datasets integration. (22239)
- Other doc enhancements and fixes. (23160, 23226, 22496, 22489, 22380)


Ray Workflow
🎉 New Features:
- Support skip checkpointing.

🔨 Fixes:
- Fix an issue where the event loop is not set.


Tune
🎉 New Features:
- Expose new checkpoint interface to users ([22741](https://github.com/ray-project/ray/pull/22741))

💫Enhancements:
- Better error msg for grpc resource exhausted error. (22806)
- Add CV support for XGB/LGBM Tune callbacks (22882)
- Make sure tune.run can run inside worker thread (https://github.com/ray-project/ray/commit/b8c28d1f2beb7a141f80a5fd6053c8e8520718b9)[#22566](https://github.com/ray-project/ray/pull/22566)[)](https://github.com/ray-project/ray/commit/b8c28d1f2beb7a141f80a5fd6053c8e8520718b9)
- Add Trainable.postprocess_checkpoint (22973)
[Trainables will now know TUNE_ORIG_WORKING_DIR (](https://github.com/ray-project/ray/commit/f5995dccdf0ab4012e511c3379b19f06f1d307b5)[#22803](https://github.com/ray-project/ray/pull/22803)[)](https://github.com/ray-project/ray/commit/f5995dccdf0ab4012e511c3379b19f06f1d307b5)
- Retry cloud sync up/down/delete on fail (22029)
- Support functools.partial names and treat as function in registry (21518)

🔨Fixes:
- [Cleanup incorrectly formatted strings (Part 2: Tune) (](https://github.com/ray-project/ray/commit/761f927720586403c642dc62fa510c033fd7ffd5)[#23129](https://github.com/ray-project/ray/pull/23129)[)](https://github.com/ray-project/ray/commit/761f927720586403c642dc62fa510c033fd7ffd5)
- fix error handling for fail_fast case. (22982)
- Remove Trainable.update_resources (22471)
- Bump flaml from 0.6.7 to 0.9.7 in /python/requirements/ml (22071)
- Fix analysis without registered trainable (21475)
- Update Lightning examples to support PTL 1.5 (20562)
- Fix WandbTrainableMixin config for rllib trainables (22063)
- [wandb] Use resume=False per default ([21892](https://github.com/ray-project/ray/pull/21892))

🏗 Refactoring:
- [Move resource updater out of trial executor (](https://github.com/ray-project/ray/commit/cc1728120f7d49b0016d190971bc8056d3245c5d)[#23178](https://github.com/ray-project/ray/pull/23178)[)](https://github.com/ray-project/ray/commit/cc1728120f7d49b0016d190971bc8056d3245c5d)
- [Preparation for deadline schedulers (](https://github.com/ray-project/ray/commit/4a15c6f8f3ebb634f7cef967a097f621462e4e50)[#22006](https://github.com/ray-project/ray/pull/22006)[)](https://github.com/ray-project/ray/commit/4a15c6f8f3ebb634f7cef967a097f621462e4e50)
- [Single wait refactor. (](https://github.com/ray-project/ray/commit/323511b716416088859967686c71889ef8425204)[#21852](https://github.com/ray-project/ray/pull/21852)[)](https://github.com/ray-project/ray/commit/323511b716416088859967686c71889ef8425204)

📖Documentation:
- Tune docs overhaul (first part) (22112)
- [Tune overhaul part II (](https://github.com/ray-project/ray/commit/372c620f58c2269dccd5a871a72aebb9df76e32c)[#22656](https://github.com/ray-project/ray/pull/22656)[)](https://github.com/ray-project/ray/commit/372c620f58c2269dccd5a871a72aebb9df76e32c)
- Note TPESampler performance issues in docs (22545)
- hyperopt notebook (22315)

Train
🎉 New Features
- Integration with PyTorch profiler. Easily enable the pytorch profiler with Ray Train to profile training and visualize stats in Tensorboard (22345).
- Automatic pipelining of host to device transfer. While training is happening on one batch of data, the next batch of data is concurrently being moved from CPU to GPU (22716, 22974)
- Automatic Mixed Precision. Easily enable PyTorch automatic mixed precision during training (22227).

💫 Enhancements
- Add utility function to enable reproducibility for Pytorch training (22851)
- Add initial support for metrics aggregation (22099)
- Add support for `trainer.best_checkpoint` and `Trainer.load_checkpoint_path`. You can now directly access the best in memory checkpoint, or load an arbitrary checkpoint path to memory. (22306)

🔨 Fixes
- Add a utility function to turn off TF autosharding (21887)
- Fix fault tolerance for Tensorflow training (22508)
- Train utility methods (`train.report()`, etc.) can now be called outside of a Train session (21969)
- Fix accuracy calculation for CIFAR example (22292)
- Better error message for placement group time out (22845)

📖 Documentation
- Update docs for ray.train.torch import (22555)
- Clarify shuffle documentation in `prepare_data_loader` (22876)
- Denote `train.torch.get_device` as a Public API (22024)
- Minor fixes on Ray Train user guide doc (22379)


Serve
🎉 New Features
- [Deployment Graph API](https://docs.ray.io/en/master/serve/deployment-graph.html) is now in alpha. It provides a way to build, test and deploy complex inference graph composed of many deployments. (#23177, 23252, 23301, 22840, 22710, 22878, 23208, 23290, 23256, 23324, 23289, 23285, 22473, 23125, 23210)
- New experimental REST API and CLI for creating and managing deployments. (
22839, 22257, 23198, 23027, 22039, 22547, 22578, 22611, 22648, 22714, 22805, 22760, 22917, 23059, 23195, 23265, 23157, 22706, 23017, 23026, 23215)
- New sets of [HTTP adapters](https://docs.ray.io/en/master/serve/http-servehandle.html#http-adapters) making it easy to build simple application, as well as [Ray AI Runtime model wrappers](https://docs.ray.io/en/master/ray-air/getting-started.html#air-serve-integration) in alpha. (22913, 22914, 22915, 22995)
- New `health_check` API for end to end user provided health check. (22178, 22121, 22297)

🔨 Fixes
- Autoscaling algorithm will now relingquish most idle nodes when scaling down (22669)
- Serve can now manage Java replicas (22628)
- Added a hands-on self-contained MLflow and Ray Serve deployment example (22192)
- Added `root_path` setting to `http_options` (21090)
- Remove `shard_key`, `http_method`, and `http_headers` in `ServeHandle` (21590)


Dashboard
🔨Fixes:
- Update CPU and memory reporting in kubernetes. (21688)

Thanks
Many thanks to all those who contributed to this release!
edoakes, pcmoritz, jiaodong, iycheng, krfricke, smorad, kfstorm, jjyyxx, rodrigodelazcano, scv119, dmatrix, avnishn, fyrestone, clarkzinzow, wumuzi520, gramhagen, XuehaiPan, iasoon, birgerbr, n30111, tbabej, Zyiqin-Miranda, suquark, pdames, tupui, ArturNiederfahrenhorst, ashione, ckw017, siddgoel, Catch-Bull, vicyap, spolcyn, stephanie-wang, mopga, Chong-Li, jjyao, raulchen, sven1977, nikitavemuri, jbedorf, mattip, bveeramani, czgdp1807, dependabot[bot], Fabien-Couthouis, willfrey, mwtian, SlowShip, Yard1, WangTaoTheTonic, Wendi-anyscale, kaushikb11, kennethlien, acxz, DmitriGekhtman, matthewdeng, mraheja, orcahmlee, richardliaw, dsctt, yupbank, Jeffwan, gjoliver, jovany-wang, clay4444, shrekris-anyscale, jwyyy, kyle-chen-uber, simon-mo, ericl, amogkam, jianoaix, rkooo567, maxpumperla, architkulkarni, chenk008, xwjiang2010, robertnishihara, qicosmos, sriram-anyscale, SongGuyang, jon-chuang, wuisawesome, valiantljk, simonsays1980, ijrsvt

1.11.1

Not secure
Patch release including fixes for the following issues:

- Ray Job Submission not working with remote `working_dir` URLs in their runtime environment (https://github.com/ray-project/ray/pull/22018)
- Ray Tune + MLflow integration failing to set MLflow experiment ID (https://github.com/ray-project/ray/pull/23662)
- Dependencies for `gym` not pinned, leading to version incompatibility issues (https://github.com/ray-project/ray/pull/23705)

1.11.0

Not secure
Highlights

🎉 Ray no longer starts Redis by default. Cluster metadata previously stored in Redis is stored in the GCS now.

Ray Autoscaler

🎉 New Features
- AWS Cloudwatch dashboard support 20266

💫 Enhancements
- Kuberay autoscaler prototype 21086

🔨 Fixes
- Ray.autoscaler.sdk import issue 21795

Ray Core

🎉 New Features
- Set actor died error message in ActorDiedError 20903
- Event stats is enabled by default 21515

🔨 Fixes
- Better support for nested tasks
- Fixed 16GB mac perf issue by limit the plasma store size to 2GB 21224
- Fix `SchedulingClassInfo.running_tasks` memory leak 21535
- Round robin during spread scheduling 19968

🏗 Architecture refactoring
- Refactor scheduler resource reporting public APIs 21732
- Refactor ObjectManager wait logic to WaitManager 21369

Ray Data Processing
🎉 New Features
- More powerful to_torch() API, providing more control over the GPU batch format. (21117)

🔨 Fixes
- Fix simple Dataset sort generating only 1 non-empty block. (21588)
- Improve error handling across sorting, groupbys, and aggregations. (21610, 21627)
- Fix boolean tensor column representation and slicing. (22358)

RLlib

🎉 New Features
- Better utils for flattening complex inputs and enable prev-actions for LSTM/attention for complex action spaces. (21330)
- `MultiAgentEnv` pre-checker (21476)
- Base env pre-checker. (21569)

🔨 Fixes
- Better defaults for QMix (21332)
- Fix contrib/MADDPG + pettingzoo coop-pong-v4. (21452)
- Fix action unsquashing causes inf/NaN actions for unbounded action spaces. (21110)
- Ignore PPO KL-loss term completely if kl-coeff == 0.0 to avoid NaN values (21456)
- `unsquash_action` and `clip_action` (when None) cause wrong actions computed by `Trainer.compute_single_action`. (21553)
- Conv2d default filter tests and add default setting for 96x96 image obs space. (21560)
- Bing back and fix offline RL(BC & MARWIL) learning tests. (21574, 21643)
- SimpleQ should not use a prio. replay buffer. (21665)
- Fix video recorder env wrapper. Added test case. (21670)

🏗 Architecture refactoring
- Decentralized multi-agent learning (21421)
- Preparatory PR for multi-agent multi-GPU learner (alpha-star style) (21652)

Ray Workflow
🔨 Fixes
- Fixed workflow recovery issue due to a bug of dynamic output 21571

Tune
🎉 New Features
- It is now possible to load all evaluated points from an experiment into a Searcher (21506)
- Add CometLoggerCallback (20766)

💫 Enhancements
- Only sync the checkpoint folder instead of the entire trial folder for cloud checkpoint. (21658)
- Add test for heterogeneous resource request deadlocks (21397)
- Remove unused `return_or_clean_cached_pg` (21403)
- Remove `TrialExecutor.resume_trial` (21225)
- Leave only one canonical way of stopping a trial (21021)

🔨 Fixes
- Replace deprecated `running_sanity_check` with `sanity_checking` in PTL integration (21831)
- Fix loading an `ExperimentAnalysis` object without a registered `Trainable` (21475)
- Fix stale node detection bug (21516)
- Fixes to allow `tune/tests/test_commands.py` to run on Windows (21342)
- Deflake PBT tests (21366)
- Fix dtype coercion in `tune.choice` (21270)

📖 Documentation
- Fix typo in `schedulers.rst` (21777)

Train
🎉 New Features
- Add PrintCallback (21261)
- Add MLflowLoggerCallback(20802)

💫 Enhancements
- Refactor Callback implementation (21468, 21357, 21262)

🔨 Fixes
- Fix Dataloader (21467)

📖 Documentation
- Documentation and example fixes (​​21761, 21689, 21464)

Serve
🎉 New Features
- Checkout our revampt end-to-end [tutorial](https://docs.ray.io/en/master/serve/end_to_end_tutorial.html) that walks through the deployment journey! (#20765)

🔨 Fixes
- Warn when serve.start() with different options (21562)
- Detect http.disconnect and cancel requests properly (21438)

Thanks
Many thanks to all those who contributed to this release!
isaac-vidas, wuisawesome, stephanie-wang, jon-chuang, xwjiang2010, jjyao, MissiontoMars, qbphilip, yaoyuan97, gjoliver, Yard1, rkooo567, talesa, czgdp1807, DN6, sven1977, kfstorm, krfricke, simon-mo, hauntsaninja, pcmoritz, JamieSlome, chaokunyang, jovany-wang, sidward14, DmitriGekhtman, ericl, mwtian, jwyyy, clarkzinzow, hckuo, vakker, HuangLED, iycheng, edoakes, shrekris-anyscale, robertnishihara, avnishn, mickelliu, ndrwnaguib, ijrsvt, Zyiqin-Miranda, bveeramani, SongGuyang, n30111, WangTaoTheTonic, suquark, richardliaw, qicosmos, scv119, architkulkarni, lixin-wei, Catch-Bull, acxz, benblack769, clay4444, amogkam, marin-ma, maxpumperla, jiaodong, mattip, isra17, raulchen, wilsonwang371, carlogrisetti, ashione, matthewdeng

1.10.0

Not secure
Highlights

- 🎉 Ray Windows support is now in beta – a significant fraction of the Ray test suite is now passing on Windows. We are eager to learn about your experience with Ray 1.10 on Windows, please file issues you encounter at https://github.com/ray-project/ray/issues. In the upcoming releases we will spend more time on making Ray Serve and Runtime Environment tests pass on Windows and on polishing things.

Ray Autoscaler
💫Enhancements:
- Add autoscaler update time to prometheus metrics (20831)
- Fewer non terminated nodes calls in autoscaler update (20359, 20623)

🔨 Fixes:
- GCP TPU autoscaling fix (20311)
- Scale-down stability fix (21204)
- Report node launch failure in driver logs (20814)


Ray Client
💫Enhancements
- Client task options are encoded with pickle instead of json (20930)


Ray Core
🎉 New Features:
- `runtime_env`’s `pip` field now installs pip packages in your existing environment instead of installing them in a new isolated environment. (20341)

🔨 Fixes:
- Fix bug where specifying runtime_env conda/pip per-job using local requirements file using Ray Client on a remote cluster didn’t work (20855)
- Security fixes for `log4j2` – the `log4j2` version has been bumped to 2.17.1 (21373)

💫Enhancements:
- Allow runtime_env working_dir and py_modules to be pathlib.Path type (20853, 20810)
- Add environment variable to skip local runtime_env garbage collection (21163)
- Change runtime_env error log to debug log (20875)
- Improved reference counting for runtime_env resources (20789)

🏗 Architecture refactoring:
- Refactor runtime_env to use protobuf for multi-language support (19511)

📖Documentation:
- Add more comprehensive runtime_env documentation (20222, 21131, 20352)


Ray Data Processing
🎉 New Features:
- Added stats framework for debugging Datasets performance (20867, 21070)
- [Dask-on-Ray] New config helper for enabling the Dask-on-Ray scheduler (21114)

💫Enhancements:
- Reduce memory usage during when converting to a Pandas DataFrame (20921)

🔨 Fixes:
- Fix slow block evaluation when splitting (20693)
- Fix boundary sampling concatenation on non-uniform blocks (20784)
- Fix boolean tensor column slicing (20905)

🏗 Architecture refactoring:
- Refactor table block structure to support more tabular block formats (20721)


RLlib

🎉 New Features:
- Support for RE3 exploration algorithm (for tf only). (19551)
- Environment pre-checks, better failure behavior and enhanced environment API. (20481, 20832, 20868, 20785, 21027, 20811)

🏗 Architecture refactoring:
- Evaluation: Support evaluation setting that makes sure `train` doesn't ever have to wait for `eval` to finish (b/c of long episodes). (20757); Always attach latest eval metrics. (21011)
- Soft-deprecate `build_trainer()` utility function in favor of sub-classing `Trainer` directly (and overriding some of its methods). (20635, 20636, 20633, 20424, 20570, 20571, 20639, 20725)
- Experimental no-flatten option for actions/prev-actions. (20918)
- Use `SampleBatch` instead of an input dict whenever possible. (20746)
- Switch off `Preprocessors` by default for `PGTrainer` (experimental). (21008)
- Toward a Replay Buffer API (cleanups; docstrings; renames; move into `rllib/execution/buffers` dir) (20552)

📖Documentation:
- Overhaul of auto-API reference pages. (19786, 20537, 20538, 20486, 20250)
- README and RLlib landing page overhaul (20249).
- Added example containing code to compute an adapted (time-dependent) GAE used by the PPO algorithm (20850).

🔨 Fixes:
- Smaller fixes and enhancements: 20704, 20541, 20793, 20743.


Tune
🎉 New Features:
- Introduce TrialCheckpoint class, making checkpoint down/upload easie (20585)
- Add random state to `BasicVariantGenerator` (20926)
- Multi-objective support for Optuna (20489)

💫Enhancements:
- Add `set_max_concurrency` to Searcher API (20576)
- Allow for tuples in _split_resolved_unresolved_values. (20794)
- Show the name of training func, instead of just ImplicitFunction. (21029)
- Enforce one future at a time for any given trial at any given time. (20783)
move `on_no_available_trials` to a subclass under `runner` (20809)
- Clean up code (20555, 20464, 20403, 20653, 20796, 20916, 21067)
- Start restricting TrialRunner/Executor interface exposures. (20656)
- TrialExecutor should not take in Runner interface. (20655)


🔨Fixes:
- Deflake test_tune_restore.py (20776)
- Fix best_trial_str for nested custom parameter columns (21078)
- Fix checkpointing error message on K8s (20559)
- Fix testResourceScheduler and testMultiStepRun. (20872)
- Fix tune cloud tests for function and rllib trainables (20536)
- Move _head_bundle_is_empty after conversion (21039)
- Elongate test_trial_scheduler_pbt timeout. (21120)


Train
🔨Fixes:
- Ray Train environment variables are automatically propagated and do not need to be manually set on every node (20523)
- Various minor fixes and improvements (20952, 20893, 20603, 20487)
📖Documentation:
- Update saving/loading checkpoint docs (20973). Thanks jwyyy!
- Various minor doc updates (20877, 20683)


Serve
💫Enhancements:
- Add validation to Serve AutoscalingConfig class (20779)
- Add Serve metric for HTTP error codes (21009)

🔨Fixes:
- No longer create placement group for deployment with no resources (20471)
- Log errors in deployment initialization/configuration user code (20620)


Jobs
🎉 New Features:
- Logs can be streamed from job submission server with `ray job logs` command (20976)
- Add documentation for ray job submission (20530)
- Propagate custom headers field to JobSubmissionClient and apply to all requests (20663)

🔨Fixes:
- Fix job serve accidentally creates local ray processes instead of connecting (20705)

💫Enhancements:
- [Jobs] Update CLI examples to use the same setup (20844)

Thanks
Many thanks to all those who contributed to this release!

dmatrix, suquark, tekumara, jiaodong, jovany-wang, avnishn, simon-mo, iycheng, SongGuyang, ArturNiederfahrenhorst, wuisawesome, kfstorm, matthewdeng, jjyao, chenk008, Sertingolix, larrylian, czgdp1807, scv119, duburcqa, runedog48, Yard1, robertnishihara, geraint0923, amogkam, DmitriGekhtman, ijrsvt, kk-55, lixin-wei, mvindiola1, hauntsaninja, sven1977, Hankpipi, qbphilip, hckuo, newmanwang, clay4444, edoakes, liuyang-my, iasoon, WangTaoTheTonic, fgogolli, dproctor, gramhagen, krfricke, richardliaw, bveeramani, pcmoritz, ericl, simonsays1980, carlogrisetti, stephanie-wang, AmeerHajAli, mwtian, xwjiang2010, shrekris-anyscale, n30111, lchu-ibm, Scalsol, seonggwonyoon, gjoliver, qicosmos, xychu, iamhatesz, architkulkarni, jwyyy, rkooo567, mattip, ckw017, MissiontoMars, clarkzinzow

1.9.2

Not secure
Patch release to bump the `log4j` version from `2.16.0` to `2.17.0`. This resolves the security issue [CVE-2021-45105](https://github.com/advisories/GHSA-p6xc-xr62-6r2g).

Page 9 of 18

© 2025 Safety CLI Cybersecurity Inc. All Rights Reserved.