The dev channel has been updated to 106.0.5231.2 for Windows, Mac & Linux.
Google Workspace Updates Weekly Recap – August 12, 2022
New updates
Unless otherwise indicated, the features below are fully launched or in the process of rolling out (rollouts should take no more than 15 business days to complete), launching to both Rapid and Scheduled Release at the same time (if not, each stage of rollout should take no more than 15 business days to complete), and available to all Google Workspace and G Suite customers.Previous announcements
The announcements below were published on the Workspace Updates blog earlier this week. Please refer to the original blog posts for complete details.
Source: Google Workspace Updates
Google Workspace Updates Weekly Recap – August 12, 2022
New updates
Unless otherwise indicated, the features below are fully launched or in the process of rolling out (rollouts should take no more than 15 business days to complete), launching to both Rapid and Scheduled Release at the same time (if not, each stage of rollout should take no more than 15 business days to complete), and available to all Google Workspace and G Suite customers.Previous announcements
The announcements below were published on the Workspace Updates blog earlier this week. Please refer to the original blog posts for complete details.
Source: Google Workspace Updates
Control visibility of admin alerts with admin role privileges
What’s changing
Getting started
- Admins: Visit the Help Center to learn more about using the alert center, granting access to the alert center, and admin roles.
- End users: There is no end user impact.
Rollout pace
- This feature is available now for all users. Availability Available to all Google Workspace customers, as well as legacy G Suite Basic and Business customers
Resources
Source: Google Workspace Updates
Google Meet call control for USB peripheral devices
What’s changing
Who’s impacted
Why it’s important
Additional Details
Supported devices
- Call control will work with most USB telephony peripherals; however, the experience may differ from device to device. You can find a listing of Meet-certified headsets or speaker microphone can be found here.
- Bluetooth devices are not supported at this time.
Getting started
- Admins: There is no admin impact.
- End users: Enabling call control for your USB peripheral:
- In the pre-call green room navigate to Audio > Call Control > Connect device. You’ll be prompted to connect your USB peripheral.
- This setting can also be accessed during a call from More options (three-dot icon) > Audio > Call Control > Connect device.
- How to use call control with your USB peripheral
- When you’re in a meeting, you can press the mute / unmute button on the physical peripheral to toggle your mute state during the meeting.
- Additionally, if your peripheral has an LED status light, the light will sync with your current mute status.
Rollout pace
- Rapid Release and Scheduled Release domains: Gradual rollout (up to 15 days for feature visibility) starting on August 22, 2022
Availability
- Available to all Google Workspace customers, as well as legacy G Suite Basic and Business customers.
Resources
Source: Google Workspace Updates
Dev Channel Update for ChromeOS
The Dev channel is being updated to 106.0.5226.0 (Platform version: 15036.0.0) for most ChromeOS devices. This build contains a number of bug fixes and security updates.
If you find new issues, please let us know one of the following ways
- File a bug
- Visit our ChromeOS communities
- General: Chromebook Help Community
- Beta Specific: ChromeOS Beta Help Community
- Report an issue or send feedback on Chrome
Interested in switching channels? Find out how.
Cole Brown,
Google ChromeOS
Source: Google Chrome Releases
Rax: Composable Learning-to-Rank Using JAX
Ranking is a core problem across a variety of domains, such as search engines, recommendation systems, or question answering. As such, researchers often utilize learning-to-rank (LTR), a set of supervised machine learning techniques that optimize for the utility of an entire list of items (rather than a single item at a time). A noticeable recent focus is on combining LTR with deep learning. Existing libraries, most notably TF-Ranking, offer researchers and practitioners the necessary tools to use LTR in their work. However, none of the existing LTR libraries work natively with JAX, a new machine learning framework that provides an extensible system of function transformations that compose: automatic differentiation, JIT-compilation to GPU/TPU devices and more.
Today, we are excited to introduce Rax, a library for LTR in the JAX ecosystem. Rax brings decades of LTR research to the JAX ecosystem, making it possible to apply JAX to a variety of ranking problems and combine ranking techniques with recent advances in deep learning built upon JAX (e.g., T5X). Rax provides state-of-the-art ranking losses, a number of standard ranking metrics, and a set of function transformations to enable ranking metric optimization. All this functionality is provided with a well-documented and easy to use API that will look and feel familiar to JAX users. Please check out our paper for more technical details.
Learning-to-Rank Using Rax
Rax is designed to solve LTR problems. To this end, Rax provides loss and metric functions that operate on batches of lists, not batches of individual data points as is common in other machine learning problems. An example of such a list is the multiple potential results from a search engine query. The figure below illustrates how tools from Rax can be used to train neural networks on ranking tasks. In this example, the green items (B, F) are very relevant, the yellow items (C, E) are somewhat relevant and the red items (A, D) are not relevant. A neural network is used to predict a relevancy score for each item, then these items are sorted by these scores to produce a ranking. A Rax ranking loss incorporates the entire list of scores to optimize the neural network, improving the overall ranking of the items. After several iterations of stochastic gradient descent, the neural network learns to score the items such that the resulting ranking is optimal: relevant items are placed at the top of the list and non-relevant items at the bottom.
![]() |
Using Rax to optimize a neural network for a ranking task. The green items (B, F) are very relevant, the yellow items (C, E) are somewhat relevant and the red items (A, D) are not relevant. |
Approximate Metric Optimization
The quality of a ranking is commonly evaluated using ranking metrics, e.g., the normalized discounted cumulative gain (NDCG). An important objective of LTR is to optimize a neural network so that it scores highly on ranking metrics. However, ranking metrics like NDCG can present challenges because they are often discontinuous and flat, so stochastic gradient descent cannot directly be applied to these metrics. Rax provides state-of-the-art approximation techniques that make it possible to produce differentiable surrogates to ranking metrics that permit optimization via gradient descent. The figure below illustrates the use of rax.approx_t12n
, a function transformation unique to Rax, which allows for the NDCG metric to be transformed into an approximate and differentiable form.
![]() |
Using an approximation technique from Rax to transform the NDCG ranking metric into a differentiable and optimizable ranking loss (approx_t12n and gumbel_t12n ). |
First, notice how the NDCG metric (in green) is flat and discontinuous, making it hard to optimize using stochastic gradient descent. By applying the rax.approx_t12n
transformation to the metric, we obtain ApproxNDCG, an approximate metric that is now differentiable with well-defined gradients (in red). However, it potentially has many local optima — points where the loss is locally optimal, but not globally optimal — in which the training process can get stuck. When the loss encounters such a local optimum, training procedures like stochastic gradient descent will have difficulty improving the neural network further.
To overcome this, we can obtain the gumbel-version of ApproxNDCG by using the rax.gumbel_t12n
transformation. This gumbel version introduces noise in the ranking scores which causes the loss to sample many different rankings that may incur a non-zero cost (in blue). This stochastic treatment may help the loss escape local optima and often is a better choice when training a neural network on a ranking metric. Rax, by design, allows the approximate and gumbel transformations to be freely used with all metrics that are offered by the library, including metrics with a top-k cutoff value, like recall or precision. In fact, it is even possible to implement your own metrics and transform them to obtain gumbel-approximate versions that permit optimization without any extra effort.
Ranking in the JAX Ecosystem
Rax is designed to integrate well in the JAX ecosystem and we prioritize interoperability with other JAX-based libraries. For example, a common workflow for researchers that use JAX is to use TensorFlow Datasets to load a dataset, Flax to build a neural network, and Optax to optimize the parameters of the network. Each of these libraries composes well with the others and the composition of these tools is what makes working with JAX both flexible and powerful. For researchers and practitioners of ranking systems, the JAX ecosystem was previously missing LTR functionality, and Rax fills this gap by providing a collection of ranking losses and metrics. We have carefully constructed Rax to function natively with standard JAX transformations such as jax.jit
and jax.grad
and various libraries like Flax and Optax. This means that users can freely use their favorite JAX and Rax tools together.
Ranking with T5
While giant language models such as T5 have shown great performance on natural language tasks, how to leverage ranking losses to improve their performance on ranking tasks, such as search or question answering, is under-explored. With Rax, it is possible to fully tap this potential. Rax is written as a JAX-first library, thus it is easy to integrate it with other JAX libraries. Since T5X is an implementation of T5 in the JAX ecosystem, Rax can work with it seamlessly.
To this end, we have an example that demonstrates how Rax can be used in T5X. By incorporating ranking losses and metrics, it is now possible to fine-tune T5 for ranking problems, and our results indicate that enhancing T5 with ranking losses can offer significant performance improvements. For example, on the MS-MARCO QNA v2.1 benchmark we are able to achieve a +1.2% NDCG and +1.7% MRR by fine-tuning a T5-Base model using the Rax listwise softmax cross-entropy loss instead of a pointwise sigmoid cross-entropy loss.
![]() |
Fine-tuning a T5-Base model on MS-MARCO QNA v2.1 with a ranking loss (softmax, in blue) versus a non-ranking loss (pointwise sigmoid, in red). |
Conclusion
Overall, Rax is a new addition to the growing ecosystem of JAX libraries. Rax is entirely open source and available to everyone at github.com/google/rax. More technical details can also be found in our paper. We encourage everyone to explore the examples included in the github repository: (1) optimizing a neural network with Flax and Optax, (2) comparing different approximate metric optimization techniques, and (3) how to integrate Rax with T5X.
Acknowledgements
Many collaborators within Google made this project possible: Xuanhui Wang, Zhen Qin, Le Yan, Rama Kumar Pasumarthi, Michael Bendersky, Marc Najork, Fernando Diaz, Ryan Doherty, Afroz Mohiuddin, and Samer Hassan.
Source: Google AI Blog
Rax: Composable Learning-to-Rank Using JAX
Ranking is a core problem across a variety of domains, such as search engines, recommendation systems, or question answering. As such, researchers often utilize learning-to-rank (LTR), a set of supervised machine learning techniques that optimize for the utility of an entire list of items (rather than a single item at a time). A noticeable recent focus is on combining LTR with deep learning. Existing libraries, most notably TF-Ranking, offer researchers and practitioners the necessary tools to use LTR in their work. However, none of the existing LTR libraries work natively with JAX, a new machine learning framework that provides an extensible system of function transformations that compose: automatic differentiation, JIT-compilation to GPU/TPU devices and more.
Today, we are excited to introduce Rax, a library for LTR in the JAX ecosystem. Rax brings decades of LTR research to the JAX ecosystem, making it possible to apply JAX to a variety of ranking problems and combine ranking techniques with recent advances in deep learning built upon JAX (e.g., T5X). Rax provides state-of-the-art ranking losses, a number of standard ranking metrics, and a set of function transformations to enable ranking metric optimization. All this functionality is provided with a well-documented and easy to use API that will look and feel familiar to JAX users. Please check out our paper for more technical details.
Learning-to-Rank Using Rax
Rax is designed to solve LTR problems. To this end, Rax provides loss and metric functions that operate on batches of lists, not batches of individual data points as is common in other machine learning problems. An example of such a list is the multiple potential results from a search engine query. The figure below illustrates how tools from Rax can be used to train neural networks on ranking tasks. In this example, the green items (B, F) are very relevant, the yellow items (C, E) are somewhat relevant and the red items (A, D) are not relevant. A neural network is used to predict a relevancy score for each item, then these items are sorted by these scores to produce a ranking. A Rax ranking loss incorporates the entire list of scores to optimize the neural network, improving the overall ranking of the items. After several iterations of stochastic gradient descent, the neural network learns to score the items such that the resulting ranking is optimal: relevant items are placed at the top of the list and non-relevant items at the bottom.
![]() |
Using Rax to optimize a neural network for a ranking task. The green items (B, F) are very relevant, the yellow items (C, E) are somewhat relevant and the red items (A, D) are not relevant. |
Approximate Metric Optimization
The quality of a ranking is commonly evaluated using ranking metrics, e.g., the normalized discounted cumulative gain (NDCG). An important objective of LTR is to optimize a neural network so that it scores highly on ranking metrics. However, ranking metrics like NDCG can present challenges because they are often discontinuous and flat, so stochastic gradient descent cannot directly be applied to these metrics. Rax provides state-of-the-art approximation techniques that make it possible to produce differentiable surrogates to ranking metrics that permit optimization via gradient descent. The figure below illustrates the use of rax.approx_t12n
, a function transformation unique to Rax, which allows for the NDCG metric to be transformed into an approximate and differentiable form.
![]() |
Using an approximation technique from Rax to transform the NDCG ranking metric into a differentiable and optimizable ranking loss (approx_t12n and gumbel_t12n ). |
First, notice how the NDCG metric (in green) is flat and discontinuous, making it hard to optimize using stochastic gradient descent. By applying the rax.approx_t12n
transformation to the metric, we obtain ApproxNDCG, an approximate metric that is now differentiable with well-defined gradients (in red). However, it potentially has many local optima — points where the loss is locally optimal, but not globally optimal — in which the training process can get stuck. When the loss encounters such a local optimum, training procedures like stochastic gradient descent will have difficulty improving the neural network further.
To overcome this, we can obtain the gumbel-version of ApproxNDCG by using the rax.gumbel_t12n
transformation. This gumbel version introduces noise in the ranking scores which causes the loss to sample many different rankings that may incur a non-zero cost (in blue). This stochastic treatment may help the loss escape local optima and often is a better choice when training a neural network on a ranking metric. Rax, by design, allows the approximate and gumbel transformations to be freely used with all metrics that are offered by the library, including metrics with a top-k cutoff value, like recall or precision. In fact, it is even possible to implement your own metrics and transform them to obtain gumbel-approximate versions that permit optimization without any extra effort.
Ranking in the JAX Ecosystem
Rax is designed to integrate well in the JAX ecosystem and we prioritize interoperability with other JAX-based libraries. For example, a common workflow for researchers that use JAX is to use TensorFlow Datasets to load a dataset, Flax to build a neural network, and Optax to optimize the parameters of the network. Each of these libraries composes well with the others and the composition of these tools is what makes working with JAX both flexible and powerful. For researchers and practitioners of ranking systems, the JAX ecosystem was previously missing LTR functionality, and Rax fills this gap by providing a collection of ranking losses and metrics. We have carefully constructed Rax to function natively with standard JAX transformations such as jax.jit
and jax.grad
and various libraries like Flax and Optax. This means that users can freely use their favorite JAX and Rax tools together.
Ranking with T5
While giant language models such as T5 have shown great performance on natural language tasks, how to leverage ranking losses to improve their performance on ranking tasks, such as search or question answering, is under-explored. With Rax, it is possible to fully tap this potential. Rax is written as a JAX-first library, thus it is easy to integrate it with other JAX libraries. Since T5X is an implementation of T5 in the JAX ecosystem, Rax can work with it seamlessly.
To this end, we have an example that demonstrates how Rax can be used in T5X. By incorporating ranking losses and metrics, it is now possible to fine-tune T5 for ranking problems, and our results indicate that enhancing T5 with ranking losses can offer significant performance improvements. For example, on the MS-MARCO QNA v2.1 benchmark we are able to achieve a +1.2% NDCG and +1.7% MRR by fine-tuning a T5-Base model using the Rax listwise softmax cross-entropy loss instead of a pointwise sigmoid cross-entropy loss.
![]() |
Fine-tuning a T5-Base model on MS-MARCO QNA v2.1 with a ranking loss (softmax, in blue) versus a non-ranking loss (pointwise sigmoid, in red). |
Conclusion
Overall, Rax is a new addition to the growing ecosystem of JAX libraries. Rax is entirely open source and available to everyone at github.com/google/rax. More technical details can also be found in our paper. We encourage everyone to explore the examples included in the github repository: (1) optimizing a neural network with Flax and Optax, (2) comparing different approximate metric optimization techniques, and (3) how to integrate Rax with T5X.
Acknowledgements
Many collaborators within Google made this project possible: Xuanhui Wang, Zhen Qin, Le Yan, Rama Kumar Pasumarthi, Michael Bendersky, Marc Najork, Fernando Diaz, Ryan Doherty, Afroz Mohiuddin, and Samer Hassan.
Source: Google AI Blog
Build apps for the new Samsung devices

Posted by Diana Wong (Android Product Manager), Kseniia Shumelchyk (Developer Relations Engineer) and Sara Vickerman (Android Developer Marketing)
This week, Samsung launched the latest devices to come to the Android ecosystem at their Galaxy Unpacked event. If you haven’t already, check out their two new foldables, the Galaxy Z Fold4 and Z Flip4, and their new lineup of watches running on Wear OS, the Galaxy Watch5 series. You can learn more about their announcements here.
With the excitement around these new devices, there's never been a better time to invest in making sure your app has an amazing experience for users, on large screens or Wear OS! Here’s what you need to know to get started:
Get your apps ready for foldables, like the Galaxy Z Fold4 and Z Flip4
With their unique foldable experience, the Galaxy Z Flip4 and Z Fold4 are great examples of how Android devices come in all shapes and sizes. The Z Fold4 is the latest in large screen devices, a category that continues to see impressive growth. Active large screen users are approaching 270 million, making it a great time to optimize your apps for tablets, foldables and Chrome OS.Last year, we launched Android 12L, a feature drop designed to make Android 12 even better on tablets and foldable devices, and Samsung’s Galaxy Z Fold4 will be the first device to run 12L out of the box! Android 12L includes UI updates tailor-made for large screens, improvements to the multitasking experience, and enhancements to compatibility mode so your app looks better out of the box. Since 12L, we also launched Android 13, which includes all these large screen updates and more.
Get started building for foldables by checking out the documentation. The Z Fold4 and Z Flip4 can be used in multiple different folded states, like Samsung’s “flex mode” where you can go hands-free when doing anything from watching a show to taking a photo. To get your app looking great however it’s folded, you can use the Jetpack WindowManager library to make your app fold aware and test your app on foldables. And finally, the large screen app quality guidelines is a comprehensive set of checklists to help make your app the best it can be across an ever expanding ecosystem of large screen devices.
Build exceptional Wear OS apps
The Wear OS platform expanded this week with the new and improved Galaxy Watch5 series. This lineup of devices builds on Samsung’s commitment to the wearable platform, which we saw last year when they launched Wear OS Powered by Samsung on the Galaxy Watch4 series.If you’re looking to get started building for the latest Galaxy Watch 5 series, or any other Wear OS device, now is a great time to check out version 1.0 of Compose for Wear OS. This is the first stable release of our modern declarative UI toolkit designed to make building apps for Wear OS easier, faster, and more intuitive. The toolkit brings the best of Jetpack Compose to Wear OS, accelerating the development process so you can create beautiful apps with fewer lines of code.
The 1.0 release streamlines UI development by following the declarative approach and offering powerful Kotlin syntax. It also provides a rich set of UI components optimized for the watch experience and is accompanied by many powerful tools in Android Studio to streamline UI iteration. That’s why Compose for Wear OS is our recommended approach for building user interfaces for Wear OS apps.
We’ve built a set of materials to help you get started with Compose for Wear OS! Check out our curated learning pathway for a step-by-step journey, documentation including a quick start guide, the Compose for Wear OS codelab for hands-on experience, and samples available on Github.
Similarly to Compose for Wear OS, we’re building Wear OS Tile Components to make it faster and easier to build tiles. Tiles provide Wear OS users glanceable access to the information and actions they need in order to get things done quickly and they are one of the most used features on Wear OS. This update brings material components and layouts so you can create Tiles that embrace the latest Material design for Wear OS. Right now this is in beta, but keep a lookout for the launch announcement!
Another launch announcement to watch out for is Android Studio Dolphin, the latest release from Android Studio. Check out these features designed to make wearable app development easier:
- Updated Wear OS emulator toolbar which now includes buttons and gestures available on Wear OS devices, such as palm and tilting and simulating two physical buttons.
- Emulator pairing assistant to pair multiple Wear OS devices with a single virtual or physical phone. Android Studio remembers pairings after being closed and allows you to see Wear devices in the Device Manager.
- Direct surface launch that allows you to create run/debug configurations for Wear OS tiles, watch faces, and complications, and launch them directly from Android Studio.
There’s never been a better time to start optimizing!
Form factors are having a major moment this year and Google is committed to helping you optimize and build across form factors with new content and tools, including sessions and workshops from this year’s Google I/O and new Android Studio features. Plus, we have Material Design guidance for large screens and Wear OS to help you in your optimization journey.From the Watch5 series to the Z Fold4, Samsung’s Galaxy Unpacked brought us innovations across screen sizes and types. Prepare your app so it looks great across the entire Android device ecosystem!
Source: Android Developers Blog
Build apps for the new Samsung devices

Posted by Diana Wong (Android Product Manager), Kseniia Shumelchyk (Developer Relations Engineer) and Sara Vickerman (Android Developer Marketing)
This week, Samsung launched the latest devices to come to the Android ecosystem at their Galaxy Unpacked event. If you haven’t already, check out their two new foldables, the Galaxy Z Fold4 and Z Flip4, and their new lineup of watches running on Wear OS, the Galaxy Watch5 series. You can learn more about their announcements here.
With the excitement around these new devices, there's never been a better time to invest in making sure your app has an amazing experience for users, on large screens or Wear OS! Here’s what you need to know to get started:
Get your apps ready for foldables, like the Galaxy Z Fold4 and Z Flip4
With their unique foldable experience, the Galaxy Z Flip4 and Z Fold4 are great examples of how Android devices come in all shapes and sizes. The Z Fold4 is the latest in large screen devices, a category that continues to see impressive growth. Active large screen users are approaching 270 million, making it a great time to optimize your apps for tablets, foldables and Chrome OS.Last year, we launched Android 12L, a feature drop designed to make Android 12 even better on tablets and foldable devices, and Samsung’s Galaxy Z Fold4 will be the first device to run 12L out of the box! Android 12L includes UI updates tailor-made for large screens, improvements to the multitasking experience, and enhancements to compatibility mode so your app looks better out of the box. Since 12L, we also launched Android 13, which includes all these large screen updates and more.
Get started building for foldables by checking out the documentation. The Z Fold4 and Z Flip4 can be used in multiple different folded states, like Samsung’s “flex mode” where you can go hands-free when doing anything from watching a show to taking a photo. To get your app looking great however it’s folded, you can use the Jetpack WindowManager library to make your app fold aware and test your app on foldables. And finally, the large screen app quality guidelines is a comprehensive set of checklists to help make your app the best it can be across an ever expanding ecosystem of large screen devices.
Build exceptional Wear OS apps
The Wear OS platform expanded this week with the new and improved Galaxy Watch5 series. This lineup of devices builds on Samsung’s commitment to the wearable platform, which we saw last year when they launched Wear OS Powered by Samsung on the Galaxy Watch4 series.If you’re looking to get started building for the latest Galaxy Watch 5 series, or any other Wear OS device, now is a great time to check out version 1.0 of Compose for Wear OS. This is the first stable release of our modern declarative UI toolkit designed to make building apps for Wear OS easier, faster, and more intuitive. The toolkit brings the best of Jetpack Compose to Wear OS, accelerating the development process so you can create beautiful apps with fewer lines of code.
The 1.0 release streamlines UI development by following the declarative approach and offering powerful Kotlin syntax. It also provides a rich set of UI components optimized for the watch experience and is accompanied by many powerful tools in Android Studio to streamline UI iteration. That’s why Compose for Wear OS is our recommended approach for building user interfaces for Wear OS apps.
We’ve built a set of materials to help you get started with Compose for Wear OS! Check out our curated learning pathway for a step-by-step journey, documentation including a quick start guide, the Compose for Wear OS codelab for hands-on experience, and samples available on Github.
Similarly to Compose for Wear OS, we’re building Wear OS Tile Components to make it faster and easier to build tiles. Tiles provide Wear OS users glanceable access to the information and actions they need in order to get things done quickly and they are one of the most used features on Wear OS. This update brings material components and layouts so you can create Tiles that embrace the latest Material design for Wear OS. Right now this is in beta, but keep a lookout for the launch announcement!
Another launch announcement to watch out for is Android Studio Dolphin, the latest release from Android Studio. Check out these features designed to make wearable app development easier:
- Updated Wear OS emulator toolbar which now includes buttons and gestures available on Wear OS devices, such as palm and tilting and simulating two physical buttons.
- Emulator pairing assistant to pair multiple Wear OS devices with a single virtual or physical phone. Android Studio remembers pairings after being closed and allows you to see Wear devices in the Device Manager.
- Direct surface launch that allows you to create run/debug configurations for Wear OS tiles, watch faces, and complications, and launch them directly from Android Studio.
There’s never been a better time to start optimizing!
Form factors are having a major moment this year and Google is committed to helping you optimize and build across form factors with new content and tools, including sessions and workshops from this year’s Google I/O and new Android Studio features. Plus, we have Material Design guidance for large screens and Wear OS to help you in your optimization journey.From the Watch5 series to the Z Fold4, Samsung’s Galaxy Unpacked brought us innovations across screen sizes and types. Prepare your app so it looks great across the entire Android device ecosystem!