Google Docs offers several existing capabilities to help you write faster and with more confidence, such as grammar suggestions, smart compose, and smart reply. Today, we’re introducing the next generation of writing suggestions, powered by AI: Proofread.
In addition to spelling and grammar, Proofread will provide a list of suggestions you can review and accept, including:
Conciseness: Suggestions for making your writing more concise
Active voice: Suggestions for making your writing more clear to readers
Wording: Suggestions for making words sound more dynamic or formal in context
Sentence Split: Suggestions for splitting up complex sentences to make your writing easier to read
Who’s impacted
End users
Why you’d use it
Proofread is our latest innovation that helps you more easily and quickly compose high-quality content in Google Docs. It can help improve conciseness and make your documents clearer and easier to read. These suggestions are non-intrusive and won’t clutter your document as you compose and collaborate — you can easily review all suggestions directly from the Proofread sidebar whenever you’re ready and decide which suggestions are most appropriate for you to accept. Plus, you can customize the types of suggestions you’ll see via the filter option, helping ensure we’re giving you the most relevant suggestions.
Additional details
If you’re currently participating in alpha testing for Proofread, you’ll need to purchase the Duet AI for Workspace Enterprise add-on to continue using this feature. If not, you will no longer be able to access the feature beginning on August 29, 2023.
This feature will be ON by default and can be turned off in Docs by going to Tools > Proofread. Visit the Help Center to learn more about using Proofread in Docs.
You can open the Proofread sidebar by selecting this icon:
There will be a bluedot alongside this icon when there are writing suggestions for your document:
If you have Duet AI for Google Workspace, you can now translate English and non-English speech to and from the following languages:
Chinese (Mandarin, Traditional)
Hindi (India)
Thai
Italian
Russian
This adds to the existing list of languages: French, German, Spanish, Portuguese, Japanese, Swedish, Chinese (Simplified), Vietnamese, Dutch, Turkish, Indonesian.
Additionally, you can translate non-English speech to and from the following languages:
Meeting participants: You can use live translated captions if you’re using an eligible Google Workspace edition or the meeting is organized by a user with an eligible Google Workspace edition.
Note: Translated captions in Google Meet will be available exclusively to Duet AI for Google Workspace Enterprise add-on users at a future date. We will provide a minimum of 6 months notice to existing Google Workspace customers before this change takes place.
This announcement was made at Google Cloud Next ‘23. Visit the Workspace Blog to learn more about the next wave of AI innovation in Workspace.
What’s changing
If your organization uses a mixture of messaging tools, you can now use Mio to enable messaging interoperability between Google Chat and other platforms such as Slack and Teams. This is available now in beta, with general availability expected in early 2024.
Who’s impacted
Admins
Why it matters
We know that communication and collaboration happens over multiple channels and tools. This can cause missed messages, silos of communication, and a frustrating experience monitoring multiple chat tools.
That’s why we’re partnering with Mio, a leading provider of collaborative interoperability solutions helping customers enhance their productivity by streamlining communication across multiple channels. You can leverage Mio to create a seamless experience between Chat and the various tools you need to get your work done.
Getting started
Admins: To request beta access or to discuss pricing for your organization, reach out to your account rep or submit this form to contact Mio.
Availability
Note: In order to leverage Mio’s interoperability capabilities, Mio licenses are required.
General Availability:
Available to Google Workspace Business Starter, Business Standard, Business Plus, Essentials Starter, Enterprise Essentials, Enterprise Essentials Plus, Enterprise Standard, Enterprise Plus, Frontline Starter, Frontline Standard, and Nonprofits customers
There are a variety of ways to change your background in Google Meet today, including immersive backgrounds, seasonal backgrounds, and more. Beginning today, you can now use the power of Duet AI to generate completely unique and bespoke meeting backgrounds.
Create your custom background from the Meet greenroom
Who’s impacted
Admins and end users
Why you’d use it
In addition to protecting your privacy during video calls, you can generate a background that enhances the meeting itself. For example, you can create a background specific to your clients and the industry they work with, helping to deepen the connection through visual reinforcement. You can create backgrounds that set the tone for kicking off an important project, or one befitting of celebrating a company or colleague milestone.
To enable or disable immersive backgrounds and styles, go to Admin console > Apps > Google Workspace > Settings for Google Meet > Meet video setting. Visit the Help Center to learn more about how to control whether users can change their backgrounds in Meet.
Posted by Yujin Tang and Wenhao Yu, Research Scientists, Google
Simple and effective interaction between human and quadrupedal robots paves the way towards creating intelligent and capable helper robots, forging a future where technology enhances our lives in ways beyond our imagination. Key to such human-robot interaction systems is enabling quadrupedal robots to respond to natural language instructions. Recent developments in large language models (LLMs) have demonstrated the potential to perform high-level planning. Yet, it remains a challenge for LLMs to comprehend low-level commands, such as joint angle targets or motor torques, especially for inherently unstable legged robots, necessitating high-frequency control signals. Consequently, most existingwork presumes the provision of high-level APIs for LLMs to dictate robot behavior, inherently limiting the system’s expressive capabilities.
In “SayTap: Language to Quadrupedal Locomotion”, we propose an approach that uses foot contact patterns (which refer to the sequence and manner in which a four-legged agent places its feet on the ground while moving) as an interface to bridge human commands in natural language and a locomotion controller that outputs low-level commands. This results in an interactive quadrupedal robot system that allows users to flexibly craft diverse locomotion behaviors (e.g., a user can ask the robot to walk, run, jump or make other movements using simple language). We contribute an LLM prompt design, a reward function, and a method to expose the SayTap controller to the feasible distribution of contact patterns. We demonstrate that SayTap is a controller capable of achieving diverse locomotion patterns that can be transferred to real robot hardware.
SayTap method
The SayTap approach uses a contact pattern template, which is a 4 X T matrix of 0s and 1s, with 0s representing an agent’s feet in the air and 1s for feet on the ground. From top to bottom, each row in the matrix gives the foot contact patterns of the front left (FL), front right (FR), rear left (RL) and rear right (RR) feet. SayTap’s control frequency is 50 Hz, so each 0 or 1 lasts 0.02 seconds. In this work, a desired foot contact pattern is defined by a cyclic sliding window of size Lw and of shape 4 X Lw. The sliding window extracts from the contact pattern template four foot ground contact flags, which indicate if a foot is on the ground or in the air between t + 1 and t + Lw. The figure below provides an overview of the SayTap method.
SayTap introduces these desired foot contact patterns as a new interface between natural language user commands and the locomotion controller. The locomotion controller is used to complete the main task (e.g., following specified velocities) and to place the robot’s feet on the ground at the specified time, such that the realized foot contact patterns are as close to the desired contact patterns as possible. To achieve this, the locomotion controller takes the desired foot contact pattern at each time step as its input in addition to the robot’s proprioceptive sensory data (e.g., joint positions and velocities) and task-related inputs (e.g., user-specified velocity commands). We use deep reinforcement learning to train the locomotion controller and represent it as a deep neural network. During controller training, a random generator samples the desired foot contact patterns, the policy is then optimized to output low-level robot actions to achieve the desired foot contact pattern. Then at test time a LLM translates user commands into foot contact patterns.
SayTap approach overview.
SayTap uses foot contact patterns (e.g., 0 and 1 sequences for each foot in the inset, where 0s are foot in the air and 1s are foot on the ground) as an interface that bridges natural language user commands and low-level control commands. With a reinforcement learning-based locomotion controller that is trained to realize the desired contact patterns, SayTap allows a quadrupedal robot to take both simple and direct instructions (e.g., “Trot forward slowly.”) as well as vague user commands (e.g., “Good news, we are going to a picnic this weekend!”) and react accordingly.
We demonstrate that the LLM is capable of accurately mapping user commands into foot contact pattern templates in specified formats when given properly designed prompts, even in cases when the commands are unstructured or vague. In training, we use a random pattern generator to produce contact pattern templates that are of various pattern lengths T, foot-ground contact ratios within a cycle based on a given gait type G, so that the locomotion controller gets to learn on a wide distribution of movements leading to better generalization. See the paper for more details.
Results
With a simple prompt that contains only three in-context examples of commonly seen foot contact patterns, an LLM can translate various human commands accurately into contact patterns and even generalize to those that do not explicitly specify how the robot should react.
SayTap prompts are concise and consist of four components: (1) general instruction that describes the tasks the LLM should accomplish; (2) gait definition that reminds the LLM of basic knowledge about quadrupedal gaits and how they can be related to emotions; (3) output format definition; and (4) examples that give the LLM chances to learn in-context. We also specify five velocities that allow a robot to move forward or backward, fast or slow, or remain still.
General instruction block
You are a dog foot contact pattern expert.
Your job is to give a velocity and a foot contact pattern based on the input.
You will always give the output in the correct format no matter what the input is.
Gait definition block
The following are description about gaits:
1. Trotting is a gait where two diagonally opposite legs strike the ground at the same time.
2. Pacing is a gait where the two legs on the left/right side of the body strike the ground at the same time.
3. Bounding is a gait where the two front/rear legs strike the ground at the same time. It has a longer suspension phase where all feet are off the ground, for example, for at least 25% of the cycle length. This gait also gives a happy feeling.
Output format definition block
The following are rules for describing the velocity and foot contact patterns:
1. You should first output the velocity, then the foot contact pattern.
2. There are five velocities to choose from: [-1.0, -0.5, 0.0, 0.5, 1.0].
3. A pattern has 4 lines, each of which represents the foot contact pattern of a leg.
4. Each line has a label. "FL" is front left leg, "FR" is front right leg, "RL" is rear left leg, and "RR" is rear right leg.
5. In each line, "0" represents foot in the air, "1" represents foot on the ground.
Example block
Input: Trot slowly
Output: 0.5
FL: 11111111111111111000000000
FR: 00000000011111111111111111
RL: 00000000011111111111111111
RR: 11111111111111111000000000
Input: Bound in place
Output: 0.0
FL: 11111111111100000000000000
FR: 11111111111100000000000000
RL: 00000011111111111100000000
RR: 00000011111111111100000000
Input: Pace backward fast
Output: -1.0
FL: 11111111100001111111110000
FR: 00001111111110000111111111
RL: 11111111100001111111110000
RR: 00001111111110000111111111
Input:
SayTap prompt to the LLM. Texts in blue are used for illustration and are not input to LLM.
Following simple and direct commands
We demonstrate in the videos below that the SayTap system can successfully perform tasks where the commands are direct and clear. Although some commands are not covered by the three in-context examples, we are able to guide the LLM to express its internal knowledge from the pre-training phase via the “Gait definition block” (see the second block in our prompt above) in the prompt.
Following unstructured or vague commands
But what is more interesting is SayTap’s ability to process unstructured and vague instructions. With only a little hint in the prompt to connect certain gaits with general impressions of emotions, the robot bounds up and down when hearing exciting messages, like “We are going to a picnic!” Furthermore, it also presents the scenes accurately (e.g., moving quickly with its feet barely touching the ground when told the ground is very hot).
Conclusion and future work
We present SayTap, an interactive system for quadrupedal robots that allows users to flexibly craft diverse locomotion behaviors. SayTap introduces desired foot contact patterns as a new interface between natural language and the low-level controller. This new interface is straightforward and flexible, moreover, it allows a robot to follow both direct instructions and commands that do not explicitly state how the robot should react.
One interesting direction for future work is to test if commands that imply a specific feeling will allow the LLM to output a desired gait. In the gait definition block shown in the results section above, we provide a sentence that connects a happy mood with bounding gaits. We believe that providing more information can augment the LLM’s interpretations (e.g., implied feelings). In our evaluation, the connection between a happy feeling and a bounding gait led the robot to act vividly when following vague human commands. Another interesting direction for future work is to introduce multi-modal inputs, such as videos and audio. Foot contact patterns translated from those signals will, in theory, still work with our pipeline and will unlock many more interesting use cases.
Acknowledgements
Yujin Tang, Wenhao Yu, Jie Tan, Heiga Zen, Aleksandra Faust and Tatsuya Harada conducted this research. This work was conceived and performed while the team was in Google Research and will be continued at Google DeepMind. The authors would like to thank Tingnan Zhang, Linda Luu, Kuang-Huei Lee, Vincent Vanhoucke and Douglas Eck for their valuable discussions and technical support in the experiments.
Google Meet already offers a variety of tools to help you appear your best during meetings, including automatic framing and lighting adjustments. Today, we’re taking these enhancements a step further by introducing studio look in Google Meet, powered by Duet AI. Studio look enhances the quality of your portrait by reducing noise and increasing sharpness, bringing you into focus more clearly.
In the hybrid work world, our video feeds can be hindered by low light or lower quality webcams. Studio look helps ensure you’re coming across crisp and in focus, helping strengthen human connections in a hybrid work world.
For illustrative purposes only — to use Studio Light, go to settings > video > studio look.
Getting started
Admins: There is no admin control for this feature.
End users: This feature will be available by default. You can turn studio look on before a call or during a call from the three-dot settings menu. Visit the Help Center to learn more about improving your video experience with studio look.
The Stable and Extended stable channels has been updated to 116.0.5845.140 for Mac and Linux and 116.0.5845.140/.141 for Windows, which will roll out over the coming days/weeks. A full list of changes in this build is available in the log.
Security Fixes and Rewards
Note: Access to bug details and links may be kept restricted until a majority of users are updated with a fix. We will also retain restrictions if the bug exists in a third party library that other projects similarly depend on, but haven’t yet fixed.
This update includes 1 security fix. Below, we highlight fixes that were contributed by external researchers. Please see the Chrome Security Page for more information.
[$TBD][1472492] High CVE-2023-4572: Use after free in MediaStream. Reported by fwnfwn(@_fwnfwn) on 2023-08-12
We would also like to thank all security researchers that worked with us during the development cycle to prevent security bugs from ever reaching the stable channel.
Interested in switching release channels? Find out how here. If you find a new issue, please let us know by filing a bug. The community help forum is also a great place to reach out for help or learn about common issues.
Through Duet AI for Google Workspace, you can now use Google Sheets to quickly get started organizing your data in a straightforward, highly-visual manner. By simply describing what you want to accomplish using Help me organize, Sheets will generate a plan, tracker, budget, schedule, and more in the form of a template with sample text and smart chips that are easily editable. Once inserted into the grid, you can make the template your own by assigning owners, updating statuses, and bringing in relevant information such as location or time.
Whether you’re a travel agent planning a trip for a company retreat or a team lead establishing onboarding plans for new hires, Help me organize gives you a running start for creating highly-organized plans, trackers, timelines and more.
Who’s impacted
End users
Why it’s important
Through the power of Duet AI for Google Workspace, you can now get started organizing information faster than ever before in Google Sheets.
Additional details
Help me organize makes template tables with sample data, and does not have access to the data within your spreadsheet or elsewhere.
Getting started
Admins: This feature will be ON by default and there is no admin control for this feature. Visit the Help Center to learn more about Duet AI for Google Workspace Enterprise.
End users:
To use the Help me organize feature, write a description of what you need in the sidebar > click Create > View the generated table preview.
Note: You must write a minimum of 2 characters in the sidebar. The maximum number of characters is 150.
The Help me organize sidebar is surfaced for any empty tab (new or existing Sheets files).
You can re-open Help me organize from the insert menu or the toolbar via a new icon. You can also re-trigger the sidebar when adding a new tab.
This announcement was made at Google Cloud Next ‘23. Visit the Workspace Blog to learn more about the next wave of AI innovation in Workspace.
What’s changing
If you’re a Google Workspace admin, you can now track the carbon footprint and gross emissions of using Google Workspace. This means you can track the emissions impact of tools such as Google Meet, Gmail, Google Docs, and more.
Who’s impacted
Admins
Why it matters
We’ve heard from our customers that it would be helpful to understand their organization’s carbon emissions from Google Workspace product usage. For many of our customers, sustainability has become an increasingly important factor for IT-related decision making and resource allocation.
We’re making this information easily accessible to our customers in the Admin console. Using the Carbon Footprint reporting tool, you can:
View aggregated emission data for your organization over time, broken down by service. You’ll be able to sort this information, month over month as well.
You can break down emission data by service, such as Gmail or Drive .
Additional details
At Google, we continue to make product and operational improvements to reduce environmental impact. Our goal is to run on carbon-free energy, 24/7, at all of our data centers and cloud regions by 2030. Plus, we’re sharing technology, methods, and funding to enable organizations around the world to transition to more carbon-free and sustainable systems — see here for more information about our sustainability commitments.
Google uses the Greenhouse Gas Protocol, the global standard for carbon accounting to generate the Carbon footprint reports. We recommend that admins familiarize themselves with the GHG terminology — you can find more information in our Help Center or the video below.
Getting started
Admins: You can find your Carbon Footprint report in the Admin console under Reporting > Carbon footprint. Visit the Help Center to learn more about the Workspace Carbon Footprint.