Control the display of user availability across Google Workspace products with a new admin control

Quick launch summary

We’re introducing a new Admin setting that controls whether a user’s Google Calendar status is displayed across Google Workspace products. For example, when the setting is turned OFF, end users will no longer see if their colleagues are out of office in products such as Google Chat or Gmail.

We hope this new setting gives admins greater control over what user information is displayed across Google Workspace products.

Getting started


Rollout pace

Availability

  • Available to Google Workspace Business Plus, Enterprise Standard, Enterprise Plus, and Education Plus
  • Not Available to Google Workspace Essentials, Business Starter, Business Standard, Enterprise Essentials, Education Fundamentals and Nonprofits, as well as G Suite Basic and Business customers

Resources

Canadian Actor Simu Liu named ambassador for Google Pixel 6

Liu working on ad and social media projects to promote Google’s flagship smartphone 




Google Canada has recruited film and TV star Simu Liu to act as the spokesperson for their latest smartphones, the Pixel 6 and Pixel 6 Pro

Liu, fresh off his star-making role playing superhero Shang-Chi in the Marvel Cinematic Universe film Shang-Chi and The Legend of The Ten Rings, will now collaborate with Google to create content for Canadian audiences that celebrates the latest generation of Pixel smartphones, which launched in Canada on October 28. In the same way that Simu has helped spearhead conversations about representation in film and TV, Simu’s vision seamlessly lines up with Google’s values of helpfulness and inclusivity. Liu will now shoot a series of digital ad and social media spots that contextualize some of the new phones’ most interesting features in a style that puts Simu’s personality front and centre. 

“In my industry, everyone tends to gravitate towards one way of communicating. But I think it’s time for me to make a change and shake things up with Google.” says Liu. “I’ve been trying out the Pixel 6 Pro for a couple weeks now, and I’m excited to share my thoughts on Pixel with the world.” 

Born in China before moving to Canada at the age of five, Liu first gained acclaim on the CBC comedy series Kim’s Convenience. In both his acting career and personal life, Liu has broken barriers around Asian representation, serving as a champion for his community. Google Canada has identified Simu’s important role in the space of representation, making him a perfect fit for the Pixel 6. 

“Simu Liu has captured the hearts of Canadians with his inspirational story and newfound stardom, and we’re excited to continue telling that story through the Pixel 6,”says Laura Pearce, Head of Marketing, Google Canada. "The Pixel 6 is our most progressive phone, bringing out what makes people unique and connecting them to those around them; perfectly aligned with the message that Simu is bringing to fans in Canada and across the world." 

The Pixel 6 and Pixel 6 Pro are a culmination of Google’s long-term investments in AI and machine learning, featuring its first System on a Chip (SoC), the Google Tensor. With the Pixel 6 being the first true all-Google phone inside and out, the powerful Google Tensor chip unlocks the full potential of the phone’s camera system, Android user interface and AI that Google is known for. 

Google Tensor allows the Pixel 6 to adapt to each individual user and provide some of the most accommodating and inclusive features ever offered on a Pixel phone. This includes Real Tone technology, which uses camera tuning models and algorithms to help more accurately highlight the nuances and details of darker skin tones, and Live Translate, which combines Pixel’s speech recognition with Google’s translation capabilities to help users to communicate in languages other than their own across a number of applications². 

Liu will film the new spots and social media content later this month, with the content beginning to appear in Canada and his social media channels (@simuliu) this holiday season. The Google Pixel 6 and Pixel 6 Pro are now available on the Google Store

1 Based on a survey conducted by Google using pre-production devices in early September 2021. 
2 Not available in all languages or countries. Not available on all media or apps. See g.co/pixel/livetranslate for more information.

2021’s Top YouTube Videos had Canadians laughing, learning, Minecrafting and singing

As Canadians navigated their way through the ongoing pandemic and the country's gradual reopening, they turned to YouTube to connect with the world, gain new skills and share their passions. 

This year's Top Canadian Creators, Top Breakout Creators and Shorts Rising Stars were all about comedy and learning. Breakthrough creators Kallmekris and Jeenie.Weenie found ways to make us laugh, while NileRed brought out our inquisitive side as we learned about all the ways that chemistry is cool. And many Canadians wanted to brush up on their English language skills turning to ENGLISH with James to get them there. Our love for Minecraft became apparent as aCookieGod and MrFudgeMonkeyz dug, built, and crafted their own virtual worlds for millions of viewers.






 Top Canadian Creators

  1. Kallmekris

  2. Linus Tech Tips

  3. MadFit

  4. aCookieGod

  5. NileRed





Top Canadian Breakout Creators

  1. Kallmekris

  2. aCookieGod

  3. Girl With The Dogs

  4. ENGLISH with James - engVid

  5. Jason Fung





Canadian Shorts Creators - Rising Stars

  1. Jeenie.Weenie

  2. Chris Ramsay

  3. NileRed Shorts

  4. Luke davidson

  5. MrFudgeMonkeyz



When it came to Top Trending Videos, we saw viewers were hungry for excitement and a thrill after being cooped up at home, turning to creators like MrBeast who spent 50 hours buried underground in a glass coffin (not for the claustrophobic!) as one of his many elaborate stunts. We laughed together as we categorized our friends into game night stereotypes, sobbed together while rooting for Nightbirde, and collectively spent an afternoon with royalty sipping tea on top of an open bus through LA.



Top Trending Videos in Canada

  1. MrBeast, I Spent 50 Hours Buried Alive

  2. Mark Rober, Glitterbomb Trap Catches Phone Scammer (who gets arrested) 

  3. Dream, Minecraft Speedrunner VS 5 Hunters 

  4. Sidemen, SIDEMEN TINDER IN REAL LIFE 3 

  5. NFL, The Weeknd's FULL Pepsi Super Bowl LV Halftime Show 

  6. The Late Late Show with James Corden, An Afternoon with Prince Harry & James Corden 

  7. America's Got Talent, Golden Buzzer: Nightbirde's Original Song Makes Simon Cowell Emotional

  8. Forge Labs, I Spent 100 Days in a Zombie Apocalypse in Minecraft ... Here's What Happened 

  9. Dude Perfect, Game Night Stereotypes 

  10. SAD-ist, Hog Hunt | Dream SMP Animation 



In Music, 2021 was the Year of The Weeknd and Justin Bieber, with both homegrown Canadian music superstars releasing top hits ranking high on the list. The Weeknd dropped his darkly comedic new video for Save Your Tears, during which he delivers a performance to a masked audience and we jammed along to Bieber's chart-topping, neon technicolor sensation “ Peaches,” featuring Canadian-talent Daniel Caesar and Giveon . Both songs were part of a soundtrack of swoon-worthy songs that Canadians had on repeat this year. From Bruno Mars to Doja Cat, you could say some of these tracks were smooth as Butter.


Olivia Rodrigo had a breakout year, and brought home the # 4 and # 5 spots as a heartbroken teen who finally gets their drivers license and as a cheerleader exacting revenge in good 4 u.

Top Music Videos in Canada


  1. TheWeekndVEVO, The Weeknd - Save Your Tears (Official Music Video) 

  2. JustinBieberVEVO, Justin Bieber - Peaches ft. Daniel Caesar, Giveon 

  3. LilNasXVEVO, Lil Nas X - MONTERO (Call Me By Your Name) (Official Video) 

  4. OliviaRodrigoVEVO, Olivia Rodrigo - drivers license (Official Video) 

  5. OliviaRodrigoVEVO, Olivia Rodrigo - good 4 u (Official Video) 

  6. HYBE LABELS, BTS (방탄 소년단) 'Butter' Official MV 

  7. Bruno Mars, Bruno Mars, Anderson .Paak, Silk Sonic - Leave the Door Open [Official Video] 

  8. Pooh Shiesty, Pooh Shiesty - Back In Blood (feat. Lil Durk) [Official Music Video] 

  9. dojacatVEVO, Doja Cat - Kiss Me More (Official Video) ft. SZA 

  10. PoloGVEVO, Polo G - RAPSTAR (Official Video)



Congratulations and thank you to all the creators that helped bring joy and entertainment to Canadians this year. Want to learn more about the creators and videos that made this year's top lists? Head to the Culture & Trends site, where you can find all the global moments that defined 2021.


The path to Malaysia’s digital potential

When the COVID-19 pandemic hit, Mohd Zaid, from Kajang, Malaysia, felt the pressure of providing for his family in an uncertain environment. To bring in some extra income, he turned first to one of his personal passions — making soy wax candles infused with scented oils — and then he turned to the internet. After learning digital marketing skills through a Grow with Google course, Zaid was able to go beyond word-of-mouth sales and promote his candles online through Google Ads and Search. His revenue jumped 70%.

Zaid is one of a growing number of Malaysian entrepreneurs embracing a more digital economy. Technology has helped Malaysians through the economic effects of the pandemic, enabling people across the country to work, learn and run their businesses in new ways. According to the latest eConomy Southeast Asia report, 81% of all Malaysian internet users now use digital services — including three million people who’ve become new ‘digital consumers’ since the pandemic began. And business owners are adopting technology at a faster pace, using digital tools to serve their customers better. Over 40% of digital merchants in Malaysia believe their businesses wouldn’t have survived the pandemic without digital platforms (the highest proportion anywhere in the region).

Technology is equally important to Malaysia’s long-term future. According to a new report released by AlphaBeta, making the most of digital opportunities could create $61.3 billion in annual economic value for Malaysia by 2030. That’s the equivalent of about 17% of Malaysia’s GDP in 2020.

So the possibilities are enormous — but right now, Malaysia has some catching up to do. Only one-third of Malaysian businesses have a website, compared with 44% globally. The digital economy is also uneven. Some industries, like manufacturing, use technology far more intensively than others, like agriculture, while small businesses face a shortage of workers with the right skills.

Malaysia’s government has developed a Digital Economy Blueprint, aiming to position Malaysia as a regional technology leader by the end of the decade, and the AlphaBeta report sets out three priorities for getting there: digitalizing the public and private sectors, building the nation’s digital talent and promoting digital trade opportunities.

To help, Google Malaysia will continue to expand programs like Mahir Digital Bersama Google, which has already trained more than 36,000 Malaysian small businesses. We’ll keep working to close digital skills gaps through initiatives like Go Digital ASEAN (supported by Google.org and focused on marginalized communities) and AirAsia academy, which provides free digital courses for local small businesses. Through YouTube, we’ll expand our efforts to help Malaysian creators find global audiences and grow revenue for their businesses. And we’ll deepen our efforts with the Ministry of Education to improve digital learning in schools, laying the ground for the next generation of talent.

After a challenging period, I know we can look to the future with confidence — and technology is at the heart of the ambitions we share for our economy and society. We’re looking forward to playing our part in advancing Malaysia’s exciting digital potential together.

Machine learning to make sign language more accessible

Google has spent over twenty years helping to make information accessible and useful in more than 150 languages. And our work is definitely not done, because the internet changes so quickly. About 15% of searches we see are entirely new every day. And when it comes to other types of information beyond words, in many ways, technology hasn’t even begun to scratch the surface of what’s possible. Take one example: sign language.

The task is daunting. There are as many sign languages as there are spoken languages around the world. That’s why, when we started exploring how we could better support sign language, we started small by researching and experimenting with what machine learning models could recognize. We also spoke with members of the Deaf community, as well as linguistic experts. We began combining several ML models to recognize sign language as a sum of its parts — going beyond just hands to include body gestures and facial expressions.

After 14 months of testing with a database of videos for Japanese Sign Language and Hong Kong Sign Language, we launched SignTown: an interactive desktop application that works with a web browser and camera.

SignTown is an interactive web game built to help people to learn about sign language and Deaf culture. It uses machine learning to detect the user's ability to perform signs learned from the game.

Project Shuwa

SignTown is only one component of a broader effort to push the boundaries of technology for sign language and Deaf culture, named “Project Shuwa” after the Japanese word for sign language (“手話”). Future areas of development we’re exploring include building a more comprehensive dictionary across more sign and written languages, as well as collaborating with the Google Search team on surfacing these results to improve search quality for sign languages.

A woman in a black top facing the camera and making a sign with her right hand.

Advances in AI and ML now allow us to reliably detect hands, body poses and facial expressions using any camera inside a laptop or mobile phone. SignTown uses the MediaPipe Holistic model to identify keypoints from raw video frames, which we then feed into a classifier model to determine which sign is the closest match. This all runs inside of the user's browser, powered by Tensorflow.js.

A grid with separate images of four people facing the camera and making signs with their hands.

We open-sourced the core models and tools for developers and researchers to build their own custom models at Google IO 2021. That means anyone who wants to train and deploy their own sign language model has the ability to do so.

At Google, we strive to help build a more accessible world for people with disabilities through technology. Our progress depends on collaborating with the right partners and developers to shape experiments that may one day become stand-alone tools. But it’s equally important that we raise awareness in the wider community to foster diversity and inclusivity. We hope our work in this area with SignTown gets us a little closer to that goal.

Stable Channel Update for Chrome OS

The Stable channel is being updated to 96.0.4664.77 (Platform version: 14268.51.0) for most Chrome OS devices.

If you find new issues, please let us know by visiting our forum or filing a bug. Interested in switching channels Find out how. You can submit feedback using ‘Report an issue...’ in the Chrome menu (3 vertical dots in the upper right corner of the browser). 

Daniel Gagnon,

Google Chrome OS 

MURAL: Multimodal, Multi-task Retrieval Across Languages

For many concepts, there is no direct one-to-one translation from one language to another, and even when there is, such translations often carry different associations and connotations that are easily lost for a non-native speaker. In such cases, however, the meaning may be more obvious when grounded in visual examples. Take, for instance, the word "wedding". In English, one often associates a bride in a white dress and a groom in a tuxedo, but when translated into Hindi (शादी), a more appropriate association may be a bride wearing vibrant colors and a groom wearing a sherwani. What each person associates with the word may vary considerably, but if they are shown an image of the intended concept, the meaning becomes more clear.

The word “wedding” in English and Hindi conveys different mental images. Images are taken from wikipedia, credited to Psoni2402 (left) and David McCandless (right) with CC BY-SA 4.0 license.

With current advances in neural machine translation and image recognition, it is possible to reduce this sort of ambiguity in translation by presenting a text paired with a supporting image. Prior research has made much progress in learning image–text joint representations for high-resource languages, such as English. These representation models strive to encode the image and text into vectors in a shared embedding space, such that the image and the text describing it are close to each other in that space. For example, ALIGN and CLIP have shown that training a dual-encoder model (i.e., one trained with two separate encoders) on image–text pairs using a contrastive learning loss works remarkably well when provided with ample training data.

Unfortunately, such image–text pair data does not exist at the same scale for the majority of languages. In fact, more than 90% of this type of web data belongs to the top-10 highly-resourced languages, such as English and Chinese, with much less data for under-resourced languages. To overcome this issue, one could either try to manually collect image–text pair data for under-resourced languages, which would be prohibitively difficult due to the scale of the undertaking, or one could seek to leverage pre-existing datasets (e.g., translation pairs) that could inform the necessary learned representations for multiple languages.

In “MURAL: Multimodal, Multitask Representations Across Languages”, presented at Findings of EMNLP 2021, we describe a representation model for image–text matching that uses multitask learning applied to image–text pairs in combination with translation pairs covering 100+ languages. This technology could allow users to express words that may not have a direct translation into a target language using images instead. For example, the word “valiha”, refers to a type of tube zither played by the Malagasy people, which lacks a direct translation into most languages, but could be easily described using images. Empirically, MURAL shows consistent improvements over state-of-the-art models, other benchmarks, and competitive baselines across the board. Moreover, MURAL does remarkably well for the majority of the under-resourced languages on which it was tested. Additionally, we discover interesting linguistic correlations learned by MURAL representations.

MURAL Architecture
The MURAL architecture is based on the structure of ALIGN, but employed in a multitask fashion. Whereas ALIGN uses a dual-encoder architecture to draw together representations of images and associated text descriptions, MURAL employs the dual-encoder structure for the same purpose while also extending it across languages by incorporating translation pairs. The dataset of image–text pairs is the same as that used for ALIGN, and the translation pairs are those used for LaBSE.

MURAL solves two contrastive learning tasks: 1) image–text matching and 2) text–text (bitext) matching, with both tasks sharing the text encoder module. The model learns associations between images and text from the image–text data, and learns the representations of hundreds of diverse languages from the translation pairs. The idea is that a shared encoder will transfer the image–text association learned from high-resource languages to under-resourced languages. We find that the best model employs an EfficientNet-B7 image encoder and a BERT-large text encoder, both trained from scratch. The learned representation can be used for downstream visual and vision-language tasks.

The architecture of MURAL depicts dual encoders with a shared text-encoder between the two tasks trained using a contrastive learning loss.

Multilingual Image-to-Text and Text-to-Image Retrieval
To demonstrate MURAL’s capabilities, we choose the task of cross-modal retrieval (i.e., retrieving relevant images given a text and vice versa) and report the scores on various academic image–text datasets covering well-resourced languages, such as MS-COCO (and its Japanese variant, STAIR), Flickr30K (in English) and Multi30K (extended to German, French, Czech), XTD (test-only set with seven well-resourced languages: Italian, Spanish, Russian, Chinese, Polish, Turkish, and Korean). In addition to well-resourced languages, we also evaluate MURAL on the recently published Wikipedia Image–Text (WIT) dataset, which covers 108 languages, with a broad range of both well-resourced (English, French, Chinese, etc.) and under-resourced (Swahili, Hindi, etc.) languages.

MURAL consistently outperforms prior state-of-the-art models, including M3P, UC2, and ALIGN, in both zero-shot and fine-tuned settings evaluated on well-resourced and under-resourced languages. We see remarkable performance gains for under-resourced languages when compared to the state-of-the-art model, ALIGN.

Mean recall on various multilingual image–text retrieval benchmarks. Mean recall is a common metric used to evaluate cross-modal retrieval performance on image–text datasets (higher is better). It measures the Recall@N (i.e., the chance that the ground truth image appears in the first N retrieved images) averaged over six measurements: Image→Text and Text→Image retrieval for N=[1, 5, 10]. Note that XTD scores report Recall@10 for Text→Image retrieval.

Retrieval Analysis
We also analyzed zero-shot retrieved examples on the WIT dataset comparing ALIGN and MURAL for English (en) and Hindi (hi). For under-resourced languages like Hindi, MURAL shows improved retrieval performance compared to ALIGN that reflects a better grasp of the text semantics.

Comparison of the top-5 images retrieved by ALIGN and by MURAL for the Text→Image retrieval task on the WIT dataset for the Hindi text, एक तश्तरी पर बिना मसाले या सब्ज़ी के रखी हुई सादी स्पगॅत्ती”, which translates to the English, “A bowl containing plain noodles without any spices or vegetables”.

Even for Image→Text retrieval in a well-resourced language, like French, MURAL shows better understanding for some words. For example, MURAL returns better results for the query “cadran solaire” (“sundial”, in French) than ALIGN, which doesn’t retrieve any text describing sundials (below).

Comparison of the top-5 text results from ALIGN and from MURAL on the Image→Text retrieval task for the same image of a sundial.

Embeddings Visualization
Previously, researchers have shown that visualizing model embeddings can reveal interesting connections among languages — for instance, representations learned by a neural machine translation (NMT) model have been shown to form clusters based on their membership to a language family. We perform a similar visualization for a subset of languages belonging to the Germanic, Romance, Slavic, Uralic, Finnic, Celtic, and Finno-Ugric language families (widely spoken in Europe and Western Asia). We compare MURAL’s text embeddings with LaBSE’s, which is a text-only encoder.

A plot of LabSE’s embeddings shows distinct clusters of languages influenced by language families. For instance, Romance languages (in purple, below) fall into a different region than Slavic languages (in brown, below). This finding is consistent with prior work that investigates intermediate representations learned by a NMT system.

Visualization of text representations of LaBSE for 35 languages. Languages are color coded based on their genealogical association. Representative languages include: Germanic (red) — German, English, Dutch; Uralic (orange) — Finnish, Estonian; Slavic (brown) — Polish, Russian; Romance (purple) — Italian, Portuguese, Spanish; Gaelic (blue) — Welsh, Irish.

In contrast to LaBSE’s visualization, MURAL’s embeddings, which are learned with a multimodal objective, shows some clusters that are in line with areal linguistics (where elements are shared by languages or dialects in a geographic area) and contact linguistics (where languages or dialects interact and influence each other). Notably, in the MURAL embedding space, Romanian (ro) is closer to the Slavic languages like Bulgarian (bg) and Macedonian (mk), which is in line with the Balkan sprachbund, than it is in LaBSE. Another possible language contact brings Finnic languages, Estonian (et) and Finnish (fi), closer to the Slavic languages cluster. The fact that MURAL pivots on images as well as translations appears to add an additional view on language relatedness as learned in deep representations, beyond the language family clustering observed in a text-only setting.

Visualization of text representations of MURAL for 35 languages. Color coding is the same as the figure above.

Final Remarks
Our findings show that training jointly using translation pairs helps overcome the scarcity of image–text pairs for many under-resourced languages and improves cross-modal performance. Additionally, it is interesting to observe hints of areal linguistics and contact linguistics in the text representations learned by using a multimodal model. This warrants more probing into different connections learned implicitly by multimodal models, such as MURAL. Finally, we hope this work promotes further research in the multimodal, multilingual space where models learn representations of and connections between languages (expressed via images and text), beyond well-resourced languages.

Acknowledgements
This research is in collaboration with Mandy Guo, Krishna Srinivasan, Ting Chen, Sneha Kudugunta, Chao Jia, and Jason Baldridge. We thank Zarana Parekh, Orhan Firat, Yuqing Chen, Apu Shah, Anosh Raj, Daphne Luong, and others who provided feedback for the project. We are also grateful for general support from Google Research teams.

Source: Google AI Blog


Do even more with your Chromebook camera

This summer, we shared an update about how we’re continuing to improve video calling on Chromebooks, thanks to performance improvements across Google Meet, Zoom and more. And the camera on your Chromebook is good for more than just video chatting. Hundreds of millions of images and videos have been captured using the Chromebook Camera app so far this year.

Today, we’re sharing a few features that make your Chromebook’s camera even more useful.

Scan documents and more

Have you ever wanted to use your Chromebook to share a physical document or image, but weren’t sure how without the help of a scanner? You can now use your Chromebook’s built-in camera to scan any document and turn it into a PDF or JPEG file. If your Chromebook comes with a front and back facing camera, you can use either of these to scan.

Open the Camera app and select “Scan” mode. When you hold out the document you want to scan in front of the camera, the edges will be automatically detected. Once it’s done, it’s easy to share through Gmail, to social media or to nearby Android phones or Chromebooks using Nearby Share.

Chromebook Camera app in “Scan” mode scanning a hard copy document.

You can now scan files using your Chromebook’s built-in camera.

Personalize your camera angle

If you use an external camera with your Chromebook, you can use the Pan-Tilt-Zoom feature to have more control over what your camera captures. You can now crop and angle your camera view exactly how you want it. Whether you want to show your furry friend napping in the background or just want to zoom in on yourself, your Chromebook’s got you covered.

With your external camera plugged in and configured, open the Camera app to adjust the angle you want to capture. Your selections will automatically save so when you jump from a Google Meet work call to making a video with your new puppy, your camera angle preferences will stay the same.

Man sitting on the floor uses the Pan-Tilt-Zoom feature open on the left hand side of the screen to adjust the camera angle.

With Pan-Tilt-Zoom you can adjust your camera angle to capture only what you want.

Try other Camera app features

In addition to taking pictures or scanning documents with your Chromebook’s camera, here are a few other features to test out:

  • Video mode. If you want to send a quick message to a loved one for their birthday, record a video by clicking on the “Video” mode.
  • Self timer. You don’t need to be within arm’s length of your laptop to take a picture. Set the timer, and you can take a few steps back to get the perfect shot.
  • QR Code. In addition to new document scanning, you can also use the “Scan” option to scan QR codes. It works just like document scanning, so use your front or back facing camera to scan a QR code.
  • Save for later. All your pictures and videos will automatically save to the “Camera” folder in your Files app for easy access later.

And coming soon…

Starting early next year, you’ll be able to create GIFs on the Camera app. Just record a five-second video dancing around with friends, hugging your loved ones, or playing with your favorite pet, and it will automatically turn into a shareable GIF.

If you’re interested in getting a sneak peak and providing feedback on Chromebook features before they launch, join our Chrome OS Beta Community. Sign-up here to be a Chrome OS Beta Tester Product Expert. Currently in Beta is a feature that integrates the Camera app with the Google Assistant. Just say “take a photo,” “record video” or “take a selfie” – you can even use Google Assistant to open the Camera app, so you don’t have to lift a finger.

We’ll be back in the new year to share more new Chromebook features.

Dev Channel Update for Desktop

The Dev channel has been updated to 98.0.4736.0 for Linux, Windows and Mac coming soon

A partial list of changes is available in the log. Interested in switching release channels? Find out how. If you find a new issue, please let us know by filing a bug. The community help forum is also a great place to reach out for help or learn about common issues.


Prudhvikumar Bommana

Google Chrome

More accessible web images arrive in 10 new languages

Images can be an integral part of many people’s online experiences. We rely on them to help bring news stories to life, see what our family and friends are up to, or help us decide which couch to buy. However, for 338 million people who are blind or have moderate to severe vision impairment, knowing what's in a web image that isn’t properly labeled can be a challenge. Screen reader technology relies on the efforts of content creators and developers who manually label images in order to make them accessible through spoken feedback or braille. Yet, billions of web images remain unlabelled, rendering them inaccessible for these users.

To help close this gap, the Chrome Accessibility and Google Research teams collaborated on developing a feature that automatically describes unlabelled images using AI. This feature was first released in 2019 supporting English only and was subsequently extended to five new languages in 2020 – French, German, Hindi, Italian and Spanish.

Today, we are expanding this feature to support ten additional languages: Croatian, Czech, Dutch, Finnish, Indonesian, Norwegian, Portuguese, Russian, Swedish and Turkish.

The major innovation behind this launch is the development of a single machine learning model that generates descriptions in each of the supported languages. This enables a more equitable user experience across languages in the sense that the generated image descriptions in any two languages can often be regarded as translations that respect the image details (Thapliyal and Soricut (2020)).

group of friends jumping on the beach

Auto-generated image descriptions can be incredibly helpful and their quality has come a long way, but it’s important to note they still can’t caption all images as well as a human. Our system was built to describe natural images and is unlikely to generate a description for other types of images, such as sketches, cartoons, memes or screenshots. We considered fairness, safety and quality when developing this feature and implemented a process to evaluate the images and captions along these dimensions before they're eligible to be shown to users.

We are excited to take this next step towards improving accessibility for more people around the world and look forward to expanding support to more languages in the future.

To activate this feature, you first need to turn on your screen reader (here's how to do that in Chrome). From there, you can activate the “Get image descriptions from Google” feature either by opening the context menu when browsing a web page or under your browser’s Accessibility settings. Chrome will then automatically generate descriptions for unlabelled web images in your preferred language.