Google Fiber is expanding to another Colorado city — Wheat Ridge.

Google Fiber is getting deeply rooted in Wheat Ridge, Colorado!

Yesterday, the Wheat Ridge City Council unanimously passed an agreement which will allow GFiber to deliver high speed internet to residents and businesses in the community.

Wheat Ridge is our third fiber-to-the-home city in the Rocky Mountain state, joining Lakewood and Westminster, which were announced within the past year. We’ve been active in the Denver area since 2017, providing gigabit speeds with Google Fiber Webpass to apartments and condominiums in the city.

Thumbnail

Wheat Ridge has long been known for growing beautiful things — in fact, it’s known as Carnation City, because of the many carnation farms that were once located here. GFiber is looking forward to growing something special for residents here as well, connecting them to high quality internet.

We expect to start construction in 2024 and begin serving customers later next year. Anyone looking for more information on service availability or the construction process can sign up for updates. Additionally, we are hiring in the Denver area — if you are interested in joining our team, please check out our website.  

Posted by Sasha Petrovic, General Manager - Southwest Region



Expanding grading systems in Google Classroom

What’s changing

Currently, teachers can set up grading systems in Google Classroom using points, grade categories, or grading periods. These options are numerical-based, meaning they correlate to points or percentages. 

In order to support other forms of grading, such as letter grades or proficiency ratings (ex. unsatisfactory to excellent), we’re introducing customizable grading scales so that a school’s or teacher’s grading preferences are better reflected in Google Classroom. With this update, teachers will have the option to select from numerical, letter grade, proficiency, four-point grading scales, or set up their own scale. 

Once the grading scale is selected and personalized (if the teacher chooses to do so), teachers can create assignments right away and any existing assignments will automatically have their newly selected grading scale. When grading assignments, teachers can enter either the grade with the points value or the level from the grading scale and Classroom will automatically display both to teachers and to students once a grade is returned.
Expanding grading systems in Google Classroom

Who’s impacted 

End users 


Why you’d use it 

This feature gives teachers more flexibility in deciding which grading scales to use for assignments and classwork. 


Getting started 

  • Admins: There is no admin control for this feature. 
  • End users: 
    • To configure grading scales including accessing grading scale templates, go to Class settings > scroll down to the Grading Scales settings > click Add. From there you can either choose a pre-set grading scale or create your own scale. 
      • Teachers also have the option to copy other classes once they have grading scales set up. 
      • Any new classes created will have the grading scale from the last edited class. 
    • Visit the Help Center to learn more about grading & returning assignments. 

Rollout pace 

Availability 

  • Available to Education Plus and the Teaching and Learning Upgrade 

Resources 

RO-ViT: Region-aware pre-training for open-vocabulary object detection with vision transformers

The ability to detect objects in the visual world is crucial for computer vision and machine intelligence, enabling applications like adaptive autonomous agents and versatile shopping systems. However, modern object detectors are limited by the manual annotations of their training data, resulting in a vocabulary size significantly smaller than the vast array of objects encountered in reality. To overcome this, the open-vocabulary detection task (OVD) has emerged, utilizing image-text pairs for training and incorporating new category names at test time by associating them with the image content. By treating categories as text embeddings, open-vocabulary detectors can predict a wide range of unseen objects. Various techniques such as image-text pre-training, knowledge distillation, pseudo labeling, and frozen models, often employing convolutional neural network (CNN) backbones, have been proposed. With the growing popularity of vision transformers (ViTs), it is important to explore their potential for building proficient open-vocabulary detectors.

The existing approaches assume the availability of pre-trained vision-language models (VLMs) and focus on fine-tuning or distillation from these models to address the disparity between image-level pre-training and object-level fine-tuning. However, as VLMs are primarily designed for image-level tasks like classification and retrieval, they do not fully leverage the concept of objects or regions during the pre-training phase. Thus, it could be beneficial for open-vocabulary detection if we build locality information into the image-text pre-training.

In “RO-ViT: Region-Aware Pretraining for Open-Vocabulary Object Detection with Vision Transformers”, presented at CVPR 2023, we introduce a simple method to pre-train vision transformers in a region-aware manner to improve open-vocabulary detection. In vision transformers, positional embeddings are added to image patches to encode information about the spatial position of each patch within the image. Standard pre-training typically uses full-image positional embeddings, which does not generalize well to detection tasks. Thus, we propose a new positional embedding scheme, called “cropped positional embedding”, that better aligns with the use of region crops in detection fine-tuning. In addition, we replace the softmax cross entropy loss with focal loss in contrastive image-text learning, allowing us to learn from more challenging and informative examples. Finally, we leverage recent advances in novel object proposals to enhance open-vocabulary detection fine-tuning, which is motivated by the observation that existing methods often miss novel objects during the proposal stage due to overfitting to foreground categories. We are also releasing the code here.


Region-aware image-text pre-training

Existing VLMs are trained to match an image as a whole to a text description. However, we observe there is a mismatch between the way the positional embeddings are used in the existing contrastive pre-training approaches and open-vocabulary detection. The positional embeddings are important to transformers as they provide the information of where each element in the set comes from. This information is often useful for downstream recognition and localization tasks. Pre-training approaches typically apply full-image positional embeddings during training, and use the same positional embeddings for downstream tasks, e.g., zero-shot recognition. However, the recognition occurs at region-level for open-vocabulary detection fine-tuning, which requires the full-image positional embeddings to generalize to regions that they never see during the pre-training.

To address this, we propose cropped positional embeddings (CPE). With CPE, we upsample positional embeddings from the image size typical for pre-training, e.g., 224x224 pixels, to that typical for detection tasks, e.g., 1024x1024 pixels. Then we randomly crop and resize a region, and use it as the image-level positional embeddings during pre-training. The position, scale, and aspect ratio of the crop is randomly sampled. Intuitively, this causes the model to view an image not as a full image in itself, but as a region crop from some larger unknown image. This better matches the downstream use case of detection where recognition occurs at region- rather than image-level.

For the pre-training, we propose cropped positional embedding (CPE) which randomly crops and resizes a region of positional embeddings instead of using the whole-image positional embedding (PE). In addition, we use focal loss instead of the common softmax cross entropy loss for contrastive learning.

We also find it beneficial to learn from hard examples with a focal loss. Focal loss enables finer control over how hard examples are weighted than what the softmax cross entropy loss can provide. We adopt the focal loss and replace it with the softmax cross entropy loss in both image-to-text and text-to-image losses. Both CPE and focal loss introduce no extra parameters and minimal computation costs.


Open-vocabulary detector fine-tuning

An open-vocabulary detector is trained with the detection labels of ‘base’ categories, but needs to detect the union of ‘base’ and ‘novel’ (unlabeled) categories at test time. Despite the backbone features pre-trained from the vast open-vocabulary data, the added detector layers (neck and heads) are newly trained with the downstream detection dataset. Existing approaches often miss novel/unlabeled objects in the object proposal stage because the proposals tend to classify them as background. To remedy this, we leverage recent advances in a novel object proposal method and adopt the localization quality-based objectness (i.e., centerness score) instead of object-or-not binary classification score, which is combined with the detection score. During training, we compute the detection scores for each detected region as the cosine similarity between the region’s embedding (computed via RoI-Align operation) and the text embeddings of the base categories. At test time, we append the text embeddings of novel categories, and the detection score is now computed with the union of the base and novel categories.

The pre-trained ViT backbone is transferred to the downstream open-vocabulary detection by replacing the global average pooling with detector heads. The RoI-Align embeddings are matched with the cached category embeddings to obtain the VLM score, which is combined with the detection score into the open-vocabulary detection score.

Results

We evaluate RO-ViT on the LVIS open-vocabulary detection benchmark. At the system-level, our best model achieves 33.6 box average precision on rare categories (APr) and 32.1 mask APr, which outperforms the best existing ViT-based approach OWL-ViT by 8.0 APr and the best CNN-based approach ViLD-Ens by 5.8 mask APr. It also exceeds the performance of many other approaches based on knowledge distillation, pre-training, or joint training with weak supervision.

RO-ViT outperforms both the state-of-the-art (SOTA) ViT-based and CNN-based methods on LVIS open-vocabulary detection benchmark. We show mask AP on rare categories (APr) , except for SOTA ViT-based (OwL-ViT) where we show box AP.

Apart from evaluating region-level representation through open-vocabulary detection, we evaluate the image-level representation of RO-ViT in image-text retrieval through the MS-COCO and Flickr30K benchmarks. Our model with 303M ViT outperforms the state-of-the-art CoCa model with 1B ViT on MS COCO, and is on par on Flickr30K. This shows that our pre-training method not only improves the region-level representation but also the global image-level representation for retrieval.

We show zero-shot image-text retrieval on MS COCO and Flickr30K benchmarks, and compare with dual-encoder methods. We report recall@1 (top-1 recall) on image-to-text (I2T) and text-to-image (T2I) retrieval tasks. RO-ViT outperforms the state-of-the-art CoCa with the same backbone.
RO-ViT open-vocabulary detection on LVIS. We only show the novel categories for clarity. RO-ViT detects many novel categories that it has never seen during detection training: “fishbowl”, “sombrero”, “persimmon”, “gargoyle”.

Visualization of positional embeddings

We visualize and compare the learned positional embeddings of RO-ViT with the baseline. Each tile is the cosine similarity between positional embeddings of one patch and all other patches. For example, the tile in the top-left corner (marked in red) visualizes the similarity between the positional embedding of the location (row=1, column=1) and those positional embeddings of all other locations in 2D. The brightness of the patch indicates how close the learned positional embeddings of different locations are. RO-ViT forms more distinct clusters at different patch locations showing symmetrical global patterns around the center patch.

Each tile shows the cosine similarity between the positional embedding of the patch (at the indicated row-column position) and the positional embeddings of all other patches. ViT-B/16 backbone is used.

Conclusion

We present RO-ViT, a contrastive image-text pre-training framework to bridge the gap between image-level pre-training and open-vocabulary detection fine-tuning. Our methods are simple, scalable, and easy to apply to any contrastive backbones with minimal computation overhead and no increase in parameters. RO-ViT achieves the state-of-the-art on LVIS open-vocabulary detection benchmark and on the image-text retrieval benchmarks, showing the learned representation is not only beneficial at region-level but also highly effective at the image-level. We hope this study can help the research on open-vocabulary detection from the perspective of image-text pre-training which can benefit both region-level and image-level tasks.


Acknowledgements

Dahun Kim, Anelia Angelova, and Weicheng Kuo conducted this work and are now at Google DeepMind. We would like to thank our colleagues at Google Research for their advice and helpful discussions.

Source: Google AI Blog


Google Ads API v12 sunset reminder

Google Ads API v12 will sunset on September 27, 2023. After this date, all v12 API requests will begin to fail. Please migrate to a newer version before September 27, 2023 to ensure your API access is unaffected.

We've prepared various resources to help you with the migration: In addition, using the Google Cloud Console, you can view the list of methods and services to which your project recently submitted requests:
  1. Open the Dashboard page (found under APIs & Services) in the Google Cloud Console.
  2. Click on Google Ads API in the table.
  3. On the METRICS subtab, you should see your recent requests plotted on each graph. At the bottom of the page, you’ll see the Methods table, where you can see which methods you’ve sent requests to. The method name includes a Google Ads API version, a service, and a method name, e.g., google.ads.googleads.v12.services.GoogleAdsService.Mutate. In this way, you can see all versions that you’ve used recently.
  4. (Optional) Click on the time frame at the top right of the page if you need to change it.
If you have questions while you’re upgrading, please reach out to us on the forum or through [email protected].

Latinitas conecta a la próxima generación de mujeres líderes a Internet de alta velocidad en el centro de Texas con el apoyo de Google Fiber

GFiber trabaja con organizaciones de todo el país para conectar a más personas con los beneficios de Internet de calidad. En Austin, Latinitas ayuda a las mujeres jóvenes a desarrollar las habilidades necesarias para vivir en un mundo digital. La directora de comunicaciones, Salwa Yordi, comparte su historia.


Nuestra misión en Latinitas es empoderar a todas las niñas para que innoven a través de los medios de comunicación y la tecnología. Brindamos programación en persona y virtual para que las estudiantes se expresen, desarrollen habilidades tecnológicas, aprendan sobre su cultura y descubran su voz única. Cada verano la Latinitas organiza “Club Latinitas” para niños de ocho a 18 años para facilitar clases sobre creatividad digital y codificación. Sin duda, se necesita Internet confiable de alta velocidad para apoyar adecuadamente a estas niñas y sus proyectos innovadores.


Thumbnail

Como Conexión Comunitaria de Google Fiber, Latinitas ha confiado en nuestro servicio de Google Fiber para brindar actividades enriquecedoras y recursos educativos que se han integrado perfectamente en las actividades de nuestro campamento, incluyendo talleres de transmisión en vivo con oradores invitados de alrededor de todo el país.


Según la Liga Nacional de Ciudades, solo el 65% de las personas Hispanas que viven en los Estados Unidos tienen acceso a una forma de conexión de banda ancha en comparación con las personas Blancas (80%) y Negras (71%). En Texas, más del 15% de los hogares carecen de banda ancha en cinco condados del centro de Texas (Bastrop, Fayette, Lee, Caldwell y Mason), según estimaciones de la Oficina del Censo de EE. UU. de 2017 a 2021. Además de la falta de acceso, también hay una falta de habilidades digitales dentro de la comunidad Latina. Para abordar esta necesidad, Google Fiber ha sido un socio constante en la financiación de las clases digitales para padres de Latinitas durante los últimos años.


Por primera vez, cada campista en Latinitas tiene la capacidad de interactuar con profesionales de la industria en sesiones virtuales en tiempo real, participar en talleres dinámicos y colaborar en proyectos multimedia sin interrupciones debido a la conectividad. Además, la confiabilidad de Internet también ha facilitado la investigación en línea, permitiendo a las niñas a profundizarse en áreas de interés y ultimamente desarrollar su confianza.



Latinitas se compromete a fomentar una generación de líderes expertos en tecnología. Este es un compromiso a largo plazo. Conectar a nuestra comunidad garantiza que la misión y la visión de nuestra organización permanezcan accesibles para todos, y estamos agradecidas por socios como Google Fiber que están ayudando a nuestras futuras generaciones a prosperar.


Publicado por Salwa Yordi, Directora de Comunicaciones de Latinitas



Latinitas connects the next generation of women leaders to high speed internet in Central Texas with support from Google Fiber

GFiber works with organizations across the country to connect more people to the benefits of quality internet. In Austin, Latinitas helps young women develop the necessary skills for living in a digital world. Communications Director Salwa Yordi shares their story.


Our mission at Latinitas is to empower all girls to innovate through media and technology. We provide in-person and virtual programming for students to express themselves, develop tech skills, learn about their culture, and discover their unique voice. Every summer the organization hosts “Club Latinitas” for kids ages eight to 18 to facilitate classes on digital creativity and coding. Without question, reliable high-speed internet is necessary in order to adequately support these children and their innovative projects.


Thumbnail

As a Google Fiber Community Connection, Latinitas has relied on our Google Fiber service to provide enriching activities and educational resources that have been seamlessly integrated into our camp activities, including live streaming workshops with guest speakers from around the country. 


According to the National League of Cities, only 65% of Hispanic individuals living in the United States have access to a form of broadband connection compared to white (80%) and Black (71%) individuals. In Texas, more than 15% of households lacked broadband in five Central Texas counties (Bastrop, Fayette, Lee, Caldwell and Mason), according to U.S. Census Bureau estimates from 2017 to 2021. In addition to the lack of access, there’s also a lack of digital skills within the Latino community. To address this need Google Fiber has been an ongoing partner in funding Latinitas Digital Parents classes for the past few years. 


For the first time, every camper at Latinitas has the ability to interact with industry professionals in real-time virtual sessions, participate in dynamic workshops and collaborate on multimedia projects without disruption due to connectivity. Additionally, the internet's reliability has also facilitated smooth online research, empowering kids to delve deeper into areas of interest and ultimately building up their confidence. 



Latinitas is committed to fostering a generation of tech-savvy leaders. This is a long-term commitment. Connecting our community ensures the mission and vision of our organization remains accessible to all, and we are thankful for partners like Google Fiber who are helping our future generations thrive.


Posted by Salwa Yordi, Communications Director at Latinitas