{"id":921,"date":"2023-12-22T19:10:27","date_gmt":"2023-12-22T19:10:27","guid":{"rendered":"https:\/\/todaysainews.com\/index.php\/2023\/12\/22\/2023-a-year-of-groundbreaking-advances-in-ai-and-computing\/"},"modified":"2025-04-27T07:30:41","modified_gmt":"2025-04-27T07:30:41","slug":"2023-a-year-of-groundbreaking-advances-in-ai-and-computing","status":"publish","type":"post","link":"https:\/\/todaysainews.com\/index.php\/2023\/12\/22\/2023-a-year-of-groundbreaking-advances-in-ai-and-computing\/","title":{"rendered":"2023: A Year of Groundbreaking Advances in AI and Computing"},"content":{"rendered":"<p> [ad_1]<br \/>\n<\/p>\n<div>\n<div class=\"article-cover\">\n<div class=\"article-cover__header\">\n<p class=\"article-cover__eyebrow glue-label\">Company<\/p>\n<dl class=\"article-cover__meta\">\n<dt class=\"glue-visually-hidden\">Published<\/dt>\n<dd class=\"article-cover__date glue-label\">\n              <time datetime=\"2023-12-22\"><br \/>\n                22 December 2023<br \/>\n              <\/time>\n            <\/dd>\n<dt class=\"glue-visually-hidden\">Authors<\/dt>\n<dd class=\"article-cover__authors\">\n<p data-block-key=\"srw3r\">By Jeff Dean, Chief Scientist, Google DeepMind &amp; Google Research, Demis Hassabis, CEO, Google DeepMind, and James Manyika, SVP, Google Research, Technology &amp; Society<\/p>\n<\/dd>\n<\/dl>\n<section class=\"glue-social glue-social--zippy share share--left article-cover__share\" data-glue-expansion-panel-expand-tooltip=\"Share: Expand to see social channels\" data-glue-expansion-panel-collapse-tooltip=\"Share: Hide social channels\" id=\"share-d818ae4f-93a7-43fe-a093-319fd713ff9a\">\n<\/section><\/div>\n<\/p><\/div>\n<div class=\"gdm-rich-text rich-text\">\n<p class=\"gdm-rich-text__subtitle\" data-block-key=\"uou0g\">This has been a year of incredible progress in the field of Artificial Intelligence (AI) research and its practical applications.<\/p>\n<p data-block-key=\"56ien\">As ongoing research pushes AI even farther, we look back to our <a href=\"https:\/\/ai.google\/static\/documents\/google-why-we-focus-on-ai.pdf\" rel=\"noopener\" target=\"_blank\">perspective<\/a> published in January of this year, titled \u201cWhy we focus on AI (and to what end),\u201d where we noted:<\/p>\n<p data-block-key=\"akp7k\"><i>We are committed to leading and setting the standard in developing and shipping useful and beneficial applications, applying ethical principles grounded in human values, and evolving our approaches as we learn from research, experience, users, and the wider community.<\/i><\/p>\n<p data-block-key=\"ftj8v\"><i>We also believe that getting AI right \u2014 which to us involves innovating and delivering widely accessible benefits to people and society, while mitigating its risks \u2014 must be a collective effort involving us and others, including researchers, developers, users (individuals, businesses, and other organizations), governments, regulators, and citizens.<\/i><\/p>\n<p data-block-key=\"flm6b\"><i>We are convinced that the AI-enabled innovations we are focused on developing and delivering boldly and responsibly are useful, compelling, and have the potential to assist and improve lives of people everywhere \u2014 this is what compels us.<\/i><\/p>\n<p data-block-key=\"dpj0b\">In this Year-in-Review post we\u2019ll go over some of Google Research\u2019s and Google DeepMind\u2019s efforts putting these paragraphs into practice safely throughout 2023.<\/p>\n<h2 data-block-key=\"2v7c7\">Advances in Products &amp; Technologies<\/h2>\n<p data-block-key=\"d0erc\">This was the year generative AI captured the world\u2019s attention, creating imagery, music, stories, and engaging conversation about everything imaginable, at a level of creativity and a speed almost implausible a few years ago.<\/p>\n<p data-block-key=\"eca9t\">In February, we <a href=\"https:\/\/blog.google\/technology\/ai\/bard-google-ai-search-updates\/\" rel=\"noopener\" target=\"_blank\">first launched<\/a> <a href=\"https:\/\/bard.google.com\/\" rel=\"noopener\" target=\"_blank\">Bard<\/a>, a tool that you can use to explore creative ideas and explain things simply. It can generate text, translate languages, write different kinds of creative content and more.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\">\n<\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">In May, we watched the results of months and years of our foundational and applied work announced on stage <a href=\"https:\/\/blog.research.google\/2023\/05\/google-research-at-io-2023.html\" rel=\"noopener\" target=\"_blank\">at Google I\/O<\/a>. Principally, this included <a href=\"https:\/\/ai.google\/discover\/palm2\/\" rel=\"noopener\" target=\"_blank\">PaLM 2<\/a>, a large language model (LLM) that brought together compute-optimal scaling, an improved dataset mixture, and model architecture to excel at advanced reasoning tasks.<\/p>\n<p data-block-key=\"avtn5\">By fine-tuning and instruction-tuning PaLM 2 for different purposes, we were able to integrate it into numerous Google products and features, including:<\/p>\n<ul>\n<li data-block-key=\"5jque\">An update to Bard, which enabled multilingual capabilities. Since its initial launch, Bard is now available in more than <a href=\"https:\/\/support.google.com\/bard\/answer\/13575153?hl=en\" rel=\"noopener\" target=\"_blank\">40 languages and over 230 countries and territories<\/a>, and <a href=\"https:\/\/blog.google\/products\/bard\/google-bard-new-features-update-sept-2023\/\" rel=\"noopener\" target=\"_blank\">with extensions<\/a>, Bard can find and show relevant information from Google tools used every day \u2014 like Gmail, Google Maps, YouTube, and more.<\/li>\n<li data-block-key=\"a9orq\"><a href=\"https:\/\/blog.google\/products\/search\/generative-ai-search\/\" rel=\"noopener\" target=\"_blank\">Search Generative Experience<\/a> (SGE), which uses LLMs to reimagine both how to organize information and how to help people navigate through it, creating a more fluid, conversational interaction model for our core Search product. This work extended the search engine experience from primarily focused on information retrieval into something much more \u2014 capable of retrieval, synthesis, creative generation and continuation of previous searches \u2014 while continuing to serve as a connection point between users and the web content they seek.<\/li>\n<li data-block-key=\"19alk\"><a href=\"https:\/\/google-research.github.io\/seanet\/musiclm\/examples\/\" rel=\"noopener\" target=\"_blank\">MusicLM<\/a>, a text-to-music model powered by <a href=\"https:\/\/ai.googleblog.com\/2022\/10\/audiolm-language-modeling-approach-to.html\" rel=\"noopener\" target=\"_blank\">AudioLM<\/a> and <a href=\"https:\/\/arxiv.org\/abs\/2208.12415\" rel=\"noopener\" target=\"_blank\">MuLAN<\/a>, which can make music from text, humming, images or video and musical accompaniments to singing.<\/li>\n<li data-block-key=\"ag6p4\">Duet AI, our AI-powered collaborator that provides users with assistance when they use Google Workspace and Google Cloud. <a href=\"https:\/\/workspace.google.com\/blog\/product-announcements\/duet-ai\" rel=\"noopener\" target=\"_blank\">Duet AI in Google Workspace<\/a>, for example, helps users write, create images, analyze spreadsheets, draft and summarize emails and chat messages, and summarize meetings. <a href=\"https:\/\/cloud.google.com\/blog\/products\/application-modernization\/introducing-duet-ai-for-google-cloud\" rel=\"noopener\" target=\"_blank\">Duet AI in Google Cloud<\/a> helps users code, deploy, scale, and monitor applications, as well as identify and accelerate resolution of cybersecurity threats.<\/li>\n<li data-block-key=\"a347i\">And many <a href=\"https:\/\/blog.google\/technology\/developers\/google-io-2023-100-announcements\/\" rel=\"noopener\" target=\"_blank\">other developments<\/a>.<\/li>\n<\/ul>\n<p data-block-key=\"aecp\">In June, following last year\u2019s release of our text-to-image generation model <a href=\"https:\/\/imagen.research.google\/\" rel=\"noopener\" target=\"_blank\">Imagen<\/a>, we released <a href=\"https:\/\/blog.research.google\/2023\/06\/imagen-editor-and-editbench-advancing.html\" rel=\"noopener\" target=\"_blank\">Imagen Editor<\/a>, which provides the ability to use region masks and natural language prompts to interactively edit generative images to provide much more precise control over the model output.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\">\n<\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">Later in the year, we released Imagen 2, which improved outputs via a specialized image aesthetics model based on human preferences for qualities such as good lighting, framing, exposure, and sharpness.<\/p>\n<p data-block-key=\"co7ql\">In October, we launched a feature that <a href=\"https:\/\/blog.research.google\/2023\/10\/google-search-can-now-help-with-english-speaking-practice.html\" rel=\"noopener\" target=\"_blank\">helps people practice speaking and improve their language skills<\/a>. The key technology that enabled this functionality was a novel deep learning model developed in collaboration with the Google Translate team, called Deep Aligner. This single new model has led to dramatic improvements in alignment quality across all tested language pairs, reducing average alignment error rate from 25% to 5% compared to alignment approaches based on <a href=\"https:\/\/aclanthology.org\/C96-2141\/\" rel=\"noopener\" target=\"_blank\">Hidden Markov models<\/a> (HMMs).<\/p>\n<p data-block-key=\"6bdnt\">In November, in partnership with <a href=\"https:\/\/blog.youtube\/inside-youtube\/ai-and-music-experiment\/\" rel=\"noopener\" target=\"_blank\">YouTube<\/a>, we announced <a href=\"https:\/\/deepmind.google\/discover\/blog\/transforming-the-future-of-music-creation\/\" rel=\"noopener\" target=\"_blank\">Lyria<\/a>, our most advanced AI music generation model to date. We released two experiments designed to open a new playground for creativity, DreamTrack and music AI tools, in concert with <a href=\"https:\/\/blog.youtube\/inside-youtube\/partnering-with-the-music-industry-on-ai\/\" rel=\"noopener\" target=\"_blank\">YouTube\u2019s Principles for partnering with the music industry on AI technology<\/a>.<\/p>\n<p data-block-key=\"fhggg\">Then in December, we launched <a href=\"https:\/\/blog.google\/technology\/ai\/google-gemini-ai\/\" rel=\"noopener\" target=\"_blank\">Gemini<\/a>, our most capable and general AI model. Gemini was built to be multimodal from the ground up across text, audio, image and videos.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\">\n<\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">Our initial family of Gemini models comes in three different sizes, Nano, Pro, and Ultra. Nano models are our smallest and most efficient models for powering on-device experiences in products like Pixel. The Pro model is highly-capable and best for scaling across a wide range of tasks. The Ultra model is our largest and most capable model for highly complex tasks.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\">\n<\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">In a <a href=\"https:\/\/storage.googleapis.com\/deepmind-media\/gemini\/gemini_1_report.pdf\" rel=\"noopener\" target=\"_blank\">technical report<\/a> about <a href=\"https:\/\/deepmind.google\/technologies\/gemini\" rel=\"noopener\" target=\"_blank\">Gemini models<\/a>, we showed that Gemini Ultra\u2019s performance exceeds current state-of-the-art results on 30 of the 32 widely-used academic benchmarks used in LLM research and development. With a score of 90.04%, Gemini Ultra was the first model to outperform human experts on <a href=\"https:\/\/arxiv.org\/abs\/2009.03300\" rel=\"noopener\" target=\"_blank\">MMLU<\/a>, and achieved a state-of-the-art score of 59.4% on the new <a href=\"https:\/\/arxiv.org\/abs\/2009.03300\" rel=\"noopener\" target=\"_blank\">MMMU<\/a> benchmark.<\/p>\n<p data-block-key=\"7mh7g\">Building on <a href=\"https:\/\/deepmind.google\/discover\/blog\/competitive-programming-with-alphacode\/\" rel=\"noopener\" target=\"_blank\">AlphaCode<\/a>, the first AI system to perform at the level of the median competitor in competitive programming, we <a href=\"https:\/\/storage.googleapis.com\/deepmind-media\/AlphaCode2\/AlphaCode2_Tech_Report.pdf\" rel=\"noopener\" target=\"_blank\">introduced AlphaCode 2<\/a> powered by a specialized version of Gemini. When evaluated on the same platform as the original AlphaCode, we found that AlphaCode 2 solved 1.7x more problems, and performed better than 85% of competition participants<\/p>\n<p data-block-key=\"scqs\">At the same time, <a href=\"https:\/\/blog.google\/products\/bard\/google-bard-try-gemini-ai\/\" rel=\"noopener\" target=\"_blank\">Bard got its biggest upgrade<\/a> with its use of the Gemini Pro model, making it far more capable at things like understanding, summarizing, reasoning, coding, and planning. In six out of eight benchmarks, Gemini Pro outperformed GPT-3.5, including in MMLU, one of the key standards for measuring large AI models, and <a href=\"https:\/\/huggingface.co\/datasets\/gsm8k\" rel=\"noopener\" target=\"_blank\">GSM8K<\/a>, which measures grade school math reasoning. Gemini Ultra will come to Bard early next year through Bard Advanced, a new cutting-edge AI experience.<\/p>\n<p data-block-key=\"3vn9i\">Gemini Pro is also available on <a href=\"https:\/\/cloud.google.com\/blog\/products\/ai-machine-learning\/gemini-support-on-vertex-ai\" rel=\"noopener\" target=\"_blank\">Vertex AI<\/a>, Google Cloud\u2019s end-to-end AI platform that empowers developers to build applications that can process information across text, code, images, and video. <a href=\"https:\/\/blog.google\/technology\/ai\/gemini-api-developers-cloud\/\" rel=\"noopener\" target=\"_blank\">Gemini Pro was also made available in AI Studio<\/a> in December.<\/p>\n<p data-block-key=\"4r2e8\">To best illustrate some of Gemini\u2019s capabilities, we produced a <a href=\"https:\/\/deepmind.google\/technologies\/gemini\/#hands-on\" rel=\"noopener\" target=\"_blank\">series of short videos<\/a> with explanations of how Gemini could:<\/p>\n<h2 data-block-key=\"3i4fd\">ML\/AI Research<\/h2>\n<p data-block-key=\"vaa5\">In addition to our advances in products and technologies, we\u2019ve also made a number of important advancements in the broader fields of machine learning and AI research.<\/p>\n<p data-block-key=\"4hb8e\">At the heart of the most advanced ML models is the Transformer model architecture, <a href=\"https:\/\/blog.research.google\/2017\/08\/transformer-novel-neural-network.html\" rel=\"noopener\" target=\"_blank\">developed by Google researchers in 2017<\/a>. Originally developed for language, it has proven useful in domains as varied as <a href=\"https:\/\/blog.research.google\/2020\/12\/transformers-for-image-recognition-at.html\" rel=\"noopener\" target=\"_blank\">computer vision<\/a>, <a href=\"https:\/\/deepmind.google\/discover\/blog\/transforming-the-future-of-music-creation\/\" rel=\"noopener\" target=\"_blank\">audio<\/a>, <a href=\"https:\/\/deepmind.google\/discover\/blog\/a-catalogue-of-genetic-mutations-to-help-pinpoint-the-cause-of-diseases\/\" rel=\"noopener\" target=\"_blank\">genomics<\/a>, <a href=\"https:\/\/deepmind.google\/technologies\/alphafold\/\" rel=\"noopener\" target=\"_blank\">protein folding<\/a>, and more. This year, our work on <a href=\"https:\/\/blog.research.google\/2023\/03\/scaling-vision-transformers-to-22.html\" rel=\"noopener\" target=\"_blank\">scaling vision transformers<\/a> demonstrated state-of-the-art results across a wide variety of vision tasks, and has also been useful in building <a href=\"https:\/\/blog.research.google\/2023\/03\/palm-e-embodied-multimodal-language.html\" rel=\"noopener\" target=\"_blank\">more capable robots<\/a>.<\/p>\n<p data-block-key=\"7hpsj\">Expanding the versatility of models requires the ability to perform higher-level and multi-step reasoning. This year, we approached this target following several research tracks. For example, <a href=\"https:\/\/blog.research.google\/2023\/08\/teaching-language-models-to-reason.html\" rel=\"noopener\" target=\"_blank\">algorithmic prompting<\/a> is a new method that teaches language models reasoning by demonstrating a sequence of algorithmic steps, which the model can then apply in new contexts. This approach improves accuracy on one middle-school mathematics benchmark from 25.9% to 61.1%.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\"><figcaption class=\"single-media__caption\">\n<p data-block-key=\"g18z5\">By providing algorithmic prompts, we can teach a model the rules of arithmetic via in-context learning.<\/p>\n<\/figcaption><\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">In the domain of visual question answering, in a collaboration with UC Berkeley researchers, we showed how we could <a href=\"https:\/\/blog.research.google\/2023\/07\/modular-visual-question-answering-via.html\" rel=\"noopener\" target=\"_blank\">better answer complex visual questions<\/a> (\u201cIs the carriage to the right of the horse?\u201d) by combining a visual model with a language model trained to answer visual questions by synthesizing a program to perform multi-step reasoning.<\/p>\n<p data-block-key=\"be1e9\">We are now using a <a href=\"https:\/\/blog.research.google\/2023\/05\/large-sequence-models-for-software.html\" rel=\"noopener\" target=\"_blank\">general model that understands many aspects of the software development life cycle<\/a> to automatically generate code review comments, respond to code review comments, make performance-improving suggestions for pieces of code (by learning from past such changes in other contexts), fix code in response to compilation errors, and more.<\/p>\n<p data-block-key=\"28ojq\">In a multi-year research collaboration with the Google Maps team, we were able to scale inverse reinforcement learning and apply it to the <a href=\"https:\/\/blog.research.google\/2023\/09\/world-scale-inverse-reinforcement.html\" rel=\"noopener\" target=\"_blank\">world-scale problem of improving route suggestions<\/a> for over 1 billion users. Our work culminated in a 16\u201324% relative improvement in global route match rate, helping to ensure that routes are better aligned with user preferences.<\/p>\n<p data-block-key=\"dd0ss\">We also continue to work on techniques to improve the inference performance of machine learning models. In work on <a href=\"https:\/\/blog.research.google\/2023\/08\/neural-network-pruning-with.html\" rel=\"noopener\" target=\"_blank\">computationally-friendly approaches to pruning connections in neural networks<\/a>, we were able to devise an approximation algorithm to the computationally intractable best-subset selection problem that is able to prune 70% of the edges from an image classification model and still retain almost all of the accuracy of the original.<\/p>\n<p data-block-key=\"bt0lq\">In work on <a href=\"https:\/\/blog.research.google\/2023\/06\/speed-is-all-you-need-on-device.html\" rel=\"noopener\" target=\"_blank\">accelerating on-device diffusion models<\/a>, we were also able to apply a variety of optimizations to attention mechanisms, convolutional kernels, and fusion of operations to make it practical to run high quality image generation models on-device; for example, enabling \u201ca photorealistic and high-resolution image of a cute puppy with surrounding flowers\u201d to be generated in just 12 seconds on a smartphone.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\">\n<\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">Advances in capable language and multimodal models have also benefited our robotics research efforts. We combined separately trained language, vision, and robotic control models into <a href=\"https:\/\/blog.research.google\/2023\/03\/palm-e-embodied-multimodal-language.html\" rel=\"noopener\" target=\"_blank\">PaLM-E<\/a>, an embodied multi-modal model for robotics, and <a href=\"https:\/\/deepmind.google\/discover\/blog\/rt-2-new-model-translates-vision-and-language-into-action\/\" rel=\"noopener\" target=\"_blank\">Robotic Transformer 2<\/a> (RT-2), a novel vision-language-action (VLA) model that <a href=\"https:\/\/deepmind.google\/discover\/blog\/robocat-a-self-improving-robotic-agent\/\" rel=\"noopener\" target=\"_blank\">learns<\/a> from both web and robotics data, and translates this knowledge into generalized instructions for robotic control.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\"><figcaption class=\"single-media__caption\">\n<p data-block-key=\"g18z5\">RT-2 architecture and training: We co-fine-tune a pre-trained vision-language model on robotics and web data. The resulting model takes in robot camera images and directly predicts actions for a robot to perform.<\/p>\n<\/figcaption><\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">Furthermore, we showed how <a href=\"https:\/\/blog.research.google\/2023\/08\/saytap-language-to-quadrupedal.html\" rel=\"noopener\" target=\"_blank\">language can also be used to control the gait of quadrupedal robots<\/a> and explored the <a href=\"https:\/\/blog.research.google\/2023\/08\/language-to-rewards-for-robotic-skill.html\" rel=\"noopener\" target=\"_blank\">use of language to help formulate more explicit reward functions<\/a> to bridge the gap between human language and robotic actions. Then, in <a href=\"https:\/\/blog.research.google\/2023\/05\/barkour-benchmarking-animal-level.html\" rel=\"noopener\" target=\"_blank\">Barkour<\/a> we benchmarked the agility limits of quadrupedal robots.<\/p>\n<h2 data-block-key=\"3l1ue\">Algorithms &amp; Optimization<\/h2>\n<p data-block-key=\"c64v7\">Designing efficient, robust, and scalable algorithms remains a high priority. This year, our work included: applied and scalable algorithms, market algorithms, system efficiency and optimization, and privacy.<\/p>\n<p data-block-key=\"5pjud\">We introduced <a href=\"https:\/\/deepmind.google\/discover\/blog\/alphadev-discovers-faster-sorting-algorithms\/\" rel=\"noopener\" target=\"_blank\">AlphaDev<\/a>, an AI system that uses reinforcement learning to discover enhanced computer science algorithms. AlphaDev uncovered a faster algorithm for sorting, a method for ordering data, which led to improvements in the LLVM libc++ sorting library that were up to 70% faster for shorter sequences and about 1.7% faster for sequences exceeding 250,000 elements.<\/p>\n<p data-block-key=\"3rv6j\">We developed a novel model to <a href=\"https:\/\/arxiv.org\/abs\/2305.12322\" rel=\"noopener\" target=\"_blank\">predict the properties of large graphs<\/a>, enabling estimation of performance for large programs. We released a new dataset, <a href=\"https:\/\/arxiv.org\/abs\/2308.13490\" rel=\"noopener\" target=\"_blank\">TPUGraphs<\/a>, to accelerate <a href=\"https:\/\/www.kaggle.com\/competitions\/predict-ai-model-runtime\" rel=\"noopener\" target=\"_blank\">open research in this area<\/a>, and showed how we can use <a href=\"https:\/\/blog.research.google\/2023\/12\/advancements-in-machine-learning-for.html\" rel=\"noopener\" target=\"_blank\">modern ML to improve ML efficiency<\/a>.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\"><figcaption class=\"single-media__caption\">\n<p data-block-key=\"g18z5\">The TPUGraphs dataset has 44 million graphs for ML program optimization.<\/p>\n<\/figcaption><\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">We developed a new <a href=\"https:\/\/en.wikipedia.org\/wiki\/Load_balancing_(computing)\" rel=\"noopener\" target=\"_blank\">load balancing<\/a> algorithm for distributing queries to a server, called <a href=\"https:\/\/arxiv.org\/abs\/2312.10172\" rel=\"noopener\" target=\"_blank\">Prequal<\/a>, which minimizes a combination of requests-in-flight and estimates the latency. Deployments across several systems have saved CPU, latency, and RAM significantly. We also designed a new <a href=\"https:\/\/arxiv.org\/abs\/2305.02508\" rel=\"noopener\" target=\"_blank\">analysis framework<\/a> for the classical caching problem with capacity reservations.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\"><figcaption class=\"single-media__caption\">\n<p data-block-key=\"g18z5\">Heatmaps of normalized CPU usage transitioning to Prequal at 08:00.<\/p>\n<\/figcaption><\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">We improved state-of-the-art in clustering and <a href=\"https:\/\/en.wikipedia.org\/wiki\/Graph_neural_network\" rel=\"noopener\" target=\"_blank\">graph algorithms<\/a> by developing new techniques for<a href=\"https:\/\/arxiv.org\/abs\/2106.05513\" rel=\"noopener\" target=\"_blank\"> computing minimum-cut<\/a>, <a href=\"https:\/\/arxiv.org\/abs\/2309.17243\" rel=\"noopener\" target=\"_blank\">approximating correlation clustering<\/a>, and <a href=\"https:\/\/arxiv.org\/abs\/2308.00503\" rel=\"noopener\" target=\"_blank\">massively parallel graph clustering<\/a>. Additionally, we introduced<a href=\"https:\/\/arxiv.org\/abs\/2308.03578\" rel=\"noopener\" target=\"_blank\"> TeraHAC<\/a>, a novel hierarchical clustering algorithm for trillion-edge graphs, designed a <a href=\"https:\/\/blog.research.google\/2023\/11\/best-of-both-worlds-achieving.html\" rel=\"noopener\" target=\"_blank\">text clustering algorithm<\/a> for better scalability while maintaining quality, and designed the most efficient <a href=\"https:\/\/arxiv.org\/abs\/2307.03043\" rel=\"noopener\" target=\"_blank\">algorithm for approximating the Chamfer Distance<\/a>, the standard similarity function for multi-embedding models, offering &gt;50\u00d7 speedups over highly-optimized exact algorithms and scaling to billions of points.<\/p>\n<p data-block-key=\"b7ch5\">We continued optimizing Google\u2019s large embedding models (LEMs), which power many of our core products and recommender systems. Some new techniques include <a href=\"https:\/\/arxiv.org\/abs\/2305.12102\" rel=\"noopener\" target=\"_blank\">Unified Embedding<\/a> for battle-tested feature representations in web-scale ML systems and <a href=\"https:\/\/arxiv.org\/abs\/2209.14881\" rel=\"noopener\" target=\"_blank\">Sequential Attention<\/a>, which uses attention mechanisms to discover high-quality sparse model architectures during training.<\/p>\n<p data-block-key=\"5vbd6\">Beyond auto-bidding systems, we also studied auction design in other complex settings, such as <a href=\"https:\/\/arxiv.org\/abs\/2204.01962\" rel=\"noopener\" target=\"_blank\">buy-many mechanisms<\/a>, <a href=\"https:\/\/arxiv.org\/abs\/2207.09429\" rel=\"noopener\" target=\"_blank\">auctions for heterogeneous bidders<\/a>, <a href=\"https:\/\/arxiv.org\/abs\/2309.10766\" rel=\"noopener\" target=\"_blank\">contract designs<\/a>, and innovated <a href=\"https:\/\/dl.acm.org\/doi\/10.5555\/3618408.3618478\" rel=\"noopener\" target=\"_blank\">robust online bidding algorithms<\/a>. Motivated by the application of generative AI in collaborative creation (e.g., joint ad for advertisers), we proposed <a href=\"https:\/\/arxiv.org\/abs\/2310.10826\" rel=\"noopener\" target=\"_blank\">a novel token auction model<\/a> where LLMs bid for influence in the collaborative AI creation. Finally, we show how to <a href=\"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3580507.3597702\" rel=\"noopener\" target=\"_blank\">mitigate personalization effects in experimental design<\/a>, which, for example, may cause recommendations to drift over time.<\/p>\n<p data-block-key=\"9g6md\">The Chrome Privacy Sandbox, a multi-year collaboration between Google Research and Chrome, has publicly launched several APIs, including for <a href=\"https:\/\/privacysandbox.com\/intl\/en_us\/learning-hub\/#protected-audience\" rel=\"noopener\" target=\"_blank\">Protected Audience<\/a>, <a href=\"https:\/\/privacysandbox.com\/intl\/en_us\/learning-hub\/#topics\" rel=\"noopener\" target=\"_blank\">Topics<\/a>, and <a href=\"https:\/\/privacysandbox.com\/intl\/en_us\/learning-hub\/#attribution-reporting\" rel=\"noopener\" target=\"_blank\">Attribution Reporting<\/a>. This is a major step in protecting user privacy while supporting the open and free web ecosystem. These efforts have been facilitated by fundamental research on <a href=\"https:\/\/arxiv.org\/abs\/2304.07210\" rel=\"noopener\" target=\"_blank\">re-identification risk<\/a>, <a href=\"https:\/\/arxiv.org\/abs\/2301.05605\" rel=\"noopener\" target=\"_blank\">private streaming computation<\/a>, <a href=\"https:\/\/blog.research.google\/2023\/12\/summary-report-optimization-in-privacy.html\" rel=\"noopener\" target=\"_blank\">optimization<\/a> of privacy caps and budgets, <a href=\"https:\/\/arxiv.org\/pdf\/2308.13510.pdf\" rel=\"noopener\" target=\"_blank\">hierarchical aggregation<\/a>, and training models with <a href=\"https:\/\/arxiv.org\/pdf\/2312.05659.pdf\" rel=\"noopener\" target=\"_blank\">label privacy<\/a>.<\/p>\n<h2 data-block-key=\"6q7ma\">Science and Society<\/h2>\n<p data-block-key=\"6gi3n\">In the not too distant future, there is a very real possibility that AI applied to scientific problems can accelerate the rate of discovery in certain domains by 10\u00d7 or 100\u00d7, or more, and lead to major advances in diverse areas including bioengineering, <a href=\"https:\/\/deepmind.google\/discover\/blog\/millions-of-new-materials-discovered-with-deep-learning\" rel=\"noopener\" target=\"_blank\">materials science<\/a>, <a href=\"https:\/\/deepmind.google\/discover\/blog\/graphcast-ai-model-for-faster-and-more-accurate-global-weather-forecasting\/\" rel=\"noopener\" target=\"_blank\">weather prediction<\/a>, <a href=\"https:\/\/blog.google\/outreach-initiatives\/sustainability\/google-ai-climate-change-solutions\/\" rel=\"noopener\" target=\"_blank\">climate forecasting<\/a>, <a href=\"https:\/\/blog.research.google\/2023\/09\/google-research-embarks-on-effort-to.html\" rel=\"noopener\" target=\"_blank\">neuroscience<\/a>, <a href=\"https:\/\/blog.research.google\/2023\/04\/an-ml-based-approach-to-better.html\" rel=\"noopener\" target=\"_blank\">genetic medicine<\/a>, and <a href=\"https:\/\/health.google\/health-research\/publications\/\" rel=\"noopener\" target=\"_blank\">healthcare<\/a>.<\/p>\n<h3 data-block-key=\"bhlt3\">Sustainability and Climate Change<\/h3>\n<p data-block-key=\"9mu4m\">In <a href=\"https:\/\/blog.google\/outreach-initiatives\/sustainability\/google-ai-reduce-greenhouse-emissions-project-greenlight\/\" rel=\"noopener\" target=\"_blank\">Project Green Light<\/a>, we partnered with 13 cities around the world to help improve traffic flow at intersections and reduce stop-and-go emissions. Early numbers from these partnerships indicate a potential for up to 30% reduction in stops and up to 10% reduction in emissions.<\/p>\n<p data-block-key=\"9uos7\">In our <a href=\"https:\/\/sites.research.google\/contrails\/\" rel=\"noopener\" target=\"_blank\">contrails work<\/a>, we analyzed large-scale weather data, historical satellite images, and past flights. We <a href=\"https:\/\/blog.google\/technology\/ai\/ai-airlines-contrails-climate-change\/\" rel=\"noopener\" target=\"_blank\">trained an AI model<\/a> to predict where contrails form and reroute airplanes accordingly. In partnership with American Airlines and Breakthrough Energy, we used this system to demonstrate contrail reduction by 54%.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\"><figcaption class=\"single-media__caption\">\n<p data-block-key=\"a4olz\">Contrails detected over the United States using AI and GOES-16 satellite imagery.<\/p>\n<\/figcaption><\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">We are also developing novel technology-driven approaches to <a href=\"https:\/\/blog.google\/outreach-initiatives\/sustainability\/google-ai-climate-change-solutions\/\" rel=\"noopener\" target=\"_blank\">help communities with the effects of climate change<\/a>. For example, we have <a href=\"https:\/\/blog.google\/outreach-initiatives\/sustainability\/flood-hub-ai-flood-forecasting-more-countries\/\" rel=\"noopener\" target=\"_blank\">expanded our flood forecasting coverage to 80 countries<\/a>, which directly impacts more than 460 million people. We have initiated a <a href=\"https:\/\/blog.research.google\/2023\/10\/looking-back-at-wildfire-research-in.html\" rel=\"noopener\" target=\"_blank\">number of research efforts<\/a> to help mitigate the increasing danger of wildfires, including <a href=\"https:\/\/blog.research.google\/2023\/02\/real-time-tracking-of-wildfire.html\" rel=\"noopener\" target=\"_blank\">real-time tracking of wildfire boundaries<\/a> using satellite imagery, and work that <a href=\"https:\/\/blog.research.google\/2023\/10\/improving-traffic-evacuations-case-study.html\" rel=\"noopener\" target=\"_blank\">improves emergency evacuation plans<\/a> for communities at risk to rapidly-spreading wildfires. Our <a href=\"https:\/\/www.americanforests.org\/article\/american-forests-unveils-updates-for-tree-equity-score-tool-to-address-climate-justice\/\" rel=\"noopener\" target=\"_blank\">partnership<\/a> with American Forests puts data from our <a href=\"https:\/\/insights.sustainability.google\/places\/ChIJVTPokywQkFQRmtVEaUZlJRA\/trees?hl=en-US\" rel=\"noopener\" target=\"_blank\">Tree Canopy<\/a> project to work in their <a href=\"https:\/\/treeequityscore.org\/\" rel=\"noopener\" target=\"_blank\">Tree Equity Score<\/a> platform, helping communities identify and address unequal access to trees.<\/p>\n<p data-block-key=\"39t7\">Finally, we continued to develop better models for weather prediction at longer time horizons. Improving on <a href=\"https:\/\/blog.research.google\/2020\/03\/a-neural-weather-model-for-eight-hour.html\" rel=\"noopener\" target=\"_blank\">MetNet<\/a> and <a href=\"https:\/\/blog.research.google\/2021\/11\/metnet-2-deep-learning-for-12-hour.html\" rel=\"noopener\" target=\"_blank\">MetNet-2<\/a>, in this year\u2019s work on <a href=\"https:\/\/blog.research.google\/2023\/11\/metnet-3-state-of-art-neural-weather.html\" rel=\"noopener\" target=\"_blank\">MetNet-3<\/a>, we now outperform traditional numerical weather simulations up to twenty-four hours. In the area of medium-term, global weather forecasting, our work on <a href=\"https:\/\/deepmind.google\/discover\/blog\/graphcast-ai-model-for-faster-and-more-accurate-global-weather-forecasting\/\" rel=\"noopener\" target=\"_blank\">GraphCast<\/a> showed significantly better prediction accuracy for up to 10 days compared to <a href=\"https:\/\/en.wikipedia.org\/wiki\/Integrated_Forecast_System\" rel=\"noopener\" target=\"_blank\">HRES<\/a>, the most accurate operational deterministic forecast, produced by the <a href=\"https:\/\/www.ecmwf.int\/\" rel=\"noopener\" target=\"_blank\">European Centre for Medium-Range Weather Forecasts<\/a> (ECMWF). In collaboration with ECMWF, we released <a href=\"https:\/\/blog.research.google\/2023\/08\/weatherbench-2-benchmark-for-next.html\" rel=\"noopener\" target=\"_blank\">WeatherBench-2<\/a>, a benchmark for evaluating the accuracy of weather forecasts in a common framework.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\"><figcaption class=\"single-media__caption\">\n<p data-block-key=\"85azw\">A selection of GraphCast\u2019s predictions rolling across 10 days showing specific humidity at 700 hectopascals (about 3 km above surface), surface temperature, and surface wind speed.<\/p>\n<\/figcaption><\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<h3 data-block-key=\"uou0g\">Health and the Life Sciences<\/h3>\n<p data-block-key=\"f7c09\">The potential of AI to dramatically improve processes in healthcare is significant. Our initial <a href=\"https:\/\/www.nature.com\/articles\/s41586-023-06291-2\" rel=\"noopener\" target=\"_blank\">Med-PaLM<\/a> model was the first model capable of achieving a passing score on the U.S. medical licensing exam. Our more recent <a href=\"https:\/\/blog.google\/technology\/health\/ai-llm-medpalm-research-thecheckup\/\" rel=\"noopener\" target=\"_blank\">Med-PaLM 2 model<\/a> improved by a further 19%, achieving an expert-level accuracy of 86.5%. These <a href=\"https:\/\/sites.research.google\/med-palm\/\" rel=\"noopener\" target=\"_blank\">Med-PaLM models<\/a> are language-based, enable clinicians to ask questions and have a dialogue about complex medical conditions, and are <a href=\"https:\/\/cloud.google.com\/blog\/topics\/healthcare-life-sciences\/introducing-medlm-for-the-healthcare-industry\" rel=\"noopener\" target=\"_blank\">available<\/a> to healthcare organizations as part of <a href=\"https:\/\/cloud.google.com\/vertex-ai\/docs\/generative-ai\/medlm\/overview\" rel=\"noopener\" target=\"_blank\">MedLM<\/a> through Google Cloud.<\/p>\n<p data-block-key=\"dqdjd\">In the same way our general language models are evolving to handle multiple modalities, we have recently shown research on a <a href=\"https:\/\/blog.research.google\/2023\/08\/multimodal-medical-ai.html\" rel=\"noopener\" target=\"_blank\">multimodal version of Med-PaLM<\/a> capable of interpreting medical images, textual data, and other modalities, <a href=\"https:\/\/arxiv.org\/abs\/2307.14334\" rel=\"noopener\" target=\"_blank\">describing a path<\/a> for how we can realize the exciting potential of AI models to help advance real-world clinical care.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\"><figcaption class=\"single-media__caption\">\n<p data-block-key=\"e98d9\">Med-PaLM M is a large multimodal generative model that flexibly encodes and interprets biomedical data including clinical language, imaging, and genomics with the same model weights.<\/p>\n<\/figcaption><\/figure>\n<figure class=\"single-media single-media--large\"><figcaption class=\"single-media__caption\">\n<p data-block-key=\"g18z5\">Med-PaLM M is a large multimodal generative model that flexibly encodes and interprets biomedical data including clinical language, imaging, and genomics with the same model weights.<\/p>\n<\/figcaption><\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">We have also been working on <a href=\"https:\/\/deepmind.google\/discover\/blog\/codoc-developing-reliable-ai-tools-for-healthcare\/\" rel=\"noopener\" target=\"_blank\">how best to harness AI models in clinical workflows<\/a>. We have shown that <a href=\"https:\/\/blog.research.google\/2023\/03\/learning-from-deep-learning-case-study.html\" rel=\"noopener\" target=\"_blank\">coupling deep learning with interpretability methods<\/a> can yield new insights for clinicians. We have also shown that self-supervised learning, with careful consideration of privacy, safety, fairness and ethics, <a href=\"https:\/\/blog.research.google\/2023\/04\/robust-and-efficient-medical-imaging.html\" rel=\"noopener\" target=\"_blank\">can reduce the amount of de-identified data needed<\/a> to train clinically relevant medical imaging models by 3\u00d7\u2013100\u00d7, reducing the barriers to adoption of models in real clinical settings. We also released an <a href=\"https:\/\/blog.research.google\/2023\/11\/enabling-large-scale-health-studies-for.html\" rel=\"noopener\" target=\"_blank\">open source mobile data collection platform<\/a> for people with chronic disease to provide tools to the community to build their own studies.<\/p>\n<p data-block-key=\"80ahk\">AI systems can also discover completely new signals and biomarkers in existing forms of medical data. In work on <a href=\"https:\/\/blog.research.google\/2023\/03\/detecting-novel-systemic-biomarkers-in.html\" rel=\"noopener\" target=\"_blank\">novel biomarkers discovered in retinal images<\/a>, we demonstrated that a number of systemic biomarkers spanning several organ systems (e.g., kidney, blood, liver) can be predicted from external eye photos. In other work, we showed that combining <a href=\"https:\/\/blog.research.google\/2023\/04\/developing-aging-clock-using-deep.html\" rel=\"noopener\" target=\"_blank\">retinal images and genomic information<\/a> helps identify some underlying factors of aging.<\/p>\n<p data-block-key=\"c8a35\">In the genomics space, we worked with 119 scientists across 60 institutions to create a <a href=\"https:\/\/blog.research.google\/2023\/05\/building-better-pangenomes-to-improve.html\" rel=\"noopener\" target=\"_blank\">new map of the human genome<\/a>, or pangenome. This more equitable pangenome better represents the genomic diversity of global populations. Building on our ground-breaking <a href=\"https:\/\/www.nature.com\/articles\/s41586-021-03819-2\" rel=\"noopener\" target=\"_blank\">AlphaFold<\/a> work, our work on <a href=\"https:\/\/deepmind.google\/discover\/blog\/a-catalogue-of-genetic-mutations-to-help-pinpoint-the-cause-of-diseases\/\" rel=\"noopener\" target=\"_blank\">AlphaMissense<\/a> this year provides a catalog of predictions for 89% of all 71 million possible <a href=\"https:\/\/en.wikipedia.org\/wiki\/Missense_mutation\" rel=\"noopener\" target=\"_blank\">missense variants<\/a> as either likely pathogenic or likely benign.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\"><figcaption class=\"single-media__caption\">\n<p data-block-key=\"g18z5\">Examples of AlphaMissense predictions overlaid on AlphaFold predicted structures (red \u2013 predicted as pathogenic; blue \u2013 predicted as benign; grey \u2013 uncertain). Red dots represent known pathogenic missense variants, blue dots represent known benign variants. <b>Left:<\/b> HBB protein. Variants in this protein can cause sickle cell anaemia. <b>Right:<\/b> CFTR protein. Variants in this protein can cause cystic fibrosis.<\/p>\n<\/figcaption><\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">We also shared <a href=\"https:\/\/deepmind.google\/discover\/blog\/a-glimpse-of-the-next-generation-of-alphafold\/\" rel=\"noopener\" target=\"_blank\">an update<\/a> on progress towards the next generation of AlphaFold. Our latest model can now generate predictions for nearly all molecules in the <a href=\"https:\/\/www.wwpdb.org\/\" rel=\"noopener\" target=\"_blank\">Protein Data Bank<\/a> (PDB), frequently reaching atomic accuracy. This unlocks new understanding and significantly improves accuracy in multiple key biomolecule classes, including ligands (small molecules), proteins, nucleic acids (DNA and RNA), and those containing post-translational modifications (PTMs).<\/p>\n<p data-block-key=\"8ef3u\">On the neuroscience front, we <a href=\"https:\/\/blog.research.google\/2023\/09\/google-research-embarks-on-effort-to.html\" rel=\"noopener\" target=\"_blank\">announced a new collaboration<\/a> with Harvard, Princeton, the NIH, and others to map an entire mouse brain at synaptic resolution, beginning with a first phase that will focus on the <a href=\"https:\/\/en.wikipedia.org\/wiki\/Hippocampal_formation\" rel=\"noopener\" target=\"_blank\">hippocampal formation<\/a> \u2014 the area of the brain responsible for memory formation, spatial navigation, and other important functions.<\/p>\n<h3 data-block-key=\"cojep\">Quantum computing<\/h3>\n<p data-block-key=\"2pnb6\">Quantum computers have the potential to solve big, real-world problems across science and industry. But to realize that potential, they must be significantly larger than they are today, and they must reliably perform tasks that cannot be performed on classical computers.<\/p>\n<p data-block-key=\"9739m\">This year, we took an important step towards the development of a large-scale, useful quantum computer. Our breakthrough is the first demonstration of <a href=\"https:\/\/blog.research.google\/2023\/02\/suppressing-quantum-errors-by-scaling.html\" rel=\"noopener\" target=\"_blank\">quantum error correction<\/a>, showing that it\u2019s possible to reduce errors while also increasing the number of qubits. To enable real-world applications, these qubit building blocks must perform more reliably, lowering the error rate from ~1 in 103 typically seen today, to ~1 in 108.<\/p>\n<h2 data-block-key=\"9ugt\">Responsible AI Research<\/h2>\n<h3 data-block-key=\"9scj2\">Design for Responsibility<\/h3>\n<p data-block-key=\"4imna\">Generative AI is having a transformative impact in a wide range of fields including healthcare, education, security, energy, transportation, manufacturing, and entertainment. Given these advances, the importance of designing technologies consistent with our <a href=\"https:\/\/ai.google\/responsibility\/principles\/\" rel=\"noopener\" target=\"_blank\">AI Principles<\/a> remains a top priority. We also recently published case studies of <a href=\"https:\/\/blog.research.google\/2023\/11\/emerging-practices-for-society-centered.html\" rel=\"noopener\" target=\"_blank\">emerging practices in society-centered AI<\/a>. And in our annual <a href=\"https:\/\/storage.googleapis.com\/gweb-uniblog-publish-prod\/documents\/2023_Google_AI_Principles_Progress_Update.pdf\" rel=\"noopener\" target=\"_blank\">AI Principles Progress Update<\/a>, we offer details on how our Responsible AI research is integrated into products and risk management processes.<\/p>\n<p data-block-key=\"bgtem\">Proactive design for Responsible AI begins with identifying and documenting potential harms. For example, we recently <a href=\"https:\/\/deepmind.google\/discover\/blog\/evaluating-social-and-ethical-risks-from-generative-ai\/\" rel=\"noopener\" target=\"_blank\">introduced<\/a> a <a href=\"https:\/\/arxiv.org\/abs\/2310.11986\" rel=\"noopener\" target=\"_blank\">three-layered<\/a> context-based framework for comprehensively evaluating the social and ethical risks of AI systems. During model design, harms can be mitigated with the use of <a href=\"https:\/\/blog.research.google\/2023\/11\/responsible-ai-at-google-research_16.html\" rel=\"noopener\" target=\"_blank\">responsible datasets<\/a>.<\/p>\n<p data-block-key=\"f167\">We are <a href=\"https:\/\/blog.google\/technology\/research\/project-elevate-black-voices-google-research\/\" rel=\"noopener\" target=\"_blank\">partnering with Howard University<\/a> to build high quality African-American English (AAE) datasets to improve our products and make them work well for more people. Our research on <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3593013.3594016\" rel=\"noopener\" target=\"_blank\">globally inclusive cultural representation<\/a> and our publication of the <a href=\"https:\/\/skintone.google\/\" rel=\"noopener\" target=\"_blank\">Monk Skin Tone scale<\/a> furthers our commitments to equitable representation of all people. The insights we gain and techniques we develop not only help us improve our own models, they also power <a href=\"https:\/\/blog.google\/intl\/en-in\/company-news\/using-ai-to-study-demographic-representation-in-indian-tv\/\" rel=\"noopener\" target=\"_blank\">large-scale studies of representation in popular media<\/a> to inform and inspire more inclusive content creation around the world.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\"><figcaption class=\"single-media__caption\">\n<p data-block-key=\"g18z5\">Monk Skin Tone (MST) Scale. See more at skintone.google.<\/p>\n<\/figcaption><\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">With advances in generative image models, <a href=\"https:\/\/blog.research.google\/2023\/08\/responsible-ai-at-google-research.html\" rel=\"noopener\" target=\"_blank\">fair and inclusive representation of people<\/a> remains a top priority. In the development pipeline, we are working to <a href=\"https:\/\/blog.research.google\/2023\/07\/using-societal-context-knowledge-to.html\" rel=\"noopener\" target=\"_blank\">amplify underrepresented voices and to better integrate social context knowledge<\/a>. We proactively address potential harms and bias using <a href=\"https:\/\/arxiv.org\/pdf\/2306.06135.pdf\" rel=\"noopener\" target=\"_blank\">classifiers and filters<\/a>, <a href=\"https:\/\/arxiv.org\/pdf\/2311.17259.pdf\" rel=\"noopener\" target=\"_blank\">careful dataset analysis<\/a>, and in-model mitigations such as fine-tuning, <a href=\"https:\/\/arxiv.org\/abs\/2310.16523\" rel=\"noopener\" target=\"_blank\">reasoning<\/a>, <a href=\"https:\/\/arxiv.org\/abs\/2306.14308\" rel=\"noopener\" target=\"_blank\">few-shot prompting<\/a>, <a href=\"https:\/\/arxiv.org\/abs\/2310.16959\" rel=\"noopener\" target=\"_blank\">data augmentation<\/a> and <a href=\"https:\/\/arxiv.org\/abs\/2310.17022\" rel=\"noopener\" target=\"_blank\">controlled decoding<\/a>, and our research showed that generative AI enables <a href=\"https:\/\/arxiv.org\/abs\/2302.06541\" rel=\"noopener\" target=\"_blank\">higher quality safety classifiers<\/a> to be developed with far less data. We also released <a href=\"https:\/\/developers.googleblog.com\/2023\/10\/make-with-makersuite-part-2-tuning-llms.html\" rel=\"noopener\" target=\"_blank\">a powerful way to better tune models with less data<\/a> giving developers more control of responsibility challenges in generative AI.<\/p>\n<p data-block-key=\"7brqf\">We have developed new<a href=\"https:\/\/arxiv.org\/abs\/2303.08114\" rel=\"noopener\" target=\"_blank\"> state-of-the-art explainability methods<\/a> to identify the role of training data on model behaviors. By <a href=\"https:\/\/arxiv.org\/abs\/2302.06598\" rel=\"noopener\" target=\"_blank\">combining training data attribution methods with agile classifiers<\/a>, we found that we can identify mislabelled training examples. This makes it possible to reduce the noise in training data, leading to significant improvements in model accuracy.<\/p>\n<p data-block-key=\"60k7g\">We initiated several efforts to improve safety and transparency about online content. For example, we introduced <a href=\"https:\/\/deepmind.google\/discover\/blog\/identifying-ai-generated-images-with-synthid\/\" rel=\"noopener\" target=\"_blank\">SynthID<\/a>, a tool for watermarking and identifying AI-generated images. SynthID is imperceptible to the human eye, doesn&#8217;t compromise image quality, and allows the watermark to remain detectable, even after modifications like adding filters, changing colors, and saving with various lossy compression schemes.<\/p>\n<p data-block-key=\"bm3hv\">We also launched <a href=\"https:\/\/blog.google\/products\/search\/google-search-new-fact-checking-features\/\" rel=\"noopener\" target=\"_blank\">About This Image<\/a> to help people assess the credibility of images, showing information like an image&#8217;s history, how it&#8217;s used on other pages, and available metadata about an image. And we <a href=\"https:\/\/arxiv.org\/abs\/2210.03535\" rel=\"noopener\" target=\"_blank\">explored safety methods<\/a> that have been developed in other fields, learning from established situations where there is low-risk tolerance.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\"><figcaption class=\"single-media__caption\">\n<p data-block-key=\"g18z5\">SynthID generates an imperceptible digital watermark for AI-generated images.<\/p>\n<\/figcaption><\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">Privacy remains an essential aspect of our commitment to Responsible AI. We continued improving our state-of-the-art privacy preserving learning algorithm <a href=\"https:\/\/arxiv.org\/abs\/2103.00039\" rel=\"noopener\" target=\"_blank\">DP-FTRL<\/a>, developed the DP-Alternating Minimization algorithm (<a href=\"https:\/\/arxiv.org\/pdf\/2310.15454.pdf\" rel=\"noopener\" target=\"_blank\">DP-AM<\/a>) to enable personalized recommendations with rigorous privacy protection, and defined a new <a href=\"https:\/\/blog.research.google\/2023\/09\/differentially-private-median-and-more.html\" rel=\"noopener\" target=\"_blank\">general paradigm<\/a> to reduce the privacy costs for many aggregation and learning tasks. We also proposed a scheme for <a href=\"https:\/\/openreview.net\/pdf?id=q15zG9CHi8\" rel=\"noopener\" target=\"_blank\">auditing differentially private machine learning systems<\/a>.<\/p>\n<p data-block-key=\"q0ia\">On the applications front we demonstrated that <a href=\"https:\/\/arxiv.org\/pdf\/2308.10888.pdf\" rel=\"noopener\" target=\"_blank\">DP-SGD offers a practical solution<\/a> in the large model fine-tuning regime and showed that images generated by DP diffusion models are <a href=\"https:\/\/arxiv.org\/pdf\/2302.13861.pdf\" rel=\"noopener\" target=\"_blank\">useful for a range of downstream tasks<\/a>. We <a href=\"https:\/\/blog.research.google\/2023\/12\/sparsity-preserving-differentially.html\" rel=\"noopener\" target=\"_blank\">proposed<\/a> a new algorithm for DP training of large embedding models that provides efficient training on TPUs without compromising accuracy.<\/p>\n<p data-block-key=\"6t5ja\">We also teamed up with a broad group of academic and industrial researchers to organize the <a href=\"https:\/\/unlearning-challenge.github.io\/\" rel=\"noopener\" target=\"_blank\">first Machine Unlearning Challenge<\/a> to address the scenario in which training images are forgotten to protect the privacy or rights of individuals. We shared a mechanism for <a href=\"https:\/\/arxiv.org\/pdf\/2311.17035.pdf\" rel=\"noopener\" target=\"_blank\">extractable memorization<\/a>, and <a href=\"https:\/\/arxiv.org\/abs\/2302.03874\" rel=\"noopener\" target=\"_blank\">participatory systems<\/a> that give users more control over their sensitive data.<\/p>\n<p data-block-key=\"bqp1j\">We continued to expand the world\u2019s largest corpus of atypical speech recordings to &gt;1M utterances in <a href=\"https:\/\/sites.research.google\/euphonia\/about\/\" rel=\"noopener\" target=\"_blank\">Project Euphonia<\/a>, which enabled us to train a <a href=\"https:\/\/blog.research.google\/2023\/03\/universal-speech-model-usm-state-of-art.html\" rel=\"noopener\" target=\"_blank\">Universal Speech Model<\/a> to <a href=\"https:\/\/blog.research.google\/2023\/06\/responsible-ai-at-google-research-ai.html\" rel=\"noopener\" target=\"_blank\">better recognize atypical speech by 37%<\/a> on real-world benchmarks.<\/p>\n<p data-block-key=\"6u1mb\">We also built an <a href=\"https:\/\/blog.research.google\/2023\/08\/study-socially-aware-temporally-causal.html\" rel=\"noopener\" target=\"_blank\">audiobook recommendation system<\/a> for students with reading disabilities such as dyslexia.<\/p>\n<h3 data-block-key=\"6ooj6\">Adversarial Testing<\/h3>\n<p data-block-key=\"osru\">Our work in adversarial testing <a href=\"https:\/\/blog.research.google\/2023\/03\/responsible-ai-at-google-research.html\" rel=\"noopener\" target=\"_blank\">engaged community voices<\/a> from historically marginalized communities. We partnered with groups such as the <a href=\"https:\/\/arxiv.org\/abs\/2303.08177\" rel=\"noopener\" target=\"_blank\">Equitable AI Research Round Table<\/a> (EARR) to ensure we represent the diverse communities who use our models and <a href=\"https:\/\/dynabench.org\/tasks\/adversarial-nibbler\" rel=\"noopener\" target=\"_blank\">engage with external users<\/a> to identify potential harms in generative model outputs.<\/p>\n<p data-block-key=\"e0027\">We <a href=\"https:\/\/blog.research.google\/2023\/11\/responsible-ai-at-google-research_16.html\" rel=\"noopener\" target=\"_blank\">established a dedicated Google AI Red Team<\/a> focused on testing AI models and products for security, privacy, and abuse risks. We showed that attacks such as \u201c<a href=\"https:\/\/arxiv.org\/pdf\/2302.10149.pdf?isApp=1\" rel=\"noopener\" target=\"_blank\">poisoning<\/a>\u201d or <a href=\"https:\/\/arxiv.org\/pdf\/2306.15447.pdf\" rel=\"noopener\" target=\"_blank\">adversarial examples<\/a> can be applied to production models and surface additional risks such as memorization in both <a href=\"https:\/\/www.usenix.org\/system\/files\/usenixsecurity23-carlini.pdf\" rel=\"noopener\" target=\"_blank\">image<\/a> and <a href=\"https:\/\/arxiv.org\/pdf\/2311.17035.pdf\" rel=\"noopener\" target=\"_blank\">text generative models<\/a>. We also demonstrated that defending against such attacks can be challenging, as merely applying defenses can cause other <a href=\"https:\/\/arxiv.org\/pdf\/2309.05610.pdf\" rel=\"noopener\" target=\"_blank\">security and privacy leakages<\/a>. We also introduced model evaluation for <a href=\"https:\/\/arxiv.org\/abs\/2305.15324\" rel=\"noopener\" target=\"_blank\">extreme risks<\/a>, such as offensive cyber capabilities or strong manipulation skills.<\/p>\n<h3 data-block-key=\"5douj\">Democratizing AI Though Tools and Education<\/h3>\n<p data-block-key=\"384ub\">As we advance the state-of-the-art in ML and AI, we also want to ensure people can understand and apply AI to specific problems. We released <a href=\"https:\/\/makersuite.google.com\/\" rel=\"noopener\" target=\"_blank\">MakerSuite<\/a> (now Google AI Studio), a web-based tool that enables AI developers to quickly iterate and build lightweight AI-powered apps. To help AI engineers better understand and debug AI, we released <a href=\"https:\/\/pair-code.github.io\/lit\/\" rel=\"noopener\" target=\"_blank\">LIT 1.0<\/a>, a state-of-the-art, open-source debugger for machine learning models.<\/p>\n<p data-block-key=\"c8g29\"><a href=\"https:\/\/colab.google\/\" rel=\"noopener\" target=\"_blank\">Colab<\/a>, our tool that helps developers and students access powerful computing resources right in their web browser, reached over 10 million users. We\u2019ve just added <a href=\"https:\/\/blog.google\/technology\/ai\/democratizing-access-to-ai-enabled-coding-with-colab\/\" rel=\"noopener\" target=\"_blank\">AI-powered code assistance<\/a> to all users at no cost \u2014 making Colab an even more helpful and integrated experience in data and ML workflows.<\/p>\n<\/div>\n<figure class=\"single-media single-media--inline\"><figcaption class=\"single-media__caption\">\n<p data-block-key=\"g18z5\">One of the most used features is \u201cExplain error\u201d \u2014 whenever the user encounters an execution error in Colab, the code assistance model provides an explanation along with a potential fix.<\/p>\n<\/figcaption><\/figure>\n<div class=\"gdm-rich-text rich-text\">\n<p data-block-key=\"uou0g\">To ensure AI produces accurate knowledge when put to use, we also recently introduced <a href=\"https:\/\/deepmind.google\/discover\/blog\/funsearch-making-new-discoveries-in-mathematical-sciences-using-large-language-models\/\" rel=\"noopener\" target=\"_blank\">FunSearch<\/a>, a new approach that generates verifiably true knowledge in mathematical sciences using evolutionary methods and large language models.<\/p>\n<p data-block-key=\"f0ui9\">For AI engineers and product designers, we\u2019re updating the <a href=\"https:\/\/pair.withgoogle.com\/guidebook\/\" rel=\"noopener\" target=\"_blank\">People + AI Guidebook<\/a> with generative AI best practices, and we continue to design <a href=\"https:\/\/pair.withgoogle.com\/explorables\/\" rel=\"noopener\" target=\"_blank\">AI Explorables<\/a>, which includes <a href=\"https:\/\/pair.withgoogle.com\/explorables\/uncertainty-ood\/\" rel=\"noopener\" target=\"_blank\">how and why models sometimes make incorrect predictions confidently<\/a>.<\/p>\n<h3 data-block-key=\"2b0ne\">Community Engagement<\/h3>\n<p data-block-key=\"faeko\">We continue to advance the fields of AI and computer science by publishing much of our work and participating in and organizing conferences. We have published more than 500 papers so far this year, and have strong presences at conferences like ICML (see the <a href=\"https:\/\/blog.research.google\/2023\/07\/google-at-icml-2023.html\" rel=\"noopener\" target=\"_blank\">Google Research<\/a> and <a href=\"https:\/\/deepmind.google\/discover\/blog\/google-deepmind-research-at-icml-2023\/\" rel=\"noopener\" target=\"_blank\">Google DeepMind<\/a> posts), ICLR (<a href=\"https:\/\/blog.research.google\/2023\/04\/google-at-iclr-2023.html\" rel=\"noopener\" target=\"_blank\">Google Research<\/a>, <a href=\"https:\/\/deepmind.google\/discover\/blog\/deepminds-latest-research-at-iclr-2023\/\" rel=\"noopener\" target=\"_blank\">Google DeepMind<\/a>), NeurIPS (<a href=\"https:\/\/blog.research.google\/2023\/12\/google-at-neurips-2023.html\" rel=\"noopener\" target=\"_blank\">Google Research<\/a>, <a href=\"https:\/\/deepmind.google\/discover\/blog\/google-deepmind-at-neurips-2023\/\" rel=\"noopener\" target=\"_blank\">Google DeepMind<\/a>), <a href=\"https:\/\/blog.research.google\/2023\/10\/google-at-iccv-2023.html\" rel=\"noopener\" target=\"_blank\">ICCV<\/a>, <a href=\"https:\/\/blog.research.google\/2023\/06\/google-at-cvpr-2023.html\" rel=\"noopener\" target=\"_blank\">CVPR<\/a>, <a href=\"https:\/\/blog.research.google\/2023\/07\/google-at-acl-2023.html\" rel=\"noopener\" target=\"_blank\">ACL<\/a>, <a href=\"https:\/\/blog.research.google\/2023\/04\/google-at-chi-2023.html\" rel=\"noopener\" target=\"_blank\">CHI<\/a>, and <a href=\"https:\/\/blog.research.google\/2023\/08\/google-at-interspeech-2023.html\" rel=\"noopener\" target=\"_blank\">Interspeech<\/a>. We are also working to support researchers around the world, participating in events like the <a href=\"https:\/\/deeplearningindaba.com\/2023\/google-outreach-mentorship-programme\/\" rel=\"noopener\" target=\"_blank\">Deep Learning Indaba<\/a>, <a href=\"https:\/\/khipu.ai\/khipu2023\/khipu-2023-speakers2023\/\" rel=\"noopener\" target=\"_blank\">Khipu<\/a>, supporting <a href=\"https:\/\/blog.google\/around-the-globe\/google-latin-america\/phd-fellowship-research-latin-america\/\" rel=\"noopener\" target=\"_blank\">PhD Fellowships in Latin America<\/a>, and more. We also worked with partners from 33 academic labs to pool data from 22 different robot types and create the <a href=\"https:\/\/deepmind.google\/discover\/blog\/scaling-up-learning-across-many-different-robot-types\/\" rel=\"noopener\" target=\"_blank\">Open X-Embodiment dataset and RT-X model<\/a> to better advance responsible AI development.<\/p>\n<p data-block-key=\"42h3n\">Google has spearheaded an industry-wide effort to develop <a href=\"https:\/\/mlcommons.org\/working-groups\/ai-safety\/ai-safety\/\" rel=\"noopener\" target=\"_blank\">AI safety benchmarks<\/a> under the <a href=\"https:\/\/mlcommons.org\/\" rel=\"noopener\" target=\"_blank\">MLCommons<\/a> standards organization with participation from several major players in the generative AI space including OpenAI, Anthropic, Microsoft, Meta, Hugging Face, and more. Along with others in the industry we also <a href=\"https:\/\/blog.google\/outreach-initiatives\/public-policy\/google-microsoft-openai-anthropic-frontier-model-forum\/\" rel=\"noopener\" target=\"_blank\">co-founded<\/a> the <a href=\"https:\/\/www.frontiermodelforum.org\/\" rel=\"noopener\" target=\"_blank\">Frontier Model Forum<\/a> (FMF), which is focused on ensuring safe and responsible development of frontier AI models. With our FMF partners and other philanthropic organizations, we launched a $10 million <a href=\"https:\/\/blog.google\/outreach-initiatives\/public-policy\/google-microsoft-anthropic-open-ai-frontier-model-forum-executive-director\/\" rel=\"noopener\" target=\"_blank\">AI Safety Fund<\/a> to advance research into the ongoing development of the tools for society to effectively test and evaluate the most capable AI models.<\/p>\n<p data-block-key=\"8rh2k\">In close partnership with <a href=\"http:\/\/google.org\/\" rel=\"noopener\" target=\"_blank\">Google.org<\/a>, we <a href=\"https:\/\/blog.google\/technology\/ai\/google-ai-data-un-global-goals\/\" rel=\"noopener\" target=\"_blank\">worked with the United Nations<\/a> to build the <a href=\"https:\/\/unstats.un.org\/UNSDWebsite\/undatacommons\/sdgs\" rel=\"noopener\" target=\"_blank\">UN Data Commons for the Sustainable Development Goals<\/a>, a tool that tracks metrics across the 17 <a href=\"https:\/\/sdgs.un.org\/goals\" rel=\"noopener\" target=\"_blank\">Sustainable Development Goals<\/a>, and <a href=\"https:\/\/globalgoals.withgoogle.com\/globalgoals\/supported-organizations\" rel=\"noopener\" target=\"_blank\">supported projects<\/a> from NGOs, academic institutions, and social enterprises on <a href=\"https:\/\/blog.google\/outreach-initiatives\/google-org\/httpsbloggoogleoutreach-initiativesgoogle-orgunited-nations-global-goals-google-ai-\/\" rel=\"noopener\" target=\"_blank\">using AI to accelerate progress on the SDGs<\/a>.<\/p>\n<p data-block-key=\"7v1kb\">The items highlighted in this post are a small fraction of the research work we have done throughout the last year. Find out more at the <a href=\"https:\/\/blog.research.google\/\" rel=\"noopener\" target=\"_blank\">Google Research<\/a> and <a href=\"https:\/\/deepmind.google\/discover\/blog\/\" rel=\"noopener\" target=\"_blank\">Google DeepMind<\/a> blogs, and our <a href=\"https:\/\/research.google\/pubs\/\" rel=\"noopener\" target=\"_blank\">list of publications<\/a>.<\/p>\n<h2 data-block-key=\"9qski\">Future Vision<\/h2>\n<p data-block-key=\"4q23p\">As multimodal models become even more capable, they will empower people to make incredible progress in areas from science to education to entirely new areas of knowledge.<\/p>\n<p data-block-key=\"cd4q2\">Progress continues apace, and as the year advances, and our products and research advance as well, people will find more and interesting creative uses for AI.<\/p>\n<p data-block-key=\"380h5\">Ending this Year-in-Review where we began, as we say in <a href=\"https:\/\/ai.google\/static\/documents\/google-why-we-focus-on-ai.pdf\" rel=\"noopener\" target=\"_blank\"><i>Why We Focus on AI (and to what end)<\/i><\/a>:<\/p>\n<p data-block-key=\"7l9m5\"><i>If pursued boldly and responsibly, we believe that AI can be a foundational technology that transforms the lives of people everywhere \u2014 this is what excites us!<\/i><\/p>\n<\/div>\n<aside class=\"notes\">\n<div class=\"glue-page\">\n<div class=\"gdm-rich-text notes__inner\">\n<p data-block-key=\"empri\">This Year-in-Review is cross-posted on both the <a href=\"https:\/\/blog.research.google\/2023\/12\/2023-year-of-groundbreaking-advances-in.html\" rel=\"noopener\" target=\"_blank\">Google Research Blog<\/a> and the <a href=\"https:\/\/deepmind.google\/discover\/blog\/2023-a-year-of-groundbreaking-advances-in-ai-and-computing\/\" rel=\"noopener\" target=\"_blank\">Google DeepMind Blog<\/a>.<\/p>\n<\/p><\/div>\n<\/p><\/div>\n<\/aside>\n<aside class=\"related-posts\">\n<\/aside><\/div>\n<p>[ad_2]<br \/>\n<br \/><a href=\"https:\/\/deepmind.google\/discover\/blog\/2023-a-year-of-groundbreaking-advances-in-ai-and-computing\/\">Source link <\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>[ad_1] Company Published 22 December 2023 Authors By Jeff Dean, Chief Scientist, Google DeepMind &amp; Google Research, Demis<\/p>\n","protected":false},"author":2,"featured_media":922,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[21],"tags":[],"class_list":["post-921","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-deepmind-ai"],"_links":{"self":[{"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/posts\/921","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/comments?post=921"}],"version-history":[{"count":1,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/posts\/921\/revisions"}],"predecessor-version":[{"id":2592,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/posts\/921\/revisions\/2592"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/media\/922"}],"wp:attachment":[{"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/media?parent=921"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/categories?post=921"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/tags?post=921"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}