{"id":344,"date":"2023-02-11T01:41:21","date_gmt":"2023-02-11T01:41:21","guid":{"rendered":"https:\/\/todaysainews.com\/index.php\/2023\/02\/11\/algorithmic-advances-google-ai-blog\/"},"modified":"2025-04-27T07:34:18","modified_gmt":"2025-04-27T07:34:18","slug":"algorithmic-advances-google-ai-blog","status":"publish","type":"post","link":"https:\/\/todaysainews.com\/index.php\/2023\/02\/11\/algorithmic-advances-google-ai-blog\/","title":{"rendered":"Algorithmic advances \u2013 Google AI Blog"},"content":{"rendered":"<p> [ad_1]<br \/>\n<\/p>\n<div id=\"post-body-126309359572930839\">\n<span class=\"byline-author\">Posted by Vahab Mirrokni, VP and Google Fellow, Google Research<\/span><\/p>\n<p><img decoding=\"async\" src=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEgXanejmCPJF5TnYAK2e-ifvooB_HJ7BY3EPSNlwBFphjLB-je2Ka-9flly5P8a73kxK5sT0wqQI7IMAvmcYat8wSZeRCygJU81zw2kGKR1P52MgsbjXeAxcW9cMLNbDBDOoJzQzrrYmiTXiixOrsGchLMHdHt6qZZhTc5PBIJavA0Jrdg5JUzR2NKs1w\/s1038\/Algos-hero.png\" style=\"display: none;\"\/><a name=\"more\"\/> <\/p>\n<div style=\"margin-left: auto; margin-right: auto; text-align: center;\"><span style=\"font-size: small; text-align: center;\"><i>(This is Part 5 in our series of posts covering different topical areas of research at Google. You can find other posts in the series <a href=\"#ToC\">here.)<\/a><\/i><\/span>\n<\/div>\n<p><a name=\"top\"> <\/a><\/p>\n<p>\nRobust algorithm design is the backbone of systems across Google, particularly for our ML and AI models. Hence, developing algorithms with improved efficiency, performance and speed remains a high priority as it empowers services ranging from Search and Ads to Maps and YouTube. Google Research has been at the forefront of this effort, developing many innovations from privacy-safe recommendation systems to scalable solutions for large-scale ML. In 2022, we continued this journey, and advanced the state-of-the-art in several related areas. Here we highlight our progress in a subset of these, including scalability, privacy, market algorithms, and algorithmic foundations.\n<\/p>\n<p><\/p>\n<p><a name=\"Theme1\"> <\/a><\/p>\n<p><\/p>\n<h2>Scalable algorithms: Graphs, clustering, and optimization<\/h2>\n<p>\nAs the need to handle large-scale datasets increases, scalability and reliability of complex  algorithms that also exhibit improved explainability, robustness, and speed remain a high priority. We continued our efforts in developing new algorithms for handling large datasets in various areas, including <a href=\"https:\/\/en.wikipedia.org\/wiki\/Unsupervised_learning\">unsupervised<\/a> and <a href=\"https:\/\/en.wikipedia.org\/wiki\/Semi-Supervised_Learning#Semi-supervised_learning\">semi-supervised learning<\/a>, <a href=\"https:\/\/en.wikipedia.org\/wiki\/Graph_neural_network\">graph-based learning<\/a>, <a href=\"https:\/\/en.wikipedia.org\/wiki\/Cluster_analysis\">clustering<\/a>, and large-scale optimization.\n<\/p>\n<p>\nAn important component of such systems is to build a <em><a href=\"https:\/\/en.wikipedia.org\/wiki\/Nearest_neighbor_graph\">similarity graph<\/a><\/em> \u2014 a nearest-neighbor graph that represents similarities between objects. For scalability and speed, this graph should be sparse without compromising quality. We proposed a <a href=\"https:\/\/neurips.cc\/Conferences\/2022\/ScheduleMultitrack?event=53141\">2-hop spanner technique<\/a>, called STAR<em>,<\/em> as an efficient and distributed graph building strategy, and showed how it significantly decreases the number of similarity computations in theory and practice, building much sparser graphs while producing high-quality graph learning or clustering outputs. As an example, for graphs with 10T edges, we demonstrate ~100-fold improvements in pairwise similarity comparisons and significant running time speedups with negligible quality loss. We had previously applied this idea to develop massively parallel algorithms for <a href=\"https:\/\/proceedings.mlr.press\/v162\/cohen-addad22b.html\">metric<\/a>, and <a href=\"https:\/\/epubs.siam.org\/doi\/10.1137\/1.9781611977073.66\">minimum-size clustering<\/a>. More broadly in the context of clustering, we developed the first linear-time <a href=\"https:\/\/proceedings.mlr.press\/v139\/dhulipala21a.html\">hierarchical agglomerative clustering<\/a> (HAC) algorithm as well as<a href=\"https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/view\/16902\"> DBSCAN<\/a>, the first parallel <a href=\"http:\/\/go\/parhac-paper\">algorithm for HAC with logarithmic depth<\/a>, which achieves 50x speedup on 100B-edge graphs. We also designed improved sublinear algorithms for different flavors of clustering problems such as <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3519935.3519979\">geometric linkage clustering<\/a>, <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3519935.3519979\">constant-round correlation clustering<\/a>, and <a href=\"https:\/\/epubs.siam.org\/doi\/abs\/10.1137\/1.9781611977554.ch101\">fully dynamic k-clustering<\/a>.\n<\/p>\n<p>\nInspired by the success of multi-core processing (e.g.,\u00a0<a href=\"https:\/\/github.com\/ParAlg\/gbbs\">GBBS<\/a>), we embarked on a mission to develop graph mining algorithms that can handle graphs with 100B edges <em>on a single multi-core machine<\/em>. The big challenge here is to achieve fast (e.g., sublinear) parallel running time (i.e., depth).<em> <\/em>Following our previous work for<a href=\"http:\/\/go\/correlation-clustering-paper\"> community detection<\/a> and correlation clustering, we developed an algorithm for HAC, called <a href=\"http:\/\/go\/parhac-paper\">ParHAC<\/a>, which has provable polylogarithmic depth and near-linear work and achieves a 50x speedup. As an example, it took ParHAC only ~10 minutes to find an <a href=\"https:\/\/proceedings.neurips.cc\/paper\/2017\/file\/2e1b24a664f5e9c18f407b2f9c73e821-Paper.pdf\">approximate affinity hierarchy<\/a> over a graph of over 100B edges, and ~3 hours to find the full HAC on a single machine. Following <a href=\"https:\/\/research.google\/pubs\/pub50343\/\">our previous work on distributed HAC<\/a>, we use these multi-core algorithms as a subroutine within our distributed algorithms in order to handle tera-scale graphs.\n<\/p>\n<p>\nWe also had a number of interesting results on <a href=\"https:\/\/en.wikipedia.org\/wiki\/Graph_neural_network\">graph neural networks<\/a> (GNN) in 2022. We provided a <a href=\"https:\/\/www.jmlr.org\/papers\/volume23\/20-852\/20-852.pdf\">model-based taxonomy that unified many graph learning methods<\/a>. In addition, we discovered insights for GNN models from their <a href=\"https:\/\/dl.acm.org\/doi\/abs\/10.1145\/3534678.3539203\">performance across thousands of graphs<\/a> with varying structure (shown below). We also proposed a <a href=\"https:\/\/ojs.aaai.org\/index.php\/AAAI\/article\/download\/20548\/version\/18845\/20307\">new hybrid architecture<\/a> to overcome the depth requirements of existing GNNs for solving fundamental graph problems, such as shortest paths and the <a href=\"https:\/\/en.wikipedia.org\/wiki\/Minimum_spanning_tree\">minimum spanning tree<\/a>.\n<\/p>\n<table align=\"center\" cellpadding=\"0\" cellspacing=\"0\" class=\"tr-caption-container\" style=\"margin-left: auto; margin-right: auto;\">\n<tbody>\n<tr>\n<td style=\"text-align: center;\"><a href=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEgCSf6FQjhHgF9yseDxsdPINYzQHVU6DBbPMKSLBbNb_NRvIXI5JemxUdcVpMYyA0dWx9IwhF8L6Ig_UIro4GtaAm-78kuOrRj0Ci1yV09UAfeWXTx1I-1wbnRCchhMTeUp0r4YdFjQ4vlWja6kBvyujW9LUGAJ0gfiK6epmL2FWqDa_aqeMtSmoKwY\/s800\/image3.gif\" style=\"margin-left: auto; margin-right: auto;\"><img loading=\"lazy\" decoding=\"async\" border=\"0\" data-original-height=\"683\" data-original-width=\"800\" height=\"546\" src=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEgCSf6FQjhHgF9yseDxsdPINYzQHVU6DBbPMKSLBbNb_NRvIXI5JemxUdcVpMYyA0dWx9IwhF8L6Ig_UIro4GtaAm-78kuOrRj0Ci1yV09UAfeWXTx1I-1wbnRCchhMTeUp0r4YdFjQ4vlWja6kBvyujW9LUGAJ0gfiK6epmL2FWqDa_aqeMtSmoKwY\/w640-h546\/image3.gif\" width=\"640\"\/><\/a><\/td>\n<\/tr>\n<tr>\n<td class=\"tr-caption\" style=\"text-align: center;\">Relative performance results of three GNN variants (<a href=\"https:\/\/arxiv.org\/abs\/1609.02907\">GCN<\/a>, <a href=\"https:\/\/arxiv.org\/abs\/1810.05997\">APPNP<\/a>, <a href=\"https:\/\/arxiv.org\/abs\/1906.12192\">FiLM<\/a>) across 50,000 distinct node classification datasets in <a href=\"https:\/\/ai.googleblog.com\/2022\/05\/graphworld-advances-in-graph.html\">GraphWorld<\/a>. We find that academic GNN benchmark datasets exist in regions where model rankings do not change. GraphWorld can discover previously unexplored graphs that reveal new insights about GNN architectures.<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p>\nFurthermore, to bring some of these many advances to the broader community, we had three releases of our flagship modeling library for building <a href=\"https:\/\/github.com\/tensorflow\/gnn\">graph neural networks in TensorFlow<\/a> (TF-GNN). Highlights include a model library and model orchestration API to make it easy to compose GNN solutions. Following our <a href=\"https:\/\/gm-neurips-2020.github.io\/\">NeurIPS\u201920 workshop<\/a> on <em>Mining and Learning with Graphs at Scale<\/em>, we ran a <a href=\"https:\/\/docs.google.com\/presentation\/d\/1s-Tnbhgx5iwSPLHuFhS7MUOThqFitZCVLyrM1yGlVkg\/edit#slide=id.g13da81c65c5_0_0\">workshop on graph-based learning<\/a> at <a href=\"https:\/\/icml.cc\/Conferences\/2022\">ICML\u201922<\/a>, and a <a href=\"https:\/\/www.youtube.com\/watch?v=tZfJbgU86Ag\">tutorial for GNNs in TensorFlow<\/a> at <a href=\"https:\/\/nips.cc\/Conferences\/2022\">NeurIPS\u201922<\/a>.\n<\/p>\n<p>\nIn \u201c<a href=\"https:\/\/ai.googleblog.com\/2022\/02\/robust-routing-using-electrical-flows.html\">Robust Routing Using Electrical Flows<\/a>\u201d, we presented a <a href=\"https:\/\/dl.acm.org\/doi\/pdf\/10.1145\/3474717.3483961\">recent paper<\/a> that proposed a Google Maps solution to efficiently compute alternate paths in road networks that are resistant to failures (e.g., closures, incidents). We demonstrate how it significantly outperforms the state-of-the-art <a href=\"https:\/\/drops.dagstuhl.de\/opus\/volltexte\/2013\/4248\/\">plateau and penalty methods<\/a> on real-world road networks.\n<\/p>\n<table align=\"center\" cellpadding=\"0\" cellspacing=\"0\" class=\"tr-caption-container\" style=\"margin-left: auto; margin-right: auto;\">\n<tbody>\n<tr>\n<td style=\"text-align: center;\"><a href=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEh2bOHjg2t4y7jTjCfIIwgL84rYC5HlWAopPganz6FobFRf1KNMgKiPCQdLPZxlBBYs3fERCiFp8qsUQBkT_8NHf-AqwsRQvIsaE3UDFyWpLB_IiaA2EqEdlPHvxsj-XvMLhoiv1opHZ2OhFpS8t4VcehjzYa0ptlsUB5O_SVRl2rv6QJeVOFH-n1Zt\/s1250\/image2.png\" style=\"margin-left: auto; margin-right: auto;\"><img decoding=\"async\" border=\"0\" data-original-height=\"341\" data-original-width=\"1250\" src=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEh2bOHjg2t4y7jTjCfIIwgL84rYC5HlWAopPganz6FobFRf1KNMgKiPCQdLPZxlBBYs3fERCiFp8qsUQBkT_8NHf-AqwsRQvIsaE3UDFyWpLB_IiaA2EqEdlPHvxsj-XvMLhoiv1opHZ2OhFpS8t4VcehjzYa0ptlsUB5O_SVRl2rv6QJeVOFH-n1Zt\/s16000\/image2.png\"\/><\/a><\/td>\n<\/tr>\n<tr>\n<td class=\"tr-caption\" style=\"text-align: center;\">Example of how we construct the electrical circuit corresponding to the road network. The current can be decomposed into three flows, i<sub>1<\/sub>, i<sub>2<\/sub> and i<sub>3<\/sub>, each of which corresponds to a viable alternate path from Fremont, CA to San Rafael, CA.<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p>\nOn the optimization front, we <a href=\"https:\/\/ai.googleblog.com\/2023\/02\/open-source-vizier-towards-reliable-and.html\">open-sourced<\/a> <a href=\"https:\/\/proceedings.mlr.press\/v188\/song22a.html\">Vizier<\/a>, our flagship <a href=\"https:\/\/en.wikipedia.org\/wiki\/Derivative-free_optimization\">blackbox optimization<\/a> and hyperparameter tuning library at Google. We also developed new techniques for<a href=\"https:\/\/en.wikipedia.org\/wiki\/Linear_programming\"> linear programming<\/a> (LP) solvers that address scalability limits caused by their reliance on matrix factorizations, which restricts the opportunity for parallelism and distributed approaches. To this end, we open-sourced a <a href=\"https:\/\/www.cs.umd.edu\/~tomg\/projects\/pdhg\/\">primal-dual hybrid gradient<\/a> (PDHG) solution for LP called <a href=\"https:\/\/github.com\/google\/or-tools\">primal-dual linear programming<\/a> (PDLP), a new first-order solver for large-scale LP problems. PDLP has been used to solve real-world problems with as many as 12B non-zeros (and an internal distributed version scaled to 92B non-zeros). PDLP&#8217;s effectiveness is due to a combination of <a href=\"https:\/\/arxiv.org\/abs\/2105.12715\">theoretical developments<\/a> and <a href=\"https:\/\/papers.nips.cc\/paper\/2021\/file\/a8fbbd3b11424ce032ba813493d95ad7-Paper.pdf\">algorithm engineering<\/a>.\n<\/p>\n<table align=\"center\" cellpadding=\"0\" cellspacing=\"0\" class=\"tr-caption-container\" style=\"margin-left: auto; margin-right: auto;\">\n<tbody>\n<tr>\n<td style=\"text-align: center;\"><a href=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEh4GXw0trA84pvfBYc8bK0EtaAC0TP0tCqEUHP2ZZUCvSgg2mtM4xlmdjpFH7yxnQ2SXoafsw_BWk4DDgLtG6tGGiyApasFM9e59seyWH9j_JqhgpKYzRUEFsO_aC_-GJRKF-6zY7ndnWP5sJOhh0Mzz0owLmpnxwBUy5nZxO5nOmWZyswvcRZiNsOR\/s1433\/image1.gif\" style=\"margin-left: auto; margin-right: auto;\"><img decoding=\"async\" border=\"0\" data-original-height=\"723\" data-original-width=\"1433\" src=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEh4GXw0trA84pvfBYc8bK0EtaAC0TP0tCqEUHP2ZZUCvSgg2mtM4xlmdjpFH7yxnQ2SXoafsw_BWk4DDgLtG6tGGiyApasFM9e59seyWH9j_JqhgpKYzRUEFsO_aC_-GJRKF-6zY7ndnWP5sJOhh0Mzz0owLmpnxwBUy5nZxO5nOmWZyswvcRZiNsOR\/s16000\/image1.gif\"\/><\/a><\/td>\n<\/tr>\n<tr>\n<td class=\"tr-caption\" style=\"text-align: center;\">With OSS Vizier, multiple clients each send a \u201cSuggest\u201d request to the Service API, which produces Suggestions for the clients using <a href=\"https:\/\/oss-vizier.readthedocs.io\/en\/latest\/guides\/developer\/writing_algorithms.html\">Pythia policies<\/a>. The clients evaluate these suggestions and return measurements. All transactions are stored to allow fault-tolerance.<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p align=\"right\"><a href=\"#top\">Top<\/a><\/p>\n<p><a name=\"Theme2\"> <\/a><\/p>\n<h2>Privacy and federated learning<\/h2>\n<p>\nRespecting user privacy while providing high-quality services remains a top priority for all Google systems. Research in this area spans many products and uses principles from <a href=\"https:\/\/en.wikipedia.org\/wiki\/Differential_privacy\">differential privacy<\/a> (DP) and <a href=\"https:\/\/pair.withgoogle.com\/explorables\/federated-learning\/\">federated learning<\/a>.\n<\/p>\n<p>\nFirst of all, we have made a variety of algorithmic advances to address the problem of training large neural networks with DP. Building on our <a href=\"https:\/\/arxiv.org\/abs\/2103.00039\">earlier work<\/a>, which <a href=\"https:\/\/ai.googleblog.com\/2022\/02\/federated-learning-with-formal.html\">enabled us<\/a> to launch a DP neural network based on the <a href=\"https:\/\/arxiv.org\/abs\/2103.00039\">DP-FTRL<\/a> algorithm, we developed the <a href=\"https:\/\/arxiv.org\/abs\/2202.08312\">matrix factorization DP-FTRL approach<\/a>. This work demonstrates that one can design a mathematical program to optimize over a large set of possible DP mechanisms to find those best suited for specific learning problems. We also establish <a href=\"https:\/\/arxiv.org\/abs\/2204.10376\">margin guarantees<\/a> that are independent of the input feature dimension for DP learning of neural networks and kernel-based methods. We further <a href=\"https:\/\/arxiv.org\/abs\/2211.06530\">extend<\/a> this concept to a broader range of ML tasks, matching baseline performance with 300x less computation. For fine-tuning of large models, we <a href=\"https:\/\/openreview.net\/forum?id=FR--mkQu0dw\">argued<\/a> that once pre-trained, these models (even with DP) essentially operate over a low-dimensional subspace, hence circumventing the <a href=\"https:\/\/en.wikipedia.org\/wiki\/Curse_of_dimensionality\">curse of dimensionality<\/a> that DP imposes.\n<\/p>\n<div class=\"separator\" style=\"clear: both; text-align: center;\"><a href=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEj4iIlXD12ZpIyFydDTBaCCUNat0SzyDUTYKW9hhDPPEiwFlIDG2SX5Z1kivA9FDSgYbJnD0nFhYz-zYVwdXaFi1VMwFR6MM-D30of8jJ7ZrSY5QmZR63ctzN_PTcvubU98iSYq7dDiIPTr1DgNso0Dw3YWVXuD9phAARC2EkTxhAubup0fEk5eKpcL\/s1974\/image4.png\" style=\"margin-left: 1em; margin-right: 1em;\"><img decoding=\"async\" border=\"0\" data-original-height=\"496\" data-original-width=\"1974\" src=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEj4iIlXD12ZpIyFydDTBaCCUNat0SzyDUTYKW9hhDPPEiwFlIDG2SX5Z1kivA9FDSgYbJnD0nFhYz-zYVwdXaFi1VMwFR6MM-D30of8jJ7ZrSY5QmZR63ctzN_PTcvubU98iSYq7dDiIPTr1DgNso0Dw3YWVXuD9phAARC2EkTxhAubup0fEk5eKpcL\/s16000\/image4.png\"\/><\/a><\/div>\n<p>\n  On the algorithmic front, for estimating the entropy of a high-dimensional distribution, we obtained <a href=\"https:\/\/openreview.net\/forum?id=OZEmgSbRQW\">local DP mechanisms<\/a> (that work even when as little as one bit per sample is available) and efficient <a href=\"https:\/\/arxiv.org\/abs\/2210.15178\">shuffle DP mechanisms<\/a>. We proposed a more accurate <a href=\"https:\/\/arxiv.org\/abs\/2201.12333\">method<\/a> to simultaneously estimate the top-<em>k<\/em> most popular items in the database in a private manner, which we employed in the <a href=\"https:\/\/arxiv.org\/abs\/2201.11603\">Plume<\/a> library. Moreover, we showed a near-optimal approximation <a href=\"https:\/\/openreview.net\/forum?id=9lQmaKMxIUD\">algorithm for DP clustering<\/a> in the <a href=\"https:\/\/ai.googleblog.com\/2021\/03\/massively-parallel-graph-computation.html\">massively parallel computing<\/a> (MPC) model, which further improves on our previous work for <a href=\"https:\/\/ai.googleblog.com\/2021\/10\/practical-differentially-private.html\">scalable<\/a> and <a href=\"https:\/\/arxiv.org\/abs\/2206.08646\">distributed<\/a> settings.<\/p>\n<p>\nAnother exciting research direction is the intersection of privacy and <a href=\"https:\/\/en.wikipedia.org\/wiki\/Streaming_algorithm\">streaming<\/a>. We obtained a near-optimal approximation-space <a href=\"https:\/\/arxiv.org\/abs\/2301.05605\">trade-off<\/a> for the private frequency moments and a new <a href=\"https:\/\/arxiv.org\/abs\/2211.11718\">algorithm<\/a> for privately counting distinct elements in the <a href=\"https:\/\/link.springer.com\/chapter\/10.1007\/978-0-387-47534-9_8\">sliding window streaming<\/a> model. We also presented a general hybrid <a href=\"https:\/\/arxiv.org\/abs\/2107.14527\">framework<\/a> for studying adversarial streaming.\n<\/p>\n<p>\nAddressing applications at the intersection of security and privacy, we developed new <a href=\"https:\/\/petsymposium.org\/popets\/2022\/popets-2022-0019.pdf\">algorithms<\/a> that are secure, private, and communication-efficient, for measuring cross-publisher reach and frequency. The <a href=\"https:\/\/wfanet.org\/\">World Federation of Advertisers<\/a> has adopted these algorithms as part of their <a href=\"https:\/\/github.com\/world-federation-of-advertisers\/cross-media-measurement\">measurement system<\/a>. In subsequent work, we developed new <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3548606.3559383\">protocols<\/a> that are secure and private for computing sparse histograms in the two-server model of DP. These protocols are efficient from both computation and communication points of view, are substantially better than what standard methods would yield, and combine tools and techniques from sketching, cryptography and multiparty computation, and DP.\n<\/p>\n<p>\nWhile we have trained <a href=\"https:\/\/arxiv.org\/abs\/2108.01624\">BERT<\/a> and <a href=\"https:\/\/aclanthology.org\/2022.findings-acl.171\/\">transformers<\/a> with DP, understanding training example memorization in large language models (LLMs) is a heuristic way to evaluate their privacy. In particular, we <a href=\"https:\/\/arxiv.org\/abs\/2207.00099\">investigated<\/a> when and why LLMs forget (potentially memorized) training examples during training. Our findings suggest that earlier-seen examples may observe privacy benefits at the expense of examples seen later. We also <a href=\"https:\/\/arxiv.org\/abs\/2202.07646\">quantified<\/a> the degree to which LLMs emit memorized training data.\n<\/p>\n<p align=\"right\"><a href=\"#top\">Top<\/a><\/p>\n<p><a name=\"Theme3\"> <\/a><\/p>\n<h2>Market algorithms and causal inference<\/h2>\n<p>\nWe also continued our <a href=\"https:\/\/research.google\/teams\/market-algorithms\/\">research in improving online marketplaces<\/a> in 2022. For example, an important recent area in ad auction research is the study of auto-bidding online advertising where the majority of bidding happens via proxy bidders that optimize higher-level objectives on behalf of advertisers. The complex dynamics of users, advertisers, bidders, and ad platforms leads to non-trivial problems in this space. Following our earlier work in <a href=\"https:\/\/link.springer.com\/chapter\/10.1007\/978-3-030-35389-6_2\">analyzing<\/a> and <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3442381.3450052\">improving<\/a> mechanisms under auto-bidding auctions, we continued our research in improving online marketplaces in the context of automation while taking different aspects into consideration, such as <a href=\"https:\/\/research.google\/pubs\/pub52089\/\">user experience<\/a> and <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3490486.3538333\">advertiser budgets<\/a>. Our findings suggest that properly incorporating <a href=\"https:\/\/research.google\/pubs\/pub51170\/\">ML advice<\/a> and <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3485447.3512062\">randomization techniques<\/a>, even in <a href=\"https:\/\/arxiv.org\/abs\/2207.03630\">non-truthful auctions<\/a>, can robustly improve the overall welfare at equilibria among auto-bidding algorithms.\n<\/p>\n<table align=\"center\" cellpadding=\"0\" cellspacing=\"0\" class=\"tr-caption-container\" style=\"margin-left: auto; margin-right: auto;\">\n<tbody>\n<tr>\n<td style=\"text-align: center;\"><a href=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEg6UmPqxmAEMkqO0OzCO66JdWXThOahWxxhDJ4PzbC6ganvNxYjs1VMsTKIE8J2SpE-TJu3qnJaGSe-wJDVZPjMU1WNyxCVdX6OURgGDNZkxVWbTfRI_dpneRtHuFYMp9KBRFsK22tsWLZieyzE4kv5Qp3XCnwbZBZLRZ1DLqsICFLCv9ScLLAaX8uv\/s1999\/image5.png\" style=\"margin-left: auto; margin-right: auto;\"><img loading=\"lazy\" decoding=\"async\" border=\"0\" data-original-height=\"1085\" data-original-width=\"1999\" height=\"347\" src=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEg6UmPqxmAEMkqO0OzCO66JdWXThOahWxxhDJ4PzbC6ganvNxYjs1VMsTKIE8J2SpE-TJu3qnJaGSe-wJDVZPjMU1WNyxCVdX6OURgGDNZkxVWbTfRI_dpneRtHuFYMp9KBRFsK22tsWLZieyzE4kv5Qp3XCnwbZBZLRZ1DLqsICFLCv9ScLLAaX8uv\/w640-h347\/image5.png\" width=\"640\"\/><\/a><\/td>\n<\/tr>\n<tr>\n<td class=\"tr-caption\" style=\"text-align: center;\">Structure of auto-bidding online ads system.<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p>\nBeyond auto-bidding systems, we also studied auction improvements in complex environments, e.g., settings where buyers are represented by <a href=\"https:\/\/drops.dagstuhl.de\/opus\/volltexte\/2022\/15597\/\">intermediaries<\/a>, and with <a href=\"https:\/\/arxiv.org\/abs\/2206.02948\">Rich Ads<\/a> where each ad can be shown in one of several possible variants. We summarize our work in this area in a recent <a href=\"https:\/\/www.sigecom.org\/exchanges\/volume_20\/2\/BHAWALKAR.pdf\">survey<\/a>. Beyond auctions, we also investigate the use of contracts in <a href=\"https:\/\/arxiv.org\/abs\/2211.05434\">multi-agent<\/a>\u00a0and <a href=\"https:\/\/arxiv.org\/abs\/2010.06742\">adversarial<\/a> settings.\n<\/p>\n<p>\nOnline stochastic optimization remains an important part of online advertising systems with application in optimal bidding and budget pacing. Building on our long-term research in online allocation, we recently blogged about <a href=\"https:\/\/ai.googleblog.com\/2022\/09\/robust-online-allocation-with-dual.html\">dual mirror descent<\/a>, a new <a href=\"https:\/\/pubsonline.informs.org\/doi\/abs\/10.1287\/opre.2021.2242\">algorithm<\/a> for online allocation problems that is simple, robust, and flexible. This state-of-the-art algorithm is robust against a wide range of adversarial and stochastic input distributions and can optimize important objectives beyond economic efficiency, such as <a href=\"https:\/\/proceedings.mlr.press\/v139\/balseiro21a.html\">fairness<\/a>. We also show that by tailoring dual mirror descent to the special structure of the increasingly popular <a href=\"https:\/\/arxiv.org\/abs\/2208.13713\">return-on-spend<\/a> constraints, we can optimize advertiser value. Dual mirror descent has a wide range of applications and has been used over time to help advertisers obtain more value through better algorithmic decision making.\n<\/p>\n<table align=\"center\" cellpadding=\"0\" cellspacing=\"0\" class=\"tr-caption-container\" style=\"margin-left: auto; margin-right: auto;\">\n<tbody>\n<tr>\n<td style=\"text-align: center;\"><a href=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEgg80Fhc3fVmKiS1DwLkOjogBP4aGQJAirqbqmBSCIXQrYe82DaJPKer_5RWn3u5fgZ_IML1uzstZ3fbYR5LbBWFARVaS1YfEz5bwFGDDRIbeuB8C6FWJh7uVNNBXlsDrjFgS8DRIR2cUb2Uqc6OTUSIAaPb4wyzvxg7XGH1deYdVBgRdLluIc51ILC\/s1273\/image6.png\" style=\"margin-left: auto; margin-right: auto;\"><img loading=\"lazy\" decoding=\"async\" border=\"0\" data-original-height=\"1139\" data-original-width=\"1273\" height=\"572\" src=\"https:\/\/blogger.googleusercontent.com\/img\/b\/R29vZ2xl\/AVvXsEgg80Fhc3fVmKiS1DwLkOjogBP4aGQJAirqbqmBSCIXQrYe82DaJPKer_5RWn3u5fgZ_IML1uzstZ3fbYR5LbBWFARVaS1YfEz5bwFGDDRIbeuB8C6FWJh7uVNNBXlsDrjFgS8DRIR2cUb2Uqc6OTUSIAaPb4wyzvxg7XGH1deYdVBgRdLluIc51ILC\/w640-h572\/image6.png\" width=\"640\"\/><\/a><\/td>\n<\/tr>\n<tr>\n<td class=\"tr-caption\" style=\"text-align: center;\">An overview of the dual mirror descent algorithm.<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<p>\nFurthermore, following our <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3470442\">recent work<\/a> at the interplay of ML, mechanism design and markets, we investigated <a href=\"https:\/\/proceedings.mlr.press\/v162\/duan22a.html\">transformers for asymmetric auction design<\/a>, designed utility-maximizing strategies for <a href=\"https:\/\/proceedings.mlr.press\/v178\/mansour22a.html\">no-regret learning buyers<\/a>, and developed new <a href=\"https:\/\/arxiv.org\/abs\/2109.03173\">learning algorithms to bid<\/a> or to <a href=\"https:\/\/arxiv.org\/abs\/2107.07725\">price<\/a> in auctions.\n<\/p>\n<p>\nA critical component of any sophisticated online service is the ability to experimentally measure the response of users and other players to new interventions. A major challenge of estimating these causal effects accurately is handling complex interactions \u2014 or interference \u2014 between the control and treatment units of these experiments. We combined our graph clustering and causal inference expertise to expand the results of <a href=\"https:\/\/proceedings.neurips.cc\/paper\/2019\/hash\/bc047286b224b7bfa73d4cb02de1238d-Abstract.html\">our previous work<\/a> in this area, with <a href=\"https:\/\/dl.acm.org\/doi\/abs\/10.1145\/3490486.3538269\">improved results under a flexible response model<\/a> and <a href=\"https:\/\/openreview.net\/pdf?id=hqtSdpAK39W\">a new experimental design<\/a> that is more effective at reducing these interactions when treatment assignments and metric measurements occur on the same side of a bipartite platform. We also showed how <a href=\"https:\/\/proceedings.neurips.cc\/paper\/2021\/hash\/48d23e87eb98cc2227b5a8c33fa00680-Abstract.html\">synthetic control and optimization techniques<\/a> can be combined to design more powerful experiments, especially in small data regimes.\n<\/p>\n<p align=\"right\"><a href=\"#top\">Top<\/a><\/p>\n<p><a name=\"Theme4\"> <\/a><\/p>\n<h2>Algorithmic foundations and theory<\/h2>\n<p>\nFinally, we continued our fundamental algorithmic research by tackling long-standing open problems. A surprisingly concise <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3519935.3520054\">paper<\/a> affirmatively resolved a four-decade old <a href=\"https:\/\/www.sciencedirect.com\/science\/article\/abs\/pii\/0022053183900480\">open question<\/a> on whether there is a mechanism that guarantees a constant fraction of the <a href=\"https:\/\/en.wikipedia.org\/wiki\/Myerson%E2%80%93Satterthwaite_theorem\">gains-from-trade attainable<\/a> whenever buyer&#8217;s value weakly exceeds seller&#8217;s cost. Another recent <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3519935.3520011\">paper<\/a> obtained the state-of-the-art approximation for the classic and highly-studied <em><a href=\"https:\/\/en.wikipedia.org\/wiki\/K-means_clustering\">k-means problem<\/a><\/em>. We also improved the best approximation for <a href=\"https:\/\/arxiv.org\/abs\/2207.10889\">correlation clustering<\/a> breaking the barrier approximation factor of 2. Finally, our work on dynamic data structures to solve <a href=\"https:\/\/dl.acm.org\/doi\/10.1145\/3406325.3451108\">min-cost<\/a> and other <a href=\"https:\/\/ieeexplore.ieee.org\/abstract\/document\/9719724\">network flow<\/a> problems has contributed to a breakthrough line of work in adapting continuous optimization techniques to solve classic discrete optimization problems.\n<\/p>\n<p align=\"right\"><a href=\"#top\">Top<\/a><\/p>\n<h2>Concluding thoughts<\/h2>\n<p>\nDesigning effective algorithms and mechanisms is a critical component of many Google systems that need to handle tera-scale data robustly with critical privacy and safety considerations. Our approach is to develop algorithms with solid theoretical foundations that can be deployed effectively in our product systems. In addition, we are bringing many of these advances to the broader community by open-sourcing some of our most novel developments and by publishing the advanced algorithms behind them. In this post, we covered a subset of algorithmic advances in privacy, market algorithms, scalable algorithms, graph-based learning, and optimization. As we move toward an AI-first Google with further automation, developing robust, scalable, and privacy-safe ML algorithms remains a high priority. We are excited about developing new algorithms and deploying them more broadly.\n<\/p>\n<p><\/p>\n<h2>Acknowledgements<\/h2>\n<p>\n<em>This post summarizes research from a large number of teams and benefited from input from several researchers including Gagan Aggarwal, Amr Ahmed, David Applegate, Santiago Balseiro, Vincent Cohen-addad, Yuan Deng, Alessandro Epasto, Matthew Fahrbach, Badih Ghazi, Sreenivas Gollapudi, Rajesh Jayaram, Ravi Kumar, Sanjiv Kumar, Silvio Lattanzi, Kuba Lacki, Brendan McMahan, Aranyak Mehta, Bryan Perozzi, Daniel Ramage, Ananda Theertha Suresh,  Andreas Terzis, Sergei Vassilvitskii, Di Wang, and Song Zuo. Special thanks to Ravi Kumar for his contributions to this post.<\/em>\n<\/p>\n<p><a name=\"ToC\"> <\/a><br \/>\n<\/p>\n<h2>Google Research, 2022 &amp; beyond<\/h2>\n<p>\nThis was the fifth blog post in the \u201cGoogle Research, 2022 &amp; Beyond\u201d series. Other posts in this series are listed in the table below:\n<\/p>\n<p><\/p>\n<div style=\"margin-left: 10%; margin-right: 10%; text-align: center;\">\n<table align=\"center\" cellpadding=\"0\" cellspacing=\"0\" class=\"tr-caption-container\" style=\"margin-left: auto; margin-right: auto;\">\n<tbody>\n<tr>\n<td class=\"tr-caption\" style=\"text-align: center;\">* Articles will be linked as they are released.<\/td>\n<\/tr>\n<\/tbody>\n<\/table>\n<\/div>\n<\/div>\n<p>[ad_2]<br \/>\n<br \/><a href=\"http:\/\/ai.googleblog.com\/2023\/02\/google-research-2022-beyond-algorithmic.html\">Source link <\/a><\/p>\n","protected":false},"excerpt":{"rendered":"<p>[ad_1] Posted by Vahab Mirrokni, VP and Google Fellow, Google Research (This is Part 5 in our series<\/p>\n","protected":false},"author":2,"featured_media":345,"comment_status":"closed","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[20],"tags":[],"class_list":["post-344","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-google-ai"],"_links":{"self":[{"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/posts\/344","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/comments?post=344"}],"version-history":[{"count":1,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/posts\/344\/revisions"}],"predecessor-version":[{"id":2900,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/posts\/344\/revisions\/2900"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/media\/345"}],"wp:attachment":[{"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/media?parent=344"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/categories?post=344"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/todaysainews.com\/index.php\/wp-json\/wp\/v2\/tags?post=344"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}