{"id":14436,"date":"2026-05-04T15:32:47","date_gmt":"2026-05-04T15:32:47","guid":{"rendered":"https:\/\/techtrendfeed.com\/?p=14436"},"modified":"2026-05-04T15:32:47","modified_gmt":"2026-05-04T15:32:47","slug":"rushing-up-ai-bringing-google-colossus-to-pytorch-by-way-of-gcsfs-and-fast-bucket","status":"publish","type":"post","link":"https:\/\/techtrendfeed.com\/?p=14436","title":{"rendered":"Rushing Up AI: Bringing Google Colossus to PyTorch by way of GCSFS and Fast Bucket"},"content":{"rendered":"<p> <br \/>\n<\/p>\n<div>\n<p data-block-key=\"60gb6\">As we speak, we&#8217;re asserting a serious efficiency enhance for AI\/ML workloads utilizing the PyTorch ecosystem on Google Cloud. By integrating Fast Storage, powered by <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/cloud.google.com\/blog\/products\/storage-data-transfer\/a-peek-behind-colossus-googles-file-system?e=48754805\">Google\u2019s Colossus<\/a> storage structure, instantly with<b> PyTorch<\/b> by way of the industry-standard <code>fsspec<\/code> interface, we&#8217;re enabling researchers and builders to maintain their GPUs busier than ever earlier than.<\/p>\n<h2 data-block-key=\"ji2j0\" id=\"the-challenge:-keeping-gpus-fed\"><b>The problem: Holding GPUs fed<\/b><\/h2>\n<p data-block-key=\"ep32l\">As mannequin sizes develop, information loading and checkpointing typically turn out to be the first bottlenecks in coaching. Knowledge preparation actions to coach fashions contain fetching and processing terabytes and petabytes of information from distant storage mechanisms like object storage. Customary REST-based storage entry can wrestle to satisfy the intense throughput and low-latency necessities of recent distributed coaching, losing helpful GPU sources.<\/p>\n<h2 data-block-key=\"gs1lc\" id=\"rapid-bucket:-rapid-storage-via-bi-di-grpc\"><b>Fast Bucket: Fast Storage by way of bi-di gRPC<\/b><\/h2>\n<p data-block-key=\"37spj\">Our new <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/docs.cloud.google.com\/storage\/docs\/rapid\/rapid-bucket\"><b>Fast Bucket<\/b><\/a> resolution gives high-performance object storage in devoted zonal buckets. By bypassing legacy REST APIs and using persistent gRPC bidirectional streams, we\u2019ve introduced the ability of Colossus, filesystem stateful protocols that energy YouTube and Google Search, on to the PyTorch ecosystem.<\/p>\n<h3 data-block-key=\"zhx09\" id=\"key-performance-metrics-of-rapid-storage\"><b>Key efficiency metrics of Fast Storage<\/b><\/h3>\n<ul>\n<li data-block-key=\"ftv4b\"><b>Excessive Throughput:<\/b> <b>15+ TiB\/s<\/b> mixture throughput.<\/li>\n<li data-block-key=\"9nivo\"><b>Extremely-Low Latency:<\/b> &lt;1ms for random reads and append writes.<\/li>\n<li data-block-key=\"5m8df\"><b>Excessive QPS:<\/b> Fast Bucket gives 20M+ QPS.<\/li>\n<\/ul>\n<h2 data-block-key=\"f6vph\" id=\"fsspec-pytorch's-pythonic-file-interface\"><b>Fsspec &#8211; PyTorch\u2019s Pythonic file interface<\/b><\/h2>\n<p data-block-key=\"fuvd6\"><code>fsspec<\/code> is the pervasive Pythonic interface for file methods within the PyTorch ecosystem. It&#8217;s already used for:<\/p>\n<ul>\n<li data-block-key=\"5eavu\"><b>Knowledge preparation:<\/b> Dask, Pandas, Hugging Face Datasets, Ray Knowledge<\/li>\n<li data-block-key=\"aa8ha\"><b>Checkpoints:<\/b> PyTorch Lightning, Torch.dist, Weights &amp; Biases<\/li>\n<li data-block-key=\"fv134\"><b>Inference:<\/b> vLLM<\/li>\n<\/ul>\n<\/div>\n<div>\n<p data-block-key=\"3f9ty\">There are numerous backend implementations of fsspec for a lot of totally different storage methods, which might all be built-in beneath a single layer, eliminating the necessity to write particular code for every backend. By integrating Fast Storage with <code>gcsfs<\/code> (the Google Cloud Storage implementation of fsspec), builders can leverage velocity positive factors offered by Fast with a easy <code>fsspec.open()<\/code> name \u2014 no advanced code rewrites required.<\/p>\n<h2 data-block-key=\"97f5g\" id=\"under-the-hood:-leveraging-colossus\">Beneath the hood: Leveraging Colossus<\/h2>\n<p data-block-key=\"lfp4\">To realize a efficiency enhance with Fast Buckets, we optimized all the information path:<\/p>\n<ol>\n<li data-block-key=\"6mdmu\"><b>Stateful grpc-based streaming:<\/b> gRPC bi-directional streaming retains the connection alive, minimizing per-operation overhead like connection setup, auth, metadata and many others., and enabling environment friendly, stateful information alternate for a number of reads or appends inside a single object.<\/li>\n<li data-block-key=\"bkpk3\"><b>Direct path:<\/b> Google Cloud Storage(GCS) Fast Bucket makes use of <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/docs.cloud.google.com\/storage\/docs\/direct-connectivity\">direct connectivity<\/a> for its gRPC bi-directional streaming APIs (BidiReadObject, BidiWriteObject) to attain most efficiency by connecting purchasers on to underlying Colossus information. Non-Fast site visitors to GCS would sometimes have extra community hops than direct paths, making learn\/write latencies over Fast considerably decrease. For extra particulars, see <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/cloud.google.com\/blog\/products\/storage-data-transfer\/how-the-colossus-stateful-protocol-benefits-rapid-storage?e=48754805\">Fast storage inner working.<\/a><\/li>\n<li data-block-key=\"ehq5k\"><b>Zonal co-location:<\/b> By inserting storage in the identical zone as your compute (e.g., <code>us-central1-a<\/code>), we remove cross-zone latency. Previous to Fast buckets, information in a regional bucket and compute(accelerators) could be in numerous zones and entry the info induced latency.<\/li>\n<li data-block-key=\"d9tip\"><b>No-Op Person Migration:<\/b> Preserved the present <code>fsspec<\/code> API whereas completely upgrading inner site visitors from HTTP to BiDi-gRPC for Fast buckets. By including bucket-type auto-detection to gcsfs, PyTorch and different <code>fsspec<\/code> purchasers transparently make the most of Fast with zero handbook configuration.<\/li>\n<\/ol>\n<h2 data-block-key=\"t8fdn\" id=\"results\">Outcomes<\/h2>\n<p data-block-key=\"ai3u9\">A dataset of 134M rows totaling round 451GB was loaded onto 16 GKE nodes, every containing eight A4 GPUs. Coaching was performed in 100 steps, with a checkpoint after each 25 steps utilizing PyTorch Lightning. We benchmarked the efficiency of complete coaching time, together with the info load instances, and we noticed a <b>efficiency achieve of 23% utilizing Fast Bucket in contrast with Customary regional bucket.<\/b><\/p>\n<\/div>\n<div>\n<p data-block-key=\"3f9ty\">Microbenchmarking \u2014 that&#8217;s, measuring the efficiency of a constructing block like I\/O or useful resource utilization \u2014 confirms these positive factors. Throughput improved by 4.8x for reads (each sequential and random) and a couple of.8x for writes. These checks used 16MB IO sizes throughout 48 processes. You&#8217;ll find extra particulars at <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/github.com\/fsspec\/gcsfs\/blob\/main\/docs\/source\/rapid_storage_support.rst#performance-benchmarks\">GCSFS-performance-benchmarks<\/a>.<\/p>\n<h2 data-block-key=\"nkjuq\" id=\"get-started\">Get began<\/h2>\n<p data-block-key=\"a3tju\">Getting began with <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/github.com\/fsspec\/gcsfs\">GCSFS<\/a> on Fast Bucket is simple. Your current code and scripts stay the identical. You simply want to alter the bucket to a <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/docs.cloud.google.com\/storage\/docs\/rapid\/rapid-bucket\">Fast Bucket<\/a> to benefit from the efficiency enhance.<\/p>\n<p data-block-key=\"8sca4\"><b>To put in:<\/b><\/p>\n<p data-block-key=\"3to40\">Fast Bucket integration is offered from model <a rel=\"nofollow\" target=\"_blank\" href=\"https:\/\/github.com\/fsspec\/gcsfs\/releases\/tag\/2026.3.0\">2026.3.0.<\/a><\/p>\n<\/div>\n<div>\n<p data-block-key=\"6cb9x\"><b>Code pattern to learn\/write from GCS Fast:<\/b><\/p>\n<\/div>\n<div>\n<pre><code class=\"language-python\">import gcsfs&#13;\n&#13;\n# Initialize the filesystem&#13;\nfs = gcsfs.GCSFileSystem()&#13;\n&#13;\n# Writing to a Fast bucket&#13;\nwith fs.open('my-zonal-rapid-bucket\/information\/checkpoint.pt', 'wb') as f:&#13;\n   f.write(b\"mannequin information...\")&#13;\n&#13;\n# Appending to an current object (Native Fast function)&#13;\nwith fs.open('my-zonal-rapid-bucket\/information\/checkpoint.pt', 'ab') as f:&#13;\n   f.write(b\"appended information...\")<\/code><\/pre>\n<p>\n        Python\n    <\/p>\n<\/div>\n\n","protected":false},"excerpt":{"rendered":"<p>As we speak, we&#8217;re asserting a serious efficiency enhance for AI\/ML workloads utilizing the PyTorch ecosystem on Google Cloud. By integrating Fast Storage, powered by Google\u2019s Colossus storage structure, instantly with PyTorch by way of the industry-standard fsspec interface, we&#8217;re enabling researchers and builders to maintain their GPUs busier than ever earlier than. The problem: [&hellip;]<\/p>\n","protected":false},"author":2,"featured_media":14438,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[56],"tags":[1458,8939,8937,8938,81,6749,4931,8936],"class_list":["post-14436","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-software","tag-bringing","tag-bucket","tag-colossus","tag-gcsfs","tag-google","tag-pytorch","tag-rapid","tag-speeding"],"_links":{"self":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/14436","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=14436"}],"version-history":[{"count":1,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/14436\/revisions"}],"predecessor-version":[{"id":14437,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/posts\/14436\/revisions\/14437"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=\/wp\/v2\/media\/14438"}],"wp:attachment":[{"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=14436"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=14436"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/techtrendfeed.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=14436"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}<!-- This website is optimized by Airlift. Learn more: https://airlift.net. Template:. Learn more: https://airlift.net. Template: 69d9690a190636c2e0989534. Config Timestamp: 2026-04-10 21:18:02 UTC, Cached Timestamp: 2026-05-04 17:21:08 UTC -->