<?xml version="1.0" encoding="utf-8" standalone="yes"?><rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom"><channel><title>PyTorch on The Coders Blog</title><link>https://thecodersblog.com/tag/pytorch/</link><description>Recent content in PyTorch on The Coders Blog</description><generator>Hugo</generator><language>en-us</language><lastBuildDate>Wed, 06 May 2026 22:22:11 +0000</lastBuildDate><atom:link href="https://thecodersblog.com/tag/pytorch/index.xml" rel="self" type="application/rss+xml"/><item><title>Google Colossus on PyTorch via GCSF: Speeding Up AI Training</title><link>https://thecodersblog.com/speeding-up-ai-with-google-colossus-on-pytorch-via-gcsf-2026/</link><pubDate>Wed, 06 May 2026 22:22:11 +0000</pubDate><guid>https://thecodersblog.com/speeding-up-ai-with-google-colossus-on-pytorch-via-gcsf-2026/</guid><description>&lt;p&gt;Your GPUs are starving. They&amp;rsquo;re idling, waiting for data or, worse, for model checkpoints to be saved. For anyone wrestling with terabyte and petabyte-scale datasets in AI/ML, this GPU starvation is a familiar, frustrating bottleneck, often exacerbated by the inherent limitations of standard REST-based object storage.&lt;/p&gt;
&lt;h3 id="the-core-problem-storage-bottlenecks-in-large-scale-ai"&gt;The Core Problem: Storage Bottlenecks in Large-Scale AI&lt;/h3&gt;
&lt;p&gt;The traditional approach of accessing massive datasets and saving frequent checkpoints via standard cloud object storage APIs often becomes a choke point. For complex models and extensive datasets, the latency and throughput limitations of these APIs simply cannot keep pace with the demands of high-performance computing clusters. This leads to inefficient resource utilization, longer training times, and increased costs.&lt;/p&gt;</description></item><item><title>Critical Alert: Shai-Hulud Malware Discovered in PyTorch Lightning Dependencies</title><link>https://thecodersblog.com/shai-hulud-malware-in-pytorch-lightning-2026/</link><pubDate>Fri, 01 May 2026 07:48:47 +0000</pubDate><guid>https://thecodersblog.com/shai-hulud-malware-in-pytorch-lightning-2026/</guid><description>&lt;p&gt;Stop what you&amp;rsquo;re doing. A critical alert has been raised around the &amp;lsquo;Shai-Hulud Malware&amp;rsquo;, a sophisticated supply chain attack targeting the &lt;code&gt;lightning&lt;/code&gt; PyPI package, specifically versions &lt;code&gt;2.6.2&lt;/code&gt; and &lt;code&gt;2.6.3&lt;/code&gt;. This isn&amp;rsquo;t theoretical; your enterprise ML pipelines could be replicating a credential-stealing worm with every &lt;code&gt;pip install&lt;/code&gt;. This incident is a harsh lesson: the era of implicit trust in open-source ML libraries is irrevocably over for enterprise environments.&lt;/p&gt;
&lt;p&gt;The &amp;ldquo;Shai-Hulud Malware&amp;rdquo; isn&amp;rsquo;t merely a vulnerability; it&amp;rsquo;s a confirmed and active threat that has explicitly crossed from npm to compromise the PyTorch Lightning ecosystem. This attack directly hit a widely used deep-learning framework, demonstrating a sophisticated adversary&amp;rsquo;s ability to adapt and target critical infrastructure. Your next &lt;code&gt;pip install&lt;/code&gt; could be an open door.&lt;/p&gt;</description></item></channel></rss>