<?xml version="1.0" encoding="UTF-8"?><rss version="2.0"
	xmlns:content="http://purl.org/rss/1.0/modules/content/"
	xmlns:dc="http://purl.org/dc/elements/1.1/"
	xmlns:atom="http://www.w3.org/2005/Atom"
	xmlns:sy="http://purl.org/rss/1.0/modules/syndication/"
	
	xmlns:georss="http://www.georss.org/georss"
	xmlns:geo="http://www.w3.org/2003/01/geo/wgs84_pos#"
	
	>
<channel>
	<title>
	Comments on: Which GPU(s) to Get for Deep Learning: My Experience and Advice for Using GPUs in Deep Learning	</title>
	<atom:link href="https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/feed/" rel="self" type="application/rss+xml" />
	<link>https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/</link>
	<description>Making deep learning accessible.</description>
	<lastBuildDate>Wed, 10 Dec 2025 15:06:20 +0000</lastBuildDate>
	<sy:updatePeriod>
	hourly	</sy:updatePeriod>
	<sy:updateFrequency>
	1	</sy:updateFrequency>
	<generator>https://wordpress.org/?v=6.0.11</generator>
	<item>
		<title>
		By: Zoran		</title>
		<link>https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-204944</link>

		<dc:creator><![CDATA[Zoran]]></dc:creator>
		<pubDate>Mon, 14 Apr 2025 18:12:28 +0000</pubDate>
		<guid isPermaLink="false">http://timdettmers.wordpress.com/?p=4#comment-204944</guid>

					<description><![CDATA[Threadripper 3945wx works almost 2x slower than intel 12400 as per my test.]]></description>
			<content:encoded><![CDATA[<p>Threadripper 3945wx works almost 2x slower than intel 12400 as per my test.</p>
]]></content:encoded>
		
			</item>
		<item>
		<title>
		By: Zoran		</title>
		<link>https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-132150</link>

		<dc:creator><![CDATA[Zoran]]></dc:creator>
		<pubDate>Fri, 16 Feb 2024 06:01:42 +0000</pubDate>
		<guid isPermaLink="false">http://timdettmers.wordpress.com/?p=4#comment-132150</guid>

					<description><![CDATA[Where can i find the code for the 8-bit inference?]]></description>
			<content:encoded><![CDATA[<p>Where can i find the code for the 8-bit inference?</p>
]]></content:encoded>
		
			</item>
		<item>
		<title>
		By: Andrea de Luca		</title>
		<link>https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-127183</link>

		<dc:creator><![CDATA[Andrea de Luca]]></dc:creator>
		<pubDate>Mon, 18 Dec 2023 18:24:43 +0000</pubDate>
		<guid isPermaLink="false">http://timdettmers.wordpress.com/?p=4#comment-127183</guid>

					<description><![CDATA[Hi Tim. I think there is a bit of confusion in the article regarding the RTX A6000. You wrote:
&quot;The best GPUs for academic and startup servers seem to be A6000 Ada GPUs (not to be confused with A6000 Turing). &quot;

The RTX A6000 is a 48gb Ampere card, not Turing. Its performance (in any domain) is slightly better than a 3090, while in the charts it performs equal to the old Turing workstation cards (which is frankly impossible). Other than that, the Ada 48gb workstation card is officially called &quot;RTX 6000 Ada&quot; (without the &quot;A&quot;). Thanks.]]></description>
			<content:encoded><![CDATA[<p>Hi Tim. I think there is a bit of confusion in the article regarding the RTX A6000. You wrote:<br />
&#8220;The best GPUs for academic and startup servers seem to be A6000 Ada GPUs (not to be confused with A6000 Turing). &#8221;</p>
<p>The RTX A6000 is a 48gb Ampere card, not Turing. Its performance (in any domain) is slightly better than a 3090, while in the charts it performs equal to the old Turing workstation cards (which is frankly impossible). Other than that, the Ada 48gb workstation card is officially called &#8220;RTX 6000 Ada&#8221; (without the &#8220;A&#8221;). Thanks.</p>
]]></content:encoded>
		
			</item>
		<item>
		<title>
		By: Andrea de Luca		</title>
		<link>https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-127182</link>

		<dc:creator><![CDATA[Andrea de Luca]]></dc:creator>
		<pubDate>Mon, 18 Dec 2023 18:24:02 +0000</pubDate>
		<guid isPermaLink="false">http://timdettmers.wordpress.com/?p=4#comment-127182</guid>

					<description><![CDATA[Hi Tim. I think there is a bit of confusion in the article regarding the RTX A6000. You wrote:
&#060;&#062;

The RTX A6000 is a 48gb Ampere card, not Turing. Its performance (in any domain) is slightly better than a 3090, while in the charts it performs equal to the old Turing workstation cards (which is frankly impossible). Other than that, the Ada 48gb workstation card is officially called &quot;RTX 6000 Ada&quot; (without the &quot;A&quot;). Thanks.]]></description>
			<content:encoded><![CDATA[<p>Hi Tim. I think there is a bit of confusion in the article regarding the RTX A6000. You wrote:<br />
&lt;&gt;</p>
<p>The RTX A6000 is a 48gb Ampere card, not Turing. Its performance (in any domain) is slightly better than a 3090, while in the charts it performs equal to the old Turing workstation cards (which is frankly impossible). Other than that, the Ada 48gb workstation card is officially called &#8220;RTX 6000 Ada&#8221; (without the &#8220;A&#8221;). Thanks.</p>
]]></content:encoded>
		
			</item>
		<item>
		<title>
		By: Zoran		</title>
		<link>https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-118648</link>

		<dc:creator><![CDATA[Zoran]]></dc:creator>
		<pubDate>Sun, 30 Apr 2023 15:21:59 +0000</pubDate>
		<guid isPermaLink="false">http://timdettmers.wordpress.com/?p=4#comment-118648</guid>

					<description><![CDATA[Hello, 

Have you had any chance to test high end CPU only interface, for example on Intel 13900K CPU? It seemed like could be pretty fast and a dedicated server with it can be around 100 EUR per month. Much cheaper compared to GPU dedicated server or cloud.]]></description>
			<content:encoded><![CDATA[<p>Hello, </p>
<p>Have you had any chance to test high end CPU only interface, for example on Intel 13900K CPU? It seemed like could be pretty fast and a dedicated server with it can be around 100 EUR per month. Much cheaper compared to GPU dedicated server or cloud.</p>
]]></content:encoded>
		
			</item>
		<item>
		<title>
		By: David Laxer		</title>
		<link>https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-115887</link>

		<dc:creator><![CDATA[David Laxer]]></dc:creator>
		<pubDate>Tue, 17 Jan 2023 20:24:30 +0000</pubDate>
		<guid isPermaLink="false">http://timdettmers.wordpress.com/?p=4#comment-115887</guid>

					<description><![CDATA[Hi Tim,

Thanks for your posts.

Do you have any comments on Apple&#039;s M1/M2 chips for Deep Learning research?
Apple&#039; Metal Performance Shader API only supports float32 as well as 32bit complex numbers.  Do you see this as a
&#039;show stopper&#039; for Deep Learning Research for M1/M2 chips.
The M1/M2 processors do use considerably less power then NVIDIA GPUs
does this significantly change the trade-off calculus?

Thanks in advance.]]></description>
			<content:encoded><![CDATA[<p>Hi Tim,</p>
<p>Thanks for your posts.</p>
<p>Do you have any comments on Apple&#8217;s M1/M2 chips for Deep Learning research?<br />
Apple&#8217; Metal Performance Shader API only supports float32 as well as 32bit complex numbers.  Do you see this as a<br />
&#8216;show stopper&#8217; for Deep Learning Research for M1/M2 chips.<br />
The M1/M2 processors do use considerably less power then NVIDIA GPUs<br />
does this significantly change the trade-off calculus?</p>
<p>Thanks in advance.</p>
]]></content:encoded>
		
			</item>
		<item>
		<title>
		By: Tim Dettmers		</title>
		<link>https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-104394</link>

		<dc:creator><![CDATA[Tim Dettmers]]></dc:creator>
		<pubDate>Mon, 14 Mar 2022 17:09:07 +0000</pubDate>
		<guid isPermaLink="false">http://timdettmers.wordpress.com/?p=4#comment-104394</guid>

					<description><![CDATA[In reply to &lt;a href=&quot;https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-102444&quot;&gt;Paulo Ricardo&lt;/a&gt;.

Yes, I would love to see a Portuguese translation! Go for it. The only thing that I would ask you to do if you translate it section by section is to include the source somewhere in the introduction to the translation.]]></description>
			<content:encoded><![CDATA[<p>In reply to <a href="https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-102444">Paulo Ricardo</a>.</p>
<p>Yes, I would love to see a Portuguese translation! Go for it. The only thing that I would ask you to do if you translate it section by section is to include the source somewhere in the introduction to the translation.</p>
]]></content:encoded>
		
			</item>
		<item>
		<title>
		By: Paulo Ricardo		</title>
		<link>https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-102444</link>

		<dc:creator><![CDATA[Paulo Ricardo]]></dc:creator>
		<pubDate>Thu, 10 Feb 2022 14:04:39 +0000</pubDate>
		<guid isPermaLink="false">http://timdettmers.wordpress.com/?p=4#comment-102444</guid>

					<description><![CDATA[What a wonderful text.

Can I translate to Portuguese?]]></description>
			<content:encoded><![CDATA[<p>What a wonderful text.</p>
<p>Can I translate to Portuguese?</p>
]]></content:encoded>
		
			</item>
		<item>
		<title>
		By: Simon Demeule		</title>
		<link>https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-98251</link>

		<dc:creator><![CDATA[Simon Demeule]]></dc:creator>
		<pubDate>Sun, 07 Nov 2021 06:34:57 +0000</pubDate>
		<guid isPermaLink="false">http://timdettmers.wordpress.com/?p=4#comment-98251</guid>

					<description><![CDATA[In reply to &lt;a href=&quot;https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-97547&quot;&gt;Tim Dettmers&lt;/a&gt;.

Hi Tim!

Thank you for your insight. After some more research, I have found that in synthetic benchmarks the 5950X actually outperforms the 3975WX in both single and multi core performance (which is very surprising given it has half the core count) — it seems the architecture version (Gen 3 vs Gen 2) plays a large role here. It really doesn&#039;t make sense to choose Threadripper Pro over Ryzen for my use case, or at least right now. Gen 3 Threadripper Pro is likely coming out soon, as some benchmark scores have been leaked (and they are definitely scoring higher than Ryzen there). I will likely order the GPUs soon, and maybe wait a bit until the next generation Threadripper Pro CPUs are revealed.]]></description>
			<content:encoded><![CDATA[<p>In reply to <a href="https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-97547">Tim Dettmers</a>.</p>
<p>Hi Tim!</p>
<p>Thank you for your insight. After some more research, I have found that in synthetic benchmarks the 5950X actually outperforms the 3975WX in both single and multi core performance (which is very surprising given it has half the core count) — it seems the architecture version (Gen 3 vs Gen 2) plays a large role here. It really doesn&#8217;t make sense to choose Threadripper Pro over Ryzen for my use case, or at least right now. Gen 3 Threadripper Pro is likely coming out soon, as some benchmark scores have been leaked (and they are definitely scoring higher than Ryzen there). I will likely order the GPUs soon, and maybe wait a bit until the next generation Threadripper Pro CPUs are revealed.</p>
]]></content:encoded>
		
			</item>
		<item>
		<title>
		By: Tim Dettmers		</title>
		<link>https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-97547</link>

		<dc:creator><![CDATA[Tim Dettmers]]></dc:creator>
		<pubDate>Sun, 24 Oct 2021 18:51:51 +0000</pubDate>
		<guid isPermaLink="false">http://timdettmers.wordpress.com/?p=4#comment-97547</guid>

					<description><![CDATA[In reply to &lt;a href=&quot;https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-95663&quot;&gt;Simon Demeule&lt;/a&gt;.

Hi Simon! I think the choice of A5000 GPUs is great in your use-case. I think both CPU options can have their advantages. For inference, a CPU can sometimes be critical to achieving good latency, but that can depend on many factors, mainly how much pre and post-processing needs to be done on CPU. Depending on this, it could be that a good CPU has large benefits or not. On the other hand, such performance will not be very noticeable for general use, but might be critical for user-facing applications. So if you aim to build projects that are user-facing and require low latency it might be worth it to go with the 3975WX. However, since CPUs have quirks and many threads are difficult to use, in many cases you will get the same performance for the much cheaper 5950X. As such, it might also be a waste of money. I think I would probably go with the 5950X, and if the user needs to wait for an additional 50ms then the user just needs to wait for that!]]></description>
			<content:encoded><![CDATA[<p>In reply to <a href="https://timdettmers.com/2023/01/30/which-gpu-for-deep-learning/comment-page-2/#comment-95663">Simon Demeule</a>.</p>
<p>Hi Simon! I think the choice of A5000 GPUs is great in your use-case. I think both CPU options can have their advantages. For inference, a CPU can sometimes be critical to achieving good latency, but that can depend on many factors, mainly how much pre and post-processing needs to be done on CPU. Depending on this, it could be that a good CPU has large benefits or not. On the other hand, such performance will not be very noticeable for general use, but might be critical for user-facing applications. So if you aim to build projects that are user-facing and require low latency it might be worth it to go with the 3975WX. However, since CPUs have quirks and many threads are difficult to use, in many cases you will get the same performance for the much cheaper 5950X. As such, it might also be a waste of money. I think I would probably go with the 5950X, and if the user needs to wait for an additional 50ms then the user just needs to wait for that!</p>
]]></content:encoded>
		
			</item>
	</channel>
</rss>
