Well, normally Perf/W is the only important metric, simply because even the slowest performing options are still usually not too terribly slow, and with modern cloud and virtualized focused systems, the performance of any given request/job is not going to be measurably impacted by the hardware running it, the hardware just defines how many can be run on a given chunk of hardware and at what cost, so the service provider will optimize for overall (vs single request) Perf/W to optimize their costs.
But when some NN jobs might take a month to compute on existing clusters, then there's probably a willingness to give up Perf/W for faster execution, as long as the resulting power needs aren't too insane.
If they reduced a NN job from 30 days to 1 day and it took the same amount of total power (just delivered at 30X the rate), that would be an even trade and a no-brainer, as long as their local substation doesn't trip in the process.
If it took 10 or 100 times the total power, that might or might not be acceptable, even if it resulted in significant speedup. Somewhere in the middle, if your jobs are taking too long to execute, you're going to be happy to pay more money (both in hardware costs and infrastructure costs, which includes energy and cooling all the heat that energy usage generates) for a "worse" Perf/W just to get faster results, but obviously an infinite cost is not acceptable for any level of performance, so there are limits.
If Tesla is managing to be both faster AND maintain equivalent or better Perf/W, then of course, Dojo is a win on all fronts, with the exception of possibly needing to get their own private substation built...