Here comes the push.

@devbo@lemmy.world
link
fedilink
English
201Y

why do CEOs never say “lets take our time to avoid making mistakes and insure quality”?

@aelwero@lemmy.world
link
fedilink
English
181Y

That’s like saying we should all drive faster to help identify shortcomings in traffic signals or vehicle safety features…

Sounds like a flimsy ass false pretense to chase profits. Just my opinion, mind, but that doesn’t sound like something someone would posit at face value.

“Quick…we need to get everyone to buy our GPU-based bullshit before people figure out FPGA is wildly better”

Dr. Dabbles
link
fedilink
English
-11Y

Dedicated ASIC is where all the hotness lies. Flexibility of FPGA doesn’t seem to overcome its overhead for most users. Not sure if it will change when custom ASIC becomes too expensive again, and all the magic money furnaces run out of bills to burn.

ASIC are single purpose at the benefit of potential power efficiency improvements. Not at all useful something like running neutral networks, especially not when they are being retrained and updated.

FPGAs are fully (re)programmable. There’s a reason why datacenters don’t lease ASIC instances.

Dr. Dabbles
link
fedilink
English
-11Y

Not at all useful something like running neutral networks

Um. lol What? You may want to do your research here, because you’re so far off base I don’t think you’re even playing the right game.

There’s a reason why datacenters don’t lease ASIC instances.

Ok, so you should just go ahead and tell all the ASIC companies then.

https://www.allaboutcircuits.com/news/intel-and-google-collaborate-on-computing-asic-data-centers/

https://www.datacenterfrontier.com/servers/article/33005340/closer-look-metas-custom-asic-for-ai-computing

https://ieeexplore.ieee.org/document/7551392

Seriously. You realize that the most successful TPUs in the industry are ASICs, right? And that all the “AI” components in your phone are too? What are you even talking about here?

TPU units are specific to individual model frameworks, and engineers avoid using them for that reason. The most successful adoptions for them so far are vendor locked-in NN Models a la Amazon (Trainium), and Google (Coral), and neither of them has wide adoption since they have limited scopes. The GPU game being flexible in this arena is exactly why companies like OpenAI are struggling to justify the costs in using them over TPUs: it’s easy to run up front, but the cost is insane, and TPU is even more expensive in most cases. It’s also inflexible should you need to do something like multi-model inference (detection+evaluation+result…etc).

As I said, ASICs are single purpose, so you’re stuck running a limited model engine (Tensorflow) and instruction set. They also take a lot of engineering effort to design, so unless you’re going all-in on a specific engine and thinking you’re going to be good for years, it’s short sighted to do so. If you read up, you’ll see the most commonly deployed edge boards in the world are…Jetsons.

Enter FPGAs.

FPGAs have speedup improvements for certain things like transcoding and inference in the 2x-5x range for specific workloads, and much higher for ML purposes and in-memory datasets (think Apache Ignite+Arrow workloads), and at a massive reduction in power and cooling, so obviously very attractive for datacenters to put into production. The newer slew of chips out are even reprogrammable “on the fly”, meaning a simple context switch and flash can take milliseconds, and multi-purpose workloads can exist in a single application, where this was problematic before.

So unless you’ve got some articles about the most prescient AI companies currently using GPUs and moving to ASIC, the field is wide open for FPGA, and the datacenter adoption of such says it’s the path forward unless Nvidia starts kicking out more efficient devices.

Dr. Dabbles
link
fedilink
English
01Y

Now ask open AI to type for you what the draw backs of FPGA is. Also the newest slew of chips is using partially charged NAND gates instead of FPGA.

Almost all ASIC being used right now is implementing the basic math functions, activations, etc. and the higher level work is happening in more generalized silicon. You can not get the transistor densities necessary for modern accelerator work in FPGA.

Friend, I do this for a living, and I have no idea why you’re even bringing gating into the equation, because it doesn’t even matter.

I’m assuming you’re a big crypto fan, because that’s about all I could say of ASIC in an HPC type of environment to be good for. Companies who pay the insane amounts of money for “AI” right now want a CHEAP solution, and ASIC is the most short-term, e-wastey, inflexible solve to that problem. When you get a job in the industry and understand the different vectors, let’s talk. Otherwise, you’re just spouting junk.

Dr. Dabbles
link
fedilink
English
11Y

I’m assuming you’re a big crypto fan

Swing and a miss.

because that’s about all I could say of ASIC in an HPC type of environment to be good for

Really? Gee, I think switching fabrics might have a thing to tell you. For someone that does this for a living, to not know the extremely common places that ASICs are used is a bit of a shock.

want a CHEAP solution

Yeah, I already covered that in my initial comment, thanks for repeating my idea back to me.

and ASIC is the most short-term

Literally being atabled to the Intel tiles in Sapphire Rapids and beyond. Used in every switch, network card, and millions of other devices. Every accelerator you can list is an ASIC. Shit, I’ve got a Xilinx Alveo 30 in my basement at home. But yeah, because you can get an FPGA instance in AWS, you think you know that ASICs aren’t used. lmao

e-wastey

I’ve got bad news for you about ML as a whole.

inflexible

Sometimes the flexibility of a device’s application isn’t the device itself, but how it’s used. Again, if I can do thousands, tens of thousands, or hundreds of thousands of integer operations in a tenth of the power, and a tenth of the clock cycles, then load those results into a segment of activation functions that can do the same, and all I have to do is move this data with HBM and perhaps add some cheap ARM cores, bridge all of this into a single SoC product, and sell them on the open market, well then I’ve created every single modern ARM product that has ML acceleration. And also nvidia’s latest products.

Woops.

When you get a job in the industry

I’ve been a hardware engineer for longer than you’ve been alive, most likely. I built my first FPGA product in the 90s. I strongly suspect you just found this hammer and don’t actually know what the market as a whole entails, let alone the long LONG history of all of these things.

Do look up ASICs in switching, BTW. You might learn something.

@the_q@lemmy.world
link
fedilink
English
111Y

Oh is that why? It’s not because Nvidia is making bank in the AI sector and he’s just another greedy CEO?

Dr. Dabbles
link
fedilink
English
51Y

“Please buy more of my hardware so nobody finds out how deep in trouble my company is. PLEASE.”

Create a post

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


  • 1 user online
  • 186 users / day
  • 583 users / week
  • 1.37K users / month
  • 4.49K users / 6 months
  • 1 subscriber
  • 7.41K Posts
  • 84.7K Comments
  • Modlog