ML workloads will be really interesting. An ok sized GPU with access to 128GB of RAM could be faster than any other consumer device in cases where those just can't fit the model into memory. Intels B60 showed already that memory alone can make a difference.
As always a big thanks for the article. However, I would have liked to see a bit more emphasis on the new D2D interconnect. Are there any measurable differences regarding the C2C latencies or regarding power usage in idle/load compared to the good old IFoP of their Desktop counterparts?
ML workloads will be really interesting. An ok sized GPU with access to 128GB of RAM could be faster than any other consumer device in cases where those just can't fit the model into memory. Intels B60 showed already that memory alone can make a difference.
As always a big thanks for the article. However, I would have liked to see a bit more emphasis on the new D2D interconnect. Are there any measurable differences regarding the C2C latencies or regarding power usage in idle/load compared to the good old IFoP of their Desktop counterparts?