I had some email me with this question:
As I experiment with reconfigurable risc v cores, is there tensor acceleration for LLM and diffusion network ai models on the FPGA fabric?
I had some email me with this question:
As I experiment with reconfigurable risc v cores, is there tensor acceleration for LLM and diffusion network ai models on the FPGA fabric?
Hi,
Although it's technically possible for someone to implement this, I don't think it would be very efficient to do this in the FPGA cells. There are better options, which would be (for instance) to use an AMD Zynq type part (i.e. with processing subsystem) and attach an accelerator module, using say AMD PCIe IP https://www.xilinx.com/products/technology/pci-express.html . Example is Google's accelerator module, but actually obtaining that could be a challenge (due to semiconductor shortage). I've been waiting 4 months just to get hold of a small quantity (3) of the Google accelerators, and no promised date. Maybe it's easier to obtain if there is a requirement for large volumes perhaps, but I don't know : (