Chip-Chat: Challenges and Opportunities in Conversational Hardware Design
https://arxiv.org/abs/2305.13243
https://arxiv.org/abs/2305.13243
Brand new Microscaling format from authors of Block FP (Microsoft, Intel PSG, ...):
° https://www.opencompute.org/documents/ocp-microscaling-formats-mx-v1-0-spec-final-pdf
° https://arxiv.org/abs/2310.10537
° https://github.com/microsoft/microxcaling
° https://www.opencompute.org/documents/ocp-microscaling-formats-mx-v1-0-spec-final-pdf
° https://arxiv.org/abs/2310.10537
° https://github.com/microsoft/microxcaling
Jim Keller's talk at the AI Hardware Summit 2023
https://www.youtube.com/watch?v=lPX1H3jW8ZQ
https://www.youtube.com/watch?v=lPX1H3jW8ZQ
YouTube
AI Hardware w/ Jim Keller
“Our mission is to help you solve your problem in a way that is super cost-effective and available to as many people as possible.”
Listen to Tenstorrent CEO Jim Keller discuss why AI models are graphs and how we built our software and hardware to give you…
Listen to Tenstorrent CEO Jim Keller discuss why AI models are graphs and how we built our software and hardware to give you…
MaxEVA: Maximizing the Efficiency of Matrix Multiplication on Versal AI Engine
https://arxiv.org/abs/2311.04980
https://github.com/enyac-group/MaxEVA
https://arxiv.org/abs/2311.04980
https://github.com/enyac-group/MaxEVA
GitHub
GitHub - enyac-group/MaxEVA: MaxEVA: Maximizing the Efficiency of Matrix Multiplication on Versal AI Engine (accepted as full paper…
MaxEVA: Maximizing the Efficiency of Matrix Multiplication on Versal AI Engine (accepted as full paper at FPT'23) - enyac-group/MaxEVA
#Unhardware
Doug Burger (Microsoft Catapult, Brainwave, MX data formats) talks about his next project
https://thesequence.substack.com/p/the-sequence-chat-doug-burger-technical
Doug Burger (Microsoft Catapult, Brainwave, MX data formats) talks about his next project
https://thesequence.substack.com/p/the-sequence-chat-doug-burger-technical
TheSequence
The Sequence Chat: Doug Burger- Technical Fellow, Microsoft Research About Building Autonomous Agents, AutoGen and the Future of…
One of the collaborators in the AutoGen project, shares insights about its vision, architecture and the future of autonomous agents.
https://fpt2023.org/workshops.html
Workshops from Intel and AMD dedicated to their specialized hardware blocks in FPGA for ML acceleration
AI Engines and AI Tensor blocks
Yokohama, Japan, 11th - 14th December, 2023
Workshops from Intel and AMD dedicated to their specialized hardware blocks in FPGA for ML acceleration
AI Engines and AI Tensor blocks
Yokohama, Japan, 11th - 14th December, 2023
TPU v5p for Training
v5e for cost effective
v5p for powerful
https://cloud.google.com/blog/products/ai-machine-learning/introducing-cloud-tpu-v5p-and-ai-hypercomputer
v5e for cost effective
v5p for powerful
https://cloud.google.com/blog/products/ai-machine-learning/introducing-cloud-tpu-v5p-and-ai-hypercomputer
Google Cloud Blog
Introducing Cloud TPU v5p and AI Hypercomputer | Google Cloud Blog
The new TPU v5p is a core element of AI Hypercomputer, which is tuned, managed, and orchestrated specifically for gen AI training and serving.
FP6 quantization for LLMs
https://www.linkedin.com/feed/update/urn:li:activity:7141243626730176512/
https://arxiv.org/pdf/2312.08583.pdf
https://www.linkedin.com/feed/update/urn:li:activity:7141243626730176512/
https://arxiv.org/pdf/2312.08583.pdf
Linkedin
Leon Song on LinkedIn: 2312.08583.pdf
Very proud that the team has just pushed out the algorithmic investigation into FP6 and its new quantization strategy for LLMs that tackles the quality challenges from the INT4 solutions. Very soon, my team will release an ultra fast FP6 kernel design on…
Soft GPGPU for FPGA
https://arxiv.org/abs/2401.04261
https://arxiv.org/abs/2401.04261
arXiv.org
A Statically and Dynamically Scalable Soft GPGPU
Current soft processor architectures for FPGAs do not utilize the potential of the massive parallelism available. FPGAs now support many thousands of embedded floating point operators, and have...
The main FPGA News today is the new name of Intel PSG
https://youtu.be/Mb34D4f5tc8?si=NTNt5hZxYdY7JaZe
https://youtu.be/Mb34D4f5tc8?si=NTNt5hZxYdY7JaZe
YouTube
We are Altera. We are for the innovators.
Today we embark on an exciting journey as we transition to Altera, an Intel Company. In a world of endless opportunities and challenges, we are here to provide the flexibility needed by our ecosystem of customers and partners to pioneer and accelerate innovation.…
Collective Communications Library for FPGA?
https://arxiv.org/abs/2312.11742
https://arxiv.org/abs/2312.11742
arXiv.org
ACCL+: an FPGA-Based Collective Engine for Distributed Applications
FPGAs are increasingly prevalent in cloud deployments, serving as Smart NICs or network-attached accelerators. Despite their potential, developing distributed FPGA-accelerated applications remains...
Next gen facebook accelerator (MTIA v2)
https://ai.meta.com/blog/next-generation-meta-training-inference-accelerator-AI-MTIA/
https://ai.meta.com/blog/next-generation-meta-training-inference-accelerator-AI-MTIA/
Meta
Our next generation Meta Training and Inference Accelerator
We are sharing details of our next generation chip in our Meta Training and Inference Accelerator (MTIA) family. MTIA is a long-term bet to provide the most efficient architecture for Meta’s unique workloads.
FPGA Startup offers LLM performance better than Nvidia A100
https://hc2023.hotchips.org/assets/program/posters/HC2023.hyperaccel.ai.Moon.Poster.pdf
https://hc2023.hotchips.org/assets/program/posters/HC2023.hyperaccel.ai.Moon.Poster.pdf