You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Copy file name to clipboardexpand all lines: docs/source/publication_list.md
+4-3
Original file line number
Diff line number
Diff line change
@@ -1,8 +1,9 @@
1
-
Full Publications/Events (88)
1
+
Full Publications/Events (89)
2
2
==========
3
-
## 2025 (2)
3
+
## 2025 (3)
4
+
* arXiv: [Faster Inference of LLMs using FP8 on the Intel Gaudi](https://arxiv.org/abs/2503.09975) (Mar 2025)
5
+
* PyTorch landscape: [PyTorch general optimizations](https://landscape.pytorch.org/) (Mar 2025)
4
6
* Blog on SqueezeBits: [[Intel Gaudi]#4. FP8 Quantization](https://blog.squeezebits.com/intel-gaudi-4-fp8-quantization--40269) (Jan 2025)
5
-
* PyTorch landscape: [PyTorch general optimizations](https://landscape.pytorch.org/) (March 2025)
6
7
7
8
## 2024 (7)
8
9
* Blog by Microsoft: [Phi-4 quantization and inference speedup](https://techcommunity.microsoft.com/blog/machinelearningblog/phi-4-quantization-and-inference-speedup/4360047) (Dec 2024)
0 commit comments