-
REAM: Compressing Mixture-of-Experts LLMs
Merging experts in Mixture-of-Experts (MoE) LLMs to compress a 235B LLM.
-
Training LLMs Faster by Learning Neuron Interaction
-
Can we do better than Convolutional Neural Networks?
-
Spectral Graph Convolution Explained and Implemented Step By Step
-
Anisotropic, Dynamic, Spectral and Multiscale Filters Defined on Graphs