BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
semantic_scholar
The Condensate Theorem: Transformers are O(n), Not $O(n^2)$
arxiv
NEXUS: Bit-Exact ANN-to-SNN Equivalence via Neuromorphic Gate Circuits with Surrogate-Free Training
arxiv
Metriplector: From Field Theory to Neural Architecture
arxiv
Kathleen: Oscillator-Based Byte-Level Text Classification Without Tokenization or Attention
arxiv