Sparsh Garg
@_sparshgarg_
3D Perception Researcher @ Bosch Center for Artificial Intelligence | CMU Robotics
ID: 1713356440142782464
https://sparsh913.github.io/sparshgarg/ 15-10-2023 00:49:57
57 Tweet
138 Followers
990 Following
Log-linear attention — a new type of attention proposed by Massachusetts Institute of Technology (MIT) which is: - fast and efficient as linear attention - expressive as softmax It uses a small but growing number of memory slots that increases logarithmically with the sequence length. Here's how it works:
TRI's latest Large Behavior Model (LBM) paper landed on arxiv last night! Check out our project website: toyotaresearchinstitute.github.io/lbm1/ One of our main goals for this paper was to put out a very careful and thorough study on the topic to help people understand the state of the