About me
I am a 1st year Computer Science Ph.D. student in Georgia Institute of Technology.
Research
I have been exploring 2 threads of research:
- Improving the performance and efficiency of language models, with a focus on State Space Based models such as Mamba
- The application of language models to speed up and lower the entrance barrier of hardware design, including automatic generation of RTL
Publications
- [ACL ‘25] [LAMB: A Training-Free Method to Enhance the Long-Context Understanding of SSMs via Attention-Guided Token Filtering]
- [ICLR ‘25] LongMamba: Enhancing Mamba’s Long-Context Capabilities via Training-Free Receptive Field Enlargement
- [NeurIPS ‘24] 3D Gaussian Can Be Sparser Than You Thought: Efficient Rendering via Learned Fragment Pruning
- [MICRO ‘24] VGA: Hardware Accelerator for Scalable Long Sequence Model
- [DAC ‘23] A Memory-Efficient Edge Inference Accelerator with XOR-based Model Compression
