{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"Quest","owner":"mit-han-lab","isFork":false,"description":"[ICML 2024] Quest: Query-Aware Sparsity for Efficient Long-Context LLM Inference","allTopics":[],"primaryLanguage":{"name":"Cuda","color":"#3A4E3A"},"pullRequestCount":0,"issueCount":0,"starsCount":71,"forksCount":5,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,7,2,2],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-28T09:30:24.496Z"}},{"type":"Public","name":"torchsparse","owner":"mit-han-lab","isFork":false,"description":"[MICRO'23, MLSys'22] TorchSparse: Efficient Training and Inference Framework for Sparse Convolution on GPUs.","allTopics":["acceleration","pytorch"],"primaryLanguage":{"name":"Cuda","color":"#3A4E3A"},"pullRequestCount":0,"issueCount":24,"starsCount":1149,"forksCount":130,"license":"MIT License","participation":[0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,1,1,0,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T02:03:43.161Z"}}],"repositoryCount":2,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}