Yongsuk Kwon
Orcid: 0000-0002-1956-4629
According to our database1,
Yongsuk Kwon
authored at least 9 papers
between 2021 and 2025.
Collaborative distances:
Collaborative distances:
Timeline
Legend:
Book In proceedings Article PhD thesis Dataset OtherLinks
On csauthors.net:
Bibliography
2025
Accelerating Confidential Recommendation Model Inference With Near-Memory Processing.
IEEE Trans. Dependable Secur. Comput., 2025
IEEE Comput. Archit. Lett., 2025
Scalable Processing-Near-Memory for 1M-Token LLM Inference: CXL-Enabled KV-Cache Management Beyond GPU Limits.
Proceedings of the 34th International Conference on Parallel Architectures and Compilation Techniques, 2025
2024
Proceedings of the 38th ACM International Conference on Supercomputing, 2024
An LPDDR-based CXL-PNM Platform for TCO-efficient Inference of Transformer-based Large Language Models.
Proceedings of the IEEE International Symposium on High-Performance Computer Architecture, 2024
AttAcc! Unleashing the Power of PIM for Batched Transformer-based Generative Model Inference.
Proceedings of the 29th ACM International Conference on Architectural Support for Programming Languages and Operating Systems, 2024
2023
A Hardware-Friendly Tiled Singular-Value Decomposition-Based Matrix Multiplication for Transformer-Based Models.
IEEE Comput. Archit. Lett., 2023
2022
Near-Memory Processing in Action: Accelerating Personalized Recommendation With AxDIMM.
IEEE Micro, 2022
2021
Understanding and Reducing Weight-Load Overhead of Systolic Deep Learning Accelerators.
Proceedings of the 18th International SoC Design Conference, 2021