@Venkat2811
Research & Engineering - LLM Inference & High Performance systems.
π§βπ» Research & Engineering - LLM Inference & High Performance systems. |πBerlin π©πͺ | π https://venkat.eu | π¬ https://twitter.com/Venkat2811
Nothing here yet.
May 31, 2024 Β· 12 min read Β· (Image Credit: HF TGI Benchmark) Introduction As enterprises and tech enthusiasts increasingly integrate LLM applications into their daily workflows, the demand for TFLOPS is ever increasing. Apple, Microsoft, Google, and Samsung have already introdu...
Join discussion
Apr 18, 2024 Β· 22 min read Β· Introduction Modern software programming languages, compilers, and frameworks abstract away underlying complexities and details, allowing developers to focus on building systems and applications to solve business problems. This design enables enginee...
Join discussion
Apr 8, 2024 Β· 16 min read Β· Introduction In this article, we'll go through some fundamental low level details to understand why GPUs are good at Graphics, Neural Network and Deep Learning tasks and CPUs are good at wide number of sequential, complex general purpose computing ta...
Join discussion
Mar 26, 2024 Β· 7 min read Β· Intro Engineers who've built, deployed and operated backend services would've encountered this error. It usually means your service is serving real user requests - Yay π ! One possible scenario is - you need to fine-tune server OS configuration to s...
Join discussion