Large language models (LLMs) are rapidly becoming the backbone of AI-driven applications. Without proper optimization, however, LLMs can be expensive to run, slow to serve, and prone to performance bottlenecks. As the demand for real-time AI applications grows, along comes Hands-On Serving and Optimizing LLM Models, a comprehensive guide to the complexities of deploying and optimizing LLMs at scale. In this hands-on book, authors Chi Wang and Peiheng Hu take a real-world approach backed by practical examples and code, and assemble essential strategies for designing robust infrastructures that…mehr
Large language models (LLMs) are rapidly becoming the backbone of AI-driven applications. Without proper optimization, however, LLMs can be expensive to run, slow to serve, and prone to performance bottlenecks. As the demand for real-time AI applications grows, along comes Hands-On Serving and Optimizing LLM Models, a comprehensive guide to the complexities of deploying and optimizing LLMs at scale. In this hands-on book, authors Chi Wang and Peiheng Hu take a real-world approach backed by practical examples and code, and assemble essential strategies for designing robust infrastructures that are equal to the demands of modern AI applications. Whether you're building high-performance AI systems or looking to enhance your knowledge of LLM optimization, this indispensable book will serve as a pillar of your success. * Learn the key principles for designing a model-serving system tailored to popular business scenarios * Understand the common challenges of hosting LLMs at scale while minimizing costs * Pick up practical techniques for optimizing LLM serving performance * Build a model-serving system that meets specific business requirements * Improve LLM serving throughput and reduce latency * Host LLMs in a cost-effective manner, balancing performance and resource efficiency
Chi Wang has over 17 years of experience in the tech industry, with a particular focus on artificial intelligence and distributed systems. For the past 8 years, Chi has been a key contributor at Salesforce's Einstein AI group, where he leads the development of AI platforms and infrastructures that support millions of Salesforce customers and power hundreds of AI features. Currently, as the Director of Engineering, Chi oversees two critical teams: one focused on model serving and optimization solutions, and the other on data science environments. Chi also filed 12 patents in areas such as dataset management, model serving and optimization, data access authorization, and networking management. In addition, he holds an Artificial Intelligence Graduate Certificate from Stanford University, which he completed in 2020.
Es gelten unsere Allgemeinen Geschäftsbedingungen: www.buecher.de/agb
Impressum
www.buecher.de ist ein Internetauftritt der buecher.de internetstores GmbH
Geschäftsführung: Monica Sawhney | Roland Kölbl | Günter Hilger
Sitz der Gesellschaft: Batheyer Straße 115 - 117, 58099 Hagen
Postanschrift: Bürgermeister-Wegele-Str. 12, 86167 Augsburg
Amtsgericht Hagen HRB 13257
Steuernummer: 321/5800/1497
USt-IdNr: DE450055826