Skip to main content

Print Edition Cover Revealed: Generative AI with Local LLM

 Disclaimer: this post is not a usual technical article. If you are not interested, you can skip it. I just want to share a few recent moments with my readers.


After several months of work, the print edition of my book Generative AI with Local LLM has reached a new milestone — the cover design for the print version is now complete.

This edition is a translated version of the original book and is planned for publication through BomBora Publisher in Eastern Europe. The publisher has adapted the content specifically for non-technical readers, making the material more accessible while keeping its practical focus. The translation and adaptation process took about eight months and included multiple rounds of review, proofreading, and refinement to ensure clarity and quality.

The book is designed to help readers understand and apply generative AI through practical and accessible methods. It emphasizes clarity, simplicity, and real-world application.

Key Features of the Book:

  • Clear and Concise Explanations: Complex AI concepts are broken down into simple, step-by-step guidance, making them accessible to readers of all backgrounds.

  • Hands-On Projects: Each chapter includes guided projects, from environment setup to model deployment.

  • Real-World Applications: The examples focus on solving real problems, providing practical experience in applying AI methods.

  • Essential Tools and Libraries: The book introduces and uses tools such as Langchain, Vanna, TensorFlow, and PyTorch.

  • Project-Based Learning: Readers work through projects that range from image recognition to advanced large language model fine-tuning.

The book aims to serve as a concise and practical starting point for anyone interested in generative AI and local LLM development.

More details about the original edition are available at https://leanpub.com/quickstartwithai

Comments

Popular posts from this blog

8 things every developer should know about the Apache Ignite caching

Any technology, no matter how advanced it is, will not be able to solve your problems if you implement it improperly. Caching, precisely when it comes to the use of a distributed caching, can only accelerate your application with the proper use and configurations of it. From this point of view, Apache Ignite is no different, and there are a few steps to consider before using it in the production environment. In this article, we describe various technics that can help you to plan and adequately use of Apache Ignite as cutting-edge caching technology. Do proper capacity planning before using Ignite cluster. Do paperwork for understanding the size of the cache, number of CPUs or how many JVMs will be required. Let’s assume that you are using Hibernate as an ORM in 10 application servers and wish to use Ignite as an L2 cache. Calculate the total memory usages and the number of Ignite nodes you have to need for maintaining your SLA. An incorrect number of the Ignite nodes can become a b...

Analyse with ANT - a sonar way

After the Javaone conference in Moscow, i have found some free hours to play with Sonar . Here is a quick steps to start analyzing with ANT projects. Sonar provides Analyze with ANT document to play around with ANT, i have just modify some parts. Here is it. 1) Download the Sonar Ant Task and put it in your ${ANT_HOME}/lib directory 2) Modify your ANT build.xml as follows: <?xml version = '1.0' encoding = 'windows-1251'?> <project name="abc" default="build" basedir="."> <!-- Define the Sonar task if this hasn't been done in a common script --> <taskdef uri="antlib:org.sonar.ant" resource="org/sonar/ant/antlib.xml"> <classpath path="E:\java\ant\1.8\apache-ant-1.8.0\lib" /> </taskdef> <!-- Out-of-the-box those parameters are optional --> <property name="sonar.jdbc.url" value="jdbc:oracle:thin:@xyz/sirius.xyz" /> <property na...

Apache Ignite Baseline Topology by Examples

Ignite Baseline Topology or BLT represents a set of server nodes in the cluster that persists data on disk. Where, N1-2 and N5 server nodes are the member of the Ignite clusters with native persistence which enable data to persist on disk. N3-4 and N6 server nodes are the member of the Ignite cluster but not a part of the baseline topology. The nodes from the baseline topology are a regular server node, that store's data in memory and on the disk, and also participates in computing tasks. Ignite clusters can have different nodes that are not a part of the baseline topology such as: Server nodes that are not used Ignite native persistence to persist data on disk. Usually, they store data in memory or persists data to a 3rd party database or NoSQL. In the above equitation, node N3 or N4 might be one of them. Client nodes that are not stored shared data. To better understand the baseline topology concept, let’s start at the beginning and try to understand its goal and what ...