Fine-tune a large language model (llm) for multi-turn conversations and run it on a Text Generation Inference (TGI) server

This blog post delves into the initial fine-tuning process for large language models (LLMs) for multi-turn conversations and their deployment on Text Generation Inference (TGI) servers. It covers topics such as use cases, data formats, training data preparation, server setup, and evaluation frameworks. The goal is to guide readers through the process of fine-tuning and deploying LLMs.

How to create a watsonx.ai REST client in Spring Boot?

This blog post demonstrates the Java Spring Boot implementation to invoke a watsonx.ai endpoint. It outlines the classes and steps involved, including building and sending requests, handling prompts, and extracting answers. The post also provides sample code for invoking the endpoint and using RestTemplate. Overall, it offers a comprehensive guide on utilizing watsonx.ai in a Spring Boot application.

Unleash your creativity and design a custom visualization for the Shelly 3EM device with Grafana

The blog post details an example implementation of a connection server using Shelly 3EM, IBM Cloud Cloudant, and Grafana. It aims to store historical data for visualizing electricity consumption. The project involves detailed architecture, environment setup, Python, FastAPI, Podman, and more usage. The setup covers Raspberry Pi, Podman Compose, and IBM Cloud Code Engine environments, with prerequisites and detailed configurations. The approach allows users to monitor and visualize power consumption efficiently and cost-effectively using Grafana.

Blog at WordPress.com.

Up ↑