This blog post delves into the initial fine-tuning process for large language models (LLMs) for multi-turn conversations and their deployment on Text Generation Inference (TGI) servers. It covers topics such as use cases, data formats, training data preparation, server setup, and evaluation frameworks. The goal is to guide readers through the process of fine-tuning and deploying LLMs.
Writing an HTML-to-Text converter can be the first task in an AI pipeline with the JSOUP Java Library
This blog post is about the powerful JSOUP Java Library, which allows you to convert an HTML to plain formatted text based on your requirements by extracting and inspecting HTML elements in various ways. We check two methods to do this in this example.
Show the collection IDs of IBM Cloud Watson Discovery projects using cURL
This blog post is a simple example (cheat sheet) of listing the collections for a project in Watson Discovery using cURL and the IBM Cloud Watson Discovery API V2. You can get more details in the IBM Cloud Watson Discovery API documentation. 1. Log on to IBM Cloud ibmcloud login (-sso) REGION=us-south GROUP=default ibmcloud target -r... Continue Reading →
Cheat Sheet: Get started with Data Science and Python
This blog post is a short cheat sheet when it comes to getting started with data science and Python and is not a full list.
Some thoughts about ChatGPT and AI
Everyone is now talking about this new way of using AI in an interactive form of communication. When we talk about free or open and AI, these three questions immediately came to my mind: “If you're not paying for the product, then you are the product.” That's a quote from Daniel Hövermann in The Social Dilemma. What will be the business model? "Will my Job be replaced?" "Can I trust, and what is my remaining responsibility?"
How to create a model container image for Watson NLP for Embed
This longer blog post shows how to : … build a model init container with a custom model for Watson NLP for Embed. … upload the model init container to the IBM Cloud container registry. … deploy the model init container and the Watson NLP runtime to an IBM Cloud Kubernetes Cluster. … test Watson NLP runtime with the loaded model using the REST API.
Run Watson Speech to Text for Embed on an IBM Cloud Kubernetes cluster in a Virtual Private Cloud environment
This blog post is about to deploy the IBM Watson Speech to Text Library for Embed to an IBM Cloud Kubernetes cluster. IBM Cloud Kubernetes cluster is a “certified, managed Kubernetes solution, built for creating a cluster of compute hosts to deploy and manage containerized apps on IBM Cloud“.
Create a custom dictionary model for Watson NLP
This blog post is about, how to create a custom dictionary model for Watson NLP. One capability of the Watson NLP is the "Entity extraction to find mentions of entities (like person, organization, or date)." We will adapt the Watson NLP model to extract entities from a given text to find single entities like names and locations which are identified by an entry and its label.
Run Watson NLP for Embed in a KServe ModelMesh Serving environment on an IBM Cloud Kubernetes cluster in a VPC environment
This blog post is about to run Watson NLP for Embed example in a KServe ModelMesh Serving environment on an IBM Cloud Kubernetes cluster in a Virtual Private Cloud environment and reuses parts of the IBM Watson Libraries for Embed documentation.
Run Watson NLP for Embed on an IBM Cloud Kubernetes cluster in a Virtual Private Cloud environment
This blog post is about to deploy the IBM Watson Natural Language Processing Library for Embed to an IBM Cloud Kubernetes cluster in a Virtual Private Cloud (VPC) environment and is related to my blog post Run Watson NLP for Embed on IBM Cloud Code Engine. IBM Cloud Kubernetes cluster is a “certified, managed Kubernetes solution, built for creating a cluster of compute hosts to deploy and manage containerized apps on IBM Cloud“.
