Over 6,000 cloud edge servers revealed! Nvidia Tesla T10 real-world performance sharing

Tesla T10 GPU-Z

Recently, a special display card has emerged in the Chinese market during the Tesla T10 launch event — a GPU originally designed by NVIDIA exclusively for cloud gaming services, primarily used in the GeForce NOW cloud gaming platform. These retired display cards have now entered the secondary market, with current prices on Chinese platforms around 1,350 RMB (approximately 190 USD). Due to their affordability, I purchased two to examine their performance.

Leveraging Nvidia GPU on Kubernetes for LLM Chatbot Deployment

在 Kubernetes 上利用 Nvidia GPU 部署 LLM Chatbot

With the rapid advancement of artificial intelligence and large language models, more and more companies and developers are eager to integrate language models into their own chatbot systems (LLM Chatbot). This article aims to guide readers through deploying a high-performance LLM chatbot in a Kubernetes environment using Nvidia GPU, covering everything from essential installation and tools to detailed deployment steps.

Understanding and Applying Linux PSI (Pressure Stall Information) Metrics

如何監控 container 的 PSI 資訊

Introduction When CPU, memory, or I/O resources experience contention, workloads may suffer from increased latency, degraded performance, and even face abrupt OOM (Out of Memory) termination. Without proper monitoring to detect such contention, users risk exhausting their hardware resources, leading to frequent crashes due to over-provisioning. Since Linux kernel 4.20, the Linux kernel has introduced PSI (Pressure Stall Information), a metric that enables users to precisely understand how resource shortages impact overall system performance. This article will briefly explain PSI and how to interpret its data.

Deploying Charmed Kubernetes with OpenStack Integrator

部署 Charmed Kubernetes with OpenStack Integrator

Introduction Charmed Kubernetes is a Kubernetes deployment solution provided by Canonical, enabling deployment of Kubernetes across various environments through juju. This article will walk through deploying Charmed Kubernetes onto OpenStack, and leveraging the OpenStack Integrator to provide Persistent Volumes and Load Balancers to Kubernetes.

AMD GPU and Deep Learning: Practical Learning Guide

AMD GPU 與深度學習:實用教學指南

Historically, AMD GPUs have been considered less suitable for deep learning tasks, leading many deep learning users to prefer Nvidia GPUs. However, recently, LLMs (Large Language Models) have gained significant attention, and numerous research teams have released models based on LLaMA, prompting me to feel inspired and eager to experiment. I have several AMD GPUs with ample VRAM, so I’ve decided to test these cards for running LLMs.

Introduction to Kubernetes Cluster-API

Kubernetes Cluster-API 介紹

Kubernetes has been developing for many years in the cloud-native world and has also evolved numerous specialized projects related to managing its lifecycle, such as Kops and Rancher. VMware, on the other hand, has launched a project named Cluster API to leverage Kubernetes' own capabilities for managing other Kubernetes clusters. This article will briefly introduce the Cluster API project.