Skip to content

Commit 9b28572

Browse files
authored
Merge pull request #9 from LronDC/main
Update README.md
2 parents ba721b4 + 30eb8e6 commit 9b28572

File tree

7 files changed

+49
-16
lines changed

7 files changed

+49
-16
lines changed

README.md

Lines changed: 49 additions & 16 deletions
Original file line numberDiff line numberDiff line change
@@ -1,41 +1,74 @@
1-
![**DTX Logo**](https://raw.githubusercontent.com/DataTunerX/datatunerx-controller/main/assets/logo/Logo_DataTunerX%20-%20Horizontal%20-%20Color%20Light.png#gh-dark-mode-only)
2-
![**DTX Logo**](https://raw.githubusercontent.com/DataTunerX/datatunerx-controller/main/assets/logo/Logo_DataTunerX%20-%20Horizontal%20-%20Color%20Dark.png#gh-light-mode-only)
1+
![DTX Logo](https://raw.githubusercontent.com/DataTunerX/datatunerx-controller/main/assets/logo/Logo_DataTunerX%20-%20Horizontal%20-%20Color%20Light.png#gh-dark-mode-only)
2+
![DTX Logo](https://raw.githubusercontent.com/DataTunerX/datatunerx-controller/main/assets/logo/Logo_DataTunerX%20-%20Horizontal%20-%20Color%20Dark.png#gh-light-mode-only)
33

44
![Kubernetes](https://img.shields.io/badge/kubernetes-%23326ce5.svg?style=flat&logo=kubernetes&logoColor=white)
55
![release](https://img.shields.io/badge/version-0.1.0-blue)
66
![owner](https://img.shields.io/badge/daocloud.io-green)
77
![fine-tuning](https://img.shields.io/badge/fine--tuning-8B3E3)
88
# Welcome 👋
99

10-
**Description**:
11-
*DataTunerX*, short for *DTX*, is a powerful tool.
10+
***DataTunerX (DTX)*** is designed as a cloud-native solution integrated with distributed computing frameworks. Leveraging scalable *GPU* resources, it's a platform built for efficient fine-tuning *LLMs* with a focus on practical utility. Its core strength lies in facilitating batch fine-tuning tasks, enabling users to conduct multiple tasks concurrently within a single ***experiment***. ***DTX*** encompasses essential capabilities such as ${\color{#D8CBBE}dataset \space management}$, ${\color{#BDE7BD}hyperparameter \space control}$, ${\color{#F1A7A7}fine-tuning \space workflows}$, ${\color{#BADBF4}model \space management}$, ${\color{#F4CEDB}model \space evaluation}$, ${\color{#D2E3EE}model \space comparison \space inference}$, and a ${\color{#F9E195}modular \space plugin \space system}$.
1211

13-
**Technology stack**:
14-
Built using *Go*.
12+
**Technology stack**:
13+
14+
***DTX*** is built on cloud-native principles, employing a variety of [*Operators*](https://www.redhat.com/en/topics/containers/what-is-a-kubernetes-operator) that consist of distinct *Custom Resource Definitions (CRDs)* and *Controller* logic. Developed primarily in *Go*, the implementation utilizes the [*operator-sdk*](https://github.com/operator-framework/operator-sdk) toolkit. Operating within a [*Kubernetes (K8s)*](https://github.com/kubernetes/kubernetes) environment, ***DTX*** relies on the operator pattern for *CRD* development and management. Furthermore, ***DTX*** integrates with [*kuberay*](https://github.com/ray-project/kuberay) to harness distributed execution and inference capabilities.
1515

1616
**Status**:
17+
1718
*Alpha (v0.1.0)* - Early development phase. [CHANGELOG](CHANGELOG.md) for details on recent updates.
1819

19-
**Links to production or demo instances**:
20+
**Quick Demo & More Documentation**:
2021

21-
[Demo Instance](https://github.com/DataTunerX/datatunerx-controller) (COMING SOON)
22+
- [Demo](https://github.com/DataTunerX/datatunerx-controller) (COMING SOON)
2223

23-
[Documentation](https://github.com/DataTunerX/datatunerx-controller) (COMING SOON)
24+
- [Documentation](https://github.com/DataTunerX/datatunerx-controller) (COMING SOON)
25+
26+
**Screenshot**:
2427

25-
**Screenshot**:
2628
![**DTX Screenshot**](https://raw.githubusercontent.com/DataTunerX/datatunerx-controller/main/assets/screenshot/Job_Details.png)
2729

28-
# Dependencies 🖇️
30+
# What DTX can do? 💪
2931

30-
Describe any dependencies that must be installed for this software to work. This includes programming languages, databases, build tools, etc. Specify versions if necessary.
32+
***DTX*** empowers users with a robust set of features designed for efficient fine-tuning of large language models. Dive into the capabilities that make ***DTX*** a versatile platform:
3133

32-
# Installation 📦
34+
## 1. Dataset Management 🗄️
35+
Effortlessly manage datasets by supporting both *S3* protocol (*http* is coming) and local dataset uploads. Datasets are organized with splits such as test, validate, and training. Additionally, feature mapping enhances flexibility for fine-tuning jobs.
3336

34-
Detailed instructions on how to install, configure, and run the project are available in the [*INSTALL*](INSTALL.md) document.
37+
## 2. Fine-Tuning Experiments 🧪
38+
Conduct fine-tuning experiments by creating multiple fine-tuning jobs. Each job can employ different llms, datasets, and hyperparameters. Evaluate the fine-tuned models uniformly through the experiment's evaluation unit to identify the fine-tuning results.
39+
<div align="center">
40+
<img src="https://raw.githubusercontent.com/DataTunerX/datatunerx-controller/main/assets/design/finetune.png" alt="FineTune" width="30%" />
41+
<img src="https://raw.githubusercontent.com/DataTunerX/datatunerx-controller/main/assets/design/finetunjobe.png" alt="FineTuneJob" width="30%" />
42+
<img src="https://raw.githubusercontent.com/DataTunerX/datatunerx-controller/main/assets/design/finetuneexperiment.png" alt="FineTuneExperiment" width="30%" />
43+
</div>
44+
45+
## 3. Job Insights 📊
46+
Gain detailed insights into each fine-tuning job within an experiment. Explore job details, logs, and metric visualizations, including learning rate trends, training loss, and more.
47+
48+
## 4. Model Repository 🗃️
49+
Store llms in the model repository, facilitating efficient management and deployment of inference services.
50+
51+
## 5. Hyperparameter Group Management 🧰
52+
Utilize a rich parameter configuration system with support for diverse parameters and template-based differentiation.
3553

36-
# Configuration ⚙️
54+
## 6. Inference Services 🚀
55+
Deploy inference services for multiple models simultaneously, enabling straightforward comparison and selection of the best-performing model.
3756

38-
If the software is configurable, describe the configuration options in detail, either here or in other linked documentation. (COMING SOON)
57+
## 7. Plugin System 🧩
58+
Leverage the plugin system for datasets and evaluation units, allowing users to integrate specialized datasets and evaluation methods tailored to their unique requirements.
59+
60+
## 8. More Coming 🤹‍♀️
61+
DTX offers a comprehensive suite of tools, ensuring a seamless fine-tuning experience with flexibility and powerful functionality. Explore each feature to tailor your fine-tuning tasks according to your specific needs.
62+
63+
# Why DTX? 🤔
64+
65+
# Architecture 🏛️
66+
67+
Introducing the architectural design provides an overview of how DataTunerX is structured. This includes details on key components, their interactions, and how they contribute to the system's functionality.
68+
69+
# Installation 📦
70+
71+
Detailed instructions on how to install, configure, and run the project are available in the [*INSTALL*](INSTALL.md) document.
3972

4073
# Usage 🖥️
4174

assets/design/batch.png

171 KB
Loading

assets/design/datasetplugin.png

230 KB
Loading

assets/design/evalandinference.png

137 KB
Loading

assets/design/finetune.png

214 KB
Loading
383 KB
Loading

assets/design/finetunejob.png

210 KB
Loading

0 commit comments

Comments
 (0)