SwathiManikya's picture
Update README.md
b1eb5a6 verified
---
dataset_name: "hlo-feature-dataset"
pretty_name: "HLO Feature Dataset for Deep Learning Resource Estimation"
dataset_type: "graph-and-tabular"
license: "apache-2.0"
task_categories:
- graph-ml
- tabular-regression
language: "en"
tags:
- HPC
- resource-prediction
- XLA
- compiler-features
- deep-learning
- graph-learning
- scheduling
size_categories:
- 1K<n<10K
source_datasets:
- custom
dataset_summary: >
The HLO Feature Dataset contains High-Level Optimizer (HLO) graph features and metadata extracted
from deep learning training workloads. It is designed for tasks such as runtime prediction, resource
estimation, and graph-based machine learning in HPC environments.
Each entry pairs model configuration metadata with compiler graph data stored in `.npz` format.
Ideal for ML system optimization studies, GNN research, and AI workload scheduling.
structured_data:
features:
- name: "batch"
type: "integer"
- name: "epochs"
type: "integer"
- name: "learn_rate"
type: "float"
- name: "gpu_core_count"
type: "integer"
- name: "gpu_memory_size"
type: "integer"
- name: "fit_time"
type: "float"
- name: "npz_path"
type: "string"
graph_data:
node_features: "node_feat"
edge_index: "edge_index"
additional_keys:
- "node_opcode"
- "node_config_ids"
- "node_splits"
usage_example: |
```python
from datasets import load_dataset
import numpy as np
dataset = load_dataset("your-username/hlo-feature-dataset")
sample = dataset['train'][0]
graph_data = np.load(sample['npz_path'])
node_features = graph_data['node_feat']
edges = graph_data['edge_index']
---
# HLO Feature Dataset for Deep Learning Resource Estimation
[![Dataset](https://img.shields.io/badge/HuggingFace-Dataset-blue)](https://huggingface.co/datasets/your-username/hlo-feature-dataset)
## Dataset Summary
The **HLO Feature Dataset** is a collection of compiler-level graph features (HLO graphs) extracted from deep learning training workloads. Alongside detailed metadata (model configs, GPU stats), this dataset enables machine learning approaches for:
- ⏱️ **Training Time Prediction**
- 📉 **Resource Consumption Estimation**
-**HPC and GPU Scheduling Optimization**
- 🧩 **Graph-based Neural Architecture Analysis**
This dataset is ideal for experimenting with regression models (e.g., XGBoost) and Graph Neural Networks (GNNs) using compiler features.
---
## Supported Tasks
- **⚙️ Runtime & Resource Prediction**: Predict training time (`fit_time`) based on HLO features.
- **📊 ML for Systems Optimization**: Use tabular + graph data for AI workload management.
- **🔗 Graph Representation Learning**: Apply GNNs on HLO graphs (`node_feat`, `edge_index`).
---
## Dataset Structure
Each entry includes:
- **Metadata**: From `dataset-new.csv` (model, optimizer, GPU specs, timing metrics, etc.)
- **HLO Graph Features**: `.npz` files containing:
- `node_opcode`, `node_feat`, `edge_index`, `node_config_ids`, `node_splits`
---
## Usage Example
This example demonstrates how to load metadata, preprocess features, and train an XGBoost model to predict training time (`fit_time`), as shown in the Colab notebook.
```python
import pandas as pd
import numpy as np
from sklearn.model_selection import train_test_split
from sklearn.metrics import mean_squared_error
from xgboost import XGBRegressor
# Load metadata CSV
df = pd.read_csv('dataset-new.csv')
# Example feature selection (drop non-numeric/categorical handling needed)
X = df[['batch', 'epochs', 'learn_rate', 'gpu_core_count', 'gpu_memory_size']]
y = df['fit_time']
# Train-test split
X_train, X_test, y_train, y_test = train_test_split(X, y, test_size=0.2, random_state=42)
# Initialize XGBoost Regressor
xgb_model = XGBRegressor(n_estimators=100, learning_rate=0.1, max_depth=6, random_state=42)
xgb_model.fit(X_train, y_train)
# Evaluate
preds = xgb_model.predict(X_test)
rmse = mean_squared_error(y_test, preds, squared=False)
print(f"RMSE: {rmse}")
```
---
### Example Notebooks
#### 🚀 Baseline: XGBoost for Resource Estimation
A sample baseline implementation using **XGBoost** is provided to demonstrate how to predict resource metrics such as `fit_time` using the dataset's metadata.
📥 **Download the notebook** from the repository:
[Baseline_XGBoost_Resource_Estimation.ipynb](https://huggingface.co/datasets/ICICLE-AI/ResourceEstimation_HLOGenCNN/blob/main/Baseline_XGBoost_Resource_Estimation.ipynb)
This notebook covers:
- Loading and preprocessing metadata from `dataset-new.csv`
- Training an XGBoost regressor to predict training time
- Evaluating model performance (e.g., RMSE)
> ⚡ **Note:** Make sure to adjust paths if cloning the dataset locally or integrating with Hugging Face `datasets` API.
---
### Loading HLO Graph Features
For graph-based ML tasks, load the `.npz` files:
```python
npz_file = df.iloc[0]['npz_path']
graph_data = np.load(npz_file)
node_features = graph_data['node_feat']
edges = graph_data['edge_index']
print("Node Feature Shape:", node_features.shape)
print("Edge Index Shape:", edges.shape)
```
---
<!-- ## Citation
If you use this dataset, please cite:
```
@misc{hlofeatures2025,
title={HLO Feature Dataset for AI Resource Estimation},
author={Your Name},
year={2025},
url={https://huggingface.co/datasets/your-username/hlo-feature-dataset}
} -->
```
---
## License
Specify your license here (e.g., MIT, Apache-2.0).
---
## Contributions
Open to contributions! Feel free to suggest improvements or share your models trained on this dataset.