PrimeHub
  • Introduction
  • Installation
  • Tiers and Licenses
  • End-to-End Tutorial
    • 1 - MLOps Introduction and Scoping the Project
    • 2 - Train and Manage the Model
    • 3 - Compare, Register and Deploy the Model
    • 4 - Build the Web Application
    • 5 - Summary
  • User Guide
    • User Portal
    • Notebook
      • Notebook Tips
      • Advanced Settings
      • PrimeHub Notebook Extension
      • Submit Notebook as Job
    • Jobs
      • Job Artifacts
      • Tutorial
        • (Part1) MNIST classifier training
        • (Part2) MNIST classifier training
        • (Advanced) Use Job Submission to Tune Hyperparameters
        • (Advanced) Model Serving by Seldon
        • Job Artifacts Simple Usecase
    • Models
      • Manage and Deploy Model
      • Model Management Configuration
    • Deployments
      • Pre-packaged servers
        • TensorFlow server
        • PyTorch server
        • SKLearn server
        • Customize Pre-packaged Server
        • Run Pre-packaged Server Locally
      • Package from Language Wrapper
        • Model Image for Python
        • Model Image for R
        • Reusable Base Image
      • Prediction APIs
      • Model URI
      • Tutorial
        • Model by Pre-packaged Server
        • Model by Pre-packaged Server (PHFS)
        • Model by Image built from Language Wrapper
    • Shared Files
    • Datasets
    • Apps
      • Label Studio
      • MATLAB
      • MLflow
      • Streamlit
      • Tutorial
        • Create Your Own App
        • Create an MLflow server
        • Label Dataset by Label Studio
        • Code Server
    • Group Admin
      • Images
      • Settings
    • Generate an PrimeHub API Token
    • Python SDK
    • SSH Server Feature
      • VSCode SSH Notebook Remotely
      • Generate SSH Key Pair
      • Permission Denied
      • Connection Refused
    • Advanced Tutorial
      • Labeling the data
      • Notebook as a Job
      • Custom build the Seldon server
      • PrimeHub SDK/CLI Tools
  • Administrator Guide
    • Admin Portal
      • Create User
      • Create Group
      • Assign Group Admin
      • Create/Plan Instance Type
      • Add InfuseAI Image
      • Add Image
      • Build Image
      • Gitsync Secret for GitHub
      • Pull Secret for GitLab
    • System Settings
    • User Management
    • Group Management
    • Instance Type Management
      • NodeSelector
      • Toleration
    • Image Management
      • Custom Image Guideline
    • Volume Management
      • Upload Server
    • Secret Management
    • App Settings
    • Notebooks Admin
    • Usage Reports
  • Reference
    • Jupyter Images
      • repo2docker image
      • RStudio image
    • InfuseAI Images List
    • Roadmap
  • Developer Guide
    • GitHub
    • Design
      • PrimeHub File System (PHFS)
      • PrimeHub Store
      • Log Persistence
      • PrimeHub Apps
      • Admission
      • Notebook with kernel process
      • JupyterHub
      • Image Builder
      • Volume Upload
      • Job Scheduler
      • Job Submission
      • Job Monitoring
      • Install Helper
      • User Portal
      • Meta Chart
      • PrimeHub Usage
      • Job Artifact
      • PrimeHub Apps
    • Concept
      • Architecture
      • Data Model
      • CRDs
      • GraphQL
      • Persistence Storages
      • Persistence
      • Resources Quota
      • Privilege
    • Configuration
      • How to configure PrimeHub
      • Multiple Jupyter Notebook Kernels
      • Configure SSH Server
      • Configure Job Submission
      • Configure Custom Image Build
      • Configure Model Deployment
      • Setup Self-Signed Certificate for PrimeHub
      • Chart Configuration
      • Configure PrimeHub Store
    • Environment Variables
Powered by GitBook
On this page
  • Prerequisites
  • Creating New Dataset
  • Browse Dataset
  • Access from Jupyter
  • Editing Dataset
  • Deleting Dataset
  1. User Guide

Datasets

PreviousShared FilesNextApps

Last updated 2 years ago

Datasets feature provides the capabilities of managing datasets resources such as create, update, and delete actions.

Datasets feature is based on to store dataset files with additional metadata. Group members can browse dataset files and upload files for sharing with other group members.

Prerequisites

Datasets feature is only enabled if is enabled.

Creating New Dataset

Click New Dataset to add a dataset and it will pop up the dialog of New Dataset.

Step 1: Name and Tags

  • Dataset Name: Name of the dataset (only the alphanumeric characters, hyphen -, or underscore _ are allowed).

  • Tags: Tags for the dataset.

Step 2: Uploading Files

Click or drop files to start uploading files. Once the files are uploaded, they will appear in the list of files.

Click Done to finish the upload.

All uploaded files are stored in the dataset.

Browse Dataset

Dataset List

  • Name: Name of the dataset.

  • Created By: User who created the dataset.

  • Last Modified: Last modified time of the dataset.

  • Tags: Tags for the dataset.

  • Delete button: Delete the dataset.

Dataset - Data Tab

The Data tab shows the list of files in the dataset.

Dataset - Information Tab

  • Dataset Name: Name of the dataset.

  • Created By: User who created the dataset.

  • Last Modified: Last modified time of the dataset.

  • Tags: Tags for the dataset.

  • Size: Size of the dataset.

  • Edit button: Edit the dataset.

Access from Jupyter

Open a Notebook and access the dataset from the path of phfs/datasets/<dataset_name>.

Editing Dataset

Go to the Information tab and click Edit to edit the dataset.

Deleting Dataset

In the Datasets list, click Delete button to delete the dataset.

Shared Files
PHFS