PrimeHub
  • Introduction
  • Installation
  • Tiers and Licenses
  • End-to-End Tutorial
    • 1 - MLOps Introduction and Scoping the Project
    • 2 - Train and Manage the Model
    • 3 - Compare, Register and Deploy the Model
    • 4 - Build the Web Application
    • 5 - Summary
  • User Guide
    • User Portal
    • Notebook
      • Notebook Tips
      • Advanced Settings
      • PrimeHub Notebook Extension
      • Submit Notebook as Job
    • Jobs
      • Job Artifacts
      • Tutorial
        • (Part1) MNIST classifier training
        • (Part2) MNIST classifier training
        • (Advanced) Use Job Submission to Tune Hyperparameters
        • (Advanced) Model Serving by Seldon
        • Job Artifacts Simple Usecase
    • Models
      • Manage and Deploy Model
      • Model Management Configuration
    • Deployments
      • Pre-packaged servers
        • TensorFlow server
        • PyTorch server
        • SKLearn server
        • Customize Pre-packaged Server
        • Run Pre-packaged Server Locally
      • Package from Language Wrapper
        • Model Image for Python
        • Model Image for R
        • Reusable Base Image
      • Prediction APIs
      • Model URI
      • Tutorial
        • Model by Pre-packaged Server
        • Model by Pre-packaged Server (PHFS)
        • Model by Image built from Language Wrapper
    • Shared Files
    • Datasets
    • Apps
      • Label Studio
      • MATLAB
      • MLflow
      • Streamlit
      • Tutorial
        • Create Your Own App
        • Create an MLflow server
        • Label Dataset by Label Studio
        • Code Server
    • Group Admin
      • Images
      • Settings
    • Generate an PrimeHub API Token
    • Python SDK
    • SSH Server Feature
      • VSCode SSH Notebook Remotely
      • Generate SSH Key Pair
      • Permission Denied
      • Connection Refused
    • Advanced Tutorial
      • Labeling the data
      • Notebook as a Job
      • Custom build the Seldon server
      • PrimeHub SDK/CLI Tools
  • Administrator Guide
    • Admin Portal
      • Create User
      • Create Group
      • Assign Group Admin
      • Create/Plan Instance Type
      • Add InfuseAI Image
      • Add Image
      • Build Image
      • Gitsync Secret for GitHub
      • Pull Secret for GitLab
    • System Settings
    • User Management
    • Group Management
    • Instance Type Management
      • NodeSelector
      • Toleration
    • Image Management
      • Custom Image Guideline
    • Volume Management
      • Upload Server
    • Secret Management
    • App Settings
    • Notebooks Admin
    • Usage Reports
  • Reference
    • Jupyter Images
      • repo2docker image
      • RStudio image
    • InfuseAI Images List
    • Roadmap
  • Developer Guide
    • GitHub
    • Design
      • PrimeHub File System (PHFS)
      • PrimeHub Store
      • Log Persistence
      • PrimeHub Apps
      • Admission
      • Notebook with kernel process
      • JupyterHub
      • Image Builder
      • Volume Upload
      • Job Scheduler
      • Job Submission
      • Job Monitoring
      • Install Helper
      • User Portal
      • Meta Chart
      • PrimeHub Usage
      • Job Artifact
      • PrimeHub Apps
    • Concept
      • Architecture
      • Data Model
      • CRDs
      • GraphQL
      • Persistence Storages
      • Persistence
      • Resources Quota
      • Privilege
    • Configuration
      • How to configure PrimeHub
      • Multiple Jupyter Notebook Kernels
      • Configure SSH Server
      • Configure Job Submission
      • Configure Custom Image Build
      • Configure Model Deployment
      • Setup Self-Signed Certificate for PrimeHub
      • Chart Configuration
      • Configure PrimeHub Store
    • Environment Variables
Powered by GitBook
On this page
  • Browse
  • Upload
  • Create/Update a Dataset
  1. User Guide

Shared Files

PreviousModel by Image built from Language WrapperNextDatasets

Last updated 2 years ago

The Shared Files feature is based on to store shared files. Group members are able to browse shared files and able to upload files for sharing with other group members.

Browse

It lists files and directories with columns of Name, Size, and Last Modified. The list can be sorted by these columns.

  • <Directory>/: navigate into the directory by clicking it.

Some directories are created by PrimeHub automatically, such as jobArtifacts/.

  • <File>: there are actions we can perform over files.

    • View file: view the file in an new tab if the file format is supported. E.g. image, text and notebook(.ipynb).

    • Download file

    • Copy PHFS URI: URI looks like phfs:///<path/to/file> (triple /); which can be used in Jobs.

    • Delete

Upload

Click Upload to pop up an uploader.

Click or drag files to start uploading them.

During uploading, the progress bar will show the progress of uploading.

We can click Upload in Background to move the uploading action to the background.

Once the uploading is done, the file will be listed in the Shared Files page.

Click OK to close the dialogue or click/drag to upload additional files.

Uploaded files can been only accessed by same group; please confirm the selected working group if files are not listed.

Create directory

Firstly, click the + icon.

Then input the name of sub-directory and click Confirm.

Example of typing the path/to/file.

The nested directories won't be created until a file is uploaded into, in other words, when showing No data, at this moment, no directories is created yet.

Create/Update a Dataset

We can directly choose some files or directories from the Shared Files list, it will pop-up dialogue to Create new dataset or Add to existing dataset.

Create new dataset

Select Create new dataset, it will pop-up dialogue and we can input the Dataset Name, specify Tags, and view the file contents.

  • Dataset Name: name of the dataset.

  • Tags: tags of the dataset.

  • Contents: the file contents of the current creation.

After specifying the values, click Create Dataset.

The dataset image-test is created successfully.

Add to existing dataset

In addition to creating a new dataset from the Shared Files, we can also add the files to an existing dataset.

Select Add to existing dataset, it will pop-up dialogue and we can select the existing dataset and view the file contents.

View selectable datasets. Here we select cv-test.

We can add files to the cv-test dataset directly or even add a sub-directory to the cv-test dataset by clicking the Add folder.

Input the sub-directory name view and click Confirm.

Ensure the update location and file contents are correct, click Add files to Dataset.

The dataset cv-test is updated successfully.

The Shared Files is the whole files storage and the is the specified collection of files that based on the Shared Files.

Datasets
PHFS Storage