Skip to content

Mutinifni/hb_starlite

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

19 Commits
 
 
 
 
 
 

Repository files navigation

Starlite: A Programmability Testing Environment for HammerBlade

This repository contains a setup for evaluating the programming infrastructure for the HammerBlade project. The main component is a Docker container setup that includes TVM, PyTorch, and GraphIt. There are also tools for estimating the energy consumption of programs implemented using this infrastructure.

Running the Container

To run the container, use our launch script to pull it from Docker Hub:

$ ./docker/bash.sh mutinifni/sdh-tvm-pytorch:latest

You might need root permissions based on your local Docker setup (so try this with sudo if it doesn't work without it).

You can also see the Dockerfile source for details on how this is set up. For other (standard) containers for TVM (including GPU support), see the official Dockerfiles, and related instructions.

Development

Inside the container, you can use PyTorch, TVM, and GraphIt.

PyTorch

For machine learning and dense linear algebra development, we recommend you use PyTorch. To get started, begin with the "60-minute blitz" PyTorch tutorial. Then, depending on the task you want to accomplish, explore the full set of PyTorch tutorials to find examples that are closest to your domain.

TVM

You can also opt to use TVM directly, especially via its MXNet frontend. See the extensive list of tutorials that come with TVM.

GraphIt

To develop graph processing kernels, use GraphIt. Begin by following the Getting Started guide, which walks you through the implementation of the PageRank-Delta algorithm. (You can skip the initial setup instructions; the compiler is already installed in the container for you.) Then, check out the language manual for more details.

HBPintool

The tool only supports being run from the source root of hbpintool (/hbpintool/ in this image).

From the hbpintool directory run these commands to profile your GraphIt program:

  1. source SOURCE_THIS
  2. ./hbpintool /path/to/your/graphit/program [your-graphit-program-arguments]

You should find the output in a file called hbpintool.out

Preventing Compiler Inlining

For accuracy the hbpintool relies on the GraphIt generated edgeset_apply function being defined in the program executable.

However, depending on how the GraphIt C++ output was compiled, this function may have been inlined.

To keep this from happening, open the .cpp file generated by graphitc.py and add __attribute__((noinline)) to the definition of the templated edgeset_apply function (the exact name depends on the GraphIt schedules used).

Serial Code Only

This tool is only written to profile serial code (Don't compile the GraphIt program with -fopenmp). The models it uses to calculate energy factors in the parallelism for the relevant hardware.

Intel 64

This tool can only run on Intel x86_64 processors.

Profiling

We recommend that testers use the PyTorch frontend for development. Then, to profile an implementation you've created, we recommend you use TVM. To do so, you will need to export the PyTorch model as ONNX and import into TVM. Follow these steps:

  1. Export the model. Use torch.onnx.export to save your model to an ONNX file.
  2. Import the model into TVM and compile. See the TVM tutorial about importing ONNX models.

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Contributors 3

  •  
  •  
  •