Skip to content

join sparkcmd

join sparkcmd #199

Workflow file for this run

# CodeCoverage result can be found at https://coveralls.io/github/HDFGroup/hermes
name: GitHub Actions
# Controls when the action will run. Triggers the workflow on push or pull request
# events but only for the master branch
on:
push:
pull_request:
branches: [ master ]
workflow_dispatch:
inputs:
debug_enabled:
description: 'Run the build with tmate debugging enabled'
required: false
default: false
env:
# Customize the CMake build type here (Release, Debug, RelWithDebInfo, etc.)
BUILD_TYPE: Debug
LOCAL: local
# A workflow run is made up of one or more jobs that can run sequentially or in parallel
jobs:
build:
# The type of runner that the job will run on
runs-on: ubuntu-20.04
# Steps represent a sequence of tasks that will be executed as part of the job
steps:
# Checks-out your repository under $GITHUB_WORKSPACE, so your job can access it
- name: Get Sources
uses: actions/checkout@v2
- name: Setup python
uses: actions/setup-python@v4
- name: Install Jarvis
run: bash ci/install_jarvis.sh
- name: Run pylint
run: bash ci/lint.sh
- name: Cache Spack packages
uses: actions/cache@v3
id: spack-cache
with:
path: |
~/spack
~/.spack
key: ${{ runner.os }}-${{ hashFiles('ci/install_deps.sh') }}
- name: Install APT Dependencies
run: bash ci/install_deps.sh
- name: Install Spack Dependencies
if: steps.spack-cache.outputs.cache-hit != 'true'
run: bash ci/install_spack.sh
- name: Test
run: bash ci/run_tests.sh
# - name: Multi-node Test
# run: pushd ci/cluster && ./multi_node_ci_test.sh
- name: Coveralls
uses: coverallsapp/github-action@master
with:
path-to-lcov: lcov.info
github-token: ${{ secrets.GITHUB_TOKEN }}