Skip to content

v1.2

Compare
Choose a tag to compare
@anzarabbas anzarabbas released this 15 Jun 12:46
· 696 commits to main since this release
7d72bbe

OpenWillis v1.2

Release date: June 14th, 2023

The v1.2 release improves OpenWillis’ speech analysis capabilities and improves processing workflows.

If you have feedback or questions, please do reach out.

Contributors

General updates

  1. For better accessibility, all method description documentation has been moved from Google Docs to the repo’s wiki––a much more appropriate place for it.
  2. The example uses from the notebook included in the code have been moved to the same methods description documents in the wiki, consolidating this information in one place.

Repository updates

We have restructured the folder organization: Functions are now categorized based on the modality of data they process. This will feel more intuitive to independent contributors.

Function updates

We've separated speech transcription into two functions:

  1. Speech transcription v1.1: This uses locally executable models for speech transcription, maintaining the functionality of the previous version of the same method.
  2. Speech transcription cloud v1.0: This new function uses cloud-based models for speech transcription, specifically incorporating Amazon Transcribe. Users must input their own AWS credentials for this. A notable feature of this version is its ability to label speakers in a dual-speaker audio file. In the case of clinical interview recordings, speakers can also be identified as 'clinician' or 'participant', with these labels included in the outputted JSON.

The speaker separation function has been updated to accommodate both transcription workflows:

  1. The locally executable models that separate speakers remain the same, the difference being that they use the JSON output from the speech transcription v1.1 function for improved efficiency.
  2. For when the user employs the speech transcription cloud v1.0 function to get a JSON with speaker labels included, the speaker separation function can simply use those labels to separate the audio into individual files for each speaker. This is a much faster option.

In response to these function modifications, we are also releasing speech characteristics v1.1, which enables users to choose which speaker they wish to calculate speech characteristics from thanks to the labeling included in the output JSON file from the cloud-based speech transcription function.