Ntj

Meghanani 2021

Last edited: August 8, 2025

DOI: 10.3389/fcomp.2021.624558

One-Liner

analyzed spontaneous speech transcripts (only!) from TD and AD patients with fastText and CNN; best was \(83.33\%\) acc.

Novelty

  • threw the NLP kitchen sink to transcripts
    • fastText
    • CNN (with vary n-gram kernel 2,3,4,5 sizes)

Notable Methods

  • embeddings seaded by GloVe
  • fastText are much faster, but CNN won out

Key Figs

the qual results

PAR (participant), INV (investigator)

Notes

Hey look a review of the field:

Parvin 2020

Last edited: August 8, 2025

DOI: 10.3389/fnagi.2020.605317

One-Liner

An excercize scheme has had some measured effect on theta/alpha ratio and Brain wave frequency on AD patients; prognosis of AD not controlled for.

Novelty

  • Leveraged physical training scheme and measured EEG effects by quantifying theta/alpha ratio

Notable Methods

  • Used theta/alpha ratio as assay for improvement, and found the exercise scheme did so p<0.05
  • Only tested patients with AD w/o a control for stage

Key Figs

Figure 1

This figure tells us th N number of participants through the study

Sadeghian 2021

Last edited: August 8, 2025

DOI: 10.3389/fcomp.2021.624594

(Sadeghian, Schaffer, and Zahorian 2021)

One-Liner

Using a genetic algorithm, picked features to optimize fore; achieved \(94\%\) with just MMSE data alone (ok like duh me too). Developed ASR tool to aid.

Novelty

  • Developed an ASR methodology for speech, complete with punctuations
  • Used a genetic algorithm to do feature selection; NNs performed worse because “space is smaller???”

Notable Methods

Used a GRU to insert punctuations

The paper leveraged the nuke that is a bidirectional GRU, ATTENTION,

Shah 2021

Last edited: August 8, 2025

DOI: 10.3389/fcomp.2021.624659

One-Liner

Multi-feature late fusion of NLP results (by normalizing text and n-gram processing) with OpenSMILE embedding results.

Novelty

NLP transcript normalization (see methods) and OpenSMILE; otherwise similar to Martinc 2021. Same gist but different data-prep.

Notable Methods

  • N-gram processed the input features
  • Used WordNet to replace words with roots

Key Figs

New Concepts

Yuan 2021

Last edited: August 8, 2025

DOI: 10.3389/fcomp.2020.624488

One-Liner

Used an ERNIE trained on transcripts for classification; inclusion of pause encoding made results better.

Novelty

  • Instead of just looking at actual speech content, look at pauses specific as a feature engineering task
  • \(89.6\%\) on the ADReSS Challenge dataset

Notable Methods

Applied FA with pause encoding with standard .cha semantics (short pauses, medium pauses, long pauses). Shoved all of this into an ERNIE.

Assay for performance was LOO