🪴 jaden lorenc

Search

Search IconIcon to open search

Merkelbach Paper

Last updated Feb 27, 2024 Edit Source

#work/patientsim #paper_notes

# Merkelbach version

Based on the MIMIC dataset.

# data prep

# GRU

(it’s bidirectional) processes the whole of the data in both directions. Outputs are averaged over all the time steps, presumable leaving a single vector the length of the number of features.

# autoencoder

The single vector is put through a dense layer, making the bottleneck. That bottleneck is “repeated through time”, which I just think means it’s copy/pasted as many times as there are timesteps, with the related timestep features added on.

# decoder gru

Takes in the expanded bottleneck, turns it into another set (size of n timesteps) of vectors. A dense layer turns that into a copy of the original data.

# pic

Pasted image 20240125124146.png

# my issues

Triplet Sparsity Reduction is my proposed solution to those issues.