[Methods 2024]Automatic ICD-10-CM coding via Lambda-Scaled attention based deep learning model

论文网址:Automatic ICD-10-CM coding via Lambda-Scaled attention based deep learning model - ScienceDirect

英文是纯手打的!论文原文的summarizing and paraphrasing。可能会出现难以避免的拼写错误和语法错误,若有发现欢迎评论指正!文章偏向于笔记,谨慎食用

目录

1. 心得

2. 论文逐段精读

2.1. Abstract

2.2. Introduction

2.3. Methodology

2.3.1. Overview of workflow

2.3.2. Creating clinical Pool of liver transplant (CPLT) database

2.3.3. Model architecture

2.3.4. Web application deployment

2.4. Experiments

2.4.1. Datasets

2.4.2. Evaluation metrics

2.4.3. Parameter setting

2.5. Experimental results

2.5.1. Baseline models

2.5.2. Results

2.6. Discussion

2.7. Conclusion

1. 心得

(1)设计的比较简单

2. 论文逐段精读

2.1. Abstract

        ①Task: automatic International Classification of Diseases (ICD) (ICD-10-CM) coding

2.2. Introduction

        ①Version of ICD: ICD-9, ICD-10, ICD-11 etc.

        ②Challenge: lack of kownledge system will caused suboptimal diagnosis result

        ③Introduced relevant works

2.3. Methodology

2.3.1. Overview of workflow

        ①Pipeline:

2.3.2. Creating clinical Pool of liver transplant (CPLT) database

        ①The authors annotated MIMLT with 1380 ICD-9-CM samples then named it as “Clinical Pool of Liver Transplant” (CPLT)

        ②They transfer ICD-9-CM to ICD-10-CM by https://www.aapc.com/icd-10/codes/(啊,这可靠吗)

        ③They directly accepted one-to-one transfer but found experts to precisely classify the cases of one-to-many

2.3.3. Model architecture

(1)Embedding layer

        ①Employ Word2Vec to mapping original code/words w=\{w_{1},w_{2},\cdots,w_{i}\} in one clinical text \mathbb{C} to i vectors E=\{ew_1,ew_2,ew_3,\cdots,ew_i\}\in\mathbb{R}^{i\times d^c} with dimension of d^e=100 

(2)Deep bi-directional LSTM layer

       ①The processes of bi-LSTM:

\overrightarrow{h_f}=\overrightarrow{LSTM}\left(x_n,\overrightarrow{h_{n-1}}\right)

\overset{\leftarrow}{\operatorname*{h_b}}=\overset{\leftarrow}{\operatorname*{LSTM}}\left(x_n,\overset{\leftarrow}{\operatorname*{h_{n+1}}}\right)

h_{bi-lstm}=\overrightarrow{h_f}\oplus\overleftarrow{h_b}

where H=[h_1,h_2,h_3,\cdots,h_m]\in\mathbb{R}^{2q\times m}m是根据时间步自适应的?)

(3)Multi-scale CNN layer

        ①This MS-CNN is constructed by a max pooling layer and mean pooling layer after CNN 

        ②Concatenate all output from MS-CNN they get \mathcal{C}_{r}\in\mathbb{R}^{1\times\sum d_{k}^{r}}:

\mathcal{C}=[\mathcal{C}_1;\mathcal{C}_2;\mathcal{C}_3;\mathcal{C}_4;\mathcal{C}_5]

(4)Lambda-Scaled attention layer

        ①They further scale features:

\begin{aligned} & u_{j}=\text{tanh}(w_{o}\mathcal{C}_{r}+b_{o}) \\ & a_{t}=softmax(w_{a}u_{j}) \\ & a_{t}^{max}=max\left[a_{1},a_{2},a_{3},\cdots,a_{t}\right] \\ & a_{t}^{\prime}=\frac{a_{t}}{a_{t}^{max}};\quad\text{s.t.}a_{t}^{\prime}\leq1 \\ & T_{R}=\sum_{t=1}^{l}a_{t}^{\prime}\mathcal{C}_{r} \end{aligned}

(5)Classification

        ①Classification by fully connected layers with Sigmoid 

        ②Optimizer: Adam

        ③Binary cross entropy loss:

Loss=-\frac{1}{G}\sum_{l=1}^{G}y_{l}\odot\log\bar{y}_{l}+(1-y_{l})\odot\log\left(1-\bar{y}_{l}\right)

2.3.4. Web application deployment

        ①They build a web to predict ICD code

2.4. Experiments

2.4.1. Datasets

        ①Pretraining on MIMIC III and test on MIMIC III-Top 50 and CPLT

        ②Statistics:

(1)MIMIC III

        ①Sample: 53423

        ②Pre-training: stopword removal, tokenization, lowercase conversion, and removal of numbers, punctuation, and symbols by the Natural Language Toolkit (NLTK) library

        ③Limited record length: 2500

(2)CPLT

        ①Samples: 1380

        ②Preview: 

        ③Data split: 1104 for training, 138 for testing and 138 for val

        ④Max record length: 150

2.4.2. Evaluation metrics

        ①Micro F1 and Macro F1 for imbalanced data

2.4.3. Parameter setting

        ①Hidden dim of bi-LSTM: 64 for CPLT and 128 for MIMC III-Top 50

        ②Batch size: 8

        ③Epoch: 50

        ④Dropout rate: 0.5

2.5. Experimental results

2.5.1. Baseline models

        ~

        

2.5.2. Results

(1)Complete distribution of ICD-10-CM codes in the CPLT database

        ①34 ICD code in CPLT dataset:

(2)Comparison of DRCNN-ATT model with baselines on CPLT database

        ①Performance comparison table on CPLT:

(3)Comparison of DRCNN-ATT model with baselines on MIMIC III-Top 50 database

        ①Performance on MIMIC III-Top 50:

(4)Medical code Predictor web application

        ①Application preview:

(5)Ablation study

        ①Attention module ablation:

2.6. Discussion

        ~

2.7. Conclusion

        ~

评论
成就一亿技术人!
拼手气红包6.0元
还能输入1000个字符
 
红包 添加红包
表情包 插入表情
 条评论被折叠 查看
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值