[2507.06996] Generating Multi-Table Time Series EHR from Latent Space with Minimal Preprocessing

[2507.06996] Generating Multi-Table Time Series EHR from Latent Space with Minimal Preprocessing

arXiv - Machine Learning 3 min read

About this article

Abstract page for arXiv paper 2507.06996: Generating Multi-Table Time Series EHR from Latent Space with Minimal Preprocessing

Computer Science > Machine Learning arXiv:2507.06996 (cs) [Submitted on 9 Jul 2025 (v1), last revised 2 Mar 2026 (this version, v2)] Title:Generating Multi-Table Time Series EHR from Latent Space with Minimal Preprocessing Authors:Eunbyeol Cho, Jiyoun Kim, Minjae Lee, Sungjin Park, Edward Choi View a PDF of the paper titled Generating Multi-Table Time Series EHR from Latent Space with Minimal Preprocessing, by Eunbyeol Cho and 4 other authors View PDF HTML (experimental) Abstract:Electronic Health Records (EHR) are time-series relational databases that record patient interactions and medical events over time, serving as a critical resource for healthcare research and applications. However, privacy concerns and regulatory restrictions limit the sharing and utilization of such sensitive data, necessitating the generation of synthetic EHR datasets. Unlike previous EHR synthesis methods, which typically generate medical records consisting of expert-chosen features (e.g. a few vital signs or structured codes only), we introduce RawMed, the first framework to synthesize multi-table, time-series EHR data that closely resembles raw EHRs. Using text-based representation and compression techniques, RawMed captures complex structures and temporal dynamics with minimal preprocessing. We also propose a new evaluation framework for multi-table time-series synthetic EHRs, assessing distributional similarity, inter-table relationships, temporal dynamics, and privacy. Validated on two open...

Originally published on March 03, 2026. Curated by AI News.

Related Articles

Machine Learning

[D] MXFP8 GEMM: Up to 99% of cuBLAS performance using CUDA + PTX

New blog post by Daniel Vega-Myhre (Meta/PyTorch) illustrating GEMM design for FP8, including deep-dives into all the constraints and des...

Reddit - Machine Learning · 1 min ·
UMKC Announces New Master of Science in Artificial Intelligence
Ai Infrastructure

UMKC Announces New Master of Science in Artificial Intelligence

UMKC announces a new Master of Science in Artificial Intelligence program aimed at addressing workforce demand for AI expertise, set to l...

AI News - General · 4 min ·
[2603.15159] To See is Not to Master: Teaching LLMs to Use Private Libraries for Code Generation
Llms

[2603.15159] To See is Not to Master: Teaching LLMs to Use Private Libraries for Code Generation

Abstract page for arXiv paper 2603.15159: To See is Not to Master: Teaching LLMs to Use Private Libraries for Code Generation

arXiv - AI · 4 min ·
[2602.07374] TernaryLM: Memory-Efficient Language Modeling via Native 1.5-Bit Quantization with Adaptive Layer-wise Scaling
Llms

[2602.07374] TernaryLM: Memory-Efficient Language Modeling via Native 1.5-Bit Quantization with Adaptive Layer-wise Scaling

Abstract page for arXiv paper 2602.07374: TernaryLM: Memory-Efficient Language Modeling via Native 1.5-Bit Quantization with Adaptive Lay...

arXiv - AI · 4 min ·
More in Ai Infrastructure: This Week Guide Trending

No comments

No comments yet. Be the first to comment!

Stay updated with AI News

Get the latest news, tools, and insights delivered to your inbox.

Daily or weekly digest • Unsubscribe anytime