<?xml version="1.0" encoding="UTF-8"?><rss xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:atom="http://www.w3.org/2005/Atom" version="2.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0"><channel><title><![CDATA[Standard Model]]></title><description><![CDATA[Standard Model]]></description><link>https://blog.standardmodel.bio</link><image><url>https://substackcdn.com/image/fetch/$s_!M-rU!,w_256,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F515c2dee-6942-4f20-ae48-491f0d4cc29a_817x817.png</url><title>Standard Model</title><link>https://blog.standardmodel.bio</link></image><generator>Substack</generator><lastBuildDate>Wed, 08 Apr 2026 14:04:32 GMT</lastBuildDate><atom:link href="https://blog.standardmodel.bio/feed" rel="self" type="application/rss+xml"/><copyright><![CDATA[Standard Model Biomedicine, Inc.]]></copyright><language><![CDATA[en]]></language><webMaster><![CDATA[info@standardmodel.bio]]></webMaster><itunes:owner><itunes:email><![CDATA[info@standardmodel.bio]]></itunes:email><itunes:name><![CDATA[Arda Pekis]]></itunes:name></itunes:owner><itunes:author><![CDATA[Arda Pekis]]></itunes:author><googleplay:owner><![CDATA[info@standardmodel.bio]]></googleplay:owner><googleplay:email><![CDATA[info@standardmodel.bio]]></googleplay:email><googleplay:author><![CDATA[Arda Pekis]]></googleplay:author><itunes:block><![CDATA[Yes]]></itunes:block><item><title><![CDATA[The Patient is Not a Document: Moving from LLMs to a World Model for Oncology (Part 2)]]></title><description><![CDATA[We Tested Our Model with Memorial Sloan Kettering Cancer Center&#8217;s (MSK&#8217;s) Data while Participating in the MSK iHub Challenge 2025 Cohort Program, and We&#8217;re Releasing the Model Weights to the Public]]></description><link>https://blog.standardmodel.bio/p/the-patient-is-not-a-document-moving-f88</link><guid isPermaLink="false">https://blog.standardmodel.bio/p/the-patient-is-not-a-document-moving-f88</guid><dc:creator><![CDATA[Zach Chen]]></dc:creator><pubDate>Wed, 17 Dec 2025 15:51:55 GMT</pubDate><enclosure url="https://substack-post-media.s3.amazonaws.com/public/images/ec36a4bb-02e0-4978-a067-66fed3016d3c_3600x1881.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>TL;DR: We validated our Standard Model on real-world oncology data from Memorial Sloan Kettering Cancer Center (MSK) during the 2025 iHub Challenge Cohort program, proving that a specialized architecture designed to simulate biological dynamics </strong><em><strong>S(t)</strong></em><strong> delivers superior predictive performance on complex, non-linear clinical tasks&#8212;and <a href="https://huggingface.co/standardmodelbio/SMB-v1-1.7B-Structure">we&#8217;re making those model weights available today</a>.</strong></p><p>In Part 1, we introduced our <strong><a href="https://blog.standardmodel.bio/p/the-patient-is-not-a-document-moving">Standard Model v1</a></strong>, our biological world model built on top of the <a href="https://arxiv.org/pdf/2301.08243">Joint-Embedding Predictive Architecture</a> (JEPA). The core hypothesis of this design is that by training on cause-and-effect pairs (State at time <em>t</em> + Intervention &#8594; State <em>t+1</em>), the model learns a high-fidelity <em>patient state embedding</em> that captures the underlying biological dynamics of the disease.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!n38D!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff29fcbd6-bb22-47a5-b44a-cd882ea2a286_4962x3351.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!n38D!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff29fcbd6-bb22-47a5-b44a-cd882ea2a286_4962x3351.png 424w, https://substackcdn.com/image/fetch/$s_!n38D!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff29fcbd6-bb22-47a5-b44a-cd882ea2a286_4962x3351.png 848w, https://substackcdn.com/image/fetch/$s_!n38D!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff29fcbd6-bb22-47a5-b44a-cd882ea2a286_4962x3351.png 1272w, https://substackcdn.com/image/fetch/$s_!n38D!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff29fcbd6-bb22-47a5-b44a-cd882ea2a286_4962x3351.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!n38D!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff29fcbd6-bb22-47a5-b44a-cd882ea2a286_4962x3351.png" width="1456" height="983" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/f29fcbd6-bb22-47a5-b44a-cd882ea2a286_4962x3351.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:983,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1235641,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://blog.standardmodel.bio/i/181196973?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff29fcbd6-bb22-47a5-b44a-cd882ea2a286_4962x3351.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!n38D!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff29fcbd6-bb22-47a5-b44a-cd882ea2a286_4962x3351.png 424w, https://substackcdn.com/image/fetch/$s_!n38D!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff29fcbd6-bb22-47a5-b44a-cd882ea2a286_4962x3351.png 848w, https://substackcdn.com/image/fetch/$s_!n38D!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff29fcbd6-bb22-47a5-b44a-cd882ea2a286_4962x3351.png 1272w, https://substackcdn.com/image/fetch/$s_!n38D!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Ff29fcbd6-bb22-47a5-b44a-cd882ea2a286_4962x3351.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: <strong>The Standard Model Architecture</strong>. It utilizes a Joint-Embedding Predictive Architecture (JEPA) to conditionally predict patient trajectories within the latent space rather than classifying static snapshots.</figcaption></figure></div><p>However, model validation presents a nuance. Logic suggests we should ask the model to generate the &#8220;future trajectory&#8221; (e.g. predicted text or synthesize a follow-up CT scan) and judge the output. But our Standard Model projects the patient journey <strong>in the latent space</strong>, not in the pixel space. Therefore, evaluating based on raw signal generation is a category mismatch.</p><p><strong>Instead</strong><em><strong> </strong></em><strong>of</strong><em><strong> judging raw signal generation</strong></em><strong>, our Standard Model tests the quality of the patient representation itself.</strong></p><p>In a functioning World Model, <em>S(t)</em> is not just a summary of the past; it is a predictive vector that explicitly encodes the trajectory of the future. To verify this, we use a <strong>frozen encoder</strong> protocol: we freeze the model&#8217;s weights and attach a simple linear probe. Think of this linear probe as a &#8220;truth serum.&#8221; It is too mathematically simple to learn complex disease patterns on its own. For the probe to succeed, the prognosis must already be structured within the patient&#8217;s embedding.</p><h3><strong>The Proving Ground: The MSKCC Oncology Cohort</strong></h3><p>To stress-test a World Model, one needs more than just &#8220;labels&#8221;; one needs high-density longitudinal history. We utilized a massive, high-fidelity cohort from Memorial Sloan Kettering Cancer Center (MSKCC) during the <a href="https://www.mskcc.org/commercialization/programs-accelerators/msk-innovation-hub/msk-ihub-challenge">iHub Challenge</a> comprising <strong>23,319 patients</strong> and over <strong>323,000 patient-years</strong> of data.</p><p>What makes this dataset unique for a JEPA-based architecture is its <strong>temporal density</strong>. With a median follow-up of over <strong>50 months</strong> and an average of <strong>127 clinical events per patient</strong>, the dataset provides a &#8220;high-resolution&#8221; movie of cancer progression rather than a grainy snapshot. The coverage is near-universal: 100% of patients have confirmed pathology and outcomes, and 95.6% have deep genomic biomarker data. This depth allows the Standard Model to move beyond simple text-matching and actually model the interaction between systemic therapies (94.8% coverage) and the biological state of the tumor.</p><h3><strong>Evaluation on MSK&#8217;s Cancer Dataset</strong></h3><p>Our evaluation pipeline is designed to prove that the model has internalized clinical logic rather than just memorized data.</p><h4><strong>The Workflow: From History to Embedding</strong> </h4><ul><li><p><strong>Generate:</strong> We feed the de-identified patient&#8217;s longitudinal history up to a specific time point <em>t</em> into the model. Using our <a href="https://github.com/standardmodelbio/smb-biopan-utils">utility function</a>, we convert raw clinical events into a chronological narrative (e.g., [2024-01-10]: Glucose: 110. [2024-01-12]: Dyspnea, etc.).</p></li></ul><ul><li><p><strong>Embed:</strong> The model acts as a <strong>Frozen Backbone</strong> and outputs a single, fused patient state embedding <em>S(t)</em> to ensure we are testing the <em>existing</em> representation.</p></li><li><p><strong>Probe:</strong> We train a simple, lightweight linear survival head (e.g., using <a href="https://arxiv.org/abs/1606.00931">CoxPH loss</a> as learning objective) on top of this embedding to predict future outcomes.</p></li></ul><h4><strong>The Methodology: &#8220;Point-in-Time&#8221; Construction</strong></h4><blockquote><p><em>Oncology is not a static classification problem; it is a trajectory of shifting risk.</em> </p></blockquote><p>A patient&#8217;s prognosis changes dramatically between diagnosis, remission, and recurrence. To capture this, we moved beyond &#8220;one-patient-one-label&#8221; and adopted a <strong>point-in-time </strong>evaluation framework.</p><ul><li><p><strong>Selecting Indexing Dates (</strong><em><strong>t=0</strong></em><strong>): </strong>We slice a patient&#8217;s timeline into multiple examples based on clinically significant <em>decision nodes</em>. These are moments where the state of the patient undergoes a potential transition, and where a World Model&#8217;s predictive capability is most critical. At each indexing date, we mask all future data. The model must predict the trajectory solely based on the history available at <em>t=0</em>. Our generator scans the MSKCC patients&#8217; event logs for five specific triggers to establish an indexing date:</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2" target="_blank" href="https://substackcdn.com/image/fetch/$s_!VSLJ!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89097330-a7f6-4030-b7a5-5053359a4f43_4860x1504.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!VSLJ!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89097330-a7f6-4030-b7a5-5053359a4f43_4860x1504.png 424w, https://substackcdn.com/image/fetch/$s_!VSLJ!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89097330-a7f6-4030-b7a5-5053359a4f43_4860x1504.png 848w, https://substackcdn.com/image/fetch/$s_!VSLJ!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89097330-a7f6-4030-b7a5-5053359a4f43_4860x1504.png 1272w, https://substackcdn.com/image/fetch/$s_!VSLJ!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89097330-a7f6-4030-b7a5-5053359a4f43_4860x1504.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!VSLJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89097330-a7f6-4030-b7a5-5053359a4f43_4860x1504.png" width="728" height="225.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/89097330-a7f6-4030-b7a5-5053359a4f43_4860x1504.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:451,&quot;width&quot;:1456,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:441179,&quot;alt&quot;:&quot;wew&quot;,&quot;title&quot;:&quot;wew&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.standardmodel.bio/i/181196973?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F0410b7fa-0e78-4dd4-b3cf-188c7b89432f_4860x1767.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-normal" alt="wew" title="wew" srcset="https://substackcdn.com/image/fetch/$s_!VSLJ!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89097330-a7f6-4030-b7a5-5053359a4f43_4860x1504.png 424w, https://substackcdn.com/image/fetch/$s_!VSLJ!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89097330-a7f6-4030-b7a5-5053359a4f43_4860x1504.png 848w, https://substackcdn.com/image/fetch/$s_!VSLJ!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89097330-a7f6-4030-b7a5-5053359a4f43_4860x1504.png 1272w, https://substackcdn.com/image/fetch/$s_!VSLJ!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89097330-a7f6-4030-b7a5-5053359a4f43_4860x1504.png 1456w" sizes="100vw" loading="lazy"></picture><div></div></div></a><figcaption class="image-caption">Table 1: <strong>Indexing Date Triggers</strong>. The generator scans patient event logs for these five specific clinical triggers to establish a <em>decision node </em>(t=0) where future data is masked.</figcaption></figure></div><ul><li><p><strong>The Downstream Tasks (The &#8220;Future&#8221;): </strong>To ensure the embeddings <em>S(t)</em> are robust, we evaluate them against five distinct clinical dimensions as below using the linear probe:</p></li></ul><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!mh-8!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904d8fd7-cb63-47f1-aa1f-c82494941ff9_7446x1902.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!mh-8!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904d8fd7-cb63-47f1-aa1f-c82494941ff9_7446x1902.png 424w, https://substackcdn.com/image/fetch/$s_!mh-8!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904d8fd7-cb63-47f1-aa1f-c82494941ff9_7446x1902.png 848w, https://substackcdn.com/image/fetch/$s_!mh-8!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904d8fd7-cb63-47f1-aa1f-c82494941ff9_7446x1902.png 1272w, https://substackcdn.com/image/fetch/$s_!mh-8!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904d8fd7-cb63-47f1-aa1f-c82494941ff9_7446x1902.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!mh-8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904d8fd7-cb63-47f1-aa1f-c82494941ff9_7446x1902.png" width="1456" height="372" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/904d8fd7-cb63-47f1-aa1f-c82494941ff9_7446x1902.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:372,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1004030,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.standardmodel.bio/i/181196973?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fb6145e8e-d7c2-4fb1-abc4-3f61c8d544ef_7446x2160.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!mh-8!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904d8fd7-cb63-47f1-aa1f-c82494941ff9_7446x1902.png 424w, https://substackcdn.com/image/fetch/$s_!mh-8!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904d8fd7-cb63-47f1-aa1f-c82494941ff9_7446x1902.png 848w, https://substackcdn.com/image/fetch/$s_!mh-8!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904d8fd7-cb63-47f1-aa1f-c82494941ff9_7446x1902.png 1272w, https://substackcdn.com/image/fetch/$s_!mh-8!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F904d8fd7-cb63-47f1-aa1f-c82494941ff9_7446x1902.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Table 2: <strong>Downstream Clinical Evaluation Tasks</strong>. To validate the robustness of the Point-in-Time embeddings, we evaluate performance across five distinct clinical dimensions using a linear probe.</figcaption></figure></div><ul><li><p><strong>Task Filtering: </strong>We don&#8217;t just test on every available label. To ensure statistical power, our framework only evaluates tasks that meet a high bar: a minimum of 500 total samples and at least 50 positive cases, ensuring the model isn&#8217;t just &#8220;guessing&#8221; on rare events.</p></li></ul><h3><strong>Ensuring Scientific Rigor</strong></h3><p>Evaluating on Real-World Data (RWD) introduces risks of data leakage that do not exist in curated academic datasets. To ensure our results reflected genuine predictive power, we implemented strict guardrails:</p><ul><li><p><strong>The Patient-Level Split: </strong>A common pitfall in clinical AI is splitting data by <em>records</em>. If a patient has ten visits, and eight are in training while two are in testing, the model might &#8220;cheat&#8221; by recognizing that specific patient&#8217;s ID or baseline. We perform a strict <strong>Patient-Level Split</strong>. We shuffle the unique <em>subject ID</em> list and allocate 85% of patients for training and 15% for a completely &#8220;unseen&#8221; test set. If a patient is in the test set, the model has never seen a single day of their medical history during training.</p></li></ul><ul><li><p><strong>Focusing on Progression:</strong> To ensure our predictions reflect true changes in health status, we remove data from the week immediately preceding a mortality. This prevents the system from basing its predictions on the administrative processes that occur during end-of-life care.</p></li><li><p><strong>Strict Temporal Separation:</strong> To prevent intraday leakage (e.g., a toxicity event recorded at 4:00 PM influencing a prediction made at 9:00 AM on the same day), we enforce a 24-hour buffer. The target window explicitly begins the day <em>after</em> the Indexing Date.</p></li><li><p><strong>Multi-Dimensional Metric Breakdowns:</strong> We don&#8217;t settle for a single AUROC number. Our framework automatically &#8220;slices&#8221; performance across two dimensions to find where the model excels or fails:</p><ul><li><p><strong>By Task:</strong> Comparing how it predicts mortality vs. line-of-therapy changes.</p></li><li><p><strong>By Indication:</strong> Specifically analyzing high-heterogeneity diseases like Sarcoma vs. more &#8220;structured&#8221; cancers like Prostate.</p></li></ul></li></ul><h3><strong>Evidence</strong></h3><p>To test this, we benchmarked the Standard Model (SMB-v1-1.7B) against three distinct classes of baselines:</p><ul><li><p><strong>Classic Clinical Baselines:</strong> The industry standards - Logistic Regression, Random Forests, and Gradient Boosting.</p></li><li><p><strong>General-Purpose Multimodal LLMs:</strong> <a href="https://arxiv.org/abs/2511.21631">Qwen3-VL</a> (4B and 8B), to test if general-purpose AI models with massive scale can substitute for domain specialization.</p></li><li><p><strong>Internal Controls (Ablations):</strong></p><ul><li><p><em>SMB-EHR-4B:</em> Trained only on <strong>public EHR data</strong>.</p></li><li><p><em>SMB-v1-1.7B (sft):</em> Trained <strong>without the JEPA objective</strong> (supervised fine-tuning only).</p></li></ul></li></ul><p>Here is what the data reveals.</p><h4><strong>1. A Structural Break in Clinical Reasoning</strong></h4><p>The most compelling evidence for the World Model hypothesis is found not in the aggregate scores, but in <em>where</em> the model wins.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!rVTg!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F464b6609-72c4-457f-91c0-93a2fa2c0115_7938x3156.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!rVTg!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F464b6609-72c4-457f-91c0-93a2fa2c0115_7938x3156.png 424w, https://substackcdn.com/image/fetch/$s_!rVTg!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F464b6609-72c4-457f-91c0-93a2fa2c0115_7938x3156.png 848w, https://substackcdn.com/image/fetch/$s_!rVTg!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F464b6609-72c4-457f-91c0-93a2fa2c0115_7938x3156.png 1272w, https://substackcdn.com/image/fetch/$s_!rVTg!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F464b6609-72c4-457f-91c0-93a2fa2c0115_7938x3156.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!rVTg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F464b6609-72c4-457f-91c0-93a2fa2c0115_7938x3156.png" width="1456" height="579" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/464b6609-72c4-457f-91c0-93a2fa2c0115_7938x3156.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:579,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:319995,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:&quot;&quot;,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.standardmodel.bio/i/181196973?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F464b6609-72c4-457f-91c0-93a2fa2c0115_7938x3156.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!rVTg!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F464b6609-72c4-457f-91c0-93a2fa2c0115_7938x3156.png 424w, https://substackcdn.com/image/fetch/$s_!rVTg!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F464b6609-72c4-457f-91c0-93a2fa2c0115_7938x3156.png 848w, https://substackcdn.com/image/fetch/$s_!rVTg!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F464b6609-72c4-457f-91c0-93a2fa2c0115_7938x3156.png 1272w, https://substackcdn.com/image/fetch/$s_!rVTg!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F464b6609-72c4-457f-91c0-93a2fa2c0115_7938x3156.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 2: <strong>Performance Decomposition by Clinical Task Category</strong>.</figcaption></figure></div><p>Figure 2 reveals a clear &#8220;staircase&#8221; of performance that correlates with biological complexity:</p><ul><li><p><strong>Static Tasks (e.g., Mortality):</strong> On simple outcomes like death, generalist models are competitive.</p></li><li><p><strong>Dynamic Tasks (e.g., Treatment Change):</strong> As soon as the task requires understanding the vector of the disease, e.g., predicting disease progression or if a treatment will fail within a year, the generalist models collapse.</p><ul><li><p>Gradient Boosting struggles at <strong>~0.66</strong>.</p></li><li><p>Qwen3-VL improves slightly to <strong>~0.70</strong>.</p></li><li><p><strong>SMB-v1-1.7B (SFT+JEPA) dominates with an AUROC approaching 0.78.</strong></p></li></ul></li></ul><p><strong>This confirms our hypothesis.</strong> The SMB-v1 thrives here because it isn&#8217;t just looking at the patient&#8217;s current condition; it is simulating the treatment&#8217;s collision with the tumor&#8217;s trajectory.</p><p><strong>2. Validating the &#8220;World Model&#8221; Architecture (SFT vs. JEPA)</strong></p><p>To prove that our performance is a result of the architecture and not just the data, we ran an ablation study comparing pure supervised fine-tuning (SFT) against our hybrid architecture.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!GB-R!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F482f4a0a-1220-4c77-bda8-e603b19d4a7e_11172x3408.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!GB-R!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F482f4a0a-1220-4c77-bda8-e603b19d4a7e_11172x3408.png 424w, https://substackcdn.com/image/fetch/$s_!GB-R!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F482f4a0a-1220-4c77-bda8-e603b19d4a7e_11172x3408.png 848w, https://substackcdn.com/image/fetch/$s_!GB-R!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F482f4a0a-1220-4c77-bda8-e603b19d4a7e_11172x3408.png 1272w, https://substackcdn.com/image/fetch/$s_!GB-R!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F482f4a0a-1220-4c77-bda8-e603b19d4a7e_11172x3408.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!GB-R!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F482f4a0a-1220-4c77-bda8-e603b19d4a7e_11172x3408.png" width="1456" height="444" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/482f4a0a-1220-4c77-bda8-e603b19d4a7e_11172x3408.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:444,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:2681312,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.standardmodel.bio/i/181196973?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F482f4a0a-1220-4c77-bda8-e603b19d4a7e_11172x3408.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!GB-R!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F482f4a0a-1220-4c77-bda8-e603b19d4a7e_11172x3408.png 424w, https://substackcdn.com/image/fetch/$s_!GB-R!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F482f4a0a-1220-4c77-bda8-e603b19d4a7e_11172x3408.png 848w, https://substackcdn.com/image/fetch/$s_!GB-R!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F482f4a0a-1220-4c77-bda8-e603b19d4a7e_11172x3408.png 1272w, https://substackcdn.com/image/fetch/$s_!GB-R!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F482f4a0a-1220-4c77-bda8-e603b19d4a7e_11172x3408.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 3: <strong>Benchmark Comparison across Model Architectures</strong>.</figcaption></figure></div><ul><li><p><strong>The Baseline (SFT Only):</strong> The SMB-v1-1.7B (sft) model achieves a strong overall AUROC of <strong>0.715</strong>.</p></li><li><p><strong>The Breakthrough (SFT + JEPA):</strong> When we add the JEPA objective that forces the model to predict future patient states, performance jumps to <strong>0.727</strong>.</p></li></ul><p>However, the aggregate score hides the true signal. As the error bars in Figure 3 reveal, the lift is not uniform. It is specific to complexity.</p><p>In relatively homogenous indications where text patterns correlate strongly with outcomes (e.g., ovarian cancer), the SFT model is already near the ceiling. The JEPA architecture shines where traditional pattern-matching fails: in highly heterogeneous, aggressive diseases.</p><ul><li><p><strong>Sarcoma (the </strong><em><strong>heterogeneity</strong></em><strong> test):</strong> Sarcomas are notoriously diverse and difficult to subtype from text alone. Here, the SFT model struggles (~0.71), but the JEPA model delivers a massive lift to <strong>~0.77</strong>. The fact that JEPA lifts performance here suggests the model is ignoring the messy syntax of the notes and successfully embedding the underlying <em>phenotype</em>. It is learning &#8220;sarcoma-ness&#8221; from the trajectory of vitals/labs/imaging, rather than the text label &#8220;sarcoma.&#8221;</p></li></ul><p><strong>Upper-GI &amp; Prostate:</strong> We see similarly distinct gains in upper-GI and prostate cancers. In these indications, the disease trajectory is often non-linear. The SFT model creates a static risk score, but the JEPA model successfully simulates the vector of the disease, resulting in superior stratification.</p><h4><strong>3. Specialization Beats Scale (The &#8220;Generalist&#8221; Trap)</strong></h4><p>A common assumption is that massive general-purpose models will eventually render specialized models obsolete. Our results suggest otherwise: domain grounding matters more than parameter count.</p><ul><li><p><strong>The Generalist:</strong> Qwen3-VL 8B achieves an Overall AUROC of <strong>0.687</strong>.</p></li><li><p><strong>The Public Specialist:</strong> Interestingly, our SMB-EHR-4B, which trained only on public data, surpasses the generalist with an AUROC of <strong>0.708</strong>.</p></li></ul><p><strong>Why this happens:</strong> Qwen3-VL sees a lab value of &#8220;Creatinine: 1.8&#8221; as a sequence of text tokens to predict. The Standard Model understands it as a <strong>biological signal of renal function</strong> that interacts with hydration and chemotherapy toxicity.</p><h4><strong>4. Deep Dive: Granularity Beyond the Aggregate</strong></h4><p>Aggregate AUROC scores can obscure clinical nuance. Does the model actually understand anatomy, or is it just guessing risk?</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nauS!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e33ec4-2707-48b1-a6a0-a19153060fe0_5034x5100.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nauS!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e33ec4-2707-48b1-a6a0-a19153060fe0_5034x5100.png 424w, https://substackcdn.com/image/fetch/$s_!nauS!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e33ec4-2707-48b1-a6a0-a19153060fe0_5034x5100.png 848w, https://substackcdn.com/image/fetch/$s_!nauS!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e33ec4-2707-48b1-a6a0-a19153060fe0_5034x5100.png 1272w, https://substackcdn.com/image/fetch/$s_!nauS!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e33ec4-2707-48b1-a6a0-a19153060fe0_5034x5100.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nauS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e33ec4-2707-48b1-a6a0-a19153060fe0_5034x5100.png" width="1456" height="1475" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/69e33ec4-2707-48b1-a6a0-a19153060fe0_5034x5100.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:1475,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1744717,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.standardmodel.bio/i/181196973?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F390eb822-887d-41b9-8d89-d4306d9ec30d_5034x5253.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!nauS!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e33ec4-2707-48b1-a6a0-a19153060fe0_5034x5100.png 424w, https://substackcdn.com/image/fetch/$s_!nauS!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e33ec4-2707-48b1-a6a0-a19153060fe0_5034x5100.png 848w, https://substackcdn.com/image/fetch/$s_!nauS!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e33ec4-2707-48b1-a6a0-a19153060fe0_5034x5100.png 1272w, https://substackcdn.com/image/fetch/$s_!nauS!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F69e33ec4-2707-48b1-a6a0-a19153060fe0_5034x5100.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 4: <strong>Model Performance Benchmark for Pancreatic Cancer</strong>.</figcaption></figure></div><p>Aggregate AUROC scores can obscure clinical nuance. Does the model actually understand anatomy, or is it just guessing risk?</p><p>Using pancreas cancer as a representative demonstration, we broke down performance into specific anatomical tasks.</p><ul><li><p><strong>Anatomical Specificity:</strong> The model successfully differentiates between specific progression sites. <strong>For</strong> <strong>progression in the body of the pancreas</strong>, the Standard Model provides a <strong>+12.6% improvement</strong> over baselines. <strong>For</strong> <strong>progression in the liver</strong>, we see a <strong>+7.7% lift</strong>.</p></li><li><p><strong>Consistent Superiority:</strong> Across 45 distinct tasks, ranging from high-noise predictions like &#8220;Discontinuation due to Toxicity&#8221; to &#8220;Line of Therapy Transfer&#8221;, the Standard Model consistently outperforms the baselines.</p></li></ul><h3><strong>Conclusion</strong></h3><p>The hierarchy of performance is unambiguous: <strong>generalist models &lt; specialized public models &lt; specialized privately fine-tuned models &lt; specialized World Models.</strong></p><p><strong>These results indicate that while domain-specific data provides a necessary baseline, the architectural approach determines the performance ceiling. </strong>The measurable lift of the JEPA objective over pure SFT confirms that enforcing causal structure in the latent space yields more robust patient representations than autoregressive text generation alone. Ultimately, achieving state-of-the-art clinical prediction requires modeling the dynamics of the disease, not just the syntax of the data.</p><h3><strong>Announcing the Availability of Our Model Weights</strong></h3><p><strong>Standard Model offers a foundational step toward a broader vision: a reshaped oncology ecosystem where foundational models act as a partner in complex clinical decision making.</strong> By effectively modeling the patient state <em>S(t)</em> to simulate potential trajectories, we aim to support the entire care continuum - from refining early detection to guiding personalized treatment selection through <em>what-if</em> analysis.</p><p><strong>In the longer term, this approach holds significant promise for accelerating clinical research.</strong> Specifically, the generation of universal patient embeddings enables synthetic control arms that could reduce reliance on large control groups, thereby decreasing trial duration and costs.<strong> </strong>Ultimately, our goal is to facilitate a systemic shift from &#8220;general approach&#8221; to precise healthcare.</p><p><strong>That is why we are opening our model weights to the public. </strong>Visit Hugging Face <a href="https://huggingface.co/standardmodelbio/SMB-v1-1.7B-Structure">here</a> to access our weights and download them for use in your own workflow.</p><h4><strong>Acknowledgements</strong></h4><p>We would like to thank the MSK team that we worked with through the 2025 MSK iHub Challenge program, and especially the following for painstakingly contributing to dataset development.</p><ul><li><p>John Philip, MS, Senior Director, Clinical &amp; Translational Research Informatics and Data Strategy, MSK, New York, NY.</p></li><li><p>Neil J. Shah, MBBS, Assistant Attending Physician, MSK, New York, NY.</p></li><li><p>Nadia Bahadur, MS, Clinical &amp; Translational Research Informatics, MSK, New York, NY.</p></li><li><p>Andrew Niederhausern, Bioinformatics Manager, Clinical &amp; Translational Research Informatics, MSK, New York, NY.</p></li><li><p>Bryan Tran Van der Stegen, MBA, Business Analyst, Clinical &amp; Translational Research Informatics, MSK, New York, NY.</p></li><li><p>Haiyu Zheng, MS, Business Analyst, Clinical &amp; Translational Research Informatics, MSK, New York, NY.</p></li></ul><p><em><strong><a href="https://blog.standardmodel.bio/subscribe">Subscribe</a> to our Substack to be notified when publish new posts, or follow us on <a href="https://www.linkedin.com/company/standardmodel-bio/posts/?feedView=all">LinkedIn</a>, <a href="https://x.com/smbiomedicine">Twitter/X</a> and <a href="https://huggingface.co/standardmodelbio">Huggingface</a> for more.</strong></em></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.standardmodel.bio/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Thanks for reading Standard Model! Subscribe for free to receive new posts and support my work.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div>]]></content:encoded></item><item><title><![CDATA[The Patient is Not a Document: Moving from LLMs to a World Model for Oncology (Part 1)]]></title><description><![CDATA[Introducing the Standard Model: A new paradigm to build multimodal intelligence for oncology.]]></description><link>https://blog.standardmodel.bio/p/the-patient-is-not-a-document-moving</link><guid isPermaLink="false">https://blog.standardmodel.bio/p/the-patient-is-not-a-document-moving</guid><dc:creator><![CDATA[Zach Chen]]></dc:creator><pubDate>Thu, 04 Dec 2025 19:46:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!L00l!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p><strong>TL;DR: We built a multimodal foundation model for oncology that replaces the text-prediction of LLMs with the state-prediction of a World Model.</strong></p><p>For the last three years, the game of biomedical AI has been simple: take a massive general-purpose model (e.g., GPT, Claude, Gemini), feed it medical text, and watch it crush the<a href="https://www.usmle.org/about-usmle#:~:text=Why%20is%20the%20USMLE%20important,correlates%20to%20improved%20patient%20outcomes."> USMLE</a>. It worked. We saw models passing licensing exams with expert-level scores and reasoning through complex vignettes.</p><p>But as the hype settles, the data reveals a sobering reality. When these same models are tested on realistic patient cases requiring actual treatment decisions,<a href="https://www.nature.com/articles/s43018-025-00991-6"> GPT-4 achieves just 30.3% completeness</a>. Models struggle to measure tumor dimensions accurately through CT scans or to track disease progression without heavy reliance on external tools.<strong> In short, general-purpose AI models demonstrate exceptional retrieval capabilities on standardized exams, yet lack the grounded utility required for complex medical practice.</strong></p><p>So, what did we miss?</p><h3><strong>Language Cannot Represent Biological Complexity on Its Own</strong></h3><p>The disconnect of general purpose AI for medical treatment lies in a single, flawed assumption:<strong> that language is a sufficient proxy for disease biology. </strong>The industry has assumed that if it built a high-fidelity map (the text), we would understand the territory (the patient).</p><p>Today&#8217;s AI models were never natively trained on these foundational biological signals. The latest models from Google, OpenAI, and Anthropic have never &#8220;seen&#8221; an entire CT volume or &#8220;read&#8221; a whole-genome sequence; they have only processed text descriptions of them. And they certainly haven&#8217;t connected them to other signals across time and scale.</p><p><strong>Human biology is not a linguistic problem; it is inherently multimodal. </strong>And yet, today&#8217;s AI is not. Biological signals exist at every scale, forming a complex, interacting hierarchy: from molecular variations in genomic sequences to cellular structures in histopathology slides, up to anatomical changes in CT volumes and longitudinal lab patterns in EHR data.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!SbSy!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5bb438b-cf7c-4957-a222-f62494323d44_1960x1044.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!SbSy!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5bb438b-cf7c-4957-a222-f62494323d44_1960x1044.png 424w, https://substackcdn.com/image/fetch/$s_!SbSy!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5bb438b-cf7c-4957-a222-f62494323d44_1960x1044.png 848w, https://substackcdn.com/image/fetch/$s_!SbSy!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5bb438b-cf7c-4957-a222-f62494323d44_1960x1044.png 1272w, https://substackcdn.com/image/fetch/$s_!SbSy!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5bb438b-cf7c-4957-a222-f62494323d44_1960x1044.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!SbSy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5bb438b-cf7c-4957-a222-f62494323d44_1960x1044.png" width="1456" height="776" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c5bb438b-cf7c-4957-a222-f62494323d44_1960x1044.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:776,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:153701,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:false,&quot;topImage&quot;:true,&quot;internalRedirect&quot;:&quot;https://zachchen2.substack.com/i/179684092?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5bb438b-cf7c-4957-a222-f62494323d44_1960x1044.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!SbSy!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5bb438b-cf7c-4957-a222-f62494323d44_1960x1044.png 424w, https://substackcdn.com/image/fetch/$s_!SbSy!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5bb438b-cf7c-4957-a222-f62494323d44_1960x1044.png 848w, https://substackcdn.com/image/fetch/$s_!SbSy!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5bb438b-cf7c-4957-a222-f62494323d44_1960x1044.png 1272w, https://substackcdn.com/image/fetch/$s_!SbSy!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc5bb438b-cf7c-4957-a222-f62494323d44_1960x1044.png 1456w" sizes="100vw" fetchpriority="high"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 1: <strong>Image sourced from Stanford&#8217;s 2025 AI Index report</strong>. Today medical-specific models are trained on 75-250x fewer tokens than general purpose models.</figcaption></figure></div><h3><strong>Even Specialized Models Hit a &#8220;Description Ceiling&#8221;</strong></h3><p>Moving beyond general AI models, the last 18 months have seen the rise of purpose-built foundational models in biomedicine specifically for oncology, known as the &#8220;savant&#8221; models. For example,<a href="https://www.nature.com/articles/s41746-025-01780-2"> Memorial Sloan Kettering&#8217;s Woollie</a> detects cancer progression with a stunning 97% AUROC.<a href="https://www.nature.com/articles/s41586-024-08378-w"> Stanford&#8217;s MUSK</a> predicts melanoma relapse with 83% accuracy.</p><p>These more specialized models are impressive, but they hit a ceiling because of two distinct challenges in the current landscape:</p><ul><li><p><strong>The &#8220;Text Bridge&#8221; Issue:</strong> Most systems today, like Woollie, function as LLM agents (or pure LLMs) that invoke external tools (e.g., segmentation models or mutation classifiers) and synthesize the results via text.</p><ul><li><p><strong>The Challenge:</strong> They treat cancer progression as a<a href="https://www.nature.com/articles/s41746-024-01366-4"> linguistic probability problem</a>. For example, Woollie is trained on radiology <em>impressions</em> and summaries written by clinical professionals. It optimizes for semantic consistency with the doctor, not biological fidelity to the disease. By forcing biological reality through a text bottleneck, these methods suffer from <strong>lossy compression</strong>:<strong> </strong>they discard any biological signal the human observer failed to articulate (e.g., subtle textural changes in a raw CT volume). It cannot reason about biological signals that were never written down, e.g., a raw whole-genome sequence or the complex spatial topology of a tumor microenvironment<strong>.</strong></p></li></ul></li></ul><ul><li><p><strong>The &#8220;Snapshot&#8221; Limitation:</strong> Current models hit a ceiling because they reconstruct static snapshots of a patient state rather than <em>predict</em> a journey through time.</p><ul><li><p><strong>The Challenge:</strong> Models like MUSK learn by reconstructing missing pixels in static CT images or aligning a histopathology slide to a static report. While they may output the probability of a future event (e.g., high risk of relapse) and identify that a tumor looks &#8220;dangerous,&#8221; they arrive at these conclusions by correlating static patterns, such as lymphocyte infiltration. These models effectively &#8220;skip&#8221; the causal chain of clinical events. Consequently, it is difficult for them to simulate how a tumor would dynamically evolve under Treatment A versus Treatment B because they optimize for <strong>pattern matching</strong>, not <strong>causal effect</strong>.</p></li></ul></li></ul><h3><strong>Focus Needs to Shift from Description to Dynamics</strong></h3><p><strong>The fundamental problem isn&#8217;t that state-of-the-art LLMs need more medical text or that existing models need more data. The problem is the </strong><em><strong>learning paradigm</strong></em><strong>.</strong></p><p>Clinical oncology is not a series of static snapshots; it is an evolving biological trajectory.</p><p>A clinician does not simply ask &#8220;Is this cancer?&#8221; (classification). Rather, they ask:</p><blockquote><p>&#8220;Given the patient&#8217;s current state <em>S(t)</em> and this specific intervention <em>I</em>, what will their state be in 6 months <em>S(t+1)</em>?&#8221;.</p></blockquote><p>We need an architecture that bridges the gap: one that possesses the semantic high-level reasoning of an LLM but is anchored in the multimodal &#8220;ground truth&#8221; of disease biology - genomics, proteomics, imaging, EHR and longitudinal outcomes.</p><h3><strong>Standard Model&#8217;s Approach: A Biological World Model</strong></h3><p>This is the motivation behind our<a href="https://www.standardmodel.bio/"> </a><strong><a href="https://www.standardmodel.bio/">Standard Model</a></strong>. By moving from autoregressive text generation and masked reconstruction to a<a href="https://arxiv.org/pdf/2301.08243"> </a><strong><a href="https://arxiv.org/pdf/2301.08243">Joint-Embedding Predictive Architecture (JEPA)</a></strong>, we are ending the game of predicting words and beginning the complex work of modeling biological dynamics.</p><h4><strong>Architectural Deep Dive</strong></h4><p>Our Standard Model is a <strong>biological world model<a class="footnote-anchor" data-component-name="FootnoteAnchorToDOM" id="footnote-anchor-1" href="#footnote-1" target="_self">1</a></strong> that operates as a temporal loop, transforming disparate clinical signals into a cohesive, evolving &#8220;digital twin.&#8221; The architecture is defined by four specific design heuristics:</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!L00l!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!L00l!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png 424w, https://substackcdn.com/image/fetch/$s_!L00l!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png 848w, https://substackcdn.com/image/fetch/$s_!L00l!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png 1272w, https://substackcdn.com/image/fetch/$s_!L00l!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!L00l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png" width="1456" height="983" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:983,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:1235641,&quot;alt&quot;:&quot;&quot;,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://zachchen2.substack.com/i/179684092?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" title="" srcset="https://substackcdn.com/image/fetch/$s_!L00l!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png 424w, https://substackcdn.com/image/fetch/$s_!L00l!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png 848w, https://substackcdn.com/image/fetch/$s_!L00l!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png 1272w, https://substackcdn.com/image/fetch/$s_!L00l!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F81f231b2-4107-450d-ae2d-1dfd1da9e79a_4962x3351.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Figure 2: <strong>The Standard Model Architecture</strong>. It utilizes a Joint-Embedding Predictive Architecture (JEPA) to conditionally predict patient trajectories within the latent space rather than classifying static snapshots.</figcaption></figure></div><h4><strong>1. The Inputs: Modality Ingestion &amp; Fusion (Time = </strong><em><strong>t</strong></em><strong>)</strong></h4><p>Rather than relying on text as a lossy compression layer, Standard Model captures a 360-degree view of the patient. We ingest raw signals&#8212;such as genomics, proteomics, high-resolution imaging, and EHR data&#8212;and pass them through modality-specific encoders.</p><ul><li><p><strong>The Modality Fusion Heuristic:</strong> A specialized <strong>projector</strong> maps these raw encodings into an universal latent space of a <strong>state encoder</strong>. We don&#8217;t freeze the image encoders; we align the raw biological signals directly with the encoder&#8217;s strong world knowledge foundation. This yields a &#8220;fused&#8221; patient state embedding that retains both high-level semantic context and low-level biological granularity at time <em>t</em>.</p></li></ul><h4><strong>2. The Engine: The &#8220;What-If&#8221; Simulation</strong></h4><p>This is the core differentiator. Once the fused patient state embedding is generated, the model does not simply output a diagnosis. Instead, it functions as a <strong>state predictor</strong>.</p><ul><li><p><strong>Input:</strong> Current Fused Embedding <em>S(t)</em> + Intervention <em>A(t)</em>.</p></li><li><p><strong>Output:</strong> Predicted Next-Time-Point Embedding <em>S(t+1)</em>.</p></li><li><p><strong>Cause-and-Effect Data Structuring:</strong> To enforce causal learning, the training data is not fed as a static batch. It is structured in a strict <strong>cause-and-effect format</strong>: (Pre-State + Intervention) &#8594; (Post-State). By explicitly modeling a given <em>Intervention</em> as the catalyst for state change, the model learns to distinguish between the natural history of the disease and the specific impact of a treatment.</p></li></ul><blockquote><p><strong>The Shift:</strong> We do not ask the model to generate the pixels of the next CT scan or the text of the next report. We ask it to predict the <em>future state representation</em> in the latent space. This forces the model to learn the <strong>causal trajectory</strong> of the disease rather than the texture of the data.</p></blockquote><h4><strong>3. The Anchor: Hybrid Optimization</strong></h4><p>Pure JEPA models often suffer from &#8220;<a href="https://arxiv.org/pdf/2511.08544">training collapse</a>&#8221; (outputting constant representations) or drift. To prevent this and ensure clinical grounding, we employ a <strong>hybrid learning objective</strong>.</p><ul><li><p><strong>The Strategy:</strong> We combine supervised fine-tuning (SFT) with JEPA objectives by assigning mixed weights to these learning signals.</p></li><li><p><strong>The Result:</strong> The SFT component anchors the model to ground-truth clinical outcomes (e.g., &#8220;Did the patient respond to drug A?&#8221;), while the JEPA component forces the model to learn the underlying dynamics of <em>how</em> that state was reached.</p></li></ul><h4><strong>4. The Optimization Loop (Time = </strong><em><strong>t+1</strong></em><strong>)</strong></h4><p>We trained our <a href="https://huggingface.co/standardmodelbio/SMB-v1-1.7B-Structure">StandardModel-v1 (available on Hugging Face soon)</a> on real-world longitudinal oncology data. The model compares its <strong>predicted future state</strong> against the <strong>actual patient state</strong> (derived from the ground truth follow-up data). The error signal updates the model, teaching it the causal laws of how tumors progress and respond to therapy.</p><h3><strong>Conclusion: Modeling State, Not Tokens</strong></h3><p>For decades, we treated medical AI as a library problem; reading more, summarizing better, passing exams. <strong>But patients are not textbooks</strong>. They are dynamic, evolving biological systems.</p><p>Our Standard Model represents a shift from <strong>Reading</strong> to <strong>Reasoning</strong>, and from <strong>Description</strong> to <strong>Dynamics</strong>.</p><h3><strong>Next: Evaluation</strong></h3><p>Though we argued that oncology AI needs to move from describing a snapshot to projecting a journey. But how do you validate a projection? If we simply let the model generate text, we fall back into the trap of measuring syntax, not biology.</p><p>Instead, we test the <strong>quality of the state </strong>that the model learned<strong>.</strong> If our Standard Model is truly simulating the patient&#8217;s future (the territory), then the representation of that patient at any given moment <em>S(t</em>) should contain the &#8220;future&#8221; within it.</p><p><em>In Part 2, we will move from architecture to evidence. We will detail how we validated our Standard Model v1 on real-world longitudinal oncology datasets, defining &#8220;ground truth&#8221; not as static labels, but as dynamic time-windows for progression, toxicity, response, and survival. <strong><a href="https://blog.standardmodel.bio/subscribe">Subscribe</a> to our Substack to be notified when we publish the second part of this series, or follow us on <a href="https://www.linkedin.com/company/standardmodel-bio/posts/?feedView=all">LinkedIn</a> and <a href="https://x.com/smbiomedicine">Twitter/X</a> for more.</strong></em></p><p class="button-wrapper" data-attrs="{&quot;url&quot;:&quot;https://blog.standardmodel.bio/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe now&quot;,&quot;action&quot;:null,&quot;class&quot;:null}" data-component-name="ButtonCreateButton"><a class="button primary" href="https://blog.standardmodel.bio/subscribe?"><span>Subscribe now</span></a></p><div class="footnote" data-component-name="FootnoteToDOM"><a id="footnote-1" href="#footnote-anchor-1" class="footnote-number" contenteditable="false" target="_self">1</a><div class="footnote-content"><p>In robotics and autonomous driving, researchers and engineers realized that predicting text or classifying images wasn&#8217;t enough to navigate reality. They developed <strong><a href="https://arxiv.org/abs/1803.10122">World Models</a></strong>.</p><p>A World Model is not a knowledge base; it is a simulation engine. It learns an internal representation of how the environment works and, crucially, how actions change that environment.</p><ul><li><p><strong>A Classifier asks:</strong> &#8220;Is this a picture of a road?&#8221;</p></li><li><p><strong>A World Model asks:</strong> &#8220;If I turn the wheel 10 degrees left at this speed, where will the car be in 3 seconds?&#8221;</p></li></ul><p>This ability to simulate future states based on interventions is the &#8220;missing link&#8221; in oncology. A patient is not a static image; they are a dynamic environment. By treating the patient&#8217;s biology as the &#8220;world&#8221; and the treatment as the &#8220;steering wheel,&#8221; we can move from describing the cancer to simulating its trajectory.</p></div></div>]]></content:encoded></item><item><title><![CDATA[A Deep Dive Into Our Paper on Patient-Specific Biomolecular Instruction Tuning of Graph LLMs ]]></title><description><![CDATA[Combining patient-specific biomolecular interaction networks with instruction-tuned LLMs results in a system that understands the biological story behind each individual patient's data.]]></description><link>https://blog.standardmodel.bio/p/a-deep-dive-into-our-paper-on-patient</link><guid isPermaLink="false">https://blog.standardmodel.bio/p/a-deep-dive-into-our-paper-on-patient</guid><dc:creator><![CDATA[Irsyad]]></dc:creator><pubDate>Thu, 16 Oct 2025 13:02:32 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!iD48!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2913c17-aa07-43a4-ae39-f19b4c0eb408_1600x814.png" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>Two weeks ago, <a href="https://blog.standardmodel.bio/p/introducing-standard-model-biomedicine">Standard Model Biomedicine came out of stealth</a> with three new papers, each of which illustrates a different combination of the biological scales our multimodal foundation model operates across. One of these papers was titled &#8220;Patient-Specific Biomolecular Instruction Tuning of Graph LLMs&#8221;. The full text is available on arXiv: <a href="https://arxiv.org/pdf/2509.22853">https://arxiv.org/pdf/2509.22853</a>.</p><p>In this post, we will share further thoughts on this publication and our work behind it.</p><h3><strong>Biomolecular Instruction Tuning Can Reveal Advanced Molecular Interplays</strong></h3><p><strong>Omics data is a direct manifestation of biological processes occurring in the human body.</strong> While we can measure thousands of biomolecules in cancer patients, understanding what these measurements mean for each individual patient is incredibly complex. These biomolecules (e.g. proteins) aren&#8217;t produced in isolation; the measurements that we see are the result of hundreds to thousands of interactions triggering cascades of biological activity.</p><p>For example, the p53 tumor suppressor protein orchestrates DNA damage response by activating proteins such as MDM2 and p21 to coordinate cell cycle arrest and DNA repair, preventing the spread of genomic damage. Traditional analysis methods (e.g. univariate t-tests, fold-change ranking, principal component analysis, k-means clustering, logistic regression) analyze these proteomics levels independently, missing this crucial molecular interplay. Even a standard neural network would struggle to learn this distinction, as it must discover from scratch that these specific proteins should be analyzed as a connected group rather than independent features.</p><p>This is where biomolecular instruction tuning becomes imperative for analyzing diseases in the human body.<br><br><strong>By teaching AI models to understand the patient-specific protein levels within the context of their biomolecular interaction networks, we can capture a more complete biological story. </strong>At Standard Model Biomedicine, we expanded upon this idea and developed KRONOS: Knowledge Representation of patient Omics Networks in Oncology via Structured tuning. KRONOS creates an individualized molecular map for every patient by integrating their proteomics data with biomolecular networks, capturing both protein levels and their interaction dynamics within the human body. KRONOS then merges this enriched proteomics representation into an LLM, creating a multi-modal language model that empowers physicians to deliver further personalized care by mapping distinct protein signatures that characterize each patient&#8217;s malignancy.</p><p>By integrating representations grounded with biological relevance into an LLM, the model captures the complex biology underlying omics measurements, moving beyond pattern recognition towards intricate understanding of molecular networks, dependencies, and cascading effects that drive oncogenesis.</p><h3><strong>Biomolecular Graphs Boost Deep Learning of Patient Data</strong></h3><p>While traditional deep learning approaches in biomedicine assume feature independence, integrating biomolecular networks such as STRING PPI enables denoising of multi-omics data and learning of meaningful signals by utilizing these networks as structural priors. Early work such as EMOGI used graph neural networks (GNNs), leveraging these interaction maps to predict cancer genes, while subsequent methods like GNN-SubNet [<a href="https://academic.oup.com/bioinformatics/article/38/Supplement_2/ii120/6702000">1</a>] harnessed explainable AI to identify disease-specific subnetworks across a patient cohort. Recent approaches adopt larger and more heterogeneous biomolecular interaction networks (TREE [<a href="https://www.nature.com/articles/s41551-024-01312-5">2</a>]) to better capture these omics complexes. <strong>Biomedical AI literature demonstrates that leveraging these molecular networks alongside omics improves learning of meaningful signals in prediction and biological interpretability of diseases.</strong></p><p>Alongside these advances in graph omics learning, AI research is adopting new paradigms for enhancing reasoning in LLMs. The introduction of instruction tuning has enabled AI interpretation of modalities in biology and medicine that LLMs may not be already accustomed to.</p><p><strong>By training LLMs on medical instructions, researchers enable AI to reason across an out-of-scope modality for complex tasks.</strong> MIMIC-Instr [<a href="https://neurips.cc/virtual/2024/poster/97801">3</a>] allowed LLMs to reason about intricate longitudinal EHR data, while LLaVA-Med [<a href="https://arxiv.org/abs/2306.00890">4</a>] learned to expertly understand medical images. MEIT [<a href="https://arxiv.org/abs/2403.04945">5</a>] decoded ECG signals into clinical insights, and Me-LLaMA ingested 129 billion biomedical tokens to master medical language. <strong>These aren&#8217;t just pattern-matching tools anymore, they were AI systems that could understand context, follow clinical reasoning, and translate between different biomedical modalities.</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!iD48!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2913c17-aa07-43a4-ae39-f19b4c0eb408_1600x814.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!iD48!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2913c17-aa07-43a4-ae39-f19b4c0eb408_1600x814.png 424w, https://substackcdn.com/image/fetch/$s_!iD48!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2913c17-aa07-43a4-ae39-f19b4c0eb408_1600x814.png 848w, https://substackcdn.com/image/fetch/$s_!iD48!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2913c17-aa07-43a4-ae39-f19b4c0eb408_1600x814.png 1272w, https://substackcdn.com/image/fetch/$s_!iD48!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2913c17-aa07-43a4-ae39-f19b4c0eb408_1600x814.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!iD48!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2913c17-aa07-43a4-ae39-f19b4c0eb408_1600x814.png" width="728" height="370.5" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d2913c17-aa07-43a4-ae39-f19b4c0eb408_1600x814.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:false,&quot;imageSize&quot;:&quot;normal&quot;,&quot;height&quot;:741,&quot;width&quot;:1456,&quot;resizeWidth&quot;:728,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:&quot;center&quot;,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!iD48!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2913c17-aa07-43a4-ae39-f19b4c0eb408_1600x814.png 424w, https://substackcdn.com/image/fetch/$s_!iD48!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2913c17-aa07-43a4-ae39-f19b4c0eb408_1600x814.png 848w, https://substackcdn.com/image/fetch/$s_!iD48!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2913c17-aa07-43a4-ae39-f19b4c0eb408_1600x814.png 1272w, https://substackcdn.com/image/fetch/$s_!iD48!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd2913c17-aa07-43a4-ae39-f19b4c0eb408_1600x814.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Model architecture of KRONOS.</figcaption></figure></div><h3><strong>KRONOS: Knowledge Representation of Patient Omics Networks in Oncology via Structured Tuning</strong></h3><p>While proteomics is imperative to understanding a patient&#8217;s disease progression, LLMs haven&#8217;t inherently learned how to navigate through this complex modality, let alone how to reason and generate prognostic predictions for a patient&#8217;s disease state.</p><p><strong>As mentioned above, our paper introduces KRONOS as a framework that integrates enriched patient-specific network representations with modern LLM training paradigms, enabling models to capture complex proteomics signals and achieve competitive prognostic performance.</strong></p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!i_mb!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff93e782-b103-46d6-8106-f04d7adead3e_1600x590.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!i_mb!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff93e782-b103-46d6-8106-f04d7adead3e_1600x590.png 424w, https://substackcdn.com/image/fetch/$s_!i_mb!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff93e782-b103-46d6-8106-f04d7adead3e_1600x590.png 848w, https://substackcdn.com/image/fetch/$s_!i_mb!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff93e782-b103-46d6-8106-f04d7adead3e_1600x590.png 1272w, https://substackcdn.com/image/fetch/$s_!i_mb!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff93e782-b103-46d6-8106-f04d7adead3e_1600x590.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!i_mb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff93e782-b103-46d6-8106-f04d7adead3e_1600x590.png" width="1456" height="537" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/ff93e782-b103-46d6-8106-f04d7adead3e_1600x590.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:537,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!i_mb!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff93e782-b103-46d6-8106-f04d7adead3e_1600x590.png 424w, https://substackcdn.com/image/fetch/$s_!i_mb!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff93e782-b103-46d6-8106-f04d7adead3e_1600x590.png 848w, https://substackcdn.com/image/fetch/$s_!i_mb!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff93e782-b103-46d6-8106-f04d7adead3e_1600x590.png 1272w, https://substackcdn.com/image/fetch/$s_!i_mb!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fff93e782-b103-46d6-8106-f04d7adead3e_1600x590.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">CPTAC-PROTSTRUCT instruction generation pipeline.</figcaption></figure></div><p>KRONOS grounds patient-specific proteomics representation learning in biomedical interaction networks, generating biologically-relevant embeddings ingested by the LLM. To teach the model this unfamiliar modality, we constructed CPTAC-PROTSTRUCT, a dataset curated from the largest U.S. proteomics cancer study (CPTAC), which forces the LLM to align proteomics signals with its language system and unlocks true reasoning capability. The steps to generate the CPTAC-PROTSTRUCT schema alignment and clinical reasoning instruction pair sets are outlined in the figure above.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!dS-6!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89745d89-69db-45c5-b857-bb65899233ad_2520x1092.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!dS-6!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89745d89-69db-45c5-b857-bb65899233ad_2520x1092.png 424w, https://substackcdn.com/image/fetch/$s_!dS-6!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89745d89-69db-45c5-b857-bb65899233ad_2520x1092.png 848w, https://substackcdn.com/image/fetch/$s_!dS-6!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89745d89-69db-45c5-b857-bb65899233ad_2520x1092.png 1272w, https://substackcdn.com/image/fetch/$s_!dS-6!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89745d89-69db-45c5-b857-bb65899233ad_2520x1092.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!dS-6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89745d89-69db-45c5-b857-bb65899233ad_2520x1092.png" width="1456" height="631" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/89745d89-69db-45c5-b857-bb65899233ad_2520x1092.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:631,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:405561,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/png&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.standardmodel.bio/i/176200482?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89745d89-69db-45c5-b857-bb65899233ad_2520x1092.png&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!dS-6!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89745d89-69db-45c5-b857-bb65899233ad_2520x1092.png 424w, https://substackcdn.com/image/fetch/$s_!dS-6!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89745d89-69db-45c5-b857-bb65899233ad_2520x1092.png 848w, https://substackcdn.com/image/fetch/$s_!dS-6!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89745d89-69db-45c5-b857-bb65899233ad_2520x1092.png 1272w, https://substackcdn.com/image/fetch/$s_!dS-6!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F89745d89-69db-45c5-b857-bb65899233ad_2520x1092.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Model generation comparison of schema alignment and clinical reasoning.</figcaption></figure></div><p>By aligning proteomics with the LLM and enabling reasoning through instruction tuning, the model learns the semantic nuances of this newly-introduced modality. It can then generate proteomics-grounded text and, more importantly, develop rich internal representations that serve as a foundation for various downstream tasks. Once we train KRONOS, we validate the learned representations by probing its hidden layers against ground truth, using a frozen LLM backbone that was not explicitly trained on these downstream tasks during training, shown below.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!pA_L!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc760e606-3258-4926-bfa1-c9564696bd91_1600x724.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!pA_L!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc760e606-3258-4926-bfa1-c9564696bd91_1600x724.png 424w, https://substackcdn.com/image/fetch/$s_!pA_L!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc760e606-3258-4926-bfa1-c9564696bd91_1600x724.png 848w, https://substackcdn.com/image/fetch/$s_!pA_L!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc760e606-3258-4926-bfa1-c9564696bd91_1600x724.png 1272w, https://substackcdn.com/image/fetch/$s_!pA_L!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc760e606-3258-4926-bfa1-c9564696bd91_1600x724.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!pA_L!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc760e606-3258-4926-bfa1-c9564696bd91_1600x724.png" width="1456" height="659" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/c760e606-3258-4926-bfa1-c9564696bd91_1600x724.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:659,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!pA_L!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc760e606-3258-4926-bfa1-c9564696bd91_1600x724.png 424w, https://substackcdn.com/image/fetch/$s_!pA_L!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc760e606-3258-4926-bfa1-c9564696bd91_1600x724.png 848w, https://substackcdn.com/image/fetch/$s_!pA_L!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc760e606-3258-4926-bfa1-c9564696bd91_1600x724.png 1272w, https://substackcdn.com/image/fetch/$s_!pA_L!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fc760e606-3258-4926-bfa1-c9564696bd91_1600x724.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Overall performance of KRONOS against prognostic tasks. </figcaption></figure></div><p>For comparison, we benchmarked KRONOS against classical and deep learning methods in biomedical literature for predicting prognostic outcomes from omics data. These models range from traditional MLPs, to graph classification of proteomics-injected PPI networks, to multi-modal LLMs with various encoder backbones. Unlike methods specifically trained to predict prognostic outcomes from proteomics data, KRONOS learns general-purpose representations through multi-modal semantic alignment, without task-specific fine-tuning.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!nD-P!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3553f8a-cf21-4aa2-9e76-a1678591ed6d_1267x414.png" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!nD-P!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3553f8a-cf21-4aa2-9e76-a1678591ed6d_1267x414.png 424w, https://substackcdn.com/image/fetch/$s_!nD-P!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3553f8a-cf21-4aa2-9e76-a1678591ed6d_1267x414.png 848w, https://substackcdn.com/image/fetch/$s_!nD-P!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3553f8a-cf21-4aa2-9e76-a1678591ed6d_1267x414.png 1272w, https://substackcdn.com/image/fetch/$s_!nD-P!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3553f8a-cf21-4aa2-9e76-a1678591ed6d_1267x414.png 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!nD-P!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3553f8a-cf21-4aa2-9e76-a1678591ed6d_1267x414.png" width="1267" height="414" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/d3553f8a-cf21-4aa2-9e76-a1678591ed6d_1267x414.png&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:414,&quot;width&quot;:1267,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:null,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:null,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:null,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!nD-P!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3553f8a-cf21-4aa2-9e76-a1678591ed6d_1267x414.png 424w, https://substackcdn.com/image/fetch/$s_!nD-P!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3553f8a-cf21-4aa2-9e76-a1678591ed6d_1267x414.png 848w, https://substackcdn.com/image/fetch/$s_!nD-P!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3553f8a-cf21-4aa2-9e76-a1678591ed6d_1267x414.png 1272w, https://substackcdn.com/image/fetch/$s_!nD-P!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2Fd3553f8a-cf21-4aa2-9e76-a1678591ed6d_1267x414.png 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a><figcaption class="image-caption">Comparative performance of general-purpose KRONOS embeddings against various task-trained models.</figcaption></figure></div><h3><strong>Where We Go From Here</strong></h3><p>KRONOS represents more than just another biomedical AI tool&#8212;it&#8217;s proof that we can teach language models to fluently speak the language of molecular biology.</p><p><strong>Our results show something remarkable: when you combine patient-specific protein networks with instruction-tuned LLMs, you don&#8217;t just get better predictions, you get a system that truly understands the biological story behind each patient&#8217;s data</strong>. The model&#8217;s ability to outperform traditional approaches across mortality prediction, cancer typing, and survival analysis demonstrates exciting progress towards even more patient-centric care.</p><p>The most exciting implication of our paper is that KRONOS demonstrates the gap between molecular biology and clinical medicine can be bridged by artificial intelligence. This isn&#8217;t about replacing doctors or biologists; it&#8217;s about giving them a tool that can instantly clinically contextualize complex molecular measurements. <strong>As we refine this technology and expand to more biologically-relevant modalities, we move toward a future where treatments are guided by deeper understanding of individual patient molecular landscapes</strong>.</p><p><em>For deep dives into our other recent publications and to receive future news and updates from Standard Model Biomedicine, subscribe to our Substack, follow us on X and LinkedIn, or check out our website at <a href="http://www.standardmodel.bio">www.standardmodel.bio</a>.</em></p>]]></content:encoded></item><item><title><![CDATA[Introducing Standard Model Biomedicine]]></title><description><![CDATA[We're building a multimodal foundation model that integrates any type of data for biopharma, academic medical centers, and biobanks.]]></description><link>https://blog.standardmodel.bio/p/introducing-standard-model-biomedicine</link><guid isPermaLink="false">https://blog.standardmodel.bio/p/introducing-standard-model-biomedicine</guid><dc:creator><![CDATA[Kevin Brown]]></dc:creator><pubDate>Tue, 30 Sep 2025 16:30:56 GMT</pubDate><enclosure url="https://substackcdn.com/image/fetch/$s_!yLBq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic" length="0" type="image/jpeg"/><content:encoded><![CDATA[<p>We are building the <em><strong>Standard Model for Biomedicin</strong></em><strong>e</strong>.</p><p>The Standard Model is a multimodal foundation model that integrates any biomedical measurement at any scale, from the molecular to the whole-patient level, into a common, shared representation of a human patient. This is a fundamentally human endeavor that respects the inherent complexity of patients.</p><p>We expect biomedical AI researchers will tailor the Standard Model for downstream use cases as diverse as medicine itself. In biopharma, we anticipate the Standard Model powering tasks ranging from digital twins for prognosis to site enrollment and trial planning and execution. In academic medicine, we foresee our model pushing forward AI-driven radiation toxicity assessments in early lung cancer patients using electrocardiograms and radiology scans, or patient genomic profiles for targeted therapy recommendations.</p><p>Broadly speaking, the Standard Model will accelerate AI development timelines and increase performance of downstream applications in biomedical AI, just as other foundation models have done in language and vision in other industries.</p><p>We are backed on this mission by VC firms <a href="https://www.arkitektventures.com/">Arkitekt Ventures</a> and <a href="https://www.virtuevc.com/">Virtue</a>, and we&#8217;ve grown <a href="https://www.standardmodel.bio/">a team of AI and biopharma leaders with deep domain expertise</a> to build our model.</p><h2><strong>Biomedicine Needs a Multimodal Foundation Model</strong></h2><p>We are driven by four beliefs about what it takes to accelerate performant AI for human biology:</p><ol><li><p><strong>Performance Is Everything in Biomedicine. </strong>Small differences in clinical efficacy relative to standard of care drive billions of dollars in revenue and change millions of lives. Every marginal improvement in performance can define clinical, regulatory, or commercial success.</p></li></ol><ol start="2"><li><p><strong>Data Enables Performance. </strong>The primacy of scale is well accepted in traditional machine learning and AI; the most performant large language models are trained on nearly exhaustive amounts of human text, and likewise for vision models. The largest biomedical models have not yet been trained at this scale.</p></li></ol><ol start="3"><li><p><strong>Data is Naturally Siloed in Biomedicine</strong>. Any organization with ultimately limited data access will likely fail, because they will never scale to the level required. Because biopharma and academic researchers typically collect data in narrow bands for specific questions, the data required for training at scale never amasses within one organization.</p></li></ol><ol start="4"><li><p><strong>The Most Performant Foundation Models Will Not Be Siloed. </strong>The best foundation models will be trained on data across modalities as well as disease areas; data that is naturally siloed in the medical world today. Those siloes are detrimental to foundation model training. Patient data is inherently multimodal because patients are inherently multimodal; relevant measurements even for very specific indications span multiple ways of measuring patients.</p></li></ol><h2><strong>Where We Are Headed: From a Universal Patient Representation to a Biological Reasoning Engine</strong></h2><p><strong>We are building a universal foundation model for biomedicine. We do this by ingesting data from any modality and mapping it to a shared representation space. </strong>Each patient&#8217;s data is translated into a set of embeddings across time, representing patient state. Much like physics represents particle systems in appropriate coordinates, we develop a new coordinate system for patients.</p><div class="captioned-image-container"><figure><a class="image-link image2 is-viewable-img" target="_blank" href="https://substackcdn.com/image/fetch/$s_!yLBq!,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic" data-component-name="Image2ToDOM"><div class="image2-inset"><picture><source type="image/webp" srcset="https://substackcdn.com/image/fetch/$s_!yLBq!,w_424,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic 424w, https://substackcdn.com/image/fetch/$s_!yLBq!,w_848,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic 848w, https://substackcdn.com/image/fetch/$s_!yLBq!,w_1272,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic 1272w, https://substackcdn.com/image/fetch/$s_!yLBq!,w_1456,c_limit,f_webp,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic 1456w" sizes="100vw"><img src="https://substackcdn.com/image/fetch/$s_!yLBq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic" width="1456" height="962" data-attrs="{&quot;src&quot;:&quot;https://substack-post-media.s3.amazonaws.com/public/images/63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic&quot;,&quot;srcNoWatermark&quot;:null,&quot;fullscreen&quot;:null,&quot;imageSize&quot;:null,&quot;height&quot;:962,&quot;width&quot;:1456,&quot;resizeWidth&quot;:null,&quot;bytes&quot;:340192,&quot;alt&quot;:null,&quot;title&quot;:null,&quot;type&quot;:&quot;image/heic&quot;,&quot;href&quot;:null,&quot;belowTheFold&quot;:true,&quot;topImage&quot;:false,&quot;internalRedirect&quot;:&quot;https://blog.standardmodel.bio/i/174938176?img=https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic&quot;,&quot;isProcessing&quot;:false,&quot;align&quot;:null,&quot;offset&quot;:false}" class="sizing-normal" alt="" srcset="https://substackcdn.com/image/fetch/$s_!yLBq!,w_424,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic 424w, https://substackcdn.com/image/fetch/$s_!yLBq!,w_848,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic 848w, https://substackcdn.com/image/fetch/$s_!yLBq!,w_1272,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic 1272w, https://substackcdn.com/image/fetch/$s_!yLBq!,w_1456,c_limit,f_auto,q_auto:good,fl_progressive:steep/https%3A%2F%2Fsubstack-post-media.s3.amazonaws.com%2Fpublic%2Fimages%2F63c6d26a-1916-4abe-b20a-306d0f54de6b_3172x2096.heic 1456w" sizes="100vw" loading="lazy"></picture><div class="image-link-expand"><div class="pencraft pc-display-flex pc-gap-8 pc-reset"><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container restack-image"><svg role="img" width="20" height="20" viewBox="0 0 20 20" fill="none" stroke-width="1.5" stroke="var(--color-fg-primary)" stroke-linecap="round" stroke-linejoin="round" xmlns="http://www.w3.org/2000/svg"><g><title></title><path d="M2.53001 7.81595C3.49179 4.73911 6.43281 2.5 9.91173 2.5C13.1684 2.5 15.9537 4.46214 17.0852 7.23684L17.6179 8.67647M17.6179 8.67647L18.5002 4.26471M17.6179 8.67647L13.6473 6.91176M17.4995 12.1841C16.5378 15.2609 13.5967 17.5 10.1178 17.5C6.86118 17.5 4.07589 15.5379 2.94432 12.7632L2.41165 11.3235M2.41165 11.3235L1.5293 15.7353M2.41165 11.3235L6.38224 13.0882"></path></g></svg></button><button tabindex="0" type="button" class="pencraft pc-reset pencraft icon-container view-image"><svg xmlns="http://www.w3.org/2000/svg" width="20" height="20" viewBox="0 0 24 24" fill="none" stroke="currentColor" stroke-width="2" stroke-linecap="round" stroke-linejoin="round" class="lucide lucide-maximize2 lucide-maximize-2"><polyline points="15 3 21 3 21 9"></polyline><polyline points="9 21 3 21 3 15"></polyline><line x1="21" x2="14" y1="3" y2="10"></line><line x1="3" x2="10" y1="21" y2="14"></line></svg></button></div></div></div></a></figure></div><p>Once this fundamental representation &#8211; the Standard Model &#8211; is established, it can be fed to other reasoning engines. These engines can then drive decisions in diverse use cases across biopharma and academic medicine, such as:</p><ul><li><p>analyzing natural histories of patient disease </p></li><li><p>optimizing care pathways </p></li><li><p>forecasting clinical trials in silico</p></li><li><p>optimizing inclusion/exclusion criteria </p></li><li><p>modeling the probability of technical and regulatory success</p></li><li><p>and beyond</p></li></ul><p><strong>Critically, we intend to serve as the quiet backbone of these systems</strong>. Biopharma and academic medical centers should spend time bringing domain expertise to downstream applications, context-specific benchmarking (e.g. clinical trials), and curating specialized datasets. Standard Model Biomedicine offers the benefits of AI performance that come with scale and allows experts to optimize for the last 10 percent specific to their applications.</p><h2><strong>Where We Are Today</strong></h2><p>Human biology operates meaningfully across several scales, from the incredibly small, such as genomic mutations, to the incredibly large, such as lifestyle choices or surgical interventions. <strong>We firmly believe the best foundation model will integrate data from every scale.</strong></p><p>To that end, we&#8217;re releasing three new papers focused on molecular, cellular, and whole-patient scales. These publications build on the paper we quietly published last year to showcase the strength of our model: <em><a href="https://doi.org/10.48550/arXiv.2406.09454">Advancing High Resolution Vision-Language Models in Biomedicine</a></em>.</p><p>Taken together, we have published state-of-the-art methods at every level of human biology. Ultimately, we link models at every scale into the Standard Model of Biomedicine.</p><h2>1. Oncology &amp; Whole Genome Sequencing</h2><p>The ability to encode genomic sequences is the most basic and fundamental aspect of the Standard Model. Genomic mutations are fundamental to oncology and other diseases; any model of biomedicine must represent them in some way.</p><p>In this paper, we trained GenVarFormer (GVF), a whole genome sequencing foundation model, by predicting the functional consequence of variants on gene expression to achieve state-of-the-art performance on downstream tasks.</p><p><strong>Read the full paper here: </strong><em><strong><a href="https://arxiv.org/abs/2509.25573">GenVarFormer: Predicting Gene Expression From Long-Range Mutations in Cancer</a></strong></em></p><h2>2. A Molecular Language Model</h2><p>Linking molecular and cellular foundation models to a shared representation space, much like vision-language models, is a fundamental challenge for building the Standard Model. To that end, we developed a test case linking proteomic graph neural networks to language. This approach is applicable beyond proteomics to any graph-based representations at a cellular level.</p><p><strong>Read the full paper here: </strong><em><strong><a href="https://arxiv.org/abs/2509.22853">Patient-Specific Biomolecular Instruction Tuning of Graph-LLMs</a></strong></em></p><h2>3. Using EHRs to Predict Next Medical Codes</h2><p>Any multimodal foundation model must have the ability to ingest text, in particular longitudinal EHRs, the broadest and highest-level modality of input. This work reframes EHRs as timestamped chains of clinical events and fine&#8209;tunes large language models to predict the next event, improving temporal reasoning over disease trajectories.</p><p><strong>Read the full paper here: </strong><em><strong><a href="https://arxiv.org/abs/2509.25591">Building the EHR Foundation Model via Next Event Prediction</a></strong></em></p><h2><strong>Our Vision for the Future of Biomedicine</strong></h2><p>Going forward, Standard Model Biomedicine will continue to drive cutting-edge research, deploy our model to partners across the ecosystem, and collaborate with the highest-quality data sources in biomedicine. We will combine all modalities and scales of data into a single overarching Standard Model of human biology.</p><p><strong>We&#8217;re always looking for people to join us in building the future of biomedical research.</strong></p><ul><li><p>If you&#8217;re a biopharma company that wants to partner or a data source that wants to drive value from your data, reach out <a href="mailto:info@standardmodel.bio">here</a></p></li><li><p>If you&#8217;re a medical AI researcher that wants to partner, reach out <a href="mailto:info@standardmodel.bio">here</a></p></li><li><p>If you&#8217;re both technical and biologically inclined and would like to join our team, reach out <a href="mailto:hiring@standardmodel.bio">here</a></p></li><li><p>If you&#8217;d just like to chat about foundation models in biology, definitely reach out <a href="mailto:hello@standardmodel.bio">here</a></p></li></ul><p>We&#8217;re looking forward to sharing more about Standard Model Biomedicine and our work. To stay updated, follow us on <a href="https://www.linkedin.com/company/standardmodel-bio">LinkedIn</a> and <a href="https://x.com/smbiomedicine">X</a>, <a href="https://blog.standardmodel.bio/subscribe">subscribe</a> to our Substack, or reach out to us. <br><br><strong>The foundation for the future of biomedical research and drug development is being laid now, one datapoint at a time</strong>.<br></p><div class="subscription-widget-wrap-editor" data-attrs="{&quot;url&quot;:&quot;https://blog.standardmodel.bio/subscribe?&quot;,&quot;text&quot;:&quot;Subscribe&quot;,&quot;language&quot;:&quot;en&quot;}" data-component-name="SubscribeWidgetToDOM"><div class="subscription-widget show-subscribe"><div class="preamble"><p class="cta-caption">Subscribe for free to receive new posts from Standard Model Biomedicine.</p></div><form class="subscription-widget-subscribe"><input type="email" class="email-input" name="email" placeholder="Type your email&#8230;" tabindex="-1"><input type="submit" class="button primary" value="Subscribe"><div class="fake-input-wrapper"><div class="fake-input"></div><div class="fake-button"></div></div></form></div></div><p></p>]]></content:encoded></item></channel></rss>