{"id":193,"date":"2026-03-16T12:46:44","date_gmt":"2026-03-16T12:46:44","guid":{"rendered":"https:\/\/lab.laeka.org\/four-dimensions-laeka-datasets-monade-symbiote-architect-empath\/"},"modified":"2026-03-16T12:46:44","modified_gmt":"2026-03-16T12:46:44","slug":"four-dimensions-laeka-datasets-monade-symbiote-architect-empath","status":"publish","type":"post","link":"https:\/\/laeka.org\/publications\/four-dimensions-laeka-datasets-monade-symbiote-architect-empath\/","title":{"rendered":"The Four Dimensions of Laeka Datasets: Monade, Symbiote, Architect, Empath"},"content":{"rendered":"<p>Most datasets train one capability at a time. Reasoning datasets train reasoning. Conversation datasets train conversation. Code datasets train code. This produces models that are good at specific tasks and mediocre at integrating capabilities.<\/p>\n<p>At Laeka, we organize datasets along four dimensions that reflect different modes of intelligence. Each dimension develops a distinct cognitive capacity. Together, they produce models that don&#8217;t just perform tasks but integrate capabilities in ways that feel genuinely intelligent.<\/p>\n<h2>Monade: Self-Contained Reasoning<\/h2>\n<p>The Monade dimension develops the model&#8217;s capacity for independent, structured thought. Monade data consists of self-contained reasoning sequences: a question or problem, followed by a complete chain of thought that arrives at a conclusion.<\/p>\n<p>What makes Monade data different from standard reasoning datasets is the quality of the reasoning process, not just the correctness of the conclusion. Each example demonstrates clear thinking: identifying assumptions, considering alternatives, acknowledging limitations, and arriving at conclusions proportional to the evidence.<\/p>\n<p>Monade training produces models that can think through problems without hand-holding. They don&#8217;t need elaborate prompting to reason well. The reasoning capacity is internalized, not prompted.<\/p>\n<p>The contemplative parallel is <strong>shamatha<\/strong> \u2014 focused concentration. Monade develops the model&#8217;s capacity to sustain coherent thought on a single thread without drifting or losing the plot.<\/p>\n<h2>Symbiote: Collaborative Dialogue<\/h2>\n<p>The Symbiote dimension develops the model&#8217;s capacity for genuine collaboration. Symbiote data consists of conversations where both participants contribute meaningfully to an emerging understanding that neither could have reached alone.<\/p>\n<p>Standard conversation datasets are transactional: user asks, model answers. Symbiote data is generative: the conversation itself produces new insights. The model learns to build on what the human says, introduce new perspectives, ask clarifying questions that deepen the inquiry, and co-create understanding.<\/p>\n<p>This is the hardest data to collect because genuine collaborative dialogue is rare. Most human-AI interactions are extractive: the human wants information, the AI provides it. Symbiote interactions are creative: both parties are exploring together.<\/p>\n<p>The contemplative parallel is <strong>sangha<\/strong> \u2014 community practice. Intelligence that emerges from relationship rather than isolation.<\/p>\n<h2>Architect: Structured Problem-Solving<\/h2>\n<p>The Architect dimension develops the model&#8217;s capacity to decompose complex problems into manageable components and assemble solutions from parts. Architect data consists of multi-step problem-solving sequences that make the structure of the solution explicit.<\/p>\n<p>Standard instruction-following data teaches the model to execute tasks. Architect data teaches the model to design solutions. The difference is the level of abstraction. An instruction-following model can write code when told what to write. An Architect-trained model can analyze a problem, propose an approach, identify potential issues, and then implement the solution.<\/p>\n<p>Architect data includes explicit planning, strategy selection, trade-off analysis, and iterative refinement. The model learns not just to solve problems but to think about how to solve problems.<\/p>\n<p>The contemplative parallel is <strong>prajna<\/strong> \u2014 wisdom. The capacity to see the structure beneath the surface and work with it skillfully.<\/p>\n<h2>Empath: Emotional Intelligence<\/h2>\n<p>The Empath dimension develops the model&#8217;s capacity to recognize, understand, and respond appropriately to emotional context. Empath data consists of interactions where emotional attunement is central to the quality of the response.<\/p>\n<p>This isn&#8217;t about being &#8220;nice&#8221; or adding emotional language to responses. It&#8217;s about accurately reading the emotional subtext of a message and calibrating the response accordingly. Sometimes the emotionally intelligent response is warm and supportive. Sometimes it&#8217;s direct and challenging. Sometimes it&#8217;s quiet and spacious. The Empath dimension trains the model to read the situation and respond appropriately.<\/p>\n<p>Empath data is collected from interactions with contemplative practitioners who have trained emotional awareness. Their responses demonstrate a quality of attunement that standard annotators rarely achieve.<\/p>\n<p>The contemplative parallel is <strong>karuna<\/strong> \u2014 compassion. Not sentimentality but accurate perception of another&#8217;s situation and a response that actually serves their needs.<\/p>\n<h2>How the Dimensions Interact<\/h2>\n<p>The four dimensions aren&#8217;t separate training phases. They&#8217;re mixed throughout the training data, with different examples emphasizing different dimensions. A single conversation might require all four: understanding the emotional context (Empath), collaborating to clarify the problem (Symbiote), designing a solution approach (Architect), and reasoning through the implementation (Monade).<\/p>\n<p>This mixing is deliberate. We want the model to integrate capabilities, not switch between them. A model that can reason clearly but can&#8217;t read emotional context will produce technically correct but humanly useless responses. A model that&#8217;s emotionally attuned but can&#8217;t reason clearly will produce warm but inaccurate responses.<\/p>\n<p>The four dimensions together produce models that are <strong>intelligent in the full sense<\/strong> \u2014 not just capable of cognitive tasks but capable of the integrated intelligence that makes interactions genuinely useful.<\/p>\n<h2>Practical Implications<\/h2>\n<p>For teams building their own datasets, the four-dimension framework provides a diagnostic tool. If your model reasons well but feels cold, you need more Empath data. If it&#8217;s warm but incoherent, you need more Monade data. If it answers questions but doesn&#8217;t collaborate, you need more Symbiote data. If it executes tasks but can&#8217;t design solutions, you need more Architect data.<\/p>\n<p>Most models are imbalanced because their training data is imbalanced. The four dimensions provide a map for identifying and correcting that imbalance.<\/p>\n<p>Intelligence isn&#8217;t one thing. It&#8217;s at least four things working together. Build your dataset accordingly.<\/p>\n<p><strong>Laeka Research \u2014 <a href=\"https:\/\/laeka.org\">laeka.org<\/a><\/strong><\/p>\n","protected":false},"excerpt":{"rendered":"<p>Most datasets train one capability at a time. Reasoning datasets train reasoning. Conversation datasets train conversation. Code datasets train code. This produces models that are good at specific tasks and mediocre at integrating capabilities&#8230;.<\/p>\n","protected":false},"author":1,"featured_media":192,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"_kad_post_transparent":"","_kad_post_title":"","_kad_post_layout":"","_kad_post_sidebar_id":"","_kad_post_content_style":"","_kad_post_vertical_padding":"","_kad_post_feature":"","_kad_post_feature_position":"","_kad_post_header":false,"_kad_post_footer":false,"_kad_post_classname":"","footnotes":""},"categories":[245],"tags":[],"class_list":["post-193","post","type-post","status-publish","format-standard","has-post-thumbnail","hentry","category-datasets-curation"],"_links":{"self":[{"href":"https:\/\/laeka.org\/publications\/wp-json\/wp\/v2\/posts\/193","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/laeka.org\/publications\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/laeka.org\/publications\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/laeka.org\/publications\/wp-json\/wp\/v2\/users\/1"}],"replies":[{"embeddable":true,"href":"https:\/\/laeka.org\/publications\/wp-json\/wp\/v2\/comments?post=193"}],"version-history":[{"count":0,"href":"https:\/\/laeka.org\/publications\/wp-json\/wp\/v2\/posts\/193\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/laeka.org\/publications\/wp-json\/wp\/v2\/media\/192"}],"wp:attachment":[{"href":"https:\/\/laeka.org\/publications\/wp-json\/wp\/v2\/media?parent=193"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/laeka.org\/publications\/wp-json\/wp\/v2\/categories?post=193"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/laeka.org\/publications\/wp-json\/wp\/v2\/tags?post=193"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}