Personal Knowledge Engineering (PKE) Systems ... A Manifesto
Success is not solely about hard work and the diligency of showing up every day to just bang away fixing the same old shit.
Rather, SUCCESS is about having one's eyes on the horizon RATHER THAN THE REAR VIEW MIRROW ... coming up with a credible knowledge-based [rather than tool-based or equipment-based] plan for adaptability in the future ... and that means working harder, in a disciplined manner, up front preparing that plan, developing the tools and systems, like PKE systems, to have FLEXIBLE SYSTEMS THINKING in place before the trouble ever shows up ... not just to dig a well before one is thirsty, but to have the knowledge and capacity to engineer well-drilling systems when well-drilling systems are needed ... at its core, PKE is really about intelligently understanding systems and working in the development of knowledge-based systems for future needs, it's NOT personal knowledge management (PKM) -- which is important, but about the management of collections notes, ideas, plans, artifacts, references -- PKE is more forward-thinking, ie thinking about the development of future systems.
This means that we need tools and technology for NEW INTELLIGENCE, for intelligence and knowledge that hasn't come into existence yet.
Doing this is about systems that have the capacity to gather much more intelligence and knowledge than we currently do. It's also about making the attempts to apply the knowledge that our intelligence gathering ops have obtained more efficient and rapid. PKE is in the system of applying the knowledge and testing assumptions; PKE is about our methodology and thinking that we use to identify causal relationships and validate their veracity in order to utilize shortcuts to overcome challenges and free up time for pursuing larger goals ... and, if we love doing this the LARGER goal or reward will be that we get to improve the PKE implementation!
This Manifesto attempts to give an overview of the primary goals of the 100 modules in our 100-day project to level up our game in PKE systems, as well as outline the core principles of PKE systems and to explain something about what the measures of success will be for this project. You could say that this 100-module plan is really about implementing something akin Marcus du Sautoy's "Thinking Better: The Art of the Shortcut" because a big part of it is a celebration of how mathematical and statistical thinking helps us to solve problems more efficiently in everyday life, in producing anything, in design.
People might get sidetracked by the fact that du Sautoy's a mathematician, but this is most definitely not JUST mathematics, although mathematics is invaluable for implementing the art of the elegant, stable equilibrium solution. It really about understanding systems in order to find elegant and efficient solutions to complex problems by recognizing patterns and developing general algorithms ... rather than band-aids or cobbled-up, likely to fail fixes. Elegance is about solutions that stay fixed or heal and get better over time.
It is worth emphasizing that elegant thinking "shortcuts" are NOT at all about taking unethical or lazy approaches, but rather about developing a deeper understanding of problems to find more intelligent and clever ways to navigate them. The whole point of developing and using more advanced personal knowledge engineering (PKE) systems is not for PKE itself [although THAT is the goal of the 100 module plan] but to understand systems and genuinely "think better." Getting past the bandaid or likely-to-just-break-down-and-fail-again fix is about adopting not just a mindset but an entire PKE arsenal that allows one to understand, seek out and leverage the more clever solutions, recognizing that efficiency and deeper understanding can lead to more fulfilling and impactful achievements.
Primary Goals
- The core objective is progressive, to advance beyond the transition from the passive practice of Personal Knowledge Management (PKM) and make PKM note-gather more the mere collection of random notes and notetaking apps ... TOWARD ... a more actively evolving or extensible, disciplined system of AI-assisted Personal Knowledge Engineering (PKE) ... which presents all kind of opportunities that enhance our capacity to contribute to significant work in extensible open-source technologies.
- Fostering meaningful new professional connections and friendships across different disciplines in virtual venues [where people would not otherwise meet in the halls of the departments or R&D labs of their corporations]; the general goal of AI-assisted PKM and PKE is to accelerate the continuous learning and development processes, to spark new creative work, and, most importantly, to meet new friends by sharing this journey of building PKE technology to accelerate the continuous learning process in public
- As we learn more, we will attempt to better transform atomic notes, likely collected in simple MarkDown files used for this mdBook knowledgebase, from a static archive or just an online book into a more dynamic, programmable publishing AI engine, ready for sharing, collaboration, querying and other advanced augmentation with AI extensiions ... but in order to do this, we must articulating and embody the goals and principles of a systematic PKE framework to accelerate our own autodidactic education ... which is key in understanding the details of research in new systems at the forefront of technological innovation in various disciplines.
Core principles
- There's always going to be someone, somewhere who developing a better feature ... not necessarily a better system, but a compelling feature to understand and appraise. We want to be aware of what's happening with shiny new features, but we want to understand whether or not they represent generally better system. The Rust programming language's core feature, for example, is in its ownership and borrowing system, enforced by the RustLang borrow checker at compile time resulting in greater safety and simplicity in code, while retaining the speed of C language. The Modular Platform, with Mojo, Max, and the MLIR compiler framework, offers a different approach, particularly focused on high-performance AI development and leveraging advancements in compiler technology. Mojo, inspired by Rust and built on a next-generation MLIR compiler framework, aims to provide even higher levels of performance, particularly for AI applications, outperforming Rust in certain scenarios, like DNA parsing, through features like eager destruction, efficient Single Instruction, Multiple Data (SIMD) ergonomics, and advanced compiler optimizations. We love the RustLang system and developer ecosystem, ie RustLang is why this book uses mdBook ... but over time, we might find that we like Mojo and the Modular platform even more.
- The extensibility of open source enables its key feature, the strengthening and deepening of the interaction in the development community surrounding an open source project. One learns most, best, fastest by attempting to teach others and trying to understand their learning process. People will fail to understand, fail to adopt, fail to use because the technology is inheritly failure prone, but our intention must be to learn from failure -- in fact, the principle must be to fail fast, in order to learn faster. Everything in this curriculum is an experiment intended push envelopes in order to court failure.
- **Dogfooding new technology is the best way to learn how to develop new technology** and to meet people who are also serious about this objective.
- This 100-day plan adopts a documenation-first direct-to-book methodology, which means that instead of developing a better private note-taking app because so many others are doing that, our focus is on this 100-day plan as the central artifact presented as a living, version-controlled technical book, built with mdBook.. mdBook's key selling point is its speed, safety, and simplicity, its integrated search support and focus on atomic Markdown-based, locally controlled documentation, particularly for technical projects and for getting involved int the Rust programming language and it growing developer ecosystem.
- We are attempting to build something cyclonic, which means that it's ok to spin it up slow somewhere in the hinterlands in total isolation, but maintaining rotational inertia has to matter, ie the PKE system has to be built to feed back useful knowledge to help PKE developers dev the PKE system ... at first, we get the flywheel moving, then maybe try to spin the flywheel a little faster ... but even when we struggle, we stay at it and keep the flywheel spinning every day.
- Reuse, adapt, polish, master FIRST, rather than inventing your own. Instead of rolling our own that's just so or spending money on something with extra features, we will rely upon the GitHub Discussion and Issue and Project functionality, fully exploiting GitHub's ongoing GitHub Projects [along with Discussions and Issues] because these things are sufficient and an adaptable collection of pages, presenting your data, which you can view as a table, a kanban board, or a roadmap and that stays up-to-date with GitHub data. Your GitHub projects can track issues, pull requests, and ideas that you note down AND ... they can, of course, be linked to our own .md pages ... AND anybody else in the world that finds our material useful also has full access to everything GitHub puts out there.
Success Metrics
-
At first, it's simple -- just a matter about completing today's module, while looking forward 10-20 days ahead to see how the work in this Phase sets up the next Phase ... then completing the Phase, looking at full horizon of all 100-days ahead ... thus, generally, not just looking ahead, but updating and revising the 100-module strategic curriculum, and maybe going back and correcting what should have been included in earlier modules ... with a long-term view, informed by the daily experience of showing up, rather than on temporary impatience or whim ... in other words, success of PKE system is not exactly just about how it helps only one highly experienced multi-disciplinary systems engineer, although that's enough ... hopefully, the process will help engineering new opportunities to dogfood something of greater value for others.
-
The primary focus is on this PKE development journey of being much more seriously intentional about the technology of autodidactic learning and dogfooding the technology in order to continually learn better ways to learn and meet new colleagues who share that desire to accelerate learning. The whole point of open source PKE technologies assembled and developed during this journey serving goes beyond the enabling toolkit, but actually uses the process of dogfooding the PKE as well as a means of meeting more colleagues and making new friendships with people who enjoy the journey of continual learning.
-
Whether one is successful in the development of PKE technology will be tough to measure until after the PKE technology has been used, adopted, improved. Success along the way is a matter of just showing up every day to keep the flywheel spinning. The rotational inertia of developing the PKE technology necessarily must be transitted through the larger roadmap and staying focused on that larger picture [which will change as the PKE technology is built].
The 100-Day Personal Knowledge Engineering Curriculum Overview
Phase | Module Range | Core Objective | Key Deliverables |
---|---|---|---|
Phase 1: Foundation & Systems Architecture | Modules 1-20 | To design and build the core infrastructure of the PKES around a publication-first, mdBook-centric workflow. | A fully configured mdBook project serving as a "personal library"; automated content pipelines; a public-facing professional identity hub. |
Phase 2: Horizon Scanning & Deep Learning | Modules 21-50 | To systematically identify, compare, and learn emerging technologies relevant to personal and professional goals through hands-on, failure-tolerant projects documented as book chapters. | An automated tech-trend dashboard; deep-dive projects in selected domains (e.g., Generative AI, Neuromorphic Computing); refreshed mathematical foundations. |
Phase 3: Creation & Contribution | Modules 51-80 | To translate learned knowledge into tangible public artifacts and contribute to the open-source community, using creation as a vehicle for connection. | Multiple open-source project contributions; a portfolio of projects on GitHub; published models on Hugging Face; a series of technical tutorials published in the book. |
Phase 4: Connection & Synthesis | Modules 81-100 | To leverage the published book and other artifacts for networking, establish thought leadership, and synthesize career experience into high-value knowledge products that foster community. | A targeted networking strategy; a personal CRM built as an mdBook extension; a plan for an online tech discussion group; tools for tracking professional opportunities. |
Conclusion
This 100-module curriculum provides a rigorous and systematic pathway for an experienced engineer to build a Personal Knowledge Engineering System centered on the principles of autodidacticism and community. By progressing through the four phases—Foundation, Learning, Creation, and Connection—the engineer will not only acquire skills in the most important modern technologies but will also construct a sustainable, integrated system for continuous professional growth and friendship. The emphasis on rapid, failure-tolerant experimentation, open-source contribution, and value-driven networking is designed to combat the sense of being overwhelmed by providing a clear, actionable framework. The final deliverable is more than a collection of notes and projects; it is a fully operational flywheel that transforms a lifetime of experience into a source of ongoing learning, discoverability, and meaningful connection within the global technology community.
Phase 1: Foundation & Systems Architecture (Modules 1-20)
Objective: To design and build the core technical and philosophical infrastructure of the Personal Knowledge Engineering System. This phase focuses on creating a robust, extensible, and future-proof "personal library" using mdBook, which will serve as the central hub for all subsequent learning, creation, and networking activities. The architectural choices made here are paramount, prioritizing open standards, data ownership, and extensibility to create a system that is not merely used, but can be actively developed and customized over time.
Module 1: Defining the Philosophy - From PKM to PKE
-
Tasks: The initial step is to establish a guiding philosophy. This involves reading and synthesizing seminal texts on modern knowledge work. Critically analyze the distinction between methodologies focused on resource management, such as Tiago Forte's Building a Second Brain (BASB), which excels at organizing information for project-based work, and those focused on idea generation, like Niklas Luhmann's Zettelkasten Method (ZKM), which is a system for working with ideas themselves.[1] The BASB approach is explicitly project-oriented, speaking the "language of action," while the ZKM is project-agnostic, speaking the "language of knowledge".[1] Draft a personal "Knowledge Engineering Manifesto" that codifies the principles for this 100-day endeavor. This document should outline primary goals (e.g., "Learn a new technology stack and meet three new developers through a shared project"), core principles (e.g., "Default to learning in public," "Bias for action and rapid failure over perfect planning," "Prioritize connections over collections"), and success metrics (e.g., "Publish one new chapter per month," "Initiate three 'coffee chat' conversations with new contacts").
-
Deliverable: A MANIFESTO.md file, which will serve as the first chapter of the new mdBook project. This document serves as the strategic charter for the entire system.
Module 2: Architecting the Personal Library
-
Tasks: Design the foundational information architecture for your mdBook project. Instead of a freeform network, mdBook encourages a structured, hierarchical approach from the outset. Use the P.A.R.A. method (Projects, Areas, Resources, Archive) as a conceptual guide to organize the top-level chapters and sections within your book's src directory. For example, create main sections for Areas (long-term interests like "AI Engineering") and Projects (short-term efforts). The Zettelkasten concept of atomic notes can be adapted; each self-contained idea or piece of research becomes a .md page within the book's structure, linked hierarchically in the SUMMARY.md file.
-
Deliverable: A defined folder structure within the mdBook's src directory and a METHODOLOGY.md chapter. This document will detail the rules for creating new pages, the strategy for structuring chapters, and the lifecycle of information as it moves from a rough draft to a published chapter.
Module 3: Tool Selection & Core Setup - mdBook as the Core
-
Tasks: Install Rust and mdBook. Initialize a new book project which will become your central PKES. Familiarize yourself with the core components: the book.toml configuration file, the src directory for Markdown content, and the SUMMARY.md file that defines the book's structure. This "publication-first" approach aligns with the goal of moving directly from notes to a shareable format. As part of this module, create an ARCHITECTURE_ROADMAP.md chapter to brainstorm future extensions, such as building custom Rust-based preprocessors for mdBook to add new features (e.g., special syntax for callouts, dynamic content generation) or exploring high-performance stacks like Modular's Mojo/Max platform for future AI integrations.
-
Deliverable: A functional mdBook project, version-controlled with a private GitHub repository, and an ARCHITECTURE_ROADMAP.md chapter outlining future development paths for the PKES itself.
Module 4: Automating Capture - The Editorial Funnel
-
Tasks: Engineer a pipeline to capture external information for potential inclusion in your book. Since mdBook lacks a direct clipper plugin ecosystem, the workflow will be more deliberate. Create a separate inbox directory outside the mdBook src folder. Configure tools like an RSS reader (e.g., Feedly) with IFTTT/Zapier or custom scripts to automatically save interesting articles, paper abstracts, or email newsletters as raw Markdown files into this inbox. This creates an "editorial funnel." The manual process of reviewing these drafts, refining them, and then consciously moving them into the src directory and adding them to SUMMARY.md becomes a key part of the engineering process, ensuring only curated content makes it into the final publication.
-
Deliverable: An automated information capture pipeline that centralizes external content into a dedicated inbox folder, ready for editorial review and integration into the main mdBook project.
Modules 5-6: Building the Public Face - GitHub and HuggingFace
-
Tasks:
-
Day 5 (GitHub): Treat the GitHub profile as a professional landing page. Overhaul the profile README.md to be a dynamic "brag document".[10] Create distinct sections: "Current Focus," "Core Competencies," "Open Source Contributions," and "Let's Connect." Link prominently to your mdBook (once public), LinkedIn, and Hugging Face profile.
-
Day 6 (Hugging Face): Establish a professional presence on Hugging Face.[12] Create a profile mirroring the branding on GitHub. Explore Models, Datasets, and Spaces. Create a placeholder "Space" to demystify the deployment process.
-
-
Deliverable: Interconnected, professional profiles on GitHub and Hugging Face that serve as the primary public interfaces for the knowledge and artifacts generated by the PKES.
Modules 7-10: The AI-Powered Research Assistant
-
Tasks:
-
Day 7 (arXiv & Alerting): Systematize research monitoring. Use tools like ArXiv Sanity Preserver or a Python script for keyword alerts (e.g., "agentic AI," "neuromorphic computing").[14, 15] Configure these alerts to be saved into your inbox directory from Module 4.
-
Day 8 (AI Summarization): Build a summarization tool with an LLM API (e.g., Gemini). Write a Python script that processes a URL or PDF, extracts key sections, and generates a concise summary in Markdown format, ready to be moved into your book.
-
Day 9 (Papers with Code Integration): Automate tracking state-of-the-art advancements. Use the Papers With Code API to write a script that generates a weekly digest of trending papers in your field as a new Markdown file in your inbox.
-
Day 10 (Building the Research Dashboard): Create a Research Dashboard.md chapter in your mdBook. Since there's no dynamic plugin like Dataview, write a simple Python or shell script that scans your inbox directory for new files or files with a #summarize tag in their frontmatter, and generates a summary list. This script can be run manually to update the dashboard page.
-
-
Deliverable: A semi-automated system for identifying, capturing, summarizing, and tracking relevant scientific literature, feeding a structured editorial pipeline for your knowledge book.
Modules 11-15: Skill Refreshment & Foundational Tooling
-
Tasks:
-
Day 11 (Docker, containerization, setting up Python environments, k8s orchestration, buildah, cloudkernel, Modular platform, MLIR compiler frameworks): Create a standardized, but minimal Dockerfile build process for a data science container (Python, common libraries, PyTorch) to ensure all future projects are harmoniously pythonic and reproducible.
-
Day 12 (Pythonic ecosystem): Explore the pythonic ecosystem, including: a) NumPy, the library for numerical computing and tools for handling large, multi-dimensional arrays and matrices, as well as functions for mathematical operations b) pandas, the library for data manipulation and analysis, providing data structures for handling tabular data, time series data, and more. pandas also includes functions for data cleaning, merging, and reshaping c) SciPy, the library for scientific computing in Python, including tools for optimization, integration, interpolation, and more d) statsmodels, the library for statistical modeling in Python; SciPy provides tools for regression analysis, time series analysis, and more. e) scikit-learn, the library for machine learning in Python. It provides tools for supervised and unsupervised learning, as well as tools for data preprocessing and model selection. f) Matplotlib, library for creating visualizations which provides tools for creating line plots, scatter plots, histograms, and more. g) seaborn, the library for creating statistical visualizations which provides tools for creating heatmaps, scatter plots, and more.
-
Day 13 (Mathematica Deep Dive, complement Pythoic ecosystem): Refresh foundational math concepts (Linear Algebra, Calculus, Probability) using Wolfram Mathematica. Create dedicated notebooks and export key visualizations and formulas as images to be embedded in new chapters of your mdBook; in the future this might involve extending mdBook or GitHub Actions to develop a seamless "write, commit, publish" workflow.
-
Day 14 (Git commands, GitHub, advanced Git, Jujutsu): Review basic Git commands including GitHub Actions, essential for open-source collaboration: interactive rebasing, cherry-picking, submodules.
-
Day 15 (Git workflows, GitButler branching workflows): Master advanced DVCS flow, complex Git/Jujutsu workflows, including GitButler and the role of semantic versioning and conventional commit messages.
-
-
Deliverable: New mdBook chapters documenting refreshed mathematical knowledge, most likely using Python, but possibly also looking at the path for similar investigations with Mathematica and using Wolfram notebooks; a reusable Docker image for ML projects; and demonstrated proficiency in advanced Git workflows.
Modules 16-20: Establishing the Content & Networking Foundation
-
Tasks:
-
Day 16 (Technical Blog Setup): Your mdBook project is your technical blog. Looking into extending the GitHub Actions workflow used to automatically build and deploy your mdBook to GitHub Pages on every push to the main branch. Don't just create a seamless "write, commit, publish" workflow but understand how to extend, alter that infrastructure-as-code.
-
Day 17 (LinkedIn & Professional Framing): Revamp your LinkedIn profile to align with the "Practitioner-Scholar" persona, framing your career as a narrative. Perhaps publish a short article announcing the 100-day learning journey and linking to your newly deployed mdBook.
-
Day 18 (Identifying Communities): Research and identify 3-5 high-signal online communities (subreddits, Discord servers, etc.). Join and observe the culture before participating.
-
Day 19 (Crafting a Mentorship / Partnership Strategy): Develop a dual-pronged mentorship/partnership plan: identify 25-50 potential partners/mentors to learn from, and outline a plan for mentoring others based on your extensive experience.
-
Day 20 (Phase 1 Review & Planning): Conduct a formal review of the first 20 modules. Write a new chapter in your mdBook evaluating the system's architecture. Create a detailed plan for Phase 2, outlining the specific technology domains for deep dives and project objectives.
-
-
Deliverable: A live technical book deployed via GitHub Pages; a professionally framed LinkedIn profile; a curated list of target communities; a formal mentorship strategy chapter; and a detailed, actionable plan for Phase 2.
Module 1: Defining the Philosophy - From PKM to PKE
Deliverables:
First Rev of the MANIFESTO.md File
The MANIFESTO.md file will serve as the landing page for a new mdBook project found at https://markbruns.github.io/PKE/ ... as such, this file will serve as strategic, living document charter for the entire system and we should expect that it will be updated along the way. In a nutshell, the Manifesto describes the reason for the 100-module program which is entirely about attempting to build upon the best Resource Management Methodologies In Personal Knowledge Engineering(PKE), which in turn are basically implementations of improvements in Note Capturing Systems In Personal Knowledge Management(PKM). In other words, the Manifesto describes the approach we will use to improve upon the best practices of PKE by adding AI-assistance, in some cases going back to the best, simplest fundmental note-capturing methods of PKM.
Tasks
The initial step is to establish the basis of the guiding philosophy that will ground the work of all 100 modules ... the purpose of the deliverable MANIFESTO.md file is to lay out the roadmap, even though we know that the roadmap will need to change as we work through the 100 modules.
At first, understanding something about personal knowledge management involves learning about why learning as an adult is so hard or why the way that you were taught to learn in school is now obsolete because there is SO MUCH more new information to learn, so much more knowledge to assimilate every day just to stay even. When we start to understand something about learning to learn ... what learning to learn actually means ... the five core dimensions of learning ... how to diagnose which one or two of these dimension is your biggest learning rate limiter ... and how to start improving on the rate limiting areas immediately, so that we can begin to uncover a new rate limiter ... when it comes to learning, we need to think in terms of learning processes and SYSTEMS ... holistically -- proactively manage factors, barriers, surprises ... prioritizing repeatability -- avoid depending on willpower and motivation ... avoid the temporary quick fix; remove all existing band-aid solutions, ie change habits [which will involve the discomfort of transformation].
After we BEGIN TO understand the systems behind how we learn and how we don't learn ... because all individuals are slightly different and the effectiveness of different processes changes over time, with skill, age, etc ... only THEN we can start to think about why learning now HAS TO include technologies that help us manage time, squeeze more from the time we have and how to not only use but develop or dogfood our new technologies, like various different forms of AI, as aids to synthesize large bodies of seminal texts and collected "wisdom" of crowds.
Given an understanding of why continual learning is so demanding and requires knowledge management technologies, we want to critically analyze the distinction between methodologies focused on resource management for project-based knowledge work, such as Tiago Forte's Building a Second Brain (BASB); Forte teaches these methods using the CirclePlus community learning platform to help subscribers excel at organizing information for project-based work, and different, perhaps what seems to be superficially simpler or more personal approach found in notetaking methodologies focused on idea generation, like Niklas Luhmann's Zettelkasten Method (ZKM) for the hypertextual features of learning now, which is a notetaking system for working directly with ideas themselves.
It is worth spending some time on these different methodologies for resource mgmt and notetaking understanding key patterns, especially something about the key evolutionary patterns in methodologies focused on resource management for project-based knowledge work as well as the the universal patterns of knowledgework that we see in all notetaking methodologies focused on idea generation.
The BASB approach is explicitly a project-oriented system, speaking the "language of action," while the ZKM is project-agnostic, speaking the "language of knowledge" and delves into the details of makeing notes look good ... this is why instead of getting lost in pretty notes with ZKM, we will uses something akin to the BASB systems ... because the BASB method systematically manages information differently than just notetaking apps ... PROJECTS, have goals, reqmts and deadlines ... AREAS are about roles/responsibilities or obligations or capabilities that need to be earnestly developed/upgraded continually ... RESOURCES, mostly finished AREAS for references, but also curated material on ongoing interests, assets, future inspiration or bucket lists, may req continual maintenance and refactoring but, for now, are backburnerable, but usable ... ARCHIVES, inactive matl from P A R that is still relevant [maybe as an example of a bad/rejected idea] but material that shouldn't be used, except for informational purposes.
Understanding the key patterns and their evolution over time helps us understand WHY the technologies that enable, support, sustain these methodologies were almost necessarily extended or dogfooded by people who could not afford to be satisfied with the proprietary technologies that had been built for previous generations of knowledge work.
Modern knowledge work is now necessarily even more competitive and even more aggressively fast-paced than it has been in the old days, ie before 2025. One has to use, develop and extend technology to have a command of deeper and broader realms of knowledge. There is simply no other substitute for continuously developing and dogfooding even better technologies and more efficient, more effective applications of AI-assistance that can be brought to bear on the tasks of knowledge engineering resource management and idea generation.
Module 2: Architecting the Personal Library
Deliverables:
A defined P.A.R.A. method folder structure within the mdBook's src directory and a METHODOLOGY.md which details the rules for creating new pages, the strategy for structuring chapters, and the lifecycle of issues and materials as they progress in the development cycle.
Tasks:
Design the foundational information architecture for your mdBook project. Instead of a freeform network for mdBook adapted to the content of an application, we will use a structured, hierarchical approach from the outset. Use the P.A.R.A. method (Projects, Areas, Resources, Archives) as a conceptual guide to organize the top-level chapters and sections within your book's src directory.
The Zettelkasten concept of atomic notes is also adapted using the GitHub; each self-contained idea or piece of research that is not dismissed at the Issue level becomes a .md page within the book's structure, linked hierarchically in the SUMMARY.md file, starting in Projects folder, then moved.
We will rely upon the GitHub Discussion and Issue and Project functionality, fully exploiting GitHub's support of this this ... BEFORE graduating something to Project status in our PARA mdBook framework ... thus, it's important to understand the distinctions in the progression from ... Discussion ...to... Issue ...to... Project.
Discussions are mainly for just discussing something, to clarify terminology or ask questions or for just generally speculative thinking out loud.
Issues are for things that somebody really needs to look into and possibly turn into more of a Project.
The GitHub Project functionality is concurrent with the PROJECT status in our mdBook ... GitHub rather than mdBook is used because a GitHub project is adaptable collection of items that you can view as a table, a kanban board, or a roadmap and that stays up-to-date with GitHub data. Your GitHub projects can track issues, pull requests, and ideas that you note down and they can, of course, be linked to the .md page that an mdBook PROJECT has. PROJECTS are for big issues which are things that somebody really needs to look into and attempt to develop as more of a Project.
Graduating to Project status is the start of a bigger development commitment and the basis of the P.A.R.A. method of the Building a Second Brain (BASB) methodology.
P.A.R.A and BASB
As you will recall, the Building a Second Brain BASB method systematically manages information differently than just notetaking apps ... PROJECTS, have goals, reqmts and deadlines ... AREAS are about roles/responsibilities or obligations or capabilities that need to be earnestly developed ... RESOURCES, mostly finished AREAS, but also ongoing interests, assets, future inspiration, may req continual maintenance and refactoring but, for now, are backburnerable ... ARCHIVES, inactive matl from P A R that shouldn't be used, except for informational purposes.
Module 3: Tool Selection & Core Setup - mdBook as the Core
Deliverables
The mdBook project is now a minimally-functional work-in-progress at https://markbruns.github.io/PKE/ ... as such, it is version-controlled with a public GitHub repository at https://github.com/MarkBruns/PKE.
It includes ROADMAP.md chapter outlines the architecture of this specific PKE project itself, as well as the future development path outlined in the more intensive year-long CLOUDKERNEL.md development course for putting together pivotally-important AI/ML ops infrastructure that PKES will be use as the base of all its development.
Tasks
Explore the Rust ecoystem, particularly Hermit OS and various AI-aware Rust-based development communities to brainstorm future extensions, such as building custom Rust-based preprocessors for mdBook to add new features (e.g., special syntax for callouts, dynamic content generation).
AI-Aware Rust-Based Development Communities for ML/AI Infrastructure
Rust's ecosystem is increasingly supporting AI/ML through its focus on performance, safety, and concurrency, making it ideal for infrastructure that enhances ML/AI operations (MLOps) in areas like speed (e.g., via efficient computation and unikernels), security (e.g., memory safety and verifiable code), monitoring (e.g., observability tools), robustness (e.g., reliable pipelines), and predictability (e.g., deterministic execution). Below, I list as many distinct communities as possible, drawn from active open-source projects, forums, and curated resources. These are "AI-aware" in that they explicitly target or integrate ML/AI workloads, often with extensions for GPUs, distributed systems, or MLOps tools. Each entry includes the community's focus, relation to ML/AI ops improvements, and engagement details (e.g., GitHub activity, contributors, discussions).
I've prioritized diversity across infrastructure layers: kernels/unikernels (for secure, lightweight execution), frameworks/libraries (for model building/training), tools (for MLOps pipelines), and meta-communities (curated lists/forums). Communities are serious, with ongoing development, contributors, and issues/discussions.
1. Hermit OS Community (Example Provided)
- Focus: Rust-based lightweight unikernel for scalable, virtual execution environments, including kernel, bootloader, and hypervisors like uhyve.
- AI/ML Relation: Enhances speed and security for AI/ML via GPU acceleration (e.g., Cricket for RustyHermit) and minimal attack surfaces; suitable for predictable, robust cloud/edge AI ops.
- Community Details: GitHub (https://github.com/hermit-os) with 102+ issues (5 "help wanted"), 45 in uhyve; active contributors (~10-20 across repos); discussions via Zulip (https://hermit.zulipchat.com/); RWTH Aachen University-backed, open for PRs.
2. Linfa Community
- Focus: Comprehensive Rust ML framework with algorithms for clustering, regression, and more; akin to scikit-learn but optimized for Rust's safety.
- AI/ML Relation: Improves robustness and predictability via type-safe, performant implementations; supports monitoring through integrated metrics; used for faster ML ops in production (e.g., 25x speedup over Python equivalents).
- Community Details: GitHub (https://github.com/rust-ml/linfa) with 740+ issues (28% open), 150+ contributors; active forks (450+); discussions on Rust forums (e.g., https://users.rust-lang.org/t/is-rust-good-for-deep-learning-and-artificial-intelligence/22866); tutorials and workshops encourage contributions.
3. Burn Community
- Focus: Dynamic deep learning framework in Rust, supporting tensors, autodiff, and GPU backends.
- AI/ML Relation: Boosts speed (GPU/CPU optimization) and security (memory-safe); enables robust, monitorable training pipelines; targets MLOps for scalable AI inference.
- Community Details: GitHub (https://github.com/burn-rs/burn) with 740+ issues (28% open), 150+ contributors; Discord for discussions; integrated with Rust ML working group; high activity (9.1K stars, regular updates).
4. Candle Community (Hugging Face Rust ML)
- Focus: Minimalist ML framework by Hugging Face, emphasizing ease and performance for inference.
- AI/ML Relation: Enhances speed (GPU support) and predictability (static compilation); secure for edge AI ops; used in MLOps for lightweight, monitorable deployments.
- Community Details: GitHub (https://github.com/huggingface/candle) with active issues/PRs; part of Hugging Face's Rust ecosystem (e.g., tokenizers-rs); discussions on Hugging Face forums and Rust ML channels; 150+ contributors.
5. Tract Community (ONNX Runtime in Rust)
- Focus: Rust implementation of ONNX runtime for model inference.
- AI/ML Relation: Improves speed and robustness for cross-framework AI ops; secure, predictable execution; supports monitoring via perf tools.
- Community Details: GitHub (https://github.com/snipsco/tract) with issues/PRs; integrated with Rust ML lists; discussions on Rust users forum; smaller but active (280+ stars).
6. DF DX Community
- Focus: Shape-checked tensors and neural networks in Rust.
- AI/ML Relation: Enhances predictability (compile-time checks) and security (no runtime errors); faster for DL ops; robust for MLOps pipelines.
- Community Details: GitHub (https://github.com/coreylowman/dfdx) with 1.7K stars, issues; Rust ML Discord; contributions via PRs (1.7K stars, active).
7. Unikraft Community
- Focus: Posix-like unikernel with Rust support, modular for custom OS builds.
- AI/ML Relation: Faster, secure AI ops via minimal kernels; GPU extensions for ML; robust, monitorable for cloud AI.
- Community Details: GitHub (https://github.com/unikraft/unikraft) with 140+ issues (31% open), 28 contributors; Xen Project incubator; Discord for discussions; active (growing community).
8. RustyHermit Community
- Focus: Extension of Hermit with enhanced features like GPU support.
- AI/ML Relation: Secure, predictable unikernel for AI/ML; focuses on robustness in HPC/AI environments.
- Community Details: GitHub forks/extensions of Hermit; discussions in Rust internals (https://internals.rust-lang.org/t/unikernels-in-rust/2494); community via Zulip; academic contributions.
9. Enzyme Community
- Focus: High-performance auto-differentiation for LLVM/MLIR in Rust.
- AI/ML Relation: Speeds up ML training (autodiff); robust for predictable gradients; secure via no_std.
- Community Details: GitHub (https://github.com/EnzymeAD/Enzyme) with 1.3K stars, issues; Rust ML forums; contributions encouraged (1.3K stars).
10. Rain Community
- Focus: Framework for large distributed pipelines in Rust.
- AI/ML Relation: Robust, monitorable ML ops; faster distributed training; secure for scalable AI.
- Community Details: GitHub (https://github.com/rain-ml/rain) with 750 stars, issues; part of Rust ML ecosystem; discussions on forums.
11. Rust ML Working Group
- Focus: Unofficial group advancing ML in Rust, curating resources.
- AI/ML Relation: Oversees infrastructure for faster, secure ML ops; promotes robustness via standards.
- Community Details: GitHub (https://github.com/rust-ml); forums (https://users.rust-lang.org/c/domain/machine-learning); active threads on AI/Rust integration.
12. Awesome-Rust-MachineLearning Community
- Focus: Curated list of Rust ML libraries, blogs, and resources.
- AI/ML Relation: Aggregates tools for secure, fast MLOps; aids predictability via best practices.
- Community Details: GitHub (https://github.com/vaaaaanquish/Awesome-Rust-MachineLearning); contributions via PRs; discussions on Reddit/Rust forums; 1K+ stars.
13. Best-of-ML-Rust Community
- Focus: Ranked awesome list of Rust ML libraries.
- AI/ML Relation: Highlights tools for robust, monitorable AI infra; focuses on performance/security.
- Community Details: GitHub (https://github.com/e-tornike/best-of-ml-rust); PRs for updates; tied to Rust ML discussions; 230+ projects curated.
14. AreWeLearningYet Community
- Focus: Comprehensive guide to Rust ML ecosystem.
- AI/ML Relation: Catalogs frameworks/tools for faster, secure ops; emphasizes robustness.
- Community Details: Website (https://www.arewelearningyet.com/); GitHub for contributions; forums for ecosystem growth.
Additional Notes
- Trends (as of Aug 2025): Rust's ML adoption is growing (e.g., xAI uses Rust for AI infra); communities emphasize unikernels for edge AI security/speed.
- Engagement Tips: Join Rust Discord/ML channels or Reddit (r/rust, r/MachineLearning with Rust tags) for cross-community discussions.
- Table of Infrastructure Layers:
Layer | Communities | Key Improvements |
---|---|---|
Kernels/Unikernels | Hermit, Unikraft, RustyHermit | Speed (minimal overhead), Security (isolated), Predictability (deterministic boot) |
Frameworks/Libraries | Linfa, Burn, Candle, Tract, DF DX, Enzyme | Robustness (type safety), Monitoring (metrics), Speed (GPU/autodiff) |
Tools/Pipelines | Rain | Monitorable (distributed), Robust (fault-tolerant) |
Meta/Curated | Rust ML WG, Awesome-Rust-ML, Best-of-ML-Rust, AreWeLearningYet | Overall ecosystem for secure, efficient MLOps |
AI-Aware Development Communities for Modular Platform, Mojo, Max, and MLIR
The ecosystem around Modular AI's technologies (Mojo programming language, Max inference platform, and the broader Modular Platform) and MLIR (Multi-Level Intermediate Representation, foundational to many AI compilers) is focused on unifying AI infrastructure. These communities emphasize performance (e.g., GPU/CPU optimizations), security (e.g., verifiable code transformations), monitoring (e.g., traceable compilations), robustness (e.g., extensible dialects), and predictability (e.g., deterministic optimizations). Mojo, as a Python superset, targets seamless AI development; Max accelerates deployment; MLIR enables reusable compiler stacks. Communities are active but emerging, with Modular's tools launched in 2023-2025 and MLIR since 2019 ... the following dev communities are active as of August 2025.
1. Modular Forum Community
- Focus: Official discussion hub for Mojo, Max, and Modular Platform; covers language features, inference optimizations, and ecosystem tools.
- AI/ML Relation: Drives faster AI ops via Mojo's 35,000x Python speedups and Max's GPU scaling; enhances security/robustness through community-driven patches; monitorable via integrated tracing in compilations.
- Community Details: https://forum.modular.com/; 100+ categories (e.g., Installation, Community Projects); active with 1K+ threads, monthly meetings; contributions via PRs to GitHub.
2. Modular Discord Community
- Focus: Real-time chat for developers building with Mojo/Max; includes channels for debugging, feature requests, and hackathons.
- AI/ML Relation: Supports predictable AI workflows (e.g., porting PyTorch to Mojo); secure via shared best practices; robust for distributed training/inference.
- Community Details: Linked from forum.modular.com; 10K+ members; channels like #mojo-general, #max-support; high activity with daily discussions and Q&A.
3. Modular GitHub Organization
- Focus: Open-source repos for Modular Platform (includes Max & Mojo); collaborative development of AI libraries/tools.
- AI/ML Relation: Accelerates ML ops with open-sourced code (450K+ lines in 2025); robust/predictable via MLIR-based transformations; monitorable through benchmarks.
- Community Details: https://github.com/modular; 5K+ stars across repos; 200+ issues/PRs; contributors ~100; tied to community license for extensions.
4. Modular Community Meetings (YouTube/Forum)
- Focus: Monthly livestreams/recaps on updates like Mojo regex optimizations, GSplat kernels, Apple GPU support.
- AI/ML Relation: Focuses on faster/more robust AI (e.g., large-scale batch inference); predictable via roadmaps; monitorable with demos/benchmarks.
- Community Details: YouTube channel (e.g., Modular Community Meeting #15); forum announcements; 2-5K views per video; interactive Q&A.
5. Reddit r/ModularAI (Unofficial)
- Focus: Discussions on Mojo in real projects, comparisons to Julia/Rust, and Max licensing.
- AI/ML Relation: Explores secure/robust AI frameworks; community critiques hype vs. performance for predictable ops.
- Community Details: https://www.reddit.com/r/modularai/; 1K+ members; threads like "Mojo/Modular in real projects" (Sep 2024); cross-posts from r/MachineLearning.
6. MLIR LLVM Community
- Focus: Core MLIR development under LLVM; dialects, optimizations, and integrations.
- AI/ML Relation: Foundational for AI compilers (e.g., TensorFlow/XLA); enables faster ops via multi-level transformations; secure/robust with meritocratic contributions; monitorable through tracepoints.
- Community Details: https://mlir.llvm.org/community/; Discourse forums, mailing lists (mlir-dev@lists.llvm.org), Discord; GitHub (llvm/llvm-project); 1K+ contributors; monthly meetings.
7. OpenXLA Community
- Focus: Collaborative MLIR-based compiler for AI (e.g., JAX/TensorFlow/PyTorch).
- AI/ML Relation: Democratizes AI compute with hardware-independent optimizations; faster/secure via open partnerships; robust for GenAI.
- Community Details: https://openxla.org/; GitHub (openxla/xla); monthly meetings; partners like Google/AMD; active issues/PRs.
8. TensorFlow MLIR Integration Community
- Focus: MLIR dialects for TensorFlow graphs, quantization, and deployment.
- AI/ML Relation: Boosts predictable/monitorable ML ops (e.g., perf counters); robust for edge AI; secure via unified IR.
- Community Details: https://www.tensorflow.org/mlir; GitHub (tensorflow/mlir); forums tied to TensorFlow Discourse; 500+ contributors.
9. Tenstorrent MLIR Compiler Community (tt-mlir)
- Focus: MLIR dialects for Tenstorrent AI accelerators; graph transformations.
- AI/ML Relation: Speeds up AI hardware abstraction; robust/predictable for custom chips; monitorable via compiler tools.
- Community Details: https://github.com/tenstorrent/tt-mlir; 100+ stars; issues/PRs; part of broader MLIR users.
10. AMD MLIR-AIE Community
- Focus: MLIR for AMD AI Engines (AIE); configurable compute.
- AI/ML Relation: Enhances robust/scalable AI on FPGAs; faster via hardware-specific opts; predictable with end-to-end flows.
- Community Details: Part of mlir.llvm.org/users; GitHub extensions; papers/forums on AMD devs.
11. PolyMage Labs Community
- Focus: MLIR-based PolyBlocks for AI frameworks (PyTorch/TensorFlow/JAX).
- AI/ML Relation: Modular compiler blocks for faster/multi-hardware AI; secure/robust via abstractions.
- Community Details: https://www.polymagelabs.com/; GitHub repos; community-driven extensions; IISc-incubated.
12. Google MLIR Users/Researchers
- Focus: MLIR in XLA/TFLite; research on AI infrastructure.
- AI/ML Relation: Addresses Moore's Law end with reusable stacks; faster/secure for billions of devices.
- Community Details: Google Blog posts; arXiv papers; tied to LLVM/MLIR forums; collaborative with Modular.
Additional Notes
- Trends (August 2025): Modular's 25.5 release emphasizes scalable inference; MLIR sees growth in GenAI (e.g., CUDA alternatives). Communities overlap (e.g., Modular uses MLIR); X discussions highlight Mojo's Python edge for AI.
- Engagement Tips: Join Modular Forum/Discord for starters; LLVM Discourse for MLIR deep dives.
- Table of Infrastructure Layers:
Layer | Communities | Key Improvements |
---|---|---|
Language/Platform (Mojo/Max) | Modular Forum, Discord, GitHub, Community Meetings, Reddit r/ModularAI | Speed (35Kx Python), Robustness (extensible), Predictability (roadmaps) |
Compiler Infrastructure (MLIR) | MLIR LLVM, OpenXLA, TensorFlow MLIR, tt-mlir, MLIR-AIE, PolyMage | Security (verifiable IR), Monitoring (traceable opts), Scalability (hardware-agnostic) |
Research/Extensions | Google MLIR Users | Overall AI ops unification for efficiency/robustness |
Module 4: Automating Capture - The Editorial Funnel
Deliverable:
Start with a GitHub Discussion on "Engineering an automated information capture pipeline for mdBook" ... then, after some rumination, selected the best approach, explain why minimalist approach was selected and open a GitHub Issue ... and immediately upgrade this particular issue to a larger GitHub Project, to develop a roadmap ... and also explore the practicality of GitHub Discussion, Issue and Project functionailities, ie one Deliverable is only a meta-deliverable, just gaining experience using GitHub for this.
What is really needed, first of all, is roadmap that articulates or lays out the specifics of most rudimentary beginnings of thinking on an automated information capture pipeline after developing and using a minimalist pipeline in a more manual fashion, these tasks will eventually be automated in a way that centralizes external content into a dedicated inbox folder, ready for editorial review and integration into the main mdBook project.
Tasks:
In general, the assignment is to engineer an automated information capture pipeline to capture external information for potential inclusion in your book. Since mdBook lacks a direct clipper plugin ecosystem, the workflow will be more deliberate. Create a separate inbox directory outside the mdBook src folder. Configure tools like an RSS reader (e.g., Feedly) with IFTTT/Zapier or CRM pipeline tools or custom scripts to automatically save interesting articles, paper abstracts, or email newsletters as raw Markdown files into this inbox. This creates an "editorial funnel." The manual process of reviewing these drafts, refining them, and then consciously moving them into the src directory and adding them to SUMMARY.md becomes a key part of the engineering process, ensuring only curated content makes it into the final publication.
Module 5 of Modules 5-6: Building the Public Face - GitHub and HuggingFace
Deliverable
Interconnected, professional profiles on GitHub and Hugging Face that serve as the primary public interfaces for the knowledge and artifacts generated by the PKES.
-
Tasks:
- Day 5 (GitHub): Treat the GitHub profile as a professional landing page. Overhaul the profile README.md to be a dynamic "brag document". Create distinct sections: "Current Focus," "Core Competencies," "Open Source Contributions," and "Let's Connect." Link prominently to your mdBook (once public), LinkedIn, and Hugging Face profile ... leading to the interconnected, professional profiles on GitHub and Hugging Face that serve as the primary public interfaces for the knowledge and artifacts generated by the PKES ... ideally the GitHub profile README would pull from PKE repos such as https://github.com/MarkBruns/daily-arXiv-ai-enhanced that were config'd to produce something like https://dw-dengwei.github.io/daily-arXiv-ai-enhanced/
Module 6 of Modules 5-6: Building the Public Face - GitHub and HuggingFace
Deliverable
Interconnected, professional profiles on GitHub and Hugging Face that serve as the primary public interfaces for the knowledge and artifacts generated by the PKES.
Tasks:
- Day 6 (Hugging Face): Establish a professional presence on Hugging Face. Create a profile mirroring the branding on GitHub. Explore Models, Datasets, and Spaces. Create a placeholder "Space" to demystify the deployment process. Treat both profiles as a professional landing page. Link prominently to your mdBook (once public), LinkedIn, and Hugging Face profile ... leading to the interconnected, professional profiles on GitHub and Hugging Face that serve as the primary public interfaces for the knowledge and artifacts generated by the PKES ... ideally the GitHub profile README would pull from PKE repos such as https://github.com/MarkBruns/daily-arXiv-ai-enhanced that were config'd to produce something like https://dw-dengwei.github.io/daily-arXiv-ai-enhanced/
Module 7 of Modules 7-10: The AI-Powered Research Assistant
-
Tasks:
- Day 7 (arXiv & Alerting): Systematize research monitoring. Use tools like ArXiv Sanity Preserver or a Python script for keyword alerts (e.g., "agentic AI," "neuromorphic computing").[14, 15] Configure these alerts to be saved into your inbox directory from Module 4.
-
Deliverable: A semi-automated system for identifying, capturing, summarizing, and tracking relevant scientific literature, feeding a structured editorial pipeline for your knowledge book.
Module 8 of Modules 7-10: The AI-Powered Research Assistant
-
Tasks:
- Day 8 (AI Summarization): Build a summarization tool with an LLM API (e.g., Gemini). Write a Python script that processes a URL or PDF, extracts key sections, and generates a concise summary in Markdown format, ready to be moved into your book.
-
Deliverable: A semi-automated system for identifying, capturing, summarizing, and tracking relevant topics in ML/AI Ops Engineering, robotics, tech job search and relevant areas of emerging interest in scientific literature, feeding a structured editorial pipeline for your knowledge book.
The following list is an example of top trends in ML/AI Ops Engineering.
-
Observability in AI pipelines has gained significant attention due to the increasing complexity of deploying large-scale models, necessitating real-time monitoring to detect anomalies and ensure reliable performance in production environments.
-
LLM-aware data engineering is crucial as it addresses the unique data requirements of large language models, optimizing ingestion and processing pipelines to handle vast, unstructured datasets efficiently.
-
Cost optimization for AI inference is a hot topic because escalating computational expenses in cloud environments demand innovative strategies like model compression and efficient resource allocation to make AI scalable for enterprises.
-
Context engineering for agents is significant as it enhances the ability of AI agents to maintain and utilize long-term memory, improving decision-making in dynamic, multi-step tasks.
-
Workflow engineering in AI systems is worthy of attention for streamlining the integration of various AI components, reducing latency and errors in end-to-end automation processes.
-
Mixture-of-Experts (MoE) models are heavily discussed for their efficiency in scaling AI capabilities by activating only subsets of parameters, leading to better performance with lower resource usage.
-
Self-evolving agents are gaining traction as they enable AI systems to adapt and improve autonomously through reinforcement learning, pushing the boundaries of unsupervised intelligence.
-
Continuous training (CT) pipelines are essential for keeping models up-to-date with evolving data distributions, preventing performance degradation in real-world applications.
-
Feature stores for ML are significant because they centralize feature management, ensuring consistency across training and inference while accelerating development cycles.
-
Data quality assurance for foundation models is critical as poor data can lead to biased or inaccurate outputs, prompting advancements in validation techniques to build more trustworthy AI.
-
Vector databases like Chroma are popular for their role in enabling fast similarity searches in high-dimensional spaces, foundational for retrieval-augmented generation in AI systems.
-
Modular platforms (e.g., Mammoth, Mojo, Max) are noteworthy for providing flexible, composable tools that speed up AI development and deployment across diverse hardware.
-
Rust for LLM security and speed is gaining attention due to its memory-safe properties and performance advantages, making it ideal for building secure, efficient AI infrastructure.
-
Topics beyond LLMs, such as traditional ML ops, are significant as they remind practitioners of foundational practices like ensemble methods and classical algorithms that remain vital in hybrid systems.
-
ArXiv hot topics in MLOps highlight emerging research on scalable orchestration, influencing industry practices by bridging academic innovations with practical implementations.
-
Decentralized model training is important for privacy-preserving AI, allowing collaborative learning without central data aggregation, especially in regulated sectors like healthcare.
-
Edge AI deployment is a key focus as it enables low-latency inference on devices, reducing reliance on cloud resources and expanding AI applications in IoT and mobile environments.
-
Multi-modal AI operations are significant for handling diverse data types like text, images, and audio, enabling more comprehensive AI solutions in fields like autonomous driving.
-
Agentic AI frameworks are worthy of inclusion due to their support for building autonomous agents that can plan and execute complex tasks, revolutionizing automation.
-
Retrieval-Augmented Generation (RAG) scaling is hot because it addresses context limitations in LLMs by integrating external knowledge bases, improving accuracy in knowledge-intensive tasks.
-
Model routing in production is crucial for directing queries to the most appropriate models, optimizing for cost, speed, and expertise in multi-model ecosystems.
-
Security guardrails for LLMs are gaining attention amid rising concerns over adversarial attacks and data leaks, ensuring safer deployment in sensitive applications.
-
Evaluation techniques for agents are significant as they provide robust metrics for assessing reasoning, adaptability, and reliability in non-deterministic environments.
-
Monitoring AI drift is essential to detect shifts in data or model behavior over time, maintaining long-term efficacy in deployed systems.
-
Ethical AI in operations is a pressing topic as it integrates fairness and accountability into MLOps pipelines, responding to societal demands for responsible AI.
-
Hyper-personalization in AI is noteworthy for leveraging user data to tailor experiences, driving engagement in e-commerce and content recommendation systems.
-
Conversational AI ops are significant for managing dialogue systems at scale, improving natural language understanding and response generation in chatbots.
-
AI for creative industries is gaining traction by automating content generation in art, music, and writing, sparking debates on creativity and intellectual property.
-
Intelligent automation trends are important as they combine AI with RPA to streamline business processes, enhancing efficiency across industries.
-
AI ethics and regulation is a hot area due to upcoming global policies that require compliance frameworks in MLOps to mitigate risks like discrimination.
-
Quantum AI integration is significant for its potential to solve complex optimization problems faster, influencing future MLOps in hybrid quantum-classical setups.
-
Small Language Models (SLMs) for agents are worthy of attention for their efficiency in edge devices, offering lightweight alternatives to LLMs with comparable performance in specific domains.
-
Test-time scaling is crucial as it allows models to adapt during inference, improving robustness without retraining.
-
Long-horizon task handling is gaining focus for enabling AI agents to plan over extended sequences, vital for real-world applications like robotics.
-
Flash attention optimizations are significant for reducing memory and computational overhead in transformer models, accelerating training and inference.
-
RMSNorm in models is noteworthy as an alternative to LayerNorm, providing stability and efficiency in deep networks.
-
RoPE positional encodings are important for extending context lengths in LLMs without losing positional information.
-
YaRN for context extension is hot due to its innovative approach to handling ultra-long sequences, expanding LLM capabilities.
-
Reinforcement learning from human feedback (RLHF) is essential for aligning models with human values, improving safety and usability.
-
Deliberative alignment for safety is gaining attention as it incorporates reasoning steps to ensure ethical decision-making in AI.
-
Data pipelines in ML systems are significant for automating data flow from ingestion to modeling, reducing manual errors.
-
Schema validation in AI workflows is crucial to ensure data consistency, preventing downstream issues in model training.
-
Semantic correctness checks are worthy for verifying meaning in data, enhancing the reliability of AI outputs.
-
Human-in-the-loop validators are important for incorporating expert oversight, improving model accuracy in ambiguous cases.
-
GPU cost management is a key topic as AI workloads explode, requiring strategies to optimize utilization and reduce expenses.
-
Prompt caching is significant for speeding up repeated queries in LLMs, lowering latency and costs.
-
Multi-modal models are gaining traction for integrating vision and language, enabling advanced applications like image captioning.
-
Model adaptation techniques are essential for fine-tuning pre-trained models to specific tasks, accelerating deployment.
-
Storage for retrieval (vector/graph DBs) is noteworthy for supporting complex queries in knowledge graphs.
-
Hybrid retrieval systems are important for combining vector and keyword search, improving information retrieval accuracy.
-
RAG and agentic RAG are hot as they enhance LLMs with dynamic knowledge access, making agents more informed.
-
LLM orchestration frameworks are significant for managing multiple models in workflows, ensuring seamless integration.
-
AI agent design patterns are gaining attention for providing reusable architectures, speeding up development.
-
Multi-agent systems are crucial for collaborative AI, simulating team dynamics in problem-solving.
-
Memory in agents is worthy as it enables persistence of information across interactions, improving continuity.
-
Human-in/on-the-loop is important for balancing autonomy with oversight, ensuring safety in critical applications.
-
Agent communication protocols are significant for enabling interoperability between agents, fostering ecosystems.
-
Kubernetes for AI infra is hot due to its scalability for containerized ML workloads.
-
CI/CD for ML is essential for automating model updates, reducing deployment risks.
-
Model routing security is gaining focus to prevent unauthorized access in shared environments.
-
Guardrails and red teaming are important for testing AI vulnerabilities, enhancing robustness.
-
Voice and vision agents are noteworthy for multimodal interactions, expanding AI interfaces.
-
Robotics agents are significant for physical world applications, integrating perception and action.
-
Computer use agents are gaining traction for automating software tasks, boosting productivity.
-
CLI agents are crucial for command-line automation, aiding developers.
-
Automated prompt engineering is hot as it optimizes inputs for better LLM performance.
-
Experiment tracking is essential for reproducibility in ML research.
-
ML metadata stores are worthy for managing experiment data, facilitating collaboration.
-
Pipeline orchestration (Airflow, Kubeflow) is important for scheduling complex workflows.
-
Data and model validation is significant to ensure integrity before deployment.
-
Ad-hoc and scheduled triggers are gaining attention for flexible pipeline execution.
-
Training/serving skew is crucial to address discrepancies that affect performance.
-
Model performance profiling is noteworthy for identifying bottlenecks.
-
Latency bottlenecks are important to optimize for real-time AI.
-
Inference cost management is hot amid rising hardware demands.
-
Model quantization is essential for deploying on resource-constrained devices.
-
Distillation techniques are significant for creating efficient models from larger ones.
-
MLOps with DevOps integration is gaining traction for unified practices.
-
Autoscaling in ML is crucial for handling variable workloads.
-
Failover and redundancy are worthy for high-availability AI systems.
-
Microservices in AI are important for modular, scalable architectures.
-
Database design for ML is significant for efficient data handling.
-
Scalability patterns are gaining focus in large-scale deployments.
-
Performance metrics are essential for benchmarking AI systems.
-
Loss functions in ops are noteworthy for guiding model improvements.
-
Bias and fairness monitoring is hot due to ethical concerns.
-
Explainability in production is crucial for trust in AI decisions.
-
Hyperparameter tuning at scale is significant for optimizing models.
-
Federated learning ops are important for privacy in distributed training.
-
Privacy-preserving ML is gaining attention in data-sensitive fields.
-
Sustainable AI practices are worthy for reducing environmental impact.
-
Energy-efficient training is essential amid climate concerns.
-
Cloud-native MLOps is hot for leveraging cloud services.
-
Lakehouse for AI is significant for unified data and analytics.
-
Data mesh principles are gaining traction for decentralized data management.
-
Governance in MLOps is crucial for compliance and control.
-
Model versioning (DVC, MLflow) is important for tracking changes.
-
CI/CD for agents is noteworthy for agile agent development.
-
GPU scheduling is significant for resource sharing.
-
Monitoring tools (Prometheus, Grafana) are hot for visualizing AI metrics.
Module 9 of Modules 7-10: The AI-Powered Research Assistant
-
Tasks:
- Day 9 (PapersWithCode Integration): Automate tracking state-of-the-art advancements. Use the PapersWithCode API to write a script that generates a weekly digest of trending papers in your field as a new Markdown file in your inbox.
-
Deliverable: A semi-automated system for identifying, capturing, summarizing, and tracking relevant scientific literature, feeding a structured editorial pipeline for your knowledge book.
##Module 10 of Modules 7-10: The AI-Powered Research Assistant**
Deliverable:
A Research Dashboard for coordinating a semi-automated system for identifying, capturing, summarizing, and tracking relevant scientific literature, feeding a structured editorial pipeline for your knowledge book.
Tasks:
- Create the Research Dashboard chapter in your mdBook. Since there's no dynamic plugin like Dataview, write a simple Python or shell script that scans your inbox directory for new files or files with a #summarize tag in their frontmatter, and generates a summary list. This script can be run manually to update the dashboard page.
Module 11 of Modules 11-15: Skill Refreshment & Foundational Tooling
-
Tasks:
- Day 11 (Docker, containerization, setting up Python environments, k8s orchestration, buildah, cloudkernel, Modular platform, MLIR compiler frameworks): Create a standardized, but minimal Dockerfile build process for a data science container (Python, common libraries, PyTorch) to ensure all future projects are harmoniously pythonic and reproducible.
-
Deliverable: New mdBook chapters documenting refreshed mathematical knowledge, most likely using Python, but possibly also looking at the path for similar investigations with Mathematica and using Wolfram notebooks; a reusable Docker image for ML projects; and demonstrated proficiency in advanced Git workflows.
Module 12 of Modules 11-15: Skill Refreshment & Foundational Tooling
-
Tasks:
- Day 12 (Pythonic ecosystem): Explore the pythonic ecosystem, including: a) NumPy, the library for numerical computing and tools for handling large, multi-dimensional arrays and matrices, as well as functions for mathematical operations b) pandas, the library for data manipulation and analysis, providing data structures for handling tabular data, time series data, and more. pandas also includes functions for data cleaning, merging, and reshaping c) SciPy, the library for scientific computing in Python, including tools for optimization, integration, interpolation, and more d) statsmodels, the library for statistical modeling in Python; SciPy provides tools for regression analysis, time series analysis, and more. e) scikit-learn, the library for machine learning in Python. It provides tools for supervised and unsupervised learning, as well as tools for data preprocessing and model selection. f) Matplotlib, library for creating visualizations which provides tools for creating line plots, scatter plots, histograms, and more. g) seaborn, the library for creating statistical visualizations which provides tools for creating heatmaps, scatter plots, and more.
-
Deliverable: New mdBook chapters documenting refreshed mathematical knowledge, most likely using Python, but possibly also looking at the path for similar investigations with Mathematica and using Wolfram notebooks; a reusable Docker image for ML projects; and demonstrated proficiency in advanced Git workflows.
Module 13 of Modules 11-15: Skill Refreshment & Foundational Tooling
-
Tasks:
- Day 13 (Mathematica Deep Dive, complement Pythoic ecosystem): Refresh foundational math concepts (Linear Algebra, Calculus, Probability) using Wolfram Mathematica. Create dedicated notebooks and export key visualizations and formulas as images to be embedded in new chapters of your mdBook; in the future this might involve extending mdBook or GitHub Actions to develop a seamless "write, commit, publish" workflow.
-
Deliverable: New mdBook chapters documenting refreshed mathematical knowledge, most likely using Python, but possibly also looking at the path for similar investigations with Mathematica and using Wolfram notebooks; a reusable Docker image for ML projects; and demonstrated proficiency in advanced Git workflows.
Module 14 of Modules 11-15: Skill Refreshment & Foundational Tooling
-
Tasks:
- Day 14 (Git commands, GitHub, advanced Git, Jujutsu): Review basic Git commands including GitHub Actions, essential for open-source collaboration: interactive rebasing, cherry-picking, submodules.
-
Deliverable: New mdBook chapters documenting refreshed mathematical knowledge, most likely using Python, but possibly also looking at the path for similar investigations with Mathematica and using Wolfram notebooks; a reusable Docker image for ML projects; and demonstrated proficiency in advanced Git workflows.
Module 15 of Modules 11-15: Skill Refreshment & Foundational Tooling
-
Tasks:
- Day 15 (Git workflows, GitButler branching workflows): Master advanced DVCS flow, complex Git/Jujutsu workflows, including GitButler and the role of semantic versioning and conventional commit messages.
-
Deliverable: New mdBook chapters documenting refreshed mathematical knowledge, most likely using Python, but possibly also looking at the path for similar investigations with Mathematica and using Wolfram notebooks; a reusable Docker image for ML projects; and demonstrated proficiency in advanced Git workflows.
Module 16 of Modules 16-20: Establishing the Content & Networking Foundation
-
Tasks:
- Day 16 (Technical Blog Setup): Your mdBook project is your technical blog. Looking into extending the GitHub Actions workflow used to automatically build and deploy your mdBook to GitHub Pages on every push to the main branch. Don't just create a seamless "write, commit, publish" workflow but understand how to extend, alter that infrastructure-as-code.
-
Deliverable: A live technical book deployed via GitHub Pages; a professionally framed LinkedIn profile; a curated list of target communities; a formal mentorship strategy chapter; and a detailed, actionable plan for Phase 2.
Module 17 of Modules 16-20: Establishing the Content & Networking Foundation
-
Tasks:
- Day 17 (LinkedIn & Professional Framing): Revamp your LinkedIn profile to align with the "Practitioner-Scholar" persona, framing your career as a narrative. Perhaps publish a short article announcing the 100-day learning journey and linking to your newly deployed mdBook.
-
Deliverable: A live technical book deployed via GitHub Pages; a professionally framed LinkedIn profile; a curated list of target communities; a formal mentorship strategy chapter; and a detailed, actionable plan for Phase 2.
Module 18 of Modules 16-20: Establishing the Content & Networking Foundation
-
Tasks:
- Day 18 (Identifying Communities): Research and identify 3-5 high-signal online communities (subreddits, Discord servers, etc.). Join and observe the culture before participating.
-
Deliverable: A live technical book deployed via GitHub Pages; a professionally framed LinkedIn profile; a curated list of target communities; a formal mentorship strategy chapter; and a detailed, actionable plan for Phase 2.
Module 19 of Modules 16-20: Establishing the Content & Networking Foundation
-
Tasks:
-
Day 19 (Crafting a Mentorship / Partnership Strategy): Develop a dual-pronged mentorship/partnership plan: identify 25-50 potential partners/mentors to learn from, and outline a plan for mentoring others based on your extensive experience.
-
Deliverable: A live technical book deployed via GitHub Pages; a professionally framed LinkedIn profile; a curated list of target communities; a formal mentorship strategy chapter; and a detailed, actionable plan for Phase 2.
-
Module 20 of Modules 16-20: Establishing the Content & Networking Foundation
-
Tasks:
- Day 20 (Phase 1 Review & Planning): Conduct a formal review of the first 20 modules. Write a new chapter in your mdBook evaluating the system's architecture. Create a detailed plan for Phase 2, outlining the specific technology domains for deep dives and project objectives.
-
Deliverable: A live technical book deployed via GitHub Pages; a professionally framed LinkedIn profile; a curated list of target communities; a formal mentorship strategy chapter; and a detailed, actionable plan for Phase 2.
Phase 2: Horizon Scanning & Deep Learning (Modules 21-50)
Objective: To systematically explore and gain hands-on proficiency in a curated set of emerging technologies. This phase emphasizes active, project-based learning over passive consumption, with a core tenet of embracing rapid failure as a learning mechanism. Each module is designed to produce a tangible artifact—a piece of code, a trained model, a working demo—which serves as both a learning tool and a potential portfolio piece, thereby energizing the PKES flywheel.
Sub-theme: Generative AI & LLMs (Modules 21-30)
This sub-theme focuses on building practical skills in the dominant technology trend of the 2020s. The projects move from foundational theory to building and deploying sophisticated AI applications.
- Module 21: Refresher: Linear Algebra with Python/Mathematica: Revisit Jupyter and Mathematica notebooks from Day 12-13. Focus specifically on the concepts underpinning transformer architectures: vector spaces, dot products (as a measure of similarity), matrix multiplication, and Singular Value Decomposition (SVD). Implement a simple attention mechanism calculation in a notebook to solidify the mathematical intuition.
- Module 22: Building a RAG Application with LlamaIndex: Follow a tutorial to build a complete Retrieval-Augmented Generation (RAG) application.32 Use a personal dataset, such as a collection of past technical reports, articles, or even the notes from this 100-day plan. The goal is to create a question-answering system over this private data. Deploy it locally using a simple FastAPI wrapper. This project provides immediate personal utility and a powerful demonstration of context-augmented LLMs.
- Module 23: Fine-Tuning a Foundational Model: Gain hands-on experience with model customization. Using a framework like Hugging Face's transformers library and a platform with free GPU access like Google Colab, fine-tune a small, open-source LLM (e.g., a member of the Llama 3 or Mistral family) on a specific, narrow task.35 A practical project is to create a dataset of your own commit messages from a key project and fine-tune the model to generate new commit messages in your personal style. This demonstrates an understanding of the full training and tuning loop.
- Module 24: Building an AI Agent with LangChain: Construct a basic autonomous agent that can reason and use tools. Using LangChain or LangGraph, define two tools: a search tool (e.g., Tavily Search) and a code execution tool (e.g., a Python REPL). Create an agent that can answer a question like, "What is the current price of Apple stock and what is its P/E ratio?" by first searching for the price and then using the REPL to calculate the ratio. This project demonstrates the core concepts of agentic workflows.38
- Module 25: Exploring Generative AI in the SDLC: Dedicate a full day to integrating Generative AI into a typical software development workflow. Select an AI-native code editor like Cursor or use GitHub Copilot extensively within your preferred IDE.41 Take on a small coding task (e.g., building a simple web app) and use the AI assistant for every stage: generating boilerplate, writing functions, creating unit tests, explaining unfamiliar code, and writing documentation. Meticulously document the experience in your PKES, noting productivity changes, quality of generated code, and points of friction. This provides a first-hand, critical evaluation of how GenAI is transforming the development lifecycle.43
- Modules 26-30: Project: Build an "AI Research Analyst" Agent: Synthesize the skills from this sub-theme into a multi-day project. Build an autonomous agent that fully automates the workflow designed in Modules 7-10. The agent's task, triggered daily, is to: 1) Fetch new papers from your arXiv feed. 2) For each paper, decide if it's relevant based on a set of criteria. 3) If relevant, summarize the paper using the LLM tool. 4) Check Papers With Code for an associated implementation. 5) Compile the findings into a structured daily brief in Markdown format. 6) Push the Markdown file to a dedicated GitHub repository that powers a section of your technical blog.
Sub-theme: Modern Data Engineering (Modules 31-35)
This sub-theme addresses the shift in data architecture, moving beyond monolithic data warehouses to more flexible, scalable, and decentralized paradigms. For a senior engineer, understanding these system-level trends is crucial.46
- Module 31: End-to-End MLOps with MLflow: Go beyond a simple model.fit() call and embrace the discipline of MLOps. Using a classic dataset like the UCI Wine Quality dataset, train a scikit-learn model, but with a focus on the operational aspects.47 Set up a local MLflow tracking server. In your training script, log hyperparameters, evaluation metrics (e.g., RMSE, MAE), and the trained model itself as an artifact. Use the MLflow UI to compare several runs with different hyperparameters. Finally, register the best-performing model in the MLflow Model Registry, promoting it to a "Staging" or "Production" tag. This project covers the core lifecycle of a managed ML model.48
- Module 32: Data Mesh Proof-of-Concept: Build a small-scale simulation of a data mesh architecture to understand its core principles. Create two separate Python scripts or services. The first, the "Users Domain," generates mock user data and exposes it via a simple API as a "data product." The second, the "Orders Domain," does the same for mock order data. Create a third "Analytics" service that acts as a data consumer, pulling data from both domain APIs to answer a business question (e.g., "What is the average order value for users in California?"). This hands-on exercise demonstrates the principles of decentralized data ownership and data-as-a-product, contrasting it with a centralized data warehouse approach.52
- Modules 33-35: Project: Real-Time Data Processing Pipeline (Comparative Study): Build a small but complete real-time data pipeline. Use a public streaming data source. The core task is to implement a simple consumer and transformation process twice, first using a traditional message queue like Apache Kafka and then using a unified processing framework like Apache Beam. Document the architectural differences, development overhead, and performance trade-offs in your PKES. This comparative approach deepens understanding beyond a single tool.
Sub-theme: The Next Frontiers (Modules 36-45)
This section focuses on gaining conceptual and practical fluency in technologies that represent significant long-term shifts in computing.55 The objective is not mastery but the ability to understand the fundamentals and identify potential future applications.
- Module 36: Quantum Computing Fundamentals (Comparative Study): Demystify the core concepts of quantum computation. Using IBM's Qiskit open-source framework, implement a simple algorithm like creating an entangled Bell state. Then, repeat the same exercise using Google's Cirq framework. Document the differences in syntax, circuit construction, and overall developer experience. This provides a concrete understanding of concepts like superposition and entanglement from the perspective of two major ecosystems.
- Modules 37-38: Neuromorphic & Brain-Computer Interfaces: Shift focus from quantum to another frontier: brain-inspired computing.
- Day 37 (Neuromorphic Concepts): Research the principles of neuromorphic computing and spiking neural networks (SNNs). Investigate current hardware like Innatera's Pulsar and IBM's NorthPole. Create a detailed summary in your PKES comparing the architecture of these chips to traditional von Neumann architectures.
- Day 38 (BCI Exploration): Explore the open-source Brain-Computer Interface (BCI) landscape. Research the hardware and software stacks of OpenBCI 91 and commercial platforms like Emotiv. The goal is to understand the types of data (EEG, EMG) they capture and the kinds of projects the communities are building.
- Modules 39-40: AR/VR for Education & Training: Replace the Web3 focus with an exploration of immersive technologies for learning, aligning with interests in simulation and education.
- Day 39 (Intro to WebXR): Set up a basic development environment for WebXR. Work through a "Hello, World" tutorial to render a simple 3D object in a browser that can be viewed in VR or AR on a compatible device. This provides a low-barrier entry into immersive development.97
- Day 40 (Educational AR/VR Prototype): Brainstorm and create a simple proof-of-concept for an educational AR/VR experience. For example, an AR app that displays a 3D model of a molecule when the phone camera is pointed at a marker, or a simple VR scene that visualizes a mathematical concept. The focus is on rapid prototyping, not a polished application.99
- Modules 41-45: Project: Advanced Frontier Exploration: Select one of the frontier topics (Generative AI, BCI, or AR/VR) and build a more in-depth project.
- AI Option: Build and deploy a multi-modal application (e.g., an image captioning model) to a Hugging Face Space, making it publicly accessible.
- BCI Option: Download a public EEG dataset and use Python libraries to perform basic signal processing and visualization, attempting to identify simple patterns (e.g., eye blinks).
- AR/VR Option: Expand the educational prototype from Day 40, adding more interactivity or information overlays to create a more comprehensive learning module.
Sub-theme: Review & Synthesis (Modules 46-50)
Sub-theme: Review & Synthesis (Modules 46-50)
- Tasks: This process is now even more natural with mdBook. For each major technology explored, create a main chapter that serves as a "Map of Content" (MOC), linking to all the sub-pages (project notes, tutorials, etc.) you've written on the topic. This makes your book's structure itself a tool for synthesis.
- Deliverable: A set of highly organized, interconnected chapters within your mdBook. This transforms the raw learning experience into a structured, searchable, and reusable knowledge asset.
Phase 3: Creation & Contribution (Modules 51-80)
Objective: To transition from internal learning to external creation and contribution. This phase is dedicated to applying the skills and knowledge from Phase 2 to produce public artifacts and make meaningful contributions to the open-source ecosystem. This directly addresses the core goals of becoming "more useful" and "discoverable" by demonstrating expertise through tangible work. The "fail fast, learn faster" philosophy is critical here; the goal is to ship, gather feedback, and iterate.
Sub-theme: Finding Your Niche (Modules 51-55)
The approach for a senior engineer should be strategic, focusing on building relationships and making impactful contributions rather than simply collecting commits. This requires careful selection of a project and a gradual, respectful entry into its community.27
- Module 51: Open Source Contribution Strategy: Identify 3-5 open-source projects that are personally or professionally relevant. These should be tools used daily or libraries central to the technologies explored in Phase 2 (e.g., LangChain, LlamaIndex, MLflow, dbt). For each candidate project, conduct a thorough investigation. Read the CONTRIBUTING.md file, join their primary communication channels (Discord, Slack, mailing list), and observe the dynamics of the community. Analyze the project's governance model to understand how decisions are made and who the key maintainers are.24
- Module 52: Identifying "Good First Issues": Use platforms like goodfirstissue.dev and forgoodfirstissue.github.io or search directly on GitHub for labels like good first issue, help wanted, or beginner-friendly within the target projects.62 The purpose of this exercise is not necessarily to solve these issues, but to analyze them. This provides insight into the project's backlog, the types of tasks available for new contributors, and the clarity of their issue tracking.
- Module 53: Beyond "Good First Issues" - The User-Contributor Path: For an experienced developer, a more impactful entry point is often to solve a problem they have personally encountered while using the software. Spend the day using one of the target projects intensively. Identify a bug, a gap in the documentation, or a minor feature that would improve the user experience. Create a detailed, reproducible issue report on GitHub. This approach leads to authentic contributions that are highly valued by maintainers.
- Module 54: Your First Non-Code Contribution: Make a contribution that builds social capital within the community. Options include: thoroughly improving a section of the official documentation that was confusing, providing a detailed and helpful answer to another user's question in the project's Discord or forum, or taking an existing bug report and adding more detail, such as a minimal reproducible example or root cause analysis. This demonstrates commitment and an understanding of the project without requiring a code change.
- Module 55: Your First Code Contribution: Select a small, well-defined issue—ideally the one identified in Module 53. Follow the project's contribution workflow precisely: fork the repository, create a new branch, make the code changes, add or update tests, and submit a pull request.66 The pull request description should be clear, linking to the original issue and explaining the change and its justification. Be prepared to engage constructively with feedback from maintainers.
Sub-theme: The Creator Track - Technical Content (Modules 56-65)
This sub-theme focuses on leveraging the user's deep experience to teach others, which is a powerful method for solidifying knowledge and building a professional reputation.68
- Modules 56-58: Writing Your First Technical Tutorial: Select one of the hands-on projects from Phase 2 (e.g., "Building a RAG Application with LlamaIndex") and transform the project notes from your PKES into a comprehensive, step-by-step tutorial. The structure should follow best practices: start by explaining the "why" and showing the final result, then walk through the process with clear code snippets and explanations.70 Publish the final article on the technical blog established in Phase 1.
- Modules 59-60: Promoting Your Content: Actively distribute the published tutorial. Share a link on LinkedIn with a summary of what readers will learn. Post it to relevant subreddits or forums, being mindful of community rules on self-promotion. The key is to frame the post as a helpful resource, not an advertisement. Monitor these channels and engage thoughtfully with all comments and questions.
- Modules 61-65: Creating a Video Tutorial: Repurpose the written tutorial into a video format to reach a different audience.
- Day 61: Write a concise script based on the blog post.
- Day 62: Prepare the coding environment for recording (e.g., increase font size, clean up the desktop). Record the screen and audio, walking through the project step-by-step.73
- Day 63-64: Perform basic video editing (e.g., using DaVinci Resolve or Descript) to remove mistakes and add simple titles or callouts.
- Day 65: Upload the video to YouTube, with a clear title, detailed description, and a link back to the original blog post.
Sub-theme: The Builder Track - Capstone Project (Modules 66-80)
This three-week block is dedicated to building a single, more substantial project that synthesizes skills from multiple modules and serves as a significant portfolio piece.
- Project Definition: Personalized arXiv Assistant:
- Modules 66-70 (Data Ingestion & Processing): Build a robust data pipeline that fetches daily papers from a custom arXiv RSS feed. The pipeline should parse the XML, extract metadata (title, authors, abstract), and store it in a local database (e.g., SQLite).
- Modules 71-73 (Custom Classification): Use the skills from Module 23. Create a small, labeled dataset by manually classifying 100-200 abstracts from your feed as "highly relevant," "somewhat relevant," or "not relevant." Fine-tune a small classification model (e.g., a BERT-based model) on this dataset. Integrate this model into your pipeline to automatically tag new papers.
- Modules 74-76 (Conversational Interface - Comparative Study): Build two prototype chat interfaces for the RAG system. First, use a rapid development framework like Streamlit or Gradio for quick iteration.101 Second, build a more performant, desktop-native prototype using a modern stack like
Tauri with a Rust backend and a Svelte frontend.79 Document the trade-offs in development speed, performance, and complexity. - Modules 77-80 (Deployment & Documentation): Package the most promising prototype (or both) using the Docker skills from Module 14. Deploy the containerized application as a Hugging Face Space, making it publicly accessible.13 Write a comprehensive
README.md on GitHub for the project, explaining the architecture, setup instructions, and how to use the application.
- Deliverable: A publicly deployed, interactive AI application that solves a real personal problem and demonstrates expertise across the entire machine learning lifecycle, from data engineering to model fine-tuning and a comparative analysis of application deployment frameworks.
Phase 4: Connection & Synthesis (Modules 81-100)
Objective: To actively leverage the knowledge base and artifacts created in the previous phases to build a professional network, establish a reputation for expertise, and synthesize 40 years of experience into high-value, shareable assets. The strategy shifts from building and learning to connecting and influencing, using the created work as the foundation for all interactions.
Sub-theme: Strategic Networking & Friendship (Modules 81-90)
For a senior engineer, effective networking is not about volume but about the quality of connections. The goal is to build a network based on mutual respect and shared technical interests, allowing opportunities and new friendships to emerge organically.21
- Module 81: Activating Your Network: Begin with existing connections. Share the capstone project from Phase 3 on LinkedIn, tagging any relevant technologies or companies. Send personalized messages to a select group of 5-10 trusted former colleagues, briefly explaining the project and asking for their expert feedback.
- Module 82: Engaging in Communities: Transition from passive observation to active participation in the online communities identified in Day 18. The key is to lead with value. When someone asks a question that your capstone project or a tutorial can help answer, share your work as a resource. Participate in technical discussions, drawing upon the deep knowledge synthesized in your PKES.
- Module 83: Conference & Meetup Strategy: Identify one key virtual or in-person conference or a series of local meetups to attend. Before the event, study the speaker list and agenda. Identify 2-3 speakers or project maintainers with whom you want to connect. Prepare specific, insightful questions about their work that demonstrate you have engaged with it deeply. The goal is to have a memorable, substantive conversation, not just to exchange contact information.23
- Module 84: The Art of the "Coffee Chat": From the interactions in online communities or events, invite 2-3 people for a 30-minute virtual "coffee chat." The explicit goal of this meeting should be to learn about their work and interests. Be prepared with questions about their challenges, their perspective on industry trends, and their career journey. This approach, focused on genuine curiosity, is the most effective way to build lasting professional relationships and friendships.21
- Modules 85-90: Project: Personal CRM Engineering with mdBook: Systematize relationship management by building a tool directly into your publishing pipeline. The project is to design and build a custom mdBook preprocessor in Rust. This preprocessor will parse special syntax within your Markdown files (e.g., @[Contact Name](contact_id)) and automatically generate a "Contacts" chapter, cross-linking individuals to the projects and ideas you've discussed with them. This is a perfect "closer-to-the-metal" project that enhances your core tool and directly serves the goal of fostering connections.
Sub-theme: Opportunity Engineering (Modules 91-95)
- Modules 91-93: Gig & Project Tracking System: Build a tool to analyze the freelance and independent project market.
- Day 91 (API Exploration): Research and get API keys for platforms like Upwork and Freelancer.com.106 Understand their data structures for job postings, required skills, and pricing.
- Day 92-93 (Dashboard Build): Write a Python script to pull data from these APIs based on keywords relevant to your skills. Create a simple dashboard (using a tool of your choice from Module 74-76) to visualize trends in demand, popular technologies, and typical project rates.
- Modules 94-95: Talent & Collaborator Discovery: Extend the previous tool to identify potential collaborators. Write a script to scan GitHub or other platforms for developers contributing to open-source projects in your areas of interest. The goal is to build a system that helps you find interesting people to connect with for potential side hustles or independent projects.
Sub-theme: Mentorship & Knowledge Synthesis (Modules 96-100)
This final sub-theme focuses on the highest-leverage activities: codifying and sharing the unique wisdom gained over a 40-year career to build community.
- Module 96: Becoming a Mentor: Actively seek a mentorship opportunity. This could be through a formal platform like MentorCruise or CodePath, or informally within one of the open-source communities you have joined.75 Offering to guide a junior developer through their first open-source contribution is an excellent way to give back and solidify your own understanding.
- Module 97: The "Brag Document" Synthesis Project: Dedicate a focused effort to creating a comprehensive "Brag Document" as outlined by GitHub's career guides.10 This document is an internal-facing narrative of your entire career. Structure it by key projects or roles. For each, detail the business problem, the technical solution you engineered, the skills you applied, and—most importantly—the quantifiable business outcome.
- Modules 98-99: Podcasting & Community Building:
- Day 98 (Autodidactic Podcasting): Plan a small, focused podcast or webcast series. The theme could be a "Technical Journal Club" where you and a guest discuss a recent arXiv paper. Outline the first 3-5 episodes. Research and set up a minimal audio recording/editing workflow.108 The goal is to learn the process through a hands-on, "Toastmasters" style of disciplined practice.
- Day 99 (Pilot Episode & Online Discussion Group): Record a short pilot episode. Use this as a catalyst to start an online discussion group (e.g., on Discord or a dedicated forum) for people interested in discussing cutting-edge tech papers, creating a space for the friendships and connections you aim to foster.
- Module 100: The 100-Day Review & The Next 100 Days: Conduct a final, formal review of the entire 100-day journey. Use your PKES to write a detailed retrospective. Analyze the system you have built, the new skills you have acquired, the portfolio of artifacts you have created, and the new relationships you have formed. The ultimate measure of success for this curriculum is not its completion, but its continuation. Use the final day to leverage the full power of your new Personal Knowledge Engineering System to plan the next 100 days of learning, creating, and connecting.
Deliverables
Pipeline
At first, this page will just lay out the roadmap or thinking for completing the assingment.
In general, the assignment was to engineer an automated information capture pipeline to capture external information for potential inclusion in your book. Since mdBook lacks a direct clipper plugin ecosystem, the workflow will be more deliberate. Create a separate inbox directory outside the mdBook src folder. Configure tools like an RSS reader (e.g., Feedly) with IFTTT/Zapier or custom scripts to automatically save interesting articles, paper abstracts, or email newsletters as raw Markdown files into this inbox. This creates an "editorial funnel." The manual process of reviewing these drafts, refining them, and then consciously moving them into the src directory and adding them to SUMMARY.md becomes a key part of the engineering process, ensuring only curated content makes it into the final publication.
Four approaches are being considered. I am leaning toward Approach 4, but I would like to capture as much of the advantages as possible from the other three approaches as I adapt Approach 4 going forward.
Approach 1: Adapt an Existing Open-Source Self-Hosted RSS Reader (e.g., NewsBlur or Alternatives)
NewsBlur can be seen as a potential starting point or stalking horse for a starting point until something better is identified, this approach focuses on self-hosting it or a similar tool, then extending it with custom scripts for Markdown export and GitHub integration. NewsBlur is a Python/Django-based RSS reader that supports feed aggregation, story filtering (e.g., by tags, keywords, authors), and self-hosting via Docker. While it doesn't natively export to Markdown, its open-source nature allows modification. Alternatives like FreshRSS (PHP-based, lightweight, customizable with extensions) or Miniflux (Go-based, minimalistic, supports OPML imports and API for exports) could be easier to adapt if the development of NewsBlur feels too heavy.
Steps:
- Set Up the Reader: Clone and deploy NewsBlur using Docker (run
make nb
for containers including databases and web servers). For alternatives, install FreshRSS via Docker or a web server—it's simpler with built-in mobile app support. - Configure Feeds: Add RSS sources for articles, paper abstracts (e.g., arXiv feeds), and newsletters. Use filters to auto-tag or highlight relevant content.
- Extend for Export: Write a custom Python script (using libraries like feedparser for RSS parsing and markdownify for HTML-to-Markdown conversion) to query the reader's API/database, convert saved/favorited items to raw Markdown files. Schedule this with cron jobs to run periodically.
- Push to Inbox: Use the GitHub API (via PyGitHub library) in the script to commit Markdown files to your PKE repo's
src/1.Projects/inbox
subfolder (create it if needed). This keeps it outside the main src but within Projects for development. - Curation Workflow: Manually review files in the inbox, refine them (e.g., add metadata like tags or links to SUMMARY.md), and move to appropriate src sections. For automation, integrate an LLM script (e.g., using Hugging Face models) to summarize or classify content before pushing.
- AI Integration Path: Once stable, hook into your MCP vision by treating the inbox as a RAG (Retrieval-Augmented Generation) source for AI agents that curate and suggest additions to the mdBook.
Pros:
- Leverages proven RSS functionality (e.g., NewsBlur's social features for potential collaboration).
- Fully open-source and customizable, aligning with your PKE principles of extensibility.
- Alternatives like Miniflux have APIs that make scripting easier than NewsBlur's setup.
Cons:
- Self-hosting requires server resources (e.g., VPS for Docker); NewsBlur's setup involves multiple containers, which might be overkill initially.
- Initial extension work needed for Markdown export.
This builds on existing wheels like NewsBlur, as you suggested, and fits your preference for open-source tools similar to Feedly.
Approach 2: Use No-Code Integrations with IFTTT/Zapier for RSS-to-GitHub Automation
If you want a quicker start without heavy coding, use no-code platforms like IFTTT or Zapier to handle RSS ingestion and file creation in GitHub. These can act as your "editorial funnel" by triggering on new feed items and saving them as Markdown. For a free alternative, use Actionsflow (a GitHub Actions-based Zapier clone) to keep everything in your repo ecosystem.
Steps:
- Set Up Triggers: In Zapier/IFTTT, create a "Zap" or "Applet" with RSS as the trigger (e.g., new item in a feed from arXiv or newsletters). Filter by keywords to capture only pertinent content.
- Convert to Markdown: Use built-in formatters or a intermediate step (e.g., Zapier's code block with JavaScript) to extract title, summary, and content, then format as basic Markdown (e.g.,
# Title\n\nExcerpt...
). - Push to GitHub: Connect to GitHub integration to create a new file in your PKE repo (e.g.,
src/1.Projects/inbox/new-article.md
). IFTTT has direct RSS-to-GitHub applets for creating issues or commits; Zapier can append to files or create pull requests. - Inbox Management: Files land in the inbox for manual review. Use GitHub Actions in your repo to auto-label or notify you of new files.
- Enhance with Scripts: For better Markdown quality, add a custom GitHub Action (e.g., from repos like keiranlovett/rss-feed-to-markdown) that runs on push to refine files.
- Towards Automation: Upgrade to AI-assisted curation by integrating Zapier with an LLM API (e.g., OpenAI) to summarize/refine before saving. This aligns with your MCP goal, where the mdBook becomes context for AI-driven filtering.
Pros:
- Minimal setup time; no self-hosting needed.
- Handles automation like saving abstracts or newsletters out-of-the-box.
- Free tiers available (e.g., IFTTT for basic RSS triggers); Actionsflow is fully free and GitHub-native.
Cons:
- Limited customization (e.g., Zapier might not handle complex Markdown conversion perfectly).
- Dependency on third-party services, which contrasts with your open-source preference—mitigate with Actionsflow.
This is ideal for prototyping your funnel before building custom elements.
Approach 3: Build a Custom Script-Based Pipeline with Python and GitHub Actions
For full control within your mdBook ecosystem, create a bespoke pipeline using Python scripts and GitHub Actions. This leverages your PKE repo directly, treating the inbox as a staging area in src/1.Projects
. Tools like feedparser (for RSS) and GitHub Actions ensure it's automated and extensible.
Steps:
- Script Development: Write a Python script using feedparser to fetch RSS feeds, markdownify to convert HTML content to Markdown, and frontmatter to add metadata (e.g., source URL, date). Save as individual .md files locally.
- Scheduling: Run the script via cron on a local machine/server or as a GitHub Action workflow (e.g., scheduled daily). Use repos like myquay/feedmd as a base—it's a CLI for converting feeds to Markdown digests.
- GitHub Integration: In the script or Action, use Git commands or the GitHub API to push files to
src/1.Projects/inbox
. Configure the workflow to commit only if new content matches criteria (e.g., via regex filters). - Review Process: Use mdBook's preview server to view inbox files separately. Manually move refined files to src and update SUMMARY.md.
- Automation Evolution: Add AI layers (e.g., integrate with torch or sympy for content analysis) to auto-curate: classify relevance, generate summaries, or even propose SUMMARY.md updates. This directly supports your vision of the mdBook as a foundation model, where scripts feed into MCP for AI-assisted engineering.
- Expansion: Incorporate email newsletters via IMAP parsing in the script, or web scraping for non-RSS sources.
Pros:
- Highly tailored to PKE's structure (e.g., P.A.R.A. organization) and your AI goals.
- No external hosting; runs on GitHub for free.
- Easy to version-control the pipeline itself in the repo.
Cons:
- Requires scripting knowledge, though starting with existing repos minimizes this.
- Manual setup for feeds and filters initially.
This approach emphasizes deliberate workflow, as mdBook lacks plugins, and scales to your automated curation objective.
Approach 4: Hybrid mdBook-Centric System with Browser Clippers and AI Preprocessing
To stay as close as possible to mdBook without external readers, use browser-based clippers combined with scripts for ingestion. This treats your toolchain as an "editorial funnel" extension of mdBook, potentially forking mdBook for custom preprocessors later.
Steps:
- Clipping Tools: Use open-source clippers like MarkDownload (browser extension that saves web pages as Markdown) or adapt Obsidian's web clipper. Configure to save clips to a local folder synced with GitHub (e.g., via Git).
- RSS Integration: Pair with a simple RSS poller script (Python with feedparser) that fetches items, uses requests to get full content, converts to Markdown, and saves to the synced inbox.
- GitHub Sync: Use GitHub Desktop or Actions to pull/push the inbox folder in
src/1.Projects
. - Preprocessing: Develop a Rust-based mdBook preprocessor (as hinted in your curriculum's Phase 4) to scan the inbox, apply AI filters (e.g., via local models), and suggest integrations into SUMMARY.md.
- Full Automation: Evolve to use IFTTT for clipping triggers or Zapier for RSS, but route everything through scripts that enforce curation rules.
- MCP Tie-In: Design the pipeline to output structured data (e.g., YAML frontmatter in MD files) that serves as context for AI models in your MCP infrastructure.
Pros:
- Keeps everything within mdBook's ecosystem, per your preference.
- Flexible for non-RSS sources like emails or abstracts.
- Directly advances your AI-assisted knowledge engineering goal.
Cons:
- More fragmented initially (clipper + scripts vs. unified reader).
- Requires building/forking mdBook extensions for seamless integration.
These approaches start simple (no-code) and scale to complex (custom AI), aligning with your 100-day PKE curriculum's phases—e.g., foundation in Phase 1, deep learning in Phase 2, and synthesis in Phase 4. Begin with Approach 2 for quick wins, then transition to 3 or 1 for longevity.
Research Dashboard
At first, this page will just lay out the roadmap or thinking for completing the assingment.
In general, the assignment was to create the Research Dashboard chapter in your mdBook. Since there's no dynamic plugin like Dataview, write a simple Python or shell script that scans your inbox directory for new files or files with a #summarize tag in their frontmatter, and generates a summary list. This script can be run manually to update the dashboard page.
Grok was asked to give suggestions on how to complete this task of building a research dashboard.
Existing Developments
While there isn't a direct equivalent to Obsidian's Dataview plugin specifically for mdBook (which would allow querying Markdown files like a database and generating dynamic views such as tables or lists), some related tools and plugins are in development or available that could serve as starting points or inspirations for your Personal Knowledge Engineering (PKE) system. Based on recent searches:
-
mdbook-template: This is a prototypical method for building preprocessor plugin that enables dynamic text generation by allowing you to include Markdown files with customizable arguments (e.g., passing variables to templates for conditional or parameterized content). A simple mdbook-preprocessor or mdbook-plugins for rendering content in interactive tabs, which adds a layer of dynamic presentation to static Markdown. This isn't query-based but demonstrates how plugins can manipulate content structure during build. This does not immediately yield a full query engine like Dataview, but it supports basic dynamic inclusion and could be extended for metadata-based generation. mdbook-template was actively maintained as a crate on crates.io and available on GitHub as themdbook-template archive repo. One feasible approach would be to fork archived GH repo for your PKE repo to add query-like features, such as scanning frontmatter or tags.
-
Community discussions on extending mdBook (e.g., via preprocessors for custom features) are ongoing, but no full Dataview clone is under active open development as of mid-2025. Anyone interested in collaborating or forking extending mdBook should check Rust forums or GitHub issues for mdBook extensions.
For a comprehensive list of mdBook plugins, refer to the official third-party plugins wiki, though it doesn't highlight any exact Dataview matches. If none fit, building your own is feasible given mdBook's extensible architecture.
Approaches to Building a Custom mdBook Dynamic Plugin
Here are several practical approaches to create Dataview-like functionality in mdBook for your PKE system. These build on mdBook's preprocessor system (which processes content before rendering) and can handle dynamic generation based on metadata, tags, or queries in your Markdown files. Your PKE repo appears to be a GitHub Pages-hosted mdBook site focused on knowledge management concepts, so these could integrate via custom chapters or automated builds.
1. Custom Preprocessor with Query Syntax (Server-Side Build-Time Generation)
This is the most direct way to mimic Dataview: Create a preprocessor that scans your book's Markdown files, parses queries, and generates content during the mdbook build
process.
- Steps:
- Define a custom syntax in your Markdown, e.g., fenced code blocks like:
```pke-query TABLE title, tags, summary FROM folder:notes WHERE tags CONTAINS #project
- Write the preprocessor in Rust (or any language, e.g., Python via a script) that:
- Receives the book's JSON structure via stdin.
- Scans all chapters for frontmatter (YAML metadata like tags, dates) or inline elements.
- Parses the query (use libraries like
serde
for JSON/YAML, orpest
for query parsing in Rust). - Queries the content (e.g., filter files by tags, folders, or properties).
- Generates Markdown/HTML output (e.g., a table) and replaces the query block.
- Configure in
book.toml
:[preprocessor.pke-dataview] command = "./target/release/mdbook-pke-dataview" # Or path to your script
- Define a custom syntax in your Markdown, e.g., fenced code blocks like:
- Pros: Fully integrated, no runtime overhead; works offline.
- Cons: Build-time only (not live updates); requires recompiling for changes.
- Tools/Libs: In Rust, use
mdbook::preprocess
crate; for Python, parse JSON input and usepandas
for querying data. - Extension for PKE: Start by extracting metadata from your existing notes in the repo, then generate index pages dynamically.
2. JavaScript-Based Client-Side Dynamics (Post-Render Manipulation)
For interactive queries without rebuilding the book each time, embed JavaScript to query and manipulate the DOM after the HTML is generated.
- Steps:
- In your mdBook theme (customize
theme/index.hbs
or add JS viaadditional-js
inbook.toml
), include a script that loads all page data (e.g., via a pre-generated JSON index of metadata). - Pre-build a metadata index: Use a script to scan Markdown files and output a
data.json
with entries like{ "path": "notes/project.md", "tags": ["#project"], "summary": "..." }
. - In Markdown, add placeholders like
<div class="pke-query" data-query="FROM #project"></div>
. - JS code (e.g., with vanilla JS or a lib like DataTables) fetches the JSON, filters based on the query, and injects tables/lists.
- Example JS snippet:
document.querySelectorAll('.pke-query').forEach(el => { const query = el.dataset.query; fetch('/data.json').then(res => res.json()).then(data => { // Filter data based on query logic const results = data.filter(item => item.tags.includes('#project')); // Generate and insert table HTML el.innerHTML = generateTable(results); }); });
- In your mdBook theme (customize
- Pros: Interactive (e.g., sortable tables); no full rebuild needed for minor changes.
- Cons: Requires JS enabled; heavier for large books; data must be static or pre-indexed.
- Tools/Libs: Use
lunr.js
for search indexing oralasql
for SQL-like queries on JSON. - Extension for PKE: This could add real-time filtering to your GitHub Pages site, enhancing knowledge navigation.
3. Hybrid Pre-Build Scripting with External Tools
Run scripts before mdbook build
to generate dynamic content, treating your Markdown as a database.
- Steps:
- Use tools like
jq
(for JSON) orawk
to process files, or a full script in Python/Node.js. - Example: A bash/Python script that:
- Recursively scans
.md
files for frontmatter/tags. - Builds a database (e.g., SQLite or JSON).
- Executes queries and outputs generated Markdown files (e.g., auto-create an "index.md" with tables).
- Recursively scans
- Integrate via a Makefile or GitHub Actions workflow:
make generate && mdbook build
. - For queries, mimic Dataview with a custom DSL parsed by your script.
- Use tools like
- Pros: Flexible; leverage existing tools (e.g., combine with
pandoc
for advanced processing). - Cons: Adds build steps; not as seamless as a native plugin.
- Tools/Libs: Python with
frontmatter
lib for metadata;sqlite3
for querying. - Extension for PKE: Automate this in your repo's CI to regenerate views on push, keeping your knowledge base up-to-date.
4. Integration with External Frameworks or Generators
Embed mdBook within a larger system for advanced dynamics, especially if your PKE evolves beyond static sites.
- Steps:
- Use mdBook as a content source, but render via a dynamic framework like Next.js (with MDX for Markdown).
- Example: Fork something like "MDNext" (a Next.js starter for MDX) to add query layers.
- Parse mdBook output into a Next.js site, adding server-side querying.
- Or, sync your Markdown to a tool like Obsidian (for Dataview) and export back, but this is roundabout.
- For GitHub Pages, use Jekyll plugins if migrating, but stick to mdBook for Rust ecosystem benefits.
- Use mdBook as a content source, but render via a dynamic framework like Next.js (with MDX for Markdown).
- Pros: Scales to full apps; adds features like search APIs.
- Cons: Increases complexity; may require rewriting parts of your PKE setup.
- Tools/Libs: Next.js with
next-mdx-remote
; or Rust alternatives like Leptos for web apps. - Extension for PKE: If your system grows, this could turn your static book into a web app with user queries.
Start with the preprocessor approach for closest integration, as it's mdBook-native and aligns with your provided example. Test on a branch of your repo, and consider open-sourcing the plugin to attract contributors. If I need code snippets or help with implementation, all that I need to doe is provide more details to Grok, when I understand the specifics of what I need!
Methodology
This document, other than following the mdBook documentation, will detail the repository specific rules for creating new pages in this mdBook, the strategy for structuring chapters, and the lifecycle of information as it moves from a rough draft to a published chapter.
Specifically, the purpose of this page is to describe the design of the mdBook which catalogs the process of developing of the AI-assisted PKE system per our Manifesto.
We will use the P.A.R.A. method (Projects, Areas, Resources, Archive) as a conceptual guide to organize the top-level chapters and sections within this mdBook's src directory as the foundational information architecture for your mdBook project. In contrast to a freeform approach OR generally adaptible mdBook approach that fits appropriately to the software being documented and implemented simultaneously, this mdBook is somewhat self-referential in terms of developing a PKE, thus following the PARA structured, hierarchical approach from the outset makes sense for developing a PARA-influence PKE.
In general, an issue-driven approach will be followed as we progress working through the daily modules in this mdBook's PKE development process, using the Zettelkasten concept of atomic notes. Each new issue that arises will be given it's own self-contained piece of research or issue#.md page. At first the issue#.md page will be in the 1.Projects folder until they are dispatched or dispositioned appropriately within the book's structure, all will be linked hierarchically by the SUMMARY.md file.
The 1.Projects folder will be the landing place for new issues and thereafter for short-term, less than one week efforts which are currently underway and should be regarded as under HEAVY construction. Issues that take on a larger life as much larger, ongoing effort will go to the 2.Areas folder. Issues that are developed and completed will go to he 3.Resources folder. Issues that are dismissed, after even a minor expenditure of dev effort, will go to the 4.Archive folder.
The 2.Areas folder will be for longer-term development and ongoing efforts that will stay open, perhaps indefinitely as perhaps usable, but under ongoing development. Areas that are developed for some time and eventually completed will go to he 3.Resources folder.
The 3.Resources folder will be for usable references and material that's that have been either curated or developed and although curation might continue to add things, these items should be regarded as stable enough to be considered usable, as good as complete. In some cases, a Project or Area might graduate to being in its own development repository, but page linking to that effort will be maintained in the Resources folder.
The 4.Archive folder will be for things that in the back Area 51 parking lot and might still be valuable for informational purposes, but are basically not something anyone should use.
Project Overview
This landing page will feature a list of ongoing PROJECTS. We will develop a template after we have experience with several examples.
A Project is the start of a bigger development commitment and the basis of the P.A.R.A. method of the Building a Second Brain (BASB) methodology. The BASB method systematically manages information differently than just notetaking apps ... PROJECTS, have goals, reqmts and deadlines ... AREAS are about roles/responsibilities or obligations or capabilities that need to be earnestly developed ... RESOURCES, mostly finished AREAS, but also ongoing interests, assets, future inspiration, may req continual maintenance and refactoring but, for now, are backburnerable ... ARCHIVES, inactive matl from P A R that shouldn't be used, except for informational purposes.
GitHub Discussion, Issue, Project Functionality
We will rely upon the GitHub Discussion and Issue functionality, BEFORE graduating something to "Project" status ... when something becomes a Project on GitHub, it will simultaneously become a PROJECT in our P.A.R.A. hierarchy.
Please understand the GitHub progression from ... Discussions ...to... Issue ...to... Project.
Discussions are mainly for just discussing something, to clarify terminology or ask questions or for just generally speculative thinking out loud.
Issues are for things that somebody really needs to look into and possibly turn into more of a Project.
On GitHub a Project is an adaptable spreadsheet, task-board, and road map that integrates with your issues and pull requests on GitHub to help you plan and track your work effectively. You can create and customize multiple views by filtering, sorting, grouping your issues and pull requests, visualize work with configurable charts, and add custom fields to track metadata specific to your team. Rather than enforcing a specific methodology, a project provides flexible features you can customize to your team’s needs and processes.
Areas Overview
This landing page will feature a list of ongoing AREAS. We will develop a template after we have experience with several examples.
An AREA begins first as a PROJECT and then graduates to AREA status after it is sufficiently mature, but still not fully developed.
A Project is the start of a bigger development commitment and the basis of the P.A.R.A. method of the Building a Second Brain (BASB) methodology. The BASB method systematically manages information differently than just notetaking apps ... PROJECTS, have goals, reqmts and deadlines ... AREAS are about roles/responsibilities or obligations or capabilities that need to be earnestly developed ... RESOURCES, mostly finished AREAS, but also ongoing interests, assets, future inspiration, may req continual maintenance and refactoring but, for now, are backburnerable ... ARCHIVES, inactive matl from P A R that shouldn't be used, except for informational purposes.
GitHub Discussion, Issue, Project Functionality
We will rely upon the GitHub Discussion and Issue functionality, BEFORE graduating something to "Project" status ... when something becomes a Project on GitHub, it will simultaneously become a PROJECT in our P.A.R.A. hierarchy.
Please understand the GitHub progression from ... Discussions ...to... Issue ...to... Project.
Discussions are mainly for just discussing something, to clarify terminology or ask questions or for just generally speculative thinking out loud.
Issues are for things that somebody really needs to look into and possibly turn into more of a Project.
On GitHub a Project is an adaptable spreadsheet, task-board, and road map that integrates with your issues and pull requests on GitHub to help you plan and track your work effectively. You can create and customize multiple views by filtering, sorting, grouping your issues and pull requests, visualize work with configurable charts, and add custom fields to track metadata specific to your team. Rather than enforcing a specific methodology, a project provides flexible features you can customize to your team’s needs and processes.
New Frontiers for Scientific Foundation Models
We describe the motivation for our curated portfolio of 100 Foundational Model Concepts that we are developing.
The prevailing narrative of artificial intelligence has been dominated by foundation models that emulate human cognition, primarily through language and image generation. While transformative, this represents only the first wave of a far more profound technological revolution. This report posits that the next, and arguably more significant, frontier for foundation models lies in their application to fundamental scientific discovery. The emergence of these Scientific Foundation Models (SciFMs) is the catalyst that fully realizes the "Fourth Paradigm" of science—a new era where data-intensive, AI-driven exploration becomes a primary mode of discovery, standing alongside the traditional pillars of theory, experimentation, and simulation.
This analysis identifies four key domains ripe for this transformation: the Physical Sciences, where models can navigate the vast combinatorial space of materials and molecules and even discover new physical laws; the Biological Sciences, where they can unravel the multi-scale complexity of life from genomics to whole-organism behavior; Complex System Simulation, where they can act as high-fidelity surrogates to model intractable systems like climate and turbulence; and Emergent Social Dynamics, where they can simulate in-silico societies to understand the complex interplay between individual actions and collective outcomes.
To chart a course toward this future, this report presents a curated portfolio of 100 high-potential foundation model concepts, each designed to tackle a specific, high-impact scientific challenge. However, realizing this vision is contingent on overcoming the primary bottleneck: the availability of large-scale, structured scientific data. The core strategic recommendation of this report is therefore a concerted, multi-stakeholder effort to build open, large-scale "Data Commons" for key scientific fields. This initiative, coupled with a strategy that creates a virtuous cycle between simulation and experimentation and fosters deeply integrated, cross-disciplinary research teams, forms the critical path to unlocking the unprecedented potential of these models. The successful development of SciFMs will not merely accelerate research; it will fundamentally redefine the scientific method, promising to advance human knowledge at a rate unprecedented in history.
Part I: The New Scientific Revolution - From Data to Discovery
This introductory section establishes the report's conceptual framework, differentiating scientific foundation models (SciFMs) from their well-known generative AI counterparts and positioning them as the primary tool for the new, data-intensive paradigm of science.
1.1 The Dawn of the Fourth Paradigm
The history of science can be understood as a succession of paradigms, each defined by its dominant methodology. The first paradigm was empirical, characterized by the observation and description of natural phenomena. The second was theoretical, marked by the use of models and generalizations, epitomized by Newton's laws and Maxwell's equations. The third paradigm, which emerged in the latter half of the 20th century, was computational, leveraging computer simulations to explore complex phenomena that were analytically intractable.1 Today, we are witnessing the consolidation of a fourth and profoundly different paradigm: data-intensive scientific discovery.1
This Fourth Paradigm is defined by its reliance on advanced computing capabilities to analyze, manage, and explore massive datasets generated from simulations, experiments, and sensors.1 The speed at which any scientific discipline now advances is directly dependent on its ability to harness these vast data streams.1 This shift is not merely about processing more data; it represents a fundamental change in the scientific process itself, one that requires new tools, novel research methodologies, and new modes of collaboration between domain scientists and technologists.1
While this paradigm has been emerging for over a decade, it is the recent maturation of foundation models that provides the technological catalyst to fully realize its potential. Foundation models are, by definition, large-scale models trained on vast datasets that can be adapted to a wide range of downstream tasks.5 Their ability to learn general representations from broad data makes them uniquely suited to the challenges and opportunities of data-intensive science. Therefore, it is most accurate to view foundation models not as just another tool
within the Fourth Paradigm, but as the enabling engine that makes its most ambitious goals—from autonomous discovery of physical laws to the high-fidelity simulation of entire biological organisms—achievable. This reframes the conversation from "using AI in science" to "AI as the new platform for science."
1.2 Defining the Scientific Foundation Model (SciFM): Beyond Plausibility to Physicality
A foundation model is formally defined as a large, deep learning model pre-trained on a broad spectrum of data, often through self-supervision, which can then be fine-tuned or adapted to perform a wide variety of specific tasks.5 Prominent examples like GPT-3 and BERT have demonstrated remarkable capabilities in processing and generating human language.6 However, to apply this paradigm to scientific discovery, a critical distinction must be made between these existing models and the next generation of Scientific Foundation Models (SciFMs).
Generative AI models like large language models (LLMs) are optimized for linguistic plausibility and coherence. Their objective is to generate outputs that are statistically likely given the patterns in their training data, which consists primarily of human-generated text and images. In contrast, SciFMs must be optimized for physical validity and empirical verifiability. Their success is not measured by their ability to conduct a human-like conversation but by their capacity to generate novel, valid hypotheses, accurately predict experimental outcomes, and accelerate research and development cycles.9 While an LLM might "hallucinate" plausible but incorrect information with relatively low consequence, a materials science model that hallucinates a physically impossible crystal structure or a medical model that proposes a toxic drug molecule is fundamentally failing its core purpose. SciFMs must be rigorously grounded in the laws of nature.12
This distinction gives rise to a central design tension in the development of SciFMs. On one hand, the naive application of machine learning to scientific data can lead to erroneous, "Aristotelian" conclusions, such as discovering that heavier objects fall faster because the model has not been constrained by the concept of a vacuum.12 This suggests the need for "physics-informed" AI, where known physical laws and constraints are embedded into the model's architecture and training process to ensure its outputs are physically sound. On the other hand, pre-training a model too heavily on the existing body of scientific knowledge may introduce powerful "inductive biases".14 These biases, while ensuring consistency with current theories, could fundamentally limit the model's ability to discover truly novel phenomena or physical laws that lie outside of, or even contradict, our present understanding. Navigating this trade-off—between ensuring physical realism and enabling genuine discovery—is a core research and development challenge that will define the field of scientific AI.
1.3 A Comparative Framework
To crystallize the unique characteristics of SciFMs, the following framework provides a direct comparison with the generative AI models that currently dominate the public and commercial landscape. Understanding these differences is essential for formulating appropriate R&D strategies, investment theses, and evaluation metrics for this new class of AI.
Table 1: A Comparative Framework for Foundation Models
Dimension | Generative AI (Language/Image) | Scientific Foundation Models (SciFMs) |
---|---|---|
Primary Objective | Plausibility & Coherence | Verifiability & Discovery |
Core Data Modalities | Unstructured Web Text/Images | Structured Experimental/Simulation Data (e.g., genomic sequences, sensor time-series, molecular graphs, simulation outputs) |
Validation & Grounding | Human preference & internal consistency | Empirical experimentation & adherence to physical laws |
Key Technical Challenges | Hallucination & bias mitigation | Physical constraint satisfaction, causality inference, uncertainty quantification |
Measure of Success | User engagement & task completion | Novel discoveries, predictive accuracy, accelerated R&D cycles |
Part II: A Sector-by-Sector Analysis of High-Potential Domains
This section provides the analytical core of the report, examining in detail the most promising scientific and engineering fields for the development and application of Scientific Foundation Models. Each chapter outlines the core challenges of the domain, the nature of its data landscape, and the specific opportunities for transformative impact through SciFMs.
Chapter 1: The Physical Universe - FMs for Physics, Materials, and Chemistry
1.1 The Challenge: Navigating Vast Chemical and Physical Possibilities
The fundamental sciences of physics, chemistry, and materials science are confronted by challenges of immense scale and complexity. The primary obstacle in materials and drug discovery is the combinatorially explosive nature of the design space. For example, scientists estimate that the number of potentially stable small molecules relevant for battery electrolytes could be on the order of 1060.9 Exploring this "molecular universe" through traditional methods, which rely on a combination of human intuition, trial-and-error experimentation, and computationally expensive simulations, is an inefficient and painstaking process that can take years or even decades to yield a single breakthrough.
Simultaneously, in fundamental physics, our understanding of the universe remains incomplete. The process of deriving governing equations and physical principles from complex, high-dimensional, and often noisy experimental data is a monumental intellectual challenge.12 Human cognition, with its inherent biases and limitations in processing vast datasets, may overlook subtle patterns or correlations that hold the key to new discoveries, while naive machine learning approaches risk latching onto spurious correlations without grasping the underlying causal structure of reality.12
1.2 The Data Landscape: From First Principles to High-Throughput Experiments
The physical sciences benefit from a growing landscape of high-quality, structured data, forming a fertile ground for training SciFMs. A significant portion of this data comes from large-scale computational databases generated using first-principles methods like Density Functional Theory (DFT).16 Publicly accessible repositories such as the Materials Project (MP), the Open Quantum Materials Database (OQMD), and JARVIS-DFT contain hundreds of thousands to millions of calculated structures and their associated properties, providing a clean and consistent training corpus.5
This computational data is complemented by vast archives of experimental results and scientific literature. Chemical databases like PubChem and ChEMBL contain information on tens of millions of molecules.5 Furthermore, modern laboratories are increasingly equipped with high-throughput screening and automated instrumentation that generate streams of experimental and sensor data, providing the real-world grounding necessary to validate and refine computationally derived models.13
1.3 The Foundation Model Opportunity
The application of foundation models to this data landscape opens up three distinct and increasingly ambitious opportunities.
First is predictive modeling. By training on these large databases, SciFMs can learn the intricate and non-linear relationships between the structure of a material or molecule and its emergent properties. Models with architectures like Graph Neural Networks (GNNs) and Transformers, such as GraphCL, MoLFormer, and CrystalLLM, are already being developed to predict a wide range of characteristics, including electronic conductivity, thermodynamic stability, catalytic activity, and toxicity.5 These models can serve as powerful screening tools, allowing researchers to evaluate thousands of candidate compounds in silico, dramatically reducing the time and expense required for physical synthesis and testing.
Second, and more transformative, is inverse design. This approach fundamentally reverses the traditional scientific workflow. Instead of starting with a structure and predicting its properties, an inverse design model starts with a set of desired properties and generates novel, physically viable structures that exhibit them. Generative models, such as the diffusion-based MatterGen for crystal structures and the Transformer-based GP-MoLFormer for molecules, are trained to navigate the vast chemical space and propose new candidates tailored to specific applications, such as designing a material with a target band gap for a semiconductor or a molecule with high binding affinity for a specific protein.5 This shifts the role of the scientist from a manual explorer to the architect of a discovery process.
The third and most revolutionary opportunity is the autonomous discovery of physical laws. This nascent field aims to create AI systems that can move beyond applying known physics to discovering new physics. Systems like AI-Newton have demonstrated a remarkable proof of concept: by ingesting only raw observational data from simulated experiments (e.g., the positions and times of moving bodies), the system can autonomously formulate fundamental physical concepts like mass, momentum, and kinetic energy, and then use these concepts to rediscover the symbolic, mathematical form of canonical physical laws, including Newton's second law (F=ma) and the law of universal gravitation.17 This represents a paradigm shift. For a scientific discovery to be truly integrated into human knowledge, the explanation is as crucial as the prediction. The ability of these systems to produce interpretable, symbolic outputs—actual equations—rather than just opaque neural network predictions, suggests that the most advanced SciFMs will incorporate a "symbolic grounding" layer. This makes the AI's reasoning legible, its discoveries verifiable, and its output directly integrable into the enduring edifice of scientific theory.
Chapter 2: The Biological Code - FMs for Genomics, Systems Biology, and Neuroethology
2.1 The Challenge: Unraveling the Complexity of Living Systems
Biological systems are arguably the most complex phenomena known to science, characterized by intricate, multi-scale interactions that span from the atomic level of DNA to the emergent behavior of entire organisms. A central challenge in systems biology is understanding metabolism, which is not a series of linear, isolated pathways but a densely interconnected network—a "hairball" of interactions between thousands of genes, proteins, and metabolites that defies simple, rational engineering.21 Predicting the system-wide effects of a single genetic modification remains a formidable task.
In neuroscience, a grand challenge is to understand how neural circuits generate the complex, adaptive behaviors that allow animals to thrive in natural, dynamic environments. This is the core pursuit of neuroethology.23 Traditional laboratory-based neuroscience often relies on highly constrained and simplified tasks, which limits insight into how the brain actually functions in the real world, where it must integrate sensory information, guide motor actions, and learn from experience in an unceasing, complex loop.23
2.2 The Data Landscape: A Multi-Modal Deluge
The life sciences are in the midst of a data explosion, driven by rapid technological advances across multiple fronts. High-throughput sequencing technologies generate petabytes of genomic, transcriptomic, and proteomic data, providing an unprecedented view into the molecular foundations of life.13 In the medical domain, the digitization of healthcare has created vast repositories of electronic health records and clinical trial data, which, when combined with information from consumer wearable devices, offer rich, longitudinal datasets on human health and disease at a population scale.3
In parallel, the field of neuroethology is being transformed by new observational technologies. High-resolution motion capture systems, dense multi-electrode arrays for recording neural activity from freely moving animals, and advanced sensors for environmental monitoring are creating comprehensive, multi-modal datasets that link neural dynamics, motor output, and sensory input with unprecedented fidelity.23 This data provides the raw material for building computational models of the brain in action.
2.3 The Foundation Model Opportunity
Foundation models are uniquely positioned to integrate and learn from this multi-modal biological data, opening new avenues for understanding, engineering, and healing living systems.
In metabolic engineering and synthetic biology, SciFMs can model the entire causal chain from genotype to phenotype. By training on vast datasets of genomic sequences and their corresponding metabolic outputs, these models can learn to predict the complex, system-wide consequences of genetic interventions. This enables a new paradigm of biological design, moving beyond single-gene edits to the rational, multi-variate optimization of entire metabolic networks for the production of valuable chemicals, biofuels, and pharmaceuticals.21 The goal is to transform microbial cells into programmable "factories" that can be engineered with the same predictability as traditional chemical plants.
In precision medicine, SciFMs can serve as powerful engines for integrating diverse patient data streams. A model trained on a combination of a patient's genomics, electronic health records, imaging data, and lifestyle information from wearable devices can generate highly personalized predictions for disease risk, treatment response, and optimal therapeutic strategies.3 Models like NatureLM are being developed with the ambitious goal of creating a unified representation that spans molecules, proteins, and DNA, enabling cross-domain applications such as designing a drug based on a patient's specific genetic makeup.13
A frontier application lies in computational neuroethology. Here, a foundation model can be trained on a comprehensive dataset capturing an animal's complete experience: its sensory inputs, the simultaneous activity of thousands of its neurons, and its resulting motor outputs. Such a model would learn the fundamental "language" of that organism's nervous system. It could be used to predict future behavior from patterns of neural activity, simulate the behavioral effects of specific neural perturbations, and ultimately, uncover the general computational principles that brains—even highly alien, non-mammalian brains like those of cephalopods—use to solve universal problems like navigation, foraging, and social communication.23
The convergence of these distinct modeling efforts points toward a more holistic and ambitious long-term goal: the creation of a "digital twin" of an entire organism. By integrating foundation models for genomics, metabolic networks, and neural control, it becomes possible to construct a multi-scale, in-silico simulation of a simple model organism, such as the nematode C. elegans. Such a digital twin would provide an unprecedented experimental platform, allowing scientists to conduct virtual experiments to test hypotheses about everything from the effects of a new drug to the neural basis of a specific behavior. This represents the ultimate synthesis of data-driven biology: a comprehensive, predictive, and executable model of life itself.
Chapter 3: Simulating Complex Systems - FMs for Climate, Turbulence, and Fluid Dynamics
3.1 The Challenge: The Computational Cost of Complexity
Many of the most critical systems in science and engineering—from the flow of air over an aircraft wing to the global climate system—are governed by complex, non-linear partial differential equations (PDEs).10 The Navier-Stokes equations, which describe the motion of fluids, are a prime example. While these equations are known, solving them directly for realistic, three-dimensional, and turbulent scenarios is a task of staggering computational complexity.31
This computational barrier forces practitioners to rely on approximations. In fluid dynamics, methods like Reynolds-Averaged Navier-Stokes (RANS) are used, but they introduce significant modeling errors by simplifying the physics of turbulence.33 In climate science, global models must parameterize crucial small-scale phenomena like cloud formation, leading to uncertainties in long-term projections. Performing a Direct Numerical Simulation (DNS) that resolves all scales of motion is computationally prohibitive for almost all practical engineering and scientific problems, creating a major bottleneck that slows innovation in aerospace design, energy production, and weather forecasting.32
3.2 The Data Landscape: A Firehose of Simulation and Sensor Data
Despite the cost, the scientific community has generated and continues to generate massive datasets that characterize these complex systems. High-fidelity simulations like DNS and Large Eddy Simulation (LES), while too expensive for routine use, can be run for canonical problems to create benchmark datasets of unparalleled accuracy and physical consistency.31 These simulations serve as a form of "computational experiment," providing perfect, complete data that is ideal for training machine learning models.
This simulation data is complemented by vast archives of real-world observational data. Decades of satellite imagery, global networks of weather and climate sensors, and measurements from experimental facilities like wind tunnels provide a continuous stream of information about the Earth's systems and engineering prototypes.33 The development of the Surya heliophysics model, for instance, was made possible by a training dataset comprising nine years of continuous, high-resolution solar observations from NASA's Solar Dynamics Observatory.35 This combination of pristine simulation data and noisy, complex observational data creates a rich and diverse foundation for training SciFMs.
3.3 The Foundation Model Opportunity
Foundation models offer a path to circumventing the computational bottleneck of direct simulation by learning the underlying physics from data.
The primary opportunity is the creation of physics-informed surrogate models. An SciFM can be trained on the input-output pairs from a large number of high-fidelity simulations. Once trained, the model learns a highly accurate mapping from the system's parameters (e.g., an aircraft's shape and speed) to its performance (e.g., the resulting lift and drag). This surrogate model can then make predictions in milliseconds, replacing the need to run a new, hours- or days-long simulation for every design change, thereby enabling real-time analysis and rapid design optimization.13
A specific and highly impactful application is in turbulence modeling. This has long been considered a "holy grail" of fluid dynamics. An SciFM can be trained on high-fidelity DNS and LES data to learn the complex physics of turbulent eddies. This learned knowledge can then be used to directly correct the known errors of cheaper RANS models or, more ambitiously, to derive entirely new and more accurate turbulence closure models from the data itself.31 This would represent a fundamental breakthrough in our ability to simulate and design systems involving turbulent flows.
In climate and weather forecasting, foundation models are already demonstrating their potential. Models like the Prithvi weather model and the Surya heliophysics model are designed to ingest and process vast amounts of spatio-temporal data from satellites and ground-based sensors.35 By learning complex patterns and long-range dependencies that are difficult for traditional numerical models to capture, these SciFMs can improve the accuracy of forecasts for everything from short-term weather patterns to long-term climate change impacts and space weather events like solar flares, which pose a risk to satellites and power grids.35
A profound implication of this technological shift is the democratization of high-fidelity simulation. Currently, the ability to perform large-scale, high-fidelity simulations is a strategic advantage held by large corporations, government labs, and well-funded research institutions with access to supercomputing resources.9 A foundation model, once trained on such a resource, can be deployed and run (a process called inference) at a small fraction of the computational cost. This means that a small engineering startup could leverage a pre-trained turbulence SciFM to achieve aerodynamic design insights that were previously accessible only to a major aerospace manufacturer. This leveling of the technological playing field could dramatically accelerate the pace of innovation across the entire economy, from renewable energy to advanced manufacturing.
Chapter 4: Emergent Social Dynamics - FMs for Economics and Social Science
4.1 The Challenge: Bridging the Micro-Macro Gap
The social sciences grapple with a fundamental challenge known as the "micro-macro gap".36 This refers to the difficulty of understanding how complex, large-scale social phenomena—such as the formation of social norms, the crash of a stock market, or the mobilization of a protest movement—emerge from the decentralized interactions of millions of individual agents.36 Human societies are complex adaptive systems, and their behavior is often non-linear, unpredictable, and counter-intuitive. Traditional modeling approaches, whether based on aggregate statistical analysis or simplified theoretical models, often fail to capture the rich, dynamic feedback loops between individual behavior and collective outcomes.
4.2 The Data Landscape: Digital Traces of Human Interaction
The data landscape for this domain is unique in that it can leverage the same kind of massive, unstructured textual data from the internet that powers conventional LLMs. This provides a rich source of information on human communication, beliefs, and culture. In addition to text, this field utilizes large, structured datasets from economic transactions, financial markets, demographic surveys, and controlled laboratory experiments designed to probe human decision-making and social behavior.38 This blend of structured and unstructured data provides a comprehensive, though often noisy, record of human social and economic life.
4.3 The Foundation Model Opportunity
The most exciting opportunity in this domain involves a radical new application of foundation models, particularly LLMs. Instead of being used as single, monolithic entities, they can serve as the cognitive "engines" for a large population of autonomous agents, creating sophisticated, large-scale agent-based models (ABMs).38 By endowing each agent in a simulation with a copy of an LLM, researchers can give them distinct goals, memories, behavioral heuristics, and the ability to communicate and reason using natural language. This allows for the creation of in-silico societies that are far more realistic and behaviorally rich than traditional ABMs.
Recent experiments have already demonstrated the power of this approach for simulating social convention formation. When placed in a simulated environment where they are rewarded for coordinating with each other, populations of LLM agents have been shown to spontaneously develop and universally adopt shared linguistic conventions and social norms, purely through decentralized interactions.38 These simulations provide a powerful, controllable, and repeatable experimental testbed for theories of social dynamics that have historically been difficult to verify empirically.
This methodology can be extended to explore complex economic and political dynamics. Agent-based simulations populated by LLM agents can be used to model the behavior of entire economies, supply chains, or political systems.36 By observing the emergent, macro-level behavior of the agent population—such as the formation of asset bubbles, the propagation of supply chain shocks, or the spread of misinformation—researchers can test the potential impacts of policy interventions and explore the underlying drivers of systemic phenomena in a controlled virtual environment.
Perhaps the most profound potential of this approach is to use AI as a "computational microscope" for society. One of the striking findings from recent research is that populations of individually unbiased LLM agents can, through their interaction dynamics, give rise to strong collective biases and polarized outcomes.38 This is a deeply significant and non-obvious result. It demonstrates that these FM-powered ABMs can be used to study the root causes of societal-level problems, like systemic bias or political polarization, that are not simply reducible to the psychology or intentions of individuals. This new tool allows social scientists to probe how these critical social challenges can emerge from the structure of our interactions and institutions, providing a new path toward understanding and potentially mitigating them.
Part III: A Curated Portfolio of 100 Foundational Model Concepts
This section presents the core deliverable of the report: a curated portfolio of 100 high-potential foundation model concepts. The list has been developed based on the sector-by-sector analysis in Part II, with a focus on concepts that are specific, possess transformative potential, and are feasible given the current trajectory of AI technology and data generation capabilities. This portfolio is intended to serve as a menu of actionable R&D targets and investment opportunities for stakeholders seeking to lead the development of the next generation of AI.
Table 2: Curated Portfolio of 100 Foundational Model Concepts
Physical Universe: Physics, Materials, and Chemistry
Materials Science
-
ElectrolyteExplorer - A generative foundation model conditioned on properties like ionic conductivity and electrochemical stability. It will be trained on the Materials Project database to propose novel solid-state electrolyte materials for next-generation batteries.
-
CatalystGen - An inverse design model that takes a target chemical reaction and desired efficiency as input. It will generate novel catalyst surfaces and molecular structures by exploring the vast chemical space of potential catalytic materials.
-
SuperconductorSeeker - A predictive model trained on experimental data and theoretical calculations for known superconducting materials. It will screen novel compounds to identify candidates with high critical temperatures, accelerating the search for room-temperature superconductors.
-
PolymerArchitect - A generative model for designing polymers with bespoke mechanical, thermal, and chemical properties. It will predict polymer performance based on monomer composition and chain architecture for applications in advanced manufacturing and sustainable plastics.
-
ThermoelectricDesigner - An inverse design model focused on discovering materials with high thermoelectric figures of merit (ZT). It will generate novel semiconductor compounds for efficient waste heat recovery and solid-state cooling applications.
-
PhotovoltaicOptimizer - A model trained on a comprehensive database of organic and perovskite solar cell materials. It will predict power conversion efficiency and long-term stability to guide the synthesis of next-generation photovoltaic technologies.
-
MOFBuilder - A generative model for designing novel Metal-Organic Frameworks (MOFs) with tailored pore geometries and chemical functionalities. It will be used to create materials optimized for carbon capture, hydrogen storage, and chemical separations.
-
CrystalPredictor-XRD - A model that learns to solve the phase problem in crystallography. It will take raw X-ray diffraction (XRD) pattern data as input and output the most probable crystal structures, automating a key bottleneck in materials characterization.
-
AlloyForge - A predictive model trained on metallurgical data to forecast the properties of complex alloys, such as strength, corrosion resistance, and performance at high temperatures. It will accelerate the design of new superalloys for aerospace and energy applications.
-
QuantumDotDesigner - A generative model for designing colloidal quantum dots with specific photoluminescent properties. It will predict emission spectra based on composition, size, and surface chemistry for advanced display and bio-imaging technologies.
Chemistry
-
RetroSynthAI - A foundation model for retrosynthesis that deconstructs a target molecule into simpler, commercially available precursors. It will learn from the entire corpus of chemical reaction literature to propose novel and efficient synthesis routes.
-
ReactionKineticsOracle - A predictive model that estimates the reaction rates and activation energies for chemical transformations. It will be trained on computational chemistry data to help chemists optimize reaction conditions without extensive experimentation.
-
SolventSelector - A model that predicts the effect of different solvents on reaction yield, selectivity, and rate. It will recommend optimal solvent systems to improve the efficiency and sustainability of chemical processes.
-
DrugDiscoverer - A generative model for de novo drug design that creates novel molecules with high predicted binding affinity to a specific biological target and favorable ADMET properties. It will integrate structural biology and bioactivity data to accelerate lead optimization.
-
ToxPredict - A foundation model trained on a massive database of toxicological studies. It will predict the potential toxicity of any given molecule to humans and the environment, enabling early-stage safety assessment in drug and materials development.
-
SpectraInterpreter - A multi-modal model that interprets complex analytical data from techniques like NMR, Mass Spectrometry, and IR spectroscopy. It will automatically elucidate the chemical structure of unknown compounds from their spectral fingerprints.
-
ProteinFolder-Alpha - An advanced protein structure prediction model trained on the full Protein Data Bank and metagenomic sequence databases. It will predict the 3D structure of proteins and protein complexes from their amino acid sequences with atomic-level accuracy.
-
EnzymeEvolver - A model that simulates the process of directed evolution for enzymes. It will predict the functional effects of mutations to guide the engineering of novel biocatalysts with enhanced activity, stability, and specificity.
Physics
-
QuantumNewton - An extension of the AI-Newton concept trained on experimental data from quantum mechanical systems. Its objective is to autonomously discover novel concepts and symbolic representations of quantum phenomena, potentially identifying patterns that hint at physics beyond the Standard Model.
-
CosmoSim-AI - A surrogate model trained on large-scale cosmological N-body simulations. It will provide rapid predictions of large-scale structure formation, such as the distribution of dark matter halos, for a given set of cosmological parameters.
-
ParticleColliderAnalyst - A model trained on petabytes of data from particle colliders like the LHC. It will be designed to perform real-time event classification and anomaly detection to search for new particles and rare physical processes.
-
PlasmaControl - A reinforcement learning-based foundation model for controlling plasma instabilities in real-time within a tokamak fusion reactor. It will learn control policies from simulation data to maintain stable, high-performance fusion plasmas.
-
AstroLens - A model that analyzes astronomical survey data to automatically detect and model gravitational lensing events. It will be used to map the distribution of dark matter in the universe and test theories of gravity.
-
StandardModelValidator - An unsupervised model trained on all known particle interaction data. Its purpose is to identify subtle deviations from the predictions of the Standard Model, pointing physicists toward areas where new physics may be discovered.
-
FluidMechanica - A general-purpose surrogate model for fluid dynamics, pre-trained on a vast and diverse library of canonical flow problems. It will be fine-tunable for specific engineering applications, from aerodynamics to hydraulics.
The Biological Code: Genomics, Systems Biology, and Neuroethology
Genomics/Proteomics
-
GeneRegulatorNet - A model that infers complete gene regulatory networks from single-cell RNA sequencing and ATAC-seq data. It will predict how transcription factors and non-coding DNA elements control gene expression in different cell types.
-
EpiGenomeMapper - A model that predicts the functional consequences of epigenetic modifications like DNA methylation and histone acetylation. It will help decipher how the epigenome regulates cellular identity and contributes to disease.
-
VariantInterpreter - A foundation model trained on population-scale genomic data and clinical records. It will predict the pathogenicity of novel genetic variants, aiding in the diagnosis of rare genetic diseases.
-
RNA-Struct - A model that predicts the three-dimensional structure and function of RNA molecules from their sequence. It will be crucial for understanding the roles of non-coding RNAs and for designing RNA-based therapeutics.
-
Proteome-Interactome - A model that predicts the complete network of protein-protein interactions within a cell. It will use sequence, structure, and expression data to map the cellular machinery underlying biological processes.
-
CRISPR-GuideDesigner - A model that designs optimal guide RNAs for CRISPR-based gene editing. It will predict both on-target efficiency and off-target effects to improve the safety and efficacy of gene therapies.
-
VirusEvolve - A foundation model trained on viral genomic sequences and epidemiological data. It will predict the evolutionary trajectories of viruses like influenza and coronaviruses, forecasting the emergence of new, potentially pandemic-causing variants.
-
Microbiome-Host - A model that learns the complex interactions between the human gut microbiome and host health. It will predict how changes in microbial composition affect metabolism, immunity, and disease risk.
Systems Biology
-
MetabolomeOracle - A predictive foundation model trained on multi-omics data to simulate the complete metabolic network of E. coli. It will predict the metabolic flux and product yield resulting from specific genetic interventions, accelerating metabolic engineering cycles.
-
YeastFactory - A digital twin of the Saccharomyces cerevisiae (baker's yeast) cell. It will be used to design and optimize metabolic pathways for the industrial production of pharmaceuticals, chemicals, and biofuels.
-
CellCycleSim - A dynamic model of the eukaryotic cell cycle. It will predict how perturbations to key regulatory proteins affect cell division, providing insights into cancer biology and regenerative medicine.
-
SignalingPathwayDecoder - A model that reconstructs cellular signaling pathways from phosphoproteomic and transcriptomic data. It will map how cells process information and make decisions in response to external stimuli.
-
SyntheticCircuitDesigner - A generative model for designing synthetic genetic circuits with predictable behavior. It will enable the engineering of cells with novel functions, such as biosensors or therapeutic delivery systems.
-
BiofuelOptimizer - A model focused on the metabolic engineering of photosynthetic organisms like algae and cyanobacteria. It will design genetic modifications to maximize the production of advanced biofuels from sunlight and CO2.
-
OrganoidGenesis - A model that simulates the self-organization and development of stem cells into organoids. It will help researchers understand tissue formation and create better in-vitro models for disease and drug testing.
-
Immunome-AI - A comprehensive simulation of the human immune system. It will predict the response to pathogens and vaccines, and model the dynamics of autoimmune diseases and immunotherapies.
-
TissueEngineer - A model that optimizes the conditions for tissue engineering, including scaffold design, growth factor cocktails, and mechanical stimuli. It will guide the development of lab-grown tissues and organs for transplantation.
Neuroethology
-
CephaloMind - A foundation model of the cephalopod brain, trained on neural and behavioral data from octopus and cuttlefish. It will aim to understand the principles of their distributed, non-mammalian intelligence and sophisticated camouflage abilities.
-
AvianNavigate - A model of the neural circuits underlying bird navigation. It will integrate data on head direction cells, grid cells, and magnetoreception to understand how birds perform long-distance migrations.
-
InsectBrain - A whole-brain emulation of a simpler insect, such as the fruit fly Drosophila. It will serve as a complete, executable model linking genes, neurons, and behavior in a single system.
-
PrimateSocialCognition - A model trained on neural recordings from primates engaged in social tasks. It will aim to decode the neural basis of complex social behaviors like cooperation, competition, and theory of mind.
Neuroscience
-
MotorCortex-Decoder - A foundation model for brain-computer interfaces that translates neural activity from the motor cortex into control signals for prosthetic limbs or computers. It will learn a general representation of motor intent that adapts quickly to new users.
-
MemoryTrace - A model of synaptic plasticity and memory engram formation in the hippocampus. It will simulate how memories are encoded, consolidated, and recalled at the circuit level.
-
SensoryIntegrator - A model of how the brain integrates information from multiple sensory modalities (e.g., vision, hearing, touch). It will be trained on neural responses to multi-sensory stimuli to understand the principles of perception.
-
SleepRhythm - A model of the neural circuits in the brainstem and hypothalamus that govern sleep-wake cycles. It will simulate the dynamics of sleep stages and their role in memory consolidation and brain health.
Complex Systems Simulation: Climate, Turbulence, and Engineering
Climate Science
-
GeoSurrogate-Climate - A high-fidelity surrogate for computationally expensive global climate models. It will provide rapid, ensemble-based projections of key climate variables under different emissions scenarios.
-
OceanCurrents-AI - A predictive model for global ocean circulation patterns, including phenomena like El Niño-Southern Oscillation. It will be trained on satellite altimetry, ocean buoys, and simulation data to improve seasonal forecasts.
-
AtmoChem - A surrogate model for complex atmospheric chemistry simulations. It will predict the formation and transport of pollutants like ozone and particulate matter to improve air quality forecasting.
-
Cryosphere-Melt - A model that predicts the dynamics of ice sheets in Greenland and Antarctica. It will learn from satellite data and physical models to provide more accurate projections of future sea-level rise.
-
CarbonCycle-AI - A data-driven model of the global carbon cycle. It will assimilate satellite and in-situ measurements to quantify carbon fluxes between the atmosphere, oceans, and land ecosystems.
-
ExtremeWeatherForecaster - A foundation model specifically trained to predict the genesis, intensity, and track of high-impact weather events like hurricanes, tornadoes, and atmospheric rivers. It will learn from decades of historical weather data and high-resolution simulations.
Earth Science
-
SeismicPredict - A model that analyzes continuous seismic and geodetic data streams to identify subtle precursor patterns to earthquakes. Its goal is to move beyond statistical forecasting to provide probabilistic, short-term risk assessments.
-
HydroCycle - A model of the global terrestrial water cycle. It will predict soil moisture, groundwater levels, and river flows to improve drought and flood forecasting.
-
WildfireSpread - A real-time wildfire behavior model that integrates weather forecasts, fuel maps, and topography. It will predict the rate and direction of fire spread to aid in firefighting and evacuation efforts.
-
SolarCycle-Surya - A foundation model trained on multi-modal solar observation data. It will predict solar flares and coronal mass ejections to improve space weather forecasting and protect critical infrastructure.
Fluid Dynamics
-
AeroSurrogate-1 - A physics-informed surrogate model trained on a massive dataset of high-fidelity CFD simulations and wind tunnel data for various airfoil geometries. It will provide real-time prediction of aerodynamic forces and flow fields, replacing expensive simulations in early-stage aircraft design.
-
TurbulenceClosure-AI - A model designed to discover new, more accurate, and generalizable closure models for RANS simulations. It will learn from DNS data to output symbolic equations that represent the Reynolds stresses, a fundamental challenge in fluid mechanics.
-
CombustionSim - A surrogate model for detailed chemical kinetics in combustion simulations. It will accelerate the design of more efficient and cleaner engines, gas turbines, and rocket propulsion systems.
-
MultiphaseFlow - A model for simulating complex multiphase flows, such as oil, water, and gas mixtures in pipelines or bubbly flows in chemical reactors. It will learn the dynamics of phase interfaces from experimental and simulation data.
Solid Dynamics
- StructuralIntegrity-AI - A predictive model for material fatigue and fracture mechanics. It will forecast the remaining useful life of mechanical components by learning from sensor data and simulation of crack propagation.
Acoustics
- AcousticWave - A model for predicting the generation and propagation of sound in complex environments. It will be used for applications ranging from reducing aircraft noise to designing concert hall acoustics.
Industrial Physics
- GranularFlow - A model that simulates the behavior of granular materials like sand, grains, and powders. It will be used to optimize industrial processes in agriculture, pharmaceuticals, and manufacturing.
Geophysics
- GeoMechanics - A surrogate model for geomechanical simulations. It will predict subsurface stress, deformation, and fracture propagation for applications in geothermal energy, carbon sequestration, and resource extraction.
Energy Systems
- GridStability-AI - A foundation model of the national power grid that predicts grid stability and cascading failure risk in real-time. It will be trained on sensor data from across the grid to manage the integration of intermittent renewable energy sources.
Engineering Systems
-
SupplyChain-Opt - A digital twin of global supply chains. It will simulate the flow of goods and identify vulnerabilities to disruptions from geopolitical events, climate change, or pandemics.
-
UrbanMobility - A city-scale agent-based model of traffic and public transit. It will be used by urban planners to test the impact of new infrastructure, transportation policies, and autonomous vehicle deployment.
-
ManufacturingProcess-Twin - A digital twin for complex manufacturing processes, such as semiconductor fabrication or biopharmaceutical production. It will use sensor data to predict yield, optimize process parameters, and perform predictive maintenance.
-
BuildingEnergy-Mod - A model that predicts the energy consumption of commercial and residential buildings. It will be used to design more efficient buildings and optimize the operation of HVAC systems.
-
ReservoirSim - A surrogate model for petroleum reservoir simulations. It will rapidly predict oil and gas production under different operational strategies to maximize resource recovery.
-
BatteryLifecycle - A model that predicts the degradation and aging of batteries over their lifetime. It will be used to optimize battery management systems for electric vehicles and grid storage, extending their lifespan and performance.
Emergent Social Dynamics: Economics, Social Science, and Human-System Interaction
Economics
-
MarketSim - An agent-based model foundation populated by millions of LLM agents representing consumers, producers, and investors with distinct goals and behavioral heuristics. It will be used to simulate emergent market phenomena like asset bubbles and crashes.
-
MacroEcon-AI - A foundation model that simulates the entire economy of a nation or region. It will be used to forecast the impact of fiscal and monetary policy changes on GDP, inflation, and unemployment.
-
SystemicRisk-Detector - A model of the interbank lending network and financial system. It will identify institutions that are "too connected to fail" and simulate how shocks can propagate through the system, causing financial crises.
-
ConsumerBehavior-ABM - An agent-based model that simulates consumer purchasing decisions and the adoption of new products. It will be trained on market data to predict how trends and fads emerge and spread through a population.
-
TradeFlow-AI - A dynamic model of the global trade network. It will predict how tariffs, trade agreements, and geopolitical events alter the flow of goods and impact national economies.
-
LaborMarket-Dynamics - An agent-based simulation of the labor market, with agents representing workers and firms. It will be used to study the effects of automation, minimum wage laws, and education policies on employment and inequality.
-
CryptoEcon - A model for simulating the economic dynamics and stability of decentralized finance (DeFi) protocols and cryptocurrency ecosystems. It will be used to stress-test protocols for vulnerabilities and emergent failure modes.
-
AuctionTheorist - A model that learns to design optimal auction mechanisms for specific environments. It will be used for applications like spectrum auctions and online advertising markets.
-
FirmEvolution - An agent-based model where agents are firms competing in a market. It will simulate how industries evolve over time through innovation, competition, and strategic interaction.
-
DevelopmentEcon-ABM - An agent-based model for studying economic development. It will simulate how factors like education, infrastructure, and institutional quality can help or hinder a region's escape from poverty traps.
Social Science
-
NormFormation-AI - A multi-agent simulation that models how social norms and conventions, from linguistic conventions to moral norms, emerge and stabilize in a population through local interactions. It will be used to test theories of cultural evolution.
-
OpinionDynamics - A model of how opinions and beliefs spread and evolve within a social network. It will be used to study the drivers of political polarization and the formation of echo chambers.
-
CollectiveAction-Sim - An agent-based model designed to simulate the conditions under which collective action, such as protests or social movements, emerges. It will explore the roles of social networks, grievances, and critical mass dynamics.
-
UrbanSegregation-ABM - An advanced simulation of residential segregation in cities, extending classic models like Schelling's. It will incorporate realistic agent behaviors and economic constraints to understand the drivers of and solutions to segregation.
-
GovernanceAI - A simulation environment for comparing the stability and outcomes of different systems of governance. It will model how different voting rules, institutional structures, and constitutional arrangements affect political outcomes.
-
InfoWarfare-Detector - A model that simulates the spread of misinformation and disinformation campaigns through social media networks. It will be used to understand their dynamics and test the effectiveness of different mitigation strategies.
-
CulturalEvolution - A model that simulates the long-term evolution of cultural traits, such as languages, technologies, and social structures. It will explore how demographic and environmental factors shape human cultural diversity.
-
VoterModel-AI - An agent-based model of voter behavior. It will simulate election outcomes based on demographic data, social influence, and campaign effects to understand the dynamics of democratic elections.
-
InstitutionalDesign - A model that allows for the in-silico testing of new institutional designs, such as market regulations or international treaties. It will predict the likely behavioral responses and emergent outcomes of different rule sets.
-
SocialNetwork-Evolve - A model that simulates the co-evolution of individual attributes and social network structure. It will explore how phenomena like friendship formation and social status dynamics unfold over time.
Human-System Interaction
-
AI-Collaboration-Sim - An agent-based model populated by both human and AI agents. It will be used to study the emergent dynamics of human-AI teams and identify principles for designing effective collaborative intelligence systems.
-
PlatformEcology - A model of the dynamics of online platforms like social media or e-commerce sites. It will simulate the interactions between users, content creators, and platform algorithms to understand the health and stability of digital ecosystems.
-
GigEconomy-ABM - An agent-based model of the gig economy, with agents representing workers, consumers, and platform companies. It will be used to study issues of wage dynamics, labor supply, and the impact of algorithmic management.
-
TrustDynamics - A model that simulates how public trust in institutions, technologies, and media evolves over time. It will explore the factors that lead to the erosion or building of social trust.
-
EthicalAI-ABM - A multi-agent simulation where AI agents learn and evolve their behaviors. It will be used as a testbed to study how ethical or unethical collective AI behavior can emerge, even from simple individual rules, informing AI safety research.
Part IV: Strategic Imperatives for R&D and Data Acquisition
The successful development of the foundation models outlined in this report is not merely a question of algorithmic innovation; it is fundamentally a challenge of data infrastructure and organizational strategy. The primary limiting factor for the advancement of SciFMs is the availability of large-scale, high-quality, and well-structured scientific data.2 Unlike the web-scale text and image data that fueled the generative AI boom, scientific data is often expensive to generate, difficult to standardize, and siloed within individual labs, institutions, or proprietary databases. To overcome this grand challenge and unlock the future of AI-driven science, a concerted, multi-faceted strategy is required. This strategy rests on three core pillars.
4.1 A Three-Pillar Strategy for Data Dominance
A comprehensive research and development plan must be implemented to collect the sufficient, high-quality data required to begin training these new classes of foundation models. This plan begins with the establishment of large-scale, centralized "Data Commons" for key scientific domains, which will serve as the foundational infrastructure for model training and require extensive public-private partnerships to create standards for data sharing and management.2 Concurrently, a "simulation-experimentation flywheel" must be built, where high-fidelity simulations generate massive, clean datasets for pre-training models, and these models, in turn, guide more efficient physical experiments for validation, creating a virtuous cycle of data generation and discovery. Finally, this entire effort must be driven by a new organizational paradigm of cross-disciplinary "fusion" teams, deeply integrating domain scientists with machine learning and high-performance computing experts to ensure the models are both scientifically rigorous and computationally feasible.1
4.1.1 Pillar 1: Architecting the Scientific Data Commons
The foundational imperative is to treat scientific data as a public good and a core piece of research infrastructure. This requires a strategic shift away from fragmented, project-specific data collection toward the creation of large-scale, centralized, and standardized "Data Commons" for key scientific domains.40 These platforms must go beyond simple data storage; they must be architected as integrated environments that aggregate data from myriad public and private sources and provide the necessary cloud-based, high-performance computing resources for massive-scale model training. In this model, the data itself becomes part of the shared cloud infrastructure, as essential as storage or networking.40 Successfully building these commons will necessitate significant public-private partnerships and international collaboration to establish and enforce robust standards for data quality, metadata, sharing protocols, management, and reuse, ensuring the data is not only accessible but also usable for training the next generation of SciFMs.2
4.1.2 Pillar 2: Building the Simulation-Experimentation Flywheel
The second pillar focuses on creating a self-reinforcing, virtuous cycle that dramatically accelerates the rate of data generation and scientific discovery. This "simulation-experimentation flywheel" leverages the complementary strengths of computational modeling and physical experimentation. The cycle begins with high-fidelity simulations (e.g., DNS in fluid dynamics, DFT in materials science) generating vast, clean, and physically consistent datasets that are ideal for the initial pre-training of foundation models. The pre-trained models, now imbued with a foundational understanding of the system, are then used to make rapid predictions that guide more efficient and targeted physical experiments. Instead of exploring a vast parameter space blindly, experimental resources are focused on the most promising and informative areas of inquiry identified by the model. The data from these targeted experiments is then used to validate, fine-tune, and further improve the model, which in turn enables even more powerful simulations, completing and accelerating the flywheel.
4.1.3 Pillar 3: Cultivating Cross-Disciplinary "Fusion" Teams
The third and final pillar is organizational. The complexity of building and validating SciFMs demands a new model of scientific collaboration that breaks down traditional disciplinary silos. Progress will be fastest in organizations that create deeply integrated "fusion" teams, co-locating domain scientists (e.g., physicists, biologists, chemists), machine learning researchers, and high-performance computing engineers.1 In this model, the development process is not a linear handoff from one group to the next. Instead, all three areas of expertise are brought to bear simultaneously. The domain scientist ensures the model's inputs, constraints, and outputs are scientifically meaningful; the machine learning researcher designs the model architecture and training procedures; and the HPC engineer ensures the entire workflow can scale efficiently on modern supercomputing hardware. This collaborative structure is essential for navigating the fundamental trade-offs in SciFM design and for ensuring that the resulting models are not just computationally powerful but also scientifically valid and impactful.
Conclusion: The Future of AI-Driven Science
The advent of Scientific Foundation Models marks a pivotal moment in the history of science and technology. The analysis presented in this report indicates that we are moving beyond an era where AI is merely a tool for data analysis and into one where it becomes a genuine partner in discovery. The opportunities are not incremental; they are transformative. We are on the cusp of developing AI systems that can design novel materials for clean energy, engineer microorganisms to produce life-saving drugs, provide early warnings for extreme weather events, and even discover the fundamental laws that govern our universe.
Realizing this future, however, is not inevitable. It requires a strategic and sustained commitment from leaders across academia, industry, and government. The romantic image of the lone scientific genius is being replaced by a new reality where breakthroughs are achieved by collaborative, cross-disciplinary teams leveraging vast computational resources and shared data infrastructure. The primary bottleneck is no longer a lack of computational power or algorithmic ingenuity, but a deficit of large-scale, high-quality, accessible scientific data.
Therefore, the path forward is clear. The call to action is to make the bold, long-term investments necessary to build the open data commons, foster the simulation-experimentation flywheels, and cultivate the fusion teams that will power this new scientific revolution. By embracing this Fourth Paradigm and harnessing the power of Scientific Foundation Models, we can accelerate the pace of human progress and unlock solutions to some of the most pressing challenges of our time. This is not simply the next chapter in the story of artificial intelligence; it is the beginning of a fundamental redefinition of the scientific method itself.
Works cited
- The Fourth Paradigm: Data-Intensive Scientific Discovery - Microsoft Research, accessed August 23, 2025, https://www.microsoft.com/en-us/research/publication/fourth-paradigm-data-intensive-scientific-discovery/
- The Future of Science Policy: Data-Intensive Research - Number Analytics, accessed August 23, 2025, https://www.numberanalytics.com/blog/future-of-science-policy-data-intensive-research
- The data-intensive research paradigm: challenges and responses in clinical professional graduate education - PMC, accessed August 23, 2025, https://pmc.ncbi.nlm.nih.gov/articles/PMC11842464/
- The Future of Data Science: Emerging Technologies and Trends - University of the Cumberlands, accessed August 23, 2025, https://www.ucumberlands.edu/blog/the-future-of-data-science-emerging-technologies-and-trends
- A Perspective on Foundation Models in Chemistry | JACS Au, accessed August 23, 2025, https://pubs.acs.org/doi/10.1021/jacsau.4c01160
- What are Foundation Models? - Foundation Models in Generative AI Explained - AWS, accessed August 23, 2025, https://aws.amazon.com/what-is/foundation-models/
- Foundation model - Wikipedia, accessed August 23, 2025, https://en.wikipedia.org/wiki/Foundation_model
- On the Opportunities and Risks of Foundation Models - Stanford CRFM, accessed August 23, 2025, https://crfm.stanford.edu/assets/report.pdf
- Building AI Foundation Models to Accelerate the Discovery of New Battery Materials, accessed August 23, 2025, https://www.hpcwire.com/2025/08/19/building-ai-foundation-models-to-accelerate-the-discovery-of-new-battery-materials/
- AI and the Language of Mathematics: How Artificial Intelligence is Unlocking the Universe's Most Complex Problems | by Leon Tyron | Medium, accessed August 23, 2025, https://medium.com/@leontyron/ai-and-the-language-of-mathematics-how-artificial-intelligence-is-unlocking-the-universes-most-7db2258f9af8
- The End of Physics? AI Is Discovering New Laws of the Universe - Without Us - Leximancer, accessed August 23, 2025, https://www.leximancer.com/blog/0lu21hnlp0ho7z1qxvs14jsrpx94op
- Discovery of Physics From Data: Universal Laws and Discrepancies - PMC, accessed August 23, 2025, https://pmc.ncbi.nlm.nih.gov/articles/PMC7861345/
- Foundation Models Shift Paradigms for Engineering and Energy - JPT/SPE, accessed August 23, 2025, https://jpt.spe.org/foundation-models-shift-paradigms-for-engineering-and-energy
- Archetype AI's Newton Model Masters Physics From Raw Data - HPCwire, accessed August 23, 2025, https://www.hpcwire.com/2024/10/28/archetype-ais-newton-model-masters-physics-from-raw-data/
- Can AI Discover New Laws of Physics? A Thought Experiment in Quantum Weirdness | by Sevak Avakians | Medium, accessed August 23, 2025, https://medium.com/@sevakavakians/can-ai-discover-new-laws-of-physics-a-thought-experiment-in-quantum-weirdness-a373d369858e
- Towards Foundation Models for Materials Science: The Open MatSci ML Toolkit - arXiv, accessed August 23, 2025, https://arxiv.org/pdf/2310.07864
- AI-Newton: A Concept-Driven Physical Law Discovery System without Prior Physical Knowledge - arXiv, accessed August 23, 2025, https://arxiv.org/html/2504.01538v1
- (PDF) AI-Newton: A Concept-Driven Physical Law Discovery System without Prior Physical Knowledge - ResearchGate, accessed August 23, 2025, https://www.researchgate.net/publication/390440166_AI-Newton_A_Concept-Driven_Physical_Law_Discovery_System_without_Prior_Physical_Knowledge
- AI-Newton: A Concept-Driven Physical Law Discovery System without Prior Physical Knowledge | AI Research Paper Details, accessed August 23, 2025, https://www.aimodels.fyi/papers/arxiv/ai-newton-concept-driven-physical-law-discovery
- [2504.01538] AI-Newton: A Concept-Driven Physical Law Discovery System without Prior Physical Knowledge - arXiv, accessed August 23, 2025, https://arxiv.org/abs/2504.01538
- Modeling for understanding and engineering metabolism | QRB Discovery | Cambridge Core, accessed August 23, 2025, https://www.cambridge.org/core/journals/qrb-discovery/article/modeling-for-understanding-and-engineering-metabolism/18553F7A257B68AB6403E5D4551E3B65
- MIT Open Access Articles The future of metabolic engineering and synthetic biology: Towards a systematic practice, accessed August 23, 2025, https://dspace.mit.edu/bitstream/handle/1721.1/99397/Stephanopoulos_Future%20metabolic.pdf?sequence=1&isAllowed=y
- Computational Neuroethology: Simulating Natural Behaviors - Frontiers, accessed August 23, 2025, https://www.frontiersin.org/research-topics/71920/computational-neuroethology-simulating-natural-behaviors
- Neuroethology - Wikipedia, accessed August 23, 2025, https://en.wikipedia.org/wiki/Neuroethology
- Computational Neuroethology: A Call to Action - ResearchGate, accessed August 23, 2025, https://www.researchgate.net/publication/336399298_Computational_Neuroethology_A_Call_to_Action
- Computational Neuroethology Unit | Okinawa Institute of Science ..., accessed August 23, 2025, https://www.oist.jp/research/research-units/cne
- Parts plus pipes: synthetic biology approaches to metabolic engineering - PubMed Central, accessed August 23, 2025, https://pmc.ncbi.nlm.nih.gov/articles/PMC3293987/
- Session 6: Synthetic Biology and Metabolic Engineering - iBiology, accessed August 23, 2025, https://www.ibiology.org/sessions/session-6-synthetic-biology-metabolic-engineering/
- Synthetic biology: A foundation for multi-scale molecular biology - PMC - PubMed Central, accessed August 23, 2025, https://pmc.ncbi.nlm.nih.gov/articles/PMC3037580/
- Advancement of Metabolic Engineering Assisted by Synthetic Biology - MDPI, accessed August 23, 2025, https://www.mdpi.com/2073-4344/8/12/619
- Machine learning for turbulence modeling - Monolith AI, accessed August 23, 2025, https://www.monolithai.com/blog/machine-learning-for-turbulence-modeling
- Can Artificial Intelligence Accelerate Fluid Mechanics Research? - MDPI, accessed August 23, 2025, https://www.mdpi.com/2311-5521/8/7/212
- AI for Fluid Mechanics - TU Delft, accessed August 23, 2025, https://www.tudelft.nl/en/ae/organisation/departments/flow-physics-and-technology/aerodynamics/research/ai-for-fluid-mechanics
- How Will AI Impact Computational Fluid Dynamics? - Resolved Analytics, accessed August 23, 2025, https://www.resolvedanalytics.com/ai-in-cfd/how-will-ai-impact-cfd
- IBM and NASA Release Groundbreaking Open-Source AI Model on Hugging Face to Predict Solar Weather and Help Protect Critical Technology, accessed August 23, 2025, https://newsroom.ibm.com/2025-08-20-ibm-and-nasa-release-groundbreaking-open-source-ai-model-on-hugging-face-to-predict-solar-weather-and-help-protect-critical-technology
- Understanding Emergent Social Phenomena Comparatively: The Need for Computational Simulation - ResearchGate, accessed August 23, 2025, https://www.researchgate.net/publication/255556995_Understanding_Emergent_Social_Phenomena_Comparatively_The_Need_for_Computational_Simulation
- System Theoretic Foundations for Emergent Behavior Modeling: The Case of Emergence of Human Language in a Resource-Constrained Complex Intelligent Dynamical System | Request PDF - ResearchGate, accessed August 23, 2025, https://www.researchgate.net/publication/324363181_System_Theoretic_Foundations_for_Emergent_Behavior_Modeling_The_Case_of_Emergence_of_Human_Language_in_a_Resource-Constrained_Complex_Intelligent_Dynamical_System
- Emergent social conventions and collective bias in LLM populations - PMC - PubMed Central, accessed August 23, 2025, https://pmc.ncbi.nlm.nih.gov/articles/PMC12077490/
- Emergent Social Conventions and Collective Bias in LLM Populations11footnote 1Preprint version of: Science Advances 11 (20), eadu9368 (2025). - arXiv, accessed August 23, 2025, https://arxiv.org/html/2410.08948v2
- Empirical Modeling of Complex Systems | NSF - National Science Foundation, accessed August 23, 2025, https://www.nsf.gov/events/empirical-modeling-complex-systems-0/2016-03-03
Resources Overview
This landing page will feature a list of ongoing RESOURCES. We will develop a template after we have experience with several examples.
An RESOURCE begins first as a PROJECT and which has perhaps then moved on to AREA status and then graduates to RESOURCE status after it is basically complete. In principle, a PROJECT might move directly to RESOURCE status, but it's more likely that something would get krausened in AREA status for awhile before graduating to RESOURCE status.
A Project is the start of a bigger development commitment and the basis of the P.A.R.A. method of the Building a Second Brain (BASB) methodology. The BASB method systematically manages information differently than just notetaking apps ... PROJECTS, have goals, reqmts and deadlines ... AREAS are about roles/responsibilities or obligations or capabilities that need to be earnestly developed ... RESOURCES, mostly finished AREAS, but also ongoing interests, assets, future inspiration, may req continual maintenance and refactoring but, for now, are backburnerable ... ARCHIVES, inactive matl from P A R that shouldn't be used, except for informational purposes.
GitHub Discussion, Issue, Project Functionality
We will rely upon the GitHub Discussion and Issue functionality, BEFORE graduating something to "Project" status ... when something becomes a Project on GitHub, it will simultaneously become a PROJECT in our P.A.R.A. hierarchy.
Please understand the GitHub progression from ... Discussions ...to... Issue ...to... Project.
Discussions are mainly for just discussing something, to clarify terminology or ask questions or for just generally speculative thinking out loud.
Issues are for things that somebody really needs to look into and possibly turn into more of a Project.
On GitHub a Project is an adaptable spreadsheet, task-board, and road map that integrates with your issues and pull requests on GitHub to help you plan and track your work effectively. You can create and customize multiple views by filtering, sorting, grouping your issues and pull requests, visualize work with configurable charts, and add custom fields to track metadata specific to your team. Rather than enforcing a specific methodology, a project provides flexible features you can customize to your team’s needs and processes.
Archives Overview
This landing page will feature a list of ongoing ARCHIVES. We will develop a template after we have experience with several examples.
An ARCHIVE is a PROJECT, AREA or RESOURCE that's no longer relevant or useful. It might be something that is now deprecated, even discredited or a failure or a bad idea that we regret ever bothering with, but it does not matter -- we keep things in the ARCHIVE because they might be useful for informational purposes.
A Project is the start of a bigger development commitment and the basis of the P.A.R.A. method of the Building a Second Brain (BASB) methodology. The BASB method systematically manages information differently than just notetaking apps ... PROJECTS, have goals, reqmts and deadlines ... AREAS are about roles/responsibilities or obligations or capabilities that need to be earnestly developed ... RESOURCES, mostly finished AREAS, but also ongoing interests, assets, future inspiration, may req continual maintenance and refactoring but, for now, are backburnerable ... ARCHIVES, inactive matl from P A R that shouldn't be used, except for informational purposes.
GitHub Discussion, Issue, Project Functionality
We will rely upon the GitHub Discussion and Issue functionality, BEFORE graduating something to "Project" status ... when something becomes a Project on GitHub, it will simultaneously become a PROJECT in our P.A.R.A. hierarchy.
Please understand the GitHub progression from ... Discussions ...to... Issue ...to... Project.
Discussions are mainly for just discussing something, to clarify terminology or ask questions or for just generally speculative thinking out loud.
Issues are for things that somebody really needs to look into and possibly turn into more of a Project.
On GitHub a Project is an adaptable spreadsheet, task-board, and road map that integrates with your issues and pull requests on GitHub to help you plan and track your work effectively. You can create and customize multiple views by filtering, sorting, grouping your issues and pull requests, visualize work with configurable charts, and add custom fields to track metadata specific to your team. Rather than enforcing a specific methodology, a project provides flexible features you can customize to your team’s needs and processes.
Roadmap
It has become clear that the point of this specific PKE project is actually about a Requirements elicitation process for AI/ML Ops.
The following is rough a breakdown of the key steps and considerations involved:
-
Understanding the problem and scope Clearly define the problem: Articulate the specific business problem or opportunity that the AI/ML solution aims to address. Identify the target users and their needs: Understand how the AI/ML system will impact their workflows and decision-making. Determine the desired outcomes and metrics for success: Establish clear and measurable goals for the AI/ML project.
-
Identifying key stakeholders Data scientists: Understand their needs related to data access, model development, and experimentation environments. ML engineers: Gather requirements for model deployment, monitoring, and scaling in production environments. Operations teams (IT/DevOps): Elicit needs related to infrastructure, security, and integration with existing systems. Business stakeholders: Understand the business value, impact, and desired functionality of the AI/ML solution. End-users: Gather feedback and requirements to ensure user-centricity and usability of the AI/ML system. Other departments (Marketing, Sales, HR, Legal): Recognize potential input on project purpose, scope, or goals depending on the AI project type.
-
Techniques for eliciting requirements
Develop a workable PKE system by adapting existing tech: As we use existing already-developed technology for PKE, we will be able to delve into specific needs, concerns, and expectations.
Modules as requirements workshops: The 100-module PKE course actually is about facilitate sessions, possibly including collaborators, to brainstorm, refine, and prioritize requirements with a group of stakeholders.
Surveys, polls and questionnaires: The internet, social media and discussion fora like Discord, Slack, et al give us a way to gather information from different larger audiences, especially when seeking input from diverse users or collecting data on specific aspects of the system.
Document analysis: AI helps immensely with reviewing existing documentation and process info, system specifications, roadmaps and data reports, to better identify current requirements and potential areas for improvement.
Prototyping: Create interactive mockups or early versions of the AI/ML system to gather feedback and refine requirements based on user interaction.
Observation/Ethnography: Observe users in their natural environment to gain a deeper understanding of their workflow, challenges, and unspoken needs that the AI/ML solution can address.
Brainstorming: Encourage the free flow of ideas to uncover innovative solutions and identify new requirements, especially in the early stages of a project.
Use Cases/User Stories: Capture system functionality from the perspective of different users and their interactions with the AI/ML system.
- Addressing unique challenges in AI/ML requirements elicitation
Data Quality and Availability: Elicit requirements for data collection, quality checks, governance frameworks, and security protocols to ensure reliable data for training and deploying AI/ML models.
Explainability and Interpretability: Define requirements for understanding how the AI/ML system makes decisions, especially in critical domains, to build trust and ensure accountability.
Bias and Fairness: Elicit requirements for detecting, mitigating, and monitoring potential biases in AI/ML models to ensure fair and equitable outcomes.
Scalability and Performance: Understand the need for the AI/ML solution to handle increasing workloads and complex problem-solving without compromising performance.
Integration with Existing Systems: Assess and define requirements for seamlessly integrating the AI/ML solution with legacy infrastructure and other applications.
Ethical and Regulatory Compliance: Consider and address ethical implications, privacy concerns, and compliance with data protection laws and industry regulations (e.g., GDPR) from the outset.
Evolving Requirements: Recognize the iterative nature of AI/ML development and accommodate changes and refinements throughout the project lifecycle.
- Documentation, validation, and prioritization
Document requirements clearly and consistently: Use structured formats like user stories, use cases, or requirement specifications, tailored to the project methodology (e.g., Agile, Waterfall).
Analyze and negotiate requirements: Identify potential conflicts, gaps, and redundancies in the gathered requirements and negotiate with stakeholders to prioritize based on business value, criticality, and dependencies.
Validate and verify requirements: Ensure that the documented requirements are complete, consistent, feasible, and align with business objectives.
Baseline and manage requirements: Establish a baseline for the approved requirements and implement a process for managing changes and tracking progress throughout the project lifecycle.
100 Module Plan: Developing Minimalist CloudKernel for AI/ML Acceleration
The proliferation of artificial intelligence and machine learning (AI/ML) has placed unprecedented demands on computing infrastructure. While hardware accelerators like GPUs have become the cornerstone of AI/ML computation, the operating systems that manage them remain largely general-purpose. Traditional OS kernels, designed for multitasking and fairness in interactive environments, introduce overheads—such as context switching, complex scheduling, and broad system call interfaces—that are often unnecessary for the dedicated, high-throughput workloads characteristic of AI/ML training and inference.1 This creates an opportunity for a paradigm shift: the development of a specialized, minimalist kernel architected from first principles to serve a single purpose—hosting AI/ML containers with maximum efficiency.
This document outlines an exhaustive, year-long, 100-module course designed for advanced systems engineers and researchers. The curriculum guides the student through the complete, from-scratch development of such a kernel. The objective is not merely to build an operating system, but to create a highly optimized, vertically integrated software stack. This includes a bare-metal kernel written in Rust, deep integration with NVIDIA CUDA and AMD ROCm GPU hardware at the driver level, a bespoke compiler toolchain based on modern MLIR infrastructure, and a container runtime tailored for AI/ML applications. Each module is designed to require 10-15 hours of intensive, hands-on effort, progressing from foundational bare-metal programming to the deployment of a custom kernel in a cloud environment. The final product will be a testament to low-level systems engineering: a kernel that sheds the legacy of general-purpose computing to provide a lean, powerful, and transparent foundation for the next generation of intelligent systems.
Part I: Foundations of Low-Level Systems and Kernel Bootstrapping (Modules 1-12)
This foundational part establishes the philosophical and technical groundwork for the entire course. It moves from the "why" of building a minimalist kernel to the "how" of writing the very first lines of code that will execute on bare metal. The initial modules are dedicated to architectural decisions, toolchain setup, and understanding the hardware platform, culminating in a bootable, "hello world" kernel.
Section 1.1: The Minimalist Kernel Philosophy: Unikernels vs. Microkernels vs. Monolithic (Modules 1-3)
The course begins by deconstructing the architectural trade-offs of different kernel designs to justify the selection of a specialized model for AI/ML workloads. Monolithic kernels, like Linux, integrate all services into a single address space for performance but suffer from complexity and a large attack surface. Microkernels prioritize isolation and modularity by moving services into user space, often at the cost of performance due to increased inter-process communication.
This course will pursue a unikernel-inspired, library-OS model. Unikernels are specialized, single-address-space machine images constructed by linking application code with only the necessary kernel libraries, resulting in extremely small, fast-booting, and efficient virtual machines.2 Projects like Unikraft and Nanos demonstrate this philosophy, offering POSIX- and Linux-compatible interfaces to ease application porting.4 However, these projects also highlight the primary challenge of our endeavor: integrating complex, proprietary device drivers. The effort to support NVIDIA GPUs in the Nanos unikernel, for instance, required a painstaking re-implementation of numerous Linux kernel internal features, including waitqueues, radix trees, custom
mmap callbacks, and synchronization primitives.5 This reveals an architectural paradox central to this course: while the core kernel can be minimalist, the GPU driver subsystem it must host is inherently monolithic and complex. Therefore, the kernel we build will be a "bimodal" system—a minimal core OS co-located with a highly complex driver subsystem that necessitates a significant, Linux-like ABI. This is not a failure of the minimalist philosophy but a pragmatic acknowledgment of the realities of supporting high-performance, proprietary hardware.
Furthermore, we will examine forward-looking concepts like "AI-aware" kernels, which propose integrating deep learning models directly into the kernel as Loadable Kernel Modules (LKMs) for ultra-low-latency processing.1 While our immediate goal is to efficiently
host AI containers, these advanced concepts will inform our design, encouraging an architecture that minimizes the user-kernel boundary wherever possible.
Module Breakdown:
- Module 1: Kernel Architectures. Analysis of monolithic, microkernel, and unikernel designs. A case study on the performance and security characteristics of each.
- Module 2: The Library OS Model for AI/ML. Justifying the choice of a unikernel-inspired architecture. A deep dive into the Nanos GPU driver porting effort as a case study on the challenges of driver compatibility.5
- Module 3: The AI-Native Kernel Concept. Exploring academic research on integrating AI directly into the kernel space.1 Defining the scope of our project as a high-performance host, with an eye toward future AI-native optimizations.
Section 1.2: Bare-Metal Development with Rust: no_std, Core Primitives, and Unsafe Code (Modules 4-6)
Rust is selected as the implementation language for its unique combination of performance, comparable to C/C++, and strong compile-time safety guarantees that eliminate entire classes of bugs like dangling pointers and data races.8 A key feature for OS development is Rust's standard library structure, which is split into core and alloc. The core library contains platform-independent primitives and can be used in a "freestanding" environment, ie a Rust environment with #[no_std] attribute, without an underlying OS, which is precisely our starting point.
This section focuses on the practicalities of bare-metal Rust. While the ownership model provides safety, kernel development inherently requires unsafe operations: direct memory-mapped I/O, manipulation of page tables, and handling raw pointers for DMA. A naive approach would wrap large sections of the kernel in unsafe blocks, negating Rust's benefits. The correct approach, and a central pedagogical theme of this course, is to master the art of building safe abstractions over unsafe operations. We will study patterns where minimal, well-documented unsafe code is encapsulated within a module that exposes a completely safe public API. This pattern is crucial for building a robust and maintainable kernel.
The curriculum will follow the initial steps laid out in the "Writing an OS in Rust" blog series, beginning with the creation of a freestanding binary.10 Students will set up the development environment using
rustup to manage toolchains and cargo as the build system and package manager.11 We will configure a custom cross-compilation target and survey essential libraries from the Rust OSDev ecosystem, such as the
x86_64 crate for direct access to CPU instructions and registers, and the spin crate for basic synchronization primitives.9
Module Breakdown:
- Module 4: The Rust Programming Model for Systems. Introduction to ownership, borrowing, and lifetimes. Setting up a no_std project.
- Module 5: Encapsulating Unsafe Code. Best practices for using the unsafe keyword. Building safe abstractions for hardware interaction, such as a safe VGA text buffer writer.10
- Module 6: The Rust OSDev Ecosystem. Toolchain setup for cross-compilation. Introduction to cargo, xargo, and key crates like x86_64, bootloader, and spin.9
Section 1.3: The x86_64 Architecture: Boot Process, Memory Models, and CPU Modes (Modules 7-9)
A deep, non-negotiable dive into the fundamentals of the x86_64 architecture is essential before writing any kernel code. This section covers the complete boot sequence, from power-on to the point where a bootloader hands off control to a kernel. We will examine the transition of the CPU through its various operating modes: starting in 16-bit Real Mode, transitioning to 32-bit Protected Mode, and finally entering 64-bit Long Mode, which is where our kernel will operate.
Key architectural concepts such as memory segmentation and the crucial role of paging for virtual memory will be introduced. We will study the Global Descriptor Table (GDT) and the structure of multi-level page tables. The primary hands-on platform for this section will be the QEMU emulator, which provides a flexible and debuggable environment for testing our kernel without risk to the host machine.12 Students will learn the basic QEMU command-line options for booting a kernel image, providing an initial RAM disk (
initrd), and passing kernel command-line arguments, such as console=ttyS0 to redirect console output to the terminal.12 This practical experience with QEMU provides the necessary context for understanding why bootloader tools and specific image formats are required.
Module Breakdown:
- Module 7: The x86_64 Boot Sequence. From BIOS/UEFI to the bootloader. Understanding the roles of the MBR, VBR, and the GRUB bootloader.
- Module 8: CPU Operating Modes and Segmentation. Real Mode, Protected Mode, and Long Mode. Setting up a Global Descriptor Table (GDT).
- Module 9: Introduction to Paging. The concept of virtual memory. The structure of 4-level page tables on x86_64. The role of the CR3 register.
Section 1.4: Bootstrapping the Kernel: From Bootloader to the main Function (Modules 10-12)
This section bridges the gap between the low-level, assembly-language world of the bootloader and the high-level Rust code of our kernel. Students will write the kernel's initial entry point in assembly, responsible for setting up a temporary stack and loading the GDT. This assembly code will then perform the crucial task of calling the first function written in Rust.
Once in Rust, we will parse the boot information passed by the bootloader, which typically includes a memory map detailing usable and reserved physical memory regions. This information is vital for initializing our memory manager in the next part. The final goal of this section is to create a bootable disk image that QEMU can execute, which successfully transitions into our Rust code and prints a "Hello, world!" message to the VGA buffer or serial console.10
Debugging is introduced as a core practice from the very beginning. We will leverage QEMU's built-in GDB server, using the -s and -S flags to halt the virtual machine at startup and allow a GDB client to connect.12 This enables single-stepping through the earliest assembly and Rust instructions, providing invaluable insight into the boot process and a powerful tool for troubleshooting.
Module Breakdown:
- Module 10: The Kernel Entry Point. Writing the initial assembly code to set up a stack and transition to Long Mode.
- Module 11: The First Rust Code. Calling a Rust function from assembly. Parsing bootloader information (e.g., Multiboot2 or Limine protocol).
- Module 12: Creating a Bootable Image and Debugging. Using the bootloader crate to create a bootable disk image. A hands-on lab on debugging the boot process with QEMU and GDB.12
Part II: Core Kernel Subsystems (Modules 13-30)
With the kernel successfully booting, this part focuses on constructing the essential subsystems that form the backbone of any modern operating system: memory management, scheduling, and interrupt handling. The design and implementation of these components will be guided by the principle of minimalism, tailored specifically for the single-purpose, high-performance demands of an AI/ML workload.
Section 2.1: Physical and Virtual Memory Management (Modules 13-18)
This section covers the implementation of a complete memory management subsystem from the ground up. We will begin by creating a physical frame allocator, which is responsible for tracking the usage of physical memory frames. Students will implement a bitmap-based allocator using the memory map provided by the bootloader.
Next, we will build the virtual memory manager, which involves creating and manipulating the x86_64 multi-level page tables. A critical focus of this implementation will be first-class support for HugePages (2MB and 1GB pages). For AI/ML workloads that operate on gigabytes of contiguous tensor data, the standard 4KB page size is a significant performance bottleneck. A large tensor can require hundreds of thousands of Translation Lookaside Buffer (TLB) entries, leading to frequent and costly TLB misses and page walks.14 By using 2MB HugePages, the number of required TLB entries can be reduced by a factor of 512, dramatically improving memory access performance. Consequently, our kernel's memory manager will be architected with a "fast path" for HugePage allocations, treating them as the default for large requests rather than an exotic optimization.
Finally, we will implement a kernel heap allocator, following the patterns established in the os.phil-opp.com series.10 This will enable dynamic memory allocation within the kernel itself, using the
alloc crate, which is essential for managing kernel data structures whose size is not known at compile time.
Module Breakdown:
- Module 13: Physical Memory Management. Implementing a frame allocator using a bitmap to manage the physical memory discovered at boot.
- Module 14: Paging and Virtual Address Spaces. Implementing the data structures for 4-level page tables. Creating a new, clean page table for the kernel.
- Module 15: Mapping Physical to Virtual Memory. Writing functions to map physical frames to virtual pages and translate virtual addresses to physical addresses.
- Module 16: HugePage Support. Extending the page table manager to support 2MB pages. Modifying the frame allocator to efficiently find contiguous blocks of physical memory.
- Module 17: Kernel Heap Allocator. Implementing the GlobalAlloc trait. Creating a heap region in virtual memory and backing it with physical frames.
- Module 18: Advanced Allocator Designs. Exploring and implementing more sophisticated heap allocators, such as a linked-list allocator or a fixed-size block allocator.
Section 2.2: The Scheduler and Concurrency (Modules 19-24)
This section focuses on designing and implementing a scheduler and concurrency primitives. The target workload—a single, long-running AI/ML container—does not require the complexity of a preemptive, multi-user, fairness-oriented scheduler found in general-purpose operating systems like Linux. Instead, we can build a much simpler and more efficient scheduler tailored to our needs.
We will begin by implementing a basic cooperative scheduler, where tasks voluntarily yield control of the CPU. This model is sufficient for managing the main application thread, potential background threads for I/O, and dedicated threads for submitting commands to the GPU. The modern async/await feature in Rust will be introduced as a powerful and elegant way to implement cooperative multitasking, allowing for the creation of asynchronous tasks and a simple executor to run them.10
We will then implement the foundational components for kernel-level concurrency, including context switching and kernel threads (kthreads). This will be complemented by the implementation of essential synchronization primitives, such as spinlocks and mutexes, using the spin crate and atomic CPU instructions to ensure safe access to shared data structures in a multi-core environment.9
Module Breakdown:
- Module 19: Introduction to Scheduling. Concepts of cooperative vs. preemptive multitasking. Designing a simple round-robin scheduler.
- Module 20: Context Switching. Saving and restoring CPU state (registers, instruction pointer, stack pointer). Writing the context switch function in assembly.
- Module 21: Kernel Threads. Implementing a kthread API to create and manage kernel-level threads of execution.
- Module 22: Synchronization Primitives. Implementing spinlocks and mutexes for mutual exclusion.
- Module 23: Cooperative Multitasking with async/await. Deep dive into Rust's Future trait and the state machine transformation.
- Module 24: Building a Kernel Executor. Implementing a basic executor to poll and run asynchronous tasks within the kernel.
Section 2.3: Interrupt and Exception Handling (Modules 25-27)
A robust interrupt and exception handling mechanism is critical for any stable operating system. This section covers the creation of an Interrupt Descriptor Table (IDT), which is the CPU's mechanism for dispatching interrupts and exceptions to their corresponding handler functions.
Students will write handlers for critical CPU exceptions, such as page faults and double faults. The page fault handler, in particular, is a cornerstone of the memory management system and will be essential for later implementing advanced features like demand paging and GPU Unified Memory. Handling double faults correctly is vital to prevent a system reset caused by a fatal triple fault.10
We will also implement support for hardware interrupts. This involves programming the legacy Programmable Interrupt Controller (PIC) or the modern Advanced Programmable Interrupt Controller (APIC) to receive signals from external hardware devices. As practical examples, we will configure a programmable interval timer (PIT or HPET) to generate periodic timer interrupts, which can be used as a basis for preemptive scheduling later, and handle interrupts from a PS/2 keyboard controller to receive user input.
Module Breakdown:
- Module 25: CPU Exceptions. Setting up the IDT. Implementing handlers for common exceptions like breakpoint and invalid opcode.
- Module 26: Page Faults and Double Faults. Writing a sophisticated page fault handler. Implementing an Interrupt Stack Table (IST) to handle double faults safely.10
- Module 27: Hardware Interrupts. Programming the PIC/APIC. Handling timer and keyboard interrupts.
Section 2.4: System Calls and A Rudimentary Filesystem (Modules 28-30)
This section establishes the boundary between the kernel and the user-space application it will host. We will implement a basic system call interface using the syscall and sysret instructions on x86_64. This allows the application, running in a lower privilege level (Ring 3), to request services from the kernel (running in Ring 0).
Drawing inspiration from the POSIX-compatibility layers of unikernels 4, our goal is not to replicate the entire Linux syscall API. Instead, we will implement only the minimal subset of syscalls strictly necessary for our target AI/ML runtime. This will include fundamental calls for memory management (
mmap), file operations (open, read, write), and basic process control.
To support these file operations, we will create a minimal, in-memory filesystem, similar to a Linux initramfs. This filesystem will be bundled into the kernel image at build time and will contain the AI/ML application binary, its dependencies, and any necessary configuration files. This approach avoids the complexity of implementing a full block device driver and on-disk filesystem, adhering to our minimalist design philosophy.
Module Breakdown:
- Module 28: The System Call Interface. Using the syscall/sysret instructions. Designing a system call table and dispatch mechanism.
- Module 29: Implementing Core System Calls. Hands-on implementation of a minimal set of POSIX-like syscalls (mmap, open, read, etc.).
- Module 30: Initial RAM Disk (initramfs). Creating a CPIO archive containing the application filesystem. Modifying the kernel to mount and read from this in-memory filesystem at boot.
Part III: The GPU Subsystem: A Deep Dive into NVIDIA CUDA (Modules 31-45)
This part is a cornerstone of the course, shifting focus from general OS principles to the highly specialized domain of GPU acceleration. We will dissect the NVIDIA CUDA software stack and build the kernel components necessary to communicate with an NVIDIA GPU. This endeavor involves creating a custom, minimal driver that interfaces with NVIDIA's existing OS-agnostic components, effectively replacing the Linux-specific portions of the official driver.
Section 3.1: Deconstructing the CUDA Stack: Runtime vs. Driver API (Modules 31-33)
The course begins this section with a thorough analysis of the CUDA software stack architecture. A key distinction is made between the high-level CUDA Runtime API and the lower-level Driver API.16 The Runtime API (e.g.,
cudaMalloc, cudaLaunchKernel) offers a simpler, single-source programming model, while the Driver API (e.g., cuMemAlloc, cuLaunchKernel) provides more explicit control over contexts and devices. We will trace the execution flow of a simple CUDA application, from the user-space API call down through the layers to the eventual interaction with the kernel-mode driver.
A critical component of this ecosystem is the nvcc compiler driver.17 We will study its two-phase compilation process: first, compiling CUDA C++ source code into PTX (Parallel Thread Execution), a virtual, assembly-like instruction set architecture. Second, compiling the PTX code into SASS (Shader Assembly), the architecture-specific binary code. To support forward compatibility, applications are often packaged as "fat binaries," which embed PTX code alongside binary code for several GPU architectures. The CUDA driver can then just-in-time (JIT) compile the PTX for a newer, unknown GPU, or directly load the appropriate binary if available.17 Understanding this compilation and loading process is crucial, as our kernel will ultimately be responsible for managing and loading this code onto the GPU. The canonical CUDA processing flow—copy data from host to device, CPU initiates kernel launch, GPU cores execute in parallel, copy results back—will serve as our guiding mental model.16
Module Breakdown:
- Module 31: CUDA Architecture Overview. The CUDA programming model: grids, blocks, threads. The hardware model: Streaming Multiprocessors (SMs) and CUDA cores.19
- Module 32: Runtime and Driver APIs. A comparative analysis of the two APIs. Tracing API calls using standard tools.
- Module 33: The CUDA Compilation Toolchain. Understanding nvcc, PTX, SASS, and the structure of fat binaries.17
Section 3.2: The NVIDIA Kernel-Mode Driver: Architecture and Open Modules (Modules 34-36)
Here, we delve into the architecture of the official NVIDIA Linux driver. It is not a single entity but a collection of kernel modules, primarily nvidia.ko (the core driver), nvidia-uvm.ko (for Unified Virtual Memory), nvidia-drm.ko (for display), and nvidia-modeset.ko.20 Our focus will be on the compute-related modules.
A significant development that enables this course is NVIDIA's release of open-source kernel modules.21 We will perform a detailed analysis of this source code. The most important architectural pattern to understand is the separation between the "OS-agnostic" component and the "kernel interface layer." The OS-agnostic part contains the bulk of the driver's logic and is distributed as a pre-compiled binary object (e.g.,
nv-kernel.o_binary). The kernel interface layer is a smaller, source-available component that acts as a shim, translating Linux kernel API calls and data structures into a form the OS-agnostic blob understands.21 This architecture provides a clear blueprint for our custom driver: our task is not to rewrite the entire driver, but to implement a
new kernel interface layer that connects NVIDIA's binary blob to our kernel's internal APIs (for memory allocation, interrupts, etc.), effectively replacing the Linux-specific shim. We will use the official build guides as a reference for understanding the compilation process and dependencies.22
Module Breakdown:
- Module 34: Anatomy of the NVIDIA Linux Driver. The roles of nvidia.ko, nvidia-uvm.ko, and other modules.
- Module 35: Source Code Analysis of Open Kernel Modules. A guided tour of the open-gpu-kernel-modules repository.21
- Module 36: The OS-Agnostic vs. Kernel Interface Layer. Understanding the shim architecture and its implications for porting the driver to a new kernel.
Section 3.3: The User-Kernel Interface: ioctl and GPU Command Submission (Modules 37-39)
This is the heart of the driver implementation. The primary communication mechanism between the user-space CUDA libraries and the kernel-mode driver in Linux is the ioctl system call, performed on special device files like /dev/nvidiactl and /dev/nvidia0. While the specific ioctl command codes and their associated data structures are not publicly documented by NVIDIA, their functionality can be inferred from the services the driver provides and by analyzing the open-source interface layer.
These ioctls are responsible for all fundamental GPU management tasks: creating and destroying GPU contexts, managing the GPU's virtual address space, allocating and freeing device memory, mapping memory for host access, and, most importantly, submitting kernels for execution.25 The driver's responsibilities include managing the JIT compilation of PTX, scheduling the execution of kernel grids on the device's SMs, and handling the complex state of the GPU.25 By studying the functions exported by the kernel interface layer and how they are called in response to file operations, we can piece together a functional understanding of this critical boundary.
Module Breakdown:
- Module 37: The ioctl System Call. A deep dive into how ioctl works in UNIX-like systems as a generic device control interface.
- Module 38: Inferring the NVIDIA ioctl Interface. Analyzing the open-source kernel interface layer to understand the structure of commands for context creation, memory allocation, and data transfer.
- Module 39: GPU Command Submission. Understanding the high-level process of how a kernel launch configuration (grid/block dimensions) is packaged and sent to the kernel driver for execution.
Section 3.4: Kernel-Level Memory Management: UVM, Page Faults, and DMA (Modules 40-42)
This section focuses on implementing the kernel-side logic for managing GPU memory, with a special emphasis on Unified Memory (UM). At its core, UM allows the GPU to access host memory and vice-versa within a single, unified virtual address space, simplifying programming by eliminating explicit cudaMemcpy calls.16
At a low level, this is often implemented via demand paging. When the GPU attempts to access a memory page that is not currently resident in its local VRAM, it triggers a page fault. This fault is trapped by the nvidia-uvm.ko module. The kernel driver must then handle this fault by pausing the GPU execution, allocating a page on the GPU, initiating a DMA (Direct Memory Access) transfer to copy the data from host RAM to VRAM, updating the GPU's page tables to map the virtual address to the new physical VRAM location, and finally resuming GPU execution.25
To support this, our kernel's page fault handler, developed in Part II, must be extended. It needs to be able to identify faults originating from the GPU (via the UVM driver), communicate with the NVIDIA driver components to orchestrate the page migration, and manage the underlying physical memory on both the host and device side. We will also map the concepts of different CUDA memory types—such as global, shared, and pinned memory—to the memory management primitives available in our custom kernel.26
Module Breakdown:
- Module 40: GPU Memory Models. Global, shared, constant, and texture memory. The concept of pinned (page-locked) host memory for faster DMA.27
- Module 41: Unified Memory and Demand Paging. The low-level mechanics of UM. How GPU-initiated page faults are handled by the kernel.
- Module 42: DMA and IOMMU. Understanding Direct Memory Access. Configuring the IOMMU (Input-Output Memory Management Unit) to allow the GPU to safely access host memory.
Section 3.5: Implementing a CUDA-Compatible Kernel Driver Stub (Modules 43-45)
This is the capstone project for the NVIDIA part of the course. Students will synthesize all the knowledge gained in the preceding modules to write a loadable module for our custom kernel. This "driver stub" will be the practical realization of the new "kernel interface layer."
The project will involve several key steps:
- Creating the necessary device nodes (/dev/nvidiactl, /dev/nvidia0) that the user-space driver expects to find.
- Implementing the open, close, and ioctl file operations for these device nodes.
- Handling a minimal but critical subset of ioctl commands. This will start with the commands required for device initialization and context creation.
- Implementing the ioctls for memory allocation (cuMemAlloc), which will involve calling our kernel's physical frame allocator and updating the GPU's address space.
- Implementing the ioctls for host-to-device data transfer (cuMemcpyHtoD), which will require setting up and managing a DMA transfer.
This project is a significant engineering challenge that directly applies the lessons learned from analyzing both the Nanos GPU driver port 5 and the architectural separation of NVIDIA's open-source modules.21 Successful completion demonstrates a deep, functional understanding of the user-kernel boundary for a complex hardware accelerator.
Module Breakdown:
- Module 43: Project Setup. Creating the module structure. Interfacing with our kernel's module loading and device management subsystems.
- Module 44: Implementing Initialization and Memory ioctls. Writing the handlers for device discovery, context creation, and memory allocation.
- Module 45: Implementing a Simple Data Transfer. Writing the handler for a host-to-device memory copy and verifying the data integrity on the device side using debugging tools.
Part IV: The GPU Subsystem: A Deep Dive into AMD ROCm (Modules 46-60)
This part of the curriculum mirrors the deep dive into NVIDIA CUDA but focuses on the AMD ROCm (Radeon Open Compute) platform. The key difference and pedagogical advantage of studying ROCm is its open-source nature. Unlike the CUDA stack, where parts of the driver are a "black box," the entire ROCm stack, from the user-space runtime down to the kernel driver, is open source.29 This allows for a more direct and less inferential study of the driver architecture, providing an invaluable counterpoint to the NVIDIA ecosystem.
Table: Comparative Analysis of CUDA and ROCm Kernel-Level Interfaces
To frame the upcoming modules, it is essential to establish a high-level comparison between the two GPU ecosystems. This strategic overview highlights the architectural differences that will directly impact our kernel development process.
Feature | NVIDIA CUDA Approach | AMD ROCm Approach | Implications for Custom Kernel |
---|---|---|---|
Driver Licensing | Proprietary core with open-source kernel interface layers.21 | Fully open source (FOSS) under permissive licenses (MIT, etc.).29 | ROCm allows for direct code porting and adaptation; CUDA requires interfacing with binary blobs. |
User/Kernel Interface | Undocumented ioctl commands on /dev/nvidia* nodes.25 | Documented ioctl interface via kfd_ioctl.h (for compute) and DRM/amdgpu interfaces.31 | ROCm development is API-driven and transparent; CUDA development requires more reverse engineering and inference. |
Memory Management | Unified Virtual Memory (UVM) is a custom, fault-based system implemented in nvidia-uvm.ko.16 | Leverages standard Linux Heterogeneous Memory Management (HMM) and HSA architecture.33 | Porting ROCm memory management involves implementing HMM-like features; CUDA requires handling specific UVM-related faults. |
Compilation Target | PTX (Parallel Thread Execution), a stable virtual ISA, provides forward compatibility.17 | GCN/CDNA ISA, a hardware-specific instruction set. HIP provides a source-level compatibility layer.29 | Kernel does not need to be aware of PTX, but for ROCm, it handles hardware-specific code objects directly. |
Driver Modularity | Multiple modules (nvidia.ko, nvidia-uvm.ko, etc.) with distinct roles.20 | A single, monolithic DRM driver (amdgpu.ko) handles graphics, display, and compute.29 | Interfacing with ROCm means interacting with a single, large driver module's compute-specific subset (KFD). |
Section 4.1: The ROCm Open-Source Stack: ROCk, ROCt, ROCr, and HIP (Modules 46-48)
This section provides a comprehensive overview of the entire ROCm software stack, emphasizing its modular, open-source composition, which is often compared to the UNIX philosophy of small, interoperable tools.29 We will trace the relationships between the key components:
- HIP (Heterogeneous-computing Interface for Portability): A C++ runtime API and kernel language that allows developers to write portable applications that can run on both AMD and NVIDIA GPUs. It acts as a crucial compatibility layer, often by source-to-source translation of CUDA code via tools like HIPIFY.29
- ROCr (ROC Runtime): The user-space runtime library that implements the HSA (Heterogeneous System Architecture) runtime API. It is responsible for discovering devices, managing memory, and launching compute kernels.29
- ROCt (ROC Thunk): A user-space library that acts as the "thunk" layer, translating ROCr API calls into the specific ioctl commands required by the kernel-mode driver.29
- ROCk (Kernel Driver): This is not a separate driver but rather the compute-specific functionality within the upstream Linux amdgpu kernel module.29
Understanding this layered architecture is key to seeing how a high-level call in a HIP application is progressively lowered until it becomes a command submitted to the kernel.
Module Breakdown:
- Module 46: ROCm Architecture and Philosophy. The HSA foundation. The roles of the various ROC libraries.29
- Module 47: The HIP Programming Model. Writing and compiling HIP applications. The role of the hipcc compiler wrapper.
- Module 48: Source-to-Source Translation. A practical lab using the HIPIFY tool to convert a simple CUDA application to HIP.29
Section 4.2: The amdgpu Kernel Driver: KFD, ioctls, and Command Queues (Modules 49-52)
This section is a deep dive into the amdgpu Linux kernel module, which is a comprehensive Direct Rendering Manager (DRM) driver responsible for all aspects of AMD GPU operation. For our purposes, we will focus on the KFD (Kernel Fusion Driver) interface, which is the component of amdgpu dedicated to handling compute workloads.
Unlike the opaque NVIDIA interface, the KFD interface is a public, stable API defined in the kfd_ioctl.h header file, which is part of the Linux kernel source.31 This section will be a line-by-line analysis of this header. Students will learn the specific
ioctl commands and their associated data structures for fundamental operations:
- Querying device properties and version information.
- Creating and destroying process address spaces.
- Allocating and managing memory (both VRAM and system memory accessible to the GPU).
- Creating and destroying hardware command queues.
- Submitting jobs (command packets) to a queue.
- Managing events and synchronization.
We will also examine the extensive list of module parameters that the amdgpu driver exposes via sysfs, which allow for fine-grained tuning and debugging of everything from memory sizes (vramlimit, vm_size) to scheduler timeouts (lockup_timeout) and power management (dpm).38
Module Breakdown:
- Module 49: The Linux DRM and KFD Subsystems. An overview of the Direct Rendering Manager framework and KFD's role within it.
- Module 50: The kfd_ioctl.h API (Part 1). A detailed study of the ioctls for device and process management.
- Module 51: The kfd_ioctl.h API (Part 2). A detailed study of the ioctls for memory and event management.
- Module 52: The kfd_ioctl.h API (Part 3). A detailed study of the ioctls for queue creation and job submission.
Section 4.3: Kernel-Level Heterogeneous Memory Management (HMM) (Modules 53-56)
Here, we will study AMD's approach to unified memory, which contrasts with NVIDIA's custom UVM solution. AMD's implementation is designed to integrate more closely with standard Linux kernel features, specifically HMM (Heterogeneous Memory Management). HMM allows a device driver to mirror a process's page tables, enabling the device (the GPU) to directly access the process's memory and trigger page faults on non-resident pages, which the kernel can then handle.34
We will analyze the different levels of unified memory support available in ROCm, from basic unified virtual addressing (where CPU and GPU share an address space but require explicit data copies) to true, demand-paged HMM where memory pages are automatically migrated between host and device on-fault.33 We will trace how a user-space call like
hipMallocManaged interacts with the underlying KFD memory management ioctls to allocate memory that is visible to both the CPU and GPU, and how the kernel's page fault handler is involved in orchestrating migrations.
Module Breakdown:
- Module 53: Introduction to Linux HMM. The core concepts of HMM and its role in CPU-device memory coherence.
- Module 54: ROCm Unified Memory. How hipMallocManaged works. The distinction between fine-grained and coarse-grained coherence.33
- Module 55: Demand Paging on AMD GPUs. Tracing a GPU-initiated page fault through the amdgpu driver and the kernel's MMU notifier subsystem.
- Module 56: Memory Allocation Strategies. Analyzing the amdgpu driver's internal memory manager and how it handles VRAM and GTT (Graphics Translation Table) memory.
Section 4.4: GPU Scheduling and Command Processor Interaction (Modules 57-58)
This section focuses on the mechanics of job submission. We will examine how the amdgpu driver's scheduler takes command packets from user space and submits them to the GPU's hardware command processors. A key concept here is the AQL (Architected Queuing Language) packet format, which provides a standardized way for the host to enqueue work for the GPU.36
User-space runtimes like ROCr create command queues in VRAM, which are structured as ring buffers. The application writes AQL packets into this queue, and then informs the kernel (via an ioctl) that new work is available by updating a write pointer. The kernel then instructs the GPU's command processor to fetch and execute these packets from the ring buffer. We will study the amdgpu_cs (Command Submission) ioctl and its role in this process, and the interrupt mechanisms used by the GPU to signal job completion back to the kernel.41
Module Breakdown:
- Module 57: Hardware Command Queues and Ring Buffers. The architecture of user-space accessible command queues.
- Module 58: The AQL Packet Format and Command Submission. Dissecting the structure of AQL packets. Tracing the amdgpu_cs ioctl from user space to hardware submission.
Section 4.5: Implementing a ROCm-Compatible Kernel Driver Stub (Modules 59-60)
As the capstone project for the AMD section, students will implement a driver module for our custom kernel that provides the KFD ioctl interface. Because the entire ROCm stack is open source and the KFD API is public, this project will be fundamentally different from the NVIDIA one. The task is less about reverse engineering and more about careful porting and adaptation.
Students will use the kfd_ioctl.h header as a formal specification. The project will involve implementing the kernel-side handlers for the core KFD ioctls, such as those for queue creation, memory allocation, and mapping. The internal logic of these handlers will call upon the respective subsystems of our custom kernel (e.g., the memory manager, scheduler). This project requires a deep understanding of both the KFD API and our kernel's internal architecture, and successful completion will result in a driver capable of initializing an AMD GPU and managing its basic resources from our custom OS.
Module Breakdown:
- Module 59: Project Setup. Defining the ioctl dispatch table in our kernel. Mapping KFD data structures to our kernel's native types.
- Module 60: Implementing Core KFD ioctls. Writing the handlers for KFD_IOC_CREATE_QUEUE and KFD_IOC_ALLOC_MEMORY_OF_GPU. Testing the implementation by writing a simple user-space program that calls these ioctls.
Part V: The Toolchain: Compilers and Build Systems (Modules 61-70)
Having built the core kernel and the GPU driver interfaces, this part shifts focus from the operating system itself to the essential tools required to build and optimize applications for it. A specialized kernel deserves a specialized toolchain. The goal of this section is to move beyond using off-the-shelf compilers and instead create a custom, domain-specific compiler toolchain using modern, modular infrastructure.
Section 5.1: Introduction to Compiler Infrastructure: LLVM and MLIR (Modules 61-63)
This section introduces the foundational technologies for our custom toolchain: LLVM and MLIR (Multi-Level Intermediate Representation). LLVM is a collection of modular and reusable compiler and toolchain technologies, famous for its well-defined Intermediate Representation (LLVM IR), which serves as a universal language for optimizers and backends.42
MLIR is a newer project within the LLVM ecosystem that provides a novel infrastructure for building compilers. Its key innovation is the concept of a "multi-level" IR, which can represent code at various levels of abstraction within a single framework.43 This is exceptionally well-suited for AI/ML, as it can model everything from a high-level TensorFlow or PyTorch computation graph down to low-level, hardware-specific instructions. MLIR is not a single IR, but a framework for creating new IRs, known as "dialects".42 This extensibility is what allows us to build a compiler that is perfectly tailored to our custom kernel. We will explore MLIR's design philosophy as a "compiler construction kit" that aims to reduce the cost of building domain-specific compilers and improve compilation for heterogeneous hardware.44
Module Breakdown:
- Module 61: The LLVM Project. Architecture of LLVM. The role of LLVM IR. The separation of frontend, middle-end (optimizer), and backend (code generator).
- Module 62: Introduction to MLIR. The motivation for a multi-level IR. Core concepts: Operations, Attributes, Types, and Regions.43
- Module 63: MLIR Dialects. Understanding the dialect as a mechanism for extensibility. A survey of existing dialects (e.g., func, affine, scf, gpu).
Section 5.2: Designing an MLIR Dialect for Kernel-Level AI Operations (Modules 64-66)
This is a hands-on section where students will apply the concepts of MLIR to design and implement their own custom dialect. This dialect will serve as the high-level interface for our compiler, defining custom operations that map directly to the unique primitives and system calls provided by our minimalist kernel.
For example, we might define operations like:
- aikernel.gpu.submit_commands: An operation that takes a buffer of GPU commands and lowers to the specific system call that submits work to our custom GPU driver.
- aikernel.mem.alloc_pinned: An operation to allocate page-locked host memory, which lowers to the corresponding memory management syscall in our kernel.
The implementation process will heavily utilize TableGen, a declarative language used by LLVM and MLIR to define records that can be processed by a backend to generate large amounts of C++ boilerplate code.45 By defining our operations, types, and attributes in
TableGen, we can automatically generate the C++ classes and verification logic, significantly accelerating the development of our dialect.
Module Breakdown:
- Module 64: Dialect Design Principles. Defining the semantics of our custom operations. Structuring the dialect for progressive lowering.
- Module 65: Implementing a Dialect with TableGen. Writing .td files to define operations, attributes, and interfaces.
- Module 66: Building and Testing the Custom Dialect. Integrating the new dialect into an MLIR-based tool. Writing test cases using mlir-opt and FileCheck.
Section 5.3: Lowering MLIR to LLVM IR and Target-Specific Code (Modules 67-68)
A dialect on its own is just a representation. To be useful, it must be transformable into something that can eventually be executed. This section focuses on writing the compiler passes that perform this "lowering." Lowering is the process of progressively transforming operations from a higher-level, more abstract dialect into operations in a lower-level dialect.
Students will write C++ passes that match on operations from our custom aikernel dialect and replace them with equivalent semantics expressed in standard MLIR dialects. For example, a high-level GPU submission operation might be lowered into a sequence of standard function calls that implement the system call ABI of our kernel. This process continues until the entire program is represented in dialects that have a direct lowering path to LLVM IR. Once in LLVM IR, we can leverage LLVM's mature ecosystem of optimizers and code generators to produce highly optimized x86_64 machine code for the host-side application. This demonstrates the full power of MLIR's multi-level pipeline: representing domain-specific concepts at a high level and systematically compiling them down to efficient, low-level code.44
Module Breakdown:
- Module 67: Writing MLIR Transformation Passes. The structure of a rewrite pass. Using the Declarative Rewrite Rule (DRR) framework.
- Module 68: Lowering to LLVM IR. The MLIR LLVM dialect as the bridge to the LLVM ecosystem. Generating the final executable object file.
Section 5.4: Integrating the Rust Compiler (Modules 69-70)
The final step in building our toolchain is to ensure it interoperates smoothly with Rust, our kernel's implementation language. While the compiler passes are written in C++, the end user (the application developer) will be writing Rust. This requires creating a seamless bridge between the two ecosystems.
We will use tools like bindgen or cxx to automatically generate safe Rust wrappers around the C++ APIs of our custom MLIR-based compiler. This will allow Rust code to invoke the compiler programmatically. Furthermore, we will leverage Cargo's powerful build script functionality (build.rs). The build script will be configured to run our custom compiler on specific source files (e.g., files defining the AI model's computation graph) during the application's build process, and then link the resulting object files into the final Rust binary. This deep integration makes the custom toolchain a natural part of the standard Rust development workflow.8
Module Breakdown:
- Module 69: Bridging C++ and Rust. Using bindgen to create safe FFI (Foreign Function Interface) bindings for the compiler's C++ API.
- Module 70: Cargo Build Scripts and Integration. Writing a build.rs script to invoke the custom compiler. Linking the generated object files into a Rust application.
Part VI: Containerization and AI/ML Workload Support (Modules 71-80)
This part brings all the preceding work together, focusing on the ultimate goal: running a containerized AI/ML application on top of our custom kernel. We will implement the necessary container primitives, create the runtime interface to support a real AI/ML framework, and explore unique optimizations that are only possible because we control the entire software stack.
Section 6.1: Implementing Container Primitives (Modules 71-73)
The term "container" in the context of a general-purpose OS like Linux refers to a combination of kernel features: namespaces for isolation (PID, network, mount, etc.) and cgroups for resource control. For our single-application, single-tenant unikernel, this level of complexity is unnecessary. Our implementation of "container primitives" will be far simpler and tailored to our specific use case.
The primary goal is to provide a filesystem boundary for the application. We will implement a chroot-like mechanism that confines the application's view of the filesystem to the initramfs we created earlier. This prevents the application from accessing any kernel-internal structures or devices that are not explicitly exposed. We will also implement a rudimentary form of resource limiting, akin to a simplified cgroup, to control the maximum amount of memory the application can allocate. This provides a basic level of containment and security without the overhead of full namespace virtualization.
Module Breakdown:
- Module 71: Filesystem Isolation. Implementing a chroot-style environment for the application process.
- Module 72: Resource Management. Designing and implementing a simple resource controller to limit the application's memory usage.
- Module 73: Loading and Running an ELF Binary. Writing the kernel code to parse an ELF executable from the initramfs, load its segments into virtual memory, set up its stack, and transfer control to its entry point in user mode.
Section 6.2: The AI/ML Runtime Interface (Modules 74-76)
This section focuses on creating the "glue" layer that allows a standard AI/ML framework, such as PyTorch or TensorFlow, to execute on our custom kernel. The goal is not to recompile the entire framework, but to intercept its calls to the underlying GPU driver libraries (like libcuda.so or librocm.so) and redirect them to our kernel's custom system call interface.
This can be achieved by creating a shared library that implements the public API of the CUDA or ROCm runtime and preloading it into the application's environment. When the application calls a function like cudaMalloc, our library's implementation will be invoked. Instead of communicating with the standard NVIDIA driver, it will execute our custom system call, which in turn invokes our kernel's memory manager. This shim layer effectively translates the standard AI framework's requests into the language of our minimalist kernel. This requires a deep understanding of how these frameworks manage GPU memory, track device state, and launch kernels, often through tools like PyTorch's memory snapshot visualizer.46
Module Breakdown:
- Module 74: The CUDA/ROCm User-Mode Driver API. A deep dive into the functions exported by libcuda.so and their purpose.
- Module 75: Building a Shim Library. Creating a shared library that implements a subset of the CUDA Driver API.
- Module 76: Intercepting and Redirecting API Calls. Using LD_PRELOAD to load our shim library. Translating API calls into our kernel's system calls.
Section 6.3: Optimizing for AI: Direct GPU Scheduling and Memory Pinning (Modules 77-78)
Now that we control the entire stack, from the application's API call down to the kernel's interaction with the hardware, we can implement powerful, cross-layer optimizations that would be difficult or impossible in a general-purpose OS.
One such optimization is direct GPU scheduling. In a traditional model, every kernel launch requires a system call, which involves a costly context switch into the kernel. We can design a more efficient mechanism where the user-space runtime and the kernel's GPU driver share a region of memory that acts as a command buffer or ring buffer. The runtime can write kernel launch commands directly into this shared memory region and then use a single, lightweight system call (or even an atomic memory operation) to notify the kernel that new work is available. The kernel driver, which is already polling or waiting on an interrupt, can then consume these commands directly, bypassing the overhead of the ioctl path for every launch.
This approach is inspired by the ideas of bringing computation closer to the kernel, as proposed in research on AI-native kernels 1, but applied pragmatically to the user-kernel communication boundary.
Module Breakdown:
- Module 77: Shared-Memory Command Submission. Designing a ring buffer in shared memory for user-kernel communication. Implementing the kernel and user-space logic to manage it.
- Module 78: Optimizing Memory Transfers. Implementing highly efficient memory pinning via our custom mmap syscall to prepare host memory for zero-copy DMA transfers.
Section 6.4: Packaging the Kernel: Minimalist "Distroless" Container Images (Modules 79-80)
The final step in preparing our application is to package it for deployment in a way that aligns with our minimalist philosophy. This means creating the smallest possible container image that contains only the application binary and its absolute essential runtime dependencies.
We will employ multi-stage Docker builds as a best practice.47 The first stage, the "build stage," will use a full development environment containing our custom MLIR-based toolchain and the Rust compiler. This stage will compile the application. The second stage, the "final stage," will start from a truly minimal base image, such as
scratch or a "distroless" image provided by Google.49 We will then use the
COPY --from=<build_stage> instruction to copy only the compiled application binary and our custom GPU runtime shim library from the build stage into the final image.52
This technique ensures that no compilers, build tools, package managers, shells, or other unnecessary utilities are present in the final production image. The result is a container image that is typically an order of magnitude smaller than a traditional one, which reduces storage costs, speeds up deployment, and significantly minimizes the potential attack surface.50
Module Breakdown:
- Module 79: Multi-Stage Docker Builds. The syntax and benefits of multi-stage builds. Creating a Dockerfile for our AI application.
- Module 80: Distroless and Scratch Images. Understanding distroless concepts. Creating a final container image from scratch containing only our application binary and its essential shared libraries.
Part VII: Testing, Profiling, and Deployment (Modules 81-100)
The final part of the course is dedicated to ensuring the kernel is robust, performant, and deployable in a real-world cloud environment. This involves building a comprehensive testing framework, integrating advanced, low-overhead profiling capabilities, systematically benchmarking and tuning performance, and mastering the process of deploying a custom OS to major cloud providers.
Section 7.1: A Kernel Testing Framework with QEMU and GDB (Modules 81-85)
A robust testing strategy is non-negotiable for kernel development. We will build a comprehensive testing suite that combines unit tests for individual modules and integration tests that run the entire kernel within the QEMU emulator.
We will leverage Rust's support for custom test frameworks, which allows us to define how tests are discovered and executed.10 This enables us to write test functions directly within our
no_std kernel code. For integration tests, cargo test will be configured to compile the kernel, package it into a bootable image, and run it under QEMU. QEMU can be configured with a special device that allows the guest OS to signal a success or failure code back to the host upon completion, which integrates seamlessly with the test runner.54
The QEMU testing environment will be our primary tool for development and debugging.55 We will make extensive use of GDB, connecting to the QEMU GDB server to debug panics, step through code, and inspect memory.12 This rigorous, automated testing framework is essential for maintaining code quality and catching regressions as the kernel's complexity grows.
Module Breakdown:
- Module 81: Unit Testing in no_std. Setting up a unit testing framework for kernel modules.
- Module 82: Integration Testing with QEMU. Configuring cargo test to run the kernel in QEMU.
- Module 83: Reporting Test Results. Using QEMU's isa-test-device or qemu_exit mechanism to communicate test outcomes to the host.
- Module 84: Advanced GDB Debugging. Using GDB scripts provided by the kernel source for advanced debugging tasks, such as inspecting page tables or scheduler state.12
- Module 85: Building a Continuous Integration (CI) Pipeline. Setting up a CI workflow (e.g., using GitHub Actions) to automatically build and test the kernel on every commit.
Section 7.2: Advanced Profiling with eBPF (Modules 86-90)
Traditional profiling tools like Nsight are invaluable for deep, offline analysis but often introduce significant performance overhead, making them unsuitable for continuous monitoring in production environments.56 To address this, we will integrate a modern, low-overhead profiling framework into our kernel, inspired by Linux's revolutionary eBPF (Extended Berkeley Packet Filter) technology.
eBPF allows safe, sandboxed programs to be attached to hooks within the kernel, enabling powerful and programmable tracing with minimal performance impact.56 We will implement a simplified version of this concept in our kernel. This will involve defining stable tracepoints at critical locations in our code, such as system call entry/exit points, scheduler decisions, and, most importantly, the entry and exit points of our GPU driver
ioctl handlers. We can then attach small, safe "probe" programs to these tracepoints to gather detailed performance data, such as the frequency and latency of kernel launches or memory transfers.
This approach provides a form of zero-instrumentation observability, allowing us to understand the behavior of an AI application's interaction with the GPU in real-time, without modifying the application's code.56 By building this capability into the kernel from day one, we are creating a system that is designed to be transparent and debuggable in production, a significant advantage over treating the OS as an opaque black box.
Module Breakdown:
- Module 86: Introduction to eBPF. The architecture and principles of eBPF on Linux.
- Module 87: Designing a Kernel Tracing Framework. Defining tracepoints and a simple, safe in-kernel virtual machine for running probes.
- Module 88: Probing the GPU Driver. Adding tracepoints to our custom CUDA and ROCm driver stubs to monitor memory allocations, data transfers, and kernel launches.58
- Module 89: Collecting and Exporting Telemetry. Writing the user-space tooling to load probes and collect data from the kernel via ring buffers.
- Module 90: Visualizing Performance Data. Exporting the collected data to standard observability tools like Prometheus and Grafana.59
Section 7.3: Benchmarking and Performance Tuning (Modules 91-95)
With a robust testing and profiling framework in place, this section focuses on systematic performance analysis and optimization. We will run a suite of standard AI/ML benchmarks, starting with micro-benchmarks like matrix multiplication (GEMM) and progressing to small but complete models like a simple transformer for inference.
Using the eBPF-inspired profiling tools we built, we will analyze the performance of these benchmarks on our kernel. We will identify bottlenecks by measuring the latency and frequency of critical operations. This data-driven approach will guide our tuning efforts. For example, we might discover that our scheduler is causing unnecessary delays, that our memory allocator is leading to fragmentation under load, or that the GPU command submission pipeline has higher-than-expected overhead. Students will then systematically tune these subsystems, measure the impact of their changes, and iterate until performance goals are met.
Module Breakdown:
- Module 91: Selecting and Implementing Benchmarks. Porting standard ML benchmarks (e.g., from mlperf) to run on our kernel.
- Module 92: Bottleneck Analysis. Using our custom profiler to identify performance hotspots in the kernel and driver.
- Module 93: Tuning the Scheduler. Experimenting with different scheduling policies and time slices.
- Module 94: Optimizing the Memory Manager. Tuning the HugePage allocation strategy and heap allocator performance.
- Module 95: End-to-End Performance Analysis. Comparing the final performance of our custom kernel against a standard Linux kernel running the same workload in a container.
Section 7.4: Deploying the Custom Kernel and Container Runtime to Cloud Infrastructure (Modules 96-100)
The final project of the course is to deploy the complete, custom-built system to a major cloud provider like Amazon Web Services (AWS) or Microsoft Azure. This demonstrates the end-to-end viability of the kernel and provides invaluable experience with real-world deployment challenges.
The process involves taking our bootable kernel image and initramfs and packaging them into a custom machine image (e.g., an Amazon Machine Image or AMI). This requires understanding the cloud provider's specific procedures for using user-provided kernels.60 A key step is configuring the bootloader (typically GRUB) within the image to load our custom kernel instead of the provider's default Linux kernel.62 We must also ensure that our
initramfs contains the necessary drivers for the cloud environment's virtualized hardware, especially for networking (e.g., the ENA driver on AWS) and storage (e.g., the NVMe driver for EBS volumes), to ensure the instance can boot and be accessed remotely.
Once the custom image is created, we will launch a GPU-enabled virtual instance from it. The final test is to deploy and run our containerized AI/ML application on this instance, verifying that it can successfully initialize the GPU via our custom driver and execute a workload. This capstone project validates the entire year's work, from the first line of boot code to a fully functional, specialized AI/ML operating system running in the cloud.
Module Breakdown:
- Module 96: Cloud Virtualization and Drivers. Understanding the virtualized hardware environment of cloud providers (networking, storage).
- Module 97: Building a Custom Amazon Machine Image (AMI). The process of bundling our kernel and initramfs and registering a new AMI.60
- Module 98: Configuring the GRUB Bootloader. Modifying the grub.cfg to chain-load our custom kernel and provide the correct command-line arguments.
- Module 99: Deploying to Microsoft Azure. A parallel module covering the process for creating and deploying a custom image on Azure.63
- Module 100: Final Project: End-to-End Cloud Deployment. Launching a GPU instance with our custom kernel, deploying the AI container, running a benchmark, and verifying the results.
Works cited
- Composable OS Kernel Architectures for Autonomous Intelligence - arXiv, accessed August 16, 2025, https://arxiv.org/html/2508.00604v1
- r/UniKernel - Reddit, accessed August 16, 2025, https://www.reddit.com/r/UniKernel/
- Introducing Unikraft - Lightweight Virtualization Using Unikernels - KubeSimplify blog, accessed August 16, 2025, https://blog.kubesimplify.com/introducing-unikraft-lightweight-virtualization-using-unikernels
- Compatibility - Unikraft, accessed August 16, 2025, https://unikraft.org/docs/concepts/compatibility
- GPU-accelerated Computing with Nanos Unikernels - NanoVMs, accessed August 16, 2025, https://nanovms.com/dev/tutorials/gpu-accelerated-computing-nanos-unikernels
- Composable OS Kernel Architectures for Autonomous ... - arXiv, accessed August 16, 2025, https://arxiv.org/pdf/2508.00604
- Composable OS Kernel Architectures for Autonomous Intelligence | AI Research Paper Details - AIModels.fyi, accessed August 16, 2025, https://aimodels.fyi/papers/arxiv/composable-os-kernel-architectures-autonomous-intelligence
- Rust Programming Language, accessed August 16, 2025, https://www.rust-lang.org/
- Rust - OSDev Wiki, accessed August 16, 2025, https://wiki.osdev.org/Rust
- Writing an OS in Rust, accessed August 16, 2025, https://os.phil-opp.com/
- Getting started - Rust Programming Language, accessed August 16, 2025, https://www.rust-lang.org/learn/get-started
- Booting a Custom Linux Kernel in QEMU and Debugging It With GDB, accessed August 16, 2025, https://nickdesaulniers.github.io/blog/2018/10/24/booting-a-custom-linux-kernel-in-qemu-and-debugging-it-with-gdb/
- How to build the Linux kernel and test changes locally in qemu - GitHub Gist, accessed August 16, 2025, https://gist.github.com/ncmiller/d61348b27cb17debd2a6c20966409e86
- Configuring HugePages for Enhanced Linux Server Performance - WafaTech Blogs, accessed August 16, 2025, https://wafatech.sa/blog/linux/linux-security/configuring-hugepages-for-enhanced-linux-server-performance/
- xmap: Transparent, Hugepage-Driven Heap Extension over Fast Storage Devices - EuroSys 2024, accessed August 16, 2025, https://2024.eurosys.org/posters/eurosys24posters-paper21.pdf
- CUDA - Wikipedia, accessed August 16, 2025, https://en.wikipedia.org/wiki/CUDA
- NVIDIA CUDA Compiler Driver Process | by ztex, Tony, Liu | Medium, accessed August 16, 2025, https://ztex.medium.com/nvidia-cuda-compiler-driver-process-cuda-kernel-deployment-from-code-to-gpu-execution-f94fdc41c8fe
- Nvidia CUDA in 100 Seconds - YouTube, accessed August 16, 2025, https://www.youtube.com/watch?v=pPStdjuYzSI
- Exploring CUDA Architecture: A Deep Dive - Metric Coders, accessed August 16, 2025, https://www.metriccoders.com/post/exploring-cuda-architecture-a-deep-dive
- How do you build out-of-tree modules, e.g., nvidia modules for customized kernel?, accessed August 16, 2025, https://discussion.fedoraproject.org/t/how-do-you-build-out-of-tree-modules-e-g-nvidia-modules-for-customized-kernel/77295
- NVIDIA/open-gpu-kernel-modules: NVIDIA Linux open ... - GitHub, accessed August 16, 2025, https://github.com/NVIDIA/open-gpu-kernel-modules
- NVIDIA Jetson Linux Developer Guide : Kernel Customization, accessed August 16, 2025, https://docs.nvidia.com/jetson/l4t/Tegra%20Linux%20Driver%20Package%20Development%20Guide/kernel_custom.html
- Compiling the Kernel (Kernel 5.10) | NVIDIA Docs - NVIDIA Developer, accessed August 16, 2025, https://developer.nvidia.com/docs/drive/drive-os/6.0.8/public/drive-os-linux-sdk/common/topics/sys_programming/compiling_the_kernel_linux.html
- Compiling the Kernel (Kernel 5.15) | NVIDIA Docs - NVIDIA Developer, accessed August 16, 2025, https://developer.nvidia.com/docs/drive/drive-os/6.0.7/public/drive-os-linux-sdk/common/topics/sys_programming/compiling-the-kernel-kernel-515.html
- What does the nVIDIA CUDA driver do exactly? - Stack Overflow, accessed August 16, 2025, https://stackoverflow.com/questions/9764591/what-does-the-nvidia-cuda-driver-do-exactly
- CUDA Series: Memory and Allocation | by Dmitrij Tichonov - Medium, accessed August 16, 2025, https://medium.com/@dmitrijtichonov/cuda-series-memory-and-allocation-fce29c965d37
- Understanding CUDA Memory Usage: A Practical Guide | by Hey Amit - Medium, accessed August 16, 2025, https://medium.com/@heyamit10/understanding-cuda-memory-usage-a-practical-guide-6dbb85d4da5a
- Memory management - Numba, accessed August 16, 2025, https://numba.pydata.org/numba-doc/dev/cuda/memory.html
- ROCm - Wikipedia, accessed August 16, 2025, https://en.wikipedia.org/wiki/ROCm
- AMD ROCm™ Software - GitHub Home, accessed August 16, 2025, https://github.com/ROCm/ROCm
- Support and limitations — ROCdbgapi 0.77.2 Documentation, accessed August 16, 2025, https://rocm.docs.amd.com/projects/ROCdbgapi/en/docs-6.4.2/reference/known-issues.html
- git.kernel.dk Git - include/uapi/linux/kfd_ioctl.h - kernel.dk, accessed August 16, 2025, https://git.kernel.dk/?p=linux-2.6-block.git;a=blobdiff;f=include/uapi/linux/kfd_ioctl.h;fp=include/uapi/linux/kfd_ioctl.h;h=32913d674d38bb0434bacc18c5d04a45dcb64360;hp=2da5c3ad71bd0f7448e97dc4c9f24eba0f8ed603;hb=4f98cf2baf9faee5b6f2f7889dad7c0f7686a787;hpb=ba3c87fffb79311f54464288c66421d19c2c1234
- Unified memory management — HIP 7.1.0 Documentation, accessed August 16, 2025, https://rocm.docs.amd.com/projects/HIP/en/docs-develop/how-to/hip_runtime_api/memory_management/unified_memory.html
- HMM is, I believe, a Linux feature. AMD added HMM support in ROCm 5.0 according - Hacker News, accessed August 16, 2025, https://news.ycombinator.com/item?id=37309442
- AMD ROCm™ installation - AMD GPUOpen, accessed August 16, 2025, https://gpuopen.com/learn/amd-lab-notes/amd-lab-notes-rocm-installation-readme/
- AMDKFD Kernel Driver - LWN.net, accessed August 16, 2025, https://lwn.net/Articles/619581/
- src/dev/hsa/kfd_ioctl.h · lab_4_solution · Simon Jakob Feldtkeller / ProSec Lab - NOC GitLab, accessed August 16, 2025, https://git.noc.ruhr-uni-bochum.de/feldts4p/prosec-lab/-/blob/lab_4_solution/src/dev/hsa/kfd_ioctl.h?ref_type=heads
- drm/amdgpu AMDgpu driver — The Linux Kernel documentation, accessed August 16, 2025, https://www.kernel.org/doc/html/v4.20/gpu/amdgpu.html
- drm/amdgpu AMDgpu driver — The Linux Kernel documentation, accessed August 16, 2025, https://www.kernel.org/doc/html/v5.9/gpu/amdgpu.html
- drm/amdgpu AMDgpu driver — The Linux Kernel 5.10.0-rc1+ documentation, accessed August 16, 2025, https://www.infradead.org/~mchehab/kernel_docs/gpu/amdgpu.html
- amdgpu/amdgpu_cs.c - chromiumos/third_party/libdrm - Git at Google, accessed August 16, 2025, https://chromium.googlesource.com/chromiumos/third_party/libdrm/+/refs/heads/master/amdgpu/amdgpu_cs.c
- Understanding LLVM v/s MLIR: A Comprehensive Comparison Overview | by Prince Jain, accessed August 16, 2025, https://medium.com/@princejain_77044/understanding-llvm-v-s-mlir-a-comprehensive-comparison-overview-9afc0214adc1
- MLIR (software) - Wikipedia, accessed August 16, 2025, https://en.wikipedia.org/wiki/MLIR_(software)
- MLIR, accessed August 16, 2025, https://mlir.llvm.org/
- MLIR: A Compiler Infrastructure for the End of Moore's Law | Hacker News, accessed August 16, 2025, https://news.ycombinator.com/item?id=22429107
- Understanding CUDA Memory Usage — PyTorch 2.7 documentation, accessed August 16, 2025, https://pytorch.org/docs/stable/torch_cuda_memory.html
- Multi-stage builds | Docker Docs, accessed August 16, 2025, https://docs.docker.com/get-started/docker-concepts/building-images/multi-stage-builds/
- Multi-stage | Docker Docs, accessed August 16, 2025, https://docs.docker.com/build/building/multi-stage/
- Base images - Docker Docs, accessed August 16, 2025, https://docs.docker.com/build/building/base-images/
- Distroless Docker Images: A Guide to Security, Size and Optimization - BellSoft, accessed August 16, 2025, https://bell-sw.com/blog/distroless-containers-for-security-and-size/
- Is Your Container Image Really Distroless? - Docker, accessed August 16, 2025, https://www.docker.com/blog/is-your-container-image-really-distroless/
- How to Build Slim and Fast Docker Images with Multi-Stage Builds - freeCodeCamp, accessed August 16, 2025, https://www.freecodecamp.org/news/build-slim-fast-docker-images-with-multi-stage-builds/
- Using official Python base images and packaging into distroless later on #1543 - GitHub, accessed August 16, 2025, https://github.com/GoogleContainerTools/distroless/issues/1543
- Building a testing platform for my kernel? : r/osdev - Reddit, accessed August 16, 2025, https://www.reddit.com/r/osdev/comments/t6cnt9/building_a_testing_platform_for_my_kernel/
- Testing in QEMU, accessed August 16, 2025, https://www.qemu.org/docs/master/devel/testing/main.html
- Snooping on your GPU: Using eBPF to Build Zero-instrumentation ..., accessed August 16, 2025, https://dev.to/ethgraham/snooping-on-your-gpu-using-ebpf-to-build-zero-instrumentation-cuda-monitoring-2hh1
- The Silent Revolution: eBPF Is Hacking Your GPU (For Good) | by kcl17 | Jul, 2025 - Medium, accessed August 16, 2025, https://medium.com/@kcl17/the-silent-revolution-ebpf-is-hacking-your-gpu-for-good-b986ff11e3a2
- Inside CUDA: Building eBPF uprobes for GPU Monitoring | by kcl17 | Jul, 2025 - Medium, accessed August 16, 2025, https://medium.com/@kcl17/inside-cuda-building-ebpf-uprobes-for-gpu-monitoring-449519b236ed
- Auto-instrumentation for GPU performance using eBPF - DevConf.CZ 2025 - YouTube, accessed August 16, 2025, https://www.youtube.com/watch?v=gGe9QvSpSf8
- User provided kernels - Amazon Linux 2 - AWS Documentation, accessed August 16, 2025, https://docs.aws.amazon.com/linux/al2/ug/UserProvidedKernels.html
- Use Your Own Kernel with Amazon EC2 | AWS News Blog, accessed August 16, 2025, https://aws.amazon.com/blogs/aws/use-your-own-kernel-with-amazon-ec2/
- How to rebuild Amazon Linux kernel in Amazon Linux - Artem Butusov Blog, accessed August 16, 2025, https://www.artembutusov.com/how-to-rebuild-amazon-linux-kernel-in-amazon-linux/
- How to Deploy Semantic Kernel to Azure in Minutes - Microsoft Developer Blogs, accessed August 16, 2025, https://devblogs.microsoft.com/semantic-kernel/how-to-deploy-semantic-kernel-to-azure-in-minutes/
References
- How to Increase Knowledge Productivity: Combine the Zettelkasten ..., accessed August 12, 2025, https://zettelkasten.de/posts/building-a-second-brain-and-zettelkasten/
- My Personal Knowledge Management System As a Software ..., accessed August 12, 2025, https://thewordyhabitat.com/my-personal-knowledge-management-system/
- Personal Knowledge Management (PKM) - Data Engineering Blog, accessed August 12, 2025, https://www.ssp.sh/brain/personal-knowledge-management-pkm/
- Combine Your Second Brain with Zettelkasten - Sudo Science, accessed August 12, 2025, https://sudoscience.blog/2024/12/27/combine-your-second-brain-with-zettelkasten/
- FOR COMPARISON with mdBook ... Obsidian - Sharpen your thinking, accessed August 12, 2025, https://obsidian.md/
- FOR COMPARISON with mdBook... Developers - Obsidian Help, accessed August 12, 2025, https://help.obsidian.md/developers
- FOR COMPARISON with mdBook ... Home - Developer Documentation - Obsidian, accessed August 12, 2025, https://docs.obsidian.md/Home
- Managing my personal knowledge base · tkainrad, accessed August 12, 2025, https://tkainrad.dev/posts/managing-my-personal-knowledge-base/
- Engineering - Notion, accessed August 12, 2025, https://www.notion.com/help/guides/category/engineering
- Junior to senior: An action plan for engineering career success ..., accessed August 12, 2025, https://github.com/readme/guides/engineering-career-success
- AswinBarath/AswinBarath: A quick bio about myself - GitHub, accessed August 12, 2025, https://github.com/AswinBarath/AswinBarath
- What Is Hugging Face? | Coursera, accessed August 12, 2025, https://www.coursera.org/articles/what-is-hugging-face
- Hugging Face : Revolutionizing AI Collaboration in the Machine Learning Community | by Yuvraj kakkar | Medium, accessed August 12, 2025, https://medium.com/@yuvrajkakkar1/hugging-face-revolutionizing-ai-collaboration-in-the-machine-learning-community-28d9c6e94ddb
- "Operator-Based Machine Intelligence: A Hilbert Space Framework ..., accessed August 12, 2025, https://www.reddit.com/r/singularity/comments/1mkwxzk/operatorbased_machine_intelligence_a_hilbert/
- [2505.23723] ML-Agent: Reinforcing LLM Agents for Autonomous Machine Learning Engineering - arXiv, accessed August 12, 2025, https://arxiv.org/abs/2505.23723
- Getting Started with Papers With Code – IT Exams Training ..., accessed August 12, 2025, https://www.pass4sure.com/blog/getting-started-with-papers-with-code/
- Wolfram Mathematica: Modern Technical Computing, accessed August 12, 2025, https://www.wolfram.com/mathematica/
- Mathematica & Wolfram Language Tutorial: Fast Intro for Math Students, accessed August 12, 2025, https://www.wolfram.com/language/fast-introduction-for-math-students/en/
- How to start a tech blog in 6 steps - Wix.com, accessed August 12, 2025, https://www.wix.com/blog/how-to-start-a-tech-blog
- How to Start a Tech Blog: Easy Guide for Beginners - WPZOOM, accessed August 12, 2025, https://www.wpzoom.com/blog/how-to-start-tech-blog/
- Networking for Engineers: 8 Strategies to Expand Your Professional ..., accessed August 12, 2025, https://staffing.trimech.com/networking-for-engineers-8-strategies-to-expand-your-professional-circle/
- Mastering Networking as a Software Developer: Strategies for Success : r/software_soloprenures - Reddit, accessed August 12, 2025, https://www.reddit.com/r/software_soloprenures/comments/1m363gv/mastering_networking_as_a_software_developer/
- The Software Developer's Guide to Networking - Simple Programmer, accessed August 12, 2025, https://simpleprogrammer.com/software-developers-networking/
- Participating in Open Source Communities - Linux Foundation, accessed August 12, 2025, https://www.linuxfoundation.org/resources/open-source-guides/participating-in-open-source-communities
- How To Grow Your Career With a Software Engineering Mentor - Springboard, accessed August 12, 2025, https://www.springboard.com/blog/software-engineering/software-engineer-mentor/
- Where to Find a Software Engineer Mentor (and How to Benefit From Them) | HackerNoon, accessed August 12, 2025, https://hackernoon.com/where-to-find-a-software-engineer-mentor-and-how-to-benefit-from-them
- Improve your open source development impact | TODO Group // Talk ..., accessed August 12, 2025, https://todogroup.org/resources/guides/improve-your-open-source-development-impact/
- Self-Directed Learning: A Four-Step Process | Centre for Teaching ..., accessed August 12, 2025, https://uwaterloo.ca/centre-for-teaching-excellence/catalogs/tip-sheets/self-directed-learning-four-step-process
- 25 New Technology Trends for 2025 - Simplilearn.com, accessed August 12, 2025, https://www.simplilearn.com/top-technology-trends-and-jobs-article
- Emerging Technology Trends - J.P. Morgan, accessed August 12, 2025, https://www.jpmorgan.com/content/dam/jpmorgan/documents/technology/jpmc-emerging-technology-trends-report.pdf
- 5 AI Trends Shaping Innovation and ROI in 2025 | Morgan Stanley, accessed August 12, 2025, https://www.morganstanley.com/insights/articles/ai-trends-reasoning-frontier-models-2025-tmt
- Llamaindex RAG Tutorial | IBM, accessed August 12, 2025, https://www.ibm.com/think/tutorials/llamaindex-rag
- Build Your First AI Application Using LlamaIndex! - DEV Community, accessed August 12, 2025, https://dev.to/pavanbelagatti/build-your-first-ai-application-using-llamaindex-1f9
- LlamaIndex - LlamaIndex, accessed August 12, 2025, https://docs.llamaindex.ai/
- Fine-Tuning LLMs: A Guide With Examples | DataCamp, accessed August 12, 2025, https://www.datacamp.com/tutorial/fine-tuning-large-language-models
- The Ultimate Guide to LLM Fine Tuning: Best Practices & Tools - Lakera AI, accessed August 12, 2025, https://www.lakera.ai/blog/llm-fine-tuning-guide
- Fine-tuning LLMs Guide | Unsloth Documentation, accessed August 12, 2025, https://docs.unsloth.ai/get-started/fine-tuning-llms-guide
- Building AI Agents Using LangChain and OpenAI APIs: A Step-by ..., accessed August 12, 2025, https://sen-abby.medium.com/building-ai-agents-using-langchain-47ba4012a8a1
- LangGraph - LangChain, accessed August 12, 2025, https://www.langchain.com/langgraph
- Build an Agent - ️ LangChain, accessed August 12, 2025, https://python.langchain.com/docs/tutorials/agents/
- With AI at the core, Heizen has a new model for software development at scale, accessed August 12, 2025, https://economictimes.indiatimes.com/small-biz/security-tech/technology/with-ai-at-the-core-heizen-has-a-new-model-for-software-development-at-scale/articleshow/123156453.cms
- 10 Best AI code generators in 2025 [Free & Paid] - Pieces App, accessed August 12, 2025, https://pieces.app/blog/9-best-ai-code-generation-tools
- Generative AI In Software Development Life Cycle (SDLC) - V2Soft, accessed August 12, 2025, https://www.v2soft.com/blogs/generative-ai-in-sdlc
- How an AI-enabled software product development life cycle will fuel innovation - McKinsey, accessed August 12, 2025, https://www.mckinsey.com/industries/technology-media-and-telecommunications/our-insights/how-an-ai-enabled-software-product-development-life-cycle-will-fuel-innovation
- Generative AI in SDLC: Can GenAI Be Utilized throughout the Software Development Life Cycle? - EPAM Startups & SMBs, accessed August 12, 2025, https://startups.epam.com/blog/generative-ai-in-sdlc
- Future of Data Engineering: Trends for 2025 - Closeloop Technologies, accessed August 12, 2025, https://closeloop.com/blog/data-engineering-key-trends-to-watch/
- Tutorial - MLflow, accessed August 12, 2025, https://www.mlflow.org/docs/2.7.1/tutorials-and-examples/tutorial.html
- 10 MLOps Projects Ideas for Beginners to Practice in 2025 - ProjectPro, accessed August 12, 2025, https://www.projectpro.io/article/mlops-projects-ideas/486
- Tutorials and Examples - MLflow, accessed August 12, 2025, https://mlflow.org/docs/latest/ml/tutorials-and-examples/
- Your First MLflow Model: Complete Tutorial, accessed August 12, 2025, https://mlflow.org/docs/latest/ml/getting-started/logging-first-model/
- End-to-End MLOps Pipeline: A Comprehensive Project ..., accessed August 12, 2025, https://www.geeksforgeeks.org/machine-learning/end-to-end-mlops-pipeline-a-comprehensive-project/
- Snowflake Data Mesh: The Ultimate Setup Guide (2025) - Atlan, accessed August 12, 2025, https://atlan.com/snowflake-data-mesh-how-to-guide/
- What Is Data Mesh? Complete Tutorial - Confluent Developer, accessed August 12, 2025, https://developer.confluent.io/courses/data-mesh/intro/
- Data Mesh Implementation: Your Blueprint for a Successful Launch - Ascend.io, accessed August 12, 2025, https://www.ascend.io/blog/data-mesh-implementation-your-blueprint-for-a-successful-launch
- Ten More Top Emerging Technologies In 2025 - Forrester, accessed August 12, 2025, https://www.forrester.com/report/ten-more-top-emerging-technologies-in-2025/RES183100
- What Is Quantum Computing? | IBM, accessed August 12, 2025, https://www.ibm.com/think/topics/quantum-computing
- Introduction to Qiskit | IBM Quantum Documentation, accessed August 12, 2025, https://quantum.cloud.ibm.com/docs/guides/
- Quantum computing - Wikipedia, accessed August 12, 2025, https://en.wikipedia.org/wiki/Quantum_computing
- Introduction to quantum computing, accessed August 12, 2025, https://thequantuminsider.com/introduction-to-quantum-computing/
- Introduction to Qiskit | IBM Quantum Documentation, accessed August 12, 2025, https://quantum.cloud.ibm.com/docs/guides
- How do people do Open Source Contributions ? : r/csharp - Reddit, accessed August 12, 2025, https://www.reddit.com/r/csharp/comments/1bxprbo/how_do_people_do_open_source_contributions/
- Good First Issue: Make your first open-source contribution, accessed August 12, 2025, https://goodfirstissue.dev/
- For Good First Issue | Make your next open-source contribution matter. - GitHub, accessed August 12, 2025, https://forgoodfirstissue.github.com/
- MunGell/awesome-for-beginners: A list of awesome beginners-friendly projects. - GitHub, accessed August 12, 2025, https://github.com/MunGell/awesome-for-beginners
- For Good First Issue: Introducing a new way to contribute - The GitHub Blog, accessed August 12, 2025, https://github.blog/open-source/social-impact/for-good-first-issue-introducing-a-new-way-to-contribute/
- How to Contribute to Open Source, accessed August 12, 2025, https://opensource.guide/how-to-contribute/
- Find Open Source Projects to Contribute: A Developer's Guide, accessed August 12, 2025, https://osssoftware.org/blog/find-open-source-projects-to-contribute-a-developers-guide/
- A Software Developer's Guide to Writing - DEV Community, accessed August 12, 2025, https://dev.to/tyaga001/a-software-developers-guide-to-writing-bgj
- Building an Online Presence In Tech 101 - SheCanCode, accessed August 12, 2025, https://shecancode.io/building-an-online-presence-in-tech-101/
- How to write a coding tutorial | Yost's Posts, accessed August 12, 2025, https://www.ryanjyost.com/how-to-write-a-coding-tutorial/
- Creating the Best Video Programming Tutorials | Vue Mastery, accessed August 12, 2025, https://www.vuemastery.com/blog/creating-the-best-video-programming-tutorials/
- A tutorial on creating coding tutorials - LogRocket Blog, accessed August 12, 2025, https://blog.logrocket.com/a-tutorial-on-creating-front-end-tutorials-2b13d8e94df9/
- How to Create a Technical Video Tutorial | Elastic Blog, accessed August 12, 2025, https://www.elastic.co/blog/elastic-contributor-program-how-to-create-a-video-tutorial
- How to Make Engaging Programming Videos - Real Python, accessed August 12, 2025, https://realpython.com/how-to-make-programming-videos/
- One-on-one mentorship with software engineers - CodePath, accessed August 12, 2025, https://www.codepath.org/career-services/mentorship
- Find a Software Engineering mentor - MentorCruise, accessed August 12, 2025, https://mentorcruise.com/filter/softwareengineering/
- Logseq vs. Obsidian: first impressions - Share & showcase, accessed August 13, 2025, https://forum.obsidian.md/t/logseq-vs-obsidian-first-impressions/56854
- 6 ways Logseq is the perfect Obsidian alternative - XDA Developers, accessed August 13, 2025, https://www.xda-developers.com/ways-logseq-is-the-perfect-obsidian-alternative/
- Electron vs Tauri - Coditation, accessed August 13, 2025, https://www.coditation.com/blog/electron-vs-tauri
- Framework Wars: Tauri vs Electron vs Flutter vs React Native - Moon Technolabs, accessed August 13, 2025, https://www.moontechnolabs.com/blog/tauri-vs-electron-vs-flutter-vs-react-native/
- Modular: A Fast, Scalable Gen AI Inference Platform, accessed August 13, 2025, https://www.modular.com/
- MAX: AI Compute Platform - Modular, accessed August 13, 2025, https://www.modular.com/max
- apache beam vs apache kafka: Which Tool is Better for Your Next Project? - ProjectPro, accessed August 13, 2025, https://www.projectpro.io/compare/apache-beam-vs-apache-kafka
- Apache Beam over Apache Kafka Stream processing - Codemia, accessed August 13, 2025, https://codemia.io/knowledge-hub/path/apache_beam_over_apache_kafka_stream_processing
- Apache Beam: Introduction to Batch and Stream Data Processing - Confluent, accessed August 13, 2025, https://www.confluent.io/learn/apache-beam/
- Quantum Programming Languages: A Beginner's Guide for 2025 - BlueQubit, accessed August 13, 2025, https://www.bluequbit.io/quantum-programming-languages
- What are the best-known quantum programming languages (e.g., Qiskit, Quipper, Cirq)?, accessed August 13, 2025, https://milvus.io/ai-quick-reference/what-are-the-bestknown-quantum-programming-languages-eg-qiskit-quipper-cirq
- Hello Many Worlds in Seven Quantum Languages - IonQ, accessed August 13, 2025, https://ionq.com/docs/hello-many-worlds-seven-quantum-languages
- Neuromorphic Hardware Guide, accessed August 13, 2025, https://open-neuromorphic.org/neuromorphic-computing/hardware/
- Embedded Neuromorphic Computing Systems - MCSoC-2025, accessed August 13, 2025, https://mcsoc-forum.org/site/index.php/embedded-neuromorphic-computing-systems/
- OpenBCI – Open-source EEG, accessed August 13, 2025, https://www.opensourceimaging.org/project/openbci/
- Community Page Projects - OpenBCI Documentation, accessed August 13, 2025, https://docs.openbci.com/Examples/CommunityPageProjects/
- Example Projects - OpenBCI Documentation, accessed August 13, 2025, https://docs.openbci.com/Examples/ExamplesLanding/
- EEG Headsets and Software for Education - EMOTIV, accessed August 13, 2025, https://www.emotiv.com/pages/education
- EEG Monitoring – EMOTIV, accessed August 13, 2025, https://www.emotiv.com/blogs/glossary/eeg-monitoring
- EEG Headset - Emotiv, accessed August 13, 2025, https://www.emotiv.com/blogs/glossary/eeg-headset
- Developing AR/VR/MR/XR Apps with WebXR, Unity & Unreal - Coursera, accessed August 13, 2025, https://www.coursera.org/learn/develop-augmented-virtual-mixed-extended-reality-applications-webxr-unity-unreal
- WebXR Academy, accessed August 13, 2025, https://webxracademy.com/
- Top VR Education Companies in 2025 - Axon Park, accessed August 13, 2025, https://www.axonpark.com/top-vr-education-companies-in-2025/
- The Future of VR in Education: Immersive Learning Experiences, accessed August 13, 2025, https://www.immersivelearning.news/2025/06/19/the-future-of-vr-in-education-immersive-learning-experiences/
- Streamlit vs FastAPI: Choosing the Right Tool for Deploying Your Machine Learning Model | by Pelumi Ogunlusi | Jul, 2025 | Medium, accessed August 13, 2025, https://medium.com/@samuelogunlusi07/streamlit-vs-fastapi-choosing-the-right-tool-for-deploying-your-machine-learning-model-1d16d427e130
- Compare Streamlit vs. Tauri in 2025, accessed August 13, 2025, https://slashdot.org/software/comparison/Streamlit-vs-Tauri/
- Monica: Personal CRM done right, accessed August 13, 2025, https://www.monicahq.com/
- monicahq/monica: Personal CRM. Remember everything about your friends, family and business relationships. - GitHub, accessed August 13, 2025, https://github.com/monicahq/monica
- rust-lang/mdBook: Create book from markdown files. Like Gitbook but implemented in Rust, accessed August 13, 2025, https://github.com/rust-lang/mdBook
- Freelancer API for Developers, accessed August 13, 2025, https://developers.freelancer.com/
- API Developer Freelance Jobs: Work Remote & Earn Online - Upwork, accessed August 13, 2025, https://www.upwork.com/freelance-jobs/api-development/
- How to Start a Podcast: Step-by-Step Guide & Free Checklist - Riverside, accessed August 13, 2025, https://riverside.com/blog/how-to-start-a-podcast
Resource Management Methodologies In Personal Knowledge Engineering
Building a Second Brain (BASB) has sparked renewed interest in personal knowledge management, but it represents just one approach in a rich tradition of information organization systems spanning millennia. The comprehensive survey given below identifies 133 methodologies similar to Tiago Forte's BASB that excel at organizing information for project-based work, drawn from technological, engineering, and scientific domains.
Understanding Building a Second Brain as The Baseline Methodology
Tiago Forte's Building a Second Brain (2022) is based on a very appealling notion, some would say compelling insight, that our brains are fundamentally for having ideas, not really for storing them.
BASB represented a major innovation by synthesizing productivity methodologies with digital note-taking in a way that prioritized actionability over comprehensive capture. Unlike previous systems that emphasized exhaustive documentation (like GTD) or pure linking (like Zettelkasten), BASB introduced the concept of "intermediate packets" that could be immediately useful across projects. This approach solved the common problem of knowledge management systems becoming graveyards of unused information by ensuring every piece of captured information had a clear path to creative output.
Building a Second Brain (2022) operates on the CODE method (Capture, Organize, Distill, Express) combined with the PARA organizational system (Projects, Areas, Resources, Archive). BASB's effectiveness stems from its actionability-focused organization, progressive summarization techniques, and emphasis on creative output rather than passive consumption. The system specifically supports project-based work through "intermediate packets" - discrete, reusable units of work that enable incremental progress and cross-project knowledge transfer.
Modern Digital Personal Knowledge Management Systems
-
Foam: VSCode-powered personal knowledge management and sharing system in the form of a VSCode extension for developers, the Foam system is inspired by Roam Research reduces context-switching for devs who are already using Visual Studio Code and GitHub, making it easier to build personal MarkDown wikis [and things like mdBooks] alongside code, enhancing efficiency in tech-heavy careers.
-
Roam Research: Pioneering block-level references and daily notes, the Roam writing tool enables fluid, non-hierarchical knowledge structures that mirror the interconnected nature of software development workflows. For engineers, its transclusion feature turns scattered thoughts into reusable components, much like modular code, accelerating problem-solving in fast-paced tech teams.
-
Logseq: As a local-first, privacy-focused tool with Git integration, Logseq appeals to developers by applying version control principles to personal notes. Its outliner format and query capabilities make it outstanding for managing technical documentation, ensuring knowledge remains accessible and evolvable in startup settings without cloud dependencies.
-
RemNote: Integrating spaced repetition into note-taking, RemNote automates flashcard creation from technical notes, perfect for mastering programming languages or frameworks. This fusion of learning and documentation makes it worthy of emulation for career growth, as it builds long-term retention of complex tech concepts essential for interviews and innovation.
-
Notion Databases for PKM: Transforming notes into relational databases, Notion allows dynamic views and filters for organizing project roadmaps and tech stacks. Its versatility in creating custom workflows without coding empowers startup founders to centralize knowledge, reducing context-switching and boosting team productivity.
-
Digital GTD Implementations: Using tools like Todoist with Notion, this adapts Getting Things Done for digital age, adding automation to task capture. For tech careers, it stands out by linking actions to knowledge artifacts, ensuring ideas turn into executable projects without falling through cracks.
-
GTD + Zettelkasten Hybrids: Combining task management with knowledge linking, hybrids like Obsidian with plugins bridge execution and ideation. This is exemplary for engineers, as it captures expertise during projects, creating reusable assets that compound over a career in evolving tech landscapes.
-
OmniFocus Advanced Perspectives: Customizable task views surface context-specific actions, revolutionizing how developers manage multiple roles. Its query system emulates database thinking, making it invaluable for startups where quick reconfiguration of focus areas drives agility and success.
-
Andy Matuschak's Evergreen Notes: Emphasizing atomic, declarative notes written for future self, this methodology builds timeless knowledge bases. In tech, it's outstanding for documenting evolving systems, ensuring notes remain valuable across projects and career stages.
-
Digital Gardens: Treating knowledge as cultivated spaces with maturity stages, tools like Obsidian publish thinking in progress. For startups, this normalizes public learning, fostering community feedback that accelerates product development and personal growth.
-
Obsidian Zettelkasten: This digital adaptation of Luhmann's slip-box system excels in bidirectional linking and graph visualization, making it ideal for tech professionals to uncover hidden connections in code notes and project ideas. Its plugin ecosystem allows seamless integration with Git for version-controlled knowledge bases, fostering innovation in startup environments where rapid idea iteration is crucial.
-
Dendron: Hierarchical notes with schema validation bring type safety to knowledge organization. This prevents drift in large tech knowledge bases, making it essential for maintaining structured documentation in scaling startups.
-
TiddlyWiki: Single-file wikis offer portable, serverless knowledge bases. For mobile tech workers, its self-contained nature ensures access anywhere, supporting uninterrupted ideation and reference in dynamic startup environments.
-
Zotero: Beyond citations, it scrapes web content and annotates PDFs for research. Tech professionals emulate it for curating API docs and papers, integrating literature review into development workflows.
-
Mendeley: Adding social networking to references, it discovers work through connections. In tech communities, this social filtering uncovers relevant tools and papers, expanding professional networks and knowledge.
-
EndNote: Automated formatting across styles saves time on technical writing. For engineers documenting inventions, it streamlines publication, freeing focus for innovation.
-
ReadCube Papers: Visual PDF management with enhanced reading features centralizes research consumption. This innovation suits tech careers by prioritizing PDF-based learning, common in specs and whitepapers.
-
Citavi: Combining references with planning, it supports full research workflows. Worthy for tech project managers integrating sources with tasks, ensuring evidence-based decisions.
-
JabRef: Open-source BibTeX management for LaTeX users. Its deep integration aids engineers in academic-tech crossover, maintaining open bibliographic data.
-
RefWorks: Cloud-based for accessible collaboration. Pioneering web access, it enables team knowledge sharing in distributed startups.
-
Darwin's Transmutation Notebooks: Systematic cross-referencing of observations built evolutionary theory. Emulate for tech by indexing experiments across projects, synthesizing long-term insights.
-
Einstein's Thought Experiment Documentation: Recording imaginative scenarios alongside math. For developers, this documents creative problem-solving, preserving paths to breakthroughs.
-
Einstein's Zurich Notebook: Documenting failures and successes. In startups, this complete record aids debugging and iteration, learning from all attempts.
-
Leonardo da Vinci's Multi-Topic Integration: Visual-textual fusion in notebooks. Tech emulation uses diagrams as primary carriers, enhancing system design communication.
-
Marie Curie's Laboratory Documentation: Meticulous recording including negatives. For engineers, this comprehensive history enables pattern detection in trials.
-
Edison's Invention Factory System: Witnessed notebooks for IP protection. Startups benefit from searchable solution archives, securing and reusing inventions.
-
Newton's Mathematical Notebooks: Developing notation with discoveries. Worthy for creating personal symbols to tackle complex tech problems.
-
Galileo's Observation Logs: Quantitative measurements with drawings. Establishes precision in tech observations, foundational for data-driven decisions.
-
Kepler's Calculation Notebooks: Preserving iterative refinements. Documents discovery processes, essential for refining algorithms in tech.
-
Faraday's Laboratory Notebooks: Continuous numbering for cross-referencing. Creates searchable archives, ideal for long-term tech research.
-
Pasteur's Laboratory Protocols: Standardized controls. Ensures reproducibility, critical for software testing and validation.
-
Mendel's Statistical Record-Keeping: Quantitative biology analysis. Applies stats to tech metrics, founding data-informed practices.
-
Linnaeus's Species Classification System: Hierarchical taxonomies. Organizes tech stacks hierarchically, accommodating new tools.
-
Humboldt's Integrated Field Studies: Multidisciplinary connections. Pioneers holistic views, useful for interdisciplinary tech projects.
-
Hooke's Micrographia Methods: Illustration as scientific tool. Revolutionizes visual documentation in UI/UX design.
-
Brahe's Astronomical Data Tables: Unprecedented accuracy. Emphasizes precision in tech data logging.
-
Vesalius's Anatomical Documentation: Observation over authority. Corrects assumptions in system architectures.
-
Grinnell System: Tiered field documentation. Separates observations from analysis, structuring tech logs.
-
Standard Laboratory Notebook Practices: Bound, witnessed pages for IP. Legally defensible, crucial for startup patents.
-
Electronic Laboratory Notebooks (ELNs): Digital compliance with instrument integration. Speeds development, reducing errors in tech labs.
-
CAD File Management Systems: Version control for designs. Enables parallel engineering, avoiding bottlenecks.
-
Product Data Management (PDM) Systems: Centralizes product info. Integrates departments, reducing errors in startups.
-
Six Sigma DMAIC Documentation: Statistical validation. Data-driven improvements, quantifiable for tech processes.
-
Failure Mode and Effects Analysis (FMEA): Proactive failure documentation. Prevents catastrophes in software engineering.
-
Systems Engineering Management Plans (SEMP): Technical performance tracking. Manages complex tech developments.
-
Requirements Traceability Matrices (RTM): Linking needs to implementation. Ensures complete coverage in projects.
-
Quality Management System (QMS) Documentation: ISO compliance. Standardizes quality in tech firms.
-
Document Control Systems: Revision management. Prevents errors from outdated specs.
-
Change Management Documentation: Impact analysis. Avoids cascading failures in code changes.
-
Technical Data Packages (TDP): Complete manufacturing definitions. Enables outsourcing in tech production.
-
Lean Documentation Principles: Minimize non-value docs. Reduces burden while maintaining quality.
-
Agile Engineering Documentation: Iterative refinement. Matches docs to evolving products.
-
Model-Based Systems Engineering (MBSE): Models as truth sources. Eliminates inconsistencies.
-
Digital Thread Documentation: Lifecycle connectivity. Enables predictive maintenance.
-
Configuration Management Databases (CMDB): Track interdependencies. Predicts change impacts.
-
Root Cause Analysis (RCA) Documentation: Evidence-based investigations. Prevents recurrence in bugs.
-
Jupyter Notebooks: Executable code with narratives. Democratizes data science, accessible for tech learning.
-
Observable Notebooks: Reactive computational docs. Creates interactive explanations for complex algorithms.
-
Marimo Notebooks: Deterministic execution. Ensures reproducibility in ML experiments.
-
Google Colab: Free GPU access. Democratizes deep learning for startup prototyping.
-
Pluto.jl: Reactive Julia notebooks. Guarantees reproducibility in scientific computing.
-
Literate Programming: Documentation primary, code extracted. Enhances understanding in open-source contributions.
-
Documentation-Driven Development (DDD): Docs before code. Catches API issues early.
-
README-Driven Development: User docs first. Ensures usability in tech products.
-
Software Architecture Decision Records (ADRs): Capture decisions with context. Preserves memory for team handovers.
-
Design Docs: Standardize communication. Creates searchable decision archives.
-
Request for Comments (RFC) Process: Collaborative design. Opens review, catching problems early.
-
DevOps Runbooks: Operational procedures. Codifies knowledge for reliable responses.
-
Post-Mortem Documentation: Blameless failure analysis. Improves systems psychologically safely.
-
Site Reliability Engineering (SRE) Documentation: Quantified objectives. Makes reliability engineering concern.
-
Code Review Comments as Documentation: Preserve discussions. Archives engineering rationale.
-
Pull Request Templates: Standardize changes. Improves knowledge transfer.
-
Commit Message Conventions: Machine-readable history. Automates changelogs.
-
Learning-in-Public Methodologies: Share journeys. Accelerates skills through feedback.
-
Technical Blogging Platforms: Community engagement. Motivates documentation.
-
Today I Learned (TIL) Repositories: Micro-insights. Accumulates knowledge effortlessly.
-
Static Site Generators for Documentation: Markdown to sites. Focuses on content.
-
API Documentation Generators: From annotations. Syncs docs with code.
-
Interactive Documentation: Embedded playgrounds. Improves learning outcomes.
-
Knowledge Bases as Code: Version control for docs. Ensures quality through pipelines.
-
Tana: Supertags and AI for system-based organization. Powers advanced PKM with reusable metadata for tech workflows.
-
Reflect Notes: Networked thought with tasks. Balances traditional and PKM, integrating daily notes seamlessly.
-
Heptabase: Visual canvases for ideas. Suits visual thinkers in tech, blending PKM with project management.
-
AFFiNE: Universal editor for notes and tasks. Affordable, feature-rich for boosting productivity in startups.
-
Capacities: Notes, projects, visualizations. Meets knowledge workers' needs with seamless integrations.
-
Evernote: Advanced search for notes. Classic reliability for capturing ideas in busy tech careers.
-
Microsoft OneNote: Microsoft ecosystem integration. Seamless for enterprise tech stacks.
-
Craft: Sleek collaborative design. Ideal for creatives in tech product teams.
-
Zettlr: Citation management for research. Supports academic-tech writing.
-
Milanote: Visual organization. Brainstorming boards for startup ideation.
-
Antinet Zettelkasten: Analog-first revival. Forces deep processing, countering digital overload.
-
Smart Notes Method: Thinking tool focus. Drives output from notes, essential for content creation in tech.
-
Memex Methodology: Associative trails. Inspires modern linked bases for knowledge retrieval.
-
Linking Your Thinking: Emergent maps. Organic structure for flexible tech knowledge.
-
Garden-Stream Dichotomy: Separate capture and curation. Reduces guilt, streamlines workflows.
-
Resonance Calendar: Emotion-driven tracking. Compiles insights for reflective career growth.
-
Quadrant Note-Taking: Structured analysis. Forces context, reducing storage issues.
-
Notion + Zapier + Google Drive: Automated knowledge hub. Centralizes startup ops, enhancing efficiency.
-
Obsidian + Git Integration: Version-controlled notes. Applies dev practices to PKM, ensuring durability.
-
Logseq + Whiteboards: Connected outlining with visuals. Powers brainstorming and knowledge linking for innovative tech careers.
Note Capturing Systems In Personal Knowledge Management (PKM)
The personal hyperlinked notebooks or wiki that are based on atomic notetaking as exemplified by Zettelkasten (Zkn) Method have revolutionized personal knowledge management (PKM) through ATOMIC thought notes, the "folgezettel" principle of note connectivity, and a variety of emergent open source development communities built around Zkn and all kinds of advanced Zkn PKM tools/plugins/add-ins, eg Zkn using the pomodoro technique.
Of course, Zkn is certainly not the only the pattern in personal knowledgement system worth exploring. The principles underlying modern Zettelkasten implementations have deep historical roots spanning millennia of human knowledge organization and the innovations like Zkn in the realm of PKM will certainly continue and maybe proliferate even more now.
Electronic note capturing approaches certainly matter, perhaps more than ever, in the world of AI, particularly for Human In The Loop (HITL) AI because data annotation adds important context, particularly as the human changes the approach of the AI ... so the development of note-capturing technologies become more important than ever, even as note-formating, grammar-checking and stylistic-prettification are things that be delegated to AI ... or "Ship it ...we'll fix it in post!"
As one might expect, there is a significant amount of current interest in the latest, greatest AI-assisted PKM tools, but the interest in PKM is not new -- it has been a really big deal for humans for at least 2500 years, ever since humans started using the printed word or moving beyond the limitations of storytelling and human memory which had limited the sustained development of knowledge in earlier philosophical traditions. The following comprehensive survey identifies 100 distinct systems across history and domains that share these core principles of idea generation, concept linking, and networked knowledge building. These examples span from ancient memory techniques to cutting-edge AI-powered knowledge graphs, demonstrating the universal human drive to organize, connect, and build upon ideas.
Historical foundations: Pre-digital knowledge systems
Ancient and classical systems
1. Ancient Greek Hypomnema (5th Century BCE) - Personal memory aids combining notes, reminders, and philosophical commentary for self-improvement and knowledge rediscovery, presaging modern reflective note-taking practices. Unlike the purely oral tradition that preceded it, the hypomnema represented the first systematic approach to externalizing memory for personal intellectual development rather than public performance. This innovation allowed Greeks to build cumulative personal knowledge over time, moving beyond the limitations of human memory that constrained earlier philosophical traditions.
2. Roman Commentarii - Systematic recording systems including family memorials, speech abstracts, and daily observations, creating interconnected knowledge repositories across multiple information types. While Greeks focused on philosophical reflection, the Roman system innovated by integrating diverse information types—legal, administrative, and personal—into unified knowledge collections. This represented the first comprehensive approach to managing different knowledge domains within a single organizational framework, surpassing the single-purpose records common in earlier civilizations.
3. Chinese Bamboo Strip Systems (Shang-Han Dynasty) - Individual bamboo strips containing single concepts, bound with cords and rearrangeable into different organizational structures—the ancient predecessor to atomic notes. Before bamboo strips, knowledge was carved on bones or bronze vessels in fixed, immutable arrangements that couldn't be reorganized. The modular bamboo system revolutionized Chinese knowledge management by allowing dynamic reconfiguration of information, enabling scholars to experiment with different conceptual arrangements and discover new relationships between ideas.
4. Chinese Biji Notebooks (3rd Century AD) - Non-linear collections of anecdotes, quotations, and observations organized organically, mixing diverse content types in flexible arrangements. Unlike the rigid, chronological court records and official histories that dominated Chinese writing, biji introduced personal, associative organization that followed the author's thoughts rather than institutional requirements. This innovation allowed for serendipitous connections between disparate topics, creating a more naturalistic knowledge accumulation method that reflected actual thinking processes.
5. Japanese Zuihitsu/Pillow Books (10th Century) - Personal knowledge accumulation combining observations, essays, and lists, representing lifelong intellectual development through writing. While Chinese literary traditions emphasized formal structure and classical references, zuihitsu pioneered stream-of-consciousness knowledge capture that valued personal experience equally with scholarly learning. This democratization of knowledge recording broke from the exclusively academic writing of the time, establishing that everyday observations could constitute valuable knowledge worth preserving.
Medieval knowledge technologies
6. Medieval Memory Palaces/Method of Loci - Spatial mnemonic systems associating concepts with imagined locations, creating navigable knowledge architectures in mental space. While ancient rhetoricians used simple linear sequences for memorizing speeches, medieval scholars expanded this into complex architectural spaces housing entire libraries of knowledge. This innovation transformed memory from sequential recall into spatial navigation, allowing scholars to store and retrieve vastly more information than simple rote memorization permitted, essentially creating the first virtual knowledge management system.
7. Medieval Manuscript Marginalia Systems - Sophisticated annotation networks using symbols and cross-references, connecting main texts with commentary through "signes-de-renvoi" (return signs). Previous manuscript traditions simply copied texts verbatim, but medieval scribes innovated by creating parallel knowledge layers that could dialogue with primary sources. This multi-dimensional approach to text allowed centuries of accumulated wisdom to coexist on single pages, transforming static texts into dynamic knowledge conversations across time.
8. Medieval Florilegia - Thematic compilations of excerpts from religious and classical texts, literally "gathering flowers" to preserve and organize knowledge across sources. Unlike complete manuscript copying which was expensive and time-consuming, florilegia innovated by extracting and reorganizing essential passages around themes rather than sources. This represented the first systematic approach to knowledge synthesis, allowing scholars to create new works by recombining existing wisdom in novel arrangements.
9. Ramon Lull's Ars Magna (1275-1305) - Mechanical system using rotating wheels with letters representing philosophical concepts, enabling systematic idea combination for intellectual discovery. While previous philosophical methods relied on linear argumentation, Lull's mechanical approach introduced combinatorial knowledge generation that could systematically explore all possible concept relationships. This was arguably the first algorithmic approach to knowledge discovery, prefiguring modern computational methods by seven centuries and moving beyond the limitations of sequential human reasoning.
10. Medieval Scholastic Apparatus - Layered citation and cross-referencing systems connecting biblical texts with interpretive traditions through glosses and commentaries. Earlier biblical study treated scripture as isolated text, but the scholastic apparatus innovated by creating comprehensive reference networks linking verses to centuries of interpretation. This systematic approach to textual analysis established the foundation for modern academic citation practices, transforming religious texts into interconnected knowledge webs.
Renaissance and early modern systems
11. Commonplace Books (Ancient Greece-19th Century) - Personal notebooks collecting quotes, ideas, and reflections organized by topic headings, emphasizing personal synthesis of external sources. While medieval manuscripts were typically copied verbatim, commonplace books innovated by encouraging active knowledge curation where readers selected, organized, and reflected on passages. This shift from passive copying to active synthesis represented a fundamental change in how individuals engaged with knowledge, making every reader a potential author.
12. John Locke's Commonplace Method (1706) - Systematic indexing using alphabetical arrangement with expandable sections and cross-referencing techniques for efficient knowledge retrieval. Previous commonplace books used simple topical organization that became unwieldy as they grew, but Locke's innovation introduced a scalable indexing system that could handle unlimited growth. His method transformed commonplace books from simple collections into searchable databases, solving the critical problem of information retrieval that had limited earlier systems.
13. Polish-Lithuanian Silva Rerum (16th-18th Century) - Intergenerational family knowledge repositories containing diverse document types, preserving practical wisdom across generations. Unlike individual commonplace books that died with their authors, silva rerum innovated by creating hereditary knowledge systems that accumulated family wisdom over centuries. This multi-generational approach to knowledge preservation was unique in Europe, establishing knowledge as family patrimony rather than individual achievement.
14. Renaissance Artists' Pattern Books - Collections of sketches, technical notes, and design concepts with cross-references between related techniques, supporting professional knowledge development. While medieval guild knowledge was transmitted orally through apprenticeship, pattern books innovated by codifying visual and technical knowledge in portable, shareable formats. This democratization of craft knowledge accelerated artistic innovation by allowing techniques to spread beyond traditional master-apprentice relationships.
15. Islamic Za'irjah Systems - Mechanical divination devices using Arabic letters to represent philosophical categories, combined through calculations to generate new textual insights. Unlike traditional divination relying on intuition or randomness, za'irjah introduced systematic procedures for generating meaningful text from letter combinations. This mathematical approach to knowledge generation represented an early attempt at algorithmic text creation, prefiguring modern generative AI by combining predetermined rules with combinatorial processes.
Modern digital implementations
Contemporary digital tools directly implementing or inspired by Zettelkasten principles represent the most mature expression of networked knowledge management.
Direct Zettelkasten implementations
16. Obsidian - Local-first knowledge management with bidirectional linking, graph visualization, and extensive plugin ecosystem, supporting true Zettelkasten workflows with modern enhancements. While early digital note-taking apps like Evernote focused on collection and search, Obsidian revolutionized the space by implementing true bidirectional linking and local file storage. This innovation combined the linking power of wikis with the privacy and control of local files, solving the vendor lock-in problem while enabling sophisticated knowledge networks previously impossible in digital systems.
17. Zettlr - Open-source academic writing tool specifically designed for Zettelkasten method, featuring Zotero integration, mathematical formulas, and citation management. Unlike general-purpose note apps that required complex workarounds for academic writing, Zettlr innovated by building Zettelkasten principles directly into academic workflows. This integration of reference management, mathematical notation, and interconnected notes created the first purpose-built environment for scholarly knowledge work in the digital age.
18. The Archive - Native macOS Zettelkasten application emphasizing speed and simplicity, created by the Zettelkasten.de team for faithful implementation of Luhmann's method. While other apps added features that obscured core principles, The Archive innovated through radical simplicity, proving that effective knowledge management doesn't require complex features. This minimalist approach demonstrated that constraint could enhance rather than limit knowledge work, influencing a generation of "tools for thought."
19. Zettelkasten by Daniel Lüdecke - Original digital implementation staying true to Luhmann's system with cross-references, search capabilities, and traditional slip-box organization. As the first dedicated digital Zettelkasten software, it had no direct alternatives and pioneered the translation of physical card systems to digital environments. This groundbreaking tool proved that Luhmann's analog method could be enhanced rather than replaced by digitization, establishing the template for all subsequent implementations.
20. LogSeq - Open-source block-based notes with bidirectional linking, local-first privacy, and bullet-point organization combining Roam's approach with traditional Zettelkasten principles. While Roam Research required cloud storage and subscription fees, LogSeq innovated by offering similar block-reference capabilities with complete data ownership. This democratization of advanced note-taking features while maintaining privacy represented a crucial evolution in making sophisticated knowledge management accessible to privacy-conscious users.
Networked thought platforms
21. Roam Research - Pioneering bi-directional linking tool introducing block-level references, daily notes, and graph databases to mainstream knowledge management. Previous note-taking apps treated notes as isolated documents, but Roam's innovation of block-level referencing allowed ideas to exist independently of their containers. This granular approach to knowledge atomization fundamentally changed how people thought about notes, transforming them from documents into interconnected thought networks.
22. Tana - AI-native workspace with supertags, sophisticated organization, and voice integration, representing next-generation networked thought with artificial intelligence assistance. While first-generation tools required manual linking and organization, Tana innovated by using AI to suggest connections, automate organization, and understand context. This represents the first true fusion of human knowledge management with machine intelligence, moving beyond simple search to active knowledge partnership.
23. RemNote - Hierarchical note-taking integrating spaced repetition, PDF annotation, and academic workflows, combining knowledge management with active learning techniques. Previous tools separated note-taking from study, but RemNote innovated by embedding learning science directly into knowledge capture. This integration of memory techniques with knowledge organization created the first system that not only stored but actively reinforced knowledge retention.
24. Heptabase - Visual note-taking with canvas views for complex project management, offering spatial approaches to knowledge organization and relationship visualization. While most digital tools constrained thinking to linear documents, Heptabase innovated by providing infinite canvases where spatial relationships conveyed meaning. This visual-first approach to knowledge management better matched how many people naturally think, especially for complex, multi-dimensional projects.
25. Capacities - Object-based knowledge management using structured types for organizing information, providing innovative approaches to knowledge categorization and retrieval. Unlike traditional folder or tag systems, Capacities innovated by treating different information types as distinct objects with specific properties and relationships. This object-oriented approach to knowledge brought database concepts to personal notes, enabling more sophisticated organization than simple hierarchies allowed.
Personal knowledge management tools
26. Notion - All-in-one workspace supporting collaborative knowledge management, databases, and structured content creation, though with limited true bidirectional linking capabilities. While previous tools specialized in single functions, Notion innovated by combining documents, databases, and project management in one platform. This consolidation eliminated the friction of switching between tools, though it sacrificed some specialized capabilities for versatility.
27. Reflect Notes - AI-powered networked notes with Kindle integration, encryption, and intelligent connection suggestions, emphasizing privacy and artificial intelligence augmentation. Unlike cloud-based AI tools that process data on external servers, Reflect innovated by implementing local AI processing for privacy-conscious users. This combination of intelligent features with end-to-end encryption solved the privacy-functionality trade-off that plagued earlier AI-enhanced tools.
28. Mem.ai - AI-first note-taking platform with automated organization, smart search, and intelligent content discovery, representing machine-augmented knowledge management. While traditional tools required manual organization, Mem innovated by eliminating folders and tags entirely, relying on AI to surface relevant information contextually. This paradigm shift from hierarchical to associative organization represented a fundamental reimagining of how digital knowledge should be structured.
29. Craft - Beautiful writing tool with block-based structure and Apple ecosystem integration, emphasizing design and user experience in knowledge management workflows. While most note apps prioritized functionality over aesthetics, Craft innovated by proving that beautiful design could enhance rather than distract from knowledge work. This focus on visual polish and native platform integration set new standards for what users could expect from thinking tools.
30. AFFiNE - Privacy-first collaborative workspace combining block-based editing with canvas views, supporting both individual and team knowledge management approaches. Unlike tools that chose between local-first or collaborative features, AFFiNE innovated by enabling both through conflict-free replicated data types (CRDTs). This technical breakthrough allowed true peer-to-peer collaboration without sacrificing data ownership or requiring central servers.
Academic and research methodologies
Scholarly approaches to knowledge organization provide rigorous frameworks for systematic idea development and conceptual networking.
Knowledge organization frameworks
31. Knowledge Organization Systems (KOSs) - Academic frameworks including taxonomies, ontologies, and controlled vocabularies that categorize research concepts through structured relationship hierarchies. Previous library classification systems like Dewey Decimal were rigid and hierarchical, but KOSs innovated by allowing multiple relationship types beyond simple parent-child hierarchies. This flexibility enabled representation of complex conceptual relationships that better reflected actual knowledge structures in specialized domains.
32. Citation Network Analysis - Methodologies analyzing reference patterns in scholarly literature to identify knowledge flows, research impact, and conceptual evolution over time. Before citation analysis, research impact was measured through subjective peer review, but network analysis innovated by providing quantitative, reproducible metrics of influence. This mathematical approach to understanding knowledge transmission revealed hidden patterns in scientific progress invisible to traditional literature review methods.
33. Grounded Theory and Constant Comparative Method - Systematic methodology generating theories through iterative data comparison, creating conceptual networks linking observations to broader theoretical insights. Unlike traditional hypothesis-testing that imposed predetermined frameworks, grounded theory innovated by letting patterns emerge from data itself. This bottom-up approach to theory building revolutionized qualitative research by providing rigorous methods for inductive reasoning.
34. Concept Mapping Methodologies - Structured processes for visual knowledge representation following six-step procedures: preparation, generation, structuring, representation, interpretation, and utilization. While mind mapping relied on intuitive associations, concept mapping innovated by requiring explicit relationship labels between concepts. This precision transformed fuzzy mental models into testable knowledge structures, enabling systematic comparison and evaluation of understanding.
35. Systematic Review and Meta-Analysis - Rigorous evidence synthesis approaches using explicit, reproducible methods to create comprehensive knowledge networks from distributed research findings. Traditional literature reviews were subjective and unsystematic, but systematic reviews innovated by applying scientific methodology to knowledge synthesis itself. This meta-scientific approach transformed literature review from art to science, establishing evidence hierarchies that revolutionized evidence-based practice.
Qualitative research approaches
36. Qualitative Coding and Analysis Systems - Methodologies systematically organizing data into meaningful categories through open, axial, and selective coding processes creating hierarchical concept networks. Before systematic coding, qualitative analysis relied on researcher intuition, but coding systems innovated by providing transparent, replicable procedures for pattern identification. This systematization gave qualitative research the rigor previously exclusive to quantitative methods while preserving interpretive depth.
37. Thematic Analysis - Six-step analytical framework identifying patterns across qualitative data through iterative refinement of conceptual categories and systematic connection-making. Unlike grounded theory's theory-building focus, thematic analysis innovated by providing a flexible method for pattern identification without requiring theoretical development. This accessibility made rigorous qualitative analysis available to researchers without extensive methodological training.
38. Phenomenological Research Methodology - Approaches understanding lived experiences through systematic description, building conceptual models connecting individual experiences to broader insights. While traditional psychology focused on behavior or cognition, phenomenology innovated by making subjective experience itself the object of scientific study. This legitimization of first-person data opened entirely new domains of knowledge previously considered beyond scientific investigation.
39. Framework Analysis - Systematic qualitative analysis using pre-defined frameworks while allowing emergent themes, charting data across cases to identify theoretical patterns. Unlike purely inductive or deductive approaches, framework analysis innovated by combining both in a structured yet flexible methodology. This hybrid approach enabled policy-relevant research that balanced theoretical rigor with practical applicability.
40. Document Co-Citation Analysis - Methods creating knowledge networks based on shared citation patterns, enabling identification of research communities and conceptual relationships. While traditional citation analysis examined direct references, co-citation innovated by revealing implicit relationships through shared referencing patterns. This indirect approach uncovered intellectual structures and research fronts invisible to direct citation analysis.
Visual knowledge organization systems
Visual approaches to knowledge management leverage spatial relationships and graphical representation to support insight generation and concept networking.
Mind mapping and concept mapping
41. Tony Buzan's Mind Mapping Method - Foundational visual thinking technique using central images with radiating branches, colors, and keywords to engage both brain hemispheres in knowledge organization. While traditional outlining was linear and text-based, Buzan's innovation integrated visual elements, color, and radial organization to match natural thought patterns. This synthesis of verbal and visual processing revolutionized note-taking by making it more memorable, creative, and aligned with how the brain naturally associates ideas.
42. Novak's Concept Mapping - Systematic approach using linking words to describe concept relationships, creating propositional statements and supporting cross-links between knowledge domains. Unlike mind maps' free-form associations, Novak innovated by requiring explicit relationship labels that transformed vague connections into testable propositions. This precision enabled concept maps to serve as both learning tools and assessment instruments, revolutionizing educational practice.
43. CmapTools Software - Leading concept mapping platform providing knowledge modeling capabilities, multimedia integration, and collaborative knowledge construction environments. While earlier concept mapping was paper-based and static, CmapTools innovated by enabling dynamic, multimedia-rich maps that could be collaboratively edited across the internet. This digitization transformed concept mapping from individual exercise to social knowledge construction tool.
44. Visual Thinking Strategies (VTS) - Structured approach using three questions to develop visual literacy and critical thinking through systematic observation and discussion of visual materials. Traditional art education focused on historical knowledge and technique, but VTS innovated by using art as a vehicle for developing transferable thinking skills. This pedagogical shift demonstrated that visual analysis could teach critical thinking applicable across all disciplines.
45. Knowledge Visualization Techniques - Comprehensive methods including node-link diagrams, matrix visualizations, treemaps, and interactive dashboards for exploring complex knowledge networks. While early visualization focused on static representations, modern techniques innovated through interactivity, allowing users to dynamically explore and reconfigure knowledge displays. This shift from passive viewing to active exploration transformed visualization from illustration to investigation tool.
Spatial and network visualization
46. Spatial Hypertext Systems - Approaches expressing relationships through spatial proximity and visual attributes rather than explicit links, including historical systems like VIKI and Aquanet. Traditional hypertext required explicit linking, but spatial hypertext innovated by using position, color, and proximity to convey relationships implicitly. This innovation better matched how people naturally organize physical materials, reducing the cognitive overhead of explicit relationship definition.
47. Gephi Network Analysis - Open-source platform for network visualization providing force-directed layouts, community detection algorithms, and interactive exploration capabilities for knowledge networks. Previous network visualization tools were either too simple or required programming expertise, but Gephi innovated by providing professional capabilities through an intuitive interface. This democratization of network analysis made sophisticated graph exploration accessible to non-programmers.
48. Cytoscape - Biological and general network analysis platform with extensive plugin ecosystem and advanced layout algorithms for complex relationship visualization. Originally designed for biological networks, Cytoscape innovated by creating an extensible platform that could handle any network type through plugins. This architectural flexibility transformed it from specialized tool to general-purpose network analysis environment.
49. Kumu Network Platform - Web-based collaborative network visualization with real-time editing, advanced metrics, and storytelling capabilities for knowledge network exploration. While desktop tools required software installation and file sharing, Kumu innovated by moving network visualization entirely online with real-time collaboration. This cloud-based approach enabled teams to collectively explore and annotate knowledge networks without technical barriers.
50. InfraNodus - Text-to-network visualization platform with AI analytics, converting textual content into interactive network graphs for pattern recognition and insight generation. Traditional text analysis produced statistics and word clouds, but InfraNodus innovated by revealing the network structure within text itself. This graph-based approach to text analysis uncovered conceptual relationships and structural gaps invisible to conventional text mining.
Wiki-based knowledge systems
Wiki platforms and collaborative knowledge building systems provide intuitively-extensible, organically-structured hypertextual approaches to collective intelligence and knowledge sharing that just works based on some really important Wiki design principles that re-inventors of wheels seem to try extra hard to forget.
Traditional wiki platforms
51. TiddlyWiki - Non-linear personal web notebook storing everything in a single HTML file, using WikiText notation with automatic bidirectional links between atomic "tiddler" units. While traditional wikis required server infrastructure, TiddlyWiki innovated by packaging an entire wiki system in a single HTML file that could run anywhere. This radical portability combined with its unique "tiddler" concept created the first truly personal wiki that treated information as reusable micro-content units.
52. MediaWiki - Open-source wiki software powering Wikipedia, featuring hyperlinks with automatic backlink generation, categories for organization, and semantic extensions for structured queries. Previous wiki engines were simple and limited, but MediaWiki innovated by providing enterprise-grade features while remaining open source. Its template system, category hierarchies, and extension architecture transformed wikis from simple collaborative documents to sophisticated knowledge platforms.
53. DokuWiki - File-based wiki using plain text files with clean syntax, namespace hierarchies, and plugin architecture, requiring no database while supporting collaborative editing. While most wikis required database servers, DokuWiki innovated by using plain text files for storage, making it incredibly simple to backup, version control, and deploy. This file-based approach democratized wiki hosting and made wiki content permanently accessible even without the wiki software.
54. XWiki - Second-generation wiki platform with structured data models, nested page hierarchies, form-based content creation, and application development capabilities. First-generation wikis were limited to unstructured text, but XWiki innovated by adding structured data capabilities that transformed wikis into application platforms. This evolution from content management to application development represented a fundamental reimagining of what wikis could be.
55. Confluence - Commercial collaboration platform with smart links, real-time editing, automatic link suggestions, and integration with enterprise development workflows. While open-source wikis served technical users, Confluence innovated by providing polish and integration that made wikis acceptable to non-technical corporate users. This enterprise-readiness brought wiki-based knowledge management into mainstream business practice.
Modern wiki implementations
56. Dendron - Hierarchical note-taking tool with schema support, multi-vault capabilities, and VS Code integration, combining wiki principles with developer-friendly workflows. While traditional wikis used flat namespaces, Dendron innovated through hierarchical organization with dot notation and schemas that enforced consistency. This structured approach to wiki organization solved the information architecture problems that plagued large wiki installations.
57. Foam - VS Code-based digital gardening platform using markdown files with GitHub integration, leveraging development environment ecosystems for knowledge management. Unlike standalone wiki applications, Foam innovated by building knowledge management into existing developer toolchains. This integration approach meant developers could manage knowledge using the same tools and workflows they already knew.
58. Quartz - Static site generator converting Obsidian or Roam notes into websites while maintaining links and graph visualizations for public knowledge sharing. Previous publishing solutions lost the networked nature of notes, but Quartz innovated by preserving bidirectional links and graph visualizations in published form. This fidelity to the original knowledge structure transformed publishing from extraction to exposition.
59. Digital Garden Jekyll Templates - Multiple Jekyll-based solutions providing bi-directional links, hover previews, and graph views for publishing interconnected knowledge gardens. While traditional blogs were chronological and isolated, digital garden templates innovated by bringing wiki-like interconnection to public writing. This shift from stream to garden metaphor changed how people thought about sharing knowledge online.
60. Hyperdraft - Markdown to website converter enabling real-time website generation from notes, supporting instant publishing workflows for knowledge sharing. Traditional publishing required build processes and deployment, but Hyperdraft innovated through instant, automatic publishing of markdown changes. This removal of friction between writing and publishing enabled true "working in public" approaches to knowledge sharing.
Knowledge graphs and semantic systems
Advanced knowledge representation systems leveraging formal ontologies, semantic relationships, and graph databases for sophisticated knowledge modeling.
Graph databases and platforms
61. Neo4j - Native graph database using property graphs with nodes, relationships, and properties, featuring Cypher query language and comprehensive graph algorithm libraries. Relational databases forced graph data into tables requiring complex joins, but Neo4j innovated by storing relationships as first-class citizens alongside data. This native graph storage made traversing connections orders of magnitude faster than SQL joins, enabling real-time exploration of complex knowledge networks.
62. AllegroGraph - Semantic graph database with temporal knowledge capabilities, supporting RDF triples with reasoning engines and geospatial-temporal querying. While most graph databases handled static relationships, AllegroGraph innovated by adding time as a native dimension, enabling queries about how knowledge evolved. This temporal capability transformed knowledge graphs from snapshots into historical records that could answer "what did we know when" questions.
63. Stardog - Enterprise knowledge graph platform combining graph databases with reasoning, data virtualization, and unified access across multiple information sources. Previous solutions required copying all data into the graph database, but Stardog innovated through virtual graphs that could query external sources in place. This federation capability enabled knowledge graphs to span entire enterprises without massive data migration projects.
64. ArangoDB - Multi-model database supporting graphs, documents, and key-value storage in single systems, providing native graph traversal with AQL query language. While specialized databases excelled at single models, ArangoDB innovated by supporting multiple data models in one system with a unified query language. This versatility eliminated the need for multiple databases and complex synchronization for projects requiring diverse data types.
65. PuppyGraph - Graph query engine analyzing data in open formats without ETL requirements, enabling real-time graph analysis of existing information architectures. Traditional graph analytics required expensive data extraction and transformation, but PuppyGraph innovated by querying data in place using open formats. This zero-ETL approach democratized graph analytics by eliminating the primary barrier to adoption.
Semantic web technologies
66. Apache Jena - Java framework for semantic web applications featuring TDB triple store, ARQ SPARQL engine, inference engines, and comprehensive RDF manipulation APIs. Earlier RDF tools were fragmented and incomplete, but Jena innovated by providing a complete, integrated framework for building semantic applications. This comprehensive toolkit transformed semantic web development from research project to practical reality.
67. Virtuoso Universal Server - Multi-model database supporting RDF, SQL, and XML with SPARQL endpoints, reasoning support, and linked data publication capabilities. While most databases supported single data models, Virtuoso innovated by unifying multiple models under one system with cross-model querying. This universality enabled organizations to gradually adopt semantic technologies without abandoning existing systems.
68. Protégé - Open-source ontology editor supporting OWL ontologies with visual editing interfaces, reasoning engines, SWRL rules, and extensive plugin architecture. Previous ontology development required hand-coding in formal languages, but Protégé innovated through visual interfaces that made ontology creation accessible to domain experts. This democratization of ontology engineering enabled widespread adoption of semantic technologies beyond computer science.
69. TopBraid Composer - Enterprise ontology development platform with SHACL shapes, visual modeling environments, data integration, and governance capabilities. While academic tools focused on expressiveness, TopBraid innovated by adding enterprise features like governance, versioning, and integration with business systems. This enterprise-readiness brought semantic technologies from research labs into production environments.
70. OntoText GraphDB - Semantic database for RDF and graph analytics with SPARQL compliance, full-text search integration, reasoning capabilities, and analytics workbench. Generic triple stores lacked optimization for real-world queries, but GraphDB innovated through intelligent indexing and caching that made semantic queries performant at scale. This performance breakthrough made semantic databases viable for production applications with billions of triples.
Personal knowledge management methodologies
Systematic approaches to individual knowledge work emphasizing actionable organization, iterative development, and personal knowledge network building.
Second brain methodologies
71. Building a Second Brain (BASB) - Tiago Forte's methodology using CODE framework (Capture, Organize, Distill, Express) and PARA method (Projects, Areas, Resources, Archives) for actionable knowledge management. Previous PKM focused on collection and organization, but BASB innovated by emphasizing creative output as the goal of knowledge management. This shift from consumption to production transformed how people thought about their notes, making them active tools for creation rather than passive storage.
72. Progressive Summarization - Layer-by-layer summarization technique balancing compression with context, designing notes for future discoverability through opportunistic refinement over time. Traditional summarization happened once during initial capture, but Progressive Summarization innovated by treating compression as an ongoing process triggered by actual use. This just-in-time approach to distillation ensured effort was invested only in genuinely valuable information.
73. Evergreen Notes Method - Andy Matuschak's approach emphasizing atomic, densely linked notes written to evolve and accumulate over time, focusing on concept-oriented rather than source-oriented organization. While most note-taking organized by source or chronology, Evergreen Notes innovated by organizing around concepts that could grow indefinitely. This conceptual focus created notes that improved with age rather than becoming obsolete.
74. Digital Gardens - Public knowledge sharing approach emphasizing learning in the open, non-linear growth, and three developmental stages: seedling, budding, and evergreen content. Traditional blogging demanded polished, finished posts, but Digital Gardens innovated by celebrating works-in-progress and continuous revision. This permission to publish imperfect, evolving ideas lowered barriers to sharing knowledge and enabled collaborative learning.
75. Linking Your Thinking (LYT) - Nick Milo's system using Maps of Content and ACCESS framework (Atlas, Calendar, Cards, Extra, Sources, Spaces) for creating fluid knowledge structures. While rigid hierarchies or flat tags were common, LYT innovated through "Maps of Content" that provided flexible, non-hierarchical navigation points. This middle way between structure and chaos enabled organic growth while maintaining navigability.
Specialized PKM approaches
76. PARA Method - Universal organizational system emphasizing actionability over topics, with four categories supporting action-oriented rather than collection-focused knowledge management. Traditional organization used subject categories, but PARA innovated by organizing around actionability and time horizons instead of topics. This temporal approach ensured relevant information surfaced when needed rather than being buried in topical hierarchies.
77. Johnny Decimal System - Numerical hierarchical organization preventing endless subfolder nesting through clear boundaries and Dewey Decimal System-inspired structure. While most systems allowed unlimited hierarchy depth, Johnny Decimal innovated by enforcing strict two-level depth with numerical addressing. This constraint paradoxically increased findability by preventing the deep nesting that made information irretrievable.
78. Atomic Notes Method - Systematic approach emphasizing single ideas per note, self-contained autonomy, and modular knowledge construction through reusable building blocks. Traditional notes mixed multiple ideas in single documents, but Atomic Notes innovated by enforcing one-idea-per-note discipline. This granularity enabled unprecedented reusability and recombination of ideas across different contexts.
79. Seek-Sense-Share Framework - Three-phase knowledge workflow encompassing information seeking, sense-making through analysis, and knowledge sharing with communities for complete lifecycle management. Previous PKM focused on personal benefit, but this framework innovated by making sharing an integral part of the knowledge process. This social dimension transformed PKM from individual activity to community practice.
80. Personal Learning Environment (PLE) - Ecosystem approach combining multiple tools and resources for self-directed learning through aggregation, relation, creation, and sharing workflows. While Learning Management Systems imposed institutional structures, PLEs innovated by giving learners control over their own learning tools and workflows. This learner-centric approach recognized that effective learning required personalized tool ecosystems rather than one-size-fits-all platforms.
Specialized and emerging systems
Contemporary innovations addressing specific knowledge management challenges through novel approaches to visualization, collaboration, and artificial intelligence integration.
AI-enhanced knowledge systems
81. Second Brain AI - AI-powered research assistant with document chat capabilities, memory systems, and browser integration for intelligent knowledge augmentation. Previous AI assistants lacked persistent memory, but Second Brain AI innovated by maintaining context across sessions and actively building knowledge over time. This persistent memory transformed AI from stateless tool to learning partner that grew more valuable through use.
82. Constella.App - AI-powered visual knowledge management with graph-based interfaces, retrieval optimization, and visual canvas integration for next-generation knowledge work. While most AI tools used chat interfaces, Constella innovated by combining AI with visual knowledge graphs for spatial reasoning. This visual-AI fusion enabled new forms of knowledge exploration impossible with text-only interfaces.
83. Mem.ai Enhanced - Advanced AI-first note-taking with automatic connection discovery, smart search capabilities, and machine learning-powered content organization. Traditional AI features were add-ons to existing systems, but Mem built AI into its foundation, making intelligence the primary organizing principle. This AI-native architecture enabled capabilities like self-organizing notes that would be impossible to retrofit into traditional systems.
84. Graphiti - Temporal knowledge graph framework designed for AI agents, supporting dynamic knowledge building with temporal relationships and incremental updates. Static knowledge graphs couldn't represent changing information, but Graphiti innovated by making time and change first-class concepts in knowledge representation. This temporal awareness enabled AI agents to reason about how knowledge evolved rather than just its current state.
85. Anytype - Decentralized knowledge management platform using P2P architecture with object-based organization, local-first principles, and data sovereignty features. While cloud platforms controlled user data, Anytype innovated through true decentralization where users owned their data and infrastructure. This architectural revolution returned data sovereignty to users while maintaining collaboration capabilities through peer-to-peer protocols.
Specialized domain applications
86. DevonThink - Document management system with AI classification, OCR capabilities, advanced search, and large document handling optimized for research workflows. Generic document managers struggled with research volumes, but DevonThink innovated through AI that learned from user behavior to automatically classify and connect documents. This intelligent automation transformed document management from manual filing to assisted curation.
87. Trilium Notes - Hierarchical knowledge base featuring encryption, scripting capabilities, and relationship visualization for technical users requiring advanced functionality. While most note apps targeted general users, Trilium innovated by providing programming capabilities within notes themselves. This scriptability transformed notes from static content to dynamic applications that could process and generate information.
88. Milanote - Visual project organization platform using mood boards and template-based workflows optimized for creative professional knowledge management. Traditional project management was text and timeline-based, but Milanote innovated through visual boards that matched creative thinking patterns. This visual-first approach better supported the non-linear, inspirational nature of creative work.
89. Supernotes - Card-based note-taking system emphasizing speed and cross-platform synchronization with unique card interface metaphors for knowledge organization. While most apps used document metaphors, Supernotes innovated through a card-based interface that treated notes as discrete, manipulable objects. This tactile approach to digital notes made organization feel more like arranging physical cards than managing files.
90. Athens Research - Discontinued but historically significant open-source collaborative knowledge graph demonstrating community-driven approaches to networked thought development. While commercial tools dominated, Athens innovated by proving that community-driven, open-source development could produce sophisticated knowledge tools. Though discontinued, it demonstrated the viability of alternative development models for tools for thought.
Contemporary and hybrid systems
Modern platforms combining multiple knowledge management approaches while addressing current needs for collaboration, mobility, and integration.
Integrated platforms
91. Roam Research Advanced Features - Extended capabilities including block-level references, query systems, collaborative editing, and graph database functionality representing mature networked thought. Basic Roam was revolutionary, but advanced features like datalog queries and custom JavaScript innovated by turning notes into programmable databases. This convergence of notes and code created possibilities for automated knowledge work previously requiring separate programming environments.
92. Notion Advanced Implementations - Database-driven knowledge management using relational properties, template systems, and collaborative workflows, though with limited true bidirectional linking. While Notion's basics were accessible, advanced users innovated by building complex relational systems that transformed it into a no-code database platform. These sophisticated implementations demonstrated that general-purpose tools could match specialized software through creative configuration.
93. Obsidian Plugin Ecosystem - Extended functionality through community plugins supporting spaced repetition, advanced visualization, publishing, and integration with external tools and services. The core application was powerful but limited, yet the plugin ecosystem innovated by enabling community-driven feature development without waiting for official updates. This extensibility transformed Obsidian from application to platform, with plugins adding capabilities the original developers never imagined.
94. TiddlyWiki Extensions - Plugin ecosystem including TiddlyMap for graph visualization, Projectify for project management, and numerous specialized extensions for diverse knowledge management applications. The base system was already unique, but extensions innovated by adapting TiddlyWiki to specialized domains from music composition to genealogy. This adaptability proved that a sufficiently flexible core could serve any knowledge domain through community extension.
95. Logseq Enhanced Workflows - Advanced block-based notes with Git synchronization, query systems, plugin architecture, and privacy-focused local-first development approaches. While basic Logseq competed with Roam, enhanced workflows innovated by leveraging Git for version control and collaboration without cloud dependencies. This developer-friendly approach attracted users who wanted Roam's power with complete data control.
Educational and research applications
96. Compendium - Semantic hypertext tool supporting knowledge mapping and argumentation through Issue-Based Information System (IBIS) methodology for collaborative analysis and decision-making. Traditional decision-making tools were linear, but Compendium innovated by visualizing argument structures as navigable maps. This spatial representation of reasoning made complex deliberations comprehensible and enabled systematic exploration of decision spaces.
97. Concept Explorer - Formal concept analysis tool generating concept lattices from object-attribute relationships with interactive exploration and educational interface design. Mathematical concept analysis was previously paper-based, but Concept Explorer innovated by making formal concept analysis interactive and visual. This accessibility brought rigorous mathematical knowledge analysis to non-mathematicians.
98. ConExp-ng - Concept exploration and lattice analysis platform supporting interactive concept exploration, association rule mining, and educational applications for formal concept analysis. Earlier tools required mathematical expertise, but ConExp-ng innovated through educational features that taught concept analysis while using it. This pedagogical integration made formal methods accessible to students and practitioners alike.
99. Project Xanadu - Theoretical hypertext system with bidirectional linking and transclusion capabilities, representing foundational thinking about universal information access and version control. While never fully implemented, Xanadu's innovations like transclusion, micropayments, and parallel documents influenced every subsequent hypertext system. Its vision of permanent, versioned, universally accessible information remains the theoretical ideal that current systems still strive toward.
100. Vannevar Bush's Memex - Conceptual associative information system using microfilm technology and associative trails, serving as intellectual foundation for hypertext and modern knowledge management systems. Though never built, the Memex innovated by imagining mechanical assistance for human memory and association, establishing the conceptual framework for all subsequent knowledge augmentation tools. This vision of technology amplifying human intellect rather than replacing it continues to guide knowledge system development today.
The universal patterns of knowledge work
This comprehensive survey reveals remarkable consistency in human approaches to knowledge management across cultures, time periods, and technological capabilities. From ancient bamboo strips to modern AI-enhanced knowledge graphs, successful systems consistently implement atomic information units, associative linking mechanisms, emergent organizational structures, and iterative knowledge development processes.
The evolution from physical to digital systems has amplified rather than replaced these fundamental principles. Modern implementations like Obsidian, Roam Research, and semantic knowledge graphs represent technological expressions of timeless human needs: organizing information, connecting ideas, and building upon existing knowledge to generate new insights.
Contemporary trends toward AI augmentation, visual representation, collaborative knowledge building, and privacy-conscious local-first approaches suggest continued innovation while respecting core principles of personal knowledge sovereignty and emergent understanding. The future of knowledge work will likely integrate these historical insights with advancing technologies to create even more powerful tools for human intellectual development and discovery.
These 100 systems demonstrate that effective knowledge management transcends specific tools or technologies—it requires systematic approaches to capturing, connecting, and cultivating ideas over time. Whether implemented through medieval marginalia, index cards, or graph databases, successful knowledge systems serve as thinking partners that amplify human cognitive capabilities and facilitate the discovery of unexpected connections between ideas.
Supplemental List
Notetaking is HIGHLY personal and very subjective because people have different learning styles and usually tend to favor something that they are comfortable with and already using. Below we have a supplemental list of notable Personal Knowledge Management (PKM) systems, platforms, and methodologies that were not on the first list of PKM system, but perhaps, according to some, should have made the top 100.
Some Might Include The Following On the Above List of 100 PKM
- Evernote – Once the dominant note-taking app with strong OCR, web clipping, and cross-device sync. Its decline in innovation and move to subscription-only models may have excluded it, but historically, it was the gateway to digital PKM for millions.
- Microsoft OneNote – A robust, freeform note-taking tool with deep integration into the Microsoft Office ecosystem. Perhaps omitted for its lack of atomic note philosophy, but its flexibility and multi-device sync remain powerful.
- Google Keep – Lightweight, fast, and integrated with Google Workspace; excels for quick capture. May have been excluded for its simplicity and limited linking features, but it’s ubiquitous.
- Scrivener – Writing and research environment designed for long-form projects; strong binder and corkboard metaphor. Possibly excluded because it’s writing-focused rather than link-focused, but its research and reference features qualify it as a PKM tool.
- Workflowy – Minimalist outliner with infinite nesting, mirrors, and tagging. Its laser focus on outlining may have kept it out, but it’s influential in the PKM space.
- Miro – Infinite collaborative whiteboard useful for visual PKM, mind mapping, and linking ideas spatially. Excluded perhaps for being primarily a team tool, but highly relevant for visual thinkers.
- Trello – Card/board-based project organization that can be adapted into a PKM system; great for kanban-based thinking. Likely excluded as “project management,” but it is used by many as a personal idea tracker.
Other Notable Systems, Perhaps More Specialized Or Fill Certain Niches Better, But Worth Mentioning
- Airtable – Flexible database-spreadsheet hybrid used by some for PKM with custom views, linking, and filtering.
- Coda – All-in-one document platform with database features and automation; blurs the line between documents, spreadsheets, and apps.
- Notability – Popular with iPad users for handwritten + typed notes; particularly strong for students and researchers.
- GoodNotes – Another leading handwritten note app with PDF annotation; strong for visual and tactile learners.
- Milanote – (Not in your 100 list’s version?) Visual note boards, great for creative planning.
- Scapple – From Scrivener’s creators, a freeform text + connector mapping tool for non-linear brainstorming.
- Lucidchart / Lucidspark – Diagramming + brainstorming; can integrate with text notes for conceptual mapping.
- Gingko – Card-based hierarchical writing/outlining; great for breaking down ideas.
- Quip – Collaborative docs with spreadsheets and chat, used by some for integrated PKM.
- Zoho Notebook – Free, attractive note-taking app with multimedia cards.
- Standard Notes – Encrypted, minimalist note-taking with extensible editors and tagging; strong on privacy.
- Nimbus Note – Rich note platform with nested folders, databases, and collaboration.
- Roam Highlighter + Readwise Integration – A capture-to-PKM workflow worth separate mention.
- SuperMemo – Spaced repetition + incremental reading pioneer; incredibly powerful for retention-focused PKM.
- Anki – Flashcard-based spaced repetition software; although study-focused, can serve as an evergreen knowledge store.
- Hypothesis – Social annotation tool for PDFs and the web; great for collaborative PKM.
- LiquidText – PDF/document annotation with spatial linking of notes; powerful for research synthesis.
- MarginNote – Combines mind mapping, outlining, and document annotation for integrated learning.
- TagSpaces – Local file tagging and note-taking; good for offline PKM and privacy.
- Joplin – Open-source Evernote alternative with markdown, encryption, and sync.
- Lynked.World – Visual, public graph-based knowledge sharing; newer entrant in the digital garden space.
- Memos – Lightweight self-hosted note-taking with markdown, tagging, and linking.
- Tangents – Graph-based PKM platform with a focus on concept connections.
Other Emerging Or More Specialized PKM Systems
- Muse – Card and canvas-based spatial PKM, optimized for tablets.
- Scrapbox – Wiki-like PKM with instant bidirectional linking and block references.
- Athens (Modern successor forks) – Open-source Roam alternative; some forks are active despite Athens Research ending.
- Tangent Notes – Markdown-based PKM with bidirectional linking, local-first philosophy.
- NotePlan – Calendar + daily notes + tasks; bridges PKM with GTD workflows.
- Amplenote – Combines tasks, notes, and scheduling with bidirectional links.
- Akiflow – Primarily task-focused, but integrates with PKM sources for time-blocked thinking.
- Chronicle – Long-term personal history + notes archive.
- Bangle.io – Web-based markdown note system with backlinking.
- DynaList – Outliner predecessor to Workflowy; still used for hierarchical PKM.