Post
JA EN

Repurposing Engineers' Pattern Recognition Assets in the AI Era: A Three-Layer Update for the Dual Value of Evaluation and Learning

Repurposing Engineers' Pattern Recognition Assets in the AI Era: A Three-Layer Update for the Dual Value of Evaluation and Learning
  • Target audience: Engineers thinking about how to leverage their accumulated pattern recognition skills in the AI era
  • Prerequisites: Experience using AI coding assistants in your work
  • Reading time: 22 minutes

Overview

Software engineering has always been a pattern-recognition-heavy craft. Design patterns, code smells, anti-patterns, design principles — all of them assume a mental library of patterns built up through experience, which we apply to unfamiliar problems. This capability has historically generated value along two distinct axes. On the evaluation axis — how quickly you spot a problem — Schmidt & Hunter’s 1998 85-year meta-analysis shows that general mental ability (with pattern recognition at its core) is the strongest predictor of job performance in complex roles (r≈0.51)1; this has been the single biggest predictor of who gets evaluated as a strong engineer. On the learning axis, the same library lets you learn only the delta when adopting a new language or framework, dramatically cutting acquisition cost — a phenomenon repeatedly demonstrated in research on analogical transfer2.

So what happens to this asset in the AI era? That is the question this article tackles.

The short answer: the asset still pays off, but the weighting between the two axes shifts. The empirical record splits in three directions. The GitHub Copilot study (Peng et al., 2023) reports 55.8% faster task completion3. Microsoft’s three-company RCT with 4,867 developers (Cui et al., 2024) shows +27–39% for less experienced engineers and +8–13% for experienced ones4. Meanwhile, METR’s 2025 study found that experienced open-source developers working in familiar repositories were 19% slower when using AI5. “Experience automatically pays off in the AI era” is not a claim the evidence supports.

On the evaluation axis, AI is starting to partially substitute for the “spot the problem quickly” layer. On the learning axis, the weighting actually rises: the faster the cycle of new tech, the more “learn from zero” mode falls behind, and delta-learning capacity becomes the ceiling on how fast you can refresh your skills. This article positions three components from cognitive science — perceptual learning, episodic memory, and metacognition — as the update mechanism for engineers’ existing pattern recognition. We will work down to a concrete weekly routine of 60–90 minutes, grounded in evidence, that repurposes both axes for the AI era.

Engineering practice leans heavily on pattern recognition

A quick framing first. Engineering work uses pattern recognition heavily. “Central” is too strong — real practice mixes in communication, design, requirements work, and much more — but the share of work that rides on an accumulated pattern library is clearly higher than in most other professions, and multiple studies confirm this.

The vocabulary of the field reflects it:

  • Design patterns: Since the Gang of Four in 1994, the industry has treated “reusable patterns” as a shared asset
  • Code smells: Martin Fowler’s Refactoring popularized the idea of code that “smells” — less an explicit rule than a perceptual feel for something off
  • Architecture patterns / anti-patterns: Microservices, CQRS, God Object, N+1 queries — none of these conversations work without a shared pattern vocabulary
  • Debugging: Glancing at a stack trace and concluding “circular reference” within seconds is textbook pattern matching

Research has been confirming this cognitive profile for decades. McKeithen et al. (1981) had expert and novice programmers memorize programs and found that experts encoded code in “meaningful chunks,” while novices retained the literal surface6. Follow-up work like Wiedenbeck (1986) replicated the finding repeatedly: expert programmers rely on pattern-based cognition. The structure mirrors Chase & Simon’s classic chess work (1973): the moment an expert sees the board (the code), they decompose it into units7.

That said, “pattern recognition alone makes the job work” would overstate it. Meyer et al. (2014, Microsoft Research) observed 11 professional developers over four hours each and found an average of 47 activity switches per hour, with each activity lasting just 1.6 minutes on average — work that is extremely fragmented8. Microsoft’s follow-up study (2024) shows that the time developers most want to spend is on “coding” (≈20%), “designing new systems” (≈15%), and “learning new technology,” with the remainder scattered across meetings, communication, and support9.

So the silhouette of practice looks like this:

  • Coding and design (≈35%): pattern recognition is the dominant driver
  • Communication and meetings: verbalization and interpersonal skills take the lead
  • Learning / investigating new tech: meta-learning leads — but pattern recognition acts as an accelerator here too, as we will see
  • Debugging: a mix of pattern recognition and hypothesis testing

Pattern recognition anchors 30–40% of the work, and that is where this article focuses. The other capabilities (verbalization, meta-learning, interpersonal skills) deserve their own treatments.

Pattern recognition has been doubly valuable: the evaluation axis and the learning axis

Within engineering, pattern recognition has historically produced value through two distinct channels.

Evaluation axis: the strongest predictor of “good”

Schmidt & Hunter’s 1998 85-year meta-analysis is a classic of industrial-organizational psychology. Comparing 19 personnel-selection methods, it showed that general mental ability (GMA, with pattern recognition at its core) had a job-performance validity coefficient of r≈0.51 for medium-complexity jobs — the strongest predictive power of any selection method1. For comparison, job experience comes in at r≈0.18 and education at r≈0.10. The more cognitively complex the work — engineering qualifies — the larger this gap.

Research on expert cognition points in the same direction. Chase & Simon’s chess expertise studies, and Klein’s Recognition-Primed Decision model (where firefighters and military commanders make decisions by pattern matching, not deliberation) — all converge on the claim that pattern recognition sits at the core of expertise7. Inside engineering, this is the “she takes one look at the code and finds the bug” or “he spots the design problem instantly in review” phenomenon — and that has been a powerful signal in how the field evaluates talent.

Learning axis: delta learning cuts acquisition cost

The second source of value is learning acceleration. In Gick & Holyoak’s (1980) classic study, the baseline solution rate for Duncker’s radiation problem was 10%; readers given a structurally analogous “military story” beforehand reached 30%; and when told to use the story as a hint, 75%2. New information that overlaps with existing schemas is learned and transferred dramatically faster than learning from scratch — that is the heart of analogical transfer.

Alfieri et al. (2013) meta-analyze comparison-based learning across 57 studies and show that placing multiple cases side by side and verbalizing the differences produces faster and deeper schema acquisition than studying each case in isolation10. Rohrer & Taylor’s work on interleaving similarly shows that mixing similar-but-distinct items in practice outperforms blocked practice on long-term retention and transfer11. The mechanism is that forcing the brain to ask “how are these different, and why” sharpens the boundaries between patterns.

Translated to engineering practice, this matches everyday experience:

  • The second OO language is far faster to pick up than the first
  • Knowing React drops the cost of Vue or Svelte dramatically
  • The more incidents you have lived through, the faster you recognize “this looks like that one from before”

This channel of value is independent of the “good engineer” evaluation channel, and the richer your pattern library, the cheaper new technology becomes — producing a cumulative advantage.

What changes in the AI era: a re-weighting of the two axes

In the AI era, the balance between these two axes shifts. Lining up three empirical studies makes the contour of the change visible.

  • Peng et al. (2023, GitHub Copilot): JavaScript HTTP-server implementation task. The Copilot group completed 55.8% faster, with the largest gains for less-experienced developers3
  • Cui et al. (2024, Microsoft / three-company RCT with 4,867 developers): +27–39% for the less experienced, +8–13% for the experienced. A clear experience interaction, but positive effects across the board4
  • METR (2025, RCT with 16 OSS developers and 246 tasks): Working in a familiar repository, AI use led to a 19% slowdown5

Mapping these to the two axes:

On the evaluation axis, AI is partially substituting. The “spot the problem quickly” layer — where pattern-rich engineers have been rewarded — is now covered to some extent by AI. The human role shifts toward “how do I judge whether the AI’s output is sound?” The METR result suggests this verification cost can, for experienced engineers, outweigh just writing it themselves.

On the learning axis, the weighting actually rises. New languages, frameworks, and paradigms arrive faster than ever; learning from zero doesn’t keep up. Delta-learning capacity becomes the ceiling on skill renewal. The +8–13% Cui et al. find even among experienced developers reads as the result of a group that uses an existing pattern library as scaffolding and treats AI as a source of delta information.

So to keep using your pattern recognition asset in the AI era, you need to redirect the evaluation axis toward “validating AI output” and the learning axis toward “faster delta learning.” The three-layer update mechanism below is built for exactly that.

Three layers that update the asset: perceptual learning, episodic memory, metacognition

ComponentRole for the existing assetAxis it strengthens
Perceptual learningDiversifies and refreshes the pattern libraryEvaluation + Learning
Episodic memoryEnriches retrieval keys for specific contextsEvaluation
MetacognitionControls the cross-check against AI outputEvaluation (and asking better questions)

All three are demonstrably trainable in adulthood. That they can be treated as investable capabilities rather than fixed talent is what makes them valuable in the AI era.

1. Perceptual learning to “update” the pattern library

Perceptual learning is the phenomenon by which perception itself sharpens through repeated experience. The crucial point is that it is not a talent locked down in childhood; it continues to improve in adulthood12.

Kellman and colleagues at UCLA built this into a learning intervention called Perceptual Learning Modules (PLM) and validated it across multiple mathematics domains12:

  • Algebraic-transformation PLM: Equation-solving time dropped from 28 seconds to 12 seconds, with full retention on a delayed test two weeks later
  • Linear-measurement PLM: Significant gains over controls (F=19.687, p<0.001), with effects preserved on a four-month delayed test
  • Multiple-representation PLM: Significant interaction (F=21.17), with transfer even to untrained translation problems (generation tasks)

The design principle is to train selectivity and fluency across diverse instances12. This is structurally the same principle as interleaved practice11 and comparison-based learning10 — it simultaneously trains the evaluation axis (an eye for differences) and the learning axis (faster delta acquisition).

In engineering terms, that translates to:

  • Lay three implementations of the same feature side by side and verbalize the trade-offs
  • Compare an OSS feature’s early implementation, its mature version, and its post-refactor form in time order
  • Ask AI to write the same code three times with the same prompt; identify the differences and reason about why they differ

The point is not to memorize the right answer but to build an eye that notices differences. The core of perceptual learning is training discrimination, not memorizing answers12. And that same training lifts your delta-learning capacity for new technology.

2. Episodic memory, used reconstructively

Pattern recognition does not run on abstract rules alone. Memories of concrete instances — episodic memory — provide the retrieval key that lets you say “this case resembles that one from before.” Without training this layer, abstract patterns will not surface when you need them.

Episodic memory is also trainable. Banducci et al. (2017) recruited 179 participants with an average age of 69.46 years and split them into an Active Experiencing (AE) group (n=93) and a control group (a theatre-history course; n=86), running a four-week intervention13. The AE group performed short scenes with acting partners, and in the later weeks they had to memorize scripts and reproduce them accurately. The results13:

  • Significant improvement in story recall (between-group difference = 0.26, p=0.04)
  • Larger improvement in theme recall (between-group difference = 0.49, p=0.002)

At the four-month follow-up, however, both groups had improved and the group difference had disappeared (the control group also gained over time)13. The fair reading is that the intervention works but sustained differentiation requires ongoing practice.

What stands out is that the core of AE training is not rote memorization but reproducing material with meaning and emotion inside a situation. This maps structurally onto an engineer’s “incident retrospective” or “remembering what a colleague said in a design review.” The habit of actively reconstructing “what was I thinking, and how did I decide?” is exactly what later powers the judgment of “does this AI output fit our context?”

One caveat: this study targeted older adults, so effect sizes for younger engineers need separate validation. What the study does clearly establish is that the premise “give up on your memory after a certain age” is wrong.

3. Metacognition to control the cross-check against AI output

Sharpening pattern extraction through perceptual learning and enriching retrieval keys through episodic memory still isn’t enough for the AI era. You need another layer that handles “how do I reconcile my own judgment with what the AI produced?” That layer is metacognition.

Machiko Sannomiya defines metacognition for the AI era as “the skill of stepping back and questioning your own cognition,” and warns that generative AI carries risks of “fluency-driven trust bias” and “thinking atrophy” — so the right posture is active engagement rather than passive reliance14.

Empirical work backs this up. A CHI 2024 study (Tankelevitch et al.) shows that generative AI imposes new cognitive demands on users — judging whether AI assistance helps in a given moment, evaluating the quality of the output, adjusting your own approach in response — and gives this load a name: “metacognitive demands”15.

Another study (Gerlich, 2025, N=666) reports a strong negative correlation between cognitive offloading to AI and critical thinking ability (r=-0.75), with the effect particularly pronounced for less-experienced users16. Training metacognition is the lever here. The habit of asking, “Did I actually verify this AI output, or did I just settle for the fluency of the explanation?” plausibly counteracts the slide in critical thinking.

In engineering practice, that looks like:

  • After receiving AI output, run it through a verification checklist (What is the basis for the claim? Would I make a different choice? How does this differ from past similar cases?)
  • Explicitly decouple “smoothness of explanation” from “quality of substance” — the smoother it sounds, the thinner the scrutiny tends to be14
  • Verbalize your own judgment first, then show it to the AI. With a hypothesis in hand, you drift less toward the AI’s answer16

Three loops: how they interlock

Pulling the argument together onto one page: the three components are not independent tools but a self-reinforcing loop.

flowchart TB
    A[Existing pattern assets<br>Design patterns,<br>code smells, etc.] --> B[Perceptual learning<br>updates diversity]
    B --> C[Episodic memory<br>Context accumulation]
    C --> D[Pattern recognition<br>Similarity detection]
    D --> E[Validate AI output<br>Metacognition]
    E -->|Articulate discomfort| F[Record as new instance]
    F --> C
    E -->|Correct false patterns| B
    B -.Accelerate delta learning.-> G[Lower cost of<br>learning new tech]

Key points of the loop:

  1. Start from existing assets (design patterns and so on), and run perceptual learning across diverse instances
  2. Sharpened perception makes it easier to encode concrete instances as episodic memory
  3. A richer episodic memory means more retrieval keys for pattern recognition
  4. Pattern recognition lets you check AI output, and metacognition verbalizes any “something feels off”
  5. Logging that “feels off” as a new instance updates both perceptual learning and memory
  6. The “eye for differences” trained in perceptual learning also transfers to delta learning for new technology

Conversely, the loop stalls if any single piece is missing. No exposure to new patterns (no perceptual learning), no retrospection (shallow episodic memory), no questioning of your own judgment (no metacognition) — any of these tips you toward being swept along by AI output.

Engineering practice: update an existing asset in 60–90 minutes a week

Now to translate the cognitive science into engineers’ day-to-day. Three layers of habit.

Layer 1: Perceptual learning — “compare the differences,” 10 minutes, 2–3 times a week

  • In code review, place two “similar but different” designs side by side: Compare alternative implementations of the same goal and verbalize the trade-offs
  • Diff-reading via OSS commit history: Line up the early implementation, the mature implementation, and the post-refactor form of the same feature, and stare at them until you can explain the differences out loud
  • Compare multiple AI outputs: Have AI generate the same code three times from the same prompt, then identify and reason about the differences
  • When learning new tech: Place the new technology next to a known one and read with “what is shared, what is the delta” in mind (this is delta-learning acceleration)

The point is to build an eye that notices differences121011. This works on both the evaluation axis (validating AI output) and the learning axis (delta learning).

Layer 2: Episodic memory — “reconstructive retrospection,” 15 minutes once a week

  • A narrative retrospective after an incident: Beyond “what happened,” write down “what was I thinking at that moment, and why did I choose what I chose” as if telling someone the story
  • A self-journal for design decisions: After a significant choice, spend 5 minutes writing the context, the alternatives, and the reasons you ruled them out
  • A conversation with your past self: Re-read code or decisions from three months ago and push back on the version of yourself who wrote them

The implication of AE training is that “re-enacting as a participant” beats “recording facts”13.

Layer 3: Metacognition — a daily verification loop on AI output

  • Run AI output through a verification checklist: basis for the claim / would I choose differently / what differs from past similar cases
  • Self-check for “am I just impressed by fluency?” — the smoother the explanation, the thinner the scrutiny tends to be14
  • Verbalize your judgment before showing it to the AI: writing your hypothesis first reduces how much the AI’s answer pulls you off course16

The three layers add up to roughly 60–90 minutes a week. A frequency you can actually sustain, alternating perceptual training and retrospection, is the realistic landing point.

Not just “judgment,” also “asking better questions”

Two capabilities show up repeatedly on lists of what matters in the AI era: asking good questions and making good judgments. The three layers above feed judgment (the evaluation axis) directly, but it’s worth emphasizing that they also feed asking good questions.

Metacognition research treats metacognitive monitoring — watching the state of your own understanding — as the engine for “knowing what you don’t know”1415. Throwing useful questions at an AI starts with spotting “what don’t I understand yet?” and “where is my uncertainty?” That is the work of metacognition itself.

Perceptual learning feeds the question side as well. The “something nags at me here” reaction experts notice is the work of the selectivity that perceptual learning trains12; putting words to that nagging tends to yield the right question. “This code looks wrong now that I’m reading it.” “Why is this one block written differently?” Questions grow out of that feel.

So the structure is:

  • Judgment (evaluation axis) = the direct output of all three layers
  • Asking good questions = metacognition leads, perceptual learning’s “discomfort detection” supports
  • Speed of learning (learning axis) = perceptual learning at the core, accelerated as delta learning

Whichever you emphasize, the foundation is the same three layers.

Limitations and caveats

The evidence points strongly in one direction, but a few caveats are worth stating explicitly:

  • Individual differences are real: Intervention effects always have a spread. None of these studies claim “everyone can become an expert”12
  • Studies on younger engineers are scarce: Banducci et al. (2017) studied older adults; effect sizes in younger, working engineers need separate validation13
  • Continuity is assumed: Four-month maintenance is documented, but training that stops likely fades over time1213
  • The Einstellung effect: When pattern recognition is too strong, “being dragged toward an old pattern and missing a better solution” becomes a real failure mode (Bilalić et al., 2008, chess research)17. Experts are most prone to it, and one role of the metacognition layer is to push back on it
  • The territory AI substitutes for will keep shifting: “LLMs cannot do metacognition” is too strong a claim; current LLMs are strong at statistical pattern matching but have limited self-monitoring of their reasoning process1518, and what happens next is a separate question
  • The asset itself can age: Language-specific idioms from older eras may lose value. The job of the perceptual-learning layer is precisely to keep refreshing those

With those caveats, the core claim — that the underlying cognitive mechanisms remain trainable in adulthood, and that the existing pattern recognition asset can be updated for the AI era — holds up.

Summary

The pattern recognition capacity of software engineers has been doubly valuable: as the evaluation axis (Schmidt & Hunter’s r≈0.51) and as the learning axis (delta learning via analogical transfer). The AI era is rebalancing the evaluation axis, while it actually raises the weight of the learning axis.

Three small habits to start tomorrow:

  1. 10 minutes, 2–3 times a week: Line up similar-but-distinct cases and verbalize the differences (perceptual learning — updates both evaluation and learning)
  2. 15 minutes once a week: Reconstruct a recent decision as a story (episodic memory — building retrieval keys)
  3. Daily: Push AI output through a verification checklist (metacognition — controlling the cross-check)

The total is 60–90 minutes a week. The more time AI takes off your plate, the more room there is to make this investment. You don’t need to start over from zero. Bolt on the update mechanism, and the existing asset becomes a differentiator for the AI era.

If this resonated, take a look at the related pieces:

References

References are listed in citation order corresponding to the footnote numbers in the text.

  1. The Validity and Utility of Selection Methods in Personnel Psychology: Practical and Theoretical Implications of 85 Years of Research Findings — Schmidt, F. L., & Hunter, J. E. (1998). Psychological Bulletin, 124(2), 262–274. GMA job-performance validity r≈0.51 for medium-complexity jobs; 85-year meta-analysis. [Reliability: High] ↩︎ ↩︎2

  2. Analogical Problem Solving — Gick, M. L., & Holyoak, K. J. (1980). Cognitive Psychology, 12(3), 306–355. Experimental demonstration of analogical transfer on Duncker’s radiation problem (10% → 30% → 75%). [Reliability: High] ↩︎ ↩︎2

  3. The Impact of AI on Developer Productivity: Evidence from GitHub Copilot — Peng, S., Kalliamvakou, E., Cihon, P., & Demirer, M. (2023). arXiv:2302.06590. JavaScript HTTP-server task; the Copilot group completed 55.8% faster. [Reliability: Medium–High] ↩︎ ↩︎2

  4. The Effects of Generative AI on High Skilled Work: Evidence from Three Field Experiments with Software Developers — Cui, Z., Demirer, M., Jaffe, S., Musolff, L., Peng, S., & Salz, T. (2024). SSRN Working Paper ID 4945566. RCT across three companies with 4,867 developers: +27–39% for the less experienced, +8–13% for the experienced. [Reliability: High] ↩︎ ↩︎2

  5. Measuring the Impact of Early-2025 AI on Experienced Open-Source Developer Productivity — METR (2025). RCT with n=16 and 246 tasks. In familiar repositories, AI use produced a 19% slowdown. [Reliability: High (small N noted)] ↩︎ ↩︎2

  6. Knowledge Organization and Skill Differences in Computer Programmers — McKeithen, K. B., Reitman, J. S., Rueter, H. H., & Hirtle, S. C. (1981). Cognitive Psychology, 13(3), 307–325. Classic study showing that expert programmers encode code in meaningful chunks. [Reliability: High] ↩︎

  7. Perception in Chess — Chase, W. G., & Simon, H. A. (1973). Cognitive Psychology, 4(1), 55–81. Foundational study of pattern recognition and chunking in chess experts. [Reliability: High] ↩︎ ↩︎2

  8. Software Developers’ Perceptions of Productivity — Meyer, A. N., Fritz, T., Murphy, G. C., & Zimmermann, T. (2014). FSE 2014. Observational study of 11 professional developers (4 hours each); 47 activity switches per hour and 1.6 minutes per activity on average. A 379-person survey is also used. [Reliability: High] ↩︎

  9. Time Warp: The Gap Between Developers’ Ideal vs. Actual Workdays — Microsoft Research (2024). Desired time allocation: coding ≈20%, design ≈15%, with the remainder spread across other activities. [Reliability: High] ↩︎

  10. Learning Through Case Comparisons: A Meta-Analytic Review — Alfieri, L., Nokes-Malach, T. J., & Schunn, C. D. (2013). Educational Psychologist, 48(2), 87–113. Meta-analysis of comparison-based learning across 57 studies. [Reliability: High] ↩︎ ↩︎2 ↩︎3

  11. The shuffling of mathematics problems improves learning — Rohrer, D., & Taylor, K. (2007). Instructional Science, 35, 481–498. Demonstrates that interleaved practice outperforms blocked practice on long-term retention. [Reliability: High] ↩︎ ↩︎2 ↩︎3

  12. Perceptual Learning Modules in Mathematics: Enhancing Students’ Pattern Recognition, Structure Extraction, and Fluency — Kellman, P. J., Massey, C. M., & Son, J. Y. (2010). Topics in Cognitive Science, 2(2), 285–305. DOI: 10.1111/j.1756-8765.2009.01053.x (also available via PMC). [Reliability: High] ↩︎ ↩︎2 ↩︎3 ↩︎4 ↩︎5 ↩︎6 ↩︎7 ↩︎8

  13. Active Experiencing Training Improves Episodic Memory Recall in Older Adults — Banducci, S. E., Daugherty, A. M., Biggan, J. R., Cooke, G. E., Voss, M., Noice, T., Noice, H., & Kramer, A. F. (2017). Frontiers in Aging Neuroscience, 9, 133. [Reliability: High] ↩︎ ↩︎2 ↩︎3 ↩︎4 ↩︎5 ↩︎6

  14. “Metacognition,” the Skill for Surviving the AI Era: Three Strategies to Boost Learning — Interview with Machiko Sannomiya, Adecco Group “Power of Work” (2023). [Reliability: Medium–High] ↩︎ ↩︎2 ↩︎3 ↩︎4

  15. The Metacognitive Demands and Opportunities of Generative AI — Tankelevitch, L., Kewenig, V., Simkute, A., Scott, A. E., Sarkar, A., Sellen, A., & Rintel, S. (2024). Proceedings of the 2024 CHI Conference on Human Factors in Computing Systems. [Reliability: High] ↩︎ ↩︎2 ↩︎3

  16. AI Tools in Society: Impacts on Cognitive Offloading and the Future of Critical Thinking — Gerlich, M. (2025). Societies, 15(1), 6. N=666; cognitive offloading and critical thinking correlate at r=-0.75. [Reliability: Medium–High] ↩︎ ↩︎2 ↩︎3

  17. Why good thoughts block better ones: The mechanism of the pernicious Einstellung (set) effect — Bilalić, M., McLeod, P., & Gobet, F. (2008). Cognition, 108(3), 652–661. Shows that even chess experts are pulled toward familiar patterns and miss optimal solutions. [Reliability: High] ↩︎

  18. Evidence for Limited Metacognition in LLMs — 2025 preprint examining the limits of metacognitive capacity in LLMs. [Reliability: Medium (preprint)] ↩︎

This post is licensed under CC BY 4.0 by the author.