{R}R 開発ノート


合計 128 件の記事が見つかりました。

Chapter 7: Plugins and Domain Specialization — Turning Claude Into Your Organization's Expert

Chapter 7 of Master Claude Chat, Cowork and Code explores how plugins transform Claude from a generalist into a domain expert — with pre-built plugins for Sales, Finance, Marketing, and Legal, slash commands, and organization-managed customization.
2026-03-08

Chapter 2: From OpenID to OpenID Connect — How the Industry Got This One Right

Chapter 2 of the OpenID: Modern Identity series — tracing how the industry moved from the original OpenID and SAML through OAuth 2.0 to OpenID Connect, and when to reach for each standard.
2026-03-08

Chapter 1: Why Identity Is Hard — The Trust Problem Behind Every Login

Chapter 1 of the OpenID: Modern Identity book series — why identity is a trust problem first and a technology problem second, and why authentication and authorization must never be conflated.
2026-03-07

Chapter 5: Rapid Prototyping with Artifacts — From Conversation to Live Application

Chapter 5 of Master Claude Chat, Cowork and Code explores how Claude Artifacts collapse the feedback loop between idea and execution — turning conversations into live, interactive applications in seconds.
2026-03-06

OpenID: Modern Identity for Developers and Architects — A 22-Part Blog Series

Introduction and index for the 22-part blog series based on OpenID: Modern Identity for Developers and Architects by Sho Shimoda — with links to every chapter from Why Identity Is Hard through Identity in AI Systems.
2026-03-06

Master Claude, Chapter 4: Context Persistence with Claude Projects — Solving the AI Amnesia Problem

Chapter 4 of Master Claude Chat, Cowork and Code explains how Claude Projects solve the AI amnesia problem with persistent context — custom instructions, knowledge bases, and shared team workspaces that remember your architecture, conventions, and patterns across every conversation.
2026-03-05

Master Claude, Chapter 2: The Three Pillars of Claude — Chat, Cowork, and Code

Claude is not one product — it is three. Chat for reasoning, Cowork for desktop automation, Code for terminal-based development. Chapter 2 of Master Claude Chat, Cowork and Code explains the architecture of each and the decision framework for choosing the right one.
2026-03-03

Master Claude, Chapter 1: The Evolution of Large Language Models — From Markov Chains to Context Engineering

Chapter 1 of Master Claude Chat, Cowork and Code traces the journey from statistical text prediction to reasoning engines — and explains why context engineering, not bigger models, is where the next leap in AI productivity comes from.
2026-03-02

Master Claude Chat, Cowork and Code – The Complete Blog Series

The complete index for the Master Claude Chat, Cowork and Code blog series — 20 chapter teasers covering everything from prompting fundamentals to multi-agent architectures, security governance, and the future of AI-powered work.
2026-03-01

Art of Coding, Chapter 19: Why I Still Code

The final chapter. A personal reflection on why the act of writing code remains meaningful—and why craftsmanship endures even as everything else changes.
2026-01-17

Art of Coding, Chapter 17: AI, Automation, and the Role of the Engineer

How AI changes engineering roles. Why automation removes drudgery but makes human judgment more valuable, and what "curation" means for the future programmer.
2026-01-15

Art of Coding, Chapter 16: Ethics and Longevity

How ethics and longevity intertwine in code. Why the systems you write today remain your responsibility for years, and how empathy shapes sustainable software.
2026-01-13

Art of Coding, Part VI: The Human Side of Code

The human side of code: collaboration, culture, and the practices that make software sustainable. How teams thrive when they value people as much as process.
2026-01-11

Art of Coding, Chapter 13: Testing as a Design Discipline

Testing is a design discipline. How well-written tests reveal awkward APIs, improve code clarity, and become the most reliable documentation of system behavior.
2026-01-09

Art of Coding, Part V: Tools and the Ecosystem

Tools shape the culture of how teams code. The right ecosystem amplifies clarity and craftsmanship; the wrong one creates friction and distraction.
2026-01-07

Art of Coding, Chapter 11: Architectural Thinking

Architectural thinking is the discipline of designing systems that survive real-world growth. It means asking how your code will feel to live in years from now.
2026-01-06

Art of Coding, Chapter 10: Anti-Patterns to Avoid

Anti-patterns are the structural traps that silently erode codebases. Learning to recognize them early is one of the most valuable skills a developer can have.
2026-01-05

Art of Coding, Chapter 9: Design Patterns as a Language of Developers

Design patterns compress complex architectural ideas into shared language. But they're tools for solving problems, not decorations for code.
2026-01-04

Art of Coding, Part IV: Patterns, Anti-Patterns, and Architecture

Part IV explores design patterns as language, anti-patterns as warning signs, and architecture as the invisible skeleton enabling system growth.
2026-01-03

Art of Coding, Chapter 8: Performance without Sacrificing Clarity

Chasing speed too early blinds you to real bottlenecks. Clarity first, measurement second, optimization third—that's the order.
2026-01-02

Art of Coding, Chapter 7: Error Handling and Resilience

Designing for failure, not avoiding it. How graceful error handling, clear logging, and balanced defense build systems that endure.
2026-01-01

Art of Coding, Chapter 6: Abstraction and Modularity

Drawing boundaries that make systems stronger. How to abstract without over-engineering, and design interfaces that last.
2025-12-31

Art of Coding, Part III: Practices That Shape Good Code

From principles to practice. How daily habits, small decisions, and repeated choices shape code that actually endures.
2025-12-30

Art of Coding, Chapter 5: Consistency and Style

Consistency is kindness. How coding standards, formatters, and idiomatic style shape code that teams can actually live with.
2025-12-29

Art of Coding, Chapter 4: Maintainability and Scalability

How to build code that bends instead of breaks, systems that grow without collapsing, and anticipate change without over-engineering.
2025-12-28

Art of Coding, Chapter 3: Readability First

Readability first: how naming, structure, and visual rhythm make code habitable for teams and time.
2025-12-27

Art of Coding, Part II: Principles of Clarity

Part II introduces clarity as the compass of software: readability, maintainability, and the consistency that makes teams move faster.
2025-12-26

Art of Coding, Chapter 2: The Philosophy of Clean Code

Clean code is a philosophy, not a rulebook. Explore simplicity vs. cleverness, expressiveness as communication, and code as a form of writing.
2025-12-25

Art of Coding, Chapter 1: Code That Speaks

Chapter 1 of the Art of Coding series. Why beauty in code is not decoration but survival — clarity, empathy, efficiency, and what separates code that works from code that lasts. Plus: what AI-generated code means for craftsmanship.
2025-12-24

Art of Coding, Part I: Why Code is an Art

Introducing the Art of Coding blog series: a 26-week exploration of what makes code beautiful, maintainable, and enduring in the age of AI.
2025-12-23

8.4 PCA and Spectral Methods

An intuitive, in-depth explanation of PCA, spectral clustering, and eigenvector-based data analysis. Covers covariance matrices, graph Laplacians, and why eigenvalues reveal hidden structure in data. Concludes Chapter 8 and leads naturally into SVD in Chapter 9.
2025-10-10

8.3 The QR Algorithm (High-Level Intuition)

A clear, intuitive, and comprehensive explanation of the QR algorithm—how repeated QR factorizations reveal eigenvalues, why orthogonal transformations provide stability, and how shifts and Hessenberg reductions make the method efficient. Ends with a smooth bridge to PCA and spectral methods.
2025-10-09

8.2 Rayleigh Quotient

An intuitive and comprehensive explanation of the Rayleigh quotient, why it estimates eigenvalues so accurately, how it connects to the power method and inverse iteration, and why it forms the foundation of modern eigenvalue algorithms. Ends with a natural transition to the QR algorithm.
2025-10-08

8.1 Power Method and Inverse Iteration

A clear, practical, and intuitive explanation of the power method and inverse iteration for computing eigenvalues. Covers dominance, repeated multiplication, shifted inverse iteration, and real applications in ML, PCA, and large-scale systems. Smoothly introduces the Rayleigh quotient.
2025-10-07

Chapter 8 — Eigenvalues and Eigenvectors

A deep, intuitive introduction to eigenvalues and eigenvectors for engineers and practitioners. Explains why spectral methods matter, where they appear in real systems, and how modern numerical algorithms compute eigenvalues efficiently. Leads naturally into the power method and inverse iteration.
2025-10-06

7.4 Why QR Is Often Preferred

An in-depth, accessible explanation of why QR decomposition is the preferred method for solving least squares problems and ensuring numerical stability. Covers orthogonality, rank deficiency, Householder reflections, and the broader role of QR in scientific computing, with a smooth transition into eigenvalues and eigenvectors.
2025-10-05

7.3 Least Squares Problems

A clear, intuitive, book-length explanation of least squares problems, including the geometry, normal equations, QR decomposition, and SVD. Learn why least-squares solutions are central to ML and data science, and why QR provides a stable foundation for practical algorithms.
2025-10-04

7.2 Householder Reflections

A clear, intuitive, book-length explanation of Householder reflections and why they form the foundation of modern QR decomposition. Learn how reflections overcome the numerical instability of Gram–Schmidt and enable stable least-squares solutions across ML, statistics, and scientific computing.
2025-10-03

7.1 Gram–Schmidt and Modified GS

A clear, practical, book-length explanation of Gram–Schmidt and Modified Gram–Schmidt, why classical GS fails in floating-point arithmetic, how MGS improves stability, and why real numerical systems eventually rely on Householder reflections. Ideal for ML engineers, data scientists, and numerical computing practitioners.
2025-10-02

Chapter 7 — QR Decomposition

A deep, intuitive introduction to QR decomposition, explaining why orthogonality and numerical stability make QR essential for least squares, regression, kernel methods, and large-scale computation. Covers Gram–Schmidt, Modified GS, Householder reflections, and why QR is often preferred over LU and normal equations.
2025-10-01

6.3 Applications in ML, Statistics, and Kernel Methods

A deep, intuitive explanation of how Cholesky decomposition powers real machine learning and statistical systems—from Gaussian processes and Bayesian inference to kernel methods, Kalman filters, covariance modeling, and quadratic optimization. Understand why Cholesky is essential for stability, speed, and large-scale computation.
2025-09-30

6.2 Memory Advantages

A detailed, intuitive explanation of why Cholesky decomposition uses half the memory of LU decomposition, how memory locality accelerates computation, and why this efficiency makes Cholesky essential for large-scale machine learning, kernel methods, and statistical modeling.
2025-09-29

6.1 SPD Matrices and Why They Matter

A deep, intuitive explanation of symmetric positive definite (SPD) matrices and why they are essential in machine learning, statistics, optimization, and numerical computation. Covers geometry, stability, covariance, kernels, Hessians, and how SPD structure enables efficient Cholesky decomposition.
2025-09-28

Chapter 6 — Cholesky Decomposition

A deep, narrative-driven introduction to Cholesky decomposition explaining why symmetric positive definite matrices dominate real computation. Covers structure, stability, performance, and the role of Cholesky in ML, statistics, and optimization.
2025-09-27

5.4 Practical Examples

Hands-on LU decomposition examples using NumPy and LAPACK. Learn how pivoting, numerical stability, singular matrices, and performance optimization work in real systems, with clear Python code and practical insights.
2025-09-26

5.3 LU in NumPy and LAPACK

A practical, in-depth guide to how LU decomposition is implemented in NumPy and LAPACK. Learn about partial pivoting, blocked algorithms, BLAS optimization, error handling, and how modern numerical libraries achieve both speed and stability.
2025-09-25

5.2 Numerical Pitfalls

A deep, accessible explanation of the numerical pitfalls in LU decomposition. Learn about growth factors, tiny pivots, rounding errors, catastrophic cancellation, ill-conditioning, and why LU may silently produce incorrect results without proper pivoting and numerical care.
2025-09-24

5.1 LU with and without Pivoting

A clear and practical explanation of LU decomposition with and without pivoting. Learn why pivoting is essential, how partial and complete pivoting work, where no-pivot LU fails, and why modern numerical libraries rely on pivoted LU for stability.
2025-09-23

Chapter 5 — LU Decomposition

An in-depth, accessible introduction to LU decomposition—why it matters, how it improves on Gaussian elimination, where pivoting fits in, and what modern numerical libraries like NumPy and LAPACK do under the hood. Includes a guide to stability, practical applications, and a smooth transition into LU with and without pivoting.
2025-09-22

4.4 When Elimination Fails

An in-depth, practical explanation of why Gaussian elimination fails in real numerical systems—covering zero pivots, instability, ill-conditioning, catastrophic cancellation, and singular matrices—and how these failures motivate the move to LU decomposition.
2025-09-21