Skip to content

2026's Project Sandbox: AI, ML, agentic coding, RAG, cloud infra, and the occasional shiny object.

License

Notifications You must be signed in to change notification settings

radioastronomyio/year-of-code-2026

Repository files navigation

๐Ÿงช Year of Code 2026

Personal learning repository tracking AI-assisted development across the full spectrumโ€”from autocomplete to autonomous agents.

MIT License Projects

Python TypeScript PyTorch Gymnasium AWS React PostgreSQL Neo4j pgvector


Repo Banner

๐Ÿ“– About

AI-assisted coding has moved well beyond autocomplete. The tooling landscape now spans IDE completions, chat-based problem solving, boilerplating, code generation, and spec-driven autonomous development. Tools like Kiro, Kilo Code, Cline, and others are pushing toward agents that can plan, execute, and iterate with decreasing human intervention.

That said, the modelsโ€”even with the impressive coding skills of the newest generationโ€”still require human review, domain expertise, and polish. Production code needs coders who understand what they're building. Even the most capable autonomous agents work best with a human in the loop for architecture decisions, edge cases, and quality gates.

The skeptics who dismissed AI coding entirely are now using autocomplete; the early adopters are experimenting with spec-driven workflows. The whole field is moving, and it's moving fast.

This repository is my personal focus for 2026: building fluency across these tools as they evolve. It consolidates projects that would otherwise sprawl into separate reposโ€”learning vehicles, experiments, and practical builds that don't yet warrant their own home. When something graduates to "project worthy," it moves out. Everything else lives here.

Looking back at this repo in December 2026 should be an interesting view of how the year unfolded.


๐Ÿ“‚ Repository Structure

year-of-code-2026/
โ”œโ”€โ”€ ai-agentic-coding/           # Spec-driven development and CLI agent experiments
โ”œโ”€โ”€ ai-assisted-research/        # NSB methodology for AI research agents
โ”œโ”€โ”€ ai-taskmaster/               # AI competitive game show concept
โ”œโ”€โ”€ assets/                      # Repository-level images and banners
โ”œโ”€โ”€ aws-space-weather-dashboard/ # Serverless NOAA data application
โ”œโ”€โ”€ docs/                        # Documentation standards and templates
โ”œโ”€โ”€ ml-gymnasium-explainable-ai/ # Explainable RL with Gymnasium environments
โ”œโ”€โ”€ ml-orbweaver-roguelite/      # Monte Carlo roguelite mechanics validation
โ”œโ”€โ”€ work-logs/                   # Milestone worklogs and weekly rollups
โ”œโ”€โ”€ CONTRIBUTING.md
โ”œโ”€โ”€ LICENSE
โ””โ”€โ”€ README.md

๐Ÿ“Š Project Dashboard

Project Domain Status Progress Current Phase
aws-space-weather-dashboard AWS/Serverless ๐Ÿ”„ Active Spec 01/11 Identity & Auth
ml-orbweaver-roguelite ML/Game Design โœ… Sweep Ready Engine Complete Production Sweeps
ml-gymnasium-explainable-ai ML/Education ๐Ÿ”„ Active env01 Complete YouTube Production
ai-assisted-research Methodology ๐Ÿ”„ Active v0.4 Complete Example Prompts
ai-agentic-coding Tooling ๐Ÿ“ Ongoing โ€” Collection
ai-taskmaster Creative โธ๏ธ Blocked Concept Complete TTS Selection

Legend: ๐Ÿ”„ Active Development | โœ… Milestone Complete | ๐Ÿ“ Ongoing Collection | โธ๏ธ Blocked


๐Ÿ“‚ Projects

๐Ÿ›ฐ๏ธ aws-space-weather-dashboard

alt text

A full-stack serverless application that ingests NOAA GOES satellite data, stores it in a dual-database architecture (DynamoDB for fast access, RDS PostgreSQL for historical queries), and serves it through an authenticated API to a React dashboard. The real purpose is learning production AWS patterns by building something realโ€”the stack mirrors a customer engagement I'm preparing for, with the same services, architecture decisions, and operational concerns.

The project follows a 12-spec sequence from foundation through observability, built entirely through Kiro's spec-driven autonomous workflow with the constraint of never manually touching the AWS console. Spec 00 (Foundation) established CloudTrail, IAM baseline, and cross-stack sharing patterns that carry through the remaining specs.

Stack: CDK TypeScript Lambda API Gateway Cognito DynamoDB RDS S3 EventBridge React Material UI

AI Tooling: Claude Opus 4.5, Kiro IDE, AWS MCP

Status: ๐Ÿ”„ Active (Spec 01/11 โ€” Identity & Auth)


๐Ÿ•ท๏ธ ml-orbweaver-roguelite

orb-weaver

A headless Python simulation validating roguelite game mechanics through Monte Carlo analysis before committing to Godot engine implementation. The core thesis: every bullet-hell roguelite ends the same wayโ€”you're cornered, you plant your feet, you fight until you win or die. This simulation skips to that end state and asks whether strategic decisions alone (via a "hate draft" mechanic where player and room draft from a shared modifier pool) can carry engaging gameplay.

The simulation runs 10K+ games with deterministic seeding, tracking win rates, wave survival curves, and draft decision patterns. A complete parameter sweep engine using Sobol sequence sampling and Wilson CI adaptive stopping rules enables exploration of the viable parameter space. The current v5_lowturret baseline achieves 53.5% player win rate with a smooth difficulty curve peaking at wave 10.

Stack: Python 3.11+ dataclasses CSV scipy numpy

AI Tooling: Kiro IDE (~$2.78 total for simulation + sweep engine)

Status: โœ… Sweep Ready (v5_lowturret baseline, sweep engine complete)


gymnasium-explainable-ai

Companion code for a YouTube series on explainable reinforcement learning, focused on understanding why agents make decisions rather than just that they work. The approach: train agents on Gymnasium environments, then crack them open with Q-value overlays, policy inspection, and decision visualizations. Each environment gets its own mini-series on ClusterForge ML.

The xrl library provides reusable inspection and visualization tools across environmentsโ€”DQN Q-value extraction, advantage calculation, HUD overlays, and Q-bar renderers. CartPole (env01) is complete with training scripts, explained video recording, and analysis outputs. The pattern extends to Lunar Lander, MiniGrid, and eventually continuous control environments.

Stack: Python 3.10+ Gymnasium Stable-Baselines3 PyTorch OpenCV

AI Tooling: Claude Opus 4.5

Status: ๐Ÿ”„ Active (env01-cartpole complete, YouTube production pending)


ai-assisted-research

Methodology for getting better outputs from AI research agents, developed through 1000+ Gemini Deep Research sessions. The core framework is Negative Space Bounding (NSB)โ€”a three-layer constraint architecture that collapses the search space before agentic iteration begins rather than filtering sprawl afterward.

The layered approach uses Anchors (immutable physical/economic context), Walls (hard exclusions that prevent search vectors from being generated), and Gates (conditional logic that evaluates utility during synthesis). The v0.4 methodology synthesizes earlier walls-only and gates-only approaches into an integrated architecture that transfers to any agentic research system.

Stack: Gemini Deep Research, prompt engineering

AI Tooling: Gemini 2.5/3.0 Deep Research

Status: ๐Ÿ”„ Active (v0.4 methodology complete, example prompts pending)


ai-agentic-coding

Catch-all for experiments, observations, and one-offs from the spec-driven and autonomous end of AI-assisted development. The territory between "AI autocomplete" and "fully autonomous agent" is wide and weirdโ€”this folder tracks working in it. Content includes spec-driven development patterns with Kiro, CLI agent sessions with Claude Code and Gemini CLI, workflow experiments, and methodology notes on prompt engineering and human-in-the-loop patterns.

The larger projects in this repository (aws-space-weather-dashboard, ml-orbweaver-roguelite) use agentic tooling extensively; this folder captures meta-level learnings from that work. CodeRabbit was selected as the primary AI code review integration after evaluating several options.

Stack: Various

AI Tooling: Kiro, Claude Code, Codex CLI, Gemini CLI, Kilo Code, Cline

Status: ๐Ÿ“ Ongoing collection


๐ŸŽญ ai-taskmaster

ai-taskmaster

A competitive game show format where frontier AI models compete in ambiguous, bespoke tasksโ€”judged by an imperious Taskmaster (Claude Opus 4.5). The innovation is a dual-voice system: each AI contestant has both a "table voice" (what they say to other contestants, strategic and performative) and an "audience aside" (fourth-wall address revealing true strategy). The comedy emerges from the gap between public performance and private commentary.

Think Taskmaster meets QIโ€”four AI models (Claude Sonnet, GPT-5, Gemini 3 Pro, rotating fourth) competing in tasks like "make the Taskmaster laugh in exactly 17 words" or "convince the judges you're the only real AI." Production approach uses visual novel layout with TTS voices and generated scene images. Concept doc and character avatars are complete; blocked on TTS voice synthesis selection.

Stack: Visual novel layout, TTS (pending), image generation

AI Tooling: Multiple frontier models as contestants, Claude Opus 4.5 as Taskmaster

Status: โธ๏ธ Concept complete, blocked on TTS selection


๐Ÿ“‹ Repository Conventions

Naming: Projects are prefixed by domainโ€”aws-, gcp-, azure-, ml-, ai-, etc.

Structure: Each project maintains its own README with goals, approach, and status.

Memory Banks: Projects use .kilocode/rules/memory-bank/ for persistent contextโ€”brief, architecture, product, tech, context, and tasks files that AI agents load at session start.

Promotion: Projects graduate to their own repository when they become "project worthy"โ€”a judgment call based on scope, external collaboration, or standalone value. Size alone doesn't determine this.

Sensitive Data: This is a public repo. Credentials, secrets, and sensitive configuration live in a separate private repository with manual coordination.


๐Ÿค OSS Program Acknowledgments

This repository benefits from open source programs that provide free or discounted tooling to qualifying public repositories. These aren't direct sponsorshipsโ€”they're programs that support the open source ecosystem, often auto-approving based on repository metrics. The companies may not know this specific repo exists.

Primary (Active Use)

Program Provides Use Case
CodeRabbit AI code review (Pro tier) PR review with codebase context, CLI integration with Claude Code and Gemini CLI
Atlassian Jira, Confluence, Bitbucket Pipelines (Standard tier) Project tracking, documentation, CI/CD

Secondary (Available, Wave 2)

Program Provides Planned Use
Snyk Security scanning Dependency vulnerability detection
SonarCloud Code quality analysis Static analysis, code smell detection
Sentry Error tracking Runtime error monitoring
Datadog Observability Metrics, logs, APM

๐Ÿ“„ License

MIT ยฉ 2025 VintageDon


๐Ÿค Contributing

Some projects in this repository accept contributions. See CONTRIBUTING.md for which projects are open and how to submit a PR.

AI-assisted contributions are welcomeโ€”this repo is about exploring those tools, after all. But your code must be good, and you must understand what you're submitting.


๐Ÿ“ž Contact

GitHub: @vintagedon

About

2026's Project Sandbox: AI, ML, agentic coding, RAG, cloud infra, and the occasional shiny object.

Topics

Resources

License

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published