AI

2026 May Coding Agents Landscape

May 3, 2026 AI, AI Agent No comments

This post is mainly to take a snapshot of AI Coding Agents on the market as of May 2026. There is very little I’ve done to write this post but it serves its purpose. Besides it is very interesting.

This below was prepared with help of Claude:

If it doesn’t render, you might need to open the this link directly.


No comments


Vancouver Sun Run 2026 X-Ray

April 26, 2026 AI, Vibe Coding No comments

I ran Vancouver Sun Run for the 5th time this year. This is 10k race in Vancouver, roughly 30k-60k people run this race. My official time is 50min:17sec. This isn’t my personal record (48:17), but given I didn’t train for running this is fairly decent result (probably my 3rd best 10k run).

I just vibe-coded this Vancouver Sun Run X-Ray app with two tabs. The Race Map simulates all 45,013 finishers moving along the official course. You can drag the time slider and watch the people spread out. Upload your GPX and your position appears as an orange dot. The secon tab Public Results digs into the real data corral effects, finish time distributions by age group, city breakdowns.

The race data comes from a community-parsed CSV of the official Sportstats results.

I ran the race, got home, opened Claude Code, and had a working app before dinner. That’s the pattern with vibe coding: you start with curiosity, not a spec.

Live: vancouversunrunxray.streamlit.app

GitHub: github.com/andriybuday/vancouversunrunxray

Before next Sun Run I can probably vibe-code something more viral and vibrant so that people on the run can play with it and have run. At very least I have this in mind for my 2027 run (which I already signed up for).


No comments


Conway’s Game of Life, Coded From a Mall

March 29, 2026 AI, AI Agent No comments

When I was in university I coded “Conway’s Game of Life”, which is a primitive simulation of cellular life, if there are a certain number of live cells around they produce more cells and if conditions are not favorable cells die off. I don’t know what the meaning of life is. Would at some point the argument come to say that AI is alive? This is very philosophical, this post is instead a very practical showcase of some of the most recent tools and advances in AI as I play with them.

Just have a look at this photo:

As you can see I’m somewhere in a mall, asking my Claude Cowork running on my mac at home to create a folder ~/Projects/gameoflifecowork and generate a single HTML page with Conway’s game of life implementation. I came home and there it was an html page right in that folder with the implementation. It is perfectly working. I’m adding it here. If you are reading from e-mail you would need to open the blog to see it in action.

[if you are reading from e-mail you might need to open in browser to see The Game Of Life]

Cowork is an extremely powerful (and dangerous tool). It is not yet a very smooth experience. Sometimes I need it to prompt multiple times, most of the time there is no proper feedback in the Mobile app so I don’t know what is happening. For instance, I also asked it to go to my “Downloads” folder and locate any concert tickets and tell me how much I paid for them. Unfortunately there was no way to see on the mobile app the answer to the tickets (not in any chat), but when I came home I saw a dedicated chat open on my computer that had the answer.

Under the hood Cowork runs Claude Code to implement the game of life, so I was wondering if I can compare different models and how good of a job they do, so I installed Open Code and connected Gemini API, Anthropic API, and also local LLM!

With Gemini API I generated a really great fast full screen Game Of Life, which cost me about 0.20$. Local llama3.1 unfortunately is not suitable for this task, I had to give it many additional instructions and it messed up every time until in the end I got an empty html file with some broken functionality, which I fixed with Copilot just to get it render:

Screenshot Lllama3.1 implementation of GameOfLife (with lots of manual effort)

Gemini’s Game of Life was full screen and rendered perfectly:

Gemini PRO Game Of Life

I then switched to Claude Haiku 4.5 to generate the Game of Life you can see above.

This implementation cost me about 0.33$. 

What was a small mini-assignment at university to code the Game Of Life, which I did with C++ and probably took me few days, now turns out to be just 0.20$-0.30$ throwaway code just to test different API integrations. Right now I understand why those cells live and die as I wrote this algorithm myself in C++ but I’m wondering if at one point we will not know what is happening inside implementations generated by AI.

Life remains for the most part a mystery to science, though we are getting closer and closer to understanding how it works. At the same time a reverse is happening with AI we are slowly getting further and further from understanding what goes into the beautiful implementation of Game of Life.


No comments


Completely Sold On AI

March 22, 2026 AI, AI Agent, Personal No comments

Today I’ve done multiple personal things that fully sold me on AI. No way back. Sorry.

Use Case 1: Complex Tax Filing

No question using AI for Tax purposes is useful, but today I got it to the next level. To understand my tax situation: I moved from Canada to the US in 2025, I have rental property, stocks bought and sold in both Canada and US, retirement accounts in both countries, stocks/ETFs transferred from different institutions back and forth, etc. With AI handling all of this is basically avoiding nightmare and frustration taking many days on.

For specific example, I had VFV.TO which is ETF traded on Toronto stock exchange, not only I got some purchases and sales starting 2021 they tracked through multiple brokerages and income from dividends was reinvested back into shares, not only it is over many years, banks, currencies, countries, but also dozens of statements and documents, it is also treated in special way in USA where they treat it as Passive Foreign Investment Company (PFIC). Other than that the day I moved countries I “deemed disposition” of my assets. If I had to do all these things manually I would just get petrified, but instead I can just throw about 30 documents at AI and ask it to generate a final spreadsheet with all of the transactions, and also have source documents referenced so I can double-check. It all worked really-really well.

Use Case 2: Organizing My Files since 2000

Yes, I crossed the line and allowed Claude Cowork access to my local files.

I keep files dating from the early 2000s, keeping most of the things archived in Dropbox. I would keep such things as any of my university course works, archives of code snippets I have written in 2005, archives of my blog, some old presentations, tax documents, all kinds of agreements, all kinds of things that basically track my life since it became digital. It grew over many years. I tried to organize things in some iterations in the past, but then mostly gave up on just naming things more nicely and using search. But with AI I was just able to say “Create a plan to organize files in this folder X”, then iterate over the plan and let it execute things. The best part is that this is just natural language.

Use Case 3: Organizing Life

I previously posted that I do track things in life and that I use AI running chats for each area of life, like “Health”, “Finance”, “Career”, etc. This is all good, but then I wanted to build my personal AI agent to help me with tracking, but tools like OpenClaw and Claude Cowork made that unnecessary. Now, I’m playing with migrating my life planning into simple .md files. My life is essentially one repository that I can query and manage with natural language. Nuts.

Use Case 4: Posting on Instagram

Haven’t posted my climbing videos on instagram for a while, but I thought maybe I would post one latest video. I thought that analyzing video is not great with LLMs, but no. I uploaded my climb, it recognized that I was wearing martial arts t-shirt, it recognized at what points major moves where happening and I asked it to suggest music that would work well for the video matching my style but also the pace of events in the video. This is nuts. As next step I was thinking of fully automating video edit and posting on my behalf. Nuts.

Use Case 5: Work

While I cannot talk too much about work, I would just say that there was a step-function improvement in AI use and productivity. Producing code is faster, iterating over ideas is faster, getting things out just gets faster. I need to constantly adapt so that I don’t become a dinosaur that dies with a fallen AI asteroid.

Conclusion

It appears I grew from an AI skeptic, to AI learner, to fully embracing it by now. No way back.

edited with Claude Cowork Computer Use from mobile phone


No comments


Surviving the AI Asteroid: Hype Cycles, Overreactions, and What Comes Next

March 15, 2026 AI, AI Agent, Opinion No comments

Previously I compared AI to an asteroid, with the simple premise that if your job is boilerplate CRUD, your job is a dinosaur, and if your job involves high leverage of AI, including building AI-integrated systems, your job is the mammal that survives.

AI Hype is real, it is everywhere, and honestly is somewhat annoying. My LinkedIn feed is oversaturated with all of the AI noise, I keep overhearing “AI” when walking past random people on the street. Everyone has their say on AI, including me. AI washing is a real marketing tactic used by many companies. Many people would get into AI just purely because of FOMO. And, honestly, that fear is justified, because what if you are really missing something and will be left behind. No one wants to be left behind. As a simple example, it is even hard to get a Mac mini because everyone is buying them to run their personal OpenClaw. I’m playing this game as well. I did set up OpenClaw on Docker just to see what I could do, but until I have a sustained workflow, I won’t be buying a dedicated machine. It’s easy to confuse playing with new tools for actual productivity. But, maybe, I’m missing out.

There is no smoke if there is no fire. Last week Anthropic released a labor market impact study claiming that hiring has slowed in highly AI-exposed roles since ChatGPT launch. For us, software engineers, the study claims that AI can theoretically automate about 90% of our jobs and it appears current automation is only at about 30%. If this is true and if this is happening soon, some kind of a combination of the two will happen: 1) our jobs will transform by a lot so that we are building ever new and more complex things that AI cannot and/or 2) there will be significant reductions in software engineering jobs. I don’t know if 1 or 2 would be a larger component of transformation but we should be preparing for both!

Image credit: Anthropic https://www.anthropic.com/research/labor-market-impacts 

As I use Claude Code, it becomes apparent how it becomes more and more capable over time. It is no longer a question of whether the threat is real. It is absolutely real. The asteroid is here! It has hit the ground already. The transformation is already happening and if you don’t see it, you might be in trouble (that is unless you are a plumber or someone with a low exposure job). I still see value in myself by figuring out what problems to solve and then directing the work to my AI agents, sometimes finding myself directing 4 of them at once, which is really cool to be able to make progress on 4 things at once, but it is also terrifying. Does it mean I’m now 4x productive? Does it mean we need ¼ of engineers to do the same job? Or we will simply see another instance of the Jevons Paradox. Historically, making software development cheaper/faster didn’t mean we hired fewer engineers. What we have seen is that demand for more software has increased thus increasing the number of software engineers. But still, there are so many open questions that come with this transformation: like

  • where do we get senior engineers if there is no need for juniors?
  • would we ever reach “lights out” codebases not requiring human code reviews?
  • would we hit the energy limits when it is more cost effective to pay a biological human rather than energy consuming AI? … and so many questions.
  • what’s your question?

What goes up will go down. Things happen in cycles. What is born will die. I made the mistake of buying real estate in Vancouver at the peak of the market in early 2022, my property price has not yet recovered. Do you remember the COVID tech hiring frenzy followed by layoffs? Our industry over-hired only to lay off people after that. Do you remember the 2008 financial crisis and what happened to real estate prices only to recover some years after that? I am not old enough to remember the dot com bubble but it is all the same all across. My argument here is that we will see some kind of cycle of overreaction with AI as well. Some companies will over-invest in AI and not get anything out of it. Some companies will lay off too many people only to hire back. What we are looking at are micro movements, but what is more interesting is what will this bring us long term, what is the macro movement here?

Image credit: Gemini on my prompt. The image is too colorful for my taste but it is kind of fun.

Dear reader, prepare for multiple outcomes. They say to have peace, you must prepare for war. Build a strong financial safety net, constantly stay on the lookout for what is changing, and adapt relentlessly. Every technological shift creates massive, unseen upsides. The rules of the game are changing and instead of panicking, the goal is to understand the new rules so you can keep playing.


No comments


Surviving the Half-Baked AI Hype

March 1, 2026 AI No comments

I was looking at my social feeds recently, and it feels like everyone is suddenly an AI productivity guru or all knowledgeable about AI (ironically I write about AI a lot as well). People are just throwing together these half-baked agents and AI solutions, not even checking if they solve a real problem or if any quality is there, and bragging about inventing some “10x workflow”. Well, not really bragging themselves, but rather asking AI to brag about it, which makes it even worse.

I believe this year is just so much FOMO and oversaturation fatigue for all of us, software engineers. I no longer know where it is worth spending my time. Last weekend I spent some time setting up OpenClaw because that seems to be a hot thing right now. Before that I was vibe coding different things, playing with LLM integrations, Agents, tools or whatever the latest cool thing was. You can spend lots of time learning a tool or an approach, and a month or two later, it’s completely obsolete because the next “best thing” just dropped. It’s overwhelming. For the most part, it is all just noise. It is increasingly more and more difficult to figure out what the signal is. The signal-to-noise ratio just went really really bad. These days whenever I see a post by someone I try to quickly gauge if that is typical AI text and I mostly ignore it in those cases, if not, I try to see if there is some substance to whatever is written and if there are any opinions expressed, if so, that seems to be a genuine piece of work and it draws my attention. I am starting to develop an allergy to AI generated text.

I am not an AI denier. It is extremely useful and great but the hype is just over the board. What goes up will inevitably settle down, and we just need to figure out how to ride the waves. I’ve written about this in the past. The tools will inevitably change but the underlying shift in our industry is permanent. Our software engineering jobs are destined to change. There is no question about that. There is also a lot of uncertainty over which other jobs will be displaced by AI. With the current trends, it looks like anything that has to do with text and image processing can be replaced and anything that has anything to do with operating in the physical world (think plumbing) or requiring human judgment might take longer to get replaced. I spoke to some of my non-tech friends and they express fear of being affected by AI as well.

All I can say for now is that we need to keep adapting to remain relevant. So while I don’t like all of the hype, if I don’t sample around, try things out I might miss on one of the things that wasn’t a hype and be left behind as the industry moves forward.


No comments


My 1-Hour Open Claw Setup: Docker, Llama 3.1, and Telegram

February 22, 2026 AI, AI Agent, HowTo No comments

I saw all the fuss about Open Claw online and then spoke to a colleague and she was saying she is buying a Mac mini to run Open Claw locally. I could not resist the temptation to give it a try and see how far I can get. This post is just a quick documenting what I was able to do in like one hour of setup.

If you’re like me and find it difficult to follow all the latest AI hype and missed it, Open Claw is an open-source AI agent framework that connects large language models directly to your local machine, allowing them to execute commands and automate workflows right from your terminal or your phone.

A quick preview below. This is just nuts. In one hour I was able to run OpenClaw on Docker talking to llama3.1 running locally and communicating with this via Telegram bot from my phone 🤯.

Back in the old days I would open some kind of documentation and follow steps one by one and unquestionably get stuck somewhere. This time I started with Gemini chat prompting it to guide me through the installation and configuration process. This proved to be the best and quickest way.

Decision 1: Running locally or in some isolation.

I think this one is an obvious choice. Giving hallucinating LLMs permission to modify files on my primary laptop sounds like a recipe for disaster. Decided to go with Docker container but if I find the right workflows I might buy Mac mini as well. 

Commands were fairly simple, something along these lines:

git clone https://github.com/openclaw/openclaw.git
cd openclaw
./docker-setup.sh
docker compose up -d openclaw-gateway

Decision 2: Local LLM or Connect to something (Gemini, OpenAI)

This is a more difficult decision to make. Even though I’m running an M4 with 32GB, I cannot run too large of a model. From reading online it is obvious that connecting to large LLMs has an advantage of not hallucinating and giving best results but at the same time you’ve got to share your info with it and run the risk of running into huge bills on token usage. Since this was purely for my self learning and I don’t yet have good workflows to run, I just decided to connect using a small model llama.3.1 running via Ollama. Since it was running on my local machine and not docker, I had to play a bit with configuration files but it worked just fine. And yeah, the answers I would get are really silly.

Later I found that ClawRouter is the best path forward. Basically you use a combination of locally run LLM and large LLMs you connect to. I might do this in the next iteration.

Decision 3: Skills, Tools, Workflows

This is just insane how many things are available. Because this can run any bash (yeah, in your telegram you can say “/bash rm x.files” – scary as hell) on the local the capabilities for automation with LLMs are almost limitless.

Conclusion

I can barely keep up with all of the innovations that are happening in the AI space but they are awesome and I’m inspired by the people who build them and feel like I want to vibe code so much more instead of spending my time filling-in my complex cross-border tax forms over the weekend.


No comments


From Alien Translators to Human Translators

February 15, 2026 AI, Opinion No comments

Just wanted to share some quick thoughts on AI again. It does change our jobs (see my earlier thoughts “AI is asteroid and your tech job is a dino”, “Is AI redefining software craftsmanship?”). It completely rewrites how much can be done in a short period of time (see a bunch of vibe coding posts from me: blogger agent, typing game, AI powered snake, etc). And while I expressed some doubts and expected a ceiling to its advancements, I am now more deeply convinced that the time to fully embrace AI is now. Almost any knowledge work you do with your brain can get some help from AI. And while I advocate for limited use of AI in writing (“Don’t outsource your thinking to AI”) it has undeniably changed how I do my writing and what value I think I bring or don’t bring. Writing generic advice, anything that can be searched online, is an absolute waste of time, unless it is supplemented with opinions or experiences. Writing coding blog posts with technical details, as I used to do in the past, is also worthless. The entire stack overflow is now not receiving much traffic.

I liked to think about Software Engineers as these super smart almost alien-like translators. We used to translate requirements and business needs into cryptic code that most people couldn’t understand, just to make the software work. While fundamental knowledge is still relevant and our role as translators still remains, the destination language is changing to be more English-like. Instead of typing code we orchestrate AI work. What still matters is what AI cannot do and is very unlikely to be able to do soon, which is doing human things. The things that revolve around judgment, our lived experience, and our authentic connections.

An LLM can write technical documentation, generate a summary, and write lots of code. It works perfectly for transfer of knowledge, but it still is not good at transfer of experience and understanding what we really need and mean as humans. Translators are still needed, but instead of being more alien-like we might need to be more human-like and do more human things.

P.S. I resisted the urge to use AI for this blog post.


No comments


Built a personal “AI Slop” Generator in 3 Hours

February 8, 2026 AI, Opinion No comments

Presenting you with AI Slop by Andriy Buday and Gemini: https://aislop.andriybuday.com 

I was recently challenged on why my weekly blog posts are not written by AI. I do have my strong opinions on this and arguments against it but before I delve into them I wanted to accept the challenge. So in about 3 hours of vibe coding I built an automated GitHub and Google Gemini powered workflow that picks either an idea from ideas.md file or one of my older blog posts on this website and (re-)writes a new blog post based on that and then uploads it to my dedicated aislop subdomain.

Solution based on GitHub Actions, FTP, Google Gemini

The entire project took about 3 hours from initial concept to deployment. This was pure vibe coding of ~40 git commits, a bit of setup in my bluehost, and some setup on github.

I learned about GitHub Actions fairly recently, but basically you can build a workflow based on yaml definition that would be triggered on a periodic basis. Additionally you can put your secrets into GitHub repo configuration. I placed my Gemini API key into secrets as well as I then placed my FTP access details (yes, I know it’s insecure and old school, but this is a 3-hour hack project). For FTP I created a dedicated account and only allowed a specific folder on my bluehost, where I also created a subdomain.

How it works: The workflow and Tech Specs

I asked Claude to summarize the technical details because this is what AI shines at:

Workflow:

  • GitHub Actions → triggers workflow daily → Python script picks unprocessed blog post or idea from ideas_tracking.json or blog_posts.json → load content and prompt Gemini via API with custom prompt → postprocess text to HTML → create HTML file and update blog_posts.json (database) → connect to FTP and upload files → git push changes for tracking and backup

Core Development Phases

  • Scaffolding & Engine (75m): Establishing the repository and building the Python script to handle web scraping via BeautifulSoup4 and AI integration.
  • Automation & Queueing (35m): Configuring GitHub Actions for CI/CD and implementing a JSON-based status tracker to manage 370+ URLs.
  • Refinement & UI (80m): Enhancing Gemini’s prompting for authoritative content, fixing Markdown-to-HTML rendering bugs, and building a responsive progress dashboard.

Technical Stack Highlights

  • AI & Logic: Python 3.9+, Google Generative AI (Gemini 2.0 Flash), and markdown2
  • Infrastructure: GitHub Actions for scheduling, Bluehost via FTP, and GitHub Secrets for API key management
  • Frontend & Data: Vanilla HTML/CSS/JS for the dashboard, with JSON and CSV files handling all state-tracking without a database

Thoughts (Why I am against AI generated content)

Is this the future of blogging? Maybe. Is it a future I’m excited about? Not entirely. I am definitely not going to share my AI Slop sub-blog unless that is purely to prove the point. I can barely stand all of these huge walls of text that are clearly written by AI but presented as if humans had written it. Why would you read it? You can just prompt your favorite LLM to give you answers you really need. I almost want to vomit from all this clearly AI generated text with no personal substance or real opinions. Sorry for being this vivid, but again: AI would not write that it wants to vomit because of the text it has written. 

And just to be clear, I do use LLM as a tool to help with my writing, but just not to write instead of me: Don’t Outsource Your Thinking: Why I Write Instead of Prompt

So where does the value of blog posts come from?

In my opinion the value comes from giving your own perspective, from sharing your opinions, driving your own arguments, and, yes, while bloggers can and do use LLM to find blind spots and to arrive at a stronger argument, the arguments should still come from the author, otherwise it is all just crappy AI Slop (unless that was the intention originally).

My ‘AI Slop’ bot can publish 100 posts a day, but it can’t build its own perspective. It can only synthesise perspective based on data it has received before.

My concluding argument is that efficiency in generating text does not equal value in reading text.


No comments


Building Personal AI Agent

January 11, 2026 AI, AI Agent No comments

Rich people always had access to assistants (+chief of staff) that would help them with all kinds of chores, would advise them on things, or just do things behind the scenes. We live in an interesting time of AI, where anyone has access to these fabulous LLMs that can do some of those things for us. Like, I’m sure most of us are doing our travel plans with some help from LLMs, or we make buying decision, etc. It is just mind boggling how good these things are becoming!

In parallel, as software engineers, we keep hearing about AI agents all the time. We use AI agents at work. The most useful and prominent example of LLM agents are coding agents. You are likely using Claude Code at work. I already heavily rely on LLMs to track many of my personal goals, to critique me, to give suggestions, etc.

But what if I build a personal AI Agent to avoid repeating things and to make it watch me more proactively?

Vibe Coding AI Agent

There we go! Let’s build something simple first. My use cases for LLMs are fairly simple, nothing too crazy and very closely tied to my Life Goals and areas of life. For example, I have chats with Gemini labeled like “Nutrition”, “Finance”, “Career”, etc. When I eat my breakfast I snap a picture of it and estimate my nutrition intake. When I’m considering stock buying I do research with Gemini. When I plan a trip I build an itinerary with LLM, etc. I track my weekly progress in google docs. I track my finances in spreadsheets. The more I think about this the more I realize there is a room for a personal AI agent that is highly tuned to my personal needs and would orchestrate all of this. Additionally there won’t be any ready solution online, because this is so personal, so I’ve got to build one agent for myself!

Email Digest Agent after ~3 hours of Vibe Coding

So what I’ve built in 3 hours is a “personal AI agent that automatically generates daily digest emails by fetching data from multiple sources in parallel: it retrieves stock market insights for ~X tickers, extracts the current week’s goals (with checkbox tracking) from a tabbed Google Doc, pulls my monthly focus items, pulls net worth data from a Google Sheets dashboard, then uses Gemini to generate a professionally formatted HTML email summary and sends it via SendGrid. The system is built with LangGraph for workflow orchestration, uses OAuth2 for secure Google API access, and preserves formatting details like checkboxes (✅/⬜) by converting them to email-compatible emoji before direct insertion into the final email.”

Tech Stack

  • LangGraph 0.2+ – Workflow orchestration with parallel execution
  • LangChain 0.3+ & Gemini API (gemini-2.0-flash-exp) – LLM integration
  • Google Docs/Sheets API – Data sources with OAuth2
  • Alpha Vantage API
  • SendGrid – Email delivery
  • Python 3.9+ – Runtime

Conclusion

This project is just scratching the surface of what is possible to be built very quickly for personal needs. The most exciting realization for me wasn’t the technical implementation but how accessible it is to connect things together. Setting up all of the API keys and then vibe coding all together is so straight forward that it is just unbelievable.

I will continue this project next week to make it actually properly work for my needs and then will host it on some server to send me those digests. Next steps would be to supplement it with prompts to LLMs to give me quick ideas for what I should focus on, better tracking, etc so I can achieve my goals quicker. So exciting!

Go ahead and build something for yourself!


No comments