Normal view

Today — 9 March 2026Main stream

Why do developers write such terrible git commit messages? Genuine question

I've been going through some open source repos lately and the commit history is absolutely unreadable.

"fix bug", "update", "changes", "asdfgh", "ok now it works hopefully"

Like... this is code that other people have to maintain. How does this happen even in professional teams?

I'm curious do you actually care about commit quality at your job? Does your team enforce any standard? Or is it just accepted chaos?

And honestly what's your own commit message process like? Do you think about it or just type something fast and push?

submitted by /u/Existing_Round9756 to r/webdev
[link] [comments]

What makes a web dev ‘senior’ these days?

I’ve been coding for a few years, jumped from project to project, but honestly… I still feel like a junior sometimes. I see ‘senior’ devs and wonder is it years, skills, or just confidence? Someone please explain what really separates them nowadays with all the AI bubble getting more bigger.

submitted by /u/Professional_One3573 to r/webdev
[link] [comments]

I miss Flash. What an era...

I was just reminiscing today. I really miss flash games and that creative era. I know we have all the nice open standards now; canvas, webgl, js/ts game engine libraries. But there was something special about the tool itself, how available it was to creatives instead of just software developers. And the ability to export to a single artifact (SWF).

It would be wonderful if there were a similar program that exported to a single artifact that could be played in the browser with a JS/WASM runtime.

The key point is that the program was oriented towards creatives instead of just developers. Creatives don't really care about canvas/svg/etc.

Any thoughts?

submitted by /u/kizerkizer to r/webdev
[link] [comments]
Yesterday — 8 March 2026Main stream

Advice with my developer taking down our WordPress site.

Advice with my developer taking down our WordPress site.

Looking for advice for a problem happening with my developer. I got a email stating that there was an unusually high amount of resources being pulled from our site. We own a vintage jewelry sales website that was built and hosted by this developer. They stated that facebook bots were crawling our website, and causing resources to be pulled from other sites hosted on the same server. They recommended we purchase a dedicated server to host our site. After googling this we found that there should be a solution to create a rule to limit or block Facebook bots from crawling our site. We brought this to their attention, and they said they could implement this and bill us for a half hour of work. After the successfully implemented this they then took down our site saying that they had to do it as our site was bringing down their server. Trying to find out whats going on as it feels as though my site is being held hostage unless I purchase a dedicated server.

submitted by /u/reemo4580 to r/webdev
[link] [comments]

So one forgot something 😬 🤣

So one forgot something 😬 🤣

I was just going through netlify website to publish my portfolio project, but the name was not available, so out of curiosity i checked the url ans saw this🤣. Some one forgot he was working on something. The timer has gone in negative and counting is still going on.

submitted by /u/Rarararararaviiiiii to r/webdev
[link] [comments]

Appreciation for old school web dev

I just want to talk a bit about how we used to make websites, and how epic it is that it still works and is just as viable as ever 😄

I run a popular fan site for a TTRPG that's basically an anternative to DnD. Just for context, it gets about 30k visitors per month.

It's built almost entirely using good old HTML, a little connective PHP to separate components into files, a reasonable amount of vanilla CSS to make it neat and responsive, and a tiny sprinkling of vanilla JS to enable saving (into localstorage) for pages like the character sheet. No frameworks needed. And all the data is stored in markdown and json files, because I don't need a CMS at this stage.

Because it's basically entirely static pages, it's fast, secure, responsive and accessible by default 😀 And super easy to maintain of course.

I have nothing against frameworks of course (frontend, backend, etc.); they're amazing, and I'll probably have to rebuild this using one (or a CMS) in a few months' time. But they aren't always needed; especially when a website is still new and only has 1 contributor. Keep it simple, and sites start off great by default!

submitted by /u/Droces to r/webdev
[link] [comments]

Ban posts about AI

This subreddit is supposed to be about web development. But, lately, I've seen mostly posts about AI and its impact on web development. I get the relevance. I get the fear.

I'm sorry if this is inappropriate or against the rules. I recognize the irony of this post also not being about web development. But can we go back to sharing neat tricks and tips for building websites? And answering each other's questions about pieces of code that we used our brains to write?

Please?

submitted by /u/miniversal to r/webdev
[link] [comments]
Before yesterdayMain stream

Claude...

Claude...

After metas crawler sent 11 million requests. Claude has now topped the charts with 12m in the last 15 days alone. Meta is also completely ignoring robots given the 700k requests theyve sent regardless.

Here's the IP addresses hitting the hardest. 216.73.216.x is anthropics main aws crawler. Some interesting crawlers. Wtf is ripe? The 66.249.68.x seem to be some internal google one not related to search or maybe just some gcp based crawler.

requests requests
216.73.216.36 6,285,832
216.73.216.175 4,134,384
216.73.216.81 2,008,789
74.7.243.222 1,057,218
66.249.68.128 205,373
66.249.68.136 187,573
66.249.68.135 182,093
74.7.243.245 171,290
99.246.69.10 165,425
66.249.68.129 154,764
66.249.68.133 140,394

Anyone else seeing this? the vercel bill is completely fucked. first week in were at 500+ spend. 400+ is from function duration on programmatic SEO endpoints. The industries response has been to lick the boot of cloud providers as if they arent the ones funding this circular economy pyramid scheme bs. Throwing up some cloudflare WAF to block other computers from communicating is insane. yes we know vps is cheaper, not the point.

submitted by /u/cardogio to r/webdev
[link] [comments]

Showoff Saturday: I spent a weekend building a real-time meeting cost ticker instead of dealing with my actual meeting problem

Showoff Saturday: I spent a weekend building a real-time meeting cost ticker instead of dealing with my actual meeting problem

I'm an eng manager and tech lead. I have too many meetings. Instead of cancelling any of them like a normal person, I spent a weekend building a tool that shows what they cost in real-time. Classic engineer move.

It's Ash Flow (https://ashflow.app). You add people to a meeting by job title and country, and it pulls salaries from a database I built with 80+ roles across 30+ countries. Hit start and you get a live counter ticking up showing exactly how much money is being burned.

The whole point is the shareable URL. You drop it in the Zoom or MS teams chat or pull it up on the conference room TV. Sharing the link or your screen and showing this on the side. suddenly people starting getting to the point faster, or try to reduce meetings. Thats the idea at least. So far for me, its reduced number of meetings and wasted/dead meeting time.

Tech: Basically TanStack Start and Turso for the DB for the salary data. The shared/read-only view strips out individual salary numbers so you're not accidentally doxxing what people make or who they are. no names, just job titles.. Currency detection is automatic from browser locale, conversions come from ECB exhange rates.

The salary database was honestly the hardest part. Getting reasonable numbers for a Senior Software Engineer in Germany vs India vs Brazil, across 80+ titles, is a lot of spreadsheet work. I'm sure some of it is off, which is part of why I'm posting here.

if you have opinions about TanStack Start, I spent some time with this building various types of projects with it and have thoughts.

https://ashflow.app

submitted by /u/Alter_nayte to r/webdev
[link] [comments]

I've been building Tabularis — an open-source, cross-platform database client built with Tauri + React since late January. v0.9.6 just shipped, wanted to share.

I've been building Tabularis — an open-source, cross-platform database client built with Tauri + React since late January. v0.9.6 just shipped, wanted to share.

Hey,

I've been building Tabularis — an open-source, cross-platform database client built with Tauri 2 + React — since late January.

https://github.com/debba/tabularis

What it is: SQL editor, data grid, schema management, ER diagrams, SSH tunneling, split view, visual query builder, AI assistant (OpenAI/Anthropic/Ollama), MCP server.

Runs on Windows, macOS, Linux.

The interesting Rust bit: database drivers run as external processes over JSON-RPC 2.0 stdin/stdout — language-agnostic, process-isolated, hot-installable.

We already have plugins for DuckDB, Redis and working on MongoDB and Clickhouse .

Five weeks old, rough edges exist, but the architecture is solidifying.

Happy to answer questions about technical specific choices.

Stars and feedback very welcome 🙏

submitted by /u/debba_ to r/webdev
[link] [comments]

We built an open-source alternative for website analytics

We built an open-source alternative for website analytics

Hey r/webdev,

Over the past year our small team built an analytics platform from scratch to explore high-performance event ingestion and analytical workloads.

Instead of extending an existing solution, we wanted to experiment with the architecture ourselves and see how far we could push performance and efficiency.

The backend is written in Rust and uses ClickHouse as the OLAP database for storing and querying event data. The project is open source and can be self-hosted. Most of our work went into ingestion throughput, schema design, and query optimization for large event datasets.

Over time we also added uptime monitoring and keyword tracking so traffic analytics and basic site health metrics can live in the same stack instead of being spread across multiple tools.

Our team is small (three developers), and we actively use and maintain the platform ourselves.

GitHub:
https://github.com/betterlytics/betterlytics

Demo:
https://betterlytics.io/demo

Curious what other developers think. Feedback or criticism is very welcome.

submitted by /u/WeatherD00d to r/webdev
[link] [comments]

We let strangers merge code to a live site. The community spent weeks debugging why the merge bot couldn't merge their PRs.

We let strangers merge code to a live site. The community spent weeks debugging why the merge bot couldn't merge their PRs.

OpenChaos is a repo where anyone submits a PR, the community votes with reactions, and the most-voted PR gets merged. The code IS the website - every merge changes what you see at openchaos.dev.

A contributor built the automerge bot from scratch. It ranks PRs by votes, checks CI, verifies rhyming titles (yes, PR titles must rhyme to merge), and merges the winner. The community then spent weeks fixing bugs in it:

  • Feb 21: "Mergeability detection for automerge correction"
  • Feb 24: "Three stitches for the old-age and automerge hitches"
  • Feb 28: "Fix automerge rhymes-with resolution"
  • Mar 3: "Fix automerge: skip the unmergeable surge"

Four fixes. All passed community vote. All had rhyming titles. The bot still couldn't merge community PRs.

On Wednesday the bot ran automatically for the first time. It walked through all 38 open PRs top to bottom:

ERROR: Failed to merge PR #211: Resource not accessible by integration.
ERROR: Failed to merge PR #193: Resource not accessible by integration.
ERROR: Failed to merge PR #216: Resource not accessible by integration.
ERROR: Failed to merge PR #215: Resource not accessible by integration.
ERROR: Failed to merge PR #214: Resource not accessible by integration.
ERROR: Failed to merge PR #210: Resource not accessible by integration.
ERROR: Failed to merge PR #209: Resource not accessible by integration.
ERROR: Failed to merge PR #183: Resource not accessible by integration.
ERROR: Failed to merge PR #160: Resource not accessible by integration.

9 community PRs failed. It then merged mine - ranked #29 with 1 vote - because I'm the repo owner and GITHUB_TOKEN can bypass branch protection for owner PRs.

The answer was one line: GITHUB_TOKEN -> MERGE_PAT. A fine-grained PAT that acts as the repo owner. The community built the entire automerge system and debugged it for weeks. The final fix was a permissions edge case.

That fix is now a PR that needs 10 votes to merge under the new weekly rules. If it hits 10 by today 19:00 UTC, it'll be the first truly automatic democratic merge.

2 months in: 949 stars, 3,000+ unique voters, community-built themes, a researcher from TU Delft studying the voting patterns, and a bot that's one vote away from actually working.

https://openchaos.dev | https://github.com/skridlevsky/openchaos

submitted by /u/Equivalent-Yak2407 to r/webdev
[link] [comments]

How to make my site look good on all screens?

How to make my site look good on all screens?

I want to format my site to look nice on mobile and other screens, but I don't know anything about responsive web design. You can see how bad my site looks on mobile in the 2nd pic.

My website's here: https://blackedlight.neocities.org/

If you're on a desktop browser, you can see my code by opening Developer Tools with Ctrl + Shift + I.

submitted by /u/Dependent-Hamster361 to r/webdev
[link] [comments]

We got tired of basic data grid features being behind a paywall, so we built one. Announcing LyteNyte Grid Core 2.0

We got tired of basic data grid features being behind a paywall, so we built one. Announcing LyteNyte Grid Core 2.0

I have built and used many data grids in my career. One recurring issue was paywalls for basic grid features, along with dealing with heavy libraries that always seemed to hijack state. I genuinely get upset when I think about the hours I wasted with these problems.

That's why we shipped LyteNyte Grid Core v2 for the React community. It’s free, open-source (Apache 2.0), and loaded with advanced features that other libraries keep behind paywalls.

Why Care? Well, because DX matters, at least it does to our team. Core 2.0 is fully stateless and prop-driven. You can control everything declaratively from your own state, whether that’s URL params, Redux, or server state. You can run it headless if you want control over the UI, or use our styled grid if you just want to ship.

What’s New:

  • Premium Free Features: Row grouping, aggregations, and data export are now built-in. We are also moving Cell selection (another advanced feature) to Core in v2.1.
  • Tiny Bundle Size: We reduced bundle size down to just 30KB (gzipped).
  • Modernized API: Easily extendable with your own custom properties and methods. Improved: We redid the documentation so you can understand the code easily.

If you're looking for a high-performance React data grid that won't cost you a dollar, give LyteNyte Grid a try.

We’re actively building this for the community, so we’d love your feedback. Try it out, drop feature suggestions in the comments, and if it saves you a headache, a GitHub star always helps.

submitted by /u/After_Medicine8859 to r/webdev
[link] [comments]

Is IndexedDB actually... viable in 2026? Or am I wasting my time?

Is IndexedDB actually... viable in 2026? Or am I wasting my time?

I’ve been diving into local storage options for a project that needs to handle a decent amount of data (encrypted strings and some blobs).

everyone says IDB is the "standard" for this, but honestly, is offline-mode even a thing anymore for modern web apps?

i feel like most devs just rely on constant API calls now because "everyone is always online."

also, I tried implementing fuzzy search using Fuse.js on top of the data I was pulling from IDB, and the performance was a nightmare once the dataset grew as it needs to fetch everything into the memory to perform the search on them.

so, I actually had to rip the fuzzy search out because the lag was killing the UX.

is anyone actually using indexeddb in production successfully for large datasets...or is it just a legacy headache that we should replace with better API/Cloud architecture?

submitted by /u/nhrtrix to r/webdev
[link] [comments]

I accidentally shipped a bug that improved my product

so yesterday I pushed a small update to my side project. Nothing major. just cleaning up some UI stuff before going to sleep. this morning I woke up to a few emails from users. at first I thought that “great something broke.” turns out I did break something.

the feature that lets users edit items in a list stopped saving automatically. instead of saving on every change like it was supposed to, it only saved when they pressed enter.

but here’s the weird part. three different users wrote something like that “Hey, the new enter-to-save thing feels way better.” except I never designed it that way. it was literally a bug.

apparently the old version (auto saving on every change) was annoying people because they couldn’t make multiple edits easily. my bug accidentally created a small “commit step” that made editing feel more controlled. so now I’m sitting here turning my bug into an actual feature. it made me realize something kinda funny about building software that sometimes we spend days trying to design the “perfect UX”, and a random mistake ends up being closer to what users actually want.

what’s the best bug you’ve ever shipped? and did anything similar happen in your case?

submitted by /u/Interesting_Mine_400 to r/webdev
[link] [comments]
❌
❌