Codetown ::: a software developer's community
All the tutorials and books for node.js seem to use Mongo as the database. I am not sold on 'document' databases and would like to know how difficult it is to use any version for plain old tried-and-true SQL with Node.js.
Does anybody have any experience in this area?
Tags:
One of the traditional knocks on JS is the volatility of doing sql queries from a interpreted script. Not to mention security. In other words, how do you regulate resource for results in a varying client environment. Node.js is supposed to provide a server side capability. However I would be skeptical of it its implementation of a transnational capability. A memento pattern, or ability to rollback transactions, at least until thoroughly tested. Given the fact that most discussions are coupled with no-sql db's is a clue as to what its intended usage should be. Perhaps caches for local search tools like solr. Easy to update, and rebuild, but less likely to be an efficient engine for individualized rdbms queries.
Codetown is a social network. It's got blogs, forums, groups, personal pages and more! You might think of Codetown as a funky camper van with lots of compartments for your stuff and a great multimedia system, too! Best of all, Codetown has room for all of your friends.
Created by Michael Levin Dec 18, 2008 at 6:56pm. Last updated by Michael Levin May 4, 2018.
Check out the Codetown Jobs group.

Open-source AI coding tool OpenCode features a native terminal-based UI, multi-session support, and compatibility with over 75 models, including Claude, OpenAI, Gemini, and local models. In addition to its CLI tool, OpenCode is also available as a desktop app and and an IDE extension for VS Code, Cursor, and other tools.
By Sergio De Simone
Teams rely on strong unit and integration tests instead of end-to-end tests. Using TDD, pair programming, and good design, they ship small changes often, test in production for real feedback, and use feature toggles to reduce risk, Ola Hast and Asgaut Mjølne Söderbom mentioned in their talk about continuous delivery with pair programming.
By Ben Linders
Google Cloud is bridging a critical gap for enterprises by introducing a gRPC transport package for the Model Context Protocol (MCP), enhancing integration for businesses already using gRPC. This game-changer reduces development friction, ensuring AI agents seamlessly connect with existing services while boosting performance and efficiency. Join the evolving landscape of AI integration!
By Steef-Jan Wiggers
LinkedIn's engineering team successfully upgraded its legacy ZooKeeper service discovery platform to enhance scalability and performance. By leveraging Apache Kafka and the xDS protocol, the new architecture enables eventual consistency, supports multiple languages, and allows migration without downtime. Post-upgrade, latency vastly improved, facilitating hundreds of thousands of app instances.
By Patrick Farry
Sarah Usher discusses the architectural "breaking point" where warehouses like BigQuery struggle with latency and cost. She explains the necessity of a conceptual data lifecycle (Raw, Curated, Use Case) to regain control over lineage and innovation. She shares practical strategies to design a single source of truth that empowers both ML teams and analytics without bottlenecking scale.
By Sarah Usher
© 2026 Created by Michael Levin.
Powered by