DjamgaMind: Audio Intelligence for the C-Suite (Daily AI News, Energy, Healthcare, Finance)
Full-Stack AI Intelligence. Zero Noise.The definitive audio briefing for the C-Suite and AI Architects. From Daily News and Strategic Deep Dives to high-density Industrial & Regulatory Intelligence—decoded at the speed of the AI era. . 👉 Start your specialized audio briefing today at Djamgamind.com
AI Jobs and Career
I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.
- Full Stack Engineer [$150K-$220K]
- Software Engineer, Tooling & AI Workflow, Contract [$90/hour]
- DevOps Engineer, India, Contract [$90/hour]
- More AI Jobs Opportunitieshere
| Job Title | Status | Pay |
|---|---|---|
| Full-Stack Engineer | Strong match, Full-time | $150K - $220K / year |
| Developer Experience and Productivity Engineer | Pre-qualified, Full-time | $160K - $300K / year |
| Software Engineer - Tooling & AI Workflows (Contract) | Contract | $90 / hour |
| DevOps Engineer (India) | Full-time | $20K - $50K / year |
| Senior Full-Stack Engineer | Full-time | $2.8K - $4K / week |
| Enterprise IT & Cloud Domain Expert - India | Contract | $20 - $30 / hour |
| Senior Software Engineer | Contract | $100 - $200 / hour |
| Senior Software Engineer | Pre-qualified, Full-time | $150K - $300K / year |
| Senior Full-Stack Engineer: Latin America | Full-time | $1.6K - $2.1K / week |
| Software Engineering Expert | Contract | $50 - $150 / hour |
| Generalist Video Annotators | Contract | $45 / hour |
| Generalist Writing Expert | Contract | $45 / hour |
| Editors, Fact Checkers, & Data Quality Reviewers | Contract | $50 - $60 / hour |
| Multilingual Expert | Contract | $54 / hour |
| Mathematics Expert (PhD) | Contract | $60 - $80 / hour |
| Software Engineer - India | Contract | $20 - $45 / hour |
| Physics Expert (PhD) | Contract | $60 - $80 / hour |
| Finance Expert | Contract | $150 / hour |
| Designers | Contract | $50 - $70 / hour |
| Chemistry Expert (PhD) | Contract | $60 - $80 / hour |
How does a database handle pagination?

It doesn’t. First, a database is a collection of related data, so I assume you mean DBMS or database language.
Second, pagination is generally a function of the front-end and/or middleware, not the database layer.
But some database languages provide helpful facilities that aide in implementing pagination. For example, many SQL dialects provide LIMIT and OFFSET clauses that can be used to emit up to n rows starting at a given row number. I.e., a “page” of rows. If the query results are sorted via ORDER BY and are generally unchanged between successive invocations, then that can be used to implement pagination.
That may not be the most efficient or effective implementation, though.

So how do you propose pagination should be done?
On context of web apps , let’s say there are 100 mn users. One cannot dump all the users in response.
Cache database query results in the middleware layer using Redis or similar and serve out pages of rows from that.
AI-Powered Professional Certification Quiz Platform
Web|iOs|Android|Windows
Are you passionate about AI and looking for your next career challenge? In the fast-evolving world of artificial intelligence, connecting with the right opportunities can make all the difference. We're excited to recommend Mercor, a premier platform dedicated to bridging the gap between exceptional AI professionals and innovative companies.
Whether you're seeking roles in machine learning, data science, or other cutting-edge AI fields, Mercor offers a streamlined path to your ideal position. Explore the possibilities and accelerate your AI career by visiting Mercor through our exclusive referral link:
Find Your AI Dream Job on Mercor
Your next big opportunity in AI could be just a click away!
What if you have 30, 000 rows plus, do you fetch all of that from the database and cache in Redis?
I feel the most efficient solution is still offset and limit. It doesn’t make sense to use a database and then end up putting all of your data in Redis especially data that changes a lot. Redis is not for storing all of your data.
If you have large data set, you should use offset and limit, getting only what is needed from the database into main memory (and maybe caching those in Redis) at any point in time is very efficient.
With 30,000 rows in a table, if offset/limit is the only viable or appropriate restriction, then that’s sometimes the way to go.
AI- Powered Jobs Interview Warmup For Job Seekers

⚽️Comparative Analysis: Top Calgary Amateur Soccer Clubs – Outdoor 2025 Season (Kids' Programs by Age Group)
More often, there’s a much better way of restricting 30,000 rows via some search criteria that significantly reduces the displayed volume of rows — ideally to a single page or a few pages (which are appropriate to cache in Redis.)
It’s unlikely (though it does happen) that users really want to casually browse 30,000 rows, page by page. More often, they want this one record, or these small number of records.
AI Jobs and Career
And before we wrap up today's AI news, I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.
Invest in your future today by enrolling in this Azure Fundamentals - Pass the Azure Fundamentals Exam with Ease: Master the AZ-900 Certification with the Comprehensive Exam Preparation Guide!
- AWS Certified AI Practitioner (AIF-C01): Conquer the AWS Certified AI Practitioner exam with our AI and Machine Learning For Dummies test prep. Master fundamental AI concepts, AWS AI services, and ethical considerations.
- Azure AI Fundamentals: Ace the Azure AI Fundamentals exam with our comprehensive test prep. Learn the basics of AI, Azure AI services, and their applications.
- Google Cloud Professional Machine Learning Engineer: Nail the Google Professional Machine Learning Engineer exam with our expert-designed test prep. Deepen your understanding of ML algorithms, models, and deployment strategies.
- AWS Certified Machine Learning Specialty: Dominate the AWS Certified Machine Learning Specialty exam with our targeted test prep. Master advanced ML techniques, AWS ML services, and practical applications.
- AWS Certified Data Engineer Associate (DEA-C01): Set yourself up for promotion, get a better job or Increase your salary by Acing the AWS DEA-C01 Certification.
Question: This is a general question that applies to MySQL, Oracle DB or whatever else might be out there.
I know for MySQL there is LIMIT offset,size; and for Oracle there is ‘ROW_NUMBER’ or something like that.
But when such ‘paginated’ queries are called back to back, does the database engine actually do the entire ‘select’ all over again and then retrieve a different subset of results each time? Or does it do the overall fetching of results only once, keeps the results in memory or something, and then serves subsets of results from it for subsequent queries based on offset and size?
If it does the full fetch every time, then it seems quite inefficient.
If it does full fetch only once, it must be ‘storing’ the query somewhere somehow, so that the next time that query comes in, it knows that it has already fetched all the data and just needs to extract next page from it. In that case, how will the database engine handle multiple threads? Two threads executing the same query?
something will be quick or slow without taking measurements, and complicate the code in advance to download 12 pages at once and cache them because “it seems to me that it will be faster”.
Answer: First of all, do not make assumptions in advance whether something will be quick or slow without taking measurements, and complicate the code in advance to download 12 pages at once and cache them because “it seems to me that it will be faster”.
YAGNI principle – the programmer should not add functionality until deemed necessary.
Do it in the simplest way (ordinary pagination of one page), measure how it works on production, if it is slow, then try a different method, if the speed is satisfactory, leave it as it is.
From my own practice – an application that retrieves data from a table containing about 80,000 records, the main table is joined with 4-5 additional lookup tables, the whole query is paginated, about 25-30 records per page, about 2500-3000 pages in total. Database is Oracle 12c, there are indexes on a few columns, queries are generated by Hibernate. Measurements on production system at the server side show that an average time (median – 50% percentile) of retrieving one page is about 300 ms. 95% percentile is less than 800 ms – this means that 95% of requests for retrieving a single page is less that 800ms, when we add a transfer time from the server to the user and a rendering time of about 0.5-1 seconds, the total time is less than 2 seconds. That’s enough, users are happy.
And some theory – see this answer to know what is purpose of Pagination pattern
- A Deep Dive into MVCC in PostgreSQL: How Concurrency Really Worksby Navidbarsalari (Database on Medium) on April 17, 2026 at 5:02 pm
Concurrency is one of the hardest problems in database systems — not because it’s complex in theory, but because it becomes unpredictable…Continue reading on Medium »
- The Database Is the Heart of Every Systemby Ankitsapkotaquartz (Database on Medium) on April 17, 2026 at 4:46 pm
And like every heart, it cannot stop beating.Continue reading on Medium »
- When One “Fast” Query Becomes a Production Nightmare: The Dehydration Problem in Hibernateby Vitalii Sukach (Database on Medium) on April 17, 2026 at 4:36 pm
Developers love the idea of a single, elegant SQL query that fetches everything in one go. One round‑trip to the database. One response…Continue reading on Medium »
- From Fragile Direct Writes to Scalable Real-Time Persistence: How I Upgraded My Realtime…by Atharv Mutkule (Database on Medium) on April 17, 2026 at 4:17 pm
When I started building my collaborative drawing app (an advanced Excalidraw clone), everything felt magical. Multiple users could draw…Continue reading on Medium »
- Building a Unified Data-as-a-Service Platform: How GBase Empowers Telecom Regulation with Modern…by Scale (Database on Medium) on April 17, 2026 at 3:22 pm
## IntroductionContinue reading on Medium »
- Integrating Nacos with GBase Database: A Practical Guide for Enterprise Deploymentby Scale (Database on Medium) on April 17, 2026 at 3:19 pm
# Integrating Nacos with GBase Database: A Practical Guide for Enterprise DeploymentContinue reading on Medium »
- Best MongoDB Hosting Providers in 2026: Top Picks for Developers & Startupsby Advith Aatreya (Database on Medium) on April 17, 2026 at 1:26 pm
Modern applications rely heavily on fast, scalable databases — and MongoDB has become a top choice due to its flexible, document-based…Continue reading on Medium »
- Inside the Lexer and Parser of PostgreSQLby Divya karlapudi (Database on Medium) on April 17, 2026 at 1:15 pm
From Grammar to Code.Continue reading on Medium »
- Yes, You Can Run SQL Server on Linux — Here’s How to Do It Rightby Dennis Peter Munyao (Database on Medium) on April 17, 2026 at 12:34 pm
Microsoft SQL Server on Ubuntu in 2024 is not a compromise. It’s a power move.Continue reading on Medium »
- Why Your Database Becomes the Bottleneckby Akshat Jain (Database on Medium) on April 17, 2026 at 12:31 pm
Why most backend performance issues eventually lead back to the databaseContinue reading on Medium »
- Extracting data from onestream for analytics outside the platform ,anyone figured this outby /u/cole_10 (Database) on April 17, 2026 at 4:54 am
Finance operations analyst at a company that uses onestream for financial consolidation, close management, and planning. Onestream is powerful for what it does inside the platform but getting data out of it for broader analytics is proving difficult. We need onestream consolidated financial data alongside operational data from our erp and crm in a central warehouse for combined analysis. The onestream api exists but it's not well documented for bulk data extraction use cases. It was designed more for application integration than for piping large datasets into an external warehouse. The stage tables approach lets you access the underlying sql server data but requires network level access and coordination with the onestream admin team. We've been doing manual exports from onestream reports which introduces the same stale data and human error problems we were trying to solve by having onestream in the first place. Has anyone built an automated pipeline to extract onestream financial data into a cloud warehouse? What approach did you use and how reliable has it been? submitted by /u/cole_10 [link] [comments]
- Want to Replace MS Access Form with something web basedby /u/rossmosh85 (Database) on April 16, 2026 at 10:20 pm
I have an MS Access "program" that I'd like to replace with something web based. It's cobbled together by me, a non coder. I'm looking for something web based that might do something similar. Something relatively user friendly and open source would be ideal Here's an outline of what it does: I upload 3-4 formatted CSV/Excel files to multiple individual tables. Each table holds approximately 10,000 items. They are products from my suppliers. FORM 1: Part/Product Info Combines the 4 tables mentioned above via a Query. It allows me to search through the 4 tables to find an item. It will then display the part, description, and various pricing info. I also have it calculate a Suggested Retail Price via a simple and a slightly more complicated formula. The more complicated formula is due to parts being sold individually, by case, and mixed. FORM 2: Product Assembly Form This is actually the most important form. While FORM 1 is nice, the product assembly form is really the biggest one I use these days. Long story short, it allows me to form product assemblies. I have a query that combines all of the items together. It stores a more simplified data set. I then can build a Product Assembly from the parts. It then stores the Product Assembly in it's own table. To make sure pricing is current, I have it store just the quantities of the parts and the part number and then it pulls up the current pricing as it loads. Is there any web app or program that anyone could recommend that would do this without an extensive amount of research and effort? submitted by /u/rossmosh85 [link] [comments]
- Would you use a hosted DB-over-API for MVPs, scripts, and hackathons?by /u/N_Sin (Database) on April 16, 2026 at 9:57 pm
I’m building a small hosted DB-over-API (SaaS) product and I’m trying to validate whether this is actually useful to other developers. The idea is not “replace your real database.” It’s more: if you want to store and query data quickly over HTTP without setting up a full backend, would you use something like this? The use cases I have in mind are things like: quick MVPs small scripts running across different devices hackathons tutorials and demos internal tools prototypes where you just want “data + API” without much setup Example shapes would be something like: GET{{baseurl}}/api/v1/tables/{{tableName}}/{{recordId}} Or GET{{baseurl}}/api/v1/tables/{{tableName}}?filter=done:eq:false&sort=priority:asc,created_at:desc This is not meant to replace any SQL dB for bigger or more serious projects. I’m thinking of it more as a convenience tool for cases where speed and simplicity matter more than full DB power. What I’d really like to know: Would you use something like this? For which use cases would it actually be better than just using Postgres, SQLite, Supabase, Firebase, etc.? If you had heavier usage, would you pay for it? Would you be interested in helping shape the product and giving feedback on design decisions? I would really appreciate blunt feedback, especially from people who have built quick MVPs, hackathon apps, automations, or tutorial projects. Here is a video of how quick set up is: Note that columns id, created_at, updated_at are automatically managed for every table by the api and not by the user. Also in this video example I'm using the infer schema from first write option rather than first creating a schema with the dedicated endpoint (to showcase speed). https://reddit.com/link/1snhsum/video/b792idtyjpvg1/player submitted by /u/N_Sin [link] [comments]
- Sqlite: Attaching a database for ad-hoc foreign key check?by /u/No-Security-7518 (Database) on April 16, 2026 at 7:16 pm
I have 2 Sqlite databases; Users + Inventory. I have a column in several tables in inventory.db that records which user did things such as: removing/registering a product, etc. What is the cleanest way to achieve data integrity here? 1. Users.db belongs to a library I'm declaring as a dependency. 2. Both databases are copied to a directory at startup so they're next to each other. Should I merge them at startup too? (copy schema +data)? Or use Attach Database? I understand FK checks aren't possible then. So maybe just check the userId is valid? I appreciate your input. submitted by /u/No-Security-7518 [link] [comments]
- Problem automation galaraby /u/N4NT4 (Database) on April 16, 2026 at 2:04 pm
submitted by /u/N4NT4 [link] [comments]
- Multi Vendor Insurance system best db designby /u/Abhinava21 (Database) on April 16, 2026 at 7:55 am
I am building a module in which I have to integrate multi-vendor insurance using the nestjs and mysql. Mainly our purpose is to do insurance for new E-rickshaws. So, what is the best tables schemas I can create. so, it is scalable and supports multivendor. I have created some of the columns and implemented one of the vendors. But I don't think it is scalable so need advice for the same. submitted by /u/Abhinava21 [link] [comments]
- many to many binary relationship in ER to relational model but cant doby /u/Kerem1111 (Database) on April 14, 2026 at 3:15 pm
Work assignment is connected to facility and instructors. I want to translate this into a relational model but the issue is, facility has a PK so I just need to include facilityCode in Work assignment table, but instructors or by extension staff doesn't have a PK. How am I supposed to include that? Thanks submitted by /u/Kerem1111 [link] [comments]
- A LISTEN/NOTIFY debugger that survives reconnects and keeps 10k events in local SQLiteby /u/rohithgilla (Database) on April 14, 2026 at 4:40 am
I've rewritten the same 40-line pg.Client listen.js script at least six times on three different laptops. This is the version I wish I'd built the first time. The panel: Subscribes to multiple channels on a connection Persists every event to a local SQLite file (10k per connection ring buffer, enforced in SQL not JS) Reconnects with exponential backoff capped at 30s on drop Re-subscribes to the full current channel set, not the original one (this was a bug the first time — I was losing channels added after initial connect) Quotes channel identifiers properly because LISTEN takes an identifier, not a bindable parameter Writeup with the full reconnect code + the "" identifier-quoting gotcha: https://datapeek.dev/blog/listen-notify-without-tears If anyone has a better answer than exponential backoff for reconnect on pg notification clients, I'd love to hear it. submitted by /u/rohithgilla [link] [comments]
- Advice on whether nosql is the right choice?by /u/puma905 (Database) on April 14, 2026 at 2:40 am
I’m building a mobile app where users log structured daily entries about an ongoing condition (things like symptoms, possible triggers, actions taken, and optional notes). Over time, the app generates simple summaries and pattern insights based on those logs. Each user has their own dataset, entries are append-heavy with occasional edits, and the schema may evolve as I learn more from real usage. There will be lightweight analytics and AI-driven summaries on top of the data. I would like to be able to also aggregate data across users over time to better understand trends, etc. I’m trying to decide whether a NoSQL document database is the right choice long-term, or if I should be thinking about a relational model from the start. Curious how others would approach this kind of use case. submitted by /u/puma905 [link] [comments]
- How do you prevent retroactive policy application due to timing gaps between policy updates and enforcement?by /u/homerderby (Database) on April 13, 2026 at 8:11 am
I’ve been looking into an issue where there’s a timing gap between when a policy is announced (or updated in the system) and when the actual enforcement logic is applied. In several cases, transactions that were already completed ended up being evaluated under the new policy rules, which led to inconsistencies and data integrity concerns. From what I can tell, this usually comes from mismatches between the policy DB update timing and the validation/execution layer — older state gets interpreted by a newer rules engine. One approach I’ve been considering is isolating the scope using a snapshot at the time of announcement, combined with a clear grace period to strictly separate timelines. [Attached image: timeline diagram showing policy announcement vs enforcement mismatch] For those working with transactional systems, how do you architect around this? Do you version policies, rely on event sourcing, or enforce strict temporal boundaries at the DB level? I’ve been exploring this problem in a small internal context (oncastudy), and I’m curious what patterns have worked reliably in production. https://preview.redd.it/gu6gihnn1xug1.png?width=1200&format=png&auto=webp&s=bb6bddcf7f66f8550c59e0312722c7e39f8cd170 submitted by /u/homerderby [link] [comments]
- I can finally screen-share my SQL client without leaking prod databy /u/rohithgilla (Database) on April 13, 2026 at 6:52 am
submitted by /u/rohithgilla [link] [comments]
- Keeping a Postgres queue healthyby /u/2minutestreaming (Database) on April 11, 2026 at 9:23 am
submitted by /u/2minutestreaming [link] [comments]
- Json in relational dbby /u/No_Character_2277 (Database) on April 11, 2026 at 5:34 am
Is using json or jsonb datatype for a column in relational db really works well in production environment? submitted by /u/No_Character_2277 [link] [comments]
- A new approach to database queries called GiGIby /u/quant-alliance (Database) on April 10, 2026 at 9:46 pm
Hello community, we are a team of two engineers with experience working for NASA and various other short letter agencies. We took a concept based on non Euclidean geometry called the fiber bundle and built a small database around it. We call this new type of index GiGi and you can see benchmarks and run test here: https://www.davisgeometric.com/gigi#home We are looking for some sort of direction: should we make it open source but we are extremely introverted and not sure we can manage and accumulate a community or should we go for a community Vs enterprise version? do you want to see more benchmarks? which type and what other databases? submitted by /u/quant-alliance [link] [comments]
- Drew this with AI based on a real incident. Anyone else been here at 3AM?by /u/Adela_freedom (Database) on April 10, 2026 at 2:06 pm
AI-illustrated, but the story is real. Has this happened to your team? How did you fix the access model afterward? submitted by /u/Adela_freedom [link] [comments]
- A Conversation with Paul Masurel, Creator of Tantivyby /u/philippemnoel (Database) on April 9, 2026 at 9:42 pm
Tantivy is a very popular Rust search library inspired by Apache Lucene. We sat down with Paul, the main author, to discuss how he got started with Rust and Tantivy, and his journey since then. I figured it would be interesting to folks here 🙂 submitted by /u/philippemnoel [link] [comments]
- Help with normalizing a database?by /u/Fit-Try9217 (Database) on April 9, 2026 at 9:25 pm
Hi! I'm currently working on my project for my database course. I've managed to finish my ERD and relational schema, but when I come to normalize my relational schema, I feel like nothing has changed, and I'm worried I might not be seeing something properly. You can find below the ERD and the unnormalized relational schema! Any help appreciated! https://preview.redd.it/qee3alfmf8ug1.png?width=4386&format=png&auto=webp&s=e8918062be4d316b16fdc2c3cf7aee01c713d824 https://preview.redd.it/f6oaxz9nf8ug1.png?width=1920&format=png&auto=webp&s=b13d29afe86f19fa6fead19e76778ef945ded3cb submitted by /u/Fit-Try9217 [link] [comments]
- Hi there I'm having a problem in orecal dbby /u/Im_only0101 (Database) on April 9, 2026 at 6:33 pm
hi there in using orecal sql developer. Im having table with blob data in a column and I want store it in my computer at a path and add that path to same table with new column. the orecal in on server connect through lan I don't have access to admin or dba and I want use pl/sql. do anyone over internet having idea or solutions of this problem please help #orecal #bug #developer #problemsolving #computerscience #sql #plswl submitted by /u/Im_only0101 [link] [comments]
- Where do I get started with making webscale DB projectsby /u/Adventurous_Tea_2198 (Database) on April 9, 2026 at 5:54 pm
Messed around with sqlite for a project but I can tell it's really shit. submitted by /u/Adventurous_Tea_2198 [link] [comments]
- Database Helpby /u/duskti (Database) on April 9, 2026 at 5:04 pm
I recently joined an organization that requires a robust, scalable database solution for archiving data. Storage needs are projected to reach approximately 100 TB over the next few years, so I’m looking to plan strategically. The data includes a variety of file types—PDFs, Excel files, 3D renderings, and videos—with some individual files as large as 30 GB. We currently have a NAS in place. I’m seeking recommendations for setting this up effectively, ideally with a frontend that allows technicians to upload files directly to the storage system—without incurring high monthly costs from services like AWS S3 or similar cloud providers. submitted by /u/duskti [link] [comments]
- [ Removed by Reddit ]by /u/2011wpfg (Database) on April 9, 2026 at 9:11 am
[ Removed by Reddit on account of violating the content policy. ] submitted by /u/2011wpfg [link] [comments]
- How to efficiently run and re-run mysql/mariadb-test-runby /u/OttoKekalainen (Database) on April 9, 2026 at 5:59 am
For anyone doing their first contribution to MySQL or MariaDB: start out by learning how the mysql/mariadb-test-run command works and how to efficiently rebuild the sources and re-run the test suite. submitted by /u/OttoKekalainen [link] [comments]
- Why does SWITCHOFFSET return the wrong local time when used with the timezone value returned by CURRENT_TIMEZONE?by /u/Reasonable-Job4205 (Database) on April 8, 2026 at 8:55 pm
SYSDATETIMEOFFSET(): 2026-04-08 13:49:06.4745888 -07:00 SYSUTCDATETIME(): 2026-04-08 20:49:06.4745888 CURRENT_TIMEZONE: (UTC-08:00) Pactific Time (US & Canada) SWITCHOFFSET(SYSUTCDATETIME, '-08:00'): 2026-04-08 12:49:06.4745888 -08:00 The correct local time is 13:49, but CURRENT_TIMEZONE returns -08:00, which then causes the computed local time to be 12:49, which is wrong. Why is this? submitted by /u/Reasonable-Job4205 [link] [comments]
- MongoDB Indexing Recommendationby /u/goldenuser22628 (Database) on April 8, 2026 at 1:32 pm
I’m a bit confused about how to approach indexing, and I’m not fully confident in the decisions I’m making. I know .explain() can help, and I understand that indexes should usually be based on access patterns. The problem in my case is that users can filter on almost any field, which makes it harder to know what the right indexing strategy should be. For example, imagine a collection called dummy with a schema like this: { field1: string, field2: string, field3: boolean, field4: boolean, ... fieldN: ... } If users are allowed to filter by any of these fields, what would be the recommended indexing approach or best practice in this situation? submitted by /u/goldenuser22628 [link] [comments]
- Neo4j vs ArangoDB for high volume-ingest + multi-hop traversal use case?by /u/Klutzy_Plantain1737 (Database) on April 8, 2026 at 12:16 pm
Hey all — would love to get some real-world perspectives from folks who have used Neo4j and/or ArangoDB in production. We’re currently evaluating graph databases for a use case that involves: • heavy multi-hop traversal (core requirement — this is where graph really shines for us) • modeling relationships across devices, applications, vulnerabilities, etc. • some degree of temporal/state-based data • and moderate to high write volume depending on the window From a querying and traversal perspective, Neo4j has honestly been great. The model feels natural, Cypher is intuitive, and performance on traversal-heavy queries has been solid in our testing. Where we’re running into friction is ingestion. Given our constraints (security + environment), bulk loading into Neo4j Aura hasn’t been straightforward. For large loads, the suggested patterns we’ve seen involve things like: • driver-based ingestion (which is slower for large volumes) • or building/loading externally and restoring into Aura In practice, this has made large-scale ingestion feel like a bottleneck. For heavier loads, we’ve even had to consider taking the database offline overnight to get data in efficiently, which isn’t ideal if this becomes part of regular operations. This has us questioning: • how others are handling high-volume ingestion with Neo4j (especially Aura vs self-managed EE) • whether this is just a constraint of our setup, or a broader limitation depending on architecture ⸻ At the same time, we’re also looking at ArangoDB, which seems more flexible around ingestion (online writes, bulk APIs, etc.), but we’re still trying to understand: • how it compares for deep multi-hop traversal performance • how well it handles complex graph patterns vs Neo4j • any tradeoffs in query ergonomics / modeling ⸻ Questions for the group: 1. If you’re using Neo4j at scale, how are you handling ingestion? • Are you using Kafka / streaming pipelines? • Self-managed EE vs Aura? • Any pain points with large loads? 2. Has anyone used Neo4j Aura specifically for write-heavy or high-ingest workloads? 3. For those who’ve used ArangoDB: • How does it compare for multi-hop traversal performance? • Any limitations vs Neo4j when queries get complex? 4. If you had to choose again for a use case that is: • traversal-heavy • but also requires reliable, ongoing ingestion at scale what would you pick and why? submitted by /u/Klutzy_Plantain1737 [link] [comments]





































96DRHDRA9J7GTN6