How does a database handle pagination?

How does a database handle pagination?
DjamgaMind - AI Unraveled Podcast

DjamgaMind: Audio Intelligence for the C-Suite (Daily AI News, Energy, Healthcare, Finance)

Full-Stack AI Intelligence. Zero Noise.The definitive audio briefing for the C-Suite and AI Architects. From Daily News and Strategic Deep Dives to high-density Industrial & Regulatory Intelligence—decoded at the speed of the AI era. . 👉 Start your specialized audio briefing today at Djamgamind.com


AI Jobs and Career

I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.

Job TitleStatusPay
Full-Stack Engineer Strong match, Full-time $150K - $220K / year
Developer Experience and Productivity Engineer Pre-qualified, Full-time $160K - $300K / year
Software Engineer - Tooling & AI Workflows (Contract) Contract $90 / hour
DevOps Engineer (India) Full-time $20K - $50K / year
Senior Full-Stack Engineer Full-time $2.8K - $4K / week
Enterprise IT & Cloud Domain Expert - India Contract $20 - $30 / hour
Senior Software Engineer Contract $100 - $200 / hour
Senior Software Engineer Pre-qualified, Full-time $150K - $300K / year
Senior Full-Stack Engineer: Latin America Full-time $1.6K - $2.1K / week
Software Engineering Expert Contract $50 - $150 / hour
Generalist Video Annotators Contract $45 / hour
Generalist Writing Expert Contract $45 / hour
Editors, Fact Checkers, & Data Quality Reviewers Contract $50 - $60 / hour
Multilingual Expert Contract $54 / hour
Mathematics Expert (PhD) Contract $60 - $80 / hour
Software Engineer - India Contract $20 - $45 / hour
Physics Expert (PhD) Contract $60 - $80 / hour
Finance Expert Contract $150 / hour
Designers Contract $50 - $70 / hour
Chemistry Expert (PhD) Contract $60 - $80 / hour

How does a database handle pagination?

How does a database handle pagination?

It doesn’t. First, a database is a collection of related data, so I assume you mean DBMS or database language.

Second, pagination is generally a function of the front-end and/or middleware, not the database layer.

But some database languages provide helpful facilities that aide in implementing pagination. For example, many SQL dialects provide LIMIT and OFFSET clauses that can be used to emit up to n rows starting at a given row number. I.e., a “page” of rows. If the query results are sorted via ORDER BY and are generally unchanged between successive invocations, then that can be used to implement pagination.

That may not be the most efficient or effective implementation, though.

How does a database handle pagination?

So how do you propose pagination should be done?

On context of web apps , let’s say there are 100 mn users. One cannot dump all the users in response.

Cache database query results in the middleware layer using Redis or similar and serve out pages of rows from that.

AI-Powered Professional Certification Quiz Platform
Crack Your Next Exam with Djamgatech AI Cert Master

Web|iOs|Android|Windows

Are you passionate about AI and looking for your next career challenge? In the fast-evolving world of artificial intelligence, connecting with the right opportunities can make all the difference. We're excited to recommend Mercor, a premier platform dedicated to bridging the gap between exceptional AI professionals and innovative companies.

Whether you're seeking roles in machine learning, data science, or other cutting-edge AI fields, Mercor offers a streamlined path to your ideal position. Explore the possibilities and accelerate your AI career by visiting Mercor through our exclusive referral link:

Find Your AI Dream Job on Mercor

Your next big opportunity in AI could be just a click away!

What if you have 30, 000 rows plus, do you fetch all of that from the database and cache in Redis?

I feel the most efficient solution is still offset and limit. It doesn’t make sense to use a database and then end up putting all of your data in Redis especially data that changes a lot. Redis is not for storing all of your data.

If you have large data set, you should use offset and limit, getting only what is needed from the database into main memory (and maybe caching those in Redis) at any point in time is very efficient.

With 30,000 rows in a table, if offset/limit is the only viable or appropriate restriction, then that’s sometimes the way to go.

More often, there’s a much better way of restricting 30,000 rows via some search criteria that significantly reduces the displayed volume of rows — ideally to a single page or a few pages (which are appropriate to cache in Redis.)

It’s unlikely (though it does happen) that users really want to casually browse 30,000 rows, page by page. More often, they want this one record, or these small number of records.

AI Jobs and Career

And before we wrap up today's AI news, I wanted to share an exciting opportunity for those of you looking to advance your careers in the AI space. You know how rapidly the landscape is evolving, and finding the right fit can be a challenge. That's why I'm excited about Mercor – they're a platform specifically designed to connect top-tier AI talent with leading companies. Whether you're a data scientist, machine learning engineer, or something else entirely, Mercor can help you find your next big role. If you're ready to take the next step in your AI career, check them out through my referral link: https://work.mercor.com/?referralCode=82d5f4e3-e1a3-4064-963f-c197bb2c8db1. It's a fantastic resource, and I encourage you to explore the opportunities they have available.

 

Question: This is a general question that applies to MySQL, Oracle DB or whatever else might be out there.

I know for MySQL there is LIMIT offset,size; and for Oracle there is ‘ROW_NUMBER’ or something like that.

But when such ‘paginated’ queries are called back to back, does the database engine actually do the entire ‘select’ all over again and then retrieve a different subset of results each time? Or does it do the overall fetching of results only once, keeps the results in memory or something, and then serves subsets of results from it for subsequent queries based on offset and size?

If it does the full fetch every time, then it seems quite inefficient.

If it does full fetch only once, it must be ‘storing’ the query somewhere somehow, so that the next time that query comes in, it knows that it has already fetched all the data and just needs to extract next page from it. In that case, how will the database engine handle multiple threads? Two threads executing the same query?

something will be quick or slow without taking measurements, and complicate the code in advance to download 12 pages at once and cache them because “it seems to me that it will be faster”.


AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

Answer: First of all, do not make assumptions in advance whether something will be quick or slow without taking measurements, and complicate the code in advance to download 12 pages at once and cache them because “it seems to me that it will be faster”.

YAGNI principle – the programmer should not add functionality until deemed necessary.
Do it in the simplest way (ordinary pagination of one page), measure how it works on production, if it is slow, then try a different method, if the speed is satisfactory, leave it as it is.


From my own practice – an application that retrieves data from a table containing about 80,000 records, the main table is joined with 4-5 additional lookup tables, the whole query is paginated, about 25-30 records per page, about 2500-3000 pages in total. Database is Oracle 12c, there are indexes on a few columns, queries are generated by Hibernate. Measurements on production system at the server side show that an average time (median – 50% percentile) of retrieving one page is about 300 ms. 95% percentile is less than 800 ms – this means that 95% of requests for retrieving a single page is less that 800ms, when we add a transfer time from the server to the user and a rendering time of about 0.5-1 seconds, the total time is less than 2 seconds. That’s enough, users are happy.


And some theory – see this answer to know what is purpose of Pagination pattern

  • A Deep Dive into MVCC in PostgreSQL: How Concurrency Really Works
    by Navidbarsalari (Database on Medium) on April 17, 2026 at 5:02 pm

    Concurrency is one of the hardest problems in database systems — not because it’s complex in theory, but because it becomes unpredictable…Continue reading on Medium »

  • The Database Is the Heart of Every System
    by Ankitsapkotaquartz (Database on Medium) on April 17, 2026 at 4:46 pm

    And like every heart, it cannot stop beating.Continue reading on Medium »

  • When One “Fast” Query Becomes a Production Nightmare: The Dehydration Problem in Hibernate
    by Vitalii Sukach (Database on Medium) on April 17, 2026 at 4:36 pm

    Developers love the idea of a single, elegant SQL query that fetches everything in one go. One round‑trip to the database. One response…Continue reading on Medium »

  • From Fragile Direct Writes to Scalable Real-Time Persistence: How I Upgraded My Realtime…
    by Atharv Mutkule (Database on Medium) on April 17, 2026 at 4:17 pm

    When I started building my collaborative drawing app (an advanced Excalidraw clone), everything felt magical. Multiple users could draw…Continue reading on Medium »

  • Building a Unified Data-as-a-Service Platform: How GBase Empowers Telecom Regulation with Modern…
    by Scale (Database on Medium) on April 17, 2026 at 3:22 pm

    ## IntroductionContinue reading on Medium »

  • Integrating Nacos with GBase Database: A Practical Guide for Enterprise Deployment
    by Scale (Database on Medium) on April 17, 2026 at 3:19 pm

    # Integrating Nacos with GBase Database: A Practical Guide for Enterprise DeploymentContinue reading on Medium »

  • Best MongoDB Hosting Providers in 2026: Top Picks for Developers & Startups
    by Advith Aatreya (Database on Medium) on April 17, 2026 at 1:26 pm

    Modern applications rely heavily on fast, scalable databases — and MongoDB has become a top choice due to its flexible, document-based…Continue reading on Medium »

  • Inside the Lexer and Parser of PostgreSQL
    by Divya karlapudi (Database on Medium) on April 17, 2026 at 1:15 pm

    From Grammar to Code.Continue reading on Medium »

  • Yes, You Can Run SQL Server on Linux — Here’s How to Do It Right
    by Dennis Peter Munyao (Database on Medium) on April 17, 2026 at 12:34 pm

    Microsoft SQL Server on Ubuntu in 2024 is not a compromise. It’s a power move.Continue reading on Medium »

  • Why Your Database Becomes the Bottleneck
    by Akshat Jain (Database on Medium) on April 17, 2026 at 12:31 pm

    Why most backend performance issues eventually lead back to the databaseContinue reading on Medium »

  • Extracting data from onestream for analytics outside the platform ,anyone figured this out
    by /u/cole_10 (Database) on April 17, 2026 at 4:54 am

    Finance operations analyst at a company that uses onestream for financial consolidation, close management, and planning. Onestream is powerful for what it does inside the platform but getting data out of it for broader analytics is proving difficult. We need onestream consolidated financial data alongside operational data from our erp and crm in a central warehouse for combined analysis. The onestream api exists but it's not well documented for bulk data extraction use cases. It was designed more for application integration than for piping large datasets into an external warehouse. The stage tables approach lets you access the underlying sql server data but requires network level access and coordination with the onestream admin team. We've been doing manual exports from onestream reports which introduces the same stale data and human error problems we were trying to solve by having onestream in the first place. Has anyone built an automated pipeline to extract onestream financial data into a cloud warehouse? What approach did you use and how reliable has it been? submitted by /u/cole_10 [link] [comments]

  • Want to Replace MS Access Form with something web based
    by /u/rossmosh85 (Database) on April 16, 2026 at 10:20 pm

    I have an MS Access "program" that I'd like to replace with something web based. It's cobbled together by me, a non coder. I'm looking for something web based that might do something similar. Something relatively user friendly and open source would be ideal Here's an outline of what it does: I upload 3-4 formatted CSV/Excel files to multiple individual tables. Each table holds approximately 10,000 items. They are products from my suppliers. FORM 1: Part/Product Info Combines the 4 tables mentioned above via a Query. It allows me to search through the 4 tables to find an item. It will then display the part, description, and various pricing info. I also have it calculate a Suggested Retail Price via a simple and a slightly more complicated formula. The more complicated formula is due to parts being sold individually, by case, and mixed. FORM 2: Product Assembly Form This is actually the most important form. While FORM 1 is nice, the product assembly form is really the biggest one I use these days. Long story short, it allows me to form product assemblies. I have a query that combines all of the items together. It stores a more simplified data set. I then can build a Product Assembly from the parts. It then stores the Product Assembly in it's own table. To make sure pricing is current, I have it store just the quantities of the parts and the part number and then it pulls up the current pricing as it loads. Is there any web app or program that anyone could recommend that would do this without an extensive amount of research and effort? submitted by /u/rossmosh85 [link] [comments]

  • Would you use a hosted DB-over-API for MVPs, scripts, and hackathons?
    by /u/N_Sin (Database) on April 16, 2026 at 9:57 pm

    I’m building a small hosted DB-over-API (SaaS) product and I’m trying to validate whether this is actually useful to other developers. The idea is not “replace your real database.” It’s more: if you want to store and query data quickly over HTTP without setting up a full backend, would you use something like this? The use cases I have in mind are things like: quick MVPs small scripts running across different devices hackathons tutorials and demos internal tools prototypes where you just want “data + API” without much setup Example shapes would be something like: GET{{baseurl}}/api/v1/tables/{{tableName}}/{{recordId}} Or GET{{baseurl}}/api/v1/tables/{{tableName}}?filter=done:eq:false&sort=priority:asc,created_at:desc This is not meant to replace any SQL dB for bigger or more serious projects. I’m thinking of it more as a convenience tool for cases where speed and simplicity matter more than full DB power. What I’d really like to know: Would you use something like this? For which use cases would it actually be better than just using Postgres, SQLite, Supabase, Firebase, etc.? If you had heavier usage, would you pay for it? Would you be interested in helping shape the product and giving feedback on design decisions? I would really appreciate blunt feedback, especially from people who have built quick MVPs, hackathon apps, automations, or tutorial projects. Here is a video of how quick set up is: Note that columns id, created_at, updated_at are automatically managed for every table by the api and not by the user. Also in this video example I'm using the infer schema from first write option rather than first creating a schema with the dedicated endpoint (to showcase speed). https://reddit.com/link/1snhsum/video/b792idtyjpvg1/player submitted by /u/N_Sin [link] [comments]

  • Sqlite: Attaching a database for ad-hoc foreign key check?
    by /u/No-Security-7518 (Database) on April 16, 2026 at 7:16 pm

    I have 2 Sqlite databases; Users + Inventory. I have a column in several tables in inventory.db that records which user did things such as: removing/registering a product, etc. What is the cleanest way to achieve data integrity here? 1. Users.db belongs to a library I'm declaring as a dependency. 2. Both databases are copied to a directory at startup so they're next to each other. Should I merge them at startup too? (copy schema +data)? Or use Attach Database? I understand FK checks aren't possible then. So maybe just check the userId is valid? I appreciate your input. submitted by /u/No-Security-7518 [link] [comments]

  • Problem automation galara
    by /u/N4NT4 (Database) on April 16, 2026 at 2:04 pm

    submitted by /u/N4NT4 [link] [comments]

  • Multi Vendor Insurance system best db design
    by /u/Abhinava21 (Database) on April 16, 2026 at 7:55 am

    I am building a module in which I have to integrate multi-vendor insurance using the nestjs and mysql. Mainly our purpose is to do insurance for new E-rickshaws. So, what is the best tables schemas I can create. so, it is scalable and supports multivendor. I have created some of the columns and implemented one of the vendors. But I don't think it is scalable so need advice for the same. submitted by /u/Abhinava21 [link] [comments]

  • many to many binary relationship in ER to relational model but cant do
    by /u/Kerem1111 (Database) on April 14, 2026 at 3:15 pm

    Work assignment is connected to facility and instructors. I want to translate this into a relational model but the issue is, facility has a PK so I just need to include facilityCode in Work assignment table, but instructors or by extension staff doesn't have a PK. How am I supposed to include that? Thanks submitted by /u/Kerem1111 [link] [comments]

  • A LISTEN/NOTIFY debugger that survives reconnects and keeps 10k events in local SQLite
    by /u/rohithgilla (Database) on April 14, 2026 at 4:40 am

    I've rewritten the same 40-line pg.Client listen.js script at least six times on three different laptops. This is the version I wish I'd built the first time. The panel: Subscribes to multiple channels on a connection Persists every event to a local SQLite file (10k per connection ring buffer, enforced in SQL not JS) Reconnects with exponential backoff capped at 30s on drop Re-subscribes to the full current channel set, not the original one (this was a bug the first time — I was losing channels added after initial connect) Quotes channel identifiers properly because LISTEN takes an identifier, not a bindable parameter Writeup with the full reconnect code + the "" identifier-quoting gotcha: https://datapeek.dev/blog/listen-notify-without-tears If anyone has a better answer than exponential backoff for reconnect on pg notification clients, I'd love to hear it. submitted by /u/rohithgilla [link] [comments]

  • Advice on whether nosql is the right choice?
    by /u/puma905 (Database) on April 14, 2026 at 2:40 am

    I’m building a mobile app where users log structured daily entries about an ongoing condition (things like symptoms, possible triggers, actions taken, and optional notes). Over time, the app generates simple summaries and pattern insights based on those logs. Each user has their own dataset, entries are append-heavy with occasional edits, and the schema may evolve as I learn more from real usage. There will be lightweight analytics and AI-driven summaries on top of the data. I would like to be able to also aggregate data across users over time to better understand trends, etc. I’m trying to decide whether a NoSQL document database is the right choice long-term, or if I should be thinking about a relational model from the start. Curious how others would approach this kind of use case. submitted by /u/puma905 [link] [comments]

  • How do you prevent retroactive policy application due to timing gaps between policy updates and enforcement?
    by /u/homerderby (Database) on April 13, 2026 at 8:11 am

    I’ve been looking into an issue where there’s a timing gap between when a policy is announced (or updated in the system) and when the actual enforcement logic is applied. In several cases, transactions that were already completed ended up being evaluated under the new policy rules, which led to inconsistencies and data integrity concerns. From what I can tell, this usually comes from mismatches between the policy DB update timing and the validation/execution layer — older state gets interpreted by a newer rules engine. One approach I’ve been considering is isolating the scope using a snapshot at the time of announcement, combined with a clear grace period to strictly separate timelines. [Attached image: timeline diagram showing policy announcement vs enforcement mismatch] For those working with transactional systems, how do you architect around this? Do you version policies, rely on event sourcing, or enforce strict temporal boundaries at the DB level? I’ve been exploring this problem in a small internal context (oncastudy), and I’m curious what patterns have worked reliably in production. https://preview.redd.it/gu6gihnn1xug1.png?width=1200&format=png&auto=webp&s=bb6bddcf7f66f8550c59e0312722c7e39f8cd170 submitted by /u/homerderby [link] [comments]

  • I can finally screen-share my SQL client without leaking prod data
    by /u/rohithgilla (Database) on April 13, 2026 at 6:52 am

    submitted by /u/rohithgilla [link] [comments]

  • Keeping a Postgres queue healthy
    by /u/2minutestreaming (Database) on April 11, 2026 at 9:23 am

    submitted by /u/2minutestreaming [link] [comments]

  • Json in relational db
    by /u/No_Character_2277 (Database) on April 11, 2026 at 5:34 am

    Is using json or jsonb datatype for a column in relational db really works well in production environment? submitted by /u/No_Character_2277 [link] [comments]

  • A new approach to database queries called GiGI
    by /u/quant-alliance (Database) on April 10, 2026 at 9:46 pm

    Hello community, we are a team of two engineers with experience working for NASA and various other short letter agencies. We took a concept based on non Euclidean geometry called the fiber bundle and built a small database around it. We call this new type of index GiGi and you can see benchmarks and run test here: https://www.davisgeometric.com/gigi#home We are looking for some sort of direction: should we make it open source but we are extremely introverted and not sure we can manage and accumulate a community or should we go for a community Vs enterprise version? do you want to see more benchmarks? which type and what other databases? submitted by /u/quant-alliance [link] [comments]

  • Drew this with AI based on a real incident. Anyone else been here at 3AM?
    by /u/Adela_freedom (Database) on April 10, 2026 at 2:06 pm

    AI-illustrated, but the story is real. Has this happened to your team? How did you fix the access model afterward? submitted by /u/Adela_freedom [link] [comments]

  • A Conversation with Paul Masurel, Creator of Tantivy
    by /u/philippemnoel (Database) on April 9, 2026 at 9:42 pm

    Tantivy is a very popular Rust search library inspired by Apache Lucene. We sat down with Paul, the main author, to discuss how he got started with Rust and Tantivy, and his journey since then. I figured it would be interesting to folks here 🙂 submitted by /u/philippemnoel [link] [comments]

  • Help with normalizing a database?
    by /u/Fit-Try9217 (Database) on April 9, 2026 at 9:25 pm

    Hi! I'm currently working on my project for my database course. I've managed to finish my ERD and relational schema, but when I come to normalize my relational schema, I feel like nothing has changed, and I'm worried I might not be seeing something properly. You can find below the ERD and the unnormalized relational schema! Any help appreciated! https://preview.redd.it/qee3alfmf8ug1.png?width=4386&format=png&auto=webp&s=e8918062be4d316b16fdc2c3cf7aee01c713d824 https://preview.redd.it/f6oaxz9nf8ug1.png?width=1920&format=png&auto=webp&s=b13d29afe86f19fa6fead19e76778ef945ded3cb submitted by /u/Fit-Try9217 [link] [comments]

  • Hi there I'm having a problem in orecal db
    by /u/Im_only0101 (Database) on April 9, 2026 at 6:33 pm

    hi there in using orecal sql developer. Im having table with blob data in a column and I want store it in my computer at a path and add that path to same table with new column. the orecal in on server connect through lan I don't have access to admin or dba and I want use pl/sql. do anyone over internet having idea or solutions of this problem please help #orecal #bug #developer #problemsolving #computerscience #sql #plswl submitted by /u/Im_only0101 [link] [comments]

  • Where do I get started with making webscale DB projects
    by /u/Adventurous_Tea_2198 (Database) on April 9, 2026 at 5:54 pm

    Messed around with sqlite for a project but I can tell it's really shit. submitted by /u/Adventurous_Tea_2198 [link] [comments]

  • Database Help
    by /u/duskti (Database) on April 9, 2026 at 5:04 pm

    I recently joined an organization that requires a robust, scalable database solution for archiving data. Storage needs are projected to reach approximately 100 TB over the next few years, so I’m looking to plan strategically. The data includes a variety of file types—PDFs, Excel files, 3D renderings, and videos—with some individual files as large as 30 GB. We currently have a NAS in place. I’m seeking recommendations for setting this up effectively, ideally with a frontend that allows technicians to upload files directly to the storage system—without incurring high monthly costs from services like AWS S3 or similar cloud providers. submitted by /u/duskti [link] [comments]

  • [ Removed by Reddit ]
    by /u/2011wpfg (Database) on April 9, 2026 at 9:11 am

    [ Removed by Reddit on account of violating the content policy. ] submitted by /u/2011wpfg [link] [comments]

  • How to efficiently run and re-run mysql/mariadb-test-run
    by /u/OttoKekalainen (Database) on April 9, 2026 at 5:59 am

    For anyone doing their first contribution to MySQL or MariaDB: start out by learning how the mysql/mariadb-test-run command works and how to efficiently rebuild the sources and re-run the test suite. submitted by /u/OttoKekalainen [link] [comments]

  • Why does SWITCHOFFSET return the wrong local time when used with the timezone value returned by CURRENT_TIMEZONE?
    by /u/Reasonable-Job4205 (Database) on April 8, 2026 at 8:55 pm

    SYSDATETIMEOFFSET(): 2026-04-08 13:49:06.4745888 -07:00 SYSUTCDATETIME(): 2026-04-08 20:49:06.4745888 CURRENT_TIMEZONE: (UTC-08:00) Pactific Time (US & Canada) SWITCHOFFSET(SYSUTCDATETIME, '-08:00'): 2026-04-08 12:49:06.4745888 -08:00 The correct local time is 13:49, but CURRENT_TIMEZONE returns -08:00, which then causes the computed local time to be 12:49, which is wrong. Why is this? submitted by /u/Reasonable-Job4205 [link] [comments]

  • MongoDB Indexing Recommendation
    by /u/goldenuser22628 (Database) on April 8, 2026 at 1:32 pm

    I’m a bit confused about how to approach indexing, and I’m not fully confident in the decisions I’m making. I know .explain() can help, and I understand that indexes should usually be based on access patterns. The problem in my case is that users can filter on almost any field, which makes it harder to know what the right indexing strategy should be. For example, imagine a collection called dummy with a schema like this: { field1: string, field2: string, field3: boolean, field4: boolean, ... fieldN: ... } If users are allowed to filter by any of these fields, what would be the recommended indexing approach or best practice in this situation? submitted by /u/goldenuser22628 [link] [comments]

  • Neo4j vs ArangoDB for high volume-ingest + multi-hop traversal use case?
    by /u/Klutzy_Plantain1737 (Database) on April 8, 2026 at 12:16 pm

    Hey all — would love to get some real-world perspectives from folks who have used Neo4j and/or ArangoDB in production. We’re currently evaluating graph databases for a use case that involves: • heavy multi-hop traversal (core requirement — this is where graph really shines for us) • modeling relationships across devices, applications, vulnerabilities, etc. • some degree of temporal/state-based data • and moderate to high write volume depending on the window From a querying and traversal perspective, Neo4j has honestly been great. The model feels natural, Cypher is intuitive, and performance on traversal-heavy queries has been solid in our testing. Where we’re running into friction is ingestion. Given our constraints (security + environment), bulk loading into Neo4j Aura hasn’t been straightforward. For large loads, the suggested patterns we’ve seen involve things like: • driver-based ingestion (which is slower for large volumes) • or building/loading externally and restoring into Aura In practice, this has made large-scale ingestion feel like a bottleneck. For heavier loads, we’ve even had to consider taking the database offline overnight to get data in efficiently, which isn’t ideal if this becomes part of regular operations. This has us questioning: • how others are handling high-volume ingestion with Neo4j (especially Aura vs self-managed EE) • whether this is just a constraint of our setup, or a broader limitation depending on architecture ⸻ At the same time, we’re also looking at ArangoDB, which seems more flexible around ingestion (online writes, bulk APIs, etc.), but we’re still trying to understand: • how it compares for deep multi-hop traversal performance • how well it handles complex graph patterns vs Neo4j • any tradeoffs in query ergonomics / modeling ⸻ Questions for the group: 1. If you’re using Neo4j at scale, how are you handling ingestion? • Are you using Kafka / streaming pipelines? • Self-managed EE vs Aura? • Any pain points with large loads? 2. Has anyone used Neo4j Aura specifically for write-heavy or high-ingest workloads? 3. For those who’ve used ArangoDB: • How does it compare for multi-hop traversal performance? • Any limitations vs Neo4j when queries get complex? 4. If you had to choose again for a use case that is: • traversal-heavy • but also requires reliable, ongoing ingestion at scale what would you pick and why? submitted by /u/Klutzy_Plantain1737 [link] [comments]

What is Google Workspace?
Google Workspace is a cloud-based productivity suite that helps teams communicate, collaborate and get things done from anywhere and on any device. It's simple to set up, use and manage, so your business can focus on what really matters.

Watch a video or find out more here.

Here are some highlights:
Business email for your domain
Look professional and communicate as you@yourcompany.com. Gmail's simple features help you build your brand while getting more done.

Access from any location or device
Check emails, share files, edit documents, hold video meetings and more, whether you're at work, at home or on the move. You can pick up where you left off from a computer, tablet or phone.

Enterprise-level management tools
Robust admin settings give you total command over users, devices, security and more.

Sign up using my link https://referworkspace.app.goo.gl/Q371 and get a 14-day trial, and message me to get an exclusive discount when you try Google Workspace for your business.

Google Workspace Business Standard Promotion code for the Americas 63F733CLLY7R7MM 63F7D7CPD9XXUVT 63FLKQHWV3AEEE6 63JGLWWK36CP7WM
Email me for more promo codes

Active Hydrating Toner, Anti-Aging Replenishing Advanced Face Moisturizer, with Vitamins A, C, E & Natural Botanicals to Promote Skin Balance & Collagen Production, 6.7 Fl Oz

Age Defying 0.3% Retinol Serum, Anti-Aging Dark Spot Remover for Face, Fine Lines & Wrinkle Pore Minimizer, with Vitamin E & Natural Botanicals

Firming Moisturizer, Advanced Hydrating Facial Replenishing Cream, with Hyaluronic Acid, Resveratrol & Natural Botanicals to Restore Skin's Strength, Radiance, and Resilience, 1.75 Oz

Skin Stem Cell Serum

Smartphone 101 - Pick a smartphone for me - android or iOS - Apple iPhone or Samsung Galaxy or Huawei or Xaomi or Google Pixel

Can AI Really Predict Lottery Results? We Asked an Expert.

Ace the 2025 AWS Solutions Architect Associate SAA-C03 Exam with Confidence Pass the 2025 AWS Certified Machine Learning Specialty MLS-C01 Exam with Flying Colors

List of Freely available programming books - What is the single most influential book every Programmers should read



#BlackOwned #BlackEntrepreneurs #BlackBuniness #AWSCertified #AWSCloudPractitioner #AWSCertification #AWSCLFC02 #CloudComputing #AWSStudyGuide #AWSTraining #AWSCareer #AWSExamPrep #AWSCommunity #AWSEducation #AWSBasics #AWSCertified #AWSMachineLearning #AWSCertification #AWSSpecialty #MachineLearning #AWSStudyGuide #CloudComputing #DataScience #AWSCertified #AWSSolutionsArchitect #AWSArchitectAssociate #AWSCertification #AWSStudyGuide #CloudComputing #AWSArchitecture #AWSTraining #AWSCareer #AWSExamPrep #AWSCommunity #AWSEducation #AzureFundamentals #AZ900 #MicrosoftAzure #ITCertification #CertificationPrep #StudyMaterials #TechLearning #MicrosoftCertified #AzureCertification #TechBooks

Top 1000 Canada Quiz and trivia: CANADA CITIZENSHIP TEST- HISTORY - GEOGRAPHY - GOVERNMENT- CULTURE - PEOPLE - LANGUAGES - TRAVEL - WILDLIFE - HOCKEY - TOURISM - SCENERIES - ARTS - DATA VISUALIZATION
zCanadian Quiz and Trivia, Canadian History, Citizenship Test, Geography, Wildlife, Secenries, Banff, Tourism

Top 1000 Africa Quiz and trivia: HISTORY - GEOGRAPHY - WILDLIFE - CULTURE - PEOPLE - LANGUAGES - TRAVEL - TOURISM - SCENERIES - ARTS - DATA VISUALIZATION
Africa Quiz, Africa Trivia, Quiz, African History, Geography, Wildlife, Culture

Exploring the Pros and Cons of Visiting All Provinces and Territories in Canada.
Exploring the Pros and Cons of Visiting All Provinces and Territories in Canada

Exploring the Advantages and Disadvantages of Visiting All 50 States in the USA
Exploring the Advantages and Disadvantages of Visiting All 50 States in the USA


Health Health, a science-based community to discuss human health

Today I Learned (TIL) You learn something new every day; what did you learn today? Submit interesting and specific facts about something that you just found out here.

Reddit Science This community is a place to share and discuss new scientific research. Read about the latest advances in astronomy, biology, medicine, physics, social science, and more. Find and submit new publications and popular science coverage of current research.

Reddit Sports Sports News and Highlights from the NFL, NBA, NHL, MLB, MLS, NCAA, F1, and other leagues around the world.

Turn your dream into reality with Google Workspace: It’s free for the first 14 days.
Get 20% off Google Google Workspace (Google Meet) Standard Plan with  the following codes:
Get 20% off Google Google Workspace (Google Meet) Standard Plan with  the following codes: 96DRHDRA9J7GTN6 96DRHDRA9J7GTN6
63F733CLLY7R7MM
63F7D7CPD9XXUVT
63FLKQHWV3AEEE6
63JGLWWK36CP7WM
63KKR9EULQRR7VE
63KNY4N7VHCUA9R
63LDXXFYU6VXDG9
63MGNRCKXURAYWC
63NGNDVVXJP4N99
63P4G3ELRPADKQU
With Google Workspace, Get custom email @yourcompany, Work from anywhere; Easily scale up or down
Google gives you the tools you need to run your business like a pro. Set up custom email, share files securely online, video chat from any device, and more.
Google Workspace provides a platform, a common ground, for all our internal teams and operations to collaboratively support our primary business goal, which is to deliver quality information to our readers quickly.
Get 20% off Google Workspace (Google Meet) Business Plan (AMERICAS): M9HNXHX3WC9H7YE
C37HCAQRVR7JTFK
C3AE76E7WATCTL9
C3C3RGUF9VW6LXE
C3D9LD4L736CALC
C3EQXV674DQ6PXP
C3G9M3JEHXM3XC7
C3GGR3H4TRHUD7L
C3LVUVC3LHKUEQK
C3PVGM4CHHPMWLE
C3QHQ763LWGTW4C
Even if you’re small, you want people to see you as a professional business. If you’re still growing, you need the building blocks to get you where you want to be. I’ve learned so much about business through Google Workspace—I can’t imagine working without it.
(Email us for more codes)