How does a database handle pagination?

How does a database handle pagination?

AI Dashboard is available on the Web, Apple, Google, and Microsoft, PRO version

How does a database handle pagination?

How does a database handle pagination?

It doesn’t. First, a database is a collection of related data, so I assume you mean DBMS or database language.

Second, pagination is generally a function of the front-end and/or middleware, not the database layer.

But some database languages provide helpful facilities that aide in implementing pagination. For example, many SQL dialects provide LIMIT and OFFSET clauses that can be used to emit up to n rows starting at a given row number. I.e., a “page” of rows. If the query results are sorted via ORDER BY and are generally unchanged between successive invocations, then that can be used to implement pagination.

That may not be the most efficient or effective implementation, though.

Get 20% off Google Google Workspace (Google Meet) Standard Plan with  the following codes: 96DRHDRA9J7GTN6
Get 20% off Google Workspace (Google Meet)  Business Plan (AMERICAS) with  the following codes:  C37HCAQRVR7JTFK Get 20% off Google Workspace (Google Meet) Business Plan (AMERICAS): M9HNXHX3WC9H7YE (Email us for more codes)

How does a database handle pagination?

So how do you propose pagination should be done?

On context of web apps , let’s say there are 100 mn users. One cannot dump all the users in response.

Cache database query results in the middleware layer using Redis or similar and serve out pages of rows from that.

What if you have 30, 000 rows plus, do you fetch all of that from the database and cache in Redis?

I feel the most efficient solution is still offset and limit. It doesn’t make sense to use a database and then end up putting all of your data in Redis especially data that changes a lot. Redis is not for storing all of your data.

AI Unraveled: Demystifying Frequently Asked Questions on Artificial Intelligence (OpenAI, ChatGPT, Google Gemini, Generative AI, Discriminative AI, xAI, LLMs, GPUs, Machine Learning, NLP, Promp Engineering)

If you have large data set, you should use offset and limit, getting only what is needed from the database into main memory (and maybe caching those in Redis) at any point in time is very efficient.

With 30,000 rows in a table, if offset/limit is the only viable or appropriate restriction, then that’s sometimes the way to go.

More often, there’s a much better way of restricting 30,000 rows via some search criteria that significantly reduces the displayed volume of rows — ideally to a single page or a few pages (which are appropriate to cache in Redis.)

It’s unlikely (though it does happen) that users really want to casually browse 30,000 rows, page by page. More often, they want this one record, or these small number of records.

If you are looking for an all-in-one solution to help you prepare for the AWS Cloud Practitioner Certification Exam, look no further than this AWS Cloud Practitioner CCP CLF-C02 book


Question: This is a general question that applies to MySQL, Oracle DB or whatever else might be out there.

I know for MySQL there is LIMIT offset,size; and for Oracle there is ‘ROW_NUMBER’ or something like that.

But when such ‘paginated’ queries are called back to back, does the database engine actually do the entire ‘select’ all over again and then retrieve a different subset of results each time? Or does it do the overall fetching of results only once, keeps the results in memory or something, and then serves subsets of results from it for subsequent queries based on offset and size?

If it does the full fetch every time, then it seems quite inefficient.

If it does full fetch only once, it must be ‘storing’ the query somewhere somehow, so that the next time that query comes in, it knows that it has already fetched all the data and just needs to extract next page from it. In that case, how will the database engine handle multiple threads? Two threads executing the same query?

something will be quick or slow without taking measurements, and complicate the code in advance to download 12 pages at once and cache them because “it seems to me that it will be faster”.

Answer: First of all, do not make assumptions in advance whether something will be quick or slow without taking measurements, and complicate the code in advance to download 12 pages at once and cache them because “it seems to me that it will be faster”.

YAGNI principle – the programmer should not add functionality until deemed necessary.
Do it in the simplest way (ordinary pagination of one page), measure how it works on production, if it is slow, then try a different method, if the speed is satisfactory, leave it as it is.

From my own practice – an application that retrieves data from a table containing about 80,000 records, the main table is joined with 4-5 additional lookup tables, the whole query is paginated, about 25-30 records per page, about 2500-3000 pages in total. Database is Oracle 12c, there are indexes on a few columns, queries are generated by Hibernate. Measurements on production system at the server side show that an average time (median – 50% percentile) of retrieving one page is about 300 ms. 95% percentile is less than 800 ms – this means that 95% of requests for retrieving a single page is less that 800ms, when we add a transfer time from the server to the user and a rendering time of about 0.5-1 seconds, the total time is less than 2 seconds. That’s enough, users are happy.

And some theory – see this answer to know what is purpose of Pagination pattern

Djamgatech: Build the skills that’ll drive your career into six figures: Get Djamgatech.

  • AI for the Win: Conquering Database Management with Artificial Intelligence
    by Rama Desai (Database on Medium) on May 24, 2024 at 5:10 pm

    AI for the Win: Conquering Database Management with Artificial IntelligenceContinue reading on Medium »

  • Spatial Database Capabilities in Python for Geospatial Analysis and Visualization
    by Gamze Ç. (Database on Medium) on May 24, 2024 at 5:03 pm

    PostGIS serves as a spatial database extender for the PostgreSQL object-relational database system. Essentially, PostGIS integrates GIS…Continue reading on Medium »

  • Beyond the basics: Can AI-powered SQL tools handle complex data-wrangling tasks or are they best…
    by Rama Desai (Database on Medium) on May 24, 2024 at 4:57 pm

    AI and the SQL Apprentice: Conquering Data Mountains or Just Scaling Hills?Continue reading on Medium »

  • read-only key-value store, open source
    by /u/benjamin-crowell (Database) on May 24, 2024 at 4:49 pm

    I have some open-source linguistics software I've written. If you were to download and use it in its present state, you would have to run a big computation overnight, which would generate a 600 Mb sqlite database, and then the software would be ready to run. If I actually want people to use my software, this is a big barrier to entry. I could post the sqlite file on my server and have users download it, but that's still a time-consuming extra step for them. (It does compress down to about one tenth of the original size with rzip.) So to avoid any barrier to entry it seems like a reasonable option is just to set up some kind of publicly accessible database on my server. From the point of view of a user, this would be so simple as to be invisible. The software knows how to phone home for its data, and users who don't want that can build their own copy of the database. I could set up mariadb and make it accessible over port 3306, but it seems complicated to configure it securely so as to rule out every sql operation that the software doesn't need to do, like write operations or joins. In fact, the only operation my software legitimately ever has to do is select on a certain column and read back the row, i.e., I'm really just doing key-value storage. Any suggestions on how to do this in the simplest possible way? I guess what I need is a 600 Mb read-only key-value storage, using only open-source software, which is exposed to the public internet on an open port. I don't need caching, and users are currently zero, so performance isn't a huge issue. I'm tempted to roll my own, because it seems like something you really could do with 30 lines of code, but I would prefer a solution that was more like "install this open-source database using the debian package, and then read the docs to see how to configure it for your purpose, which is simple." submitted by /u/benjamin-crowell [link] [comments]

  • Optimización del Almacenamiento de Datos en un Sistema de Ventas en Línea
    by Kevinleith7 (Database on Medium) on May 24, 2024 at 4:16 pm

    En el acelerado ámbito del comercio electrónico, la organización y almacenamiento eficiente de datos son fundamentales para tomar…Continue reading on Medium »

  • Differences Between Stream Processing Engines and Streaming Databases
    by RisingWave Labs (Database on Medium) on May 24, 2024 at 3:34 pm

    Design, Use Cases, and the FutureContinue reading on ILLUMINATION »

  • Is AI for SQL a game-changer for data newbies or a productivity booster for seasoned analysts?
    by Rama Desai (Database on Medium) on May 24, 2024 at 2:49 pm

    AI for SQL: Game Changer for Newbies or Analyst Power-Up?Continue reading on Medium »

  • SQL ninjas vs. AI apprentices: Who will reign supreme in the age of database manipulation?
    by Rama Desai (Database on Medium) on May 24, 2024 at 2:33 pm

    SQL Ninjas vs. AI Apprentices: A Collaborative Dance in the Database DojoContinue reading on Medium »

  • How to Migrate a WordPress Website to New Host
    by Jackson Lane (Database on Medium) on May 24, 2024 at 1:59 pm

    Is your WordPress website feeling sluggish? Maybe your current hosting plan doesn’t offer the features or bandwidth you need to handle…Continue reading on Medium »

  • KDB Vector DB is ranked top on DBEngines; How?
    by /u/ThinkSatisfaction971 (Database) on May 24, 2024 at 1:50 pm

    As per this ranking list -> KDB is the most popular, while the next five(rank 2-6) are pretty well known in the community, I have not come across KDB as much in the wild. Are people really using it? Anyone using it can please share some pros and cons of using anything else vs this. I would like to use it for some personal project. submitted by /u/ThinkSatisfaction971 [link] [comments]

  • SQL 101
    by Kübra Bayburtli (Database on Medium) on May 24, 2024 at 1:37 pm

    Sql nedir?Continue reading on Medium »

  • Database Replication Consistency with Single Leader
    by Dipak Kr das (Database on Medium) on May 24, 2024 at 1:12 pm

    It’s common to have one more more replica of database in the production environment. Replica ensures high availability, increased read…Continue reading on Medium »

  • Best way to store and access JSON datasets.
    by /u/qmr55 (Database) on May 24, 2024 at 4:39 am

    Hi database! I am currently working on a nodejs project that will run locally. It fetches data from an API frequently, but a subset of this data rarely changes. I want to store this subset locally in JSON datasets, but I need to be able to call to it and retrieve it extremely quick. The goal is to be able to retrieve it quicker than I would from an API call. Do you guys have any suggestions on the best way to go about this? So far I’ve come up with Redis which looks like a decent solution, as well as storing in a csv cache file. Any better ideas or suggestions? The only thing I care about is speed. Thanks! submitted by /u/qmr55 [link] [comments]

  • Which database should I use?
    by /u/learning-machine1964 (Database) on May 24, 2024 at 2:40 am

    Is there a free database that can store thousands of rows with 15 columns for free? I'm thinking more of cloud solutions because I don't want to save so many data locally on my laptop for storage reasons. Also, I want to be able to easily find similar data based on their tags. For example, if two data have similar tags, then I want to be able to easily return the other data given one of them. submitted by /u/learning-machine1964 [link] [comments]

  • Your guide to Vector Databases
    by /u/TheLostWanderer47 (Database) on May 24, 2024 at 2:37 am

    submitted by /u/TheLostWanderer47 [link] [comments]

  • Database System Concepts - 6th edition enquiry
    by /u/happybaby00 (Database) on May 23, 2024 at 10:37 pm

    Does anyone know if the practice exercises alone are enough to test your knowledge or if you are to do the full exercises. Asking because there's no way for me to get the answers to the full exercises to check answers. Thanks in advance submitted by /u/happybaby00 [link] [comments]

  • Postgres+ graph queries vs graph db
    by /u/infinitypisquared (Database) on May 23, 2024 at 9:52 pm

    I am building a social network plus product recommendation app. Been researching a ton. What are the pros and cons. Fb still using Mysql with a graph layer Insta uses Postgres + graph querying Linked user graph db (guess its Neo4j) Confused about the pros and cons of different approaches and what are different parameters I should evaluate to make decision submitted by /u/infinitypisquared [link] [comments]

  • Creating an online database with a web access for a travel agency?
    by /u/Mathue24 (Database) on May 23, 2024 at 8:30 pm

    Hey! So I've just been offered a gig. Its pretty simple really, they are a small company and are looking for something to replace their exel tables for when it comes to clients, bookings, etc. Their budget is really small though, so what's the easiest way to create the database and the web client for accessing and modifying the database with forms for the employees? EDIT: something like this submitted by /u/Mathue24 [link] [comments]

  • Has anyone tried using MongoDB Atlas or MS Cosmos DB as a single primary store instead of using ElasticSearch on top of a SQL DB?
    by /u/ScaleApprehensive926 (Database) on May 23, 2024 at 5:45 pm

    I currently work in an environment where we use SQL Server as a primary store and then transform stuff into JSON and inject into ElasticSearch for searching and fast UI needs. This approach can be a PITA with a predisposition for accumulating duplication of logic. I've heard that some companies are starting to use things like MongoDB with Atlas, or MS Cosmos as a single primary store instead of structuring everything in SQL then adding ES on top of it. Does anyone have experience with this that would like to share? If I ever end up green-fielding something from the ground up I will certainly consider this approach in hopes of eliminating lots of JSON to SQL code. I understand that there is still a place for relational SQL even if you do take this approach (perhaps for things like permissions/security), but I would want to take a stab at putting most actual data records into a primary store that used JSON natively and had all the search power of ES. submitted by /u/ScaleApprehensive926 [link] [comments]

  • Better Design Choice
    by /u/JonDream00 (Database) on May 22, 2024 at 11:17 pm

    Hi! I'm new to the sub, not sure if this kind of post is allowed, but I didn't saw anything against it in the rules, so... I'm an amateur when it comes to database designing and right now I'm designing a database for a personal project, but I'm stuck at something. So, I have two tables Abilities and AbilityEffects, and ability can have multiple ability effects so it's a one-to-many relationship (these effects are unique to each ability, that's why it's not a many-to-many). I'm trying to add another table AbilityAugment, this will have a many-to-one with Abilities (one ability can have multiple augments), but, an augment can also have multiple ability effects (basically, an augment adds effects to an ability). The problem is, of course, AbilityEffects already has a relationship with Abiltiy so I'm struggling with this last relationship. I have come to 3 alternatives, but I'm not sure which one is better. AbilityEffects Copy The first one will be simply to create a copy of AbilityEffects called AugmentEffects and add the many-to-one with AbilityAugment like this: The problem here is that as I continue to design my DB, the effects tables will likely become more complex and I'll have to add everything twice Double relationship AbilityEffects with Validation Another option is to add augmentId to AbilityEffects creating a double relationship many-to-one with Abilities and AbilityAugment but adding a validation to make sure that no record will have an abilityId and augmentId at the same time I don't know about this one, I feel like it makes my design more confusing, more likely to fail, and not a great design practice. More tables, without copies Finally, I thought about splitting AbilityEffects into AbilityEffects and Effects and adding (again) AugmentEffects. But in this case, AbilityEffects and AugmentEffects will be intermediate tables, having many-to-one relationships with Abilities and AbilityAugments and one-to-one with Effects I liked this one better because I feel like it's easier to understand and it makes it easier to expand my design. But I guess more tables will make my CRUD operations harder, so not so sure if this is the best option. So, what do you think? which one is better? Or is there a better way to accomplish this? Anyway, thank you so much if you took the time to read this long post and help me with my silly project haha! submitted by /u/JonDream00 [link] [comments]

  • Reverse engineer structure
    by /u/dino_dog (Database) on May 22, 2024 at 6:33 pm

    Hi folks, How would you reverse engineer the schema/architecture of a database? I can export all the data to excel (all attributes) but that’s about it. Any help or links to help would be great. submitted by /u/dino_dog [link] [comments]

  • Sharing Experience - Advice Accepted
    by /u/scriptgamer (Database) on May 22, 2024 at 1:21 pm

    So, I've been working with Oracle DB, PL/SQL for as long as I can remember. I do not consider myself "one of the best" but I've been able to solve every problem so far. Recently I was hired by a Fintech company, which has been growing the last years. They have a Java application that uses an Oracle DB of course. Since the beggining I noticed that they tend to have all business logic in Java and avoid the database as much as possible. The database has like 150+ "table_v1, _v2...." for each original table, they just keep creating new versions for each change the table needs and they forget about the old ones. They have multiple tipes of records on the same tables, so if you want to check LOG from the app, you need to check the "main_table" where record_type = 'LOG' basically... you want financial info? query the same table where record_type = 'FIN' and so on. Now that the company is growing, this table is growing, and they have reporting tools accessing transactional tables to generate the reports. I understand that the changes that are necessary here involve re-modeling, partitioning, better queries, separation of diffent "data-sources" for each purpose (AKA BI vs DB)... so on.... But when I try to suggest that they receive it like "it was working until now, you don't know how to fix"... So, my question is, in more than 10 years in this role, I have never seen a company reestructure the whole database, If this happens, I'll be basically the only responsible for everything. Have you guys seen something like this ? is it worth trying to change "the whole thing" ? Should I just keep trying to deal with shitty stuff that will never perform because the db is a mess ? submitted by /u/scriptgamer [link] [comments]

  • Vector Search - HNSW Explained
    by /u/Personal-Trainer-541 (Database) on May 22, 2024 at 9:39 am

    submitted by /u/Personal-Trainer-541 [link] [comments]

  • Storing knowledge in a single long plain text file
    by /u/breck (Database) on May 21, 2024 at 8:41 pm

    submitted by /u/breck [link] [comments]

  • Database design for Google Calendar: a tutorial
    by /u/squadette23 (Database) on May 21, 2024 at 5:55 pm

    submitted by /u/squadette23 [link] [comments]

  • What to learn?
    by /u/Suspicious-Fox6253 (Database) on May 21, 2024 at 5:27 pm

    I know all the basics of databases. I have worked with MySQL and I know almost all the CRUD operations. What should my next step be? Should I learn PL/SQL or should I learn SQL with python/java? I am a CS student. submitted by /u/Suspicious-Fox6253 [link] [comments]

  • Good cloud database
    by /u/Mrcool654321 (Database) on May 21, 2024 at 12:33 am

    I am making a video platform and I decided not to build servers what is a really cheap company for storage submitted by /u/Mrcool654321 [link] [comments]

  • How to clean up MS Access?
    by /u/kalilikoi (Database) on May 20, 2024 at 9:34 pm

    I posted here a while back and here I am again! I need some hand-holding. A recap: I am being tasked with taking over a referral program for my company. They have been using MS Access since 2001 and only the accountant has used it. The person who created this was contracted to assist me, but has ghosted us for nearly 2 months. This is my first time using Access/a database. Ultimately, they want me to find a new solution. Here’s what our database does: Referral Program - we have many companies & their staff signed up with a referral code (numbers are from 1-12000+). We provide them with physical cards to hand out at their places of work and when we receive those cards back from customers, our retail employees tally the amount of customers in the party on the card, and I enter them into the database with the corresponding referral code + amount of guests. At the end of each month, I am to print the report and calculate the commision for each employee/company. We manually calculate by hand this as the database has old rates of $2 a card, and now we have a tier based system that changes every season. Our accountant then cuts the checks per the updated calculations. Questions: What I cannot figure out is how to clean out all these old companies & employees who are no longer partnered with us. I also cannot figure out how to change the rates as needed. Then.. once I fix this to be streamlined, they expect me to find a new system to use and get rid of Access. I had a lot of great responses on my last post but I am honestly hoping there’s something already built out there where I can just plug in our active accounts and move on, lol. They’re still trying to convince me of Salesforce… submitted by /u/kalilikoi [link] [comments]

  • Optimizing SQL Queries - PostgreSQL best practices
    by /u/waterslurpingnoises (Database) on May 20, 2024 at 5:39 pm

    submitted by /u/waterslurpingnoises [link] [comments]

  • Should I index the hell out of my read-only table?
    by /u/Turings-tacos (Database) on May 19, 2024 at 8:59 pm

    I have a table with like 20 columns that have Boolean values if the row has a characteristic or not. This is a fairly small table (8k rows) and is rarely updated/deleted and only done so by admin. This table is used in a search page for my site to return results. Do you think I should I just index every searchable column to optimize its performance? submitted by /u/Turings-tacos [link] [comments]

  • The most space efficient database?
    by /u/nikowek (Database) on May 19, 2024 at 6:47 pm

    I am a data hoarder. I have several databases at home, including a PostgreSQL database that takes up over 20TB of space. I store everything there—from web pages I scrape (complete HTML files, organized data, and scraping logs) to sensor data and data exported from Prometheus or small files (i know i shouldn't). You can definitely call me a fan of this database because, despite its size, it still runs incredibly fast on slow HDDs (Seagate Basic). For fun, I put most of the same data into MongoDB and, to my surprise, 19,270 GB of data occupies only 3,447 GB there. I measure this by invoking db.stats(1024*1024*1024) and then comparing by looking at dataSize and totalSize. Access to most of the data is managed through the value stored in PostgreSQL. Now, my question is, is there any database that will provide me with fast access to data on a hard disk while offering better compression? I am happy to test your suggestions! As it's home lab environment, i would like to avoid paid solutions. submitted by /u/nikowek [link] [comments]

  • Bulk insert challange on MongoDB vs APIs (Oracle 23ai and FerretDB)
    by /u/riddinck (Database) on May 19, 2024 at 10:47 am

    The One Billion Row Challenge (1BRC) presented by Gunnar Morling invited Java developers to aggregate and summarize a large volume of data. At the PGDay Ankara Conference, Murat Tuncer showcased FerretDB as a seamless alternative to MongoDB, aligning with the trend of “Just use Postgres for everything.” The Oracle MongoDB API and FerretDB allow developers to use MongoDB syntax with relational database management systems. I’ve decided to put both FerretDB and the Oracle Database 23ai MongoDB API to the test to see if they can serve as alternatives to MongoDB in the realm of NoSQL databases. Seamless NoSQL Solutions submitted by /u/riddinck [link] [comments]

  • need help with erd in text format(?)
    by /u/OwOwhatsthiiis (Database) on May 19, 2024 at 9:52 am

    hi! i’m taking a beginners class in database design and i missed the lesson where we were taught to write out our entity relationship diagram in a text sort of format? it’s meant to look something like this: employee_card = (card_no) K = {{card_no}} sorry if this doesnt make a lot of sense, like i said it was a beginner class and im not very familiar with everything, i would appreciate any resources or if someone would be down to dm and help me understand! thanks!! submitted by /u/OwOwhatsthiiis [link] [comments]

  • A question about graph databases
    by /u/za3b (Database) on May 19, 2024 at 6:29 am

    Hello everyone, I'm a web developer. I'm used to MongoDB and Node.js. I want to create a pet project, a social network. My first question is: which is better for this project. Surreal DB or Memgraph (I haven't used neither before). My second question is: how can I use a graph DB in a microservice project. Do I self host the DB with each microservice? If yes, how is the data shared between them without duplicates? Note: I'm going to use Node.js for this project. And lastly, if anyone has insights. Please do share them. Thanks in advance... submitted by /u/za3b [link] [comments]

  • Where to generate data to a database (university project)
    by /u/Academic_North1040 (Database) on May 18, 2024 at 8:37 pm

    submitted by /u/Academic_North1040 [link] [comments]

Ace the 2023 AWS Solutions Architect Associate SAA-C03 Exam with Confidence Pass the 2023 AWS Certified Machine Learning Specialty MLS-C01 Exam with Flying Colors

List of Freely available programming books - What is the single most influential book every Programmers should read

#BlackOwned #BlackEntrepreneurs #BlackBuniness #AWSCertified #AWSCloudPractitioner #AWSCertification #AWSCLFC02 #CloudComputing #AWSStudyGuide #AWSTraining #AWSCareer #AWSExamPrep #AWSCommunity #AWSEducation #AWSBasics #AWSCertified #AWSMachineLearning #AWSCertification #AWSSpecialty #MachineLearning #AWSStudyGuide #CloudComputing #DataScience #AWSCertified #AWSSolutionsArchitect #AWSArchitectAssociate #AWSCertification #AWSStudyGuide #CloudComputing #AWSArchitecture #AWSTraining #AWSCareer #AWSExamPrep #AWSCommunity #AWSEducation #AzureFundamentals #AZ900 #MicrosoftAzure #ITCertification #CertificationPrep #StudyMaterials #TechLearning #MicrosoftCertified #AzureCertification #TechBooks

zCanadian Quiz and Trivia, Canadian History, Citizenship Test, Geography, Wildlife, Secenries, Banff, Tourism

Africa Quiz, Africa Trivia, Quiz, African History, Geography, Wildlife, Culture

Exploring the Pros and Cons of Visiting All Provinces and Territories in Canada.
Exploring the Pros and Cons of Visiting All Provinces and Territories in Canada

Exploring the Advantages and Disadvantages of Visiting All 50 States in the USA
Exploring the Advantages and Disadvantages of Visiting All 50 States in the USA

Health Health, a science-based community to discuss health news and the coronavirus (COVID-19) pandemic

Today I Learned (TIL) You learn something new every day; what did you learn today? Submit interesting and specific facts about something that you just found out here.

Reddit Science This community is a place to share and discuss new scientific research. Read about the latest advances in astronomy, biology, medicine, physics, social science, and more. Find and submit new publications and popular science coverage of current research.

Reddit Sports Sports News and Highlights from the NFL, NBA, NHL, MLB, MLS, and leagues around the world.

Turn your dream into reality with Google Workspace: It’s free for the first 14 days.
Get 20% off Google Google Workspace (Google Meet) Standard Plan with  the following codes:
Get 20% off Google Google Workspace (Google Meet) Standard Plan with  the following codes: 96DRHDRA9J7GTN6 96DRHDRA9J7GTN6
With Google Workspace, Get custom email @yourcompany, Work from anywhere; Easily scale up or down
Google gives you the tools you need to run your business like a pro. Set up custom email, share files securely online, video chat from any device, and more.
Google Workspace provides a platform, a common ground, for all our internal teams and operations to collaboratively support our primary business goal, which is to deliver quality information to our readers quickly.
Get 20% off Google Workspace (Google Meet) Business Plan (AMERICAS): M9HNXHX3WC9H7YE
Even if you’re small, you want people to see you as a professional business. If you’re still growing, you need the building blocks to get you where you want to be. I’ve learned so much about business through Google Workspace—I can’t imagine working without it.
(Email us for more codes)

error: Content is protected !!