• hector@sh.itjust.works
    link
    fedilink
    English
    arrow-up
    10
    arrow-down
    4
    ·
    9 months ago

    Wow! 13GB! I did some heavy stuff on my computer with like a shit ton of Docker servers running together + deployment and I never reached 13GB!

    Without disclosing private company information lol what are you doing ;)

    • ben_dover@lemmy.ml
      link
      fedilink
      English
      arrow-up
      4
      ·
      edit-2
      9 months ago

      not OP, but I have to run fronted and backend of a project in docker simultaneously (multiple postgres and redis dbs, queues, search index, etc., plus two webservers), plus a few browser tabs and two VSCode instances open, regularly pushes my machine over 15gb ram usage

      pretty much like this

      • Veraxus@lemmy.world
        link
        fedilink
        English
        arrow-up
        1
        ·
        edit-2
        8 months ago

        That is basically my use-case. You add a DB service (or two), DNS, reverse proxy, Redis, Memcached, etc… maybe some containers for additional proprietary backend services like APIs, and then the application themselves that need those things to run… it adds up FAST. The advantage is that you can have multiple projects all running simultaneously and you can add/remove/swap them pretty easily.

        RAM is cheap. There is no excuse for shipping a 8GB computer… even if it’s mostly going to be used for family photos and internet.

    • Veraxus@lemmy.world
      link
      fedilink
      English
      arrow-up
      2
      ·
      8 months ago

      Running a suite of services in containers (DBs, DNS, reverse proxy, memcached, redis, elasticsearch, shared services, etc) plus a number of discreet applications that use all those things. My day-to-day usage hovers around 20GB with spikes to 32 (my max allocation) when I run parallelized test suites.

      Dockers memory usage really adds up fast.