this post was submitted on 01 Nov 2023
44 points (94.0% liked)

Selfhosted

40337 readers
521 users here now

A place to share alternatives to popular online services that can be self-hosted without giving up privacy or locking you into a service you don't control.

Rules:

  1. Be civil: we're here to support and learn from one another. Insults won't be tolerated. Flame wars are frowned upon.

  2. No spam posting.

  3. Posts have to be centered around self-hosting. There are other communities for discussing hardware or home computing. If it's not obvious why your post topic revolves around selfhosting, please include details to make it clear.

  4. Don't duplicate the full text of your blog or github here. Just post the link for folks to click.

  5. Submission headline should match the article title (don’t cherry-pick information from the title to fit your agenda).

  6. No trolling.

Resources:

Any issues on the community? Report it using the report flag.

Questions? DM the mods!

founded 1 year ago
MODERATORS
 

(I know that this is about selfhosting, but I am forced to use cloud services due to it not being viable to selfhost because of DSL internet speeds in my house, and I need this to be accessible outside my home.)

I recently made a Linode account (and got the free credit), and I am planning on only paying $5 a month if I can. I noticed that Nextcloud AIO (from Linode "Marketplace") ran very well on the lowest shared CPU plan (1GB ram, 25GB storage, 1 CPU core (CPU seems to me an AMD Epyc?)).

Will it be okay for me to host a Wordpress website and a Nextcloud instance from the same server? I will be using Docker/Podman, and only I will be using the Nextcloud instance.

you are viewing a single comment's thread
view the rest of the comments
[–] TCB13@lemmy.world -2 points 1 year ago (5 children)

If you do a barebones install / without the Docker overhead it might work.

[–] agressivelyPassive@feddit.de 5 points 1 year ago (4 children)

Docker overhead is practically zero. It's a bit more memory usage, but that's it.

[–] TCB13@lemmy.world -3 points 1 year ago (3 children)
[–] bear 2 points 1 year ago (1 children)

Convincing argument, but unfortunately a cursory Google search will reveal he was right. There is very little CPU overhead. The only real consideration is a bite extra storage and RAM to store and load the redundant dependencies of the container.

[–] TCB13@lemmy.world 1 points 1 year ago (1 children)

You're also ignoring the amount of work the kernel has to do to shift UUIDs around, the resources that the docker daemon itself uses and amounts of redundant stuff to make sure those processes are running that would usually be handled by systemd on a clean system. Yes, containerization is much better nowadays but still overhead.

[–] stardreamer@lemmy.blahaj.zone 1 points 1 year ago

Can't comment much about the docker side since it's not something I'm familiar with.

For the kernel part, assuming what you're referring to as UUIDs is the pid namespace mechanism, I'm failing to see how that would add overhead with containers. The namespace lookups/permission checks are performed regardless of whether the process is in a container or not. There is no fast path for non-containerized processes. The worst overhead that this could add is probably one extra ptr chase in the namespace linked list.

load more comments (1 replies)
load more comments (1 replies)
load more comments (1 replies)