tehnomad

joined 2 years ago
[–] [email protected] 3 points 2 weeks ago

One thing I would do differently is setup LDAP and OIDC so you can use the same authentication credentials for different apps (at least the ones that support them). I use LLDAP and Authelia for this purpose.

[–] [email protected] 1 points 2 weeks ago

I found a VRAM calculator for LLMs here: https://huggingface.co/spaces/NyxKrage/LLM-Model-VRAM-Calculator

Wow it seems like for 128K context size you do need a lot of VRAM (~55 GB). Qwen 72B will take up ~39 GB so you would either need 4x 24GB Nvidia cards or the Mac Pro 192 GB RAM. Probably the cheapest option would be to deploy GPU instances on a service like Runpod. I think you would have to do a lot of processing before you get to the breakeven point of your own machine.

[–] [email protected] 1 points 2 weeks ago (2 children)

The context cache doesn't take up too much memory compared to the model. The main benefit of having a lot of VRAM is that you can run larger models. I think you're better off buying a 24 GB Nvidia card from a cost and performance standpoint.

[–] [email protected] 9 points 1 month ago (3 children)

I would suggest an Intel N100 mini PC if you are planning to transcode video files with Plex. Intel Quick Sync performs better than AMD for media transcoding.

[–] [email protected] 1 points 5 months ago (1 children)

I wasn't sure if it was AI or not. According to the description on GitHub:

Utilizes state-of-the-art algorithms to identify duplicates with precision based on hashing values and FAISS Vector Database using ResNet152.

Isn't ResNet152 a neural network model? I was careful to say neural network instead of AI or machine learning.

[–] [email protected] 0 points 5 months ago (3 children)

Yeah, the duplicate finder uses a neural network to find duplicates I think. I went through my wedding album that had a lot of burst shots and it was able to detect similar images well.

[–] [email protected] 5 points 5 months ago (7 children)

Not sure if you're aware, but Immich has a duplicate finder

[–] [email protected] 3 points 11 months ago* (last edited 11 months ago)

Dockge for docker compose stacks. Glances for system resource usage because it has a Homepage widget.

[–] [email protected] 5 points 11 months ago (1 children)

All good recs so far but haven't seen anyone mention Infernal Affairs.

[–] [email protected] 1 points 1 year ago

I am running my disks as mirrors.

[–] [email protected] 1 points 1 year ago (2 children)

I have a Terramaster 4-bay DAS

[–] [email protected] 4 points 1 year ago (4 children)

I'm running a mini-PC with the N100, 12GB RAM, and 2x18TB mirrored drives on ZFS and it seems to work well.

 

I use my desktop PC for Jellyfin and torrenting, but I'm looking for something that I can keep on 24/7 that draws less power and run other self-hosted services on Linux. I would like to have at least 2x 14 TB 3.5" hard drives in or attached to it with the possibility of expanding in the future.

From my research, these seem to be some good options:

  1. Mini PC like this Beelink S12 Pro + USB hard drive enclosure. The price seems reasonable for the specs and low power consumption. Not sure if USB will limit transfer speeds.
  2. ODROID HC-4 or similar SBCs. I feel like these have much lower performance for not much price savings, and it's harder to get software running up because of ARM. But it seems like they don't use too much power.
  3. Used enterprise PCs/servers. I know they can be found cheap used, but I'm a little lost at comparing the performance and power draw to other options.
  4. DIY build. I'm interested in getting a Mini-ITX case like this Jonsbo N2 and getting parts for it, but it seems like it will be the most expensive option. It does seem like the most modular and upgradable.
  5. Classic NAS products like Synology. It seems like these are falling out of favor because they are pretty under powered for the price.

What does selfhosted think about these options, and what would you recommend?

view more: next ›