• 0 Posts
  • 46 Comments
Joined 1 year ago
cake
Cake day: June 15th, 2023

help-circle


  • It might be worth taking a step back and looking at your objective with all of this and why you are doing it in the first place.

    If it’s for privacy, then unfortunately that ship has sailed when it comes to email. It’s the digital equivalent of a post card. It’s inherently not private. Nothing you do will make it private. Even services like proton Mail aren’t private–unless you only email other people on proton.

    I appreciate wanting to control your own destiny with it but there are much more productive things you could be spending your time on the improve your privacy surface area.




  • cybersandwich@lemmy.worldtoSelfhosted@lemmy.world*Permanently Deleted*
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    edit-2
    5 months ago

    GPU with a ton of vran is what you need, BUT

    An alternate solution is something like a Mac mini with an m series chip and 16gb of unified memory. The neural cores on apple silicon are actually pretty impressive and since they use unified memory the models would have access to whatever the system has.

    I only mention it because a Mac mini might be cheaper than GPU with tons of vram by a couple hundred bucks.

    And it will sip power comparatively.

    4090 with 24gb of vram is $1900 M2 Mac mini with 24gb is $1000


  • Like he was saying, it’s more than just power loss. It’s a way of “sanitizing” the power as it comes in. This is “usually” not a problem. But dirty power is arguably worse than power outages. If the voltages fluctuate or get low for whatever reason that puts a big strain on your power supplies.

    This could happen because you run a vacuum on the same circuit and your house is old, guy down the street electrocutes himself or the power coming in from the electric company is ‘dirty’ because they have an issue with transformers or up stream somewhere. It can be imperceptible to you, but your tech notices.





  • cybersandwich@lemmy.worldtoSelfhosted@lemmy.worldPost your Servernames!
    link
    fedilink
    English
    arrow-up
    4
    arrow-down
    1
    ·
    7 months ago

    “rocinante” for my proxmox host.

    “awkward, past his prime, and engaged in a task beyond his capacities.” From don Quixote’s wiki page.

    It seemed fitting considering it is a server built from old PC parts…engaged in tasks beyond its abilities.

    The rest of my servers (VMs moslty) are named for what they actually do/which vlan they are on (eg vm15) and aren’t fun or excitin names. But at least I know if I am on that VM it has access to that vlan(or that it’s segregated from my other networks).



  • I don’t have nearly that much worth backing up(5TB–and realistically only 2TB is probably critical), but I have a Synology Nas(12TB raid 1) and truenas (zfs striped/mirrored) that I back my stuff to (and they back up to each other).

    Then I have a raspberry pi with a USB drive (8tb) at my parents house 4 hours away, that my Synology backs up to (over tailscale).

    Oh, and I have a USB HDD(8tb) that I plug in and backup my Synology Nas to and throw in my fireproof safe. But thats a manual backup I do once every quarter or 6 months if I remember. That’s a very very last resort backup.

    My offsite is at my parents.

    And no, I have not tested it because I don’t know how I’m actually supposed to do that.




  • Breaking things is the best way to learn. Accidentally deleting your container data is one of the best ways to learn how to not do that AND learn about proper backups.

    Breaking things and then trying to restore from a backup that…doesn’t work. Is a great way to learn about testing backups and/or properly configuring them.

    The corrolary to this is: just do stuff. Analysis paralysis is real. You can look up a dozen “right ways” to do things and end up never starting.

    My advice: just start. If you end up backing yourself into a corner where you can’t scale or easily migrate to another solution, oh well. You either learn that lesson or figure out a way to migrate. Learning all along the way.

    Each failure or screw up is worth a hundred “best practice / how to articles”.





  • There are different audiences for demos though. It should be that way at the “working level”. When you start moving up the chain with more senior leadership in your org, it starts to make more sense to have the PM do the demos/briefs.

    Usually devs don’t particularly care or want to and sometimes they aren’t really qualified to–its not their skillset. But if it’s a good PM, that’s where they shine. That’s the value they bring to the project. They (should) know the politics, landmines, things that specific leaders would care about (and to highlight for them), and how to frame it to current business needs. They have the context to understand when a seemingly innocuous question is actually pointed. They might not know the intricacies of your code, but they (should) know the intricacies of the organization. That’s not something most developers know, and why should they? That’s not their job.

    Sometimes it even involves groundwork meetings and demos to make sure you have support from other key components in your org-- like getting your CTO excited because one of his performance goals was x and your project is the first real implementation of x. Now, you have the CTO ready to speak on your behalf in front of the CIO. As a PM you know that the CIO has been getting flack from the CFO because there hasn’t been a good way to capture costs for Y, but your system starts the org down the path to fix that. Now they are both excited about the project and in your corner. Etc etc