How do you guys quickly sync your settings (especially bash aliases and ssh keys) across your machines?

Ideally i want a simple script to run on every new server I work with. Any suggestions?

  • Baron Von J
    link
    451 year ago

    I suggest you don’t sync SSH keys. That’s just increasing the blast radius of any one of those machines being compromised.

    • @RegalPotoo
      link
      English
      101 year ago

      Exactly this. Don’t move private keys between machines. Generate them where you need them, it’s not like they cost anything

      • Baron Von J
        link
        41 year ago

        Fair point, but I would equate that with syncing the authorized_keys file rather than thinking about how to sync the keys.

    • @wmassingham
      link
      21 year ago

      Right. Use some kind of centralized authentication like freeipa.

      For bash aliases, I just pull down a .bashrc from github gists.

      • PHLAK
        link
        English
        101 year ago

        OP should just generate a unique SSH key per device (+ user).

        • Muddybulldog
          link
          fedilink
          English
          41 year ago

          Agreed. I’ve probably got 100 keys registered with GitHub and 98 of them the private key is long destroyed due to OS reinstalls or whatnot. Format machine, new key. New machine, new key.

          • PHLAK
            link
            English
            11 year ago

            FYI: You can remove the old keys from GitHub.

        • @dm_me_your_feet
          link
          21 year ago

          I d rather have 2 to 3 (for critical, mid, and test systems) ssh keys that are regularly rotated than 1 key per machine. I m not gonna balance 50 ssh keys; neither enter my password every time i jump hosts.

      • mFatOP
        link
        fedilink
        English
        11 year ago

        Is the url is easy to rember?

  • Atemu
    link
    fedilink
    181 year ago

    Dotfiles go in git, SSH keys are state.

    I’m looking to migrate to home-manager though because I use Nix on all my devices anyways.

    • @[email protected]
      link
      fedilink
      11 year ago

      I also have multiple versions of by bash_profile with syntax specific to the OS. It checks if we’re on MacOS or Linux with a kernel check and then reads the appropriate ancillary bash_profile for that platform. Anything that can live in the main bash_profile with the same command on both platforms lives there and anything that needs to be system-specific is in the other one.

      I have all my important functions as individual files that get loaded with the following:

      function loadfuncs() {
      	local funcdir="$HOME/.dotfiles/functions/"
      	[ -e "${funcdir}".DS_Store ] && rm "$HOME/.dotfiles/functions/.DS_Store"
      	local n=0
      
      	for i in "${funcdir}"*; do
      		source "$(realpath $i)"
      		n=$(( n + 1 ))
      	done
      }
      loadfuncs
      
      
      • Atemu
        link
        fedilink
        11 year ago

        Interesting way to go about it. Though when I’m at the point where I need differences between linux and darwin, I’m probably going to do that at the home-manager level.

        • @[email protected]
          link
          fedilink
          11 year ago

          Just for fun, here’s how I’m checking that (this was written in 2016 and may require adjusting as I haven’t been keeping up on Linux for a while):

          function oscheck() {
          	if [[ "$(uname -s)" == 'Darwin' ]]; then
          
          		# echo Darwin
          		osType=Darwin
          		return 0
          
          	elif
          		[[ "$(uname -s)" == 'Linux' ]]; then
          
          		# echo Linux
          		osType=Linux
          
          		grep CentOS /etc/os-release > /dev/null
          		if [[ "$?" == 0 ]]; then
          		    # echo "CentOS"
          		    export theDistro=CentOS
          		    return 0
          		else
          			:
          		fi
          
          		grep Ubuntu /etc/os-release > /dev/null
          		if [[ "$?" == 0 ]]; then
          		    export theDistro=Ubuntu
          		    return 0
          		else
          			:
          			# echo "Not Ubuntu"
          		fi
          
          		printf "  %s\n" "Error: osType tested true for Linux, but did not find CentOS or Ubuntu." ""
          		return 1
          
          	else
          		osType=Unknown
          		return 1
          	fi
          }
          oscheck
          
  • @[email protected]
    link
    fedilink
    161 year ago

    I’m surprised no one mentioned ansible yet. It’s meant for this (and more).

    By ssh keys I assume you’re talking about authorized_keys, not private keys. I agree with other posters that private keys should not be synced, just generate new ones and add them to the relevant servers authorized_keys with ansible.

    • @[email protected]
      link
      fedilink
      English
      21 year ago

      I use Ansible for this as well. It’s great. I encrypt secrets with Ansible vault and then use it to set keys, permissions, config files, etc. across my various workstations. Makes setup and troubleshooting a breeze.

    • @dinosaurdynasty
      link
      21 year ago

      If the keys are password protected… eh why not sync them.

      Also ssh certificates are a thing, they make doing that kind of stuff way easier instead of updating known hosts and authorized keys all the time

      • @ouch
        link
        11 year ago

        Passwords will be brute forced if it can be done offline.

        Private SSH keys should never leave a machine. If a key gets compromised without you knowing, in worst case you will revoke the access it has once the machine’s lifespan is over. If you copy around one key, it may get compromised on any of the systems, and you will never revoke the access it has.

        And you may not want to give all systems the same access everywhere. With one key per machine, you can have more granularity for access.

        • @dinosaurdynasty
          link
          English
          11 year ago

          Passwords will be brute forced if it can be done offline.

          Set a good high entropy password, you can even tie it to your login password with ssh-agent usually

          Private SSH keys should never leave a machine.

          If this actually matters, put your SSH key on a yubikey or something

          If a key gets compromised without you knowing, in worst case you will revoke the access it has once the machine’s lifespan is over.

          People generally don’t sit on keys, this is worthless. Also knowing people I’ve worked with… no, they won’t think to revoke it unless forced to

          and you will never revoke the access it has.

          Just replace the key in authorized_keys and resync

          And you may not want to give all systems the same access everywhere

          One of the few reasons to do this, though this tends to not match “one key per machine” and more like “one key per process that needs it”

          Like yeah, it’s decent standard advice… for corporate environments with many users. For a handful of single-user systems, it essentially doesn’t matter (do you have a different boot and login key for each computer lol, the SSH keys are not the weak point)

  • S410
    link
    fedilink
    9
    edit-2
    1 year ago

    On my devices like PCs, laptops or phones, syncthing syncs all my .rc files, configs, keys, etc.

    For things like servers, routers, etc. I rely on OpenSSH’s ability to send over environmental variables to send my aliases and functions.
    On the remote I have
    [ -n "$SSH_CONNECTION" ] && eval "$(echo "$LC_RC" | { { base64 -d || openssl base64 -d; } | gzip -d; } 2>/dev/null)"
    in whatever is loaded when I connect (.bashrc, usually)
    On the local machine
    alias ssh="$([ -z "$SSH_CONNECTION" ] && echo 'LC_RC=$(gzip < ~/.rc | base64 -w 0)') ssh'

    That’s not the best way to do that by any means (it doesn’t work with dropbear, for example), but for cases like that I have other non-generic, one-off solutions.

  • @[email protected]
    link
    fedilink
    81 year ago

    Syncthing. If you want flatpak, syncthingy.

    Its simply best, does all the annoying background things like webUI, machines, versioning, verifying etc. If you disable global discovery you can use it tough LAN only

    • @[email protected]
      link
      fedilink
      31 year ago

      I love this solution, I’ve been using it for years. I had previously just been using the home directory is a git repo approach, and it never quite felt natural to me and came with quite a few annoyances. Adding stow to the mix was exactly what I needed.

    • @Anonymouse
      link
      English
      11 year ago

      This is the only answer for me. Bonus points if your .login file does a background git pull.

    • jelloeater - Ops Mgr
      link
      English
      11 year ago

      I use it with GitHub, works amazing on multiple boxes, OSX and Linux. For SSH keys, just use KeePassXC and SyncThing.

    • macallik
      link
      fedilink
      1
      edit-2
      1 year ago

      That my solution. I have a ‘Sync’ folder on every device’s Home folder, and then I use some aliases to determine whether to grab the bash_aliases file or replace it:

      • alias dba=‘diff -s ~/.bash_aliases ~/Sync/.bash_aliases’ # compare files
      • alias s2ba=‘cp ~/Sync/.bash_aliases ~/’ # Push from Sync folder to current bash aliases
      • alias ba2s=‘cp ~/.bash_aliases ~/Sync/’ # Push from current bash aliases to Sync folder

      By far, the diff alias is the most used. It allows for a quick check on what is different between files w/o having to open them up

  • @[email protected]
    link
    fedilink
    English
    6
    edit-2
    1 year ago

    I use a git repo combined with the basic install utility. Clone the repo, run the app installer, then run the install script. For symlinks I just use a zsh script.

    • mFatOP
      link
      fedilink
      English
      11 year ago

      Thanks that’s a good idea.

    • mFatOP
      link
      fedilink
      English
      21 year ago

      I like this approach. Had never heard of those solutions. Thanks!

  • @[email protected]
    link
    fedilink
    41 year ago

    I keep my dotfiles in a got repo and just do a git pull your update them. That could definitely be a cron job if you needed.

    SSH keys are a little trickier. I’d like to tell you I have a unique key for each of my desktop machines since that would be best practice, but that’s not the case. Instead I have a Syncthing shared folder. When I get around to cleaning that up, I’ll probably do just that and keep an authorize_keys and known_hosts file in git so I can pull them to needed hosts and a cron job to keep them updated.

  • @[email protected]
    link
    fedilink
    21 year ago

    My solution is not ideal:

    I created a directory, called ~/config_sync. I create sym links for config files, like ~/.bashtc to ~/config_sync/bashrc

    However, I need to record the sym links I’ve created, and repeat this process on new machines

    • @[email protected]
      link
      fedilink
      31 year ago

      Look into using GNU stow! It’s exactly what you’re doing but it creates the symlinks for you.

  • Coelacanthus
    link
    fedilink
    21 year ago

    Use a git repo and stow tool. For updating, you only need run git pull (and stow if you create config for a new software). If you modify some config, just git add && git commit && git push.
    With this way, you can also record change history of your config.