this post was submitted on 25 Oct 2024
24 points (96.2% liked)
Programming
17433 readers
154 users here now
Welcome to the main community in programming.dev! Feel free to post anything relating to programming here!
Cross posting is strongly encouraged in the instance. If you feel your post or another person's post makes sense in another community cross post into it.
Hope you enjoy the instance!
Rules
Rules
- Follow the programming.dev instance rules
- Keep content related to programming in some way
- If you're posting long videos try to add in some form of tldr for those who don't want to watch videos
Wormhole
Follow the wormhole through a path of communities !webdev@programming.dev
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Write code on a machine you can remote into from each computer? Less commits, possibly less reverts, less chance of forgetting to git pull after switching machines…idk.
Don't even need to remote in to anything, just store your working code on a network share
True!
I mean... That's kinda what git does, in a way... Right?
Don't think git as a sync storage, more like to merge works.
If you need to share files between computers use a shared storage.
Always use the right tool for the job. Mount a shared storage or use synctools rsync, etc
Yes, and use something like GNU Screen to work seamless on the other machine again.
I have considered this approach, but there are several things I had issues with.
To address the issues you brought up:
I once used a virtual desktop in the cloud, and I could access that from anywhere. It was just a regular OS that had all my tools, and it was where my work was done changes. Ultimately, that remote desktop went away when I changed jobs. But, it would be something I would think about again for me.
There is a danger of things going poof, or not being accessible. It cannot be helped at all. But a push to a backup repo during each commit, would allow an emergency restore. Doing a snapshot every few days of the machine, for example if its on AWS or other, helps lessen the loss when and if it goes poof.
To solve the issue of the internet going out, have one of your local computers do a regular pull as a cron job of the backup repo
Your git solution still has all of these issues, as you need the git server to be alive, for number 3 use something like rsync so you keep a local copy that is backed up if you are concerned about the file share being offline.
I don't need the client computers to be alive, only the central server (which could be github.com for example, so not even a server I manage).