What if you aren’t flexible? I never couch, I just say in use Linux for my workflow, can you accommodate that?
What if you aren’t flexible? I never couch, I just say in use Linux for my workflow, can you accommodate that?
I ask before I take the interview. Location, salary range, linux laptop are prerequisites to me working for anyone. If they punt on the laptop question it means no and they are hoping you’ll want the job even without. I can promise you I won’t, and if you view that as a red flag I can promise I don’t want to work there so I don’t care.
If its a hard requirement for you just say that and say that’s for workflow and you don’t want to waste anyone’s time
I was already full Linux, but gnome is the reason I stopped messing with window managers and maybe large 4k monitors.
It finally hit enough of ‘just works’ and customizability to use my standard workflow.
The only thing I want that I don’t have right now is horizontal monitor splits for vertical monitors.
No, this is a tool that can be used in a well designed architecture. Would I do this with a single database server, probably not. Would I ever run a single database server? Also probably not.
Also, by this point, you’ve probably already kernel panicked or something. There’s not much left that can be saved and you probably needed that backup five minutes before the host came up.
Check if your motherboard has a watchdog function. If the OS can’t ping the watchdog every 5 min or whatever you set it to, the board resets.
Not in an enterprise setting, so patato potato
Third best enterprise OS.
I mean, its different for everyone.
For me I’ve done plenty of shifts where I got paged, slept or didn’t sleep and then worked a full day.
But at this point, if I go back to sleep, I won’t set an alarm, because I see no value in going to work like a zombie. If I end up at work but can’t focus because I was upnall night with he pager, I’ll just hit my couple meetings and call it. No point in sitting around pretending to work.
So they can read your code and use it for copilot
One option here is to host it internally, and then VPN or ssh tunnel to your network for access.
Keeping openssh or a VPN up to date and secure is a much simpler thing than a web framework.
Separate your network access and your services. You get in trouble trying to use your service to gate access to your network.
As a note, when you can’t find a package, go find the source, you can usually build and install in a couple commands. Its nice to use the package management of the distro, but most of the time, you could just install the deps and compile and be done with it.
Yes, WebDAV will max your local connection. Its generally not the encryption that makes ssh slow but the fact that it is designed to give real time terminal feedback. In order for you to see each letter typed in an ssh session, the buffers are really small and it intentionally sends a tone of small packets. Great for single characters bad for large file transfer.
Its OK here and then when you need to push a config file or something but moving large files is not really what its designed for and consequently, it sucks.
Well, for starters, tftp is the wrong thing for local file transfers if you want it to be fast. The only reason its still around is because its simple and offer the only file transfer protocol that is built into the firmware of the network card.
You read that right, its a simple file transfer protocol built into every network card made in the last couple decades.
Your best bet for file transfer is probably something like a WebDAV server. Which next cloud can handle for you. You can just enable normal WebDAV on something like httpd but then you gotta handle authentication yourself. (Or allow local and connect with VPN)
Second, I run a fleet of kobos for the family, they alsonwork pretty well with the libraries around the area which the kids love.
Lol, this is peak selfhosted. The obvious solution is to get a router/DHCP server that is normal enough to push out two DNS servers.
The selfhosted way is to set of keep alived or a load balancer, because why the fuck not.
*pipe breaks ?
Our water (and most in the US with a general water district, think more or less any town or city) test daily. Though in bigger places they might not test every one every day but those places will test every well and sistern every week.
I’m not entirely sure there are a ton of people/companies that are considering rhel licences vs rocky. All the companies I’ve worked for are considering debian vs rocky at this point. Not huge but 1000-5000 system type companies. I’d guess that’s a huge bulk of the market that’s using rocky, and also up steaming patches and big reports.
The answer is that it depends what you’re doing. You can have an extremely efficient dev env CLI. You can kinda brows the web cli. You probably don’t want to edit videos and pictures in the CLI.
Because a lot of foss has replaceable building blocks, you’re not going to find a ‘this is how you do things course’. You’re much more likely to find, ‘this is how you use a certain text editor in the CLI’.
So, first, I guess, figure out or articulate what you want to do in the CLI. From reading and sending mail to writing code and building a dev environment to just basic scripting to maintain your install.
After that step, you’ll want to try a couple of the building blocks that do that.
Once you find one that kinda clicks, then you can go become proficient and start to put together the pieces of your workflow.
I run arch/Ubuntu and gnome. I spend about 50-70% of my time every day in a terminal. I spend the rest in a browser. Sometimes I use files, libre office calc, gimp or the calculator app, but even combined their usage is probably a rounding error.
I run gnome term (used to do a lot of urxvt but gnome term seems to work fine these days)
From there, I start tmux. Inside of tmux, I run a few windows. One has email, a couple shells, chat and system monitoring.
The next window has my core dev env, I run nvim with a server so I can upen tabs in nvim from different terminals, in nvim I run the lsp servers for linting and code completion. I use ranger as a file browser/previewer and that’s hooked to nvim so when I select a file there it will open as a tab in nvim, additionally I can run that file in the debug pane in the bottom of the window.
Then I just have windows that I drop into to do additional tasks, ops work on multiple servers at once, a second dev env to make a quick change in a different package, or a new window to scrape up a one line script to parse a log file or data dump for processing else where.
All of this takes time, for me about 15 years probably to say ‘ok, I want(need) to do this thing in the terminal, now what’s the best way to get that done . . .’
And then, you just kinda build it.
Depends if you want to assign IP addresses or not. If you don’t, you just want your own section of the same lan, I.e.all your devices connected to your router but let dhcp pass through then you can just set itnup as an extender