I don’t think that is true. Not much at Google really bought into the UUID hype. At least not for internal interfaces. But really there is no difference between a UUID v4 and a large random number. UUID just specifies a standard formatting.
I don’t think that is true. Not much at Google really bought into the UUID hype. At least not for internal interfaces. But really there is no difference between a UUID v4 and a large random number. UUID just specifies a standard formatting.
It is true, don’t do it.
Even at huge companies like Google, lots of stuff was keyed on your email address. This was a huge problem so Google employees were not allowed to change their email for the longest time. Eventually they opened it up by request but they made it very clear that you would run into problems. So many systems and services would break. Over time I think most external services are pretty robust now, but lots of internal systems still use emails (or the username part of it) and have issues.
IIUC Google accounts now use a random number as the key. But there are still places where the email is in use, slowly being fixed at massive cost.
If you want to use proprietary apps the best option is probably dumping the APK from your “Googled” phone then sideloading it onto the new phone. However it may be difficult to keep up with updates unless you have a dedicated phone to download them.
For software to run on a computer it needs to speak the computer’s “language”. This is typically called “machine language” but differs across different hardware. For example most modern Intel and AMD processors speak x86_64. This language has ways to express different operations such as “add these two numbers” or “put this CPU core into a low power mode”. This is the fundamental way that software works, but running in this language.
There are languages that are completely different, such as ARM which is very common on mobile devices and is the language used by Apple’s new M chips. These have basically nothing in common with x86_64.
These languages also evolve over time. For example x86_64 is a significant extension to the older x86 language. For the most part this is fine, it is like the CPU now knows more words, if you use those new words the new CPU will understand them, but older CPUs won’t.
RISC-V is a new machine language. What makes it interesting is that it is a free and open specification. This means that anyone can create a new RISC-V CPU, unlike x86_64 where you need to buy a license from Intel or ARM where you need to buy a license from the ARM corporation. Most people think that this openness has major benefits, for example now anyone can create a new processor which may be better, rather than having innovation being stifled by licensing costs (if you can even get a license) or needing to create their own machine language and require huge amounts of effort to migrate software to it.
Note: It is important not to confuse “machine language” with “programming language”. When people write software they very rarely write code in machine language directly. Usually they use a programming language which is then converted into the machine language of the CPU it will run on.
I thought Apple implemented push notifications? Or did they just say they would? Either way you can file the bug with them I think.
Or wait until they allow you to install a browser that isn’t dragging it’s feet.
I wrote my own. I aimed for a different UX than most services. For my use case I have a few devices that I often share files between. So opening the tool on both devices was a bit annoying. Instead you select the file on the first device and you get a push notification on the other. Then the transfer is done over WebRTC (locally if possible). All communication is done end-to-end encrypted and over your browser’s push service.
Hosted: https://filepush.kevincox.ca/
The problem with Yubikey is that it doesn’t have a good enough management story for broad use. I do use it for a few core sites (like GitHub) but if I lose a key I need to get a replacement and register that replacement with every site I have set up U2F 2FA on. This is ok with a few core accounts but doesn’t scale to the hundreds of sites that I have an account with. I am sure to miss a few and then either I can’t log in with the new key or get completely locked out when I lose that key and get a second replacement.
Yeah, this is important to realize. Most good 2FA implementations offer TOTP which doesn’t need a proprietary app. You can store all of your 2FA secrets in whatever app or password manager you like.
So yes, it is slightly better, but in practice that difference probably doesn’t matter. If you use U2F then you may have a meaningful security increase but IMHO U2F is not practical to use on every site due to basically being impossible to manage credentials.
So yes, it is better. But for me using random passwords and a password manager it isn’t worth the bother.
It is also worth noting that Firefox Sync is end-to-end encrypted. So the amount of data the server gets is quite minimal. (This is unlike the sync of a lot of other major browsers.) So unless you want to hide your IP and activity times from the host self-hosting isn’t critical.
I am willing to help with moderate. I have minimal existing moderation experience but have a long posting history and online presence.
I will not be able to commit enough time to be a sole moderator, but can help out as part of a team.
There are dozens of first-person shooters but people love porting Doom to every device. Winamp is memes and nostalgia, I would bet that people would port it just for fun.
Yeah, just jump in.
To get started it is best to keep Windows around, then if you need to get something done urgently you can go back to what you know then figure out how to do it in Linux later. Dual-booting is probably the best option if you are gaming as GPU passthrough can be difficult to get great performance. That is the approach I took a long time ago and then at some point I realized that I hadn’t booted into Windows for months and just deleted the partition.
I’m using Kagi. I find that it does a better job at finding “legitimate” sites rather than blogspam and content marketing. However I’m not sure I will stick with it a long time. I seems like it has mostly stalled and the team is getting distracted by making a browser, non-relevant AI (I have no problem with the few AI experiments tied to searching) and other side projects. We’ll see. I really hope that they pull themselves together and focus or it might not last. But for now they seem like one of the better options available.
Bing’s new “Deep Search” where it has some sort of LLM refinement iteration process has also been helpful sometimes. Probably the best AI search product I have seen, but definitely doesn’t replace most searches for me.
I bet someone will have it running on Linux within a week.
My problem is that I do red
, blue
, green
then can’t think of any more clearly visible colours.
Because some people want to filter it out. So it gets a label.
Algorithms are like AI but accurate, predictable and usually much faster.
Oh, flac fixes for HLS. I wonder if https://github.com/jellyfin/jellyfin/issues/8722 was fixed. I’ll have to try it out today.
Yup, that “what can I start in 10min” question really ruins a lot of productivity.