Why would they do that, they can’t even sell you minecoins on the Java edition
Why would they do that, they can’t even sell you minecoins on the Java edition
Yep. On kbin I think any user can too.
Incorrect. Open source means using a license that conforms to the open source definition. You can find that here: https://opensource.org/osd
If a license forbids LLM training, it is by definition not open source.
Only if you can reasonably argue that the output is the input (even with exact matches over a certain size being auto-rejected), and that it is enough to qualify as a copyrightable work. I’d argue line completions can never be enough to be copyrightable, and even a short function barely meets the bar unless it is considered creative in some way.
Every open source license grants permission for AI training, and GitHub copilot by default rejects completions that exactly match code from its training. You can’t pretend to be pro-open source or pro-free software but at the same time be upset that people are using licensed software within its license terms.
Pretty sad all the people getting mad about an optional opt-in feature. I think this is pretty cool. If it’s free to use with my existing unlimited plan I’ll probably use it regularly, otherwise if I have to pay more I’ll probably just keep using ChatGPT since I already pay for that.
As a person who has been managing Linux servers for about a decade now, trust me that a few hours or days of learning docker now will save you weeks if not months in the future. Docker makes managing servers and dealing with updates trivial and predictable. Setting everything up in docker compose makes it easy to recover if something fails, it’s it’s self documenting because you can quickly see exactly how your applications are configured and running.
Taint analysis is a real thing that several papers have been published about, but the implementations aren’t in a state where they could be run in real time without massively hampering performance. Also they’re mostly focused on findings bugs in native applications rather than privacy on the web.
Probably not. Electron is popular not just for its cross-platform support, but also that its skills are highly transferable from existing web dev.
If it ends up being ruled that training an LLM is fair use so long as the LLM doesn’t reproduce the works it is trained on verbatim, then licensing becomes irrelevant.
If something like that were to work, a lot of effort would need to be put into minimizing the UI friction. I could see something like: uploaders add topic tags to their videos, and an AI runs in the background to generate and apply new tags based on the content (most people would not understand how to properly tag content). An AI would also be used to create a graph of related tags, where similar or closely related tags are nodes joined by an edge. Then, on first login the user is prompted to pick some tags to start with. Over time, the client uses the adjacent tag graph to fine-tune users’ tags, on device. The idea here is that we could get a decent algorithm that can recommend new stuff based on what the user watches, but keep that data processing of user-specific content local. Then, the client would also have an option the user could enable that would contribute their client’s tag information back to the global tag graph, improving the global tag graph for everybody. This data could also be combined with other users data at the instance level to somewhat anonymize the data, assuming it is a large multi-user instance. If you were to host a single user instance, you’d probably not want to contribute to the global tag graph unless you’re ok with your tag preferences being public.
It’s a bit tricky but I think a privacy preserving algorithm is possible. Simply put, the more data available, the better an algorithm can be.
I think the easy discoverability on these platforms is part of what makes them so popular. Using TikTok or similar, a user typically wants to be shown new things, it maintains a sense of novelty that keeps users constantly engaged. Having to do this manually would be a huge negative.
The algorithms are what makes these services. Most interactions aren’t searching and selecting something specific or intentional, they’re just opening a fire hose and expecting the algorithm to pick content they find entertaining for them. It requires the algorithm to have a lot of information, both about the specific user, and about similar users.
Most people shouldn’t self host. It’s a hobby for people who want to do it, and there are benefits, but spending 3 hours on a weekend fixing stuff is not how most people wish to spend their time. Furthermore, it’s not a good use of most people’s time. We split labor up into specialties, forcing people to do work outside their specialty causes pointless inefficiency. I agree with what other commenters have said in that a better approach would be to have more small businesses hosting federated together, and anyone not inclined to self host should just purchase service through one of those many small providers instead.
sudoedit copies a file to a temp directory, invokes $EDITOR with that temp file, and after the editor process exits, it copies the file back to overwrite the original. This way you get your user preferred and configured editor, but it doesn’t have any elevated privileges.
As a silver lining, at least it’s terrible at it
As far as I know there is only one SSD model that meets my criteria (Samsung 870 QVO 8TB), and at $520 right now so I’ve decided it’s best to wait. I’d like it to be quieter but not so badly as to spend $1k on it (need two).
Why this over a much more popular modern language like Rust?