[-] patatahooligan@lemmy.world 20 points 2 days ago

Not even close. The paper is questioning LLMs ability to reason. The article talks about fundamental flaws of LLMs and how we might need different approaches to achieve reasoning. The benchmark is only used to prove the point. It is definitely not the headline.

[-] patatahooligan@lemmy.world 65 points 2 weeks ago

No, there isn't really any such alternate timeline. Good honest causes are not profitable enough to survive against the startup scams. Even if the non-profit side won internally, OpenAI would just be left behind, funding would go to its competitors, and OpenAI would shut down. Unless you mean a radically different alternate timeline where our economic system is fundamentally different.

[-] patatahooligan@lemmy.world 25 points 2 months ago

And now in the r/linux thread about these news people are defending Vaxry, misrepresenting what the ban was about, and hating FDO.

Indicatively, this blatantly wrong comment chain is upvoted:

Is this the project where some red Hat dev started dropping legal threats from their corporate account over offline activities by third parties in unrelated communities years past?

Sort of. You got some details wrong but essentially, yes.

But this is downvoted and has replies telling them they're wrong:

Congratulations to the hyprland project, but I definitely will not be using or contributing to the project as long as it's an exclusionary and intolerant space.

[-] patatahooligan@lemmy.world 27 points 5 months ago

This has nothing to do with centralization. AI companies are already scraping the web for everything useful. If you took the content from SO and split it into 1000 federated sites, it would still end up in a AI model. Decentralization would only help if we ever manage to hold the AI companies accountable for the en masse copyright violations they base their industry on.

[-] patatahooligan@lemmy.world 24 points 7 months ago

Humans are not generally allowed to do what AI is doing! You talk about copying someone else's "style" because you know that "style" is not protected by copyright, but that is a false equivalence. An AI is not copying "style", but rather every discernible pattern of its input. It is just as likely to copy Walt Disney's drawing style as it is to copy the design of Mickey Mouse. We've seen countless examples of AI's copying characters, verbatim passages of texts and snippets of code. Imagine if a person copied Mickey Mouse's character design and they got sued for copyright infringement. Then they go to court and their defense was that they downloaded copies of the original works without permission and studied them for the sole purpose of imitating them. They would be admitting that every perceived similarity is intentional. Do you think they would not be found guilty of copyright infringement? And AI is this example taken to the extreme. It's not just creating something similar, it is by design trying to maximize the similarity of its output to its training data. It is being the least creative that is mathematically possible. The AI's only trick is that it threw so many stuff into its mixer of training data that you can't generally trace the output to a specific input. But the math is clear. And while its obvious that no sane person will use a copy of Mickey Mouse just because an AI produced it, the same cannot be said for characters of lesser known works, passages from obscure books, and code snippets from small free software projects.

In addition to the above, we allow humans to engage in potentially harmful behavior for various reasons that do not apply to AIs.

  • "Innocent until proven guilty" is fundamental to our justice systems. The same does not apply to inanimate objects. Eg a firearm is restricted because of the danger it poses even if it has not been used to shoot someone. A person is only liable for the damage they have caused, never their potential to cause it.
  • We care about peoples' well-being. We would not ban people from enjoying art just because they might copy it because that would be sacrificing too much. However, no harm is done to an AI when it is prevented from being trained, because an AI is not a person with feelings.
  • Human behavior is complex and hard to control. A person might unintentionally copy protected elements of works when being influenced by them, but that's hard to tell in most cases. An AI has the sole purpose of copying patterns with no other input.

For all of the above reasons, we choose to err on the side of caution when restricting human behavior, but we have no reason to do the same for AIs, or anything inanimate.

In summary, we do not allow humans to do what AIs are doing now and even if we did, that would not be a good argument against AI regulation.

[-] patatahooligan@lemmy.world 24 points 8 months ago

AI companies will probably get a free pass to ignore robots.txt even if it were enforced by law. That's what they're trying to do with copyright and it looks likely that they'll get away with it.

[-] patatahooligan@lemmy.world 37 points 9 months ago

This is great. Proton is getting a lot of testing just based on Steam's userbase and it is backed by Valve. We also have a lot of data on proton's performance and potential game-specific fixes in the form of protondb. Making sure that non-Steam launchers can use all that work and information is crucial to guaranteeing the long-term health of linux gaming. Otherwise it is easy to imagine a future where proton is doing great but the other launchers are keep running into problems and are eventually abandoned.

One thing that I am curious is how this handles the AppId. If this AppId is used to figure out which game-specific fixes are needed, then it will have to be known. Do we have a tool/database that figures out the AppId from the game you are launching outside of Steam?

[-] patatahooligan@lemmy.world 50 points 9 months ago

So help me out here, what am I missing?

You're forgetting that not all outcomes are equal. You're just comparing the probability of winning vs the probability of losing. But when you lose you lose much bigger. If you calculate the expected outcome you will find that it is negative by design. Intuitively, that means that if you do this strategy, the one time you will lose will cost you more than the money you made all the other times where you won.

I'll give you a short example so that we can calculate the probabilities relatively easily. We make the following assumptions:

  • You have $13, which means you can only make 3 bets: $1, $3, $9
  • The roulette has a single 0. This is the best case scenario. So there are 37 numbers and only 18 of them are red This gives red a 18/37 to win. The zero is why the math always works out in the casino's favor
  • You will play until you win once or until you lose all your money.

So how do we calculate the expected outcome? These outcomes are mutually exclusive, so if we can define the (expected gain * probability) of each one, we can sum them together. So let's see what the outcomes are:

  • You win on the first bet. Gain: $1. Probability: 18/37.
  • You win on the second bet. Gain: $2. Probability: 19/37 * 18/37 (lose once, then win once).
  • You win on the third bet. Gain: $4. Probability: (19/37) ^ 2 * 18/37 (lose twice, then win once).
  • You lose all three bets. Gain: -$13. Probability: (19/37) ^ 3 (lose three times).

So the expected outcome for you is:

$1 * (18/37) + 2 * (19/37 * 18/37) + ... = -$0.1328...

So you lose a bit more than $0.13 on average. Notice how the probabilities of winning $1 or $2 are much higher than the probability of losing $13, but the amount you lose is much bigger.

Others have mentioned betting limits as a reason you can't do this. That's wrong. There is no winning strategy. The casino always wins given enough bets. Betting limits just keep the short-term losses under control, making the business more predictable.

[-] patatahooligan@lemmy.world 70 points 9 months ago

Exactly this. I can't believe how many comments I've read accusing the AI critics of holding back progress with regressive copyright ideas. No, the regressive ideas are already there, codified as law, holding the rest of us back. Holding AI companies accountable for their copyright violations will force them to either push to reform the copyright system completely, or to change their practices for the better (free software, free datasets, non-commercial uses, real non-profit orgs for the advancement of the technology). Either way we have a lot to gain by forcing them to improve the situation. Giving AI companies a free pass on the copyright system will waste what is probably the best opportunity we have ever had to improve the copyright system.

[-] patatahooligan@lemmy.world 36 points 11 months ago

This is very common among big tech companies and we should start treating it as what it is, a scam.

21
submitted 1 year ago* (last edited 1 year ago) by patatahooligan@lemmy.world to c/linux@lemmy.ml

I have an SSD from a PC I no longer use. I need to keep a copy of all its data for backup purposes. The problem is that dd reports "Input/output error"s when copying from the drive. There seem to be 20-30 of them in the entire 240GB drive so it is likely that most or all of my data is still intact.

What I'm concerned about is whether these input/output errors can cause issues in the image outside of the particular bad blocks. How does dd handle these errors? Will they be eg zeroed in the output or will the simply be missing? If they are simply missing will the filesystem be corrupted because the location of data has been shifted? If so, what tool should I be using to save what can be saved?

EDIT: Thanks for the help guys. I went with ddrescue and it reports to have saved 99.99% of the data. I guess there could still be significant loss if the 0.01% happens to be on filesystem structures, but in this case maybe I can use an undeleter or similar utility to see if I can get back the files. In any case, I can work at my leisure now that I have a copy of the data on non-failing storage.

[-] patatahooligan@lemmy.world 40 points 1 year ago

Personally I don't care so much about the things that Linux does better but rather the abusive things it doesn't do. No ads, surveillance, forced updates etc. And it's not that linux happens to not do that stuff. It's that the decentralized nature of free software acts as a preventative measure against those malicious practices. On the other side, your best interests always conflict with those of a multi-billion company, practically guaranteeing that the software doesn't behave as you. So windows are as unlikely to become better in this regard as linux is to become worse.

Also the ability to build things from the ground up. If you want to customize windows you're always trying to replace or override or remove stuff. Good luck figuring out if you have left something in the background adding overhead at best and conflicting with what you actually want to use at worst. This isn't just some hypothetical. For example I've had windows make an HDD-era PC completely unusable because a background telemetry process would 100% the C: drive. It was a nightmarish experience to debug and fix this because even opening the task manager wouldn't work most of the time.

Having gotten the important stuff out of the way, I will add that even for stuff that you technically can do on both platforms, it is worth considering if they are equally likely to foster thriving communities. Sure I can replace the windows shell, but am I really given options of the same quality and longevity as the most popular linux shells? When a proprietary windows component takes an ugly turn is it as likely that someone will develop an alternative if it means they have to build it from the ground up, compared to the linux world where you would start by forking an existing project, eg how people who didn't like gnome 3 forked gnome 2? The situation is nuanced and answers like "there exists a way to do X on Y" or "it is technically possible for someone to solve this" don't fully cover it.

[-] patatahooligan@lemmy.world 35 points 1 year ago

It is copyright infringement. Nvidia (and everyone writing kernel modules) has to choose between:

  • using the GPL-covered parts of the kernel interface and sharing their own source code under the GPL (a free software license)
  • not using the GPL-covered parts of the kernel interface

Remember that the kernel is maintained by volunteers and by engineers funded by/working for many companies, including Nvidia's direct competitors, and Nvidia is worth billions of dollars. Nvidia is incredibly obnoxious to infringe on the kernel's copyright. To me it is 100% the appropriate response to show them zero tolerance for their copyright infringement.

view more: next ›

patatahooligan

joined 1 year ago