• 0 Posts
  • 21 Comments
Joined 1Y ago
cake
Cake day: Jun 17, 2023

help-circle
rss

This. I am so tired of hearing “the wheels of justice turn slowly”. If justice isn’t able to address a problem for so long that the perpetrators are allowed to continue perpetuating the same behavior an entire election cycle later, the justice system has failed, straight up. This is unacceptable.


This news is from over a month ago, and conditions have materially and dramatically changed since it’s publication. Regardless of the intent, posting this without noting a critical detail (it’s age) is at best incredibly misleading, and at worst intentionally subversive.


Hey so most of us are against this nonsense, republicans haven’t won the popular vote in well over a decade. We don’t really have a say in the matter, and haven’t in a pretty long time.


Indexing and lookups on datasets as big as companies like Google and Amazon are running also take trillions of operations to complete, especially when you take into account the constant reindexing that needs to be done. In some cases, encoding data into a neural network is actually cheaper than storing the data itself. You can see this in practice with gaussian splatting point cloud capture, where they are training networks to guide points in the cloud at runtime, rather than storing the position of trillions of points over time.


I firmly believe it will slow down significantly. My prediction for the future is that there will be a much bigger focus on a few “base” models that will be tweaked slightly for different roles, rather than “from the ground up” retraining like we see now. The industry is already starting to move in that direction.


While I agree in principle, one thing I’d like to clarify is that TRAINING is super energy intensive, once the network is trained, it’s more or less static. Actually using the network isn’t dramatically more energy than any other indexed database lookup.



I program professionally, and I copy paste all the time. The difference is when I copy paste, its 10-20 lines of code, not a line or two— and I’m not fishing for a solution to the problem. I already have the optimal solution in my head, and I am just searching for the solution I already know. It’s just faster than typing it by hand 🤷🏻


I really don’t give a shit about the trials any more. At this point, he’s gotten away with it. I’m so tired of hearing this “the wheels of justice work slowly” non-sense. We all watched him do what he was accused of, live, because he himself did it on video, and posted it. The things we all watched him do, live, are expressly prohibited of any presidential candidate, and yet, the justice system has failed to adequately sentence him, and prevent him from running again. So here we are, again, 4 years later, and he’s been allowed to run again. The justice system has failed our country on this case, they aren’t slow, they aren’t “getting there eventually”, they’ve fallen flat on their face. There is a very real chance this man will be dead of natural causes before he will face any real penalty for his actions, beyond a fine or two here or there. He’s still living in de-facto luxury, he still has millions and millions of followers, what have we actually done other than keep his face on TV for the last 4 years? Even that seems like something he wanted any way. I am just so tired.

The defeatism energy is strong, but really I am just so tired of having hope and seeing it crushed over and over again. I just have to give up hoping this situation will resolve justly at this point.


How is it scammy or gimmicky to put ads on a product you don’t pay for? I guess I’m just missing the disconnect here. This service costs Nvidia money to run.


Base pay is a meaningless figure in tech. Total comp is all that matters. The true high rollers intentionally keep their base pay low to dodge taxes. This doesn’t mean they aren’t still getting and using this money, it’s just an intentional straw man to allow them to make this exact argument.


That would be a dream come true for Youtube. Think about how that would affect them financially, they would be serving significantly less load, but they would still rake in all those Youtube premium subscriptions. It would effectively be campaigning to remove the users who aren’t monetized from the platform.


If you don’t get a physical piece of media that can be viewed offline indefinitely, you don’t own anything, you’re just renting. Services revoking even bought and paid for content is not unheard of, digital purchasing gives every streaming company the ability to do that.


Firstly, let me be clear. This is my opinion as someone who isn’t, and has never been involved with their instance in any capacity. My account is with pawb.social.

Posting legal letters may open them up to additional legal liability, and it is completely reasonable to let the community know what happened without sharing the actual letter.

Secondly, you’re putting words in my mouth. I am not saying they are censoring in good faith. I am saying the complete opposite, that they are reversing their stance in good faith. In other words, they are uncensoring content in good faith.


This doesn’t sound like censorship, it sounds like they were getting legal threats directly levied at their volunteer team. I can understand the desire to protect yourself against getting sued for your (admittedly large) side project. It sounds like they are working on it in good faith though.


I’m out and about today, so apologies if my responses don’t contain the level of detail I’d like; As for the law being collective morality, all sorts of religious prohibitions and moral scares HAVE ended up in the law. The idea is that the “collective” is large enough to dispel any niche restrictive beliefs. Whether or not you agree with that strategy aside, that is how I believe the current system works in an ideal sense (even if it works differently in practice), that’s what it is designed to protect from my perspective.

As for anti-AI artists, let me pose a situation for you to illustrate my perspective. As a prerequisite for this situation, a large part of a lawsuit, and the ability to advocate for a law is based on standing, the idea that you personally, or a group you represent has been directly, tangibly harmed by the thing you are trying to restrict. Here is the situation:

I am a furry, and a LARGE part of the fandom is based on art and artists. A core furry experience is getting art commissioned of your character from other artists. It’s commonplace for all these artists to have a very specific, identifiable signature style, so much so that it is trivial for me and other furs to be able to identify artists by their work alone at just a glance. Many of these artists have shifted to making their living full time off of creating art. With the advent of some new generational models, it is now possible to train a model exclusively off of one singular artists style, and generate art indistinguishable from the real thing without ever contacting them. This puts their livelihood directly at risk, and also muddies the waters in terms of subject matter, and what they support. Without laws regulating training, this could take away their livelihood, or even give a (very convincing, and hard to disprove) impression that they support things they don’t, like making art involving political parties, or illegal activities, which I have seen happen already. This almost approaches defamation in my opinion.

One argument you could make is that this is similar to the invention of photography, which may have directly threatened the work of painters. And while there are some comparisons you could draw from that situation, photography didn’t fundamentally replace their work verbatim, it merely provided an alternative that filled a similar role. This situation is distinct because in many cases, it’s not possible, or at least immediately apparent which pieces are authentic, or not. That is a VERY large problem the law needs to solve as soon as possible.

Further, I believe the same, or similar problems exist in LLMs, like they do in the situation involving generative image models above. Sure with enough training, those issues are lessened in impact, but where is the line of what is ok and what isn’t? Ultimately the models themselves don’t contain any copyrighted content, but they (by design) combine related ideas and patterns found in the training data, in a way that will always approximate it, depending on the depth of training data. While “overfitting” might be considered a negative in the industry, it’s still a possibility, and until there is some sort of regulations establishing the fitness of commercially available LLMs, I can envision situations in which management would cut training short once it’s “good enough”, leaving overfitting issues in place.

Lastly, with respect, I’d like to push back on both the notion that I’d like to ban AI or LLMs, as well as the notion that I’m not educated enough on the subject to adequately debate regulations on it. Both are untrue. I’m very much in favor of developing the technology, and exploring all it’s applications. It’s revolutionary, and worthy of the research attention it’s getting. I work on a variety of models across the AI and LLM space professionally, and I’ve seen how versatile it is; That said, I have also seen how over publicized it is. We’re clearly (from my perspective) in a bubble that will eventually pop. We’re claiming products use AI to do this and that across nearly every industry, and while LLMs in particular are amazing, and can be used in a ton of applications, it’s certainly not all of them— and I’m particularly cautious of putting new models in charge of dangerous or risky processes where they shouldn’t be before we develop adequate metrics, regulation, and guardrails. To summarize my position, I’m very excited to work towards developing them further, but I want to publicly express the notion that it’s not a silver bullet, and we need to develop legal frameworks for protecting people now, rather than later.


The law is (in an ideal world), the reflection of our collective morality. It is supposed to dictate what is “right” and “wrong”. That said— I see too many folks believing that it works the other way too, that what is illegal must be wrong, and what is legal must be ok. This is (decisively) not the case.

In AI terms, I do believe some of the things that LLMs and the companies behind them are doing now may turn out to be illegal under certain interpretations of the law. But further, I think a lot of the things companies are doing to train these models are seen as “immoral” (me included), and that the law should be changed to reflect that.

Sure that may mean that “stuff these companies are doing now is legal”, but that doesn’t mean we don’t have the right to be upset about it. Tons of stuff large corporations have done was fully legal until public outcry forced the government to legislate against it. The first step in many laws being passed is the public demonstrating a vested interest in it. I believe the same is happening here.


I’m aware the model doesn’t literally contain the training data, but for many models and applications, the training data is by nature small enough, and the application is restrictive enough that it is trivial to get even snippets of almost verbatim training data back out.

One of the primary models I work on involves code generation, and in those applications we’ve actually observed verbatim code being output by the model from the training data, even if there’s a fair amount of training data it’s been trained on. This has spurred concerns about license violation on open source code that was trained on.

There’s also the concept of less verbatim, but more “copied” style. Sure making a movie in the style of Wes Anderson is legitimate artistic expression, but what about a graphic designer making a logo in the “style of McDonalds”? The law is intentionally pretty murky in this department, with even some colors being trademarked for certain categories in the states. There’s not a clear line here, and LLMs are well positioned to challenge what we have on the books already. IMO this is not an AI problem, it’s a legal one that AI just happens to exacerbate.


That’s not what’s happening though, they are using that data to train their AI models, which pretty irreparably embeds identifiable aspects of it into their model. The only way to remove that data from the model would be an incredibly costly retrain. It’s not literally embedded verbatim anywhere, but it’s almost as if you took an image of a book. The data is definitely different, but if you read it (i.e. make the right prompts, or enough of them), there’s the potential to get parts of the original data back.


The key to safe AI use is to treat the AI the same as the user. Let them automate tasks on behalf of the user (after confirmation) in their scope. That way no matter how much the model is manipulated, it can only ever perform the same tasks as the user.


Facebook isn’t even recognizable as a company any more, they’ve become entirely overcome with greed, not even bothering to pretend to care about their users. All they’ve done for the past 15 years is copy other platforms, steal as much data as possible to sell to the highest bidder, and retain users by any means necessary, even if that means collapsing governments and allowing people to die. We need to do everything we can to keep them the hell away from us, because they’ve shown they will stop at nothing to make money.