Gotta let the shareholders know value is only paused. More value will be created once we finish abolishing the remaining rights of EU Citizen’s privacy.
Shareholders are always worried about their value. Gotta pause growth. Can’t stop growth. Growth is infinite. The universe is infinite, and so is the capitalism machine.
I like the new UI more in terms of actual visuals, but it unfortunately has come with more ads and more general bullshit. I just use aurora store if I need something from the play store now!
It was reading this comment that made me realize that I hardly ever use the play store over f-droid anymore, because off the top of my head I couldn’t tell you what the UI looks like
Don’t forget millions upon millions of low-information, low-empathy voters and non-voters who let it happen, many who still lazily believe that corporations are our friends, that “what’s good for walmart and Exxon is good for 'Murica”, that they are still the client and not the product, and the far-reaching implications of this.
And even sadder is that users choose to still use those companies.
At the end of the day, we all have a choice. This happens to those that allow it. We try to stay safe from burglars, traffic accidents, illness, etc., but choose to still use these companies. That’s a choice, and there are ways to improve our lives, like removing them from our lives.
Well honestly, that’s too soft. They did not decide. They were pushed by regulators to follow the law. So bowing to pressure is more appropriate in my view.
Lol, if you mean facebooks AI will skew more towards the views of American Facebook users, I’d say that’s a win for Europe. It will make the AI less valuable, creating a gap in the market for a better AI that can reflect European values or american or both.
AI does not need infinite data. They can easily licence that amount of content. They are just trying to do it cheaply with user content.
I gully expect use for AI training to become a standardized part of locencing for media and content going forwards. For a band or singer, or author, it may be they ibky get a small amoint for using their content but it won’t be stolen. There is minimal value in any one part of the content. There is value in the aggregate of lots of data.
Digitized books out of copyright have more archaic language but I expect we will see lots of media out of copyright being used also. Media organization that make movies, TV shows and publish newspapers and magazines also have a trove of content.
It’s a much much bigger issue than this. Would you rather live in a world where other countries have good AI and you do not? Would you like it if only China has powerful AI? I get the copyright issue, but some things are more important than other things. This is an arms race, and everyone slowing down isn’t exactly an option.
You could have a much more complex understanding of what they are. It isn’t nearly as simple as you are imagining. If you genuinely are curious about what you’re overlooking, then here is a link.
It seems like you severely misunderstand what “AI” as we have it nowadays is (it’s not actual AI) and what it is capable (not very much) and most importantly not capable of (most things it is advertised to do). Even if investor magazines and tech CEOs try to make it seem like that, we’re not one step away from creating HAL9000. LLMs are extremely over hyped and in the most areas they have been deployed a straight up dysfunctional scam. The only arms race that is happening right now is about who can waste the most money and violate the most privacy laws with this nonsense while all the necessary data centers and their insane power and water demands accelerate the destruction of our environment even more.
Sure, thanks for your interest. It’s an incomplete picture, but we can think of LLMs as an abstraction of all the meaningful connections within a dataset to a higher dimensional space - one that can be explored. That alone is an insane accomplishment that is changing some of the pillars of data analysis and knowledge work. But that’s just the contribution of the “Attention is All You Need” paper. Many implementations of modern generative AI combine LLM inference in agentic networks, with GANs, and with rules-based processing. Extracting connections is just one part of one part of a modern AI implementation.
The emergent properties of GPT4 are enough to point toward this exponential curve continuing. Theory of mind (and therefore deception) as well as relational spatial awareness (usually illustrated with stacking problems) developed solely from increasing the parameter count describing the neural network. These were unexpected capabilities. As a result, there is an almost literal arms race on the hardware side to see what other emergent properties exist at higher model sizes. With some poetic license, we’re rending function from form so quickly and effectively that it’s seen by some as freeing and others as a sacrilege.
Some of the most interesting work on why these capabilities emerge and how we might gain some insight (and control) from exploring the mechanisms is being done by Anthropic and by users at Hugging Face. They discovered that when specific neurons in Claude’s net are stimulated, everything it responds with will in some way become about the Golden Gate Bridge, for instance. This sort of probing is perhaps a better route to progress than blindly chasing more size (despite its recent success). But only time will tell. Certainly, Google and MS have had a lot of unforced errors fumbling over themselves to stay in what they think is the race.
I’m happy to take the time to alter your perspective, if you are open to new information.
You took some time, but spent it explaining at a fairly technical level, rather than a lamens term approach. I doubt you managed to change many people’s perspective, but you maybe reinforced some.
This is another good use case for gAI. Copy/paste the comment into a GPT and tell it to re-write the content at the desired reading or technical level. Then it’s available for follow-up clarification questions.
The term "AI" has been in use since 1956 to describe a wide variety of computer algorithms and capabilities. Neural nets and large language models fall very firmly under the term's umbrella.
What you're talking about is a specific kind of AI, artificial general intelligence (AGI). Very few people believe that an LLM on its own can become AGI and even fewer believes that current LLMs are AGI, so unfortunately you're jousting with a strawman here.
If you are genuinely open to understanding the path we are on, the new situational awareness paper would be very eye-opening. It is 160 pages, so it’s probably a bit too much to get through, but there are really good videos that explain it. Matthew Berman has a great video about it. I’m not interested in swaying you and not going to debate, I’m 100s of hours deep into this and have been absolutely obsessed with it. Nobody doubted its impact as much as me. Education on the matter will undeniably change your mind tremendously. The information is there if you want a peak at the future.
I think it’s “legally grey” in the sense that governments have largely made no policies one way or the other on the data harvesting. It’s not banned, but it’s not openly encouraged either, and there’s no real legal precedent to point to for this specific matter besides the general data harvesting big tech does.
The area with the largest similarity I feel is music sampling, and as far as I know, the music industry was very quick to ensure that data harvesting for AI had to follow the same copyright laws as sampling.
There are many laws that go entirely against the constitution of a country. You can start by looking at DMCA laws that violate a bunch of rights in MANY countries. Legally gray is falling short, those are illegal, and still get enforced because $$$
It's similar to my own reaction to the people getting angry about Reddit data being used to train AIs. As someone who's been commenting rather prolifically on Reddit for 13 years I'm actually quite pleased by the thought that my views and interests are being incorporated into the foundations of modern AI. The only downside is that all those people I argued with over that period are also getting in there. :)
Good answer, no way AI will possibly ever catch up to such brilliant responses as this. Certainly, there is no reason to want to have our views represented in the next generation of technology.
I hope this isn’t the prelude to a decline. I just ordered my third Pi over the weekend. It should arrive today. I’d hate to see the platform squandered by “make number go up” types.
RISC-V is an open instruction set, which should be what the Pi foundation (if their open source mission is to be taken at face value) would be switching to if they weren’t just a way for Broadcom to push their chips on the maker community under the guise of open source.
RISC-V, an open-source instruction set architecture (ISA), has been making waves in the world of computer architecture. “RISC-V” stands for Reduced Instruction Set Computing (RISC) and the “V” represents the fifth version of the RISC architecture. **Unlike proprietary architectures such as ARM and x86, RISC-V is an open standard, allowing anyone to implement it without the need for licensing fees. This openness has led to a surge in interest and adoption across various industries, making RISC-V a key player in the evolving landscape of computing.**At its core, an instruction set architecture defines the interface between software and hardware, dictating how a processor executes instructions. RISC-V follows the principles of RISC, emphasizing simplicity and efficiency in instruction execution. This simplicity facilitates easier chip design, reduces complexity, and allows for more straightforward optimization of hardware and software interactions. This stands in contrast to Complex Instruction Set Computing (CISC) architectures, which have more elaborate and versatile instructions, often resulting in more complex hardware designs. The open nature of RISC-V is one of its most significant strengths. The ISA is maintained by the RISC-V Foundation, a non-profit organization that oversees its development and evolution. The RISC-V Foundation owns, maintains, and publishes the RISC-V Instruction Set Architecture (ISA), an open standard for processor design. The RISC-V Foundation was founded in 2015 and comprises more than 200 members from various sectors of the industry and academia.
Much like the fediverse, we’re very early on that technology. We’re waiting for the network effect to take hold in both areas. Once it does, things will improve significantly, IMO.
INB4 trust fund babies and gormless capitalists go and ream every last fucking cent from the brand destroying it in the process before moving on to the next thing.
How much stock ownership remains with the nonprofit Raspberry Pi Foundation? And will that be enough to hold off shareholder complaints that they aren’t being evil enough?
No, shareholder interest, which - in the absence of the clear desire of the majority shareholder(s) - is assumed to be profit. So I think the question above is quite important actually
This is a common misconception based on an argument put forward my Milton Friedman. It’s based on legal cases where CEOs were taken to court for knowingly defrauding shareholders for their own personal gain (say, selling all of a companies assets of the company to a different company the ceo owns privately for a single dollar).
Friedman argued that these cases set precedent that meant all CEO were legally obligated to maximize shareholder value and could be held legally accountable for not doing so. Friedman was wrong about this, like many other things he said, as he was not a lawyer, nor a particularly good economist. No CEO has even been successfully sued for “failing to maximize shareholder value” despite some people taking Friedman’s work to heart and trying to do so.
It comes from the case against Henry Ford after he saw his company was making gobs of cash and decided to give some of that to his employees. Shareholders successfully sued him to stop this on the grounds that he has a fiduciary duty to shareholders.
He was sued for miss use of company profits, not for failing to maximize profits.
He took profits and was reinvesting in new plants and cutting car prices, while also ending dividend payments to do so. That was the crux of the case, ending dividend payments despite having money to continue paying them. This case is routinely held up as an example of shareholder primacy but has been dismissed as an example of such by most modern thinkers In the field, in large part because the court also ruled that he had final say on how to proceed with company operation. Increasing worker pay was not the issue, ending dividends to make capital investment was.
Edit: also, I should clarify, he was the majority share holder, and the minority shareholders could thus not replace him with someone willing to pay dividends. He was not being sued for failing to seek profits, he was being sued for holding those profits hostage from other shareholders.
techcrunch.com
Hot