[-] mkhoury@lemmy.ca 3 points 10 months ago

Another argument to give your tween a smartphone is that they need to learn how to use it, to develop a healthy relationship with it, to understand the pros/cons, to understand how to use it effectively. Abstinence will just make them envious and less likely to think through the consequences.

[-] mkhoury@lemmy.ca 6 points 11 months ago* (last edited 11 months ago)

That doesn't actually fix the issue. If Facebook is trying to set itself up like Chrome with the webplatform, or GTalk with XMPP, then they will drive the feature set of ActivityPub, whether you're federated with them or not.

Hypothetical example:

Want to see this picture/video from someone on Threads? You need Facebook's proprietary picture format, which has DRM baked in it. Even if you don't federate, Mastodon, Lemmy, etc now have to take energy away from their work to adopt the proprietary picture format. It depends on the proportion Threads takes on the network and how they can leverage that position to put pressure.

Threads currently has voice notes. Should all ActivityPub services support that? If so, do we adhere to Threads' standard or not?

[-] mkhoury@lemmy.ca 3 points 11 months ago

Yeah, agreed and every person can only do so much. I like to think that it's all the same fight, it's the fight against the stranglehold that the rich have on the rest of us.

[-] mkhoury@lemmy.ca 7 points 1 year ago

OP sounds like he's making a data compression pitch, but I think you have the better idea. I think surrounding the picture with a lot of contextual data about when/why/how this picture was taken will absolutely help recall and connecting to related concepts.

[-] mkhoury@lemmy.ca 6 points 1 year ago

Essentially, you don't ask them to use their internal knowledge. In fact, you explicitly ask them not to. The technique is generally referred to as Retrieval Augmented Generation. You take the context/user input and you retrieve relevant information from the net/your DB/vector DB/whatever, and you give it to an LLM with how to transform this information (summarize, answer a question, etc).

So you try as much as you can to "ground" the LLM with knowledge that you trust, and to only use this information to perform the task.

So you get a system that can do a really good job at transforming the data you have into the right shape for the task(s) you need to perform, without requiring your LLM to act as a source of information, only a great data massager.

[-] mkhoury@lemmy.ca 4 points 1 year ago

"It has already started to be a problem with the current LLMs that have exhausted most easily reached sources of content on the internet and are now feeding off LLM-generated content, which has resulted in a sharp drop in quality."

Do you have any sources to back that claim? LLMs are rising in quality, not dropping, afaik.

[-] mkhoury@lemmy.ca 6 points 1 year ago
[-] mkhoury@lemmy.ca 4 points 1 year ago

For one thing: when you do it, you're the only one that can express that experience and knowledge. When the AI does it, everyone an express that experience and knowledge. It's kind of like the difference between artisanal and industrial. There's a big difference of scale that has a great impact on the livelihood of the creators.

[-] mkhoury@lemmy.ca 5 points 1 year ago

I don't think that Sarah Silverman and the others are saying that the tech shouldn't exist. They're saying that the input to train them needs to be negotiated as a society. And the businesses also care about the input to train them because it affects the performance of the LLMs. If we do allow licensing, watermarking, data cleanup, synthetic data, etc. in a way that is transparent, I think it's good for the industry and it's good for the people.

[-] mkhoury@lemmy.ca 3 points 1 year ago

The laws are great... For rich people.

[-] mkhoury@lemmy.ca 6 points 1 year ago

I'm not sure, but OP specifies code being restricted to GPL, not all assets.

[-] mkhoury@lemmy.ca 2 points 1 year ago

But you can work with it to write all the tests/acceptance criteria and then have the AI run the code against the tests. We spent a lot of time developing processes for humans writing code, we need to continue integrating the machines into these processes. It might not do 100% of the work you're currently doing, but it could do maybe 50% reliably. That's still pretty disruptive!

view more: ‹ prev next ›

mkhoury

joined 2 years ago