Eudimorphodon
Veteran Member
Things can be taken care of simply, it needs to be a lawful misdemeanor to try and patent/trademark/copyright non-human generated stuff. Period.
Up that to “firing squad” and maybe we’ll be getting somewhere.
Things can be taken care of simply, it needs to be a lawful misdemeanor to try and patent/trademark/copyright non-human generated stuff. Period.
Fair enough, although I still don't see much of a point discussing it here.That the EU has such different laws and regulation from the U.S. (including the examples you've given yourself) demonstrates that people have a very meaningful influence over businesses.
Well, I can provide a curated dataset with only that information and train an NN only on it. Given the small space, I can even fully reason about its outputs. It's just that such a network will be completely useless for any questions not related to "POP AF" and flags.If I ask you to create an NN that will never "decide" that a flag is preserved over a POP AF instruction, you can't do that, except to wrap it in traditional code.
There are two layers of randomness introduced. The initial state is randomized once ("seed"), and the sampling process introduces additional randomness ("temperature"). Reducing the temperature substantially reduces the "creativity" of the output, making it stay closer to the question at hand.And it's made all the worse by the fact that LLMs deliberately introduce randomness in their outputs.
I know. But since your argument doesn't hold in large parts of the real world, I treat it as a theoretical argument. And my counter is that LLMs are fully deterministic (assuming pseudo-randomness, which is true for all systems I looked at), so they are in theory as inspectable as any other code.That you cannot debug some compiled code given to you by someone else is not surprising, but that's not at all what I'm talking about.
Learn by example only will not provide good rules to follow. Learning languages show that quite well, as it takes most people unreasonably large amounts of time and effort to get to a point where their output is indistinguishable from native speakers.That right there says that there's something vastly different from reason going on: if you reason out the rules for addition of numbers they always work.
Related question: Do you see this as a problem with AI or as a problem with existing copyright law?There's a person who engaged into copyrighting all possible intervals and arrangements in Western music system in USA lately, to show the absurdity of the system.
Ah, there's the issue. You're looking at a minority of the software out there, and doing you're analysis as if that's all of it.I know. But since your argument doesn't hold in large parts of the real world, I treat it as a theoretical argument.....
In practice, LLMs are a lot less inspectable than traditionally programmed computer systems, but in a world of software-is-a-remote-controlled-and-constantly-updated-service, I don't see them as inspectable either.
If I were seeing LLMs struggle in similar ways, I'd be encouraged. But they're entirely unable to do simple reasoning, so it doesn't really matter how they struggle on complex reasoning.Also, brains struggle at complex reasoning, so I expect any system modelled after brain activity to struggle in similar ways.
I wasn't going to bother weighing in on this but since I'm here anyway: it seems to me very clearly a problem with copyright law. You don't need an AI to come up with songs that are the same "attack" on other composers.Related question: Do you see this as a problem with AI or as a problem with existing copyright law?
Related question: Do you see this as a problem with AI or as a problem with existing copyright law?
I wasn't going to bother weighing in on this but since I'm here anyway: it seems to me very clearly a problem with copyright law. You don't need an AI to come up with songs that are the same "attack" on other composers.
It's a very interesting difference. Since LLM is modeled on humans, I'm imagining that humans also use this kind of thinking process unconsciously.In my experiments, that reduced the tendency to hallucinate - but increased the tendency to form loops.
Yes, I think the feature you want is called RAG ; you can download and run some AI models, some even on CPU (not GPU) but of course they will run slower.Here's a question:
I'm a writer. All of my writing is set in the same continuity or story-world. I have thousands of documents containing notes, unfinished stories, ideas, world-building, etc. Without having it be uploaded to the internet in any way shape or form, can I plug this into an AI to answer questions about my own writing for me?
Not tone or story ideas. More along the lines of "what is this character's mother's maiden name?". I have all this info stored across thousands of documents. It can be frustrating to find it.
Just a random idea I am curious about exploring.
Biological neurons have a lot of inherent randomness (sometimes they fire when they shouldn't, don't fire when they should, have a long cooldown after firing), which is not modelled by technical neurons directly. Use of alcohol (or other drugs) makes them less predictable, so maybe that's an equivalent to "high temperature sampling".Since LLM is modeled on humans, I'm imagining that humans also use this kind of thinking process unconsciously.
A RAG system provides access to data sources (some data base or your documents) to LLMs. It's an AI-enhanced search engine using your own data.All of my writing is set in the same continuity or story-world. I have thousands of documents containing notes, unfinished stories, ideas, world-building, etc. Without having it be uploaded to the internet in any way shape or form, can I plug this into an AI to answer questions about my own writing for me?
Biological neurons have a lot of inherent randomness (sometimes they fire when they shouldn't, don't fire when they should, have a long cooldown after firing)
A RAG system provides access to data sources (some data base or your documents) to LLMs. It's an AI-enhanced search engine using your own data.
You could also use an LLM directly, but if you have lots of documents, they won't fit into the context. So you would fine-tune a base model on your documents, actually changing the model by teaching it your data. The result will know your world, but it will still keep its base training; that makes it usable in more creative ways, but less trustworthy for facts (expect the usual hallucination and unreliability issues).
Running LLMs locally is reasonably easy. When run on CPUs, it is usually faster to use only a subset of cores (LLMs are limited by memory bandwidth, not compute speed), and performance is acceptable for medium-sized models. Using GPUs is obviously much faster, but the limiting factor is video memory size (the LLM plus its working memory must fit into VRAM). It is possible to combine CPU and GPU execution, but I've never tried that. When I played with it, the performance characteristics were noticably different between CPU and GPU execution ("time to first output token" was very slow on CPU implementations).
Maybe, but actually modelling them is unlikely to be necessary. Quantum effects only become relevant at very small scales, and the structures we are working with here are much larger. Modelling electricity using voltage and current is perfectly fine unless we are talking nanoscale semiconductor technology, and one rarely uses that abstraction level to model computer programs.It is still under debate whether brain contains quantum systems.
That's not true in general. For some processes, especially at very small scales, knowing the relevant factors in sufficient detail is too hard or even impossible (related: Heisenberg's uncertainty principle). In most cases, one is better off modelling these as statistical noise rather than"This is random" is a placeholder for "stuff happens when it shouldn't", which is another way of saying, "we don't have enough understanding of the situation"
Even if you fine-tune a model on your documents, it's still a large language model and has seen tons of other training data. You will need to deal with hallucinations, especially when you ask questions for which your notes do not provide a clear answer. In general, model quality scales a lot with the data quality used during both training and fine-tuning. We currently just throw tons of low-quality trash at our models and hope for the best, which shows.What exactly do you mean by "less trustworthy for facts"? I'm primarily looking for something that can find me minor details I've forgotten but are written down somewhere in the myriad of notes.
Personally, I have never trained or fine-tuned a model myself, so I don't know how much time or effort goes into doing that. Your system specs are plenty good, I'd expect outputs within seconds rather than minutes.I'm fine with giving it a few weeks to chew on the data and don't mind if it takes several minutes to answer me.
Maybe, but actually modelling them is unlikely to be necessary. Quantum effects only become relevant at very small scales, and the structures we are working with here are much larger
That's not true in general. For some processes, especially at very small scales, knowing the relevant factors in sufficient detail is too hard or even impossible (related: Heisenberg's uncertainty principle). In most cases, one is better off modelling these as statistical noise rather than
Well it looks like the worst case scenario I will be out a few hours of my time figuring out how to set all this up. And as we've established, my time is worthless. I think I'll give it a whirl and see what happens.
Good luck and please report back on how it worked for you!
There's already tools that do this, though they cost money. That's where I got the idea, we are experimenting with one at work.Yeah please do, I'm also interested in the case of private document management. For me a discovery of things in the knowledge base would be perfectly fine, compiling and deriving and combining is a secondary but it would be great if an AI could give you some sort of a procedure that was sourced from multiple KB articles that weren't written step-wise.
I also have a box to test this out, so we might as well try it together.