If you mean the Star Trek stuff, though, then the specific terms for those are AGI
Even in Star Trek only Data, Lore (and Peanut-hamper) were intelligent, all the computers ran on what is being called ‘AI’ now. Massive DBs and search algorithms.
Search algorithms are, depending on the specifics, potentially “ai” now. If we’re tokenizing out vectors and running a straight match query (i.e. postgres full text search) that’s not AI - that’s just string matching. Some of the offerings get into NN guided or LLM powered… these tend to suck though because they’re unpredictable and inconsistent. That may just be the novelty of the technology though, we’ve had decades to work on small word exclusion and language specific dictionary mapping - it’s possible the consistency will get up and, at least when it comes to searching, everything really good already uses weird heuristics so it’s not like we can reason on why specific results are preferred, we just know they’re consistent.
The ship’s computer could whip up an AGI (Moriarty) in response to a simple command. The Federation later systematized this in the form of emergency holographic officers.
You are not logged in. However you can subscribe from another Fediverse account, for example Lemmy or Mastodon. To do this, paste the following into the search field of your instance: !canada@lemmy.ca
Even in Star Trek only Data, Lore (and Peanut-hamper) were intelligent, all the computers ran on what is being called ‘AI’ now. Massive DBs and search algorithms.
Search algorithms are, depending on the specifics, potentially “ai” now. If we’re tokenizing out vectors and running a straight match query (i.e. postgres full text search) that’s not AI - that’s just string matching. Some of the offerings get into NN guided or LLM powered… these tend to suck though because they’re unpredictable and inconsistent. That may just be the novelty of the technology though, we’ve had decades to work on small word exclusion and language specific dictionary mapping - it’s possible the consistency will get up and, at least when it comes to searching, everything really good already uses weird heuristics so it’s not like we can reason on why specific results are preferred, we just know they’re consistent.
The ship’s computer could whip up an AGI (Moriarty) in response to a simple command. The Federation later systematized this in the form of emergency holographic officers.