Skip Navigation
Republicans complain Joe Biden had too much energy at SOTU
  • It's important to be aware of their opinions because they quickly become policy and rhetoric while Dems do damage control and fail to fix the underlying issues Reps exploit. In this case, having an instance where they directly contradict their own "sleepy Joe" narrative can help deconstruct the beliefs of family members who have fallen for it.

  • Americans are confused, frustrated by new tipping culture, study finds
  • I saw a tip option at a coffee shop the other day, but it's very unclear who's even getting that tip. The cashier? They're not doing anything extra, so a tip doesn't make sense. The barista? If I get a complicated drink a tip might make sense. But I genuinely wouldn't doubt it if this ambiguity is taken advantage of and the business just pockets the tip and no one sees it.

  • Americans are confused, frustrated by new tipping culture, study finds
  • It's the same reason people hate ads. If you see a poster in a restaurant advertising some service, you don't care. But ads on the internet are shoved in your face and must be dismissed to get at the content. The equivalent of a tip jar for Square would be a button that says "tip your server" next to "continue". Instead, there's no easy way to dismiss the tip prompt - you have to go into custom and choose 0, which makes it an active choice which must be made in order to even continue, as if the server held the tip jar directly in your face and you had to push it aside to pay at the till. It's an imposition, one which targets neurodivergency surrounding motivation and social anxiety (eg people pleasing and depression). They took one of my spoons!

  • Humanoid robots are here, but they're a little awkward. Do we really need them?
  • I think it's moreso a matter of evolution. We know humanoid bodies can do anything we can do, so we start with that and make incremental improvements from there. We already do have plenty of other body shapes for robots (6-axis arms, SPOT, drones, etc) but none of them are general-purpose. Also, the robot shown in the article is pretty clearly not fully humanoid, it has weird insect legs probably because it's easier to control and it doubles as a vertical lift.

  • "Web Environment Integrity" is an all-out attack on the free Internet
  • Network effects. The "we" you're referring to could only be like 100 million at most, the vast majority of people don't have the technical know-how to switch, or to articulate exactly why they feel miserable every time they log in for their daily fix.

  • Drugmakers Are Set to Pay 23andMe Millions to Access Consumer DNA
  • Considering prior authorization is predicated on the fact that if they reject enough requests inevitably some people won't fight them, meaning they don't have to pay out, I wouldn't be surprised if they use a slightly better than chance prediction as justification for denying coverage, if they even need an actual excuse to begin with.

  • A mostly paper replacement for single-use RFID tags
  • For what it's worth I don't think they're proposing it will "solve" climate change - no single thing can. It's millions of tiny (alleged) improvements like this which eventually add up to taking pressure off of the environment. I see this kind of attitude a lot with stuff like paper straws or biodegradable packaging, as if the idea of a small but meaningful step in the right direction is laughable. It's fine to criticize them for the "improvement" actually being no better than the alternative, but I worry sometimes it comes across like any sort of improvement short of "solving" climate change isn't worthwhile.

  • Trump melts down on Truth Social after judge says his denial "rings hollow and untrue"
  • I'm not sure it even matters if he was "really" referring to Cohen, the fact that he's under a gag order to protect people means he should STFU. Every word he says, intentionally or "not", is potentially dangerous to the people involved. Isn't that the whole point of a gag order? Don't talk about the specifics of the trial, you can say it's unfair or whatever but the specifics are too dangerous to reveal. Why aren't his lawyers stopping him from saying this stuff? They're going to lose by contempt of court or something at this rate. Come to think of it, Alex Jones did the same thing in his trial.

  • AI-generated child sexual abuse images could flood the internet. A watchdog is calling for action
  • If we had access to the original model, we could give it the same seed and prompt and get the exact image back. Or, we could mandate techniques like statistical fingerprinting. Without the model though, it's proven to be mathematically impossible the better models get in the coming years - and what do you do if they take a real image, compress it into an embedding, then reassemble it?

  • AI-generated child sexual abuse images could flood the internet. A watchdog is calling for action
  • I respect your boldness to ask these questions, but I don't feel like I can adequately answer them. I wrote a 6 paragraph essay but using GPT-4 as a sensitivity reader, I don't think I can post it without some kind of miscommunication or unintentional hurt. Instead, I'll answer the questions directly by presenting non-authoritative alternate viewpoints.

    1. No idea, maybe someone else knows
    2. That makes sense to me; I would think there would be a strong pressure to present fake content as real to avoid getting caught but they're already in deep legal trouble anyway and I'm sure they get off to it too. It's hard to know for sure because it's so stigmatized that the data are both biased and sparse. Good luck getting anyone to volunteer that information
    3. I consider pedophilia (ie the attraction) to be amoral but acting on it to be "evil", ala noncon, gore, necrophilia, etc. That's just from consistent application of my principles though, as I haven't humanized them enough to care that pedophilia itself is illegal. I don't think violent video games are quite comparable because humans normally abhor violence, so there's a degree of separation, whereas CP is inherently attractive to them. More research is needed, if we as a society care enough to research it.
    4. I don't quite agree, rights are hard-won and easy-lost but we seem to gain them over time. Take trans rights to healthcare for example - first it wasn't available to anyone, then it was available to everyone (trans or not), now we have reactionary denials of those rights, and soon we'll get those rights for real, like what happened with gay rights. Also, I don't see what rights are lost in arguing for the status quo that pedophilia remain criminalized? If MAPs are any indication, I'm not sure we're ready for that tightrope, and there are at least a dozen marginalized groups I'd rather see get rights first. Unlike gay people for instance, being "in the closet" is a net societal good because there's no valid way to present that publicly without harming children or eroding their protections.
  • AI-generated child sexual abuse images could flood the internet. A watchdog is calling for action
  • The legality doesn't matter, what matters is that the sites will be flooded and could be taken down if they aren't able to moderate fast enough. The only long-term viable solution is image classification, but that's a tall ask to make from scratch.

  • California judge blocks school district’s forced outing policy for trans & enby kids
  • I think a sizable portion of the Republican voter base are those whose family have voted Republican for generations and they haven't thought enough about it to change their vote out of habit - or even worse, they're voting because they value group cohesion over truth and everyone else around them is Republican and speaks in their rhetoric. Why else would they vote for a party so wildly against their own best interests?

  • *Permanently deleted*
  • I grew up with CRTs and VCRs, hard pass. There's a certain nostalgia to it all: the bum-DOOON sound as its electron gun warmed up, the smell of ozone and tingly sensation that got exponentially stronger the closer you were, crusty visuals... But they were objectively orders of magnitude worse than what we have now, if nothing else than because they don't weigh 150 pounds or make you wonder if watching Rugrats in Paris for the 30th time on this monster is giving you cancer. Maybe it's because I'm techie, I've never really had much issue with "smart" TVs. Sure, apps will slow down or crash because of memory leaks and it's not as customizable as I'd like, but I might be satiated just knowing that if push comes to shove I can plug in a spare computer and use it like a monitor for a media system.

    I'm rooting it if it starts serving me out-of-band ads, though.

  • *Permanently deleted*
  • This is less an issue of "smartness" and moreso because analog signals degrade gracefully whereas digital signals are all or nothing unless specific mitigations are put in place. HDMI hits kind of a weird spot because it's a digital protocol based on analog scanlines; if the signal gets disrupted for 0.02 ms, it might only affect the upper half and maybe shift the bits for the lower half. Digital is more contextual and it will resynchronize at least every frame, so this kind of degradation is also unstable.

  • Artificial General Intelligence Is Already Here | NOEMA
  • There’s a lot of papers which propose adding new tokens to elicit some behavior or another, though I haven't seen them catch on for some reason. A new token would mean adding a new trainable static vector which would initially be something nonsensical, and you would want to retrain it on a comparably sized corpus. This is a bit speculative, but I think the introduction of a token totally orthogonal to the original (something like eg smell, which has no textual analog) would require compressing some of the dimensions to make room for that subspace, otherwise it would have a form of synesthesia, relating that token to the original neighboring subspaces. If it was just a new token still within the original domain though, you could get a good enough initial approximation by a linear combination of existing token embeddings - eg a monkey with a hat emoji comes out, you add tokens for monkey emoji + hat emoji, then finetune it.

    Most extreme option, you could increase the embedding dimensionality so the original subspaces are unaffected and the new tokens can take up those new dimensions. This is extreme because it means resizing every matrix in the model, which even for smaller models would be many thousands of parameters, and the performance would tank until it got a lot more retraining.

    (deleted original because I got token embeddings and the embedding dimensions mixed up, essentially assuming a new token would use the "extreme option").

  • Artificial General Intelligence Is Already Here | NOEMA
  • There’s a lot of papers which propose adding new tokens to elicit some behavior or another, though I haven't seen them catch on for some reason. A new token would mean adding a new trainable static vector which would initially be something nonsensical, and you would want to retrain it on a comparably sized corpus. This is a bit speculative, but I think the introduction of a token totally orthogonal to the original (something like eg smell, which has no textual analog) would require compressing some of the dimensions to make room for that subspace, otherwise it would have a form of synesthesia, relating that token to the original neighboring subspaces. If it was just a new token still within the original domain though, you could get a good enough initial approximation by a linear combination of existing token embeddings - eg a monkey with a hat emoji comes out, you add tokens for monkey emoji + hat emoji, then finetune it.

    Most extreme option, you could increase the embedding dimensionality so the original subspaces are unaffected and the new tokens can take up those new dimensions. This is extreme because it means resizing every matrix in the model, which even for smaller models would be many thousands of parameters, and the performance would tank until it got a lot more retraining.

  • Artificial General Intelligence Is Already Here | NOEMA
  • LLMs are not expert systems, unless you characterize them as expert systems in language which is fair enough. My point is that they're applicable to a wide variety of tasks which makes them general intelligences, as opposed to an expert system which by definition can only do a handful of tasks.

    If you wanted to use an LLM as an expert system (I guess in the sense of an "expert" in that task, rather than a system which literally can't do anything else), I would say they currently struggle with that. Bare foundation models don't seem to have the sort of self-awareness or metacognitive capabilities that would be required to restrain them to their given task, and arguably never will because they necessarily can only "think" on one "level", which is the predicted text. To get that sort of ability you need cognitive architectures, of which chatbot implementations like ChatGPT are a very simple version of. If you want to learn more about what I mean, the most promising idea I've seen is the ACE framework. Frameworks like this can allow the system to automatically look up an obscure disease based on the embedded distance to a particular query, so even if you give it a disease which only appears in the literature after its training cut-off date, it knows this disease exists (and is a likely candidate) by virtue of it appearing in its prompt. Something like "You are an expert in diseases yadda yadda. The symptoms of the patient are x y z. This reminds you of these diseases: X (symptoms 1), Y (symptoms 2), etc. What is your diagnosis?" Then you could feed the answer of this question to a critical prompting, and repeat until it reports no issues with the diagnosis. You can even make it "learn" by using LoRA, or keep notes it writes to itself.

    As for poorer data distributions, the magic of large language models (before which we just had "language models") is that we've found that the larger we make them, and the more (high quality) data we feed them, the more intelligent and general they become. For instance, training them on multiple languages other than English somehow allows them to make more robust generalizations even just within English. There are a few papers I can recall which talk about a "phase transition" which happens during training where beforehand, the model seems to be literally memorizing its corpus, and afterwards (to anthropomorphize a bit) it suddenly "gets" it and that memorization is compressed into generalized understanding. This is why LLMs are applicable to more than just what they've been taught - you can eg give them rules to follow within the conversation which they've never seen before, and they are able to maintain that higher-order abstraction because of that rich generalization. This is also a major reason open source models, particularly quantizations and distillations, are so successful; the models they're based on did the hard work of extracting higher-order semantic/geometric relations, and now making the model smaller has minimal impact on performance.

  • Pathos v Logos

    How do you argue with someone who's confused a lack of emotional connection to a topic with objectivity and rationality? Say a topic profoundly affects you and those you care about, but not the other person, so you get angry and flustered and they seem to think this means you're less objective as a result and it's an easy win.

    4
    Federated Search Engine?

    Considering the potential of the fediverse, is there any version of that for search engines? Something to break up a major point of internet centralization, fragility, and inertia to change (eg Google will never, ever, offer IPFS searches). Not only would decentralization be inherently beneficial, it would mean we're no longer compelled to hand over private information to centralized unvetted corporations like Google, Microsoft, and DuckDuckGo.

    10
    Semantic Functions library
    github.com GitHub - ConsciousCode/servitor: Easy, extensible creation of "semantic functions" which use LLMs to execute natural language tasks.

    Easy, extensible creation of "semantic functions" which use LLMs to execute natural language tasks. - GitHub - ConsciousCode/servitor: Easy, extensible creation of "semantic function...

    GitHub - ConsciousCode/servitor: Easy, extensible creation of "semantic functions" which use LLMs to execute natural language tasks.

    Not sure if this is the right place to put this, but I wrote a library (MIT) for creating "semantic functions" using LLMs to execute them. It's optimized for ergonomics and opacity, so you can write your functions like: python from servitor import semantic @semantic def list_people(text) -> list[str]: """List the people mentioned in the text.""" (That's not a typo - the body of the function is just the docstring, servitor detects that it returns None and uses the docstring instead)

    Basic setup: bash $ pip install .[openai] $ pip install .[gpt4all] $ cp .env.template .env Then edit .env to have your API key or model name/path.

    I'm hoping for this to be a first step towards people treating LLMs less like agents and more like inference engines - the former is currently prevalent because ChatGPT is a chatbot, but the latter is more accurate to what they actually are.

    I designed it specifically so it's easy to switch between models and LLM providers without requiring dependencies for all of them. OpenAI is implemented because it's the easiest for me to test with, but I also implemented gpt4all support as a first local model library.

    What do you think? Can you find any issues? Implement any connectors or adapters? Any features you'd like to see? What can you make with this?

    0
    InitialsDiceBearhttps://github.com/dicebear/dicebearhttps://creativecommons.org/publicdomain/zero/1.0/„Initials” (https://github.com/dicebear/dicebear) by „DiceBear”, licensed under „CC0 1.0” (https://creativecommons.org/publicdomain/zero/1.0/)CO
    ConsciousCode @beehaw.org
    Posts 3
    Comments 191