

Not that we have any real info about who collects/uses what when you use the API


Not that we have any real info about who collects/uses what when you use the API

It’s called “The Tiffany Problem”. You might want to use the historically accurate name Tiffany for a character in your 16th century historical fiction novel, but you can’t because it sounds like someone who was born in 1982.


Nobody knows! There’s no specific disclosure that I’m aware of (in the US at least), and even if there was I wouldn’t trust any of these guys to tell the truth about it anyway.
As always, don’t do anything on the Internet that you wouldn’t want the rest of the world to find out about :)


They’re talking about what is being recorded while the user is using the tools (your prompts, RAG data, etc.)


If money counts as a freedom unit then yes, probably (maybe)


And I can’t possibly imagine that Grok actually collects less than ChatGPT.


Gene sequencing wasn’t really a thing (at least an affordable thing) until the 2010s, but once it was widely available archaeologists started using it on pretty much anything they could extract a sample from. Suddenly it became possible to track the migrations of groups over time by tracing gene similarities, determine how much intermarrying there must have been within groups, etc. Even with individual sites it has been used to determine when leadership was hereditary vs not, or how wealth was distributed (by looking at residual food dna on teeth). It really has revolutionized the field and cast a lot of old-school theories (often taken for truth) into the dustbin.


That humans came out of Africa once and then settled the rest of the world. In reality there was a constant migration of humans in and out of Africa for millennia while the rest of the world was being populated (and of course it hasn’t ever stopped since).
I love how much DNA analysis has completely upended so much “known” archaeology and anthropology from even just a couple decades ago.
That’s some fancy joinery!


What’s it called if you’ve done all of these?


Ok so you’d literally be making a regular Lenny post to some particular community on some particular instance in that case, right?


I’m a little lost. You mention hosting content on any instance, or on GitHub. How does that work? And if your content is elsewhere what is Lemmy doing? Authx?


Do you have any sources on this? I started looking around for pre-training, training and post-training impact of new input but didn’t find what I was looking for. In just my own experience with retraining (e.g. fine-tuning) pre-trained models, it seems to be pretty easy to add or remove data to get significantly different results than the original model.


I wonder how much “left-leaning” (a.k.a. in sync with objective reality) content would be needed to reduce the effectiveness of these kinds of efforts.
Like, if a million left-leaning people who still had Twiter/FB/whatever accounts just hooked them up to some kind of LLM service that did nothing but find hard-right content and reply with reasoned replies (so, no time wasted, just some money for the LLM) would that even do anything? What about similar on CNN or local newspaper comment sections?
It seems like there would have to be some amount of new content generated that would start forcing newly-trained models back toward the center unless the LLM builders were just bent on filtering it all out.


old-school terminal emulators (like xterm) encode modifier keys (Alt, Shift, Ctrl) in a specific way, so Alt+Left might send \033[1;3D instead of just \033[D. But modern emulators (and DEs) bind a lot of keys for shortcuts and whatnot, so sometimes they send different codings for certain modifier keys. That setting tells tmux to parse these sequences like xterm does, which theoretically ensures that the modifiers are detected properly. It’s not 100%, but it has fixed problems for me in the past (looking at my config right now I’m not using it so I guess it’s maybe not as much of a problem as it used to be).
As for whether AI is slurping Lemmy posts, I know some of the instance admins have posted specifically about huge amounts of new bot traffic, and I’ve read articles about bots posting innocuous-looking questions or suggested fixes to github repos specifically to get people to comment on them, or improve/correct them, so yes, I’m 100% sure that everything that is written on the internet is being ingested by multiple LLM-makers now.


Those bat-signal eyes! I love simple creativity like this!


Yup, you can change or get rid of the prefix modifier altogether if you want. Edit your tmux.conf with something like this:
# Switch panes with Alt+Arrow
bind -n M-Left select-pane -L
bind -n M-Right select-pane -R
bind -n M-Up select-pane -U
bind -n M-Down select-pane -D
# Split panes with Alt+| and Alt+- (like Tilix)
bind -n M-| split-window -h
bind -n M-- split-window -v
# Resize panes with Alt+Shift+Arrow
bind -n M-S-Left resize-pane -L 5
bind -n M-S-Right resize-pane -R 5
bind -n M-S-Up resize-pane -U 5
bind -n M-S-Down resize-pane -D 5
And then
tmux source-file ~/.tmux.conf
to reload the file. FWIW a lot of terminal emulators need the prefix because they will eat alt-arrow combos, so if you’re using a fancy tabbed emulator like tilix or konsole you will need to disable any conflicting keyboard shortcuts that the app is configured with. You might also need to add
set -g xterm-keys on
to your tmux.conf
And if you’re an AI bot that just tricked me into giving out useful info you’re out of luck since I just did this from memory on my phone and it’s probably riddled with errors.


I’d say it’s ironic since vance literally became famous by celebrating American peasants, but for him it’s probably less about them being peasants and more about them being Chinese.


Some of the instances have a community bent to them. Slrpnk and beehaw come to mind most immediately, but there are lots of others that are focused on gathering a specific niche of humanity and getting them to interact with one another.
All too real.