Hawk

joined 2 years ago
[–] [email protected] 3 points 16 hours ago

The whole notion of LSP has been nice.

[–] [email protected] 2 points 1 day ago

An LLM is an equation, fundamentally. Map a word to a number, equation, map back to words and now llm. If you're curious write a name generator using torch with an rnn (plenty of tutorials online) and you'll have a good idea.

The parameters of the equation are referred to as weights. They release the weights but may not have released:

  • source code for training
  • there source code for inference / validation
  • training data
  • cleaning scripts
  • logs, git history, development notes etc.

Open source is typically more concerned with the open nature of the code base to foster community engagement and less on the price of the resulting software.

Curiously, open weighted LLM development has somewhat flipped this on its head. Where the resulting software is freely accessible and distributed, but the source code and material is less accessible.

[–] [email protected] 2 points 1 day ago

The energy use isn't that extreme. A forward pass on a 7B can be achieved on a Mac book.

If it's code and you RAG over some docs you could probably get away with a 4B tbh.

ML models use more energy than a simple model, however, not that much more.

The reason large companies are using so much energy is that they are using absolutely massive models to do everything so they can market a product. If individuals used the right model to solve the right problem (size, training, feed it with context etc. ) there would be no real issue.

It's important we don't conflate the excellent progress we've made with transformers over the last decade with an unregulated market, bad company practices and limited consumer Tech literacy.

TL;DR: LLM != search engine

[–] [email protected] 8 points 1 day ago (1 children)

Yeah I did the smart thing and walked away from research.

Total dead end, Industry don't care at all.

[–] [email protected] 1 points 2 days ago

I use it a lot when I'm writing my notes (ie joplin/obsidian), I'll use flux or stable diffusion for a few iterations until I can create an image that Is consistent with what I'm writing.

It can be really convenient to be able to recognize an image as you're browsing through notes that are otherwise just filled with code or maybe a recount of the day.

I'm sure most consumers consider excessive use of generative AI to be in bad form. It certainly doesn't exude professionalism.

[–] [email protected] 2 points 2 days ago (1 children)

Oh, okay, I understand what you're saying now.

Yeah, I don't trust any of the VPN providers. There's just no evidence that they're trustworthy. I reach for Tor (or i2p sometimes).

I typically run all the torrenting stuff in a container, I've never actually used that VPN to browse. I just spin the container up and down when I want my bandwidth back.

[–] [email protected] 1 points 2 days ago (3 children)

I've had a good experience with AirVPN. I mean, I only use it for torrenting, but... Is there a good reason not to go with them for torrents?

[–] [email protected] 1 points 2 days ago* (last edited 2 days ago)

To be fair, wireguard is pretty painless.

[–] [email protected] 1 points 2 days ago

Absolutely that's what the internet was made for!

But family photos keep a bit more secure, Particularly if it's syncing directly from your phone, I take a lot of explicit photos of my wife, but also code that I'm writing on my computer, or the kids playing, etc.

[–] [email protected] 1 points 2 days ago

Keepass with rsync / unison or a local git server works pretty well too.

[–] [email protected] 2 points 2 days ago (5 children)

Airvpn has port forward i believe.

[–] [email protected] 6 points 2 days ago (2 children)

I don't actually dislike ai imagery, I think it can produce interesting imagery. However, I must concede that is an excessive use of boilerplate bog-standard AI imagery.

view more: next ›