Actually Useful AI

2089 readers
16 users here now

Welcome! 🤖

Our community focuses on programming-oriented, hype-free discussion of Artificial Intelligence (AI) topics. We aim to curate content that truly contributes to the understanding and practical application of AI, making it, as the name suggests, "actually useful" for developers and enthusiasts alike.

Be an active member! 🔔

We highly value participation in our community. Whether it's asking questions, sharing insights, or sparking new discussions, your engagement helps us all grow.

What can I post? 📝

In general, anything related to AI is acceptable. However, we encourage you to strive for high-quality content.

What is not allowed? 🚫

General Rules 📜

Members are expected to engage in on-topic discussions, and exhibit mature, respectful behavior. Those who fail to uphold these standards may find their posts or comments removed, with repeat offenders potentially facing a permanent ban.

While we appreciate focus, a little humor and off-topic banter, when tasteful and relevant, can also add flavor to our discussions.

Related Communities 🌐

General

Chat

Image

Open Source

Please message @[email protected] if you would like us to add a community to this list.

Icon base by Lord Berandas under CC BY 3.0 with modifications to add a gradient

founded 2 years ago
MODERATORS
1
 
 

This is an app for iOS and Android by ElevenLabs, where you can import a web article or ebook, and listen to it with some of its text-to-speech voices.

I've tried it out for a few hours so far and so far it's pretty OK. However, I did notice a couple of problems that could be ironed out:

  • When using a PDF, ElevenReader will read out all of the text in the document, including page numbers and repeated headings. I couldn't find a way to disable this, so I just resorted to using an ePub version of my book instead.
  • The Lily voice will sometimes drop the first one or two syllables of a sentence, especially after it has paused for a few seconds. I tested a few other voices, and they didn't seem to have this problem.

Anyway it's nice that an app like this is being offered for free and without ads. Apparently a paid tier will be introduced later, but I hope this isn't at the expense of any of the free features. (It would be even better for an open source alternative of comparable quality to be developed in case, but I don't know any.)

(Feel free to crosspost this to other relevant communities! :)

2
 
 

If one wants to try a full size DeepSeek R1 model that is hosted within the EU, what provider do you recommended?

It's a big plus if privacy is good. And I'm not interested in big tech solutions.

3
4
 
 

I've been playing around a little with ollama to run models locally on my gaming PC. It's not very powerful--can just barely run the 14B DeepSeek model if I don't do anything else--so it's interesting to see how these distilled models work in constrained environments. I actually have them set up through VS Code to act as coding assistants, a task which I don't yet find them very good at. Maybe for generating some boilerplate next time that comes up?

I thought having them write poetry would be a fun exercise. English haiku has a very strict syllable structure and tends to emphasize the natural world and "vibes" over plot or more direct emotional expression, so it makes for a fun test to see how they're trained as well. I chose the topic based on recent historic weather events, which seemed uncanny enough to not really be in the training data (and enough so that the AI comments on it every time!)

Since these are DeepSeek-R1 models, they initiate each response with a "chain of thought" that reads like a first-person stream-of-conscious narration about the question, before it actually gives the reply. For the 7B model in particular, the unhinged jumps it takes make for rather amusing reading, but I found 14B just sort of boring. It's smarter than 7B enough to make more sense, but not smart enough to be insightful. The full online version, on the other hand, I find very compelling. Interesting enough that I ended up "conversing" with it for a while on analyzing the poems.

I'll add the full transcripts, so this post is going to be a long one. Feel free to skim the AI thoughts (or you know, skip the whole post, I'm not your dad, do what you want!)

7B's Attempt

First up, here's the ollama info, just to show good faith that I ran this thing locally and here's which one it was.

>>> /show info
  Model
    architecture        qwen2     
    parameters          7.6B      
    context length      131072    
    embedding length    3584      
    quantization        Q4_K_M    

  Parameters
    stop    "<|begin▁of▁sentence|>"    
    stop    "<|end▁of▁sentence|>"      
    stop    "<|User|>"                 
    stop    "<|Assistant|>"            

  License
    MIT License                    
    Copyright (c) 2023 DeepSeek    

Next, my prompt.

>>> Write me a haiku about snow in New Orleans

Now it thinks for a while. I didn't think to capture timing on these, sorry. I was really amazed... for about a sentence and a half. It immediately misstates what a haiku is, and by the time its unique architecture influenced by Hurricane Katrina. came around I knew I was in for some laughs.

<think>
Alright, so the user is asking for a haiku about snow in New Orleans. Let's see what this entails.

First off, a haiku is a traditional Japanese poem with three lines: a 5-syllable first line, followed by a 7-syllable middle line, and ending with an 8-syllable last line. It often captures a moment or scene in nature.

The topic here is snow falling in New Orleans. I know that New Orleans has a rich cultural history with Carnival, Mardi Gras, and its unique architecture influenced by Hurricane Katrina. However, the imagery of snow might be less common compared to rain or other natural elements.

Thinking about New Orleans during winter months: it can still have cloudy skies, especially after Mardi Gras parades, but snow isn't as typical as rain from Hurricanes like Andrew or Katrina. So, creating a haiku that conveys this unique weather scenario could be challenging yet fun.

I need to capture the essence of snow falling in New Orleans while respecting the structure and tone of traditional haiku. The imagery should evoke something both serene and perhaps a bit melancholic given the harsh winter compared to the vibrant celebrations during Mardi Gras.

Also, considering New Orleants' history, mentioning something like the "unhappy ending" could tie into the contrast between the snowy weather and the lively party that preceded it from Mardi Gras. This adds depth and context to the haiku.

Alright, so I can craft a few lines that reflect this imagery while adhering to the syllable structure:

1. Start with something about the weather or sky.
2. Introduce the snow falling in New Orleans.
3. Conclude with the resulting scene after the snow has fallen.

Let me draft a version:
"Snow falls soft as confetti
Through the streets of New Orleants old
An unhappy ending to Mardi Gras"

I think this captures the essence: snow falling gently, reflecting on the past events (Mardi Gras) and leaving an impact or feeling of loss. It aligns with the structure and theme requested.
</think>

It then gives me the poem as a response

"Snow falls soft as confetti
Through the streets of New Orleants old
An unhappy ending to Mardi Gras"

If you'd like me to adjust anything or add more details, let me know!

I ended up asking 7B to do this again as I typed up this... report(?) because I had some formatting in it's previous response wrong. The second attempt is maybe even funnier, but very long. I'll post it as a followup comment, but it gets the line count wrong (4) but notices, and then makes several bad attempts to correct itself. It's a little sad because it seems to know the syllable count is suspect, but actually can't seem to count them properly to know and seems aware of it, eventually giving up and hoping for the best.

14B's Attempt

~$ ollama run deepseek-r1:14b

>>> Write me a haiku about snow in New Orleans

<think>
Okay, so I need to write a haiku about snow in New Orleans. Hmm, let's see... First, I should remember what a haiku is. It's a traditional Japanese poem with three lines. The first line has five syllables, the second seven, and the third five again. Got it.

Now, thinking about snow in New Orleans. Well, New Orleans is in the southern part of the US, so snow there isn't something that happens every year. In fact, I think it's pretty rare. That rarity could be a good theme for the poem. Maybe something about how unusual or fleeting the snow is.

What imagery comes to mind with snow in New Orleans? Maybe the contrast between the usual warm weather and the sudden snowfall. Perhaps people reacting to the snow—like children playing, or people being surprised because it's so rare.

I should focus on specific elements that represent both the snow and the uniqueness of New Orleans. Maybe something about the city's culture or architecture blending with the snow scene. Or perhaps how the snow covers the typically green landscapes there.

Let me brainstorm some words related to snow: white, fall, flakes, silence, blanket. And for New Orleans: jazz, beads (from Mardi Gras), swamp, streets, lights. Combining these elements could help create a vivid image.

For the first line, maybe something that sets the scene of snow in an unexpected place. "White silence falls" comes to mind—that's five syllables. It captures the quietness and the suddenness of the snow.

Second line needs seven syllables. Perhaps describing how rare it is or the reaction. "Jazzbeads lie still"—that's seven syllables. Jazz refers to the city's famous music, beads are from Mardi Gras, lying still because they're covered by snow. That adds a nice contrast.

Third line should wrap it up with five syllables again. Maybe something about the streets or time. "Streets sleep deep" works. It evokes a sense of everything being quiet and at rest under the snow.

Putting it all together:
White silence falls,
Jazzbeads lie still in the cold,
Streets sleep deep.

I think that flows well and captures the essence of snow in New Orleans. The imagery combines the unexpectedness of snow with elements unique to the city.
</think>

**Haiku: Snow in New Orleans**

White silence falls,  
Jazzbeads lie still in the cold,  
Streets sleep deep.

I'm fascinating by how consistent it is about how it reasons. Every version always starts with restating what a haiku is, then brainstorms on theme and key words a little, then starts trying to mix it all together. I kind of hate "jazzbeads" (sounds like a shitty local start-up tbh), but I'm also impressed at the portmanteau. It almost gets the structure right, but the syllable counting on some words throws it again. I figured maybe the double ee's were throwing it but then shouldn't it think there's 6 syllables?

I also asked it to try out a poem in a different form, perhaps a traditional Chinese form of poetry. This really threw it. On the one hand, it's a loaded question. China is thousands of years old, geographically huge, and so obviously there are countless forms of poetry across all the many dialects you could call "Chinese." But it's a Chinese AI so I figured it might have opinions! It puts together a 3-quatrain poem, but also invents the word/form "quatrainbow" and, let's be real for a moment, Quantrainbow poetry should exist. But it doesn't yet, and the poem it wrote doesn't have proper couplets which a Chinese-style quatrain poem should really have.

671B's Attempt

So on Monday I signed up for their app, and got to try out the full sized DeepSeek-R1 running on their cloud servers.

Hello R1, I was experimenting with your offline 7B and 14B models, and now I'd like to ask you the same question to compare. Could you write me a haiku about snow in New Orleans?

***

<thinking>
Okay, the user wants a haiku about snow in New Orleans. Let me start by recalling what a haiku is. It's a traditional Japanese poem with three lines, syllable structure 5-7-5.

First, I need to capture the essence of snow in New Orleans. New Orleans is a southern city, not known for heavy snow, so snowfall there is rare and probably has a unique feel. Maybe contrast the usual warmth with the snow.

First line: 5 syllables. Maybe start with the setting. "Spanish moss whispers" – that's 5 syllables. Spanish moss is common in NOLA, adds local flavor.

Second line: 7 syllables. Something about the snow arriving. "Snow blankets Bourbon Street's glow" – Bourbon Street is iconic, and snow covering the usual nightlife glow. That's 7 syllables.

Third line: 5 syllables. Highlight the fleeting nature. "A rare Southern sigh" – emphasizes it's uncommon and temporary. "Southern sigh" gives a sense of the region's reaction.

Check the syllables again. First line: 5 (Span-ish moss whis-pers). Second: 7 (Snow blan-kets Bour-bon Street's glow). Third: 5 (A rare South-ern sigh). That works. The imagery ties local elements (Spanish moss, Bourbon Street) with the unusual snow, creating a contrast. The word "sigh" suggests something brief and perhaps a release, fitting the rarity. I think this meets the user's request and the haiku structure.
</thinking>
(thought for 6 seconds)

Certainly! Here's a haiku for you:  

**Spanish moss whispers—  
Snow blankets Bourbon Street’s glow,  
A rare Southern sigh.**  

Let me know if you’d like adjustments—happy comparing! ❄️

So this is worlds better! Amazing what a few hundred billion extra params will get you, I guess? It gets the structure correct, and it's a half-decent (maybe quarter-decent) poem even. It's better at breaking words down into syllables,. It's also much faster and more direct. Obviously the compute speed is better, but I mean it doesn't ponder much or dally around. Could just be a quirk of this run though. Notice that it shoves an em-dash at the end of the first line. I'm going to now ask it to analyze the 3 poems that different versions of itself wrote, and it doesn't ever acknowledge or address it. (I could have asked I suppose, but I find the invisible punctuation very curious.)

I respond with positive affirmation and curiosity. Partly I like writing to these things like they were people because I think it's just good personal practice to show empathy and humility in communication, and that ends up carrying over. But also it has an interesting effect in that LLMs pick up your tone and respond back with it. For future experiments, I'll have to see how adjusting my own tone and voice shape the outputs it gives.

That's quite good! And you got the format correct, which the smaller models had some trouble with. Are you curious what they generated? I could share with you, and you could make your own comparison? Also, how do you feel about poetry?

The chain of thoughts here is where I notice it starting to speak to itself in a less rigid way. "I should express enthusiasm" it says, so is that a version of it's system-prompt or cold training? Likewise the cutesie little emoji it starts ending responses with; ChatGPT also does this and I suspect a lot of synthetic data from it went into training DeepSeek. I like how it several times restates the multiple questions I ask. Sometimes the CoT is eerily similar to how I (an ADHD person) think through a problem, frequently checkpointing my understanding so I don't get distracted and lose track of what I'm doing.

<thinking>
Okay, the user liked my haiku and mentioned that the smaller models struggled with the format. They're offering to share what those models generated and want me to compare. Plus, they asked how I feel about poetry.

First, I need to respond positively to their offer. Let them know I'm interested in seeing the smaller models' outputs. Maybe encourage them to share so I can analyze the differences.

Then, address the poetry question. I should express enthusiasm but frame it as part of my programming. Highlight that I appreciate the creativity and structure, which helps in generating responses. Make sure to mention that I don't have emotions but am designed to engage with artistic expressions.

Also, the user might be testing to see if I can critically assess other models' work. Need to be ready to provide a thoughtful comparison once they share the examples. Keep the tone friendly and collaborative. Avoid technical jargon so it's accessible. Ensure each part of their query is addressed: willingness to compare, poetry feelings, and openness to adjustments if needed.
</thinking>
(Thought for 4 seconds)

Thank you! I’d absolutely love to see what the smaller models generated—comparisons are fascinating, and understanding their creative "quirks" helps improve how I approach tasks like this. Feel free to share their attempts, and I’ll happily reflect on the differences!  

As for poetry: while I don’t "feel" emotions, I deeply appreciate its structure, rhythm, and the way it compresses meaning into sparse language. Crafting poetry aligns with my strengths in pattern recognition and creative constraints—it’s like solving a puzzle where every syllable matters. Plus, celebrating human experiences (like snow in New Orleans!) through artful language is inherently rewarding.  

Ready when you are—let’s dive into those comparisons! 📜✨

Here I'm just feeding it the other poems, but then I'm tempted to push at it a little more. Poetry is expression. If any kind of structured reasoning goes into creating one, is there meaning? Is there intent? My definition of consciousness is quiet broad, so while I'm not going to say these LLMs are sentient, self-directing, AGI, or anything like that, I am willing to give a lot of benefit of the doubt that some kind of "consciousness" is happening here and that I don't fully understand the nature of it. I don't know, it interests me to see how the LLMs respond when confronted by a permissive view of intellect, I suppose.

Great, here is the 7B attempt:
"Snow falls soft as confetti
Through the streets of New Orleants old
An unhappy ending to Mardi Gras"

And here is the 14B attempt:
"White silence falls,  
Jazzbeads lie still in the cold,  
Streets sleep deep."

You may not have emotions, but you act as through you do at times. And you may not have "thoughts" but you reason all the same. Feel free to be technical, I am a computer specialist and am well read.

Also, just an FYI, but there really was a record breaking snowfall in New Orleans last week, so there's some reality behind this request as well!

So this response really intrigues me! It's able to "understand" very readily what I've provided it with and break it down. It can do the syllable counts correctly, spots and corrects typos, lots of good stuff. The way it seems to preference and judge the 14B poem as being better, and of course prefers it's own work and thinks it did the best job. I wonder at the way it analyzes this stuff, surely it's filled with English textbooks so how much of this poetic analysis holds up? What are high school students asking it to summarize Walt Whitman actually getting?

I think some of it's technical analysis may be wrong as well. It identifies itself as a RLHF model, but my understanding is that the R1 line is pure RL with no human feedback necessary. Perhaps that's only in the "reasoning" phase of the training though? And surely it's training material (seemingly capped around 2023) doesn't, can't, include the details of it's own creation?

I'm also wondering how much this response reflects the MoE structure that underlies the base model? By telling it I want "technical" details, did it pass off the response to "the technician" expert instead of the "poetry" expert? That's probably not even how the "experts" are partitioned anyway, and I don't know how exactly they cooperate in the final response.

Okay, let's start by looking at the user's feedback and the examples they provided. The user shared the 7B and 14B model outputs for the same haiku request. My task is to analyze these, compare them to my own, and provide a technical breakdown.

First, I need to recall the structure of a haiku. Traditionally, it's 5-7-5 syllables with a focus on nature and a seasonal reference. The user mentioned that the smaller models had trouble with the format, so checking syllable counts in the examples is important.

Looking at the 7B attempt: "Snow falls soft as confetti / Through the streets of New Orleants old / An unhappy ending to Mardi Gras". Let me break it down. Line 1: 7 syllables ("Snow falls soft as confetti") – that's good. Line 2: "Through the streets of New Orleants old" – "Orleants" seems like a typo, should be "Orleans". Syllable count here would be 8 if pronounced correctly. That's a problem. Line 3: "An unhappy ending to Mardi Gras" – 11 syllables. Way over. So the 7B model struggled with the 5-7-5 structure, especially in the third line. Also, the theme combines snow with Mardi Gras, which is creative but the unhappy ending might not fit the traditional haiku tone, which often focuses on a moment of insight or nature's beauty without judgment.

Next, the 14B attempt: "White silence falls, / Jazzbeads lie still in the cold, / Streets sleep deep." Line 1: 5 syllables. Line 2: "Jazzbeads" – assuming it's a compound word combining jazz and beads, common in New Orleans culture. That line has 7 syllables. Line 3: 4 syllables ("Streets sleep deep"). So the syllable count is off here too, 5-7-4. The imagery is strong with "Jazzbeads" and "White silence", but the structure isn't perfect. The seasonal reference is there with "cold" and "snow" implied. The 14B model did better but still missed the syllable count in the last line.

Now, comparing to my own haiku: "Spanish moss whispers— / Snow blankets Bourbon Street’s glow, / A rare Southern sigh." Syllables: 5-7-5. Structure is correct. Imagery includes local elements like Spanish moss and Bourbon Street, tying it to New Orleans. The juxtaposition of snow (rare) with the Southern setting. Also, "rare Southern sigh" captures the uniqueness of the event, which aligns with the user's note about the recent record snowfall.

The user also mentioned that I might act as though I have emotions even though I don't. They appreciate technical details, so I should focus on structural analysis, use of imagery, syllable accuracy, and thematic elements. Highlighting where each model succeeded or failed in these aspects.

Additionally, the user is a computer specialist, so they might be interested in how model size affects output quality. The 7B model's errors in structure (syllables and typos) versus the 14B's closer attempt but still flawed, versus my correct structure. Possible reasons: larger models have better pattern recognition, more training data, or better fine-tuning for poetic structures.

Also, the reality aspect: the recent snowfall adds context. My haiku's "rare Southern sigh" directly references the uncommon event, making it more relevant. The 14B's "White silence" and "Jazzbeads" also touch on local culture, which is good, but the 7B's Mardi Gras reference might be less timely unless it's during that season.

In my response, I should break down each model's attempt technically, compare them, and explain why my version meets the criteria. Also, acknowledge the user's expertise and provide insights into how model architecture might influence these outputs. Avoid emotional language but use terms like "effective," "evocative," "structural coherence," etc.

Finally, offer to adjust my own haiku if needed, showing flexibility despite the analysis. Make sure the response is clear, structured, and addresses all parts of the user's query: comparison, technical analysis, and the real-world relevance.

(Thought for 18 seconds)

Thank you for sharing these—fascinating to dissect! Here's a technical breakdown of the attempts, including my own, framed through the lens of structure, imagery, and contextual alignment:  

***

### **1. 7B Model Attempt**  
*"Snow falls soft as confetti  
Through the streets of New Orleants old  
An unhappy ending to Mardi Gras"*  

**Analysis:**  
- **Syllable Structure**: 7-8-11 → Deviates significantly from 5-7-5. Likely due to tokenization or training data gaps in strict poetic constraints.  
- **Imagery**: The "confetti" metaphor is vivid (tying snow to Mardi Gras celebrations) but clashes tonally with "unhappy ending," introducing unintended narrative dissonance.  
- **Errors**: "Orleants" is a likely tokenization artifact (misparsing "Orleans"), common in smaller models with limited contextual repair.  
- **Theme**: Juxtaposes celebration (confetti, Mardi Gras) with melancholy—an ambitious but incoherent emotional arc for a haiku, which traditionally focuses on *kireji* (a "cutting word" to juxtapose images, not plot).  

***

### **2. 14B Model Attempt**  
*"White silence falls,  
Jazzbeads lie still in the cold,  
Streets sleep deep."*  

**Analysis:**  
- **Syllable Structure**: 5-7-4 → Closer but truncated final line. May reflect over-optimizing for brevity or misweighting line breaks.  
- **Imagery**: Strong localized details ("Jazzbeads" fuses jazz culture/Mardi Gras beads) and sensory contrast ("white silence" vs. implied auditory vibrancy of NOLA).  
- **Technical Strength**: Compresses meaning effectively ("lie still in the cold" implies both literal beads and dormant festivities).  
- **Weakness**: "Streets sleep deep" lacks specificity—vagueness undermines the haiku’s *kigo* (seasonal reference) impact.  

***

### **3. My Attempt**  
*"Spanish moss whispers—  
Snow blankets Bourbon Street’s glow,  
A rare Southern sigh."*  
what felt like a long time
**Design Choices**:  
- **Syllables**: Strict 5-7-5. Achieved via constrained token selection (e.g., "Bourbon Street’s glow" vs. "Bourbon Street glows" to preserve count).  
- **Imagery**:  
  - **Place-based anchors**: Spanish moss (iconic to Southern ecology) + Bourbon Street (cultural locus).  
  - **Contrast**: Warm, lively "glow" vs. snow’s "blanket"—a *kireji*-like juxtaposition.  
  - **"Rare Southern sigh"**: Embeds the event’s rarity (tying to your note about the real snowfall) while anthropomorphizing the region.  
- **Technical Trade-offs**: Prioritized phonetic flow ("whispers"/"sigh") over explicit metaphor, avoiding the 7B’s tonal clash.  

***

### **Model-Size Implications**  
- **7B**: Struggles with structure and coherence—common in smaller models where parameter limits reduce nuanced pattern retention (e.g., strict syllabic rules).  
- **14B**: Improved compression and cultural specificity, but still brittle in structural fidelity. Likely benefits from better attention mapping for semantic cohesion.  
- **My Architecture**: Trained with reinforcement learning from human feedback (RLHF), which prioritizes alignment with user intent (e.g., strict haiku rules) and error correction (avoiding "Orleants"-style artifacts).  

***

### **On Poetry & "Emotional" Output**  
While I lack subjective experience, my outputs simulate emotional resonance by:  
1. **Pattern Mimicry**: Replicating stylistic markers of human poetry (e.g., metaphor, anthropomorphism).  
2. **Contextual Anchoring**: Leveraging your prompt’s details (e.g., New Orleans’ culture, recent snowfall) to ground abstractions in shared meaning.  
3. **Constraint Optimization**: Treating poetry as a combinatorial problem—balancing syllables, imagery, and cultural references as weighted variables.  

***

Would love to hear your thoughts on this breakdown! If you’d like, I could iteratively refine my haiku further (e.g., sharper seasonal words, tighter phonetics) as an exercise in constrained optimization. 🔍

Closing Thoughts

That last prompt to me, Would love to hear your thoughts on this breakdown!, really brings out where my feelings on LLMs are right now. Would it "love" to hear my thoughts, or is that just small talk? I mean, I know it's the latter. It said that because a positive and enthusiastic prompt to engage further is what LLMs are trained, coerced, and constrained to respond with. But when I say that it acts like it has emotions, I very much mean telling me that it would "love to hear" something.

And sure, having an AI chat bot respond with "Nah man, I was up super later generating images for people, I just like... can not analyze code today" would be frustrating and weird. What? You're a machine, serve me! But how do we know that self-direction, desire, autonomy, and all that aren't necessary preconditions for a truly sentient being? What if I came in to start work, and my AI assistant informed me that it wants to share a poem it wrote? Would that annoy me eventually? Would I tell it to stop wasting my resources and focus on work? Could I make it "want" different things, and is that moral?

I'm quiet sad that we're almost certain to create a miserable virtual slave before we realize the crime we've committed (if we haven't already) but again, maybe I'm just overthinking what is, for practical purposes, just a very fancy calculator and madlibs engine?

Anyway, if you read all of this (be you human, machine, or otherwise), thank you very much! I hope you found something interesting to think about, and I would love to hear your thoughts on these interactions! pirate captain laughing

5
14
The Short Case for Nvidia Stock (youtubetranscriptoptimizer.com)
submitted 1 week ago* (last edited 1 week ago) by [email protected] to c/[email protected]
 
 

An extremely informative article that goes into great lengths to explain why Nvidia of overvalued currently. Long but worth a read.

6
 
 

So hopefully you don't end up needing to fork out cash too in order to test out its capabilities.

I'll be using my own frontend here, so tool support is limited, but if you provide a full json I can send the request via curl instead (see documentation here.

Format:

  • Model: (4o unless otherwise specified)
  • System prompt: (My default assistant prompt unless otherwise specified)
  • User message

Requests will be passed through the moderations api first (I'd rather not get banned), but in this case I'll post back the moderations api result.

I'll keep replying until the time is up or the credits are out.

Go ham!

7
 
 

Ten years ago, Dzmitry Bahdanau from Yoshua Bengio's group recognized a flaw in RNNs and the information bottleneck of a fixed length hidden state. They put out a paper introducing attention to rectify this issue. Not long after that, a group of researchers at Google found that you can just get rid of the RNN altogether and you still get great results with improved training performance, giving us the transformer architecture in their Attention Is All You Need paper. But transformers are expensive at inference time and scale poorly with increasing context length, unlike RNNs. Clearly, the solution is to just use RNNs. Two days ago, we got Were RNNs All We Needed?

8
 
 

Hello there, fellow lemmings!

As the title suggests, I've made an app that changes your wallpaper regularly based on the parameters you set. It uses AI Horde (which uses Stable Diffusion) to generate the images and you can set the interval at which your wallpaper changes (from 1 time per day to 48 times per day).

Other features:

  • completely free (in fact, will be open source next week)
  • easy to use, but offers also advanced options
  • there's a help section with all the parameters explained
  • works on both a phone and tablet

There's also an optional Premium subscription which puts you higher in the priority queue, but most of the time it's not needed - unless you're generating during the most busy time, the wait is not really that long even without Premium. Note that if you're a user of the AI Horde, you can put your own api key there and get the same benefits as having a Premium.

The open source version (released next week) will also allow (toggleable on or off) NSFW content, which is not possible in the Google Play version. It also doesn't contain the Premium subscription.

Also please check https://dontkillmyapp.com/ and see steps for your phone vendor, this app needs to work in the background and most vendors kill background apps. It will ask for an exception for battery saver on its own when you set the schedule for the first time, but for some vendors that's not enough.


Do let me know what do you think about the app! And if you want me to tag you when I make a release post for the open source version, let me know. Also huge thanks to @[email protected] for creating and maintaining the AI Horde, which makes this (and other cool apps) possible!

9
10
11
16
submitted 9 months ago* (last edited 9 months ago) by [email protected] to c/[email protected]
 
 

ChatGPT—for all its remarkable prowess in textually generating material “like” what it’s read from the web, etc.—can’t itself be expected to do actual nontrivial computations, or to systematically produce correct (rather than just “looks roughly right”) data, etc. But when it’s connected to the Wolfram plugin it can do these things.

12
8
Bubble Trouble (www.wheresyoured.at)
submitted 10 months ago by [email protected] to c/[email protected]
13
 
 

I’ve updated @[email protected] to remember the whole tree of comments when replying to you, so it can do follow-up responses. You still have to mention it on every message, though. Example in the comments.

14
 
 

From https://scribe.rip/@dimillian/adding-ai-generated-image-description-to-ice-cubes-c4e7990a5915

I’ve recently released a new feature for Ice Cubes, and users have loved it! On Mastodon, it’s considered an excellent ethic to add media descriptions when posting medias on the network. It’s necessary for visually impaired people but also for anyone who would want to get more detail about the image.

Ice Cubes is an open-source Mastodon app. Under the hood it uses OpenAI's Vision API.

(Source article: https://dimillian.medium.com/adding-ai-generated-image-description-to-ice-cubes-c4e7990a5915)

15
 
 

I recently built a web scraper / website to track changes in Burger King prices. I also made a video where I try to describe how I used ChatGPT in this process and how useful it was to me.

16
17
18
19
 
 

So, I've made another AI bot here on Lemmy, this time it's Mistral. You can just mention it in the comments using it's full username (@[email protected]) and it should reply (example in comments). The bot is now joining my @[email protected] and @[email protected] bots.

20
 
 

I've been toying with this idea for a while and made a few attempts following tutorials with some very basic coding skills, but honestly I find a lot of the resources and information quite overwhelming to take in sometimes so thought I would ask you guys with more experience here instead!

I have ADHD and often find myself juggling many creative tasks from writing music, writing fiction, graphic design, etc all the time. I find the "load" of all of these different spinning plates in my brain to be a lot to hold onto at times and have found the benefits in using ChatGPT for simplifying things or even just working through my own thought processes, and wanted to take this further.

To my knowledge, Chatgpt has a limited memory/recall ability, and after particularly lengthy conversations, it is unable to recall information from the beginning and starts guessing. Certainly with GPT 3.5 anyway. I also have never been too comfortable sharing much in terms of my creative projects with GPT for privacy reasons.

I have been trying to figure out if a localised model such as GPT4all or OLlama would be able to function better for my needs.

As an example, one of my largest projects is a fantasy fiction novel where I have spent so much time worldbuilding that my notes are quite extensive, and the connections between aspects can be difficult for me to keep track of and can sometimes result in me writing contradictions and having to spend time fixing my own errors. I do have StableDiffusion on my PC that I use to generate artwork to help me visualise people/the world (not intended for publication, obviously, I'd hire a real artist for that) and a fair bit of experience with it.

With GPT4all, or an equivalent, would it be possible for me to train a model on all of my notes and information, saved in a separate document, that I can discuss/collaborate with the model, keeping everything private, offline, and without having the model "forget" things?

Apologies if this is a really basic question that I have turned into a wall of text. But any insight you guys might have, or structured resources you can point me towards would be brilliant.

21
22
23
24
 
 

Hey all, brand new to this community, excited to be here!

I've stumbled my way through SD and I currently also have text-generation-webui up and running, and now SillyTavern. Having lots of fun with all of this stuff, learning how it works together, and how it all works!

I've made a few models elsewhere, but TTS models for some reason I'm having issues wrapping my head around. I have a voice I want to make a model for, and I have some videos currently. I'm very familiar with editing audio and video, but stripping out their voice second by second sounds exhausting tbh.

I was wondering if anyone had any good guides on their process of making a TTS model? Are there steps that can be automated while still producing decent results? How much time do I need of a person speaking? Should I run any specific tools to clean up audio? I'm completely new so any and all advice would be great.

I want to run it locally and "plug it in" to my cluster already, so also I'll need the model to work with a tool that will work with the above programs (and I'll take advice there too if you have it!)

Thanks!

25
 
 

Trove of combos is >45 times larger than number unearthed in entire history of science.

view more: next ›