smorty/maria [she/her]
- 30 Posts
- 121 Comments
smorty/maria [she/her]@lemmy.blahaj.zoneto Free Open-Source Artificial Intelligence@lemmy.world•where have a group for talk about aiEnglish2·3 months agowell what did it translate to?
smorty/maria [she/her]@lemmy.blahaj.zoneOPto AI@lemmy.ml•How are companies SO bad at seeeling AI to us? (And how to fix it)1·4 months agoyes!! u get it!! that’s exactly the way we must handle AI, jus like we do with othr tech!!!
also— i do really enjoy the idea you have there with the AI assisted new site, but if you plan on making that a product while also open-accessing the prompts, people will copy that within a day, so it’d make most sense to just open-source the entire platform.
the pen knos what im gona write!!?!?!?! :o
smorty/maria [she/her]@lemmy.blahaj.zoneto Free Open-Source Artificial Intelligence@lemmy.world•where have a group for talk about aiEnglish21·4 months agosadly yea… i think its discord onli mostly ;(
i onli use matrix so dis is a no go for me <3
wud be totally nice n comfy if peeps were to create matrix space or so <3 <3 <3
smorty/maria [she/her]@lemmy.blahaj.zoneOPto Selfhosted@lemmy.world•[Help] IPv4 address reaches website, but domain doesn't... (wrong community?)English2·4 months agobut i wanna have a website others can access too. I tried using VPNs for cool stuff already (like controlling my lil raspberry robot from work with my phone) but I want this website to be available to all the people…
should i just bite the bullet and rent some hosting service? Or is there still hope for me putting “setup home website server” on my resume?
:o
will u become like an ancient wizard? that’d be SO cool
:o
ur here!!! <3
😳🥹😖🥰
omygosh! hiiiiiiiii!!!
how do you do? srri for my late respons - i had family things to pursue…
smorty/maria [she/her]@lemmy.blahaj.zoneto Free Open-Source Artificial Intelligence@lemmy.world•How to run LLaMA (and other LLMs) on Android.English2·5 months agoYea, I did this.
Can’t WAIT for vulkan support. Imagine the speed! It could be so much faster. Currently it just slows down the model to like - 2 tokens per second
I think she stepped on a crack.
Also HIII MAXXX!!! <3
smorty/maria [she/her]@lemmy.blahaj.zoneto Free Open-Source Artificial Intelligence@lemmy.world•'A virtual DPU within a GPU': Could clever hardware hack be behind DeepSeek's groundbreaking AI efficiency?English5·5 months agoHmmm this sounds like a shitpost…
a virtual DPU on a GPU
sounds likedownload some RAM on this website
to me.Like - aren’t NPUs just more tensor cores? More matrix multiplying machines? I can’t just simulate that and expect it to be faster…
smorty/maria [she/her]@lemmy.blahaj.zoneto AI@lemmy.ml•Cohere Drops Command-R 35B 08-2024 Update, Just About a Perfect Local LLM for 24GB GPUs.1·6 months agoi totally agree… with everything. 6GB really is smol and, cuz imma crazy person, i currently try and optimize everything for llama3.2 3B Q4 model so people with even less GB VRAM can use it. i really like the idea of people just having some smollm laying around on their pc and devs being able to use it.
i really should probably opt for APIs, you’re right. the only API I ever used was Cohere, cuz yea their CR+ model is real nice. but i still wanna use smol models for a smol price if any. imma have a look at the APIs you listed. Never heard of Kobold Horde and Samba so i’ll have a look at those… or i go for the lazy route and chose depseek cuz it’s apparently unreasonably cheap for SOTA perf. so eh…
also yes! Lemmy really does seem anti AI, and i’m fine with that. i just say
yeah companies use it in obviously dum ways but the tech is super interesting
which is a reasonable argument i think.so yes, local llm go! i wanna get that new top amd gpu once that gets announced. so i’ll be able to run those spicy 32B models. for now i’ll just stick with 8B and 3B cuz they work quick and kinda do what i want.
smorty/maria [she/her]@lemmy.blahaj.zoneto AI@lemmy.ml•Cohere Drops Command-R 35B 08-2024 Update, Just About a Perfect Local LLM for 24GB GPUs.1·6 months agocould you define “right settings”?
I assuma Q4 and some context window Q8 aswell. Anything lese to tweak?
I just have a smol gtx1060 6gb VRAM, so i probably can’t fit it on mine and imma have to use cpu partly. but maybe other readers here can!(I’m just a silly ollama user, not knowing anything more complex than the tokenizer… so yea, maybe put a lil infodump in here to make us all smarter please <3 )
EDIT: brucethemoose probably refered to this model named “Medius”. there is no 14B in the name.
smorty/maria [she/her]@lemmy.blahaj.zoneto AI@lemmy.ml•Cohere Drops Command-R 35B 08-2024 Update, Just About a Perfect Local LLM for 24GB GPUs.1·6 months agoi luv command R+ so very much and now i wanna try that smoler model but also the newly released r7b model was really not the best so i got sad…
smorty/maria [she/her]@lemmy.blahaj.zoneto AI@lemmy.ml•Man learns he’s being dumped via “dystopian” AI summary of texts1·6 months agoi like how when ai summarizes a sad dramatic thing, people go :o like it’s something special and not exactly what it was trained to do.
smorty/maria [she/her]@lemmy.blahaj.zoneOPto Free Open-Source Artificial Intelligence@lemmy.world•Before you buy and courses: Read this free prompting guide! (no login required)English2·7 months agoooh, leaked prompts? which ones are you talking about?
smorty/maria [she/her]@lemmy.blahaj.zoneOPto Free Open-Source Artificial Intelligence@lemmy.world•Before you buy and courses: Read this free prompting guide! (no login required)English2·7 months agoYou are completely right and it is mostly about trial and error. I’d assume these courses mainlyl teach things you can do with the big bois, those being by the obvious big evil AI companies. It’s very much an overblown topic and companies pretend it’s actually a fancy thing to learn and be good at.
The linked guide just explains the basic concepts of few shot prompting, CoT and RAG and stuff. Even these terms I feel, make the topic more complicated than it is. Could literally be summarized to
- Use examples of what you want
- Use near-zero temperature for almost everything
- For complex tasks, tell it to provide its internal thought proccess before providing the answer (or just use the QwQ model)
- maybe SCREAM AT THE LLM IN ALLCAPS if something is really important
smorty/maria [she/her]@lemmy.blahaj.zoneOPtoAsklemmy@lemmy.ml•Is Lemmy your "main social media app"? If not, which one is it?0·8 months agoThat’s the best type of social (media)
smorty/maria [she/her]@lemmy.blahaj.zoneOPtoAsklemmy@lemmy.ml•Is Lemmy your "main social media app"? If not, which one is it?0·8 months agoOk that is fair. I commented on someone elses comment here that they should try out PeerTube, as that is a decentralized option with all the benefits of the fediverse.
smorty/maria [she/her]@lemmy.blahaj.zoneOPto Linux@lemmy.ml•We already have Linux at home!6·8 months agoOh they totally will try. Microsoft is dum enough to try it, just like they are dum enough to try to train massive LLMs, and damn, they not be showin’ successes til now :)
no… turns out MCP servers are run and downloaded using mainly npx and stuff. I would LOVE it if it were just a rest api. But nope… because stuff like file access is also possible! So that data obv can’t leave ur PC…