1
Linus Posting (lemmy.world)
submitted 1 day ago* (last edited 12 hours ago) by Smokeydope@lemmy.world to c/linuxmemes@lemmy.world

Bonus Meme:

[-] Smokeydope@lemmy.world 1 points 6 days ago* (last edited 6 days ago)

You can try to use VSCode + roo to intelligently chunk it autonomously. Get a API key from your llm provider of choice, put your data into a text file, Edit the roo agent personalites thats set to coding by default. Instead add and select a custom summarizer persona, for roo to use then tell it to summarize the text file.

[-] Smokeydope@lemmy.world 1 points 6 days ago* (last edited 6 days ago)

As other commenter said your workflow requires more than what LLMs are currently capable of.

Summarization capability in LLMs is an equation of LLMs capacity for coherence over long conversational scaling operated on by the LLMs ability to navigate and distill internal structural mappings of conceptual & contextual archetype patterns as discrete objects across a continuous ambiguity sheaf.

This technical jargon that boils down to the idea that an llms summarization capability depends on its parameter size and enough vram for context lengths. Higher parameter and less quantized models maintaining more coherence over long conversations/datasets.

While enterprise llms are able to get up to 128k tokens while maintaining some level of coherence, the local models of medium quantization can handle 16-32k reliably. Theoretically 70b could maybe handle around 64k tokens but even thats stretching it.

Then comes the problem of transformer attention. You can't just put a whole books worth of text into an LLMs input and expect it to inspect any part in real detail. For best results you have to chunk it section by section, chapter by chapter.

So local llms may not be what you're looking for. If you are willing to go enterprise then Claude sonnet and deepseek R1 might be good especially if you set up a API interface.

22
submitted 3 weeks ago* (last edited 3 weeks ago) by Smokeydope@lemmy.world to c/localllama@sh.itjust.works
11

I think i've discovered something important in the field I dabble in as a advanced hobbyist. Like this was a breakthrough and perspective shift enough for me to stay awake all night into the morning until I had to go to sleep testing it works and boilerplating the abstract paper. I constructed a theoretical framework, practical implementation, and statistically analyzed experimental results across numerous test cases. I then put my findings into as good a technical paper as I could write up. I did as much research as I could to make sure nobody else had written about this before.

At this point though I don't really know how to proceed. Im an outsider systems engineer not an academic, and arXiv requires you be endorsed/recognized as a member of the scientific community with like a college email or written recommendation by someone already known. Then whenever I look at the papers on arxiv they always look a very specific way I cant get with libreoffice writer. Theres apparently a whole bunch of rules on formatting and font and style and this and that. Its overwhelming and kind of scary.

So. What do i do here? I have something I think is important enough to get off my ass and get in touch with a local college to maybe get a recommendation. I'd like to have my name in the community and contribute.

11
submitted 4 weeks ago* (last edited 4 weeks ago) by Smokeydope@lemmy.world to c/askscience@lemmy.world

I think i've discovered something important in the field I dabble in as a advanced hobbyist. Like this was a breakthrough and perspective shift enough for me to stay awake all night into the morning until I had to go to sleep testing it works and boilerplating the abstract paper. I constructed a theoretical framework, practical implementation, and statistically analyzed experimental results across numerous test cases. I then put my findings into as good a technical paper as I could write up. I did as much research as I could to make sure nobody else had written about this before.

At this point though I don't really know how to proceed. Im an outsider systems engineer not an academic, and arXiv requires you be endorsed/recognized as a member of the scientific community with like a college email or written recommendation by someone already known. Then whenever I look at the papers on arxiv they always look a very specific way I cant get with libreoffice writer. Theres apparently a whole bunch of rules on formatting and font and style and this and that. Its overwhelming and kind of scary.

So. What do i do here? I have something I think is important enough to get off my ass and get in touch with a local college to maybe get a recommendation. I'd like to have my name in the community and contribute.

1
submitted 1 month ago* (last edited 1 month ago) by Smokeydope@lemmy.world to c/buildapc@lemmy.world

Please help! I'm trying to remove this metal plate to get to my GPU for cleaning. From what I saw it had one star pattern bolt at the end of a slope connecting to the chassis, removing it loosened it a little but I don't know how to proceed. Dont want to damage anything.

Edit: I got it figured out! Had to pull a metal tab on the back as I jostled/slid the plate. Thank you!

30
submitted 1 month ago* (last edited 1 month ago) by Smokeydope@lemmy.world to c/localllama@sh.itjust.works

Recently I've been experimenting with Claude and feeling the burn on the premium API usage. I wanted to know how much cheaper my local llm was in terms of cost-per-token output.

Claude Sonnet is a good reference with 15$ per 1 million tokens out, so I wanted to know comparatively how many tokens 15$ worth electricity powering my rig would generate.

(These calculations are just simple raw token generation by the way, in real world theres cost in initial hardware, ongoing maintenance as parts fail, and human time to setup thats much harder to factor into the equation)

So how does one even calculate such a thing? Well, you need to know

  1. how many watts your inference rig consumes at load
  2. how many tokens on average it can generate per second while inferencing (with context relatively filled up, we want conservative estimates)
  3. cost of electric you pay on the utility bill in kilowatts-per-hour

Once you have those constants you can extrapolate how many kilowatt-hours worth of runtime 15$ in electric buys then figure out the total amount of tokens you would expect to generate over that time given the TPS.

The numbers shown in the screenshot are for a fully loaded into vram model on the ol' 1070ti 8gb. But even with partially offloaded numbers for 22-32b models at 1-3tps its still a better deal overall.

I plan to offer the calculator as a tool on my site and release it under a permissive license like gpl if anyone is interested.

49
submitted 1 month ago by Smokeydope@lemmy.world to c/pics@lemmy.world
28
Homelab upgrade WIP (lemmy.world)
submitted 1 month ago* (last edited 1 month ago) by Smokeydope@lemmy.world to c/localllama@sh.itjust.works

Theres a lot more to this stuff than I thought there would be when starting out. I spent the day familiarizing with how to take apart my pc and swap gpus .Trying to piece everything together.

Apparently in order for PC to startup right it needs a graphical driver. I thought the existance of a HDMI port on the motherboard implied the existance of onboard graphics but apparently only special CPUs have that capability. My ryzen 5 2600 doesnt. The p100 Tesla does not have graphical display capabilities. So ive hit a snag where the PC isnt starting up due to not finding a graphical interface output.

I'm going to try to run multiple GPU cards together on pcie. Hope I can mix amd Rx 580 and nvidia tesla on same board fingers crossed please work.

My motherboard thankfully supports 4x4x4x4 pcie x16 bifurcation which isa very lucky break I didnt know going into this 🙏

Strangely other configs for splitting 16x lanes like 8x8 or 8x4x4 arent in my bios for some reason? So I'm planning to get a 4x bifurcstion board and plug both cards in and hope that the amd one is recognized!

According to one source The performance loss for using 4x lanes for GPUs doing the compute i m doing is 10-15 % surprisingly tolerable actually.

I never really had to think about how pcie lanes work or how to allocate them properly before.

For now I'm using two power supplies one built into the desktop and the new 850e corsair psu. I choose this one as it should work with 2-3 GPUs while being in my price range.

Also the new 12v-2x6 port supports like 600w enough for the tesla and comes with a dual pcie split which was required for the power cable adapter for Tesla. so it all worked out nicely for a clean wire solution.

Sadly I fucked up a little. The pcie release press plastic thing on the motherboard was brittle and I fat thumbed it too hard while having problems removing the GPU initially so it snapped off. I dont know if that's something fixable. It doesnt seem to affect the security of the connection too bad fortunately. I intend to grt a pcie riser extensions cable so there won't be much force on the now slightly loosened pcieconnection. Ill have the gpu and bifurcation board layed out nicely on the homelab table while testing, get them mounted somewhere nicely once I get it all working.

I need to figure out a external GPU mount system. I see people use server racks or nut and bolt meta chassis. I could get a thin plate of copper the size of the desktops glass window as a base/heatsink?

2
Homelab upgrade WIP (lemmy.world)
submitted 1 month ago* (last edited 1 month ago) by Smokeydope@lemmy.world to c/buildapc@lemmy.world

Theres a lot more to this stuff than I thought there would be when starting out. I spent the day familiarizing with how to take apart my pc and swap gpus .Trying to piece everything together.

Apparently in order for PC to startup right it needs a graphical driver. I thought the existance of a HDMI port on the motherboard implied the existance of onboard graphics but apparently only special CPUs have that capability. My ryzen 5 2600 doesnt. The p100 Tesla does not have graphical display capabilities. So ive hit a snag where the PC isnt starting up due to not finding a graphical interface output.

I'm going to try to run multiple GPU cards together on pcie. Hope I can mix amd Rx 580 and nvidia tesla on same board fingers crossed please work.

My motherboard thankfully supports 4x4x4x4 pcie x16 bifurcation which isa very lucky break I didnt know going into this 🙏

Strangely other configs for splitting 16x lanes like 8x8 or 8x4x4 arent in my bios for some reason? So I'm planning to get a 4x bifurcstion board and plug both cards in and hope that the amd one is recognized!

According to one source The performance loss for using 4x lanes for GPUs doing the compute i m doing is 10-15 % surprisingly tolerable actually.

I never really had to think about how pcie lanes work or how to allocate them properly before.

For now I'm using two power supplies one built into the desktop and the new 850e corsair psu. I choose this one as it should work with 2-3 GPUs while being in my price range.

Also the new 12v-2x6 port supports like 600w enough for the tesla and comes with a dual pcie split which was required for the power cable adapter for Tesla. so it all worked out nicely for a clean wire solution.

Sadly I fucked up a little. The pcie release press plastic thing on the motherboard was brittle and I fat thumbed it too hard while having problems removing the GPU initially so it snapped off. I dont know if that's something fixable. It doesnt seem to affect the security of the connection too bad fortunately. I intend to grt a pcie riser extensions cable so there won't be much force on the now slightly loosened pcieconnection. Ill have the gpu and bifurcation board layed out nicely on the homelab table while testing, get them mounted somewhere nicely once I get it all working.

I need to figure out a external GPU mount system. I see people use server racks or nut and bolt meta chassis. I could get a thin plate of copper the size of the desktops glass window as a base/heatsink?

24
submitted 1 month ago* (last edited 1 month ago) by Smokeydope@lemmy.world to c/asklemmy@lemmy.world

I now do some work with computers that involves making graphics cards do computational work on a headless server. The computational work it does has nothing to do with graphics.

The name is more for consumers based off the most common use for graphics cards and why they were first made in the 90s but now they're used for all sorts of computational workloads. So what are some more fitting names for the part?

I now think of them as 'computation engines' analagous to a old car engine. Its where the computational horsepower is really generated. But how would ram make sense in this analogy?

34
submitted 2 months ago* (last edited 2 months ago) by Smokeydope@lemmy.world to c/localllama@sh.itjust.works

It seems mistral finally released their own version of a small 3.1 2503 with CoT reasoning pattern embedding. Before this the best CoT finetune of Small was DeepHermes with deepseeks r1 distill patterns. According to the technical report, mistral baked their own reasoning patterns for this one so its not just another deepseek distill finetune.

HuggingFace

Blog

Magistral technical research academic paper

7
submitted 2 months ago* (last edited 2 months ago) by Smokeydope@lemmy.world to c/selfhosted@lemmy.world

Setting up a personal site on local hardware has been on my bucket list for along time. I finally bit he bullet and got a basic website running with apache on a Ubuntu based linux distro. I bought a domain name, linked it up to my l ip got SSL via lets encrypt for https and added some header rules until security headers and Mozilla observatory gave it a perfect score.

Am I basically in the clear? What more do I need to do to protect my site and local network? I'm so scared of hackers and shit I do not want to be an easy target.

I would like to make a page about the hardware its running on since I intend to have it be entirely ran off solar power like solar.lowtechmagazine and wanted to share technical specifics. But I heard somewhere that revealing the internal state of your server is a bad idea since it can make exploits easier to find. Am I being stupid for wanting to share details like computer model and software running it?

[-] Smokeydope@lemmy.world 70 points 5 months ago* (last edited 5 months ago)
[-] Smokeydope@lemmy.world 71 points 1 year ago* (last edited 1 year ago)

The day adblocks/yt-dlp finally loose to google forever is the day I kiss youtube bye-bye. No youtube premium, no 2 minute long unskippable commerical breaks. I am strong enough to break the addiction and go back to the before-fore times when we bashed rocks together and stacked CDs in towers.

Peertube, odysee, bittorrenting, IPTV. Ill throw my favorite content creators a buck or two on patreon to watch their stuff there if needed. We've got options, its a matter of how hot you need to boil the water before the lowest common denominator consumer finally has enough.

[-] Smokeydope@lemmy.world 82 points 1 year ago

Heres the template if anyone wants it

[-] Smokeydope@lemmy.world 123 points 2 years ago* (last edited 2 years ago)

The pocket of air that was where you teleported now get displaced at a very decent fraction of the speed of light while the pocket of space you once ocupied becomes a almost pure vaccum. the air moves so fast it creates a sonic boom that ruptures the ear drums. Then, a few atoms of air collide together with such incredible force the atoms split and causes a small grade nuclear explosion.

[-] Smokeydope@lemmy.world 87 points 2 years ago* (last edited 2 years ago)

My elderly parents in their 60s use linux mint daily and have never had an issue with it (admittedly I did have to set it up for them still). I just set up the desktop shortcuts for them to their websites and turn on automatic updates. The hardest part isn't using an alternative OS like mint or pop, its getting an average person to figure out how to install it. Getting into your BIOS to boot into the installation drive, re-partitioning your harddrive to free up space for dual booting or nuking windows off all together, those are the hardest parts for any first timers IMO. After youve done it a dozen times its no problemo but the first time is nerve racking at least it was to me.

[-] Smokeydope@lemmy.world 227 points 2 years ago* (last edited 2 years ago)

This is a copy/pasted message I wrote up on another thread. As long as there are people in the comments shilling kagi, I will shill my prefered engines. At least my suggestions will bring awareness to free as in freedom projects. I hope to god people paying 10$/month just to not get datacucked by search engines will also learn something and save their money.

SearX/SearXNG is a free and open source, highly customizable, and self-hostable meta search engine. SearX instances act as a middle man, they query other search engines for you, stripping all their spyware ad crap and never having your connection touch their servers. Of course you have to trust the SearX instance host with your query information, but again if you are that paranoid just self host.

I personally trust some foss loving sysadmin that host social services for free out of alturism, who also accepts hosting donations, whos server is located on the other side of the planet, with my query info over Google/Alphabet any day.

Its nice to be able to email and have a human conversation with your search engine provider thats just a knowlegable every day joe who genuinely believes in the project and freely dedicates their resources to it. Consider sending some cash their way to help with upkeep if you like the services they provide, they will probably appreciate and make use of that 10$ better than kagi.

Heres a list of all public searx instances, I personally prefer to use paulgo.io All SearX instances are configured different to index different engines. If one doesn't seem to give good results try a few others.

Did I mention it has bangs like duckduckgo? If you really need google like for maps and buisness info just use !!g in the query

search.marginalia.nu is a completely novel search engine written and hosted by one dude that aims to prioritize indexing lighter websites little to no javascript as these tend to be personal websites and homepages that have poor SEO and the big search engines won't index well. If you remember the internet of the early 2000s and want a nostalgia trip this ones for you. Its also open source and self-hostable

Finally, YaCy is another completely novel search engine that uses peer-to-peer technology to power a big webcrawler which prioritizes indexes based off user queries and feedback. Everyone can download yacy and devote a bit of their computing power to both run their own local instance and help out a collective search engine. Companies can also download yacy and use it to index their private intranets.

They have a public instance available through a web portal. To be upfront, YaCy is not a great search engine for what most people usually want, which is quick and relevant information within the first few clicks. But, it is an interesting use of technology and what a true honest-to-god community-operated search engine looks like untainted by SEO scores or corporate money-making shenanigans.

I hope this has been informative to those who believe theres only a few options to pick from, I know these options are so unknown to most people.

[-] Smokeydope@lemmy.world 87 points 2 years ago* (last edited 2 years ago)

Sometimes I think I made the right decision to just get a huge harddrive and download all my favorite entertainment in drm free format. Movies, music, games, books. I saw this coming a mile away a decade ago. The only thing that will really hurt me is if/when Steam inevitably goes full corporate cucks and starts going hard on the DRM locking down my library.

[-] Smokeydope@lemmy.world 82 points 2 years ago* (last edited 2 years ago)

Now please unremove the shroom community as next priority. Empowering open minded people with the option and knowledge to heal themselves through the use of psychadelics (and other kinds of mushrooms that can potentially help fight diseases such as cancer) that they can grow themselves without big pharma and giving them a community to share their advice+experiences is the right thing to do.

[-] Smokeydope@lemmy.world 92 points 2 years ago

Wow that was actually a cute story, not sure how legit but nice feels none the less

view more: next ›

Smokeydope

joined 2 years ago
MODERATOR OF