[-] Eyekaytee@aussie.zone 26 points 1 day ago

I was thinking that like how everyone is skipping a 5th generation fighter jet and going straight to 6th generation, why not just skip over world war v3 and go straight to world war v4 with all the fighting done by robots and drones?

108
[-] Eyekaytee@aussie.zone 2 points 1 day ago

🫡 belated happy cake day to you too :) I just noticed we both signed up within 2 days of each other

[-] Eyekaytee@aussie.zone 8 points 2 days ago

it’s never recalling information, it’s always hallucinating

With the exception of information you have put within the context window

https://generativeai.pub/what-is-the-context-window-ais-memory-limit-ce8a966f954a

[-] Eyekaytee@aussie.zone 51 points 2 days ago

Why did you short url a wikipedia link with google?

[-] Eyekaytee@aussie.zone 3 points 2 days ago

Bit of a bad lead story, if she's only been in the US for a few years and was only on a student visa she would have had to have returned to Bangladesh pretty quickly anyway?

[-] Eyekaytee@aussie.zone -2 points 3 days ago

On 13 April, he said, Hamas gunmen tried to force their way into the house of an elderly man, Jamal al-Maznan.

"They wanted to launch rockets and pipes [a derogatory term used for some of Hamas' home-made projectiles] from inside his house," the eyewitness told us.

"But he refused."

The incident soon escalated, with relatives and neighbours all coming to al-Maznan's defence. The gunmen opened fire, injuring several people, but eventually were driven out.

https://www.bbc.com/news/articles/c175z14r8pro

Tell Hamas to stop fighting out of homes and hospitals?

[-] Eyekaytee@aussie.zone 1 points 4 days ago

ah cheers, is there a way to automate the pixel placement?

[-] Eyekaytee@aussie.zone 2 points 4 days ago

LOL that flag is massive! Was that really what happened last year?

[-] Eyekaytee@aussie.zone 1 points 4 days ago

or... rewatch 🔴 Friday NIGHT Plane Spotting @ Brisbane Airport BNE w/ James & Matty + ATC ✈️🌇🌕

https://www.youtube.com/watch?v=taeYSxWlZRw

Way less stressful and doesn't involve Americans 😄

18
How to fight Israel (aussie.zone)
35

It looks like AI has followed Crypto chip wise in going CPU > GPU > ASIC

GPUs, while dominant in training large models, are often too power-hungry and costly for efficient inference at scale. This is opening new opportunities for specialized inference hardware, a market where startups like Untether AI were early pioneers.

In April, then-CEO Chris Walker had highlighted rising demand for Untether’s chips as enterprises sought alternatives to high-power GPUs. “There’s a strong appetite for processors that don’t consume as much energy as Nvidia’s energy-hungry GPUs that are pushing racks to 120 kilowatts,” Walker told CRN. Walker left Untether AI in May.

Hopefully the training part of AI goes to ASIC's to reduce costs and energy use but GPU's continue to improve inference and increase VRAM sizes to the point that AI requires nothing special to run it locally

10
submitted 2 weeks ago* (last edited 2 weeks ago) by Eyekaytee@aussie.zone to c/localllama@sh.itjust.works

Sorry team flipped the URL’s around to prevent overflow from lemmy.world users

https://fly.io/blog/youre-all-nuts/

37
9
submitted 3 weeks ago* (last edited 3 weeks ago) by Eyekaytee@aussie.zone to c/australia@aussie.zone

Could the fuel powering F1 next season also run your car? | The Business | ABC NEWS

https://www.youtube.com/watch?v=0yVBvsbOZ-0

20
20
submitted 3 weeks ago* (last edited 3 weeks ago) by Eyekaytee@aussie.zone to c/localllama@sh.itjust.works

Devstral is an agentic LLM for software engineering tasks built under a collaboration between Mistral AI and All Hands AI 🙌. Devstral excels at using tools to explore codebases, editing multiple files and power software engineering agents. The model achieves remarkable performance on SWE-bench which positionates it as the #1 open source model on this benchmark.

It is finetuned from Mistral-Small-3.1, therefore it has a long context window of up to 128k tokens. As a coding agent, Devstral is text-only and before fine-tuning from Mistral-Small-3.1 the vision encoder was removed.

Learn more about Devstral in our blog post. Key Features:

Agentic coding: Devstral is designed to excel at agentic coding tasks, making it a great choice for software engineering agents.

lightweight: with its compact size of just 24 billion parameters, Devstral is light enough to run on a single RTX 4090 or a Mac with 32GB RAM, making it an appropriate model for local deployment and on-device use.

Apache 2.0 License: Open license allowing usage and modification for both commercial and non-commercial purposes.

Context Window: A 128k context window.

Tokenizer: Utilizes a Tekken tokenizer with a 131k vocabulary size.

Learn more about Devstral in our blog post.

https://mistral.ai/news/devstral

11
submitted 4 weeks ago* (last edited 4 weeks ago) by Eyekaytee@aussie.zone to c/localllama@sh.itjust.works

Gemma 3n includes the following key features:

Audio input: Process sound data for speech recognition, translation, and audio data analysis.

Visual and text input: Multimodal capabilities let you handle vision, sound, and text to help you understand and analyze the world around you.

PLE caching: Per-Layer Embedding (PLE) parameters contained in these models can be cached to fast, local storage to reduce model memory run costs. Learn more

MatFormer architecture: Matryoshka Transformer architecture allows for selective activation of the models parameters per request to reduce compute cost and response times. Learn more

Conditional parameter loading: Bypass loading of vision and audio parameters in the model to reduce the total number of loaded parameters and save memory resources. Learn more

Wide language support: Wide linguistic capabilities, trained in over 140 languages. 32K token context: Substantial input context for analyzing data and handling processing tasks.

184

from 10b0t0mized: I miss the days when I had to go through a humiliation ritual before getting my questions answered.

Now days you can just ask your questions from an infinitely patient entity, AI is really terrible.

19
submitted 1 month ago* (last edited 1 month ago) by Eyekaytee@aussie.zone to c/localllama@sh.itjust.works

This fork introduces a Radio Station feature where AI generates continuous radio music. The process involves two key components:

LLM: Generates the lyrics for the songs. ACE: Composes the music for the generated lyrics.

Due to the limitations of slower PCs, the demo video includes noticeable gaps (approximately 4 minutes) between the generated songs.

If your computer struggles to stream songs continuously, increasing the buffer size will result in a longer initial delay but fewer gaps between songs (until the buffer is depleted again).

By default the app attempts to load the model file gemma-3-12b-it-abliterated.q4_k_m.gguf from the same directory. However, you can also use alternative LLMs. Note that the quality of generated lyrics will vary depending on the LLM's capabilities.

172
submitted 1 month ago by Eyekaytee@aussie.zone to c/world@lemmy.world
view more: next ›

Eyekaytee

joined 2 years ago