Items tagged with: ML

Search

Items tagged with: ML


😲 DeepSeek-V3-4bit runs at >20 tokens per second and <200W using MLX on an M3 Ultra with 512GB. This might be the best and most user-friendly way to run DeepSeek-V3 on consumer hardware, possibly the most affordable too. You can finally run a GPT-4o level model locally, with possibly even better quality. #LLM #AI #ML #DeepSeek #OpenAI #GPT #OpenWeight #OpenSource venturebeat.com/ai/deepseek-v3…


AirBnB migrated 3,5k React component test files from Enzyme to RTL using LLMs and automation in just six weeks. The effort was originally estimated to take 1.5 years of manual engineering time to do by hand. #LLM #AI #ML #Automation
medium.com/airbnb-engineering/…


😲 It’s only March, but there’s already been incredible progress in open-weight LLMs this year. Here are my top 5 local LLM recommendations for anyone with 24GB of VRAM to try: Phi-4-14B for speed, Mistral-Small-24B for RAG, Gemma-3-27B for general use, Qwen2.5-Coder-32B for coding, QWQ-32B for reasoning. #LLM #ML #AI
#AI #ML #llm


Now there's MLX-Audio! They have Kokoro and CSM-1B for now. It's the same dev team for MLX-VLM, so I have high hope! #TTS #AI #ML github.com/Blaizzy/mlx-audio
#AI #ML #tts


Here's another fun project! AIChat lets you generate conversations between two LLMs on any topic using OpenAI API. You can mix and match models from Ollama, MLX, Claude, OpenAI, Google AI Studio, etc. It uses Kokoro-ONNX for TTS. Check out an example + the GitHub repo:
youtube.com/watch?v=FgSZLZnYlA…
github.com/chigkim/AIChat
#LLM #TTS #AI #ML
#AI #ML #llm #tts


Gemma3 is out with 1b, 4b, 12b, 27b. It also accepts images except 1b model. It's on Ollama, but ollama run gemma3 will use small 4b model. Be sure to specify size if you want to use bigger model. For example, ollama run gemma3:27b #LLM #ML #AI ollama.com/library/gemma3/tags
#AI #ML #llm


Very interesting read from HuggingFace co-founder: For big scientific breakthroughs, "we don't need an A+ student who can answer every question with general knowledge. We need a B student who sees and questions what everyone else missed." Current AI is great at answering known questions, but struggles to challenge assumptions and push boundaries. True innovation comes from questioning and redefining knowledge—something AI isn’t quite ready for yet. #AI #LLM #ML thomwolf.io/blog/scientific-ai…
#AI #ML #llm


#AI #ML #llm


3/3 Gemini Flash 2.0 Thinking: Humans are not just social animals, but profoundly self-domesticated apes. And this self-domestication is the key to understanding our unique cognitive and social abilities.
#LLM #AI #ML
#AI #ML #llm


2/3 Deepseek R1: Humans instinctively convert selfish desires into cooperative systems by collectively pretending abstract rules, money, laws, rights, are real. These shared hallucinations act as games, where competition is secretly redirected to benefit the group, turning conflict into society's fuel.
#LLM #AI #ML
#AI #ML #llm


1/3 Three Responses from 3 top reasoning models to Lex Fridman's question, What's one truly novel insight about humans?
O1-pro: Humans are unique among known species in that they simultaneously rewrite two layers of reality, the external world and their own private mental landscapes, and then merge these two rewritten layers into a continuous personal narrative that feels objectively true.
#LLM #AI #ML
#AI #ML #llm


#AI #ML


#AI #ML #llm



#AI #ML #llm



F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching! The quality is pretty impressive for open source, and it even supports mps for Mac! I was able to get it going on my Mac with no problem. #TTS #ML #AI
github.com/SWivid/F5-TTS
@ZBennoui





😲 OMG, Audio Overview feature on NotebookLM is wild! It basically creates a podcast with two AI generated voices based on the source documents you upload. Definitely try if you haven't yet. #LLM #ML #AI blog.google/technology/ai/note…
#AI #ML #llm


#AI #ML #llm


😲 Kyle Kabasares, a Physics PhD graduate working at NASA's Ames Research Center, gave the methods section of his research paper to ChatGPT O1 Preview and asked it to generate the code based on the description. After just six prompts, it produced a working version of the code that took him a year to develop during his PhD. #ChatGPT #LLM #ML #AI youtube.com/watch?v=M9YOO7N5jF…


After a long period of inactivity for vision language models, llama.cpp merged the support for MiniCPM-V-2.5. Hopefully the support for 2.6 is also on the way soon. #LLM #Multimodal #AI #ML
huggingface.co/openbmb/MiniCPM…
huggingface.co/openbmb/MiniCPM…
github.com/ggerganov/llama.cpp…


Exciting news on open-source neural voices!
Our first experiment is complete with fantastic results! Check out the audio sample attached to this post.
For this month, @pneumasolutions provided GPU resources for training. I really appreciate their contribution.
This is just the beginning. To keep training going, I'm still accepting donations. Any amount helps.
I'm happy to receive your donations via PayPal:
paypal.me/geotts
Please mention mush42/tts in the notes.
#SpeechSynthesis #AI #ML


#AI #ML


#AI #math #ML #llm


#AI #ML #llm


Llama3-405b base model is leaked on 4chan as Miqu-2. Miqu-1 was leaked Mistral 70b model which was confirmed by Mistral CEO. The download size is 764GB, and it was briefly on Huggingface but taken down. The torrent is still working apparently. #LLM #AI #ML reddit.com/r/LocalLLaMA/commen…
#AI #ML #llm


If the rumors are true, this week could be another exciting week for opensource LLMS! Meta may release Llama-3-405b on this Tuesday. Also there could be updates to 8b and 70b models distilled from 405B. Joe Spisak, a product director at Meta says they were initially going to call Llama 3 8b and 70b a prerelease or preview because these models didn't have all the things they planned to release.
Sources:
theinformation.com/briefings/m…
x.com/AlpinDale/status/1814814…
youtu.be/r3DC_gjFCSA?feature=s…
#LLM #AI #ML
#AI #ML #llm


Time to buy the Mac Studio with 192GB memory! lol Rumor: Meta plans to release the largest Llama 3 model with 405 billion parameters on July 23, according to a Meta employee. "it will be able to understand and generate images and text." #LLM #AI #ML theinformation.com/briefings/m…
#AI #ML #llm



#AI #ML #llm


Editted to fix link. Please boost for reach if this kind of stuff interests you. Will post more on this later.

Once upon a time, there was a cool emulator frontend called Retroarch. This emulator wasn't accessible until I and a few other gamers went to them and asked about adding accessibility. An amazing person known as BarryR made it happen. Now, if you turn on accessibility mode in settings, or pass the "--accessibility" (or something like that) flag on the command line, you get spoken menus, including the emulator's pause menu, good for saving states and such. Then, using PIL and other image processing Python utilities, running a server and hooking into Retroarch, the script allowed players to move around the map, battle, talk to NPC's, ETC. The only problem was, no one wanted to test it. The blind gaming community pretty much spoke, saying that we want new games. We want cool new, easy accessibility. So that's what we have no, follow the beacon or get sighted help in the case of diablo and such. It's sad, but meh. It's what we wanted I guess. No Zelda for us. So, this is about as far as he got:

To expand on what devinprater was saying: I am working on an accessibility pack/service for Final Fantasy 1 for the NES (this was what was shown in the latest RetroArch update). The idea is similar to how Pokemon Crystal access works, but it's using the RetroArch AI Service interface to do so.
Right now, the FF1 access service is mostly done, but I need more testers to try it out and give me feedback on how it's working. Right now, you can get up to the point where you get the ship, but there's no code to deal with how the ship moves, so that still needs to be done. Likewise with the airship later on.
The service works the latest version of RetroArch, on linux and mac, but not windows. This is due to how nvda reads out the text and until the next major update to nvda (which will have a feature to fix this), it'll have to wait. If you have those, I (or maybe devinprater) can help you set it up on mac/linux to test out. The package itself is available at: (new link cause old one broke yesterday): dropbox.com/scl/fi/ggffl769fx6…
#accessibility #finalFantasy #RetroArch #blind #emulator #emulation #Python #ai #ML #MachineLearning


Start saving money for that M4 Ultra with 500GB! Maybe this could be the first open source that could surpass GPT-4! AIatMeta: "Llama 3 8B & 70B models are just the beginning of what we’re working to release for Llama 3. Our largest models currently in the works are 400B+ parameters and while they’re still in active development, we’re excited about how this work is trending." #LLM #AI #ML twitter.com/AIatMeta/status/17…
#AI #ML #llm


Apparently Meta is planning to release two small varients of Llama-3 next week "as a precursor to the launch of the biggest version of Llama 3, expected this summer." Command-r-plus, mixtral 8x22b, Google CodeGemma... All of sudden companies are releasing LLMS like crazy! Where's Apple? Maybe In WWDC 2024? lol #LLM #AI #ML theinformation.com/articles/me…
#AI #ML #llm


#AI #ML #llm


Claude 3 can summarize up to about 150,00 words, (a length similar to Harry Potter and the Deathly Hallows.) also It outperformed GPT-4 and Gemini Ultra on industry benchmark tests, such as undergraduate level knowledge, graduate level reasoning and basic mathematics. It allows users to upload images and documents for the first time. #LLm #AI #ML cnbc.com/2024/03/04/google-bac…
#AI #ML #llm


Funny Reddit thread about Tim Cook's comment: "the Mac is the best computer for AI." Apple fanboys defend it with unified ram, and Apple haters attack with Nvidia GPU speed. lol #ML #AI #Apple #Mac reddit.com/r/LocalLLaMA/commen…
#apple #AI #ML #mac


#AI #ML #llm #llama