Tracing the thoughts of a large language model
Anthropic's latest interpretability research: a new microscope to understand Claude's internal mechanismswww.anthropic.com
Anthropic's latest interpretability research: a new microscope to understand Claude's internal mechanismswww.anthropic.com
DeepSeek's free 685B-parameter AI model runs at 20 tokens/second on Apple's Mac Studio, outperforming Claude Sonnet while using just 200 watts, challenging OpenAI's cloud-dependent business model.Michael Nuñez (VentureBeat)
Airbnb recently completed our first large-scale, LLM-driven code migration, updating nearly 3.5K React component test files from Enzyme to use React Testing Library (RTL) instead. We’d originally…Charles Covey-Brandt (The Airbnb Tech Blog)
A text-to-speech (TTS) and Speech-to-Speech (STS) library built on Apple's MLX framework, providing efficient speech synthesis on Apple Silicon. - Blaizzy/mlx-audioGitHub
Generate a conversation between two models using OpenAI API and Kokoro TTS. - chigkim/AIChatGitHub
I shared a controversial take the other day at an event and I decided to write it down in a longer format: I’m afraid AI won't give us a compressed 21st century.thomwolf.io
At Sesame, our goal is to achieve “voice presence”—the magical quality that makes spoken interactions feel real, understood, and valued.Sesame
OmniParser, turn your LLM into GUI agenthuggingface.co
We’re on a journey to advance and democratize artificial intelligence through open source and open science.huggingface.co
As my findings in my #GeneticProgramming research are becoming more and more thorough, I have started to document them. The main problem in my runs is still lack of diversity.
PS: It could be nice if we could write papers in some dialect of #markdown
PS2: These writings are for myself to remember where am I and where I'm going. No academic paper yet :)
#ML #MachineLearning #WakeWordDetection
We’re on a journey to advance and democratize artificial intelligence through open source and open science.huggingface.co
Today, Meta FAIR is releasing several new research artifacts in support of our goal of achieving advanced machine intelligence (AMI) while also supporting open science and reproducibility.ai.meta.com
Die @openhomefoundation freut sich übrigens über Sprachschnipsel eurer Stimme. Es soll nur "OK NABU" eingesprochen werden, also viel einfacher als etwa bei #commonvoice
Gerne so viele unterschiedliche Sprecher wie möglich, damit später dies wakeword zuverlässig erkannt wird
ohf-voice.github.io/wake-word-…
#stt #stimme #voice #nabu #crowdsourcing #commons #nlp #ml #homeassistant #hass #iot #smarthome
Dear llama.cpp Official, Hi, I'm writing to address our new PR submission for integrating our model MiniCPM-Llama3-V 2.5 into llama.cpp, which has been trending on Huggingface for over a week a...GitHub
Přejděte na adresu paypal.me/geotts a zadejte částku. Protože jde o PayPal, je to jednoduché a bezpečné. Nemáte účet PayPal? Nevadí.PayPal.Me
To try everything Brilliant has to offer for free for a full 30 days, visit http://www.brilliant.org/howmoneyworks. You’ll also get 20% off an annual premium...YouTube
Breakthrough models AlphaProof and AlphaGeometry 2 solve advanced reasoning problems in mathematicsGoogle DeepMind
Inference code for Llama models. Contribute to meta-llama/llama development by creating an account on GitHub.GitHub
In an engaging presentation at Weights & Biases’ Fully Connected conference, Joe Spisak, Product Director of GenAI at Meta, unveiled the latest family of Lla...YouTube
Phi-3 family of small language and multi-modal models. Language models are available in short- and long-context lengths.huggingface.co
Editted to fix link. Please boost for reach if this kind of stuff interests you. Will post more on this later.
Once upon a time, there was a cool emulator frontend called Retroarch. This emulator wasn't accessible until I and a few other gamers went to them and asked about adding accessibility. An amazing person known as BarryR made it happen. Now, if you turn on accessibility mode in settings, or pass the "--accessibility" (or something like that) flag on the command line, you get spoken menus, including the emulator's pause menu, good for saving states and such. Then, using PIL and other image processing Python utilities, running a server and hooking into Retroarch, the script allowed players to move around the map, battle, talk to NPC's, ETC. The only problem was, no one wanted to test it. The blind gaming community pretty much spoke, saying that we want new games. We want cool new, easy accessibility. So that's what we have no, follow the beacon or get sighted help in the case of diablo and such. It's sad, but meh. It's what we wanted I guess. No Zelda for us. So, this is about as far as he got:
To expand on what devinprater was saying: I am working on an accessibility pack/service for Final Fantasy 1 for the NES (this was what was shown in the latest RetroArch update). The idea is similar to how Pokemon Crystal access works, but it's using the RetroArch AI Service interface to do so.
Right now, the FF1 access service is mostly done, but I need more testers to try it out and give me feedback on how it's working. Right now, you can get up to the point where you get the ship, but there's no code to deal with how the ship moves, so that still needs to be done. Likewise with the airship later on.
The service works the latest version of RetroArch, on linux and mac, but not windows. This is due to how nvda reads out the text and until the next major update to nvda (which will have a feature to fix this), it'll have to wait. If you have those, I (or maybe devinprater) can help you set it up on mac/linux to test out. The package itself is available at: (new link cause old one broke yesterday): dropbox.com/scl/fi/ggffl769fx6…
#accessibility #finalFantasy #RetroArch #blind #emulator #emulation #Python #ai #ML #MachineLearning
Anthropic on Monday debuted Claude 3, a chatbot and suite of AI models that it calls its fastest and most powerful yet.Hayden Field (CNBC)