Items tagged with: LLM

Search

Items tagged with: LLM


😲 DeepSeek-V3-4bit runs at >20 tokens per second and <200W using MLX on an M3 Ultra with 512GB. This might be the best and most user-friendly way to run DeepSeek-V3 on consumer hardware, possibly the most affordable too. You can finally run a GPT-4o level model locally, with possibly even better quality. #LLM #AI #ML #DeepSeek #OpenAI #GPT #OpenWeight #OpenSource venturebeat.com/ai/deepseek-v3…


AirBnB migrated 3,5k React component test files from Enzyme to RTL using LLMs and automation in just six weeks. The effort was originally estimated to take 1.5 years of manual engineering time to do by hand. #LLM #AI #ML #Automation
medium.com/airbnb-engineering/…


😲 It’s only March, but there’s already been incredible progress in open-weight LLMs this year. Here are my top 5 local LLM recommendations for anyone with 24GB of VRAM to try: Phi-4-14B for speed, Mistral-Small-24B for RAG, Gemma-3-27B for general use, Qwen2.5-Coder-32B for coding, QWQ-32B for reasoning. #LLM #ML #AI
#AI #ML #llm


Here's another fun project! AIChat lets you generate conversations between two LLMs on any topic using OpenAI API. You can mix and match models from Ollama, MLX, Claude, OpenAI, Google AI Studio, etc. It uses Kokoro-ONNX for TTS. Check out an example + the GitHub repo:
youtube.com/watch?v=FgSZLZnYlA…
github.com/chigkim/AIChat
#LLM #TTS #AI #ML
#AI #ML #llm #tts


Sensitive content


Gemma3 is out with 1b, 4b, 12b, 27b. It also accepts images except 1b model. It's on Ollama, but ollama run gemma3 will use small 4b model. Be sure to specify size if you want to use bigger model. For example, ollama run gemma3:27b #LLM #ML #AI ollama.com/library/gemma3/tags
#AI #ML #llm


Very interesting read from HuggingFace co-founder: For big scientific breakthroughs, "we don't need an A+ student who can answer every question with general knowledge. We need a B student who sees and questions what everyone else missed." Current AI is great at answering known questions, but struggles to challenge assumptions and push boundaries. True innovation comes from questioning and redefining knowledge—something AI isn’t quite ready for yet. #AI #LLM #ML thomwolf.io/blog/scientific-ai…
#AI #ML #llm



#AI #ML #llm


3/3 Gemini Flash 2.0 Thinking: Humans are not just social animals, but profoundly self-domesticated apes. And this self-domestication is the key to understanding our unique cognitive and social abilities.
#LLM #AI #ML
#AI #ML #llm


2/3 Deepseek R1: Humans instinctively convert selfish desires into cooperative systems by collectively pretending abstract rules, money, laws, rights, are real. These shared hallucinations act as games, where competition is secretly redirected to benefit the group, turning conflict into society's fuel.
#LLM #AI #ML
#AI #ML #llm


1/3 Three Responses from 3 top reasoning models to Lex Fridman's question, What's one truly novel insight about humans?
O1-pro: Humans are unique among known species in that they simultaneously rewrite two layers of reality, the external world and their own private mental landscapes, and then merge these two rewritten layers into a continuous personal narrative that feels objectively true.
#LLM #AI #ML
#AI #ML #llm


#AI #ML #llm



ChatGPT is fairly convincing at creating code. But, like with everything you have to be vigilant on what it suggests you do. As a test I asked ChatGPT to "Write me an example C application using libcurl using secure HTTPS connection to fetch a file and save it locally. Provide instructions on how to create a test HTTPS server with self-signed certificate, and how to configure the server and the C client application for testing."

ChatGPT was fairly good here. It provided example code that didn't outright disable certificate validation, but rather uses the self-signed certificate as the CA store:

const char *cert_file = "./server.crt"; // Self-signed certificate
...
curl_easy_setopt(curl, CURLOPT_CAINFO, cert_file); // Verify server certificate
curl_easy_setopt(curl, CURLOPT_SSL_VERIFYPEER, 1L);
curl_easy_setopt(curl, CURLOPT_SSL_VERIFYHOST, 2L);

This is a very good idea, as blanket disabling security is a big nono. The deployment instructions were also quite nice, creating a self-signed certificate with openssl, and then setting up the test website with python3 http.server like this:

mkdir -p server
echo "This is a test file." > server/testfile.txt
python3 -m http.server 8443 --bind 127.0.0.1 --certfile server.crt --keyfile server.key

Looks pretty nice, right?

Except that this is totally hallucinated and even if it wasn't, it'd be totally insecure in a multiuser system anyway.

Python3 http.server doesn't allow you to pass certfile and keyfile like specified. But lets omit that small detail and assume it did. What would be the problem then?

You'd be sharing your whole work directory to everyone else on the same host. Anyone else on the same host could grab all your files with: wget --no-check-certificate -r 127.0.0.1:8443

AI can be great, but never ever blindly trust the instructions provided by a LLM. They're not intelligent, but very good at pretending to be.

#ChatGPT #LLMs #LLM


#llm #cnb




I am having a lot of fun with reasoning models, but lately with Gemini 2.0 Thinking Experimental model. I just had it try proving to me that the earth is flat. I think it did pretty well, minus all the disclaimers. :) This may be a great tool to help you understand how your opponent may be constructing their arguments. #AI #LLM #Gemini


Hraju si teď s LLM DeepSeek R1 github.com/deepseek-ai/DeepSee… Vypadá to, že tento open-source model, který je dostupný zdarma (chat.deepseek.com a zaškrtnout Deepthink), dosahuje kvalit OpenAI-o1 modelu.
Na lokálním stroji ale ten 32B nebo dokonce 70B model (zabírají cca 40 GB na disku) rozhodně nerozjedu, laptop se mně potí u 8B modelů. :)
#DeepSeek #LLM


Jednou za čas si zkusím nějaké ty lokální #LLM, jestli by nešly prakticky používat bez přístupu k netu, a zatím teda ještě ne. :D
#llm


Thx for your link and efforts @Seirdy !

All this said, being part of a decentralized web, as pointed out in this toot, our publicly visible interaction lands on other instances and servers of the #fediVerse and can be scrapped there. I wonder if this situation actually might lead, or should lead, to a federation of servers that share the same robots.txt "ideals".

As @Matthias pointed out in his short investigation of the AI matter, this has (in my eyes) already unimagined levels of criminal and without any doubt unethical behavior, not to mention the range of options rouge actors have at hand.

It's evident why for example the elongated immediately closed down access to X's public tweets and I guess other companies did the same for the same reasons. Obviously the very first reason was to protect their advantage about the hoarded data sets to train their AI in the first place. Yet, considering the latest behavior of the new owner of #twitter, nothing less than at least the creation of #AI driven lists of "political" enemies, and not only from all the collected data on his platform, is to be expected. A international political nightmare of epical proportions. Enough material for dystopian books and articles for people like @Cory Doctorow, @Mike Masnick ✅, @Eva Wolfangel, @Taylor Lorenz, @Jeff Jarvis, @Elena Matera, @Gustavo Antúnez 🇺🇾🇦🇷, to mention a few of the #journalim community, more than one #podcast episode by @Tim Pritlove and @linuzifer, or some lifetime legal cases for @Max Schrems are at hand.

What we are facing now is the fact that we need to protect our and our users data and privacy because of the advanced capabilities of #LLM. We basically are forced to consider to change to private/restricted posts and close down our servers as not only the legal jurisdictions are way to scattered over the different countries and ICANN details, but legislation and comprehension by the legislators is simply none existent, as @Anke Domscheit-Berg could probably agree to.

Like to say, it looks like we need to go dark, a fact that will drive us even more into disappearing as people will have less chance to see what we are all about, advancing further the advantages off the already established players in the social web space.
Just like Prof. Dr. Peter Kruse stated in his take about on YT The network is challenging us min 2:42 more than 14 years ago:
"With semantic understanding we'll have the real big brother. Someone is getting the best out of it and the rest will suffer."


Estoy usando un LLM para aprender teoría de filtros. Vale, ya os podéis lanzar a mi cuello, pero antes, me comentáis de algún sitio donde haya información fiable sobre teoría de filtros (polos, ceros, FIR, IIR) en notación que resulte accesible y donde puedas hacer preguntas gilipollas como: el número de coeficientes del numerador del filtro determina el orden del filtro?

Entre tanto, he aprendido algo. Y sí, corro el riesgo de errores o imprecisiones, pero por lo menos tengo un punto de partida.

#LLM #aprendizaje #accesibilidad


Reading some people complaining about image descriptions because they help train AI. I'm not sure what interpretation is worse: that there's a new excuse for not including alt text, as if there weren't enough already; or that people may actually earnestly be this spiteful about AI so as to regard helping others as a necessary collateral damage of their Buttlerian jihad.

#llm #AI #AltText


#Slop is low-quality media - including writing and images - made using generative artificial intelligence technology.


Quelle: Wikipedia.

Open source projects have to deal with a growing number of low-quality vulnerability reports based on AI. See for example this comment from Daniel Stenberg, maintainer of #Curl:

I'm sorry you feel that way, but you need to realize your own role here. We receive AI slop like this regularly and at volume. You contribute to unnecessary load of curl maintainers and I refuse to take that lightly and I am determined to act swiftly against it. Now and going forward.

You submitted what seems to be an obvious AI slop "report" where you say there is a security problem, probably because an AI tricked you into believing this. You then waste our time by not telling us that an AI did this for you and you then continue the discussion with even more crap responses - seemingly also generated by AI.

Weiterlesen bei HackerOne: Buffer Overflow Risk in Curl_inet_ntop and inet_ntop4.

#opensource #AI #LLM #Spam


#AI #ML #llm


This is making the rounds on Finnish social media.

A large association for Finnish construction companies, #Rakennusteollisuus, decided that they needed an English version of their website but apparently they didn't want to pay an actual #translator so just used some free #LLM with hilarious results.

They've fixed it now, but for a short while there was some comedy gold to be found.

P.s. I didn't find these, I've no idea who did.


A study asked 50 doctors to make six different diagnoses for medical conditions. "Doctors who did the project without AI got an average score of 74%, doctors who used AI got an average score of 76%, and ChatGPT itself got an average score of 90%." "AI didn’t help doctors using it as much as anticipated because physicians “didn’t listen to AI when AI told them things they didn’t agree. Most doctors couldn’t be convinced a chatbot knew more than them." #LLM #AI #ChatGPT qz.com/chatgpt-beat-doctors-at…



I'm a little puzzled at the salience that is being given to the Apple conclusions on #LLM #reasoning when we have lots of prior art. For example: LLMs cannot correctly infer a is b, if their corpora only contain b is a. #Paper: arxiv.org/abs/2309.12288

#AI #MachineLearning #logic


#AIagent promotes itself to #sysadmin , trashes #boot sequence

Fun experiment, but yeah, don't pipe an #LLM raw into /bin/bash

Buck #Shlegeris, CEO at #RedwoodResearch, a nonprofit that explores the risks posed by #AI , recently learned an amusing but hard lesson in automation when he asked his LLM-powered agent to open a secure connection from his laptop to his desktop machine.
#security #unintendedconsequences

theregister.com/2024/10/02/ai_…


🆕 blog! “GitHub's Copilot lies about its own documentation. So why would I trust it with my code?”

In the early part of the 20th Century, there was a fad for "Radium". The magical, radioactive substance that glowed in the dark. The market had decided that Radium was The Next Big Thing and tried to shove it into every product. There …

👀 Read more: shkspr.mobi/blog/2024/10/githu…

#AI #github #LLM



Massive E-Learning Platform #Udemy Gave Teachers a Gen #AI 'Opt-Out Window'. It's Already Over.

Udemy will train generative AI on classes developed/users contributed on its site. It is opt-out (meaning, everyone was already opted in) with a time window... and opting out may "affect course visibility and potential earnings."

Udemy's reason for the opt-out window was reportedly because removing data from LLMs is hard. IMO, that would be the reason for making it opt-in, but here we are...

#privacy #privacymatters #llm

404media.co/massive-e-learning…



😲 OMG, Audio Overview feature on NotebookLM is wild! It basically creates a podcast with two AI generated voices based on the source documents you upload. Definitely try if you haven't yet. #LLM #ML #AI blog.google/technology/ai/note…
#AI #ML #llm


#AI #ML #llm


😲 Kyle Kabasares, a Physics PhD graduate working at NASA's Ames Research Center, gave the methods section of his research paper to ChatGPT O1 Preview and asked it to generate the code based on the description. After just six prompts, it produced a working version of the code that took him a year to develop during his PhD. #ChatGPT #LLM #ML #AI youtube.com/watch?v=M9YOO7N5jF…


#DSGVO #llm #ki


After a long period of inactivity for vision language models, llama.cpp merged the support for MiniCPM-V-2.5. Hopefully the support for 2.6 is also on the way soon. #LLM #Multimodal #AI #ML
huggingface.co/openbmb/MiniCPM…
huggingface.co/openbmb/MiniCPM…
github.com/ggerganov/llama.cpp…