December 2025

Show HN: 22 GB of Hacker News in SQLite
380 by keepamovin | 124 comments on Hacker News.
Community, All the HN belong to you. This is an archive of hacker news that fits in your browser. When I made HN Made of Primes I realized I could probably do this offline sqlite/wasm thing with the whole GBs of archive. The whole dataset. So I tried it, and this is it. Have Hacker News on your device. Go to this repo ( https://ift.tt/YUyFMEk ): you can download it. Big Query -> ETL -> npx serve docs - that's it. 20 years of HN arguments and beauty, can be yours forever. So they'll never die. Ever. It's the unkillable static archive of HN and it's your hands. That's my Year End gift to you all. Thank you for a wonderful year, have happy and wonderful 2026. make something of it.

Show HN: Z80-μLM, a 'Conversational AI' That Fits in 40KB
415 by quesomaster9000 | 95 comments on Hacker News.
How small can a language model be while still doing something useful? I wanted to find out, and had some spare time over the holidays. Z80-μLM is a character-level language model with 2-bit quantized weights ({-2,-1,0,+1}) that runs on a Z80 with 64KB RAM. The entire thing: inference, weights, chat UI, it all fits in a 40KB .COM file that you can run in a CP/M emulator and hopefully even real hardware! It won't write your emails, but it can be trained to play a stripped down version of 20 Questions, and is sometimes able to maintain the illusion of having simple but terse conversations with a distinct personality. -- The extreme constraints nerd-sniped me and forced interesting trade-offs: trigram hashing (typo-tolerant, loses word order), 16-bit integer math, and some careful massaging of the training data meant I could keep the examples 'interesting'. The key was quantization-aware training that accurately models the inference code limitations. The training loop runs both float and integer-quantized forward passes in parallel, scoring the model on how well its knowledge survives quantization. The weights are progressively pushed toward the 2-bit grid using straight-through estimators, with overflow penalties matching the Z80's 16-bit accumulator limits. By the end of training, the model has already adapted to its constraints, so no post-hoc quantization collapse. Eventually I ended up spending a few dollars on Claude API to generate 20 questions data (see examples/guess/GUESS.COM), I hope Anthropic won't send me a C&D for distilling their model against the ToS ;P But anyway, happy code-golf season everybody :)

Show HN: Ez FFmpeg – Video editing in plain English
398 by josharsh | 191 comments on Hacker News.
I built a CLI tool that lets you do common video/audio operations without remembering ffmpeg syntax. Instead of: ffmpeg -i video.mp4 -vf "fps=15,scale=480:-1:flags=lanczos" -loop 0 output.gif You write: ff convert video.mp4 to gif More examples: ff compress video.mp4 to 10mb ff trim video.mp4 from 0:30 to 1:00 ff extract audio from video.mp4 ff resize video.mp4 to 720p ff speed up video.mp4 by 2x ff reverse video.mp4 There are similar tools that use LLMs (wtffmpeg, llmpeg, ai-ffmpeg-cli), but they require API keys, cost money, and have latency. Ez FFmpeg is different: - No AI – just regex pattern matching - Instant – no API calls - Free – no tokens - Offline – works without internet It handles ~20 common operations that cover 90% of what developers actually do with ffmpeg. For edge cases, you still need ffmpeg directly. Interactive mode (just type ff) shows media files in your current folder with typeahead search. npm install -g ezff

Show HN: Vibium – Browser automation for AI and humans, by Selenium's creator
385 by hugs | 109 comments on Hacker News.
i started the selenium project 21 years ago. vibium is what i'd build if i started over today with ai agents in mind. go binary under the hood (handles browser, bidi, mcp) but devs never see it. just npm install vibium. python/java coming. for claude code: claude mcp add vibium -- npx -y vibium v1 ships today. ama.

Show HN: CineCLI – Browse and torrent movies directly from your terminal
325 by samsep10l | 111 comments on Hacker News.
Hi HN I built CineCLI — a cross-platform terminal app to browse movies, view details, and open torrents directly in your system torrent client. Features: - Search movies from the terminal - Rich UI with ratings, runtime, genres - Interactive & non-interactive modes - Magnet handling via system default client - Linux/macOS/Windows support - No ads, no tracking GitHub: https://ift.tt/e9rObSl PyPI: https://ift.tt/DCmTBja Would love feedback from terminal + Python folks

Ask HN: What are the best engineering blogs with real-world depth?
417 by nishilpatel | 125 comments on Hacker News.
I’m looking for examples of high-quality engineering blog posts—especially from tech company blogs, that go beyond surface-level explanations. Specifically interested in posts that: 1. Explain technical concepts clearly and concisely 2. Show real implementation details, trade-offs, and failures 3. Are well-structured and readable 4. Tie engineering decisions back to business or product outcomes Any standout blogs, posts, or platforms you regularly learn from?

Ask HN: Should "I asked $AI, and it said" replies be forbidden in HN guidelines?
433 by embedding-shape | 257 comments on Hacker News.
As various LLMs become more and more popular, so does comments with "I asked Gemini, and Gemini said ....". While the guidelines were written (and iterated on) during a different time, it seems like it might be time to have a discussion about if those sort of comments should be welcomed on HN or not. Some examples: - https://ift.tt/4Si9FHa - https://ift.tt/YjoO3ae - https://ift.tt/WrEl9oi Personally, I'm on HN for the human conversation, and large LLM-generated texts just get in the way of reading real text from real humans (assumed, at least). What do you think? Should responses that basically boil down to "I asked $LLM about $X, and here is what $LLM said:" be allowed on HN, and the guidelines updated to state that people shouldn't critique it (similar to other guidelines currently), or should a new guideline be added to ask people from refrain from copy-pasting large LLM responses into the comments, or something else completely?

MKRdezign

Contact Form

Name

Email *

Message *

Powered by Blogger.
Javascript DisablePlease Enable Javascript To See All Widget