Gemini Embedding 2 has been released as a multimodal embedding model (it supports text, images, etc.). Instead of handling each data type separately, GeminiâŻEmbeddingâŻ2 puts everything into a single vector space for semantic search, classification, clustering, RAG (retrievalâaugmented generation), and similar tasksâmaking the workflow more unified and less messy. It supports over 100 languages, text up toâŻ8,192âŻtokens, up to six images (PNG/JPEG) per request, video up toâŻ120âŻseconds (MP4/MOV), audio, and PDF documents up to six pages.
The model uses âMatryoshka Representation Learning,â which lets you reduce the output dimension from the defaultâŻ3,072âŻtoâŻ1,536âŻorâŻ768âŻ(allowing a tradeâoff between quality and storage/speed), which is handy for systems that need to store many vectors.
Currently, the model is in public preview through the Gemini API and VertexâŻAI, with a readyâmade notebook available for testing.
canirun.ai is a handy website that gathers various hardware setups and shows how well they run todayâs openâsource large language models. Whatâs useful is that it provides detailed tokenâgeneration speeds per second, giving you reference data if youâre looking to build a system thatâs just powerful enough to run a local model for yourself.
The other day I came across this post on Facebook:âŻhttps://www.facebook.com/share/p/1DYEgsx5Kh/.
It shows how to create a JavaScript snippet that adds a panel displaying the message list in a ChatGPT web conversation. This is really handy and I hadnât thought of it, because long messages often require endless scrolling, which gets tiring and hard to follow.
In the article the author only provides a prompt for us to "vibeâcode" it. I tried it and it works quite well, but there are still a few spots that could be improved. Soon Iâll probably set up a GitHub repository and openâsource my code since I find it pretty interesting đ.
startups.rip is a website that gathers every project that has successfully raised funding from Y Combinator (YC) â a wellâknown worldwide ventureâcapital firm. It covers many projects from 2005 to today, including both successes and failures, and importantly provides detailed information about each story, the journey, and the reasons behind the outcomes. Itâs definitely worth reading.
Both OpenAI and Anthropic provide a free sixâmonth access to their ProâtoâMax service plans for openâsource project maintainers. Anthropic requires a project to have at leastâŻ5,000âŻstars on GitHub or over oneâŻmillion monthly downloads on NPM, while OpenAIâs eligibility criteria are much more general.
If you donât meet those thresholds, you can still submit a brief âletter of intentâ explaining why youâd like the grant. This shows that the companies are willing to listen to openâsource developersâ stories.
From what Iâve seen, the program has drawn quite a bit of criticism. Some community members argue that, since the two giants train their models on openâsource repositories, six months of free access is insufficient. Others defend it, saying six months is relatively generous because many openâsource maintainers donât make any money from their projects. What do you think?
Codex for Open Source Thank you for everything you ship. Claude Max is on us
If youâre using macOS with an external monitor, be sure to check whether thereâs an option to enable this mode. I tried turning it on yesterday and the colors looked noticeably more vivid. Iâve had the ViewFinityâŻS8 all year; itâs not a bad monitor, but when I plug it in the colors look washedâout for some reasonâstill, the difference is clear when I compare it to the MacBookâs screen. Adjusting the monitorâs settings didnât help either. It turns out the problem was that button; I was being a bit clueless.
These past few days Iâve been setting up additional system alerts for some of the servers I manage.
Monit is a tool thatâs been around for decades but is still very powerful. Itâs written in C, openâsource, easy to install and configure, and it consumes very little system resources. In basic terms, Monit lets you define alert thresholds; when a threshold is crossed it automatically triggers an alertâsuch as a process that suddenly disappears, a port that goes down, and many other situations.
Because of its long history, Monit even has a âshowcaseâ page that shows how people around the world use it. See the Real-world configuration examples.
Yesterday afternoon I spent time using AI to create a set of Skills for AI Agents.
If you have ever built a feature in a CMS that runs a workflow or an application, you are surely familiar with the kind of feature requests other teams make. For example, the Marketing team might ask for a report on the effectiveness of a recent campaign, or they might want a feature that sends bulk emails to a specific user group. In those cases we try to develop the feature as efficiently as possible, optimizing it so that the requester can use it without many questions.
The feature I just built is very simple: it takes a list of users, calls an API to update them, and returns a detailed report of successes and failures. I had actually built something like this before, but it was buggy and unusable. After reviewing it, I found it too complicated and costly to rewrite from scratch. Since AI Agents are becoming mainstream and our company is actively adopting AI in our work, I decided to create a dedicated set of Skills for this task. With just one markdown file, a Bash script, and a few prompts, the problem was solved. The usage is straightforwardâany tool that supports AI Agents and can read Skills can use it.
This could be the future of CMS operations: no fancy UI, no buttons, everything described in natural language, and the results returned naturally, making it very userâfriendly. In practice, all the power resides in the Bash script, which contains the processing logic. Even without AI Agents you could run the Bash command manually, but adding AI makes it much easier for people who arenât comfortable with the command line.
Looking further ahead, each CMS feature could be packaged as a set of Skills, with a Bash file (or an even more powerful CLI application) handling the core business logic. A clear output definition means that not only AI but also humans can use it to build additional interfaces if needed. In the past, MCPs were created to fetch data for AI processing, but they were a bit complex, whereas CLIs are more approachable, so developers gravitated toward building CLIs. Nowadays, anything can be built quickly because AI does most of the work; we just describe what we want. The real advantage now lies in experience with system implementation. đ
Maybe many people could use this. Jina is a research company that offers solutions related to LLM applications. I first learned about it while exploring embedding models for semantic search.
They also provide a range of other tools, one of which converts any article link into Markdown format to be more LLMâfriendly. As you know, the .md format is becoming more popular than ever because it supplies context for large language models to process information. Now you can get article data with a simple API call.
No wonder Cloudflare is so active in fighting AI âscrapingâ of web content đ
My first post after the New Year. I took a week off to get my work back on track, and starting today Iâll be posting regularly again.
By the way, Iâm building a few agents that use agent skills to serve as assistants for running and growing the blog. Itâs still in development, and Iâll share a showcase post when it looks solid enough đ
Lately Iâve noticed that any long article online often gets tagged as âAIâgeneratedâ and then criticized for having no value. What do you think about this? Do people dislike reading AIâwritten articles, or is it just because theyâre made by AI đ€?