Kagi: Announcing The Assistant
Kagi has launched the Assistant, an AI feature enhancing search with Quick Answer and Summarize Page, prioritizing user privacy and allowing Custom Assistants, available for $25 monthly to Ultimate Plan subscribers.
Read original articleKagi has launched the Assistant, an AI-integrated feature designed to enhance the search experience by providing smarter, faster, and more intuitive results. Key functionalities include Quick Answer for instant knowledge retrieval, Summarize Page for quick highlights, and the ability to ask questions about web pages directly in search results. The Assistant prioritizes user privacy, with all threads being private by default and data not being used for model training. Users can choose from various leading LLM models and create Custom Assistants tailored to specific needs, such as car maintenance advice. The Assistant also allows mid-thread editing and branching, enabling users to refine their queries for more accurate responses. Kagi emphasizes its commitment to user data protection, with automatic thread expiration and no ads or tracking. The Assistant is available to Kagi Ultimate Plan subscribers for $25 per month, with discounts for annual subscriptions.
- Kagi Assistant integrates AI to improve search functionality with features like Quick Answer and Summarize Page.
- User privacy is prioritized, with threads being private and data not used for training models.
- Users can create Custom Assistants tailored to specific tasks and utilize various leading LLM models.
- The Assistant allows for mid-thread editing and branching to refine queries.
- Available for Kagi Ultimate Plan subscribers at $25 per month, with discounts for annual plans.
Related
Optimizing AI Inference at Character.ai
Character.AI optimizes AI inference for LLMs, handling 20,000+ queries/sec globally. Innovations like Multi-Query Attention and int8 quantization reduced serving costs by 33x since late 2022, aiming to enhance AI capabilities worldwide.
Guide to Slack AI
Slack AI integrates generative tools to summarize conversations, offering customizable daily recaps and a search feature for quick answers. It prioritizes data security and privacy, aligning billing with workspace plans.
A short update on getting answers on the modern internet
The author compares search engines, favoring Kagi and Bing Copilot over Google, DuckDuckGo, and ChatGPT. Kagi stands out for clean interface and reliable results, while Bing Copilot faces copyright concerns.
Show HN: Chrome Extension to access the OpenAI API from any textbox
A browser extension named "Barra AI" on GitHub provides free AI-generated responses using GPT model. Features include no popups, client-side processing, and open-source. Installation requires Chrome extension and OpenAI API key. Developers can contribute on GitHub.
Kagi LLM Benchmarking Project
The Kagi LLM Benchmarking Project assesses large language models on reasoning, coding, and instruction-following, providing metrics like accuracy and latency, while comparing costs and performance across various models.
- Many users appreciate Kagi's search capabilities and the integration of AI features, with some stating it reduces the need for other subscriptions.
- However, several users express concerns about the $25 monthly fee, feeling it adds to their existing expenses for similar services.
- Some users are skeptical about the effectiveness of Kagi's AI, comparing it unfavorably to other search engines and AI tools.
- There are requests for improvements in Kagi's core search functionality, with some users feeling that the focus on AI detracts from the search experience.
- Users value the option to opt-in for the AI features without affecting their current subscription plans.
I've been using Kagi for a while (almost two years now!) and it's been nothing but excellent!
Lenses are very useful (Reddit lens is on every second search), and I personally really like the AI features they are working on.
The new more advanced assistant which is able to do searches, which can also be constrained to lenses, and lets you pick an arbitrary model, is excellent, and basically means I don't need a chatgpt/claude subscription, as Kagi covers it very well.
All in all, great product which I'm happy to pay for.
I love Kagi and can’t recommend it enough. I wish I could just give them my api key for this instead of paying several different service providers for the same ai access to the same models. This is getting expensive.
One feature here that I think competitors lack is that the LLM's view of search results can be constrained by Kagi's search "lens" [0] that let you exclude various categories of results.
I use Kagi but haven't dug into lenses, anyone have experience ?
I'm currently trying to write python script interfacing with outlook's mailterm interface (win32com.client) and it's annoying. I wonder if I can restrict search results to a particular domain so it only pulls from microsoft docs...
While I am still happy with the search, I find the value of the assistant no longer worth it. At first, I thought I would be able to replace other providers with the wide range of offered models, but now I find myself often going to distinct providers. Kagi seems to have introduced a harsh and generalized input token limit for prompts (detached from what the underlying model can handle), which makes it almost impossible to work with code or documents. I think the VW example in the announcement makes it clear that the assistant is only intended to respond to prompts of a few sentences—searching by sentences instead of keywords. I don't see input for documents or things like code execution coming here, so the difference of $15 monthly can get me fancier features.
Mixing the output of the model with the limited content of approximately five websites (= Assistant with internet access), on average, is most of the time counterproductive for me. The sites mentioned in the sources of the output are often not sites I would usually visit or get information from (they usually don't correspond to the top results on Kagi web search). I have used search engines for decades now and have built a sort of index/pattern/feeling in my head for which site I am going to visit, which AI can not match for me.
The lenses are nice but are also pretty specialized. Most of the time I find myself just doing a broad query, which in most cases will already satisfy me.
I will definitely continue to use Kagi for search. There is no comparable search experience out there, and it makes searching definitely more effective.
This field is moving very quickly and since there will always be a lag in getting new features/models into the "wrappers" I'd rather get it from the source.
One random thing I wish Kagi could do is offer a way to promote the "official" website to the top of the search results. I have various up/down rankings applied but when I search for "DataDog" I want their official website at the top of the list, not under SO/Reddit/etc posts. If I'm searching for a problem then yes, I want those sites higher but it's slightly frustrating to have to scroll down 3-5 results to find the main website for a product/service/company. I feel like they should be able to differentiate between "search about X product" vs "X product".
By already having a traditional search engine this puts Kagi at a big advantage compared to someone like Perplexity, or even Claude and OpenAI who I think are all cobbling together solutions on top of Bing's API.
For the first while, the search results were inarguably better than any alternative. I was thrilled and was recommending Kagi to anyone who'd listen. In the last six months or so, things have gone downhill. More often, my results ignore some of my search terms and I have to try to "trick" Kagi into matching them all. More often, I get only one or two pages of results when other engines give me notably more relevant hits. And more often, my search hangs or fails to return at all.
I don't care if they noodle with AI or make T-shirts as long as the search is great. What I've experienced is the search getting worse while they noodle with AI and make T-shirts, though. The two may be entirely unrelated but from an outside perspective, it feels like their core offering is suffering because of these other-than-search undertakings.
So far, I'm sticking with it, but my enthusiasm has definitely diminished. A couple of my friends have canceled their subscriptions in disappointment and I've started to consider cancelling mine, as well.
Mistral and GPT - Create the same example of a flip card
Gemini - Creates glowing neon text
and Claude - Produces a pulsing dot, that enlarges and shrinks and radiates a fading white shadow. That's cool.
Stop launching new products (browser, summarizer, gpt, assistant) while your core product is still behind the competition in many areas.
I might end up disabling web search until they've tweaked it a little more. I really like kagi search so I feel pretty confident that they'll get it, but I don't think it's ready to replace the brand-name chatbots.
In my daily work as an MD it's become my reflexive go-to for looking up answers to specific to general, easy to complex clinical questions. I use it far more frequently than UpToDate (which is no less than the holy book of medicine), more than PubMed/Google Scholar searches, and definitely more than a basic web search (Google, only b/c it's a hassle to log in to Kagi every session at work).
Maybe 1 time out of 10 it won't give a correct or meaningful answer (in which case my prompt needs to be refined, or is just not suited for this kind of tool). But apart from that it will give me exactly what I need, because it uses Kagi search to source its answers. Kagi search does a decent job bringing to the top relevant journal articles (which in turn may mention other articles, adding indirectly to the trove of sources FastGPT pulls its final answer from). It shows the 5 search results it referenced at the bottom of the page, so more often than not if I don't get my answer in the direct summary, I have very relevant sources to read through.
I also don't think you need a Kagi account to use it.
The thing stopping me currently from trying this or Claude is I rely on the Opt+Space shortcut with the ChatGPT mac app.
Are there any other options for a native mac app with integration as good as the ChatGPT app?
Is it still legal do do something with a computer without involving "ai"?
This mid-thread editing feature sounds really useful, I'm curious how does it work when you switch between models in the middle of a conversation?
Like, say I start with a general search question, then halfway through I want to switch to a coding model to ask something like, "Can you create a Python dictionary of the top 10 longest city names in the UK and their populations?"
Does the context carry over smoothly, or would I need to rephrase things when switching models? Wondering how it handles tasks that require different kinds of expertise without losing track of the flow.
among other things, it randomly says I have no connectivity in an awful modal, I click to search and the keyboard doesn't open, I hit "enter" to search and nothing happens (I need to tap on the search suggestion). pretty bad so far. will try it again in a couple of weeks to see if there is any progress...
What it is though, is fast, available on all my devices, constantly upgraded, and integrated with their already excellent search engine.
When I see these sorts of announcements and read some of the comments here, it makes me worry that bad customers cause enshittification and I hope kagi stays true to their human-friendly web search product.
Update: I see now that they say it's not available for free users. Need to pay $25/month. Not sure why, they can offer it for free users with the cheaper models like they do now to generate a "quick answer". I'm not going to pay to try it out.
For many queries side-by-side with Startpage it delivered the same results word-for-word (sure you get a few sponsored links top-3 of Startpage but its no big deal to scroll past those).
For other things, it was just plain annoying, e.g. "newest $type restaurants in $large_city" half the results on the first page were from 10 years ago (e.g. dated 2014). I mean FFS I put the word "newest" in there !
They seem to have a habit of interespersing very weird Facebook links randomly in the middle of a list of results. For example I was searching for something related to a specific Prometheus function (which I explicitly named in the query, alongside the word prometheus) and Kagi insisted on interspersing the technical results with random links to Facebook pages of companies selling "girlie dresses for proms".
I approached Kagi with an open mind, but having used up the 100 free searches nothing made me say "just shut up and take my money".
> PyLLMs is a minimal Python library to connect to various Language Models (LLMs) with a built-in model performance benchmark.
how do i _ in python !chat
Related
Optimizing AI Inference at Character.ai
Character.AI optimizes AI inference for LLMs, handling 20,000+ queries/sec globally. Innovations like Multi-Query Attention and int8 quantization reduced serving costs by 33x since late 2022, aiming to enhance AI capabilities worldwide.
Guide to Slack AI
Slack AI integrates generative tools to summarize conversations, offering customizable daily recaps and a search feature for quick answers. It prioritizes data security and privacy, aligning billing with workspace plans.
A short update on getting answers on the modern internet
The author compares search engines, favoring Kagi and Bing Copilot over Google, DuckDuckGo, and ChatGPT. Kagi stands out for clean interface and reliable results, while Bing Copilot faces copyright concerns.
Show HN: Chrome Extension to access the OpenAI API from any textbox
A browser extension named "Barra AI" on GitHub provides free AI-generated responses using GPT model. Features include no popups, client-side processing, and open-source. Installation requires Chrome extension and OpenAI API key. Developers can contribute on GitHub.
Kagi LLM Benchmarking Project
The Kagi LLM Benchmarking Project assesses large language models on reasoning, coding, and instruction-following, providing metrics like accuracy and latency, while comparing costs and performance across various models.