-
-
Notifications
You must be signed in to change notification settings - Fork 12.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Internet search support for full model like DeepSeek R1 | 联网搜索支持到全模型如 DeepSeek R1 #6843
Comments
用deepseek 官方api ,r1无法触发联网emmm |
Using deepseek official API, r1 cannot trigger networking emmm |
@arvinxx yesterday I tried to make edits to make the search work. The thing is that when building a Docker image, various errors pop up: either with node_modules from packages/web-crawler (error of undefined rights) errors in the code eslint complains about spelling). I'll check now, maybe I'm doing something wrong? Do I need to run pnpm install commands after making edits to the code? When running pnpm dev locally, everything works, but it doesn't compile the Docker file |
前置判断模型,我使用了我手上大部分支持fc的模型
几个建议
|
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
这个我怀疑是入参没裁剪导致的,可能是系统bug,晚点我看看
中断是正常的,我的实现中做的逻辑是,如果输出的 chunk 类型不是 fc,就停止搜索模型的输出,并开始生成。
这个我觉得不太好,我昨天测了下其实像 4omini 都能做到简单的链式调用多个 tool。所以还是保留比较合适。
目前搜索结果的返回里的确没加上时间,我补一下吧
这个现在没加吗?我记得应该有才对 |
大部分fc的模型都是只返回fc的结果,但是部分模型,比如grok会在返回fc结果之前增加一段文本
从networks里面看好像是没有,没看代码,不知道在server有没有附加上去 |
Most of the FC models only return the results of FC, but some models, such as grok, will add a piece of text before returning the results of FC.
It seems that there is no code from the networks. I don't know if it is attached to the server. |
的确… 我昨天测 Claude 3.7 的时候也发现有这个问题。我感觉通过 prompt 约束可以解决,但昨天太晚了没测。后面有空再试试看。
你看下 第一句消息,放在 system role 里的 |
Indeed... I also found this problem when I tested Claude 3.7 yesterday. I feel that it can be solved through prompt constraints, but it was too late yesterday and it was not tested. Try it later if you have time.
Look at the first message, put in the system role |
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
This comment has been minimized.
I found the problem. After I turned off the function switch of r1, it was all normal. Thank you guys |
确实有些不稳定,也不知道怎么输入才能100%触发搜索,同一句话搜索一下 rtx5060的信息删除并重新生成第二次触发联网了 |
It is indeed a bit unstable, and I don’t know how to input it to trigger the search 100%. Search the same sentence. The information of rtx5060 is deleted and regenerated for the second trigger. |
@m0eak 智能的问题就是在如果模型不够聪明的情况下就会显得很蠢 🤡 可能强制联网仍然是一个刚需? |
能强制的话还是比较好的,因为现在有点量子力学的感觉,我也不知道他会不会联网😂 |
目前存在
这两个情况,会导致输入 tokens 很高 |
联网查询底部的引用列表能否实现,点击相应的引用能够打开对应的链接的效果 |
非常不稳定,openai 系统的 4o 4omini 辅助 r1 都联不了网 |
这么看下来,还是需要一个强制联网功能的 |
缓存的话,如果使用OpenAI / DeepSeek等API侧无感支持缓存的模型;应该是能自动启用缓存的 |
For cache, if you use OpenAI/DeepSeek and other API-side sensing-free model to support cache; it should be able to automatically enable cache. |
但是强制每次都联网/爬取多个网页的话,太烧token了——无论是辅助模型还是对话模型。对于相似问题抓取相同网页,造成大量token浪费以及时间延误的话,用户体验也不会好。而且辅助模型的token消耗不会被聊天工具栏“上下文明细”中体现,容易给用户带来“无感”的大量资源消耗。使用支持缓存的模型可以大幅缓解,但不是每个用户都会选择使用支持缓存的模型(他们的渠道是否支持缓存也是变量)。——感觉这部分是否应该在UI上有消耗大量token的风险提示? 另外,在辅助模型判断搜索意图的过程中,是否可以一并要求辅助模型对爬取内容进行去杂提纯以及总结后再返回给正式模型呢?一方面是节约对话模型token,另一方面,爬取出来的网页(特别是naive方式爬取的)实际上可能有大量冗余乃至干扰内容,知识含量低。 |
But it would be too token to force networking/crawl multiple web pages every time - whether it is an auxiliary model or a dialogue model. If crawling the same web pages with similar problems causes a lot of token waste and time delay, the user experience will not be good. Moreover, the token consumption of the auxiliary model will not be reflected in the "Context Details" of the chat toolbar, which will easily bring users a large amount of resource consumption without feeling. Using a secondary model that supports caching can be greatly mitigated, but not every user will choose to use a model that supports caching (whether their channel supports caching is also a variable). ——I feel that this part should have a risk warning of consuming a large amount of tokens on the UI? In addition, in the process of judging search intentions by auxiliary models, can the auxiliary models be required to remove and purify the crawled content and then summarize it before returning it to the formal model? On the one hand, it saves the dialogue model token. On the other hand, the crawled web pages (especially crawled in naive form) may actually have a lot of redundancy or even interfere with content, and have a low knowledge content. |
When searching online with auxiliary models, it may not be possible to bring time, which will lead to a confusing date when answering. For example, ds r1 cooperates with Gemini 2.0 flash assisted search: Today's NBA game, Gemini searches hard when calling the search plug-in Today's NBA game. The returned web page does not necessarily have an accurate date, which will lead to inaccurate results. It is recommended to add the current time prop to the auxiliary model. If you add "Today is March 12, 2025, Beijing time, what NBA games are there today?", it will be no problem |
我用下来,感觉本地 Ollama 部署的 Command R7B 相对来说更「愿意」触发搜索,但也不是很符合直觉,经常想要触发,但无动于衷。 尝试过这些:
|
I used it and felt that the Command R7B deployed by Ollama is relatively more "willing to" trigger searches, but it is not very intuitive. I often want to trigger them, but I am indifferent. Try these:
|
有没有计划添加强制使用 “搜索辅助模型” 的功能?最近体验下来,有些模型链式调用插件的能力似乎比较有限,gpt-4o-mini 这种模型训练的链式调用搜索做的挺好的,但是模型能力又差一些,可以考虑运行强制用 gpt-4o-mini 这种作为“搜索辅助模型”进行搜索,然后对话模型只负责对话。 |
Are there any plans to add a feature that forces the use of Search Assisted Models? Recently, I have experienced that some models have limited capabilities in chain call plug-ins. The chain call search trained by gpt-4o-mini is quite good, but the model capability is a bit poor. You can consider running a forced search using gpt-4o-mini as a "search auxiliary model", and then the dialogue model is only responsible for dialogue. |
Version 1.70.0 has supported online search for the full model. We welcome everyone to provide feedback on usage~
v.1.70.0 已支持到全模型的联网搜索,欢迎大家提出使用反馈~
The text was updated successfully, but these errors were encountered: