From 7218463f9e4cc9ba5c1f017275adeac4f2111837 Mon Sep 17 00:00:00 2001 From: zhayujie Date: Fri, 10 Nov 2023 16:06:58 +0800 Subject: [PATCH] docs: update README --- README.md | 6 +++++- bridge/bridge.py | 2 +- voice/linkai/linkai_voice.py | 11 ++++++----- 3 files changed, 12 insertions(+), 7 deletions(-) diff --git a/README.md b/README.md index 521fcf1..5c679e7 100644 --- a/README.md +++ b/README.md @@ -6,7 +6,7 @@ - [x] **多端部署:** 有多种部署方式可选择且功能完备,目前已支持个人微信,微信公众号和企业微信应用等部署方式 - [x] **基础对话:** 私聊及群聊的消息智能回复,支持多轮会话上下文记忆,支持 GPT-3.5, GPT-4, claude, 文心一言, 讯飞星火 -- [x] **语音识别:** 可识别语音消息,通过文字或语音回复,支持 azure, baidu, google, openai等多种语音模型 +- [x] **语音识别:** 可识别语音消息,通过文字或语音回复,支持 azure, baidu, google, openai(whisper/tts) 等多种语音模型 - [x] **图片生成:** 支持图片生成 和 图生图(如照片修复),可选择 Dall-E, stable diffusion, replicate, midjourney模型 - [x] **丰富插件:** 支持个性化插件扩展,已实现多角色切换、文字冒险、敏感词过滤、聊天记录总结、文档总结和对话等插件 - [X] **Tool工具:** 与操作系统和互联网交互,支持最新信息搜索、数学计算、天气和资讯查询、网页总结,基于 [chatgpt-tool-hub](https://github.com/goldfishh/chatgpt-tool-hub) 实现 @@ -28,6 +28,10 @@ Demo made by [Visionn](https://www.wangpc.cc/) # 更新日志 +>**2023.11.10:** [1.5.0版本](https://github.com/zhayujie/chatgpt-on-wechat/releases/tag/1.5.0),新增 `gpt-4-turbo`, `dall-e-3`, `tts` 模型接入,完善图像理解&生成、语音识别&生成的多模态能力 + +>**2023.10.16:** 支持通过意图识别使用LinkAI联网搜索、数学计算、网页访问等插件,参考[插件文档](https://docs.link-ai.tech/platform/plugins) + >**2023.09.26:** 插件增加 文件/文章链接 一键总结和对话的功能,使用参考:[插件说明](https://github.com/zhayujie/chatgpt-on-wechat/tree/master/plugins/linkai#3%E6%96%87%E6%A1%A3%E6%80%BB%E7%BB%93%E5%AF%B9%E8%AF%9D%E5%8A%9F%E8%83%BD) >**2023.08.08:** 接入百度文心一言模型,通过 [插件](https://github.com/zhayujie/chatgpt-on-wechat/tree/master/plugins/linkai) 支持 Midjourney 绘图 diff --git a/bridge/bridge.py b/bridge/bridge.py index bceca10..db59545 100644 --- a/bridge/bridge.py +++ b/bridge/bridge.py @@ -29,7 +29,7 @@ class Bridge(object): self.btype["chat"] = const.XUNFEI if conf().get("use_linkai") and conf().get("linkai_api_key"): self.btype["chat"] = const.LINKAI - if not conf().get("voice_to_text") or conf().get("voice_to_text") in ["openai"]: + if not conf().get(":voice_to_text") or conf().get("voice_to_text") in ["openai"]: self.btype["voice_to_text"] = const.LINKAI if not conf().get("text_to_voice") or conf().get("text_to_voice") in ["openai", const.TTS_1, const.TTS_1_HD]: self.btype["text_to_voice"] = const.LINKAI diff --git a/voice/linkai/linkai_voice.py b/voice/linkai/linkai_voice.py index ea16312..7dc420c 100644 --- a/voice/linkai/linkai_voice.py +++ b/voice/linkai/linkai_voice.py @@ -1,18 +1,15 @@ """ google voice service """ -import json -import os import random - -import openai import requests - +from voice import audio_convert from bridge.reply import Reply, ReplyType from common.log import logger from config import conf from voice.voice import Voice from common import const +import os import datetime class LinkAIVoice(Voice): @@ -27,6 +24,10 @@ class LinkAIVoice(Voice): model = None if not conf().get("text_to_voice") or conf().get("voice_to_text") == "openai": model = const.WHISPER_1 + if voice_file.endswith(".amr"): + mp3_file = os.path.splitext(voice_file)[0] + ".mp3" + audio_convert.any_to_mp3(voice_file, mp3_file) + voice_file = mp3_file file = open(voice_file, "rb") file_body = { "file": file