This commit is contained in:
guanjihuan 2024-06-26 03:37:41 +08:00
parent 0a246550b9
commit 2dba2caf3b
2 changed files with 96 additions and 1 deletions

View File

@ -119,6 +119,8 @@ internlm-chat-7b 模型文件下载https://huggingface.co/internlm/internlm-c
python -m streamlit run ./InternLM.py --theme.base dark --server.port 8501 python -m streamlit run ./InternLM.py --theme.base dark --server.port 8501
``` ```
#### 4. ……
### 三、使用大语言模型 API ### 三、使用大语言模型 API
#### 1. 智谱 - ChatGLM_Turbo #### 1. 智谱 - ChatGLM_Turbo
@ -204,3 +206,21 @@ pip install volcengine-python-sdk
``` ```
python -m streamlit run ./Doubao_lite_32k.py --theme.base dark --server.port 8501 python -m streamlit run ./Doubao_lite_32k.py --theme.base dark --server.port 8501
``` ```
#### 8. OpenAI - GPT_3.5_Turbo
OpenAI 的 API 接口(需要海外的 IP 地址以及海外银行卡https://platform.openai.com
需要安装 OpenAI 软件包:
```
pip install openai
```
运行命令:
```
python -m streamlit run ./GPT_3.5_Turbo.py --theme.base dark --server.port 8501
```
#### 9. ……

View File

@ -0,0 +1,75 @@
"""
This code is supported by the website: https://www.guanjihuan.com
The newest version of this code is on the web page: https://www.guanjihuan.com/archives/38502
"""
import streamlit as st
st.set_page_config(
page_title="Chat",
layout='wide'
)
import openai
API_KEY = ""
with st.sidebar:
with st.expander('参数', expanded=True):
top_p = st.slider('top_p', 0.01, 1.0, step=0.01, value=0.8, key='top_p_session')
temperature = st.slider('temperature', 0.51, 1.0, step=0.01, value=0.85, key='temperature_session')
def reset_parameter():
st.session_state['top_p_session'] = 0.8
st.session_state['temperature_session'] = 0.85
reset_parameter_button = st.button('重置', on_click=reset_parameter)
prompt = st.chat_input("在这里输入您的命令")
def clear_all():
st.session_state.messages = []
st.session_state.ai_response = []
if 'messages' not in st.session_state:
st.session_state.messages = []
if 'ai_response' not in st.session_state:
st.session_state.ai_response = []
for ai_response in st.session_state.ai_response:
with st.chat_message(ai_response["role"], avatar=ai_response.get("avatar")):
st.markdown(ai_response["content"])
prompt_placeholder = st.chat_message("user", avatar='user')
with st.chat_message("robot", avatar="assistant"):
message_placeholder = st.empty()
def response_of_gpt(prompt):
st.session_state.messages.append({'role': 'user', 'content': prompt})
client = openai.OpenAI(
api_key=API_KEY,
)
completion = client.chat.completions.create(
model="gpt-3.5-turbo-0125",
messages=st.session_state.messages,
stream=True,
temperature=temperature,
top_p=top_p,
)
full_content = ''
for chunk in completion:
response = chunk.choices[0].delta.content or ""
full_content += response
message_placeholder.markdown(full_content)
if stop_button:
break
st.session_state.messages.append({'role': 'assistant',
'content': full_content})
st.session_state.ai_response.append({"role": "robot", "content": full_content, "avatar": "assistant"})
return full_content
if prompt:
prompt_placeholder.markdown(prompt)
st.session_state.ai_response.append({"role": "user", "content": prompt, "avatar": 'user'})
stop = st.empty()
stop_button = stop.button('停止', key='break_response')
response_of_gpt(prompt)
stop.empty()
button_clear = st.button("清空", on_click=clear_all, key='clear')