Add deepseek local model deepseek-r1:1.5b
This commit is contained in:
@@ -1,12 +1,20 @@
|
|||||||
{
|
{
|
||||||
"deepseek":{
|
"deepseek-chat":{
|
||||||
"api_url": "https://api.deepseek.com",
|
"api_url": "https://api.deepseek.com",
|
||||||
"api_key": "sk-12165b127043441697a8940918e207ac"
|
"api_key": "sk-12165b127043441697a8940918e207ac",
|
||||||
|
"model":"deepseek-chat"
|
||||||
},
|
},
|
||||||
|
|
||||||
"ollama":{
|
"llama3.2:1b":{
|
||||||
"api_url": "http://localhost:11434/v1",
|
"api_url": "http://localhost:11434/v1",
|
||||||
"api_key": "ollama"
|
"api_key": "ollama",
|
||||||
|
"model":"llama3.2:1b"
|
||||||
|
},
|
||||||
|
|
||||||
|
"deepseek-r1:1.5b":{
|
||||||
|
"api_url": "http://localhost:11434/v1",
|
||||||
|
"api_key": "ollama",
|
||||||
|
"model":"deepseek-r1:1.5b"
|
||||||
},
|
},
|
||||||
|
|
||||||
"send_email": {
|
"send_email": {
|
||||||
|
|||||||
BIN
data/MNIST/raw/t10k-images-idx3-ubyte
Normal file
BIN
data/MNIST/raw/t10k-images-idx3-ubyte
Normal file
Binary file not shown.
BIN
data/MNIST/raw/t10k-images-idx3-ubyte.gz
Normal file
BIN
data/MNIST/raw/t10k-images-idx3-ubyte.gz
Normal file
Binary file not shown.
BIN
data/MNIST/raw/t10k-labels-idx1-ubyte
Normal file
BIN
data/MNIST/raw/t10k-labels-idx1-ubyte
Normal file
Binary file not shown.
BIN
data/MNIST/raw/t10k-labels-idx1-ubyte.gz
Normal file
BIN
data/MNIST/raw/t10k-labels-idx1-ubyte.gz
Normal file
Binary file not shown.
BIN
data/MNIST/raw/train-images-idx3-ubyte
Normal file
BIN
data/MNIST/raw/train-images-idx3-ubyte
Normal file
Binary file not shown.
BIN
data/MNIST/raw/train-images-idx3-ubyte.gz
Normal file
BIN
data/MNIST/raw/train-images-idx3-ubyte.gz
Normal file
Binary file not shown.
BIN
data/MNIST/raw/train-labels-idx1-ubyte
Normal file
BIN
data/MNIST/raw/train-labels-idx1-ubyte
Normal file
Binary file not shown.
BIN
data/MNIST/raw/train-labels-idx1-ubyte.gz
Normal file
BIN
data/MNIST/raw/train-labels-idx1-ubyte.gz
Normal file
Binary file not shown.
0
experiment.py
Normal file
0
experiment.py
Normal file
BIN
mnist_cnn.pth
Normal file
BIN
mnist_cnn.pth
Normal file
Binary file not shown.
BIN
mnist_cnn_improved.pth
Normal file
BIN
mnist_cnn_improved.pth
Normal file
Binary file not shown.
Binary file not shown.
Binary file not shown.
Binary file not shown.
@@ -23,7 +23,7 @@ def buffbot():
|
|||||||
st.code(f.read(), language="python")
|
st.code(f.read(), language="python")
|
||||||
st.divider()
|
st.divider()
|
||||||
# Select AI model for chatbot
|
# Select AI model for chatbot
|
||||||
model_options = ["llama3.2:1b", "deepseek-chat", ]
|
model_options = ["deepseek-r1:1.5b", "llama3.2:1b", "deepseek-chat", ]
|
||||||
# on_change callback to clear chat history when model is changed
|
# on_change callback to clear chat history when model is changed
|
||||||
selected_model = st.selectbox("**👉Please select a model to start**", model_options, on_change=clear_chat)
|
selected_model = st.selectbox("**👉Please select a model to start**", model_options, on_change=clear_chat)
|
||||||
|
|
||||||
@@ -36,26 +36,35 @@ def buffbot():
|
|||||||
|
|
||||||
# Load API credentials from config.json
|
# Load API credentials from config.json
|
||||||
# the config file contains the API key and base URL for the selected model
|
# the config file contains the API key and base URL for the selected model
|
||||||
"""
|
"""
|
||||||
{
|
{
|
||||||
"deepseek":{
|
"deepseek":{
|
||||||
"api_url": "https://api.deepseek.com",
|
"api_url": "https://api.deepseek.com",
|
||||||
"api_key": "YOUR_API_KEY"
|
"api_key": "YOUR_API_KEY",
|
||||||
},
|
"model":"deepseek-chat"
|
||||||
|
},
|
||||||
"ollama":{
|
|
||||||
"api_url": "http://localhost:11434/v1",
|
"ollama3.2:1b":{
|
||||||
"api_key": "ollama"
|
"api_url": "http://localhost:11434/v1",
|
||||||
}
|
"api_key": "ollama",
|
||||||
}
|
"model":"llama3.2:1b"
|
||||||
"""
|
},
|
||||||
|
|
||||||
|
"deepseek-r1:1.5b":{
|
||||||
|
"api_url": "http://localhost:11434/v1",
|
||||||
|
"api_key": "ollama",
|
||||||
|
"model":"deepseek-r1:1.5b"
|
||||||
|
},
|
||||||
|
}
|
||||||
|
"""
|
||||||
# The API key and base URL are loaded based on the selected model
|
# The API key and base URL are loaded based on the selected model
|
||||||
with open('app_config.json') as config_file:
|
with open('app_config.json') as config_file:
|
||||||
config = json.load(config_file)
|
config = json.load(config_file)
|
||||||
# deepseek-chat model, online API
|
# deepseek-chat model, online API
|
||||||
if selected_model == "deepseek-chat":
|
if selected_model == "deepseek-chat":
|
||||||
api_base_url = config["deepseek"]["api_url"]
|
api_base_url = config[selected_model]["api_url"]
|
||||||
api_key = config["deepseek"]["api_key"]
|
api_key = config[selected_model]["api_key"]
|
||||||
|
model = config[selected_model]["model"]
|
||||||
st.info("Powered by the online [DeepSeek](https://www.deepseek.com/) API!\
|
st.info("Powered by the online [DeepSeek](https://www.deepseek.com/) API!\
|
||||||
Just a heads up, you have 10 messages to use.")
|
Just a heads up, you have 10 messages to use.")
|
||||||
# Set the maximum number of user messages
|
# Set the maximum number of user messages
|
||||||
@@ -63,11 +72,20 @@ def buffbot():
|
|||||||
|
|
||||||
# llama3.2:1b model, local API
|
# llama3.2:1b model, local API
|
||||||
if selected_model == "llama3.2:1b":
|
if selected_model == "llama3.2:1b":
|
||||||
api_base_url = config["ollama"]["api_url"]
|
api_base_url = config[selected_model]["api_url"]
|
||||||
api_key = config["ollama"]["api_key"]
|
api_key = config[selected_model]["api_key"]
|
||||||
|
model = config[selected_model]["model"]
|
||||||
st.info("Powered by local llama3.2:1b model via [Ollama](https://ollama.com/library/llama3.2:1b)!\
|
st.info("Powered by local llama3.2:1b model via [Ollama](https://ollama.com/library/llama3.2:1b)!\
|
||||||
Just a heads up, you have 100 messages to use.")
|
Just a heads up, you have 100 messages to use.")
|
||||||
MAX_USER_MESSAGES = 100
|
MAX_USER_MESSAGES = 100
|
||||||
|
|
||||||
|
if selected_model == "deepseek-r1:1.5b":
|
||||||
|
api_base_url = config[selected_model]["api_url"]
|
||||||
|
api_key = config[selected_model]["api_key"]
|
||||||
|
model = config[selected_model]["model"]
|
||||||
|
st.info("Powered by local deepseek-r1:1.5b model via [Ollama](https://ollama.com/library/deepseek-r1:1.5b)!\
|
||||||
|
Just a heads up, you have 100 messages to use.")
|
||||||
|
MAX_USER_MESSAGES = 100
|
||||||
|
|
||||||
# Initialize OpenAI client to connect with the selected model API
|
# Initialize OpenAI client to connect with the selected model API
|
||||||
client = OpenAI(api_key=api_key, base_url=api_base_url)
|
client = OpenAI(api_key=api_key, base_url=api_base_url)
|
||||||
|
|||||||
@@ -104,7 +104,8 @@ def home():
|
|||||||
""", unsafe_allow_html=True)
|
""", unsafe_allow_html=True)
|
||||||
|
|
||||||
# Faculty Advisors
|
# Faculty Advisors
|
||||||
st.markdown("""
|
st.markdown(
|
||||||
|
"""
|
||||||
<section>
|
<section>
|
||||||
<h2>Faculty Advisors</h2>
|
<h2>Faculty Advisors</h2>
|
||||||
<div class="faculty-container">
|
<div class="faculty-container">
|
||||||
|
|||||||
@@ -1,14 +1,5 @@
|
|||||||
import streamlit as st
|
import streamlit as st
|
||||||
from webpages.SendEmail import send_email
|
from webpages.SendEmail import send_email
|
||||||
|
|
||||||
# # Example usage of the send_email function
|
|
||||||
# if __name__ == "__main__":
|
|
||||||
# send_email(
|
|
||||||
# sender_email="noreply@buffteks.org", # Sender's email address
|
|
||||||
# password="cidm4360fall2024@*", # Sender's email password
|
|
||||||
# to_email="REPLACE_WITH_RESIDENT_EMAIL", # Recipient's email address
|
|
||||||
# subject="Package Pickup Notification", # Subject of the email
|
|
||||||
# )
|
|
||||||
|
|
||||||
def testing():
|
def testing():
|
||||||
st.write("Testing Page")
|
st.write("Testing Page")
|
||||||
|
|||||||
Reference in New Issue
Block a user