File size: 1,261 Bytes
297c21b
 
9a696ee
297c21b
9a696ee
297c21b
 
 
 
 
 
 
cf12fde
d597453
2992d9f
 
e689974
9a696ee
 
884cd94
 
9a696ee
 
e689974
9a696ee
99eb7e1
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
---
title: Financial Bot
emoji: 🚀
colorFrom: red
colorTo: green
sdk: gradio
sdk_version: 4.16.0
app_file: app.py
pinned: false
license: mit
---

# Friendly Fincancial Bot

This is the Inference module of a 3-part **prod-ready** FTI feature-training-inference **RAG-framework LLMOps** course. \
In this iteration, I've **replaced Falcon 7B Instruct** with the **currently-SoTa (Jan'24) Mistral-7B-Instruct-v0.2**, \
fine-tuned using **Unsloth** on financial questions and answers generated with the help of GPT-4, quantized \
and augmented with a 4bit QLoRa. \
\
Prompt analysis and model registry is handled by **Comet LLM**, and finance news is streamed via **Bytewax** using an \
**Alpaca API**, and then sent as a vector embedding to **Qdrant**'s serverless vector store. **LangChain** chains the prompt and \
most relevant news article to provide answers with real-time finance information embedded within the output. \
\
**#TODO:** Add citations to output to show end-user which article has been used to generate the output.

I have contributed to the original MIT licensed (ka-ching!) course which can be found here:\
[https://medium.com/decoding-ml/the-llms-kit-build-a-production-ready-real-time-financial-advisor-system-using-streaming-ffdcb2b50714]