File size: 1,337 Bytes
297c21b
 
9a696ee
297c21b
9a696ee
297c21b
 
 
 
 
 
11694f8
 
d597453
262a35e
2992d9f
374e487
b1deddd
9a696ee
884cd94
140cfcf
 
 
9a696ee
e689974
9a696ee
99eb7e1
ba097ab
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
---
title: Financial Bot
emoji: 🚀
colorFrom: red
colorTo: green
sdk: gradio
sdk_version: 4.16.0
app_file: app.py
pinned: false
license: mit
---

# Friendly Fincancial Bot

This is the **Inference component** of a 3-part **prod-ready** FTI feature-training-inference **RAG-framework LLMOps** course. \
In this iteration, I've **replaced Falcon 7B Instruct** with the **currently-SoTa (Jan'24) Mistral-7B-Instruct-v0.2**, \
fine-tuned using **Unsloth** on an expanded dataset of financial questions and answers generated with the help of GPT-4, \
quantized and augmented with a 4bit QLoRa. \
\
Prompt analysis and model registry is handled by **Comet LLM**, and finance news is streamed via **Bytewax** using an \
**Alpaca API**, then parsed, cleaned, and chunked with **unstructured**, and finally sent as a vector embedding to \
**Qdrant**'s serverless vector store. **LangChain** chains the prompt and most relevant news article with real-time \
finance information, **contextualizing the output**. \
\
**#TODO:** Add citations to output to show end-user which article has been used to generate the output.

I have contributed to the original MIT licensed (ka-ching!) course which can be found here:\
https://medium.com/decoding-ml/the-llms-kit-build-a-production-ready-real-time-financial-advisor-system-using-streaming-ffdcb2b50714