| # This file was autogenerated by uv via the following command: |
| # uv pip compile pyproject.toml --python 3.10 -o requirements.txt |
| aiohappyeyeballs==2.6.1 |
| # via aiohttp |
| aiohttp==3.12.9 |
| # via |
| # langchain-community |
| # llama-index-core |
| # realtime |
| aiosignal==1.3.2 |
| # via aiohttp |
| aiosqlite==0.21.0 |
| # via llama-index-core |
| annotated-types==0.7.0 |
| # via pydantic |
| anyio==4.9.0 |
| # via |
| # groq |
| # httpx |
| # openai |
| arxiv==2.2.0 |
| # via llama-index-readers-papers |
| asttokens==3.0.0 |
| # via stack-data |
| async-timeout==4.0.3 |
| # via |
| # aiohttp |
| # langchain |
| attrs==25.3.0 |
| # via aiohttp |
| backoff==2.2.1 |
| # via langfuse |
| banks==2.1.2 |
| # via llama-index-core |
| beautifulsoup4==4.13.4 |
| # via |
| # llama-index-readers-file |
| # wikipedia |
| cachetools==5.5.2 |
| # via google-auth |
| certifi==2025.4.26 |
| # via |
| # httpcore |
| # httpx |
| # llama-cloud |
| # requests |
| charset-normalizer==3.4.2 |
| # via requests |
| click==8.2.1 |
| # via |
| # duckduckgo-search |
| # llama-cloud-services |
| # nltk |
| colorama==0.4.6 |
| # via |
| # click |
| # griffe |
| # ipython |
| # pytest |
| # tqdm |
| comm==0.2.2 |
| # via |
| # ipykernel |
| # ipywidgets |
| dataclasses-json==0.6.7 |
| # via |
| # langchain-community |
| # llama-index-core |
| debugpy==1.8.14 |
| # via ipykernel |
| decorator==5.2.1 |
| # via ipython |
| defusedxml==0.7.1 |
| # via llama-index-readers-papers |
| deprecated==1.2.18 |
| # via |
| # banks |
| # llama-index-core |
| deprecation==2.1.0 |
| # via postgrest |
| dirtyjson==1.0.8 |
| # via llama-index-core |
| distro==1.9.0 |
| # via |
| # groq |
| # openai |
| dotenv==0.9.9 |
| # via final-assignment-template (pyproject.toml) |
| duckduckgo-search==6.4.2 |
| # via llama-index-tools-duckduckgo |
| exceptiongroup==1.3.0 |
| # via |
| # anyio |
| # ipython |
| # pytest |
| executing==2.2.0 |
| # via stack-data |
| feedparser==6.0.11 |
| # via arxiv |
| filelock==3.18.0 |
| # via |
| # huggingface-hub |
| # torch |
| # transformers |
| filetype==1.2.0 |
| # via |
| # langchain-google-genai |
| # llama-index-core |
| frozenlist==1.6.2 |
| # via |
| # aiohttp |
| # aiosignal |
| fsspec==2025.5.1 |
| # via |
| # huggingface-hub |
| # llama-index-core |
| # torch |
| google-ai-generativelanguage==0.6.18 |
| # via langchain-google-genai |
| google-api-core==2.25.1 |
| # via google-ai-generativelanguage |
| google-auth==2.40.3 |
| # via |
| # google-ai-generativelanguage |
| # google-api-core |
| googleapis-common-protos==1.70.0 |
| # via |
| # google-api-core |
| # grpcio-status |
| # opentelemetry-exporter-otlp-proto-grpc |
| # opentelemetry-exporter-otlp-proto-http |
| gotrue==2.12.0 |
| # via supabase |
| greenlet==3.2.3 |
| # via sqlalchemy |
| griffe==1.7.3 |
| # via banks |
| groq==0.28.0 |
| # via langchain-groq |
| grpcio==1.72.1 |
| # via |
| # google-api-core |
| # grpcio-status |
| # opentelemetry-exporter-otlp-proto-grpc |
| grpcio-status==1.71.0 |
| # via google-api-core |
| h11==0.16.0 |
| # via httpcore |
| h2==4.2.0 |
| # via httpx |
| hf-xet==1.1.3 |
| # via |
| # final-assignment-template (pyproject.toml) |
| # huggingface-hub |
| hpack==4.1.0 |
| # via h2 |
| httpcore==1.0.9 |
| # via httpx |
| httpx==0.28.1 |
| # via |
| # gotrue |
| # groq |
| # langfuse |
| # langgraph-sdk |
| # langsmith |
| # llama-cloud |
| # llama-index-core |
| # openai |
| # postgrest |
| # storage3 |
| # supabase |
| # supafunc |
| # tavily-python |
| httpx-sse==0.4.0 |
| # via langchain-community |
| huggingface-hub==0.32.4 |
| # via |
| # final-assignment-template (pyproject.toml) |
| # langchain-huggingface |
| # llama-index-llms-huggingface-api |
| # sentence-transformers |
| # tokenizers |
| # transformers |
| hyperframe==6.1.0 |
| # via h2 |
| idna==3.10 |
| # via |
| # anyio |
| # httpx |
| # requests |
| # yarl |
| importlib-metadata==8.7.0 |
| # via opentelemetry-api |
| iniconfig==2.1.0 |
| # via pytest |
| ipykernel==6.29.5 |
| # via final-assignment-template (pyproject.toml) |
| ipython==8.37.0 |
| # via |
| # ipykernel |
| # ipywidgets |
| ipywidgets==8.1.7 |
| # via final-assignment-template (pyproject.toml) |
| jedi==0.19.2 |
| # via ipython |
| jinja2==3.1.6 |
| # via |
| # banks |
| # torch |
| jiter==0.10.0 |
| # via openai |
| joblib==1.5.1 |
| # via |
| # nltk |
| # scikit-learn |
| jsonpatch==1.33 |
| # via langchain-core |
| jsonpointer==3.0.0 |
| # via jsonpatch |
| jupyter-client==8.6.3 |
| # via ipykernel |
| jupyter-core==5.8.1 |
| # via |
| # ipykernel |
| # jupyter-client |
| jupyterlab-widgets==3.0.15 |
| # via ipywidgets |
| langchain==0.3.25 |
| # via |
| # final-assignment-template (pyproject.toml) |
| # langchain-community |
| langchain-community==0.3.25 |
| # via final-assignment-template (pyproject.toml) |
| langchain-core==0.3.65 |
| # via |
| # langchain |
| # langchain-community |
| # langchain-google-genai |
| # langchain-groq |
| # langchain-huggingface |
| # langchain-text-splitters |
| # langgraph |
| # langgraph-checkpoint |
| # langgraph-prebuilt |
| langchain-google-genai==2.1.5 |
| # via final-assignment-template (pyproject.toml) |
| langchain-groq==0.3.2 |
| # via final-assignment-template (pyproject.toml) |
| langchain-huggingface==0.3.0 |
| # via final-assignment-template (pyproject.toml) |
| langchain-text-splitters==0.3.8 |
| # via langchain |
| langfuse==3.0.0 |
| # via final-assignment-template (pyproject.toml) |
| langgraph==0.4.8 |
| # via final-assignment-template (pyproject.toml) |
| langgraph-checkpoint==2.1.0 |
| # via |
| # langgraph |
| # langgraph-prebuilt |
| langgraph-prebuilt==0.2.2 |
| # via langgraph |
| langgraph-sdk==0.1.70 |
| # via langgraph |
| langsmith==0.3.45 |
| # via |
| # langchain |
| # langchain-community |
| # langchain-core |
| llama-cloud==0.1.23 |
| # via |
| # llama-cloud-services |
| # llama-index-indices-managed-llama-cloud |
| llama-cloud-services==0.6.30 |
| # via llama-parse |
| llama-index==0.12.40 |
| # via final-assignment-template (pyproject.toml) |
| llama-index-agent-openai==0.4.9 |
| # via |
| # llama-index |
| # llama-index-program-openai |
| llama-index-cli==0.4.3 |
| # via llama-index |
| llama-index-core==0.12.40 |
| # via |
| # final-assignment-template (pyproject.toml) |
| # llama-cloud-services |
| # llama-index |
| # llama-index-agent-openai |
| # llama-index-cli |
| # llama-index-embeddings-openai |
| # llama-index-indices-managed-llama-cloud |
| # llama-index-llms-huggingface-api |
| # llama-index-llms-openai |
| # llama-index-multi-modal-llms-openai |
| # llama-index-program-openai |
| # llama-index-question-gen-openai |
| # llama-index-readers-file |
| # llama-index-readers-llama-parse |
| # llama-index-readers-papers |
| # llama-index-readers-wikipedia |
| # llama-index-tools-duckduckgo |
| # llama-index-tools-tavily-research |
| llama-index-embeddings-openai==0.3.1 |
| # via |
| # llama-index |
| # llama-index-cli |
| llama-index-indices-managed-llama-cloud==0.7.4 |
| # via llama-index |
| llama-index-llms-huggingface-api==0.5.0 |
| # via final-assignment-template (pyproject.toml) |
| llama-index-llms-openai==0.4.3 |
| # via |
| # llama-index |
| # llama-index-agent-openai |
| # llama-index-cli |
| # llama-index-multi-modal-llms-openai |
| # llama-index-program-openai |
| # llama-index-question-gen-openai |
| llama-index-multi-modal-llms-openai==0.5.1 |
| # via llama-index |
| llama-index-program-openai==0.3.2 |
| # via |
| # llama-index |
| # llama-index-question-gen-openai |
| llama-index-question-gen-openai==0.3.1 |
| # via llama-index |
| llama-index-readers-file==0.4.9 |
| # via llama-index |
| llama-index-readers-llama-parse==0.4.0 |
| # via llama-index |
| llama-index-readers-papers==0.3.2 |
| # via final-assignment-template (pyproject.toml) |
| llama-index-readers-wikipedia==0.3.0 |
| # via final-assignment-template (pyproject.toml) |
| llama-index-tools-duckduckgo==0.3.0 |
| # via final-assignment-template (pyproject.toml) |
| llama-index-tools-tavily-research==0.3.0 |
| # via final-assignment-template (pyproject.toml) |
| llama-parse==0.6.30 |
| # via llama-index-readers-llama-parse |
| markdown-it-py==3.0.0 |
| # via rich |
| markupsafe==3.0.2 |
| # via jinja2 |
| marshmallow==3.26.1 |
| # via dataclasses-json |
| matplotlib-inline==0.1.7 |
| # via |
| # ipykernel |
| # ipython |
| mdurl==0.1.2 |
| # via markdown-it-py |
| mpmath==1.3.0 |
| # via sympy |
| multidict==6.4.4 |
| # via |
| # aiohttp |
| # yarl |
| mypy-extensions==1.1.0 |
| # via typing-inspect |
| nest-asyncio==1.6.0 |
| # via |
| # ipykernel |
| # llama-index-core |
| networkx==3.4.2 |
| # via |
| # llama-index-core |
| # torch |
| nltk==3.9.1 |
| # via |
| # llama-index |
| # llama-index-core |
| numpy==2.2.6 |
| # via |
| # langchain-community |
| # llama-index-core |
| # pandas |
| # scikit-learn |
| # scipy |
| # transformers |
| openai==1.84.0 |
| # via |
| # llama-index-agent-openai |
| # llama-index-embeddings-openai |
| # llama-index-llms-openai |
| opentelemetry-api==1.34.0 |
| # via |
| # langfuse |
| # opentelemetry-exporter-otlp-proto-grpc |
| # opentelemetry-exporter-otlp-proto-http |
| # opentelemetry-sdk |
| # opentelemetry-semantic-conventions |
| opentelemetry-exporter-otlp==1.34.0 |
| # via langfuse |
| opentelemetry-exporter-otlp-proto-common==1.34.0 |
| # via |
| # opentelemetry-exporter-otlp-proto-grpc |
| # opentelemetry-exporter-otlp-proto-http |
| opentelemetry-exporter-otlp-proto-grpc==1.34.0 |
| # via opentelemetry-exporter-otlp |
| opentelemetry-exporter-otlp-proto-http==1.34.0 |
| # via opentelemetry-exporter-otlp |
| opentelemetry-proto==1.34.0 |
| # via |
| # opentelemetry-exporter-otlp-proto-common |
| # opentelemetry-exporter-otlp-proto-grpc |
| # opentelemetry-exporter-otlp-proto-http |
| opentelemetry-sdk==1.34.0 |
| # via |
| # langfuse |
| # opentelemetry-exporter-otlp-proto-grpc |
| # opentelemetry-exporter-otlp-proto-http |
| opentelemetry-semantic-conventions==0.55b0 |
| # via opentelemetry-sdk |
| orjson==3.10.18 |
| # via |
| # langgraph-sdk |
| # langsmith |
| ormsgpack==1.10.0 |
| # via langgraph-checkpoint |
| packaging==24.2 |
| # via |
| # deprecation |
| # huggingface-hub |
| # ipykernel |
| # langchain-core |
| # langfuse |
| # langsmith |
| # marshmallow |
| # pytest |
| # transformers |
| pandas==2.2.3 |
| # via llama-index-readers-file |
| parso==0.8.4 |
| # via jedi |
| pillow==11.2.1 |
| # via |
| # llama-index-core |
| # sentence-transformers |
| platformdirs==4.3.8 |
| # via |
| # banks |
| # jupyter-core |
| # llama-cloud-services |
| pluggy==1.6.0 |
| # via pytest |
| postgrest==1.0.2 |
| # via supabase |
| primp==0.15.0 |
| # via duckduckgo-search |
| prompt-toolkit==3.0.51 |
| # via ipython |
| propcache==0.3.1 |
| # via |
| # aiohttp |
| # yarl |
| proto-plus==1.26.1 |
| # via |
| # google-ai-generativelanguage |
| # google-api-core |
| protobuf==5.29.5 |
| # via |
| # google-ai-generativelanguage |
| # google-api-core |
| # googleapis-common-protos |
| # grpcio-status |
| # opentelemetry-proto |
| # proto-plus |
| psutil==7.0.0 |
| # via ipykernel |
| pure-eval==0.2.3 |
| # via stack-data |
| pyasn1==0.6.1 |
| # via |
| # pyasn1-modules |
| # rsa |
| pyasn1-modules==0.4.2 |
| # via google-auth |
| pydantic==2.11.5 |
| # via |
| # banks |
| # gotrue |
| # groq |
| # langchain |
| # langchain-core |
| # langchain-google-genai |
| # langfuse |
| # langgraph |
| # langsmith |
| # llama-cloud |
| # llama-cloud-services |
| # llama-index-core |
| # openai |
| # postgrest |
| # pydantic-settings |
| pydantic-core==2.33.2 |
| # via pydantic |
| pydantic-settings==2.9.1 |
| # via langchain-community |
| pygments==2.19.1 |
| # via |
| # ipython |
| # pytest |
| # rich |
| pyjwt==2.10.1 |
| # via gotrue |
| pypdf==5.6.0 |
| # via llama-index-readers-file |
| pytest==8.4.0 |
| # via pytest-mock |
| pytest-mock==3.14.1 |
| # via gotrue |
| python-dateutil==2.9.0.post0 |
| # via |
| # jupyter-client |
| # pandas |
| # realtime |
| # storage3 |
| python-dotenv==1.1.0 |
| # via |
| # dotenv |
| # llama-cloud-services |
| # pydantic-settings |
| pytz==2025.2 |
| # via pandas |
| pywin32==310 |
| # via jupyter-core |
| pyyaml==6.0.2 |
| # via |
| # huggingface-hub |
| # langchain |
| # langchain-community |
| # langchain-core |
| # llama-index-core |
| # transformers |
| pyzmq==27.0.0 |
| # via |
| # ipykernel |
| # jupyter-client |
| realtime==2.4.3 |
| # via supabase |
| regex==2024.11.6 |
| # via |
| # nltk |
| # tiktoken |
| # transformers |
| requests==2.32.3 |
| # via |
| # arxiv |
| # google-api-core |
| # huggingface-hub |
| # langchain |
| # langchain-community |
| # langfuse |
| # langsmith |
| # llama-index-core |
| # opentelemetry-exporter-otlp-proto-http |
| # requests-toolbelt |
| # tavily-python |
| # tiktoken |
| # transformers |
| # wikipedia |
| requests-toolbelt==1.0.0 |
| # via langsmith |
| rich==14.0.0 |
| # via final-assignment-template (pyproject.toml) |
| rsa==4.9.1 |
| # via google-auth |
| safetensors==0.5.3 |
| # via transformers |
| scikit-learn==1.7.0 |
| # via sentence-transformers |
| scipy==1.15.3 |
| # via |
| # scikit-learn |
| # sentence-transformers |
| sentence-transformers==4.1.0 |
| # via final-assignment-template (pyproject.toml) |
| sgmllib3k==1.0.0 |
| # via feedparser |
| six==1.17.0 |
| # via python-dateutil |
| sniffio==1.3.1 |
| # via |
| # anyio |
| # groq |
| # openai |
| soupsieve==2.7 |
| # via beautifulsoup4 |
| sqlalchemy==2.0.41 |
| # via |
| # langchain |
| # langchain-community |
| # llama-index-core |
| stack-data==0.6.3 |
| # via ipython |
| storage3==0.11.3 |
| # via supabase |
| strenum==0.4.15 |
| # via |
| # postgrest |
| # supafunc |
| striprtf==0.0.26 |
| # via llama-index-readers-file |
| supabase==2.15.3 |
| # via final-assignment-template (pyproject.toml) |
| supafunc==0.9.4 |
| # via supabase |
| sympy==1.14.0 |
| # via torch |
| tavily-python==0.7.5 |
| # via llama-index-tools-tavily-research |
| tenacity==9.1.2 |
| # via |
| # langchain-community |
| # langchain-core |
| # llama-index-core |
| threadpoolctl==3.6.0 |
| # via scikit-learn |
| tiktoken==0.9.0 |
| # via |
| # llama-index-core |
| # tavily-python |
| tokenizers==0.21.1 |
| # via |
| # langchain-huggingface |
| # transformers |
| tomli==2.2.1 |
| # via pytest |
| torch==2.7.1 |
| # via sentence-transformers |
| tornado==6.5.1 |
| # via |
| # ipykernel |
| # jupyter-client |
| tqdm==4.67.1 |
| # via |
| # huggingface-hub |
| # llama-index-core |
| # nltk |
| # openai |
| # sentence-transformers |
| # transformers |
| traitlets==5.14.3 |
| # via |
| # comm |
| # ipykernel |
| # ipython |
| # ipywidgets |
| # jupyter-client |
| # jupyter-core |
| # matplotlib-inline |
| transformers==4.52.4 |
| # via sentence-transformers |
| typing-extensions==4.14.0 |
| # via |
| # aiosqlite |
| # anyio |
| # beautifulsoup4 |
| # exceptiongroup |
| # groq |
| # huggingface-hub |
| # ipython |
| # langchain-core |
| # llama-index-core |
| # multidict |
| # openai |
| # opentelemetry-api |
| # opentelemetry-exporter-otlp-proto-grpc |
| # opentelemetry-exporter-otlp-proto-http |
| # opentelemetry-sdk |
| # opentelemetry-semantic-conventions |
| # pydantic |
| # pydantic-core |
| # pypdf |
| # realtime |
| # rich |
| # sentence-transformers |
| # sqlalchemy |
| # torch |
| # typing-inspect |
| # typing-inspection |
| typing-inspect==0.9.0 |
| # via |
| # dataclasses-json |
| # llama-index-core |
| typing-inspection==0.4.1 |
| # via |
| # pydantic |
| # pydantic-settings |
| tzdata==2025.2 |
| # via pandas |
| urllib3==2.4.0 |
| # via requests |
| wcwidth==0.2.13 |
| # via prompt-toolkit |
| websockets==14.2 |
| # via realtime |
| widgetsnbextension==4.0.14 |
| # via ipywidgets |
| wikipedia==1.4.0 |
| # via final-assignment-template (pyproject.toml) |
| wrapt==1.17.2 |
| # via |
| # deprecated |
| # langfuse |
| # llama-index-core |
| xxhash==3.5.0 |
| # via langgraph |
| yarl==1.20.0 |
| # via aiohttp |
| zipp==3.22.0 |
| # via importlib-metadata |
| zstandard==0.23.0 |
| # via langsmith |
|
|