id
stringlengths
14
16
text
stringlengths
31
2.73k
metadata
dict
c0cf5ef15ac2-0
.ipynb .pdf ClearML Integration Contents ClearML Integration Getting API Credentials Setting Up Scenario 1: Just an LLM Scenario 2: Creating a agent with tools Tips and Next Steps ClearML Integration# In order to properly keep track of your langchain experiments and their results, you can enable the ClearML integration. ClearML is an experiment manager that neatly tracks and organizes all your experiment runs. Getting API Credentials# We’ll be using quite some APIs in this notebook, here is a list and where to get them: ClearML: https://app.clear.ml/settings/workspace-configuration OpenAI: https://platform.openai.com/account/api-keys SerpAPI (google search): https://serpapi.com/dashboard import os os.environ["CLEARML_API_ACCESS_KEY"] = "" os.environ["CLEARML_API_SECRET_KEY"] = "" os.environ["OPENAI_API_KEY"] = "" os.environ["SERPAPI_API_KEY"] = "" Setting Up# !pip install clearml !pip install pandas !pip install textstat !pip install spacy !python -m spacy download en_core_web_sm from datetime import datetime from langchain.callbacks import ClearMLCallbackHandler, StdOutCallbackHandler from langchain.callbacks.base import CallbackManager from langchain.llms import OpenAI # Setup and use the ClearML Callback clearml_callback = ClearMLCallbackHandler( task_type="inference", project_name="langchain_callback_demo", task_name="llm", tags=["test"], # Change the following parameters based on the amount of detail you want tracked visualize=True, complexity_metrics=True, stream_logs=True ) manager = CallbackManager([StdOutCallbackHandler(), clearml_callback]) # Get the OpenAI model ready to go
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-1
# Get the OpenAI model ready to go llm = OpenAI(temperature=0, callback_manager=manager, verbose=True) The clearml callback is currently in beta and is subject to change based on updates to `langchain`. Please report any issues to https://github.com/allegroai/clearml/issues with the tag `langchain`. Scenario 1: Just an LLM# First, let’s just run a single LLM a few times and capture the resulting prompt-answer conversation in ClearML # SCENARIO 1 - LLM llm_result = llm.generate(["Tell me a joke", "Tell me a poem"] * 3) # After every generation run, use flush to make sure all the metrics # prompts and other output are properly saved separately clearml_callback.flush_tracker(langchain_asset=llm, name="simple_sequential") {'action': 'on_llm_start', 'name': 'OpenAI', 'step': 3, 'starts': 2, 'ends': 1, 'errors': 0, 'text_ctr': 0, 'chain_starts': 0, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 1, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'prompts': 'Tell me a joke'}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-2
{'action': 'on_llm_start', 'name': 'OpenAI', 'step': 3, 'starts': 2, 'ends': 1, 'errors': 0, 'text_ctr': 0, 'chain_starts': 0, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 1, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'prompts': 'Tell me a poem'} {'action': 'on_llm_start', 'name': 'OpenAI', 'step': 3, 'starts': 2, 'ends': 1, 'errors': 0, 'text_ctr': 0, 'chain_starts': 0, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 1, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'prompts': 'Tell me a joke'} {'action': 'on_llm_start', 'name': 'OpenAI', 'step': 3, 'starts': 2, 'ends': 1, 'errors': 0, 'text_ctr': 0, 'chain_starts': 0, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 1, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'prompts': 'Tell me a poem'}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-3
{'action': 'on_llm_start', 'name': 'OpenAI', 'step': 3, 'starts': 2, 'ends': 1, 'errors': 0, 'text_ctr': 0, 'chain_starts': 0, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 1, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'prompts': 'Tell me a joke'} {'action': 'on_llm_start', 'name': 'OpenAI', 'step': 3, 'starts': 2, 'ends': 1, 'errors': 0, 'text_ctr': 0, 'chain_starts': 0, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 1, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'prompts': 'Tell me a poem'}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-4
{'action': 'on_llm_end', 'token_usage_prompt_tokens': 24, 'token_usage_completion_tokens': 138, 'token_usage_total_tokens': 162, 'model_name': 'text-davinci-003', 'step': 4, 'starts': 2, 'ends': 2, 'errors': 0, 'text_ctr': 0, 'chain_starts': 0, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 2, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'text': '\n\nQ: What did the fish say when it hit the wall?\nA: Dam!', 'generation_info_finish_reason': 'stop', 'generation_info_logprobs': None, 'flesch_reading_ease': 109.04, 'flesch_kincaid_grade': 1.3, 'smog_index': 0.0, 'coleman_liau_index': -1.24, 'automated_readability_index': 0.3, 'dale_chall_readability_score': 5.5, 'difficult_words': 0, 'linsear_write_formula': 5.5, 'gunning_fog': 5.2, 'text_standard': '5th and 6th grade', 'fernandez_huerta': 133.58, 'szigriszt_pazos': 131.54, 'gutierrez_polini': 62.3, 'crawford': -0.2, 'gulpease_index': 79.8, 'osman': 116.91}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-5
{'action': 'on_llm_end', 'token_usage_prompt_tokens': 24, 'token_usage_completion_tokens': 138, 'token_usage_total_tokens': 162, 'model_name': 'text-davinci-003', 'step': 4, 'starts': 2, 'ends': 2, 'errors': 0, 'text_ctr': 0, 'chain_starts': 0, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 2, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'text': '\n\nRoses are red,\nViolets are blue,\nSugar is sweet,\nAnd so are you.', 'generation_info_finish_reason': 'stop', 'generation_info_logprobs': None, 'flesch_reading_ease': 83.66, 'flesch_kincaid_grade': 4.8, 'smog_index': 0.0, 'coleman_liau_index': 3.23, 'automated_readability_index': 3.9, 'dale_chall_readability_score': 6.71, 'difficult_words': 2, 'linsear_write_formula': 6.5, 'gunning_fog': 8.28, 'text_standard': '6th and 7th grade', 'fernandez_huerta': 115.58, 'szigriszt_pazos': 112.37, 'gutierrez_polini': 54.83, 'crawford': 1.4, 'gulpease_index': 72.1, 'osman': 100.17}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-6
{'action': 'on_llm_end', 'token_usage_prompt_tokens': 24, 'token_usage_completion_tokens': 138, 'token_usage_total_tokens': 162, 'model_name': 'text-davinci-003', 'step': 4, 'starts': 2, 'ends': 2, 'errors': 0, 'text_ctr': 0, 'chain_starts': 0, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 2, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'text': '\n\nQ: What did the fish say when it hit the wall?\nA: Dam!', 'generation_info_finish_reason': 'stop', 'generation_info_logprobs': None, 'flesch_reading_ease': 109.04, 'flesch_kincaid_grade': 1.3, 'smog_index': 0.0, 'coleman_liau_index': -1.24, 'automated_readability_index': 0.3, 'dale_chall_readability_score': 5.5, 'difficult_words': 0, 'linsear_write_formula': 5.5, 'gunning_fog': 5.2, 'text_standard': '5th and 6th grade', 'fernandez_huerta': 133.58, 'szigriszt_pazos': 131.54, 'gutierrez_polini': 62.3, 'crawford': -0.2, 'gulpease_index': 79.8, 'osman': 116.91}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-7
{'action': 'on_llm_end', 'token_usage_prompt_tokens': 24, 'token_usage_completion_tokens': 138, 'token_usage_total_tokens': 162, 'model_name': 'text-davinci-003', 'step': 4, 'starts': 2, 'ends': 2, 'errors': 0, 'text_ctr': 0, 'chain_starts': 0, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 2, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'text': '\n\nRoses are red,\nViolets are blue,\nSugar is sweet,\nAnd so are you.', 'generation_info_finish_reason': 'stop', 'generation_info_logprobs': None, 'flesch_reading_ease': 83.66, 'flesch_kincaid_grade': 4.8, 'smog_index': 0.0, 'coleman_liau_index': 3.23, 'automated_readability_index': 3.9, 'dale_chall_readability_score': 6.71, 'difficult_words': 2, 'linsear_write_formula': 6.5, 'gunning_fog': 8.28, 'text_standard': '6th and 7th grade', 'fernandez_huerta': 115.58, 'szigriszt_pazos': 112.37, 'gutierrez_polini': 54.83, 'crawford': 1.4, 'gulpease_index': 72.1, 'osman': 100.17}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-8
{'action': 'on_llm_end', 'token_usage_prompt_tokens': 24, 'token_usage_completion_tokens': 138, 'token_usage_total_tokens': 162, 'model_name': 'text-davinci-003', 'step': 4, 'starts': 2, 'ends': 2, 'errors': 0, 'text_ctr': 0, 'chain_starts': 0, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 2, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'text': '\n\nQ: What did the fish say when it hit the wall?\nA: Dam!', 'generation_info_finish_reason': 'stop', 'generation_info_logprobs': None, 'flesch_reading_ease': 109.04, 'flesch_kincaid_grade': 1.3, 'smog_index': 0.0, 'coleman_liau_index': -1.24, 'automated_readability_index': 0.3, 'dale_chall_readability_score': 5.5, 'difficult_words': 0, 'linsear_write_formula': 5.5, 'gunning_fog': 5.2, 'text_standard': '5th and 6th grade', 'fernandez_huerta': 133.58, 'szigriszt_pazos': 131.54, 'gutierrez_polini': 62.3, 'crawford': -0.2, 'gulpease_index': 79.8, 'osman': 116.91}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-9
{'action': 'on_llm_end', 'token_usage_prompt_tokens': 24, 'token_usage_completion_tokens': 138, 'token_usage_total_tokens': 162, 'model_name': 'text-davinci-003', 'step': 4, 'starts': 2, 'ends': 2, 'errors': 0, 'text_ctr': 0, 'chain_starts': 0, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 2, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'text': '\n\nRoses are red,\nViolets are blue,\nSugar is sweet,\nAnd so are you.', 'generation_info_finish_reason': 'stop', 'generation_info_logprobs': None, 'flesch_reading_ease': 83.66, 'flesch_kincaid_grade': 4.8, 'smog_index': 0.0, 'coleman_liau_index': 3.23, 'automated_readability_index': 3.9, 'dale_chall_readability_score': 6.71, 'difficult_words': 2, 'linsear_write_formula': 6.5, 'gunning_fog': 8.28, 'text_standard': '6th and 7th grade', 'fernandez_huerta': 115.58, 'szigriszt_pazos': 112.37, 'gutierrez_polini': 54.83, 'crawford': 1.4, 'gulpease_index': 72.1, 'osman': 100.17} {'action_records': action name step starts ends errors text_ctr chain_starts \
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-10
0 on_llm_start OpenAI 1 1 0 0 0 0 1 on_llm_start OpenAI 1 1 0 0 0 0 2 on_llm_start OpenAI 1 1 0 0 0 0 3 on_llm_start OpenAI 1 1 0 0 0 0 4 on_llm_start OpenAI 1 1 0 0 0 0 5 on_llm_start OpenAI 1 1 0 0 0 0 6 on_llm_end NaN 2 1 1 0 0 0 7 on_llm_end NaN 2 1 1 0 0 0 8 on_llm_end NaN 2 1 1 0 0 0 9 on_llm_end NaN 2 1 1 0 0 0 10 on_llm_end NaN 2 1 1 0 0 0 11 on_llm_end NaN 2 1 1 0 0 0 12 on_llm_start OpenAI 3 2 1 0 0 0 13 on_llm_start OpenAI 3 2 1 0 0 0
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-11
14 on_llm_start OpenAI 3 2 1 0 0 0 15 on_llm_start OpenAI 3 2 1 0 0 0 16 on_llm_start OpenAI 3 2 1 0 0 0 17 on_llm_start OpenAI 3 2 1 0 0 0 18 on_llm_end NaN 4 2 2 0 0 0 19 on_llm_end NaN 4 2 2 0 0 0 20 on_llm_end NaN 4 2 2 0 0 0 21 on_llm_end NaN 4 2 2 0 0 0 22 on_llm_end NaN 4 2 2 0 0 0 23 on_llm_end NaN 4 2 2 0 0 0 chain_ends llm_starts ... difficult_words linsear_write_formula \ 0 0 1 ... NaN NaN 1 0 1 ... NaN NaN 2 0 1 ... NaN NaN 3 0 1 ... NaN NaN 4 0 1 ... NaN NaN 5 0 1 ... NaN NaN
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-12
5 0 1 ... NaN NaN 6 0 1 ... 0.0 5.5 7 0 1 ... 2.0 6.5 8 0 1 ... 0.0 5.5 9 0 1 ... 2.0 6.5 10 0 1 ... 0.0 5.5 11 0 1 ... 2.0 6.5 12 0 2 ... NaN NaN 13 0 2 ... NaN NaN 14 0 2 ... NaN NaN 15 0 2 ... NaN NaN 16 0 2 ... NaN NaN 17 0 2 ... NaN NaN 18 0 2 ... 0.0 5.5 19 0 2 ... 2.0 6.5 20 0 2 ... 0.0 5.5 21 0 2 ... 2.0 6.5 22 0 2 ... 0.0 5.5 23 0 2 ... 2.0 6.5 gunning_fog text_standard fernandez_huerta szigriszt_pazos \ 0 NaN NaN NaN NaN
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-13
0 NaN NaN NaN NaN 1 NaN NaN NaN NaN 2 NaN NaN NaN NaN 3 NaN NaN NaN NaN 4 NaN NaN NaN NaN 5 NaN NaN NaN NaN 6 5.20 5th and 6th grade 133.58 131.54 7 8.28 6th and 7th grade 115.58 112.37 8 5.20 5th and 6th grade 133.58 131.54 9 8.28 6th and 7th grade 115.58 112.37 10 5.20 5th and 6th grade 133.58 131.54 11 8.28 6th and 7th grade 115.58 112.37 12 NaN NaN NaN NaN 13 NaN NaN NaN NaN 14 NaN NaN NaN NaN 15 NaN NaN NaN NaN 16 NaN NaN NaN NaN 17 NaN NaN NaN NaN 18 5.20 5th and 6th grade 133.58 131.54 19 8.28 6th and 7th grade 115.58 112.37 20 5.20 5th and 6th grade 133.58 131.54
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-14
21 8.28 6th and 7th grade 115.58 112.37 22 5.20 5th and 6th grade 133.58 131.54 23 8.28 6th and 7th grade 115.58 112.37 gutierrez_polini crawford gulpease_index osman 0 NaN NaN NaN NaN 1 NaN NaN NaN NaN 2 NaN NaN NaN NaN 3 NaN NaN NaN NaN 4 NaN NaN NaN NaN 5 NaN NaN NaN NaN 6 62.30 -0.2 79.8 116.91 7 54.83 1.4 72.1 100.17 8 62.30 -0.2 79.8 116.91 9 54.83 1.4 72.1 100.17 10 62.30 -0.2 79.8 116.91 11 54.83 1.4 72.1 100.17 12 NaN NaN NaN NaN 13 NaN NaN NaN NaN 14 NaN NaN NaN NaN 15 NaN NaN NaN NaN 16 NaN NaN NaN NaN 17 NaN NaN NaN NaN 18 62.30 -0.2 79.8 116.91
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-15
19 54.83 1.4 72.1 100.17 20 62.30 -0.2 79.8 116.91 21 54.83 1.4 72.1 100.17 22 62.30 -0.2 79.8 116.91 23 54.83 1.4 72.1 100.17 [24 rows x 39 columns], 'session_analysis': prompt_step prompts name output_step \ 0 1 Tell me a joke OpenAI 2 1 1 Tell me a poem OpenAI 2 2 1 Tell me a joke OpenAI 2 3 1 Tell me a poem OpenAI 2 4 1 Tell me a joke OpenAI 2 5 1 Tell me a poem OpenAI 2 6 3 Tell me a joke OpenAI 4 7 3 Tell me a poem OpenAI 4 8 3 Tell me a joke OpenAI 4 9 3 Tell me a poem OpenAI 4 10 3 Tell me a joke OpenAI 4 11 3 Tell me a poem OpenAI 4 output \ 0 \n\nQ: What did the fish say when it hit the w... 1 \n\nRoses are red,\nViolets are blue,\nSugar i...
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-16
2 \n\nQ: What did the fish say when it hit the w... 3 \n\nRoses are red,\nViolets are blue,\nSugar i... 4 \n\nQ: What did the fish say when it hit the w... 5 \n\nRoses are red,\nViolets are blue,\nSugar i... 6 \n\nQ: What did the fish say when it hit the w... 7 \n\nRoses are red,\nViolets are blue,\nSugar i... 8 \n\nQ: What did the fish say when it hit the w... 9 \n\nRoses are red,\nViolets are blue,\nSugar i... 10 \n\nQ: What did the fish say when it hit the w... 11 \n\nRoses are red,\nViolets are blue,\nSugar i... token_usage_total_tokens token_usage_prompt_tokens \ 0 162 24 1 162 24 2 162 24 3 162 24 4 162 24 5 162 24 6 162 24 7 162 24 8 162 24 9 162 24 10 162 24 11 162 24 token_usage_completion_tokens flesch_reading_ease flesch_kincaid_grade \ 0 138 109.04 1.3 1 138 83.66 4.8
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-17
1 138 83.66 4.8 2 138 109.04 1.3 3 138 83.66 4.8 4 138 109.04 1.3 5 138 83.66 4.8 6 138 109.04 1.3 7 138 83.66 4.8 8 138 109.04 1.3 9 138 83.66 4.8 10 138 109.04 1.3 11 138 83.66 4.8 ... difficult_words linsear_write_formula gunning_fog \ 0 ... 0 5.5 5.20 1 ... 2 6.5 8.28 2 ... 0 5.5 5.20 3 ... 2 6.5 8.28 4 ... 0 5.5 5.20 5 ... 2 6.5 8.28 6 ... 0 5.5 5.20 7 ... 2 6.5 8.28 8 ... 0 5.5 5.20 9 ... 2 6.5 8.28 10 ... 0 5.5 5.20
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-18
10 ... 0 5.5 5.20 11 ... 2 6.5 8.28 text_standard fernandez_huerta szigriszt_pazos gutierrez_polini \ 0 5th and 6th grade 133.58 131.54 62.30 1 6th and 7th grade 115.58 112.37 54.83 2 5th and 6th grade 133.58 131.54 62.30 3 6th and 7th grade 115.58 112.37 54.83 4 5th and 6th grade 133.58 131.54 62.30 5 6th and 7th grade 115.58 112.37 54.83 6 5th and 6th grade 133.58 131.54 62.30 7 6th and 7th grade 115.58 112.37 54.83 8 5th and 6th grade 133.58 131.54 62.30 9 6th and 7th grade 115.58 112.37 54.83 10 5th and 6th grade 133.58 131.54 62.30 11 6th and 7th grade 115.58 112.37 54.83 crawford gulpease_index osman
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-19
crawford gulpease_index osman 0 -0.2 79.8 116.91 1 1.4 72.1 100.17 2 -0.2 79.8 116.91 3 1.4 72.1 100.17 4 -0.2 79.8 116.91 5 1.4 72.1 100.17 6 -0.2 79.8 116.91 7 1.4 72.1 100.17 8 -0.2 79.8 116.91 9 1.4 72.1 100.17 10 -0.2 79.8 116.91 11 1.4 72.1 100.17 [12 rows x 24 columns]} 2023-03-29 14:00:25,948 - clearml.Task - INFO - Completed model upload to https://files.clear.ml/langchain_callback_demo/llm.988bd727b0e94a29a3ac0ee526813545/models/simple_sequential At this point you can already go to https://app.clear.ml and take a look at the resulting ClearML Task that was created. Among others, you should see that this notebook is saved along with any git information. The model JSON that contains the used parameters is saved as an artifact, there are also console logs and under the plots section, you’ll find tables that represent the flow of the chain. Finally, if you enabled visualizations, these are stored as HTML files under debug samples.
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-20
Finally, if you enabled visualizations, these are stored as HTML files under debug samples. Scenario 2: Creating a agent with tools# To show a more advanced workflow, let’s create an agent with access to tools. The way ClearML tracks the results is not different though, only the table will look slightly different as there are other types of actions taken when compared to the earlier, simpler example. You can now also see the use of the finish=True keyword, which will fully close the ClearML Task, instead of just resetting the parameters and prompts for a new conversation. from langchain.agents import initialize_agent, load_tools from langchain.agents import AgentType # SCENARIO 2 - Agent with Tools tools = load_tools(["serpapi", "llm-math"], llm=llm, callback_manager=manager) agent = initialize_agent( tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, callback_manager=manager, verbose=True, ) agent.run( "Who is the wife of the person who sang summer of 69?" ) clearml_callback.flush_tracker(langchain_asset=agent, name="Agent with Tools", finish=True) > Entering new AgentExecutor chain... {'action': 'on_chain_start', 'name': 'AgentExecutor', 'step': 1, 'starts': 1, 'ends': 0, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 0, 'llm_ends': 0, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'input': 'Who is the wife of the person who sang summer of 69?'}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-21
{'action': 'on_llm_start', 'name': 'OpenAI', 'step': 2, 'starts': 2, 'ends': 0, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 1, 'llm_ends': 0, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'prompts': 'Answer the following questions as best you can. You have access to the following tools:\n\nSearch: A search engine. Useful for when you need to answer questions about current events. Input should be a search query.\nCalculator: Useful for when you need to answer questions about math.\n\nUse the following format:\n\nQuestion: the input question you must answer\nThought: you should always think about what to do\nAction: the action to take, should be one of [Search, Calculator]\nAction Input: the input to the action\nObservation: the result of the action\n... (this Thought/Action/Action Input/Observation can repeat N times)\nThought: I now know the final answer\nFinal Answer: the final answer to the original input question\n\nBegin!\n\nQuestion: Who is the wife of the person who sang summer of 69?\nThought:'}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-22
{'action': 'on_llm_end', 'token_usage_prompt_tokens': 189, 'token_usage_completion_tokens': 34, 'token_usage_total_tokens': 223, 'model_name': 'text-davinci-003', 'step': 3, 'starts': 2, 'ends': 1, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 1, 'llm_ends': 1, 'llm_streams': 0, 'tool_starts': 0, 'tool_ends': 0, 'agent_ends': 0, 'text': ' I need to find out who sang summer of 69 and then find out who their wife is.\nAction: Search\nAction Input: "Who sang summer of 69"', 'generation_info_finish_reason': 'stop', 'generation_info_logprobs': None, 'flesch_reading_ease': 91.61, 'flesch_kincaid_grade': 3.8, 'smog_index': 0.0, 'coleman_liau_index': 3.41, 'automated_readability_index': 3.5, 'dale_chall_readability_score': 6.06, 'difficult_words': 2, 'linsear_write_formula': 5.75, 'gunning_fog': 5.4, 'text_standard': '3rd and 4th grade', 'fernandez_huerta': 121.07, 'szigriszt_pazos': 119.5, 'gutierrez_polini': 54.91, 'crawford': 0.9, 'gulpease_index': 72.7, 'osman': 92.16}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-23
I need to find out who sang summer of 69 and then find out who their wife is. Action: Search Action Input: "Who sang summer of 69"{'action': 'on_agent_action', 'tool': 'Search', 'tool_input': 'Who sang summer of 69', 'log': ' I need to find out who sang summer of 69 and then find out who their wife is.\nAction: Search\nAction Input: "Who sang summer of 69"', 'step': 4, 'starts': 3, 'ends': 1, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 1, 'llm_ends': 1, 'llm_streams': 0, 'tool_starts': 1, 'tool_ends': 0, 'agent_ends': 0} {'action': 'on_tool_start', 'input_str': 'Who sang summer of 69', 'name': 'Search', 'description': 'A search engine. Useful for when you need to answer questions about current events. Input should be a search query.', 'step': 5, 'starts': 4, 'ends': 1, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 1, 'llm_ends': 1, 'llm_streams': 0, 'tool_starts': 2, 'tool_ends': 0, 'agent_ends': 0} Observation: Bryan Adams - Summer Of 69 (Official Music Video).
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-24
Observation: Bryan Adams - Summer Of 69 (Official Music Video). Thought:{'action': 'on_tool_end', 'output': 'Bryan Adams - Summer Of 69 (Official Music Video).', 'step': 6, 'starts': 4, 'ends': 2, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 1, 'llm_ends': 1, 'llm_streams': 0, 'tool_starts': 2, 'tool_ends': 1, 'agent_ends': 0}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-25
{'action': 'on_llm_start', 'name': 'OpenAI', 'step': 7, 'starts': 5, 'ends': 2, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 1, 'llm_streams': 0, 'tool_starts': 2, 'tool_ends': 1, 'agent_ends': 0, 'prompts': 'Answer the following questions as best you can. You have access to the following tools:\n\nSearch: A search engine. Useful for when you need to answer questions about current events. Input should be a search query.\nCalculator: Useful for when you need to answer questions about math.\n\nUse the following format:\n\nQuestion: the input question you must answer\nThought: you should always think about what to do\nAction: the action to take, should be one of [Search, Calculator]\nAction Input: the input to the action\nObservation: the result of the action\n... (this Thought/Action/Action Input/Observation can repeat N times)\nThought: I now know the final answer\nFinal Answer: the final answer to the original input question\n\nBegin!\n\nQuestion: Who is the wife of the person who sang summer of 69?\nThought: I need to find out who sang summer of 69 and then find out who their wife is.\nAction: Search\nAction Input: "Who sang summer of 69"\nObservation: Bryan Adams - Summer Of 69 (Official Music Video).\nThought:'}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-26
{'action': 'on_llm_end', 'token_usage_prompt_tokens': 242, 'token_usage_completion_tokens': 28, 'token_usage_total_tokens': 270, 'model_name': 'text-davinci-003', 'step': 8, 'starts': 5, 'ends': 3, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 2, 'llm_streams': 0, 'tool_starts': 2, 'tool_ends': 1, 'agent_ends': 0, 'text': ' I need to find out who Bryan Adams is married to.\nAction: Search\nAction Input: "Who is Bryan Adams married to"', 'generation_info_finish_reason': 'stop', 'generation_info_logprobs': None, 'flesch_reading_ease': 94.66, 'flesch_kincaid_grade': 2.7, 'smog_index': 0.0, 'coleman_liau_index': 4.73, 'automated_readability_index': 4.0, 'dale_chall_readability_score': 7.16, 'difficult_words': 2, 'linsear_write_formula': 4.25, 'gunning_fog': 4.2, 'text_standard': '4th and 5th grade', 'fernandez_huerta': 124.13, 'szigriszt_pazos': 119.2, 'gutierrez_polini': 52.26, 'crawford': 0.7, 'gulpease_index': 74.7, 'osman': 84.2} I need to find out who Bryan Adams is married to. Action: Search
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-27
I need to find out who Bryan Adams is married to. Action: Search Action Input: "Who is Bryan Adams married to"{'action': 'on_agent_action', 'tool': 'Search', 'tool_input': 'Who is Bryan Adams married to', 'log': ' I need to find out who Bryan Adams is married to.\nAction: Search\nAction Input: "Who is Bryan Adams married to"', 'step': 9, 'starts': 6, 'ends': 3, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 2, 'llm_streams': 0, 'tool_starts': 3, 'tool_ends': 1, 'agent_ends': 0} {'action': 'on_tool_start', 'input_str': 'Who is Bryan Adams married to', 'name': 'Search', 'description': 'A search engine. Useful for when you need to answer questions about current events. Input should be a search query.', 'step': 10, 'starts': 7, 'ends': 3, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 2, 'llm_streams': 0, 'tool_starts': 4, 'tool_ends': 1, 'agent_ends': 0} Observation: Bryan Adams has never married. In the 1990s, he was in a relationship with Danish model Cecilie Thomsen. In 2011, Bryan and Alicia Grimaldi, his ...
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-28
Thought:{'action': 'on_tool_end', 'output': 'Bryan Adams has never married. In the 1990s, he was in a relationship with Danish model Cecilie Thomsen. In 2011, Bryan and Alicia Grimaldi, his ...', 'step': 11, 'starts': 7, 'ends': 4, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 2, 'llm_ends': 2, 'llm_streams': 0, 'tool_starts': 4, 'tool_ends': 2, 'agent_ends': 0}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-29
{'action': 'on_llm_start', 'name': 'OpenAI', 'step': 12, 'starts': 8, 'ends': 4, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 3, 'llm_ends': 2, 'llm_streams': 0, 'tool_starts': 4, 'tool_ends': 2, 'agent_ends': 0, 'prompts': 'Answer the following questions as best you can. You have access to the following tools:\n\nSearch: A search engine. Useful for when you need to answer questions about current events. Input should be a search query.\nCalculator: Useful for when you need to answer questions about math.\n\nUse the following format:\n\nQuestion: the input question you must answer\nThought: you should always think about what to do\nAction: the action to take, should be one of [Search, Calculator]\nAction Input: the input to the action\nObservation: the result of the action\n... (this Thought/Action/Action Input/Observation can repeat N times)\nThought: I now know the final answer\nFinal Answer: the final answer to the original input question\n\nBegin!\n\nQuestion: Who
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-30
to the original input question\n\nBegin!\n\nQuestion: Who is the wife of the person who sang summer of 69?\nThought: I need to find out who sang summer of 69 and then find out who their wife is.\nAction: Search\nAction Input: "Who sang summer of 69"\nObservation: Bryan Adams - Summer Of 69 (Official Music Video).\nThought: I need to find out who Bryan Adams is married to.\nAction: Search\nAction Input: "Who is Bryan Adams married to"\nObservation: Bryan Adams has never married. In the 1990s, he was in a relationship with Danish model Cecilie Thomsen. In 2011, Bryan and Alicia Grimaldi, his ...\nThought:'}
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-31
{'action': 'on_llm_end', 'token_usage_prompt_tokens': 314, 'token_usage_completion_tokens': 18, 'token_usage_total_tokens': 332, 'model_name': 'text-davinci-003', 'step': 13, 'starts': 8, 'ends': 5, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 3, 'llm_ends': 3, 'llm_streams': 0, 'tool_starts': 4, 'tool_ends': 2, 'agent_ends': 0, 'text': ' I now know the final answer.\nFinal Answer: Bryan Adams has never been married.', 'generation_info_finish_reason': 'stop', 'generation_info_logprobs': None, 'flesch_reading_ease': 81.29, 'flesch_kincaid_grade': 3.7, 'smog_index': 0.0, 'coleman_liau_index': 5.75, 'automated_readability_index': 3.9, 'dale_chall_readability_score': 7.37, 'difficult_words': 1, 'linsear_write_formula': 2.5, 'gunning_fog': 2.8, 'text_standard': '3rd and 4th grade', 'fernandez_huerta': 115.7, 'szigriszt_pazos': 110.84, 'gutierrez_polini': 49.79, 'crawford': 0.7, 'gulpease_index': 85.4, 'osman': 83.14} I now know the final answer. Final Answer: Bryan Adams has never been married.
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-32
I now know the final answer. Final Answer: Bryan Adams has never been married. {'action': 'on_agent_finish', 'output': 'Bryan Adams has never been married.', 'log': ' I now know the final answer.\nFinal Answer: Bryan Adams has never been married.', 'step': 14, 'starts': 8, 'ends': 6, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 0, 'llm_starts': 3, 'llm_ends': 3, 'llm_streams': 0, 'tool_starts': 4, 'tool_ends': 2, 'agent_ends': 1} > Finished chain. {'action': 'on_chain_end', 'outputs': 'Bryan Adams has never been married.', 'step': 15, 'starts': 8, 'ends': 7, 'errors': 0, 'text_ctr': 0, 'chain_starts': 1, 'chain_ends': 1, 'llm_starts': 3, 'llm_ends': 3, 'llm_streams': 0, 'tool_starts': 4, 'tool_ends': 2, 'agent_ends': 1} {'action_records': action name step starts ends errors text_ctr \ 0 on_llm_start OpenAI 1 1 0 0 0 1 on_llm_start OpenAI 1 1 0 0 0 2 on_llm_start OpenAI 1 1 0 0 0 3 on_llm_start OpenAI 1 1 0 0 0
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-33
4 on_llm_start OpenAI 1 1 0 0 0 .. ... ... ... ... ... ... ... 66 on_tool_end NaN 11 7 4 0 0 67 on_llm_start OpenAI 12 8 4 0 0 68 on_llm_end NaN 13 8 5 0 0 69 on_agent_finish NaN 14 8 6 0 0 70 on_chain_end NaN 15 8 7 0 0 chain_starts chain_ends llm_starts ... gulpease_index osman input \ 0 0 0 1 ... NaN NaN NaN 1 0 0 1 ... NaN NaN NaN 2 0 0 1 ... NaN NaN NaN 3 0 0 1 ... NaN NaN NaN 4 0 0 1 ... NaN NaN NaN .. ... ... ... ... ... ... ... 66 1 0 2 ... NaN NaN NaN 67 1 0 3 ... NaN NaN NaN 68 1 0 3 ... 85.4 83.14 NaN 69 1 0 3 ... NaN NaN NaN
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-34
69 1 0 3 ... NaN NaN NaN 70 1 1 3 ... NaN NaN NaN tool tool_input log \ 0 NaN NaN NaN 1 NaN NaN NaN 2 NaN NaN NaN 3 NaN NaN NaN 4 NaN NaN NaN .. ... ... ... 66 NaN NaN NaN 67 NaN NaN NaN 68 NaN NaN NaN 69 NaN NaN I now know the final answer.\nFinal Answer: B... 70 NaN NaN NaN input_str description output \ 0 NaN NaN NaN 1 NaN NaN NaN 2 NaN NaN NaN 3 NaN NaN NaN 4 NaN NaN NaN .. ... ... ... 66 NaN NaN Bryan Adams has never married. In the 1990s, h... 67 NaN NaN NaN 68 NaN NaN NaN 69 NaN NaN Bryan Adams has never been married. 70 NaN NaN NaN outputs 0 NaN 1 NaN 2 NaN 3 NaN 4 NaN .. ... 66 NaN 67 NaN 68 NaN 69 NaN 70 Bryan Adams has never been married. [71 rows x 47 columns], 'session_analysis': prompt_step prompts name \ 0 2 Answer the following questions as best you can... OpenAI
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-35
0 2 Answer the following questions as best you can... OpenAI 1 7 Answer the following questions as best you can... OpenAI 2 12 Answer the following questions as best you can... OpenAI output_step output \ 0 3 I need to find out who sang summer of 69 and ... 1 8 I need to find out who Bryan Adams is married... 2 13 I now know the final answer.\nFinal Answer: B... token_usage_total_tokens token_usage_prompt_tokens \ 0 223 189 1 270 242 2 332 314 token_usage_completion_tokens flesch_reading_ease flesch_kincaid_grade \ 0 34 91.61 3.8 1 28 94.66 2.7 2 18 81.29 3.7 ... difficult_words linsear_write_formula gunning_fog \ 0 ... 2 5.75 5.4 1 ... 2 4.25 4.2 2 ... 1 2.50 2.8 text_standard fernandez_huerta szigriszt_pazos gutierrez_polini \ 0 3rd and 4th grade 121.07 119.50 54.91 1 4th and 5th grade 124.13 119.20 52.26
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
c0cf5ef15ac2-36
2 3rd and 4th grade 115.70 110.84 49.79 crawford gulpease_index osman 0 0.9 72.7 92.16 1 0.7 74.7 84.20 2 0.7 85.4 83.14 [3 rows x 24 columns]} Could not update last created model in Task 988bd727b0e94a29a3ac0ee526813545, Task status 'completed' cannot be updated Tips and Next Steps# Make sure you always use a unique name argument for the clearml_callback.flush_tracker function. If not, the model parameters used for a run will override the previous run! If you close the ClearML Callback using clearml_callback.flush_tracker(..., finish=True) the Callback cannot be used anymore. Make a new one if you want to keep logging. Check out the rest of the open source ClearML ecosystem, there is a data version manager, a remote execution agent, automated pipelines and much more! previous Chroma next Cohere Contents ClearML Integration Getting API Credentials Setting Up Scenario 1: Just an LLM Scenario 2: Creating a agent with tools Tips and Next Steps By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/clearml_tracking.html" }
d78cc1d444c6-0
.md .pdf Graphsignal Contents Installation and Setup Tracing and Monitoring Graphsignal# This page covers how to use the Graphsignal ecosystem to trace and monitor LangChain. Installation and Setup# Install the Python library with pip install graphsignal Create free Graphsignal account here Get an API key and set it as an environment variable (GRAPHSIGNAL_API_KEY) Tracing and Monitoring# Graphsignal automatically instruments and starts tracing and monitoring chains. Traces, metrics and errors are then available in your Graphsignal dashboard. No prompts or other sensitive data are sent to Graphsignal cloud, only statistics and metadata. Initialize the tracer by providing a deployment name: import graphsignal graphsignal.configure(deployment='my-langchain-app-prod') In order to trace full runs and see a breakdown by chains and tools, you can wrap the calling routine or use a decorator: with graphsignal.start_trace('my-chain'): chain.run("some initial text") Optionally, enable profiling to record function-level statistics for each trace. with graphsignal.start_trace( 'my-chain', options=graphsignal.TraceOptions(enable_profiling=True)): chain.run("some initial text") See the Quick Start guide for complete setup instructions. previous GPT4All next Hazy Research Contents Installation and Setup Tracing and Monitoring By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/graphsignal.html" }
70e267fda1e8-0
.md .pdf Cohere Contents Installation and Setup Wrappers LLM Embeddings Cohere# This page covers how to use the Cohere ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Cohere wrappers. Installation and Setup# Install the Python SDK with pip install cohere Get an Cohere api key and set it as an environment variable (COHERE_API_KEY) Wrappers# LLM# There exists an Cohere LLM wrapper, which you can access with from langchain.llms import Cohere Embeddings# There exists an Cohere Embeddings wrapper, which you can access with from langchain.embeddings import CohereEmbeddings For a more detailed walkthrough of this, see this notebook previous ClearML Integration next DeepInfra Contents Installation and Setup Wrappers LLM Embeddings By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/cohere.html" }
d7e59d6c3514-0
.md .pdf Milvus Contents Installation and Setup Wrappers VectorStore Milvus# This page covers how to use the Milvus ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Milvus wrappers. Installation and Setup# Install the Python SDK with pip install pymilvus Wrappers# VectorStore# There exists a wrapper around Milvus indexes, allowing you to use it as a vectorstore, whether for semantic search or example selection. To import this vectorstore: from langchain.vectorstores import Milvus For a more detailed walkthrough of the Miluvs wrapper, see this notebook previous Llama.cpp next Modal Contents Installation and Setup Wrappers VectorStore By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/milvus.html" }
7a4e9581bbf6-0
.md .pdf Hazy Research Contents Installation and Setup Wrappers LLM Hazy Research# This page covers how to use the Hazy Research ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Hazy Research wrappers. Installation and Setup# To use the manifest, install it with pip install manifest-ml Wrappers# LLM# There exists an LLM wrapper around Hazy Research’s manifest library. manifest is a python library which is itself a wrapper around many model providers, and adds in caching, history, and more. To use this wrapper: from langchain.llms.manifest import ManifestWrapper previous Graphsignal next Helicone Contents Installation and Setup Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/hazy_research.html" }
dfb5fa130232-0
.md .pdf Qdrant Contents Installation and Setup Wrappers VectorStore Qdrant# This page covers how to use the Qdrant ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Qdrant wrappers. Installation and Setup# Install the Python SDK with pip install qdrant-client Wrappers# VectorStore# There exists a wrapper around Qdrant indexes, allowing you to use it as a vectorstore, whether for semantic search or example selection. To import this vectorstore: from langchain.vectorstores import Qdrant For a more detailed walkthrough of the Qdrant wrapper, see this notebook previous PromptLayer next Replicate Contents Installation and Setup Wrappers VectorStore By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/qdrant.html" }
ff63c9f9023a-0
.md .pdf Jina Contents Installation and Setup Wrappers Embeddings Jina# This page covers how to use the Jina ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Jina wrappers. Installation and Setup# Install the Python SDK with pip install jina Get a Jina AI Cloud auth token from here and set it as an environment variable (JINA_AUTH_TOKEN) Wrappers# Embeddings# There exists a Jina Embeddings wrapper, which you can access with from langchain.embeddings import JinaEmbeddings For a more detailed walkthrough of this, see this notebook previous Hugging Face next Llama.cpp Contents Installation and Setup Wrappers Embeddings By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/jina.html" }
2b2b0ca81671-0
.md .pdf Apify Contents Overview Installation and Setup Wrappers Utility Loader Apify# This page covers how to use Apify within LangChain. Overview# Apify is a cloud platform for web scraping and data extraction, which provides an ecosystem of more than a thousand ready-made apps called Actors for various scraping, crawling, and extraction use cases. This integration enables you run Actors on the Apify platform and load their results into LangChain to feed your vector indexes with documents and data from the web, e.g. to generate answers from websites with documentation, blogs, or knowledge bases. Installation and Setup# Install the Apify API client for Python with pip install apify-client Get your Apify API token and either set it as an environment variable (APIFY_API_TOKEN) or pass it to the ApifyWrapper as apify_api_token in the constructor. Wrappers# Utility# You can use the ApifyWrapper to run Actors on the Apify platform. from langchain.utilities import ApifyWrapper For a more detailed walkthrough of this wrapper, see this notebook. Loader# You can also use our ApifyDatasetLoader to get data from Apify dataset. from langchain.document_loaders import ApifyDatasetLoader For a more detailed walkthrough of this loader, see this notebook. previous Aim next AtlasDB Contents Overview Installation and Setup Wrappers Utility Loader By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/apify.html" }
094b82a097d8-0
.md .pdf Replicate Contents Installation and Setup Calling a model Replicate# This page covers how to run models on Replicate within LangChain. Installation and Setup# Create a Replicate account. Get your API key and set it as an environment variable (REPLICATE_API_TOKEN) Install the Replicate python client with pip install replicate Calling a model# Find a model on the Replicate explore page, and then paste in the model name and version in this format: owner-name/model-name:version For example, for this flan-t5 model, click on the API tab. The model name/version would be: daanelson/flan-t5:04e422a9b85baed86a4f24981d7f9953e20c5fd82f6103b74ebc431588e1cec8 Only the model param is required, but any other model parameters can also be passed in with the format input={model_param: value, ...} For example, if we were running stable diffusion and wanted to change the image dimensions: Replicate(model="stability-ai/stable-diffusion:db21e45d3f7023abc2a46ee38a23973f6dce16bb082a930b0c49861f96d1e5bf", input={'image_dimensions': '512x512'}) Note that only the first output of a model will be returned. From here, we can initialize our model: llm = Replicate(model="daanelson/flan-t5:04e422a9b85baed86a4f24981d7f9953e20c5fd82f6103b74ebc431588e1cec8") And run it: prompt = """ Answer the following yes/no question by reasoning step by step.
{ "url": "https://python.langchain.com/en/latest/ecosystem/replicate.html" }
094b82a097d8-1
prompt = """ Answer the following yes/no question by reasoning step by step. Can a dog drive a car? """ llm(prompt) We can call any Replicate model (not just LLMs) using this syntax. For example, we can call Stable Diffusion: text2image = Replicate(model="stability-ai/stable-diffusion:db21e45d3f7023abc2a46ee38a23973f6dce16bb082a930b0c49861f96d1e5bf", input={'image_dimensions'='512x512'} image_output = text2image("A cat riding a motorcycle by Picasso") previous Qdrant next Runhouse Contents Installation and Setup Calling a model By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/replicate.html" }
84c348714c08-0
.md .pdf Weaviate Contents Installation and Setup Wrappers VectorStore Weaviate# This page covers how to use the Weaviate ecosystem within LangChain. What is Weaviate? Weaviate in a nutshell: Weaviate is an open-source ​database of the type ​vector search engine. Weaviate allows you to store JSON documents in a class property-like fashion while attaching machine learning vectors to these documents to represent them in vector space. Weaviate can be used stand-alone (aka bring your vectors) or with a variety of modules that can do the vectorization for you and extend the core capabilities. Weaviate has a GraphQL-API to access your data easily. We aim to bring your vector search set up to production to query in mere milliseconds (check our open source benchmarks to see if Weaviate fits your use case). Get to know Weaviate in the basics getting started guide in under five minutes. Weaviate in detail: Weaviate is a low-latency vector search engine with out-of-the-box support for different media types (text, images, etc.). It offers Semantic Search, Question-Answer Extraction, Classification, Customizable Models (PyTorch/TensorFlow/Keras), etc. Built from scratch in Go, Weaviate stores both objects and vectors, allowing for combining vector search with structured filtering and the fault tolerance of a cloud-native database. It is all accessible through GraphQL, REST, and various client-side programming languages. Installation and Setup# Install the Python SDK with pip install weaviate-client Wrappers# VectorStore# There exists a wrapper around Weaviate indexes, allowing you to use it as a vectorstore, whether for semantic search or example selection. To import this vectorstore: from langchain.vectorstores import Weaviate
{ "url": "https://python.langchain.com/en/latest/ecosystem/weaviate.html" }
84c348714c08-1
To import this vectorstore: from langchain.vectorstores import Weaviate For a more detailed walkthrough of the Weaviate wrapper, see this notebook previous Weights & Biases next Wolfram Alpha Wrapper Contents Installation and Setup Wrappers VectorStore By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/weaviate.html" }
737c2d45c58d-0
.md .pdf CerebriumAI Contents Installation and Setup Wrappers LLM CerebriumAI# This page covers how to use the CerebriumAI ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific CerebriumAI wrappers. Installation and Setup# Install with pip install cerebrium Get an CerebriumAI api key and set it as an environment variable (CEREBRIUMAI_API_KEY) Wrappers# LLM# There exists an CerebriumAI LLM wrapper, which you can access with from langchain.llms import CerebriumAI previous Banana next Chroma Contents Installation and Setup Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/cerebriumai.html" }
3d9997ed96a6-0
.md .pdf Runhouse Contents Installation and Setup Self-hosted LLMs Self-hosted Embeddings Runhouse# This page covers how to use the Runhouse ecosystem within LangChain. It is broken into three parts: installation and setup, LLMs, and Embeddings. Installation and Setup# Install the Python SDK with pip install runhouse If you’d like to use on-demand cluster, check your cloud credentials with sky check Self-hosted LLMs# For a basic self-hosted LLM, you can use the SelfHostedHuggingFaceLLM class. For more custom LLMs, you can use the SelfHostedPipeline parent class. from langchain.llms import SelfHostedPipeline, SelfHostedHuggingFaceLLM For a more detailed walkthrough of the Self-hosted LLMs, see this notebook Self-hosted Embeddings# There are several ways to use self-hosted embeddings with LangChain via Runhouse. For a basic self-hosted embedding from a Hugging Face Transformers model, you can use the SelfHostedEmbedding class. from langchain.llms import SelfHostedPipeline, SelfHostedHuggingFaceLLM For a more detailed walkthrough of the Self-hosted Embeddings, see this notebook previous Replicate next RWKV-4 Contents Installation and Setup Self-hosted LLMs Self-hosted Embeddings By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/runhouse.html" }
7a55857a880d-0
.md .pdf Writer Contents Installation and Setup Wrappers LLM Writer# This page covers how to use the Writer ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Writer wrappers. Installation and Setup# Get an Writer api key and set it as an environment variable (WRITER_API_KEY) Wrappers# LLM# There exists an Writer LLM wrapper, which you can access with from langchain.llms import Writer previous Wolfram Alpha Wrapper next Glossary Contents Installation and Setup Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/writer.html" }
93eadda20c38-0
.md .pdf Unstructured Contents Installation and Setup Wrappers Data Loaders Unstructured# This page covers how to use the unstructured ecosystem within LangChain. The unstructured package from Unstructured.IO extracts clean text from raw source documents like PDFs and Word documents. This page is broken into two parts: installation and setup, and then references to specific unstructured wrappers. Installation and Setup# Install the Python SDK with pip install "unstructured[local-inference]" Install the following system dependencies if they are not already available on your system. Depending on what document types you’re parsing, you may not need all of these. libmagic-dev (filetype detection) poppler-utils (images and PDFs) tesseract-ocr(images and PDFs) libreoffice (MS Office docs) pandoc (EPUBs) If you are parsing PDFs using the "hi_res" strategy, run the following to install the detectron2 model, which unstructured uses for layout detection: pip install "detectron2@git+https://github.com/facebookresearch/detectron2.git@e2ce8dc#egg=detectron2" If detectron2 is not installed, unstructured will fallback to processing PDFs using the "fast" strategy, which uses pdfminer directly and doesn’t require detectron2. Wrappers# Data Loaders# The primary unstructured wrappers within langchain are data loaders. The following shows how to use the most basic unstructured data loader. There are other file-specific data loaders available in the langchain.document_loaders module. from langchain.document_loaders import UnstructuredFileLoader loader = UnstructuredFileLoader("state_of_the_union.txt") loader.load()
{ "url": "https://python.langchain.com/en/latest/ecosystem/unstructured.html" }
93eadda20c38-1
loader = UnstructuredFileLoader("state_of_the_union.txt") loader.load() If you instantiate the loader with UnstructuredFileLoader(mode="elements"), the loader will track additional metadata like the page number and text type (i.e. title, narrative text) when that information is available. previous StochasticAI next Weights & Biases Contents Installation and Setup Wrappers Data Loaders By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/unstructured.html" }
0346423fe326-0
.md .pdf Helicone Contents What is Helicone? Quick start How to enable Helicone caching How to use Helicone custom properties Helicone# This page covers how to use the Helicone ecosystem within LangChain. What is Helicone?# Helicone is an open source observability platform that proxies your OpenAI traffic and provides you key insights into your spend, latency and usage. Quick start# With your LangChain environment you can just add the following parameter. export OPENAI_API_BASE="https://oai.hconeai.com/v1" Now head over to helicone.ai to create your account, and add your OpenAI API key within our dashboard to view your logs. How to enable Helicone caching# from langchain.llms import OpenAI import openai openai.api_base = "https://oai.hconeai.com/v1" llm = OpenAI(temperature=0.9, headers={"Helicone-Cache-Enabled": "true"}) text = "What is a helicone?" print(llm(text)) Helicone caching docs How to use Helicone custom properties# from langchain.llms import OpenAI import openai openai.api_base = "https://oai.hconeai.com/v1" llm = OpenAI(temperature=0.9, headers={ "Helicone-Property-Session": "24", "Helicone-Property-Conversation": "support_issue_2", "Helicone-Property-App": "mobile", }) text = "What is a helicone?" print(llm(text)) Helicone property docs previous Hazy Research next Hugging Face Contents What is Helicone? Quick start How to enable Helicone caching How to use Helicone custom properties
{ "url": "https://python.langchain.com/en/latest/ecosystem/helicone.html" }
0346423fe326-1
Quick start How to enable Helicone caching How to use Helicone custom properties By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/helicone.html" }
3537f8401cd5-0
.md .pdf AtlasDB Contents Installation and Setup Wrappers VectorStore AtlasDB# This page covers how to use Nomic’s Atlas ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Atlas wrappers. Installation and Setup# Install the Python package with pip install nomic Nomic is also included in langchains poetry extras poetry install -E all Wrappers# VectorStore# There exists a wrapper around the Atlas neural database, allowing you to use it as a vectorstore. This vectorstore also gives you full access to the underlying AtlasProject object, which will allow you to use the full range of Atlas map interactions, such as bulk tagging and automatic topic modeling. Please see the Atlas docs for more detailed information. To import this vectorstore: from langchain.vectorstores import AtlasDB For a more detailed walkthrough of the AtlasDB wrapper, see this notebook previous Apify next Banana Contents Installation and Setup Wrappers VectorStore By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/atlas.html" }
a684b376e6ea-0
.md .pdf Petals Contents Installation and Setup Wrappers LLM Petals# This page covers how to use the Petals ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Petals wrappers. Installation and Setup# Install with pip install petals Get a Hugging Face api key and set it as an environment variable (HUGGINGFACE_API_KEY) Wrappers# LLM# There exists an Petals LLM wrapper, which you can access with from langchain.llms import Petals previous OpenSearch next PGVector Contents Installation and Setup Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/petals.html" }
4024c679745f-0
.md .pdf ForefrontAI Contents Installation and Setup Wrappers LLM ForefrontAI# This page covers how to use the ForefrontAI ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific ForefrontAI wrappers. Installation and Setup# Get an ForefrontAI api key and set it as an environment variable (FOREFRONTAI_API_KEY) Wrappers# LLM# There exists an ForefrontAI LLM wrapper, which you can access with from langchain.llms import ForefrontAI previous Deep Lake next Google Search Wrapper Contents Installation and Setup Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/forefrontai.html" }
46265652e00a-0
.md .pdf PromptLayer Contents Installation and Setup Wrappers LLM PromptLayer# This page covers how to use PromptLayer within LangChain. It is broken into two parts: installation and setup, and then references to specific PromptLayer wrappers. Installation and Setup# If you want to work with PromptLayer: Install the promptlayer python library pip install promptlayer Create a PromptLayer account Create an api token and set it as an environment variable (PROMPTLAYER_API_KEY) Wrappers# LLM# There exists an PromptLayer OpenAI LLM wrapper, which you can access with from langchain.llms import PromptLayerOpenAI To tag your requests, use the argument pl_tags when instanializing the LLM from langchain.llms import PromptLayerOpenAI llm = PromptLayerOpenAI(pl_tags=["langchain-requests", "chatbot"]) To get the PromptLayer request id, use the argument return_pl_id when instanializing the LLM from langchain.llms import PromptLayerOpenAI llm = PromptLayerOpenAI(return_pl_id=True) This will add the PromptLayer request ID in the generation_info field of the Generation returned when using .generate or .agenerate For example: llm_results = llm.generate(["hello world"]) for res in llm_results.generations: print("pl request id: ", res[0].generation_info["pl_request_id"]) You can use the PromptLayer request ID to add a prompt, score, or other metadata to your request. Read more about it here. This LLM is identical to the OpenAI LLM, except that all your requests will be logged to your PromptLayer account you can add pl_tags when instantializing to tag your requests on PromptLayer
{ "url": "https://python.langchain.com/en/latest/ecosystem/promptlayer.html" }
46265652e00a-1
you can add pl_tags when instantializing to tag your requests on PromptLayer you can add return_pl_id when instantializing to return a PromptLayer request id to use while tracking requests. PromptLayer also provides native wrappers for PromptLayerChatOpenAI and PromptLayerOpenAIChat previous Pinecone next Qdrant Contents Installation and Setup Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/promptlayer.html" }
574a8d444afa-0
.md .pdf Chroma Contents Installation and Setup Wrappers VectorStore Chroma# This page covers how to use the Chroma ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Chroma wrappers. Installation and Setup# Install the Python package with pip install chromadb Wrappers# VectorStore# There exists a wrapper around Chroma vector databases, allowing you to use it as a vectorstore, whether for semantic search or example selection. To import this vectorstore: from langchain.vectorstores import Chroma For a more detailed walkthrough of the Chroma wrapper, see this notebook previous CerebriumAI next ClearML Integration Contents Installation and Setup Wrappers VectorStore By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/chroma.html" }
e9d140915dad-0
.md .pdf NLPCloud Contents Installation and Setup Wrappers LLM NLPCloud# This page covers how to use the NLPCloud ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific NLPCloud wrappers. Installation and Setup# Install the Python SDK with pip install nlpcloud Get an NLPCloud api key and set it as an environment variable (NLPCLOUD_API_KEY) Wrappers# LLM# There exists an NLPCloud LLM wrapper, which you can access with from langchain.llms import NLPCloud previous Modal next OpenAI Contents Installation and Setup Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/nlpcloud.html" }
92289032c107-0
.md .pdf Banana Contents Installation and Setup Define your Banana Template Build the Banana app Wrappers LLM Banana# This page covers how to use the Banana ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Banana wrappers. Installation and Setup# Install with pip install banana-dev Get an Banana api key and set it as an environment variable (BANANA_API_KEY) Define your Banana Template# If you want to use an available language model template you can find one here. This template uses the Palmyra-Base model by Writer. You can check out an example Banana repository here. Build the Banana app# Banana Apps must include the “output” key in the return json. There is a rigid response structure. # Return the results as a dictionary result = {'output': result} An example inference function would be: def inference(model_inputs:dict) -> dict: global model global tokenizer # Parse out your arguments prompt = model_inputs.get('prompt', None) if prompt == None: return {'message': "No prompt provided"} # Run the model input_ids = tokenizer.encode(prompt, return_tensors='pt').cuda() output = model.generate( input_ids, max_length=100, do_sample=True, top_k=50, top_p=0.95, num_return_sequences=1, temperature=0.9, early_stopping=True, no_repeat_ngram_size=3, num_beams=5, length_penalty=1.5, repetition_penalty=1.5, bad_words_ids=[[tokenizer.encode(' ', add_prefix_space=True)[0]]] )
{ "url": "https://python.langchain.com/en/latest/ecosystem/bananadev.html" }
92289032c107-1
bad_words_ids=[[tokenizer.encode(' ', add_prefix_space=True)[0]]] ) result = tokenizer.decode(output[0], skip_special_tokens=True) # Return the results as a dictionary result = {'output': result} return result You can find a full example of a Banana app here. Wrappers# LLM# There exists an Banana LLM wrapper, which you can access with from langchain.llms import Banana You need to provide a model key located in the dashboard: llm = Banana(model_key="YOUR_MODEL_KEY") previous AtlasDB next CerebriumAI Contents Installation and Setup Define your Banana Template Build the Banana app Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/bananadev.html" }
1bada699c024-0
.md .pdf SerpAPI Contents Installation and Setup Wrappers Utility Tool SerpAPI# This page covers how to use the SerpAPI search APIs within LangChain. It is broken into two parts: installation and setup, and then references to the specific SerpAPI wrapper. Installation and Setup# Install requirements with pip install google-search-results Get a SerpAPI api key and either set it as an environment variable (SERPAPI_API_KEY) Wrappers# Utility# There exists a SerpAPI utility which wraps this API. To import this utility: from langchain.utilities import SerpAPIWrapper For a more detailed walkthrough of this wrapper, see this notebook. Tool# You can also easily load this wrapper as a Tool (to use with an Agent). You can do this with: from langchain.agents import load_tools tools = load_tools(["serpapi"]) For more information on this, see this page previous SearxNG Search API next StochasticAI Contents Installation and Setup Wrappers Utility Tool By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/serpapi.html" }
a4501437101d-0
.md .pdf Wolfram Alpha Wrapper Contents Installation and Setup Wrappers Utility Tool Wolfram Alpha Wrapper# This page covers how to use the Wolfram Alpha API within LangChain. It is broken into two parts: installation and setup, and then references to specific Wolfram Alpha wrappers. Installation and Setup# Install requirements with pip install wolframalpha Go to wolfram alpha and sign up for a developer account here Create an app and get your APP ID Set your APP ID as an environment variable WOLFRAM_ALPHA_APPID Wrappers# Utility# There exists a WolframAlphaAPIWrapper utility which wraps this API. To import this utility: from langchain.utilities.wolfram_alpha import WolframAlphaAPIWrapper For a more detailed walkthrough of this wrapper, see this notebook. Tool# You can also easily load this wrapper as a Tool (to use with an Agent). You can do this with: from langchain.agents import load_tools tools = load_tools(["wolfram-alpha"]) For more information on this, see this page previous Weaviate next Writer Contents Installation and Setup Wrappers Utility Tool By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/wolfram_alpha.html" }
3d92a14ea9de-0
.md .pdf PGVector Contents Installation Setup Wrappers VectorStore Usage PGVector# This page covers how to use the Postgres PGVector ecosystem within LangChain It is broken into two parts: installation and setup, and then references to specific PGVector wrappers. Installation# Install the Python package with pip install pgvector Setup# The first step is to create a database with the pgvector extension installed. Follow the steps at PGVector Installation Steps to install the database and the extension. The docker image is the easiest way to get started. Wrappers# VectorStore# There exists a wrapper around Postgres vector databases, allowing you to use it as a vectorstore, whether for semantic search or example selection. To import this vectorstore: from langchain.vectorstores.pgvector import PGVector Usage# For a more detailed walkthrough of the PGVector Wrapper, see this notebook previous Petals next Pinecone Contents Installation Setup Wrappers VectorStore Usage By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/pgvector.html" }
71fe33e76172-0
.md .pdf OpenAI Contents Installation and Setup Wrappers LLM Embeddings Tokenizer Moderation OpenAI# This page covers how to use the OpenAI ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific OpenAI wrappers. Installation and Setup# Install the Python SDK with pip install openai Get an OpenAI api key and set it as an environment variable (OPENAI_API_KEY) If you want to use OpenAI’s tokenizer (only available for Python 3.9+), install it with pip install tiktoken Wrappers# LLM# There exists an OpenAI LLM wrapper, which you can access with from langchain.llms import OpenAI If you are using a model hosted on Azure, you should use different wrapper for that: from langchain.llms import AzureOpenAI For a more detailed walkthrough of the Azure wrapper, see this notebook Embeddings# There exists an OpenAI Embeddings wrapper, which you can access with from langchain.embeddings import OpenAIEmbeddings For a more detailed walkthrough of this, see this notebook Tokenizer# There are several places you can use the tiktoken tokenizer. By default, it is used to count tokens for OpenAI LLMs. You can also use it to count tokens when splitting documents with from langchain.text_splitter import CharacterTextSplitter CharacterTextSplitter.from_tiktoken_encoder(...) For a more detailed walkthrough of this, see this notebook Moderation# You can also access the OpenAI content moderation endpoint with from langchain.chains import OpenAIModerationChain For a more detailed walkthrough of this, see this notebook previous NLPCloud next OpenSearch Contents Installation and Setup Wrappers LLM Embeddings Tokenizer
{ "url": "https://python.langchain.com/en/latest/ecosystem/openai.html" }
71fe33e76172-1
Contents Installation and Setup Wrappers LLM Embeddings Tokenizer Moderation By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/openai.html" }
30d28c6dccd0-0
.md .pdf Hugging Face Contents Installation and Setup Wrappers LLM Embeddings Tokenizer Datasets Hugging Face# This page covers how to use the Hugging Face ecosystem (including the Hugging Face Hub) within LangChain. It is broken into two parts: installation and setup, and then references to specific Hugging Face wrappers. Installation and Setup# If you want to work with the Hugging Face Hub: Install the Hub client library with pip install huggingface_hub Create a Hugging Face account (it’s free!) Create an access token and set it as an environment variable (HUGGINGFACEHUB_API_TOKEN) If you want work with the Hugging Face Python libraries: Install pip install transformers for working with models and tokenizers Install pip install datasets for working with datasets Wrappers# LLM# There exists two Hugging Face LLM wrappers, one for a local pipeline and one for a model hosted on Hugging Face Hub. Note that these wrappers only work for models that support the following tasks: text2text-generation, text-generation To use the local pipeline wrapper: from langchain.llms import HuggingFacePipeline To use a the wrapper for a model hosted on Hugging Face Hub: from langchain.llms import HuggingFaceHub For a more detailed walkthrough of the Hugging Face Hub wrapper, see this notebook Embeddings# There exists two Hugging Face Embeddings wrappers, one for a local model and one for a model hosted on Hugging Face Hub. Note that these wrappers only work for sentence-transformers models. To use the local pipeline wrapper: from langchain.embeddings import HuggingFaceEmbeddings To use a the wrapper for a model hosted on Hugging Face Hub: from langchain.embeddings import HuggingFaceHubEmbeddings
{ "url": "https://python.langchain.com/en/latest/ecosystem/huggingface.html" }
30d28c6dccd0-1
from langchain.embeddings import HuggingFaceHubEmbeddings For a more detailed walkthrough of this, see this notebook Tokenizer# There are several places you can use tokenizers available through the transformers package. By default, it is used to count tokens for all LLMs. You can also use it to count tokens when splitting documents with from langchain.text_splitter import CharacterTextSplitter CharacterTextSplitter.from_huggingface_tokenizer(...) For a more detailed walkthrough of this, see this notebook Datasets# The Hugging Face Hub has lots of great datasets that can be used to evaluate your LLM chains. For a detailed walkthrough of how to use them to do so, see this notebook previous Helicone next Jina Contents Installation and Setup Wrappers LLM Embeddings Tokenizer Datasets By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/huggingface.html" }
f3eccae50487-0
.md .pdf Modal Contents Installation and Setup Define your Modal Functions and Webhooks Wrappers LLM Modal# This page covers how to use the Modal ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Modal wrappers. Installation and Setup# Install with pip install modal-client Run modal token new Define your Modal Functions and Webhooks# You must include a prompt. There is a rigid response structure. class Item(BaseModel): prompt: str @stub.webhook(method="POST") def my_webhook(item: Item): return {"prompt": my_function.call(item.prompt)} An example with GPT2: from pydantic import BaseModel import modal stub = modal.Stub("example-get-started") volume = modal.SharedVolume().persist("gpt2_model_vol") CACHE_PATH = "/root/model_cache" @stub.function( gpu="any", image=modal.Image.debian_slim().pip_install( "tokenizers", "transformers", "torch", "accelerate" ), shared_volumes={CACHE_PATH: volume}, retries=3, ) def run_gpt2(text: str): from transformers import GPT2Tokenizer, GPT2LMHeadModel tokenizer = GPT2Tokenizer.from_pretrained('gpt2') model = GPT2LMHeadModel.from_pretrained('gpt2') encoded_input = tokenizer(text, return_tensors='pt').input_ids output = model.generate(encoded_input, max_length=50, do_sample=True) return tokenizer.decode(output[0], skip_special_tokens=True) class Item(BaseModel): prompt: str @stub.webhook(method="POST") def get_text(item: Item):
{ "url": "https://python.langchain.com/en/latest/ecosystem/modal.html" }
f3eccae50487-1
@stub.webhook(method="POST") def get_text(item: Item): return {"prompt": run_gpt2.call(item.prompt)} Wrappers# LLM# There exists an Modal LLM wrapper, which you can access with from langchain.llms import Modal previous Milvus next NLPCloud Contents Installation and Setup Define your Modal Functions and Webhooks Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/modal.html" }
e0efe6dfcec3-0
.md .pdf StochasticAI Contents Installation and Setup Wrappers LLM StochasticAI# This page covers how to use the StochasticAI ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific StochasticAI wrappers. Installation and Setup# Install with pip install stochasticx Get an StochasticAI api key and set it as an environment variable (STOCHASTICAI_API_KEY) Wrappers# LLM# There exists an StochasticAI LLM wrapper, which you can access with from langchain.llms import StochasticAI previous SerpAPI next Unstructured Contents Installation and Setup Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/stochasticai.html" }
da104188242f-0
.md .pdf Google Search Wrapper Contents Installation and Setup Wrappers Utility Tool Google Search Wrapper# This page covers how to use the Google Search API within LangChain. It is broken into two parts: installation and setup, and then references to the specific Google Search wrapper. Installation and Setup# Install requirements with pip install google-api-python-client Set up a Custom Search Engine, following these instructions Get an API Key and Custom Search Engine ID from the previous step, and set them as environment variables GOOGLE_API_KEY and GOOGLE_CSE_ID respectively Wrappers# Utility# There exists a GoogleSearchAPIWrapper utility which wraps this API. To import this utility: from langchain.utilities import GoogleSearchAPIWrapper For a more detailed walkthrough of this wrapper, see this notebook. Tool# You can also easily load this wrapper as a Tool (to use with an Agent). You can do this with: from langchain.agents import load_tools tools = load_tools(["google-search"]) For more information on this, see this page previous ForefrontAI next Google Serper Wrapper Contents Installation and Setup Wrappers Utility Tool By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/google_search.html" }
b323fc18b8ef-0
.md .pdf AI21 Labs Contents Installation and Setup Wrappers LLM AI21 Labs# This page covers how to use the AI21 ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific AI21 wrappers. Installation and Setup# Get an AI21 api key and set it as an environment variable (AI21_API_KEY) Wrappers# LLM# There exists an AI21 LLM wrapper, which you can access with from langchain.llms import AI21 previous LangChain Ecosystem next Aim Contents Installation and Setup Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/ai21.html" }
3634e61a0e6c-0
.md .pdf DeepInfra Contents Installation and Setup Wrappers LLM DeepInfra# This page covers how to use the DeepInfra ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific DeepInfra wrappers. Installation and Setup# Get your DeepInfra api key from this link here. Get an DeepInfra api key and set it as an environment variable (DEEPINFRA_API_TOKEN) Wrappers# LLM# There exists an DeepInfra LLM wrapper, which you can access with from langchain.llms import DeepInfra previous Cohere next Deep Lake Contents Installation and Setup Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/deepinfra.html" }
ac59dbf597c4-0
.md .pdf Llama.cpp Contents Installation and Setup Wrappers LLM Embeddings Llama.cpp# This page covers how to use llama.cpp within LangChain. It is broken into two parts: installation and setup, and then references to specific Llama-cpp wrappers. Installation and Setup# Install the Python package with pip install llama-cpp-python Download one of the supported models and convert them to the llama.cpp format per the instructions Wrappers# LLM# There exists a LlamaCpp LLM wrapper, which you can access with from langchain.llms import LlamaCpp For a more detailed walkthrough of this, see this notebook Embeddings# There exists a LlamaCpp Embeddings wrapper, which you can access with from langchain.embeddings import LlamaCppEmbeddings For a more detailed walkthrough of this, see this notebook previous Jina next Milvus Contents Installation and Setup Wrappers LLM Embeddings By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/llamacpp.html" }
38a25bce3816-0
.md .pdf SearxNG Search API Contents Installation and Setup Self Hosted Instance: Wrappers Utility Tool SearxNG Search API# This page covers how to use the SearxNG search API within LangChain. It is broken into two parts: installation and setup, and then references to the specific SearxNG API wrapper. Installation and Setup# While it is possible to utilize the wrapper in conjunction with public searx instances these instances frequently do not permit API access (see note on output format below) and have limitations on the frequency of requests. It is recommended to opt for a self-hosted instance instead. Self Hosted Instance:# See this page for installation instructions. When you install SearxNG, the only active output format by default is the HTML format. You need to activate the json format to use the API. This can be done by adding the following line to the settings.yml file: search: formats: - html - json You can make sure that the API is working by issuing a curl request to the API endpoint: curl -kLX GET --data-urlencode q='langchain' -d format=json http://localhost:8888 This should return a JSON object with the results. Wrappers# Utility# To use the wrapper we need to pass the host of the SearxNG instance to the wrapper with: 1. the named parameter searx_host when creating the instance. 2. exporting the environment variable SEARXNG_HOST. You can use the wrapper to get results from a SearxNG instance. from langchain.utilities import SearxSearchWrapper s = SearxSearchWrapper(searx_host="http://localhost:8888") s.run("what is a large language model?")
{ "url": "https://python.langchain.com/en/latest/ecosystem/searx.html" }
38a25bce3816-1
s.run("what is a large language model?") Tool# You can also load this wrapper as a Tool (to use with an Agent). You can do this with: from langchain.agents import load_tools tools = load_tools(["searx-search"], searx_host="http://localhost:8888", engines=["github"]) Note that we could optionally pass custom engines to use. If you want to obtain results with metadata as json you can use: tools = load_tools(["searx-search-results-json"], searx_host="http://localhost:8888", num_results=5) For more information on tools, see this page previous RWKV-4 next SerpAPI Contents Installation and Setup Self Hosted Instance: Wrappers Utility Tool By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/searx.html" }
9ad125954864-0
.md .pdf Deep Lake Contents Why Deep Lake? More Resources Installation and Setup Wrappers VectorStore Deep Lake# This page covers how to use the Deep Lake ecosystem within LangChain. Why Deep Lake?# More than just a (multi-modal) vector store. You can later use the dataset to fine-tune your own LLM models. Not only stores embeddings, but also the original data with automatic version control. Truly serverless. Doesn’t require another service and can be used with major cloud providers (AWS S3, GCS, etc.) More Resources# Ultimate Guide to LangChain & Deep Lake: Build ChatGPT to Answer Questions on Your Financial Data Here is whitepaper and academic paper for Deep Lake Here is a set of additional resources available for review: Deep Lake, Getting Started and Tutorials Installation and Setup# Install the Python package with pip install deeplake Wrappers# VectorStore# There exists a wrapper around Deep Lake, a data lake for Deep Learning applications, allowing you to use it as a vector store (for now), whether for semantic search or example selection. To import this vectorstore: from langchain.vectorstores import DeepLake For a more detailed walkthrough of the Deep Lake wrapper, see this notebook previous DeepInfra next ForefrontAI Contents Why Deep Lake? More Resources Installation and Setup Wrappers VectorStore By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/deeplake.html" }
d6bc958921fb-0
.md .pdf GooseAI Contents Installation and Setup Wrappers LLM GooseAI# This page covers how to use the GooseAI ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific GooseAI wrappers. Installation and Setup# Install the Python SDK with pip install openai Get your GooseAI api key from this link here. Set the environment variable (GOOSEAI_API_KEY). import os os.environ["GOOSEAI_API_KEY"] = "YOUR_API_KEY" Wrappers# LLM# There exists an GooseAI LLM wrapper, which you can access with: from langchain.llms import GooseAI previous Google Serper Wrapper next GPT4All Contents Installation and Setup Wrappers LLM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/gooseai.html" }
2f1685949587-0
.md .pdf OpenSearch Contents Installation and Setup Wrappers VectorStore OpenSearch# This page covers how to use the OpenSearch ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific OpenSearch wrappers. Installation and Setup# Install the Python package with pip install opensearch-py Wrappers# VectorStore# There exists a wrapper around OpenSearch vector databases, allowing you to use it as a vectorstore for semantic search using approximate vector search powered by lucene, nmslib and faiss engines or using painless scripting and script scoring functions for bruteforce vector search. To import this vectorstore: from langchain.vectorstores import OpenSearchVectorSearch For a more detailed walkthrough of the OpenSearch wrapper, see this notebook previous OpenAI next Petals Contents Installation and Setup Wrappers VectorStore By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/opensearch.html" }
81b004db7d7b-0
.md .pdf RWKV-4 Contents Installation and Setup Usage RWKV Model File Rwkv-4 models -> recommended VRAM RWKV-4# This page covers how to use the RWKV-4 wrapper within LangChain. It is broken into two parts: installation and setup, and then usage with an example. Installation and Setup# Install the Python package with pip install rwkv Install the tokenizer Python package with pip install tokenizer Download a RWKV model and place it in your desired directory Download the tokens file Usage# RWKV# To use the RWKV wrapper, you need to provide the path to the pre-trained model file and the tokenizer’s configuration. from langchain.llms import RWKV # Test the model ```python def generate_prompt(instruction, input=None): if input: return f"""Below is an instruction that describes a task, paired with an input that provides further context. Write a response that appropriately completes the request. # Instruction: {instruction} # Input: {input} # Response: """ else: return f"""Below is an instruction that describes a task. Write a response that appropriately completes the request. # Instruction: {instruction} # Response: """ model = RWKV(model="./models/RWKV-4-Raven-3B-v7-Eng-20230404-ctx4096.pth", strategy="cpu fp32", tokens_path="./rwkv/20B_tokenizer.json") response = model(generate_prompt("Once upon a time, ")) Model File# You can find links to model file downloads at the RWKV-4-Raven repository. Rwkv-4 models -> recommended VRAM# RWKV VRAM Model | 8bit | bf16/fp16 | fp32
{ "url": "https://python.langchain.com/en/latest/ecosystem/rwkv.html" }
81b004db7d7b-1
RWKV VRAM Model | 8bit | bf16/fp16 | fp32 14B | 16GB | 28GB | >50GB 7B | 8GB | 14GB | 28GB 3B | 2.8GB| 6GB | 12GB 1b5 | 1.3GB| 3GB | 6GB See the rwkv pip page for more information about strategies, including streaming and cuda support. previous Runhouse next SearxNG Search API Contents Installation and Setup Usage RWKV Model File Rwkv-4 models -> recommended VRAM By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/rwkv.html" }
1c652c39a815-0
.md .pdf Google Serper Wrapper Contents Setup Wrappers Utility Output Tool Google Serper Wrapper# This page covers how to use the Serper Google Search API within LangChain. Serper is a low-cost Google Search API that can be used to add answer box, knowledge graph, and organic results data from Google Search. It is broken into two parts: setup, and then references to the specific Google Serper wrapper. Setup# Go to serper.dev to sign up for a free account Get the api key and set it as an environment variable (SERPER_API_KEY) Wrappers# Utility# There exists a GoogleSerperAPIWrapper utility which wraps this API. To import this utility: from langchain.utilities import GoogleSerperAPIWrapper You can use it as part of a Self Ask chain: from langchain.utilities import GoogleSerperAPIWrapper from langchain.llms.openai import OpenAI from langchain.agents import initialize_agent, Tool from langchain.agents import AgentType import os os.environ["SERPER_API_KEY"] = "" os.environ['OPENAI_API_KEY'] = "" llm = OpenAI(temperature=0) search = GoogleSerperAPIWrapper() tools = [ Tool( name="Intermediate Answer", func=search.run, description="useful for when you need to ask with search" ) ] self_ask_with_search = initialize_agent(tools, llm, agent=AgentType.SELF_ASK_WITH_SEARCH, verbose=True) self_ask_with_search.run("What is the hometown of the reigning men's U.S. Open champion?") Output# Entering new AgentExecutor chain... Yes. Follow up: Who is the reigning men's U.S. Open champion?
{ "url": "https://python.langchain.com/en/latest/ecosystem/google_serper.html" }
1c652c39a815-1
Yes. Follow up: Who is the reigning men's U.S. Open champion? Intermediate answer: Current champions Carlos Alcaraz, 2022 men's singles champion. Follow up: Where is Carlos Alcaraz from? Intermediate answer: El Palmar, Spain So the final answer is: El Palmar, Spain > Finished chain. 'El Palmar, Spain' For a more detailed walkthrough of this wrapper, see this notebook. Tool# You can also easily load this wrapper as a Tool (to use with an Agent). You can do this with: from langchain.agents import load_tools tools = load_tools(["google-serper"]) For more information on this, see this page previous Google Search Wrapper next GooseAI Contents Setup Wrappers Utility Output Tool By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/google_serper.html" }
88cc5830a655-0
.md .pdf Pinecone Contents Installation and Setup Wrappers VectorStore Pinecone# This page covers how to use the Pinecone ecosystem within LangChain. It is broken into two parts: installation and setup, and then references to specific Pinecone wrappers. Installation and Setup# Install the Python SDK with pip install pinecone-client Wrappers# VectorStore# There exists a wrapper around Pinecone indexes, allowing you to use it as a vectorstore, whether for semantic search or example selection. To import this vectorstore: from langchain.vectorstores import Pinecone For a more detailed walkthrough of the Pinecone wrapper, see this notebook previous PGVector next PromptLayer Contents Installation and Setup Wrappers VectorStore By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/pinecone.html" }
7756f836f594-0
.md .pdf GPT4All Contents Installation and Setup Usage GPT4All Model File GPT4All# This page covers how to use the GPT4All wrapper within LangChain. It is broken into two parts: installation and setup, and then usage with an example. Installation and Setup# Install the Python package with pip install pyllamacpp Download a GPT4All model and place it in your desired directory Usage# GPT4All# To use the GPT4All wrapper, you need to provide the path to the pre-trained model file and the model’s configuration. from langchain.llms import GPT4All # Instantiate the model model = GPT4All(model="./models/gpt4all-model.bin", n_ctx=512, n_threads=8) # Generate text response = model("Once upon a time, ") You can also customize the generation parameters, such as n_predict, temp, top_p, top_k, and others. Example: model = GPT4All(model="./models/gpt4all-model.bin", n_predict=55, temp=0) response = model("Once upon a time, ") Model File# You can find links to model file downloads at the GPT4all repository. They will need to be converted to ggml format to work, as specified in the pyllamacpp repository. For a more detailed walkthrough of this, see this notebook previous GooseAI next Graphsignal Contents Installation and Setup Usage GPT4All Model File By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/gpt4all.html" }
7399f5dfd4c1-0
.ipynb .pdf Weights & Biases Weights & Biases# This notebook goes over how to track your LangChain experiments into one centralized Weights and Biases dashboard. To learn more about prompt engineering and the callback please refer to this Report which explains both alongside the resultant dashboards you can expect to see. Run in Colab: https://colab.research.google.com/drive/1DXH4beT4HFaRKy_Vm4PoxhXVDRf7Ym8L?usp=sharing View Report: https://wandb.ai/a-sh0ts/langchain_callback_demo/reports/Prompt-Engineering-LLMs-with-LangChain-and-W-B–VmlldzozNjk1NTUw#👋-how-to-build-a-callback-in-langchain-for-better-prompt-engineering !pip install wandb !pip install pandas !pip install textstat !pip install spacy !python -m spacy download en_core_web_sm import os os.environ["WANDB_API_KEY"] = "" # os.environ["OPENAI_API_KEY"] = "" # os.environ["SERPAPI_API_KEY"] = "" from datetime import datetime from langchain.callbacks import WandbCallbackHandler, StdOutCallbackHandler from langchain.callbacks.base import CallbackManager from langchain.llms import OpenAI Callback Handler that logs to Weights and Biases. Parameters: job_type (str): The type of job. project (str): The project to log to. entity (str): The entity to log to. tags (list): The tags to log. group (str): The group to log to. name (str): The name of the run. notes (str): The notes to log.
{ "url": "https://python.langchain.com/en/latest/ecosystem/wandb_tracking.html" }
7399f5dfd4c1-1
notes (str): The notes to log. visualize (bool): Whether to visualize the run. complexity_metrics (bool): Whether to log complexity metrics. stream_logs (bool): Whether to stream callback actions to W&B Default values for WandbCallbackHandler(...) visualize: bool = False, complexity_metrics: bool = False, stream_logs: bool = False, NOTE: For beta workflows we have made the default analysis based on textstat and the visualizations based on spacy """Main function. This function is used to try the callback handler. Scenarios: 1. OpenAI LLM 2. Chain with multiple SubChains on multiple generations 3. Agent with Tools """ session_group = datetime.now().strftime("%m.%d.%Y_%H.%M.%S") wandb_callback = WandbCallbackHandler( job_type="inference", project="langchain_callback_demo", group=f"minimal_{session_group}", name="llm", tags=["test"], ) manager = CallbackManager([StdOutCallbackHandler(), wandb_callback]) llm = OpenAI(temperature=0, callback_manager=manager, verbose=True) wandb: Currently logged in as: harrison-chase. Use `wandb login --relogin` to force relogin
{ "url": "https://python.langchain.com/en/latest/ecosystem/wandb_tracking.html" }
7399f5dfd4c1-2
Tracking run with wandb version 0.14.0Run data is saved locally in /Users/harrisonchase/workplace/langchain/docs/ecosystem/wandb/run-20230318_150408-e47j1914Syncing run llm to Weights & Biases (docs) View project at https://wandb.ai/harrison-chase/langchain_callback_demo View run at https://wandb.ai/harrison-chase/langchain_callback_demo/runs/e47j1914wandb: WARNING The wandb callback is currently in beta and is subject to change based on updates to `langchain`. Please report any issues to https://github.com/wandb/wandb/issues with the tag `langchain`. # Defaults for WandbCallbackHandler.flush_tracker(...) reset: bool = True, finish: bool = False, The flush_tracker function is used to log LangChain sessions to Weights & Biases. It takes in the LangChain module or agent, and logs at minimum the prompts and generations alongside the serialized form of the LangChain module to the specified Weights & Biases project. By default we reset the session as opposed to concluding the session outright. # SCENARIO 1 - LLM llm_result = llm.generate(["Tell me a joke", "Tell me a poem"] * 3) wandb_callback.flush_tracker(llm, name="simple_sequential")
{ "url": "https://python.langchain.com/en/latest/ecosystem/wandb_tracking.html" }
7399f5dfd4c1-3
wandb_callback.flush_tracker(llm, name="simple_sequential") Waiting for W&B process to finish... (success). View run llm at: https://wandb.ai/harrison-chase/langchain_callback_demo/runs/e47j1914Synced 5 W&B file(s), 2 media file(s), 5 artifact file(s) and 0 other file(s)Find logs at: ./wandb/run-20230318_150408-e47j1914/logsTracking run with wandb version 0.14.0Run data is saved locally in /Users/harrisonchase/workplace/langchain/docs/ecosystem/wandb/run-20230318_150534-jyxma7huSyncing run simple_sequential to Weights & Biases (docs) View project at https://wandb.ai/harrison-chase/langchain_callback_demo View run at https://wandb.ai/harrison-chase/langchain_callback_demo/runs/jyxma7hu from langchain.prompts import PromptTemplate from langchain.chains import LLMChain # SCENARIO 2 - Chain template = """You are a playwright. Given the title of play, it is your job to write a synopsis for that title. Title: {title} Playwright: This is a synopsis for the above play:""" prompt_template = PromptTemplate(input_variables=["title"], template=template) synopsis_chain = LLMChain(llm=llm, prompt=prompt_template, callback_manager=manager) test_prompts = [ { "title": "documentary about good video games that push the boundary of game design" }, {"title": "cocaine bear vs heroin wolf"}, {"title": "the best in class mlops tooling"}, ] synopsis_chain.apply(test_prompts)
{ "url": "https://python.langchain.com/en/latest/ecosystem/wandb_tracking.html" }
7399f5dfd4c1-4
] synopsis_chain.apply(test_prompts) wandb_callback.flush_tracker(synopsis_chain, name="agent") Waiting for W&B process to finish... (success). View run simple_sequential at: https://wandb.ai/harrison-chase/langchain_callback_demo/runs/jyxma7huSynced 4 W&B file(s), 2 media file(s), 6 artifact file(s) and 0 other file(s)Find logs at: ./wandb/run-20230318_150534-jyxma7hu/logsTracking run with wandb version 0.14.0Run data is saved locally in /Users/harrisonchase/workplace/langchain/docs/ecosystem/wandb/run-20230318_150550-wzy59zjqSyncing run agent to Weights & Biases (docs) View project at https://wandb.ai/harrison-chase/langchain_callback_demo View run at https://wandb.ai/harrison-chase/langchain_callback_demo/runs/wzy59zjq from langchain.agents import initialize_agent, load_tools from langchain.agents import AgentType # SCENARIO 3 - Agent with Tools tools = load_tools(["serpapi", "llm-math"], llm=llm, callback_manager=manager) agent = initialize_agent( tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, callback_manager=manager, verbose=True, ) agent.run( "Who is Leo DiCaprio's girlfriend? What is her current age raised to the 0.43 power?" ) wandb_callback.flush_tracker(agent, reset=False, finish=True) > Entering new AgentExecutor chain... I need to find out who Leo DiCaprio's girlfriend is and then calculate her age raised to the 0.43 power. Action: Search
{ "url": "https://python.langchain.com/en/latest/ecosystem/wandb_tracking.html" }
7399f5dfd4c1-5
Action: Search Action Input: "Leo DiCaprio girlfriend" Observation: DiCaprio had a steady girlfriend in Camila Morrone. He had been with the model turned actress for nearly five years, as they were first said to be dating at the end of 2017. And the now 26-year-old Morrone is no stranger to Hollywood. Thought: I need to calculate her age raised to the 0.43 power. Action: Calculator Action Input: 26^0.43 Observation: Answer: 4.059182145592686 Thought: I now know the final answer. Final Answer: Leo DiCaprio's girlfriend is Camila Morrone and her current age raised to the 0.43 power is 4.059182145592686. > Finished chain. Waiting for W&B process to finish... (success). View run agent at: https://wandb.ai/harrison-chase/langchain_callback_demo/runs/wzy59zjqSynced 5 W&B file(s), 2 media file(s), 7 artifact file(s) and 0 other file(s)Find logs at: ./wandb/run-20230318_150550-wzy59zjq/logs previous Unstructured next Weaviate By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/wandb_tracking.html" }
4dbc219c64e4-0
.ipynb .pdf Aim Aim# Aim makes it super easy to visualize and debug LangChain executions. Aim tracks inputs and outputs of LLMs and tools, as well as actions of agents. With Aim, you can easily debug and examine an individual execution: Additionally, you have the option to compare multiple executions side by side: Aim is fully open source, learn more about Aim on GitHub. Let’s move forward and see how to enable and configure Aim callback. Tracking LangChain Executions with AimIn this notebook we will explore three usage scenarios. To start off, we will install the necessary packages and import certain modules. Subsequently, we will configure two environment variables that can be established either within the Python script or through the terminal. !pip install aim !pip install langchain !pip install openai !pip install google-search-results import os from datetime import datetime from langchain.llms import OpenAI from langchain.callbacks.base import CallbackManager from langchain.callbacks import AimCallbackHandler, StdOutCallbackHandler Our examples use a GPT model as the LLM, and OpenAI offers an API for this purpose. You can obtain the key from the following link: https://platform.openai.com/account/api-keys . We will use the SerpApi to retrieve search results from Google. To acquire the SerpApi key, please go to https://serpapi.com/manage-api-key . os.environ["OPENAI_API_KEY"] = "..." os.environ["SERPAPI_API_KEY"] = "..." The event methods of AimCallbackHandler accept the LangChain module or agent as input and log at least the prompts and generated results, as well as the serialized version of the LangChain module, to the designated Aim run. session_group = datetime.now().strftime("%m.%d.%Y_%H.%M.%S")
{ "url": "https://python.langchain.com/en/latest/ecosystem/aim_tracking.html" }
4dbc219c64e4-1
session_group = datetime.now().strftime("%m.%d.%Y_%H.%M.%S") aim_callback = AimCallbackHandler( repo=".", experiment_name="scenario 1: OpenAI LLM", ) manager = CallbackManager([StdOutCallbackHandler(), aim_callback]) llm = OpenAI(temperature=0, callback_manager=manager, verbose=True) The flush_tracker function is used to record LangChain assets on Aim. By default, the session is reset rather than being terminated outright. Scenario 1 In the first scenario, we will use OpenAI LLM. # scenario 1 - LLM llm_result = llm.generate(["Tell me a joke", "Tell me a poem"] * 3) aim_callback.flush_tracker( langchain_asset=llm, experiment_name="scenario 2: Chain with multiple SubChains on multiple generations", ) Scenario 2 Scenario two involves chaining with multiple SubChains across multiple generations. from langchain.prompts import PromptTemplate from langchain.chains import LLMChain # scenario 2 - Chain template = """You are a playwright. Given the title of play, it is your job to write a synopsis for that title. Title: {title} Playwright: This is a synopsis for the above play:""" prompt_template = PromptTemplate(input_variables=["title"], template=template) synopsis_chain = LLMChain(llm=llm, prompt=prompt_template, callback_manager=manager) test_prompts = [ {"title": "documentary about good video games that push the boundary of game design"}, {"title": "the phenomenon behind the remarkable speed of cheetahs"}, {"title": "the best in class mlops tooling"}, ] synopsis_chain.apply(test_prompts) aim_callback.flush_tracker(
{ "url": "https://python.langchain.com/en/latest/ecosystem/aim_tracking.html" }
4dbc219c64e4-2
] synopsis_chain.apply(test_prompts) aim_callback.flush_tracker( langchain_asset=synopsis_chain, experiment_name="scenario 3: Agent with Tools" ) Scenario 3 The third scenario involves an agent with tools. from langchain.agents import initialize_agent, load_tools from langchain.agents import AgentType # scenario 3 - Agent with Tools tools = load_tools(["serpapi", "llm-math"], llm=llm, callback_manager=manager) agent = initialize_agent( tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, callback_manager=manager, verbose=True, ) agent.run( "Who is Leo DiCaprio's girlfriend? What is her current age raised to the 0.43 power?" ) aim_callback.flush_tracker(langchain_asset=agent, reset=False, finish=True) > Entering new AgentExecutor chain... I need to find out who Leo DiCaprio's girlfriend is and then calculate her age raised to the 0.43 power. Action: Search Action Input: "Leo DiCaprio girlfriend" Observation: Leonardo DiCaprio seemed to prove a long-held theory about his love life right after splitting from girlfriend Camila Morrone just months ... Thought: I need to find out Camila Morrone's age Action: Search Action Input: "Camila Morrone age" Observation: 25 years Thought: I need to calculate 25 raised to the 0.43 power Action: Calculator Action Input: 25^0.43 Observation: Answer: 3.991298452658078 Thought: I now know the final answer
{ "url": "https://python.langchain.com/en/latest/ecosystem/aim_tracking.html" }
4dbc219c64e4-3
Thought: I now know the final answer Final Answer: Camila Morrone is Leo DiCaprio's girlfriend and her current age raised to the 0.43 power is 3.991298452658078. > Finished chain. previous AI21 Labs next Apify By Harrison Chase © Copyright 2023, Harrison Chase. Last updated on Apr 08, 2023.
{ "url": "https://python.langchain.com/en/latest/ecosystem/aim_tracking.html" }
251fe73a9267-0
.ipynb .pdf Tracing Walkthrough Tracing Walkthrough# import os os.environ["LANGCHAIN_HANDLER"] = "langchain" ## Uncomment this if using hosted setup. # os.environ["LANGCHAIN_ENDPOINT"] = "https://langchain-api-gateway-57eoxz8z.uc.gateway.dev" ## Uncomment this if you want traces to be recorded to "my_session" instead of default. # os.environ["LANGCHAIN_SESSION"] = "my_session" ## Better to set this environment variable in the terminal ## Uncomment this if using hosted version. Replace "my_api_key" with your actual API Key. # os.environ["LANGCHAIN_API_KEY"] = "my_api_key" import langchain from langchain.agents import Tool, initialize_agent, load_tools from langchain.agents import AgentType from langchain.chat_models import ChatOpenAI from langchain.llms import OpenAI # Agent run with tracing. Ensure that OPENAI_API_KEY is set appropriately to run this example. llm = OpenAI(temperature=0) tools = load_tools(["llm-math"], llm=llm) agent = initialize_agent( tools, llm, agent=AgentType.ZERO_SHOT_REACT_DESCRIPTION, verbose=True ) agent.run("What is 2 raised to .123243 power?") > Entering new AgentExecutor chain... I need to use a calculator to solve this. Action: Calculator Action Input: 2^.123243 Observation: Answer: 1.0891804557407723 Thought: I now know the final answer. Final Answer: 1.0891804557407723 > Finished chain. '1.0891804557407723' # Agent run with tracing using a chat model agent = initialize_agent(
{ "url": "https://python.langchain.com/en/latest/tracing/agent_with_tracing.html" }