text
stringlengths
3
1.68M
id
stringlengths
13
169
metadata
dict
__index_level_0__
int64
0
2.21k
# Serper - Google Search API This page covers how to use the [Serper](https://serper.dev) Google Search API within LangChain. Serper is a low-cost Google Search API that can be used to add answer box, knowledge graph, and organic results data from Google Search. It is broken into two parts: setup, and then references...
langchain/docs/docs/integrations/providers/google_serper.mdx/0
{ "file_path": "langchain/docs/docs/integrations/providers/google_serper.mdx", "repo_id": "langchain", "token_count": 676 }
147
PD_PREFIX = """ You are working with Xorbits dataframe object in Python. Before importing Numpy or Pandas in the current script, remember to import the xorbits version of the library instead. To import the xorbits version of Numpy, replace the original import statement `import pandas as pd` with `import xorbits.pandas ...
langchain/libs/experimental/langchain_experimental/agents/agent_toolkits/xorbits/prompt.py/0
{ "file_path": "langchain/libs/experimental/langchain_experimental/agents/agent_toolkits/xorbits/prompt.py", "repo_id": "langchain", "token_count": 317 }
414
// Presents an interface to a storage backend such as s3 or local disk. // The interface is a simple key-value store, which maps to s3 well. // For now the interface fetches a file and stores it at a specific // location on disk. This is not ideal for s3, but it is a start. // Ideally we would support streaming the fi...
chroma/rust/worker/src/storage/s3.rs/0
{ "file_path": "chroma/rust/worker/src/storage/s3.rs", "repo_id": "chroma", "token_count": 4163 }
59
# rag-astradb This template will perform RAG using Astra DB (`AstraDB` vector store class) ## Environment Setup An [Astra DB](https://astra.datastax.com) database is required; free tier is fine. - You need the database **API endpoint** (such as `https://0123...-us-east1.apps.astra.datastax.com`) ... - ... and a **...
langchain/templates/rag-astradb/README.md/0
{ "file_path": "langchain/templates/rag-astradb/README.md", "repo_id": "langchain", "token_count": 781 }
724
from copy import deepcopy from typing import Any, Dict, List, Optional from langchain_core.agents import AgentAction, AgentFinish from langchain_core.callbacks import BaseCallbackHandler from langchain_core.outputs import LLMResult def import_aim() -> Any: """Import the aim python package and raise an error if i...
langchain/libs/community/langchain_community/callbacks/aim_callback.py/0
{ "file_path": "langchain/libs/community/langchain_community/callbacks/aim_callback.py", "repo_id": "langchain", "token_count": 6637 }
226
"""Common classes/functions for tree index operations.""" import asyncio import logging from typing import Dict, List, Optional, Sequence, Tuple from llama_index.legacy.async_utils import run_async_tasks from llama_index.legacy.callbacks.schema import CBEventType, EventPayload from llama_index.legacy.data_structs.dat...
llama_index/llama-index-legacy/llama_index/legacy/indices/common_tree/base.py/0
{ "file_path": "llama_index/llama-index-legacy/llama_index/legacy/indices/common_tree/base.py", "repo_id": "llama_index", "token_count": 4328 }
1,495
import { logVersion010MigrationWarning } from "../util/entrypoint_deprecation.js"; /* #__PURE__ */ logVersion010MigrationWarning({ oldEntrypointName: "vectorstores/mongodb_atlas", }); export * from "@langchain/community/vectorstores/mongodb_atlas";
langchainjs/langchain/src/vectorstores/mongodb_atlas.ts/0
{ "file_path": "langchainjs/langchain/src/vectorstores/mongodb_atlas.ts", "repo_id": "langchainjs", "token_count": 80 }
966
use std::collections::HashMap; use std::hash::Hash; #[derive(Default)] pub struct TrieBuilder<Label> { trie: Trie<Label>, } impl<Label: Eq + Hash + Copy> TrieBuilder<Label> { pub fn push(&mut self, element: &[Label]) { self.trie.push(element); } pub fn build(self) -> Trie<Label> { sel...
tokenizers/tokenizers/src/models/unigram/trie.rs/0
{ "file_path": "tokenizers/tokenizers/src/models/unigram/trie.rs", "repo_id": "tokenizers", "token_count": 944 }
436
#!/usr/bin/env python # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unles...
accelerate/src/accelerate/commands/test.py/0
{ "file_path": "accelerate/src/accelerate/commands/test.py", "repo_id": "accelerate", "token_count": 773 }
12
from llama_index.core.base.embeddings.base import BaseEmbedding from llama_index.embeddings.nomic import NomicEmbedding def test_embedding_class(): names_of_base_classes = [b.__name__ for b in NomicEmbedding.__mro__] assert BaseEmbedding.__name__ in names_of_base_classes
llama_index/llama-index-integrations/embeddings/llama-index-embeddings-nomic/tests/test_embeddings_nomic.py/0
{ "file_path": "llama_index/llama-index-integrations/embeddings/llama-index-embeddings-nomic/tests/test_embeddings_nomic.py", "repo_id": "llama_index", "token_count": 103 }
1,200
import argparse import json import logging import os import subprocess from pathlib import Path from typing import Dict, List, Literal, Mapping, Optional, Union, cast from langsmith import env as ls_env from langsmith import utils as ls_utils logging.basicConfig(level=logging.INFO, format="%(message)s") logger = logg...
langsmith-sdk/python/langsmith/cli/main.py/0
{ "file_path": "langsmith-sdk/python/langsmith/cli/main.py", "repo_id": "langsmith-sdk", "token_count": 5243 }
1,070
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/internal/querynodev2/mock_data.go/0
{ "file_path": "milvus/internal/querynodev2/mock_data.go", "repo_id": "milvus", "token_count": 1865 }
1,988
# Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by appl...
accelerate/examples/by_feature/cross_validation.py/0
{ "file_path": "accelerate/examples/by_feature/cross_validation.py", "repo_id": "accelerate", "token_count": 4458 }
5
"""Test Google PaLM Chat API wrapper.""" import pytest from langchain_core.messages import AIMessage, HumanMessage, SystemMessage from langchain_community.chat_models.google_palm import ( ChatGooglePalm, ChatGooglePalmError, _messages_to_prompt_dict, ) def test_messages_to_prompt_dict_with_valid_message...
langchain/libs/community/tests/unit_tests/chat_models/test_google_palm.py/0
{ "file_path": "langchain/libs/community/tests/unit_tests/chat_models/test_google_palm.py", "repo_id": "langchain", "token_count": 1707 }
368
package predicates import ( "testing" "github.com/stretchr/testify/suite" ) type PredicateSuite struct { suite.Suite } func (s *PredicateSuite) TestValueEqual() { p := ValueEqual("key", "value") s.Equal("key", p.Key()) s.Equal("value", p.TargetValue()) s.Equal(PredTargetValue, p.Target()) s.Equal(PredTypeEq...
milvus/internal/kv/predicates/predicate_test.go/0
{ "file_path": "milvus/internal/kv/predicates/predicate_test.go", "repo_id": "milvus", "token_count": 303 }
1,994
"""Chain that calls SerpAPI. Heavily borrowed from https://github.com/ofirpress/self-ask """ import os import sys from typing import Any, Dict, Optional, Tuple import aiohttp from langchain_core.pydantic_v1 import BaseModel, Extra, Field, root_validator from langchain_core.utils import get_from_dict_or_env class Hi...
langchain/libs/community/langchain_community/utilities/serpapi.py/0
{ "file_path": "langchain/libs/community/langchain_community/utilities/serpapi.py", "repo_id": "langchain", "token_count": 4034 }
304
from langchain_community.llms.fake import FakeListLLM, FakeStreamingListLLM __all__ = ["FakeListLLM", "FakeStreamingListLLM"]
langchain/libs/langchain/langchain/llms/fake.py/0
{ "file_path": "langchain/libs/langchain/langchain/llms/fake.py", "repo_id": "langchain", "token_count": 43 }
546
from llama_index.tools.metaphor.base import MetaphorToolSpec __all__ = ["MetaphorToolSpec"]
llama_index/llama-index-integrations/tools/llama-index-tools-metaphor/llama_index/tools/metaphor/__init__.py/0
{ "file_path": "llama_index/llama-index-integrations/tools/llama-index-tools-metaphor/llama_index/tools/metaphor/__init__.py", "repo_id": "llama_index", "token_count": 33 }
1,572
import { logVersion010MigrationWarning } from "../util/entrypoint_deprecation.js"; /* #__PURE__ */ logVersion010MigrationWarning({ oldEntrypointName: "llms/gradient_ai", }); export * from "@langchain/community/llms/gradient_ai";
langchainjs/langchain/src/llms/gradient_ai.ts/0
{ "file_path": "langchainjs/langchain/src/llms/gradient_ai.ts", "repo_id": "langchainjs", "token_count": 74 }
944
#!/usr/bin/env python """An example that shows how to create a custom agent executor like Runnable. At the time of writing, there is a bug in the current AgentExecutor that prevents it from correctly propagating configuration of the underlying runnable. While that bug should be fixed, this is an example shows how to c...
langserve/examples/configurable_agent_executor/server.py/0
{ "file_path": "langserve/examples/configurable_agent_executor/server.py", "repo_id": "langserve", "token_count": 1757 }
1,059
--- sidebar_label: Google AI --- import CodeBlock from "@theme/CodeBlock"; # Google Generative AI You can access Google's generative AI embeddings models through `@langchain/google-genai` integration package. Get an API key here: https://ai.google.dev/tutorials/setup You'll need to install the `@langchain/google-g...
langchainjs/docs/core_docs/docs/integrations/text_embedding/google_generativeai.mdx/0
{ "file_path": "langchainjs/docs/core_docs/docs/integrations/text_embedding/google_generativeai.mdx", "repo_id": "langchainjs", "token_count": 220 }
736
# coding=utf-8 # Copyright 2023 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
transformers/tests/models/mixtral/test_modeling_mixtral.py/0
{ "file_path": "transformers/tests/models/mixtral/test_modeling_mixtral.py", "repo_id": "transformers", "token_count": 10571 }
737
"""Test formatting functionality.""" from langchain_core.language_models.base import _get_token_ids_default_method class TestTokenCountingWithGPT2Tokenizer: def test_tokenization(self) -> None: # Check that the tokenization is consistent with the GPT-2 tokenizer assert _get_token_ids_default_meth...
langchain/libs/langchain/tests/integration_tests/test_schema.py/0
{ "file_path": "langchain/libs/langchain/tests/integration_tests/test_schema.py", "repo_id": "langchain", "token_count": 273 }
615
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/internal/rootcoord/dml_channels.go/0
{ "file_path": "milvus/internal/rootcoord/dml_channels.go", "repo_id": "milvus", "token_count": 4438 }
1,851
# Consistency Decoder Consistency decoder can be used to decode the latents from the denoising UNet in the [`StableDiffusionPipeline`]. This decoder was introduced in the [DALL-E 3 technical report](https://openai.com/dall-e-3). The original codebase can be found at [openai/consistencydecoder](https://github.com/ope...
diffusers/docs/source/en/api/models/consistency_decoder_vae.md/0
{ "file_path": "diffusers/docs/source/en/api/models/consistency_decoder_vae.md", "repo_id": "diffusers", "token_count": 242 }
178
<jupyter_start><jupyter_code>!pip install -q git+https://github.com/huggingface/transformers.git !pip install -q git+https://github.com/huggingface/peft.git !pip install -q git+https://github.com/huggingface/accelerate.git@main !pip install huggingface_hub !pip install bitsandbytes !pip install SentencePiece import os ...
peft/examples/multi_adapter_examples/PEFT_Multi_LoRA_Inference.ipynb/0
{ "file_path": "peft/examples/multi_adapter_examples/PEFT_Multi_LoRA_Inference.ipynb", "repo_id": "peft", "token_count": 1328 }
319
# AWS EC2 Basic Deployment This is an example deployment to AWS EC2 Compute using [terraform](https://www.terraform.io/). This deployment will do the following: - Create a security group with required ports open (22 and 8000) - Create EC2 instance with Ubuntu 22 and deploy Chroma using docker compose - Create a data...
chroma/examples/deployments/aws-terraform/README.md/0
{ "file_path": "chroma/examples/deployments/aws-terraform/README.md", "repo_id": "chroma", "token_count": 1515 }
40
"""Data Connectors for LlamaIndex. This module contains the data connectors for LlamaIndex. Each connector inherits from a `BaseReader` class, connects to a data source, and loads Document objects from that data source. You may also choose to construct Document objects manually, for instance in our `Insert How-To Gui...
llama_index/llama-index-legacy/llama_index/legacy/readers/__init__.py/0
{ "file_path": "llama_index/llama-index-legacy/llama_index/legacy/readers/__init__.py", "repo_id": "llama_index", "token_count": 1329 }
1,758
# flake8: noqa PREFIX = """ You are working with a pandas dataframe in Python. The name of the dataframe is `df`. You should use the tools below to answer the question posed of you:""" MULTI_DF_PREFIX = """ You are working with {num_dfs} pandas dataframes in Python named df1, df2, etc. You should use the tools below...
langchain/libs/experimental/langchain_experimental/agents/agent_toolkits/pandas/prompt.py/0
{ "file_path": "langchain/libs/experimental/langchain_experimental/agents/agent_toolkits/pandas/prompt.py", "repo_id": "langchain", "token_count": 409 }
435
from langchain_core.messages import HumanMessageChunk from langchain_core.outputs import ChatGenerationChunk, GenerationChunk def test_generation_chunk() -> None: assert GenerationChunk(text="Hello, ") + GenerationChunk( text="world!" ) == GenerationChunk( text="Hello, world!" ), "Generati...
langchain/libs/core/tests/unit_tests/test_outputs.py/0
{ "file_path": "langchain/libs/core/tests/unit_tests/test_outputs.py", "repo_id": "langchain", "token_count": 1005 }
408
[build-system] build-backend = "poetry.core.masonry.api" requires = ["poetry-core"] [tool.codespell] check-filenames = true check-hidden = true skip = "*.csv,*.html,*.json,*.jsonl,*.pdf,*.txt,*.ipynb" [tool.llamahub] classes = ["HiveReader"] contains_example = false import_path = "llama_index.readers.hive" [tool.myp...
llama_index/llama-index-integrations/readers/llama-index-readers-hive/pyproject.toml/0
{ "file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-hive/pyproject.toml", "repo_id": "llama_index", "token_count": 687 }
1,457
from abc import ABC, abstractmethod from typing import List, Optional class Constraint(ABC): r"""Abstract base class for all constraints that can be applied during generation. It must define how the constraint can be satisfied. All classes that inherit Constraint must follow the requirement that ```...
transformers/src/transformers/generation/beam_constraints.py/0
{ "file_path": "transformers/src/transformers/generation/beam_constraints.py", "repo_id": "transformers", "token_count": 8310 }
595
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/pkg/util/uniquegenerator/unique_int_generator_test.go/0
{ "file_path": "milvus/pkg/util/uniquegenerator/unique_int_generator_test.go", "repo_id": "milvus", "token_count": 405 }
1,856
import { withJsonFormsControlProps } from "@jsonforms/react"; import PlusIcon from "../assets/PlusIcon.svg?react"; import TrashIcon from "../assets/TrashIcon.svg?react"; import { rankWith, and, schemaMatches, Paths, isControl, } from "@jsonforms/core"; import { AutosizeTextarea } from "./AutosizeTextarea"; im...
langserve/langserve/playground/src/components/ChatMessageTuplesControlRenderer.tsx/0
{ "file_path": "langserve/langserve/playground/src/components/ChatMessageTuplesControlRenderer.tsx", "repo_id": "langserve", "token_count": 2541 }
1,129
import logging import os from typing import Any, Callable, Optional, Tuple, Union from llama_index.core.llms.generic_utils import get_from_param_or_env from tenacity import ( before_sleep_log, retry, retry_if_exception_type, stop_after_attempt, stop_after_delay, wait_exponential, wait_rando...
llama_index/llama-index-integrations/embeddings/llama-index-embeddings-openai/llama_index/embeddings/openai/utils.py/0
{ "file_path": "llama_index/llama-index-integrations/embeddings/llama-index-embeddings-openai/llama_index/embeddings/openai/utils.py", "repo_id": "llama_index", "token_count": 1373 }
1,284
"""Mbox parser. Contains simple parser for mbox files. """ import logging from pathlib import Path from typing import Any, Dict, List, Optional from llama_index.legacy.readers.base import BaseReader from llama_index.legacy.schema import Document logger = logging.getLogger(__name__) class MboxReader(BaseReader): ...
llama_index/llama-index-legacy/llama_index/legacy/readers/file/mbox_reader.py/0
{ "file_path": "llama_index/llama-index-legacy/llama_index/legacy/readers/file/mbox_reader.py", "repo_id": "llama_index", "token_count": 1676 }
1,606
<!--Copyright 2024 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
diffusers/docs/source/ko/using-diffusers/reusing_seeds.md/0
{ "file_path": "diffusers/docs/source/ko/using-diffusers/reusing_seeds.md", "repo_id": "diffusers", "token_count": 1922 }
200
# Copyright 2023-present the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or...
peft/src/peft/tuners/lora/__init__.py/0
{ "file_path": "peft/src/peft/tuners/lora/__init__.py", "repo_id": "peft", "token_count": 413 }
349
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/internal/core/src/index/VectorMemIndex.cpp/0
{ "file_path": "milvus/internal/core/src/index/VectorMemIndex.cpp", "repo_id": "milvus", "token_count": 15202 }
1,648
from typing import List, Optional import pytest from langchain_core.documents import Document from langchain_community.vectorstores import SQLiteVSS from tests.integration_tests.vectorstores.fake_embeddings import ( FakeEmbeddings, fake_texts, ) def _sqlite_vss_from_texts( metadatas: Optional[List[dict]...
langchain/libs/community/tests/integration_tests/vectorstores/test_sqlitevss.py/0
{ "file_path": "langchain/libs/community/tests/integration_tests/vectorstores/test_sqlitevss.py", "repo_id": "langchain", "token_count": 768 }
360
# EfficientNet **EfficientNet** is a convolutional neural network architecture and scaling method that uniformly scales all dimensions of depth/width/resolution using a *compound coefficient*. Unlike conventional practice that arbitrary scales these factors, the EfficientNet scaling method uniformly scales network wi...
pytorch-image-models/docs/models/efficientnet.md/0
{ "file_path": "pytorch-image-models/docs/models/efficientnet.md", "repo_id": "pytorch-image-models", "token_count": 4897 }
354
<jupyter_start><jupyter_text>OpenRouter OpenRouter provides a standardized API to access many LLMs at the best price offered. You can find out more on their [homepage](https://openrouter.ai).If you're opening this Notebook on colab, you will probably need to install LlamaIndex 🦙.<jupyter_code>%pip install llama-index-...
llama_index/docs/examples/llm/openrouter.ipynb/0
{ "file_path": "llama_index/docs/examples/llm/openrouter.ipynb", "repo_id": "llama_index", "token_count": 1075 }
1,090
""" BEiT: BERT Pre-Training of Image Transformers (https://arxiv.org/abs/2106.08254) Model from official source: https://github.com/microsoft/unilm/tree/master/beit @inproceedings{beit, title={{BEiT}: {BERT} Pre-Training of Image Transformers}, author={Hangbo Bao and Li Dong and Songhao Piao and Furu Wei}, booktitle=...
pytorch-image-models/timm/models/beit.py/0
{ "file_path": "pytorch-image-models/timm/models/beit.py", "repo_id": "pytorch-image-models", "token_count": 12467 }
361
"""Dataset generation from documents.""" from __future__ import annotations import asyncio import re from typing import List from llama_index.legacy import Document, ServiceContext, SummaryIndex from llama_index.legacy.async_utils import DEFAULT_NUM_WORKERS, run_jobs from llama_index.legacy.core.response.schema impo...
llama_index/llama-index-legacy/llama_index/legacy/llama_dataset/generator.py/0
{ "file_path": "llama_index/llama-index-legacy/llama_index/legacy/llama_dataset/generator.py", "repo_id": "llama_index", "token_count": 4596 }
1,502
#!/usr/bin/env bash python run_asr.py \ --output_dir="./wav2vec2-large-lv60-timit-asr" \ --num_train_epochs="30" \ --per_device_train_batch_size="2" \ --per_device_eval_batch_size="2" \ --gradient_accumulation_steps="4" \ --evaluation_strategy="steps" \ --save_steps="500" \ --eval_steps="100" \ --logging_steps="50" \ -...
transformers/examples/research_projects/wav2vec2/finetune_large_lv60_timit_asr.sh/0
{ "file_path": "transformers/examples/research_projects/wav2vec2/finetune_large_lv60_timit_asr.sh", "repo_id": "transformers", "token_count": 276 }
579
python_tests()
llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-openai/tests/BUILD/0
{ "file_path": "llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-openai/tests/BUILD", "repo_id": "llama_index", "token_count": 5 }
1,342
{ "compilerOptions": { "target": "es5", "lib": ["dom", "dom.iterable", "esnext"], "allowJs": true, "skipLibCheck": true, "strict": true, "forceConsistentCasingInFileNames": true, "noEmit": true, "esModuleInterop": true, "module": "esnext", "moduleResolution": "node", "resol...
langchainjs/environment_tests/test-exports-vercel/tsconfig.json/0
{ "file_path": "langchainjs/environment_tests/test-exports-vercel/tsconfig.json", "repo_id": "langchainjs", "token_count": 290 }
835
# for backwards compatibility from llama_index.legacy.schema import QueryBundle, QueryType __all__ = ["QueryBundle", "QueryType"]
llama_index/llama-index-legacy/llama_index/legacy/indices/query/schema.py/0
{ "file_path": "llama_index/llama-index-legacy/llama_index/legacy/indices/query/schema.py", "repo_id": "llama_index", "token_count": 39 }
1,604
import type { BaseLanguageModelInterface } from "@langchain/core/language_models/base"; import { z } from "zod"; import { Example, interpolateFString, FewShotPromptTemplate, } from "@langchain/core/prompts"; import { InputValues } from "@langchain/core/utils/types"; import { QueryTransformer, TraverseType } from ...
langchainjs/langchain/src/chains/query_constructor/index.ts/0
{ "file_path": "langchainjs/langchain/src/chains/query_constructor/index.ts", "repo_id": "langchainjs", "token_count": 1759 }
882
/* eslint-disable no-instanceof/no-instanceof */ import { CacheDelete, CacheListFetch, CacheListPushBack, ICacheClient, InvalidArgumentError, CollectionTtl, } from "@gomomento/sdk-core"; import { BaseListChatMessageHistory } from "@langchain/core/chat_history"; import { BaseMessage, StoredMessage, map...
langchainjs/libs/langchain-community/src/stores/message/momento.ts/0
{ "file_path": "langchainjs/libs/langchain-community/src/stores/message/momento.ts", "repo_id": "langchainjs", "token_count": 2080 }
1,070
# Scale Tests ## Goal Scale tests are designed to check the scalability of Milvus. For instance, if the dataNode pod expands from one to two: - verify the consistency of existing data - verify that the DDL and DML operation is working ## Prerequisite - Kubernetes Cluster - Milvus Operator (refer to [M...
milvus/tests/python_client/scale/README.md/0
{ "file_path": "milvus/tests/python_client/scale/README.md", "repo_id": "milvus", "token_count": 421 }
1,973
import os import threading import time import glob from chaos import constants from yaml import full_load from utils.util_log import test_log as log from delayed_assert import expect import pytest def check_config(chaos_config): if not chaos_config.get("kind", None): raise Exception("kind must be specifie...
milvus/tests/python_client/chaos/chaos_commons.py/0
{ "file_path": "milvus/tests/python_client/chaos/chaos_commons.py", "repo_id": "milvus", "token_count": 1836 }
1,952
from typing import Any, List from langchain_community.document_loaders.unstructured import ( UnstructuredFileLoader, validate_unstructured_version, ) class UnstructuredOrgModeLoader(UnstructuredFileLoader): """Load `Org-Mode` files using `Unstructured`. You can run the loader in one of two modes: "s...
langchain/libs/community/langchain_community/document_loaders/org_mode.py/0
{ "file_path": "langchain/libs/community/langchain_community/document_loaders/org_mode.py", "repo_id": "langchain", "token_count": 628 }
255
--- title: chat-ui emoji: 🔥 colorFrom: purple colorTo: purple sdk: docker pinned: false license: apache-2.0 base_path: /chat app_port: 3000 failure_strategy: rollback --- # Chat UI ![Chat UI repository thumbnail](https://huggingface.co/datasets/huggingface/documentation-images/resolve/main/chatui-websearch.png) A c...
chat-ui/README.md/0
{ "file_path": "chat-ui/README.md", "repo_id": "chat-ui", "token_count": 8361 }
83
python_sources()
llama_index/llama-index-integrations/readers/llama-index-readers-pandas-ai/llama_index/readers/pandas_ai/BUILD/0
{ "file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-pandas-ai/llama_index/readers/pandas_ai/BUILD", "repo_id": "llama_index", "token_count": 6 }
1,422
python_sources()
llama_index/llama-index-integrations/readers/llama-index-readers-snscrape-twitter/llama_index/readers/snscrape_twitter/BUILD/0
{ "file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-snscrape-twitter/llama_index/readers/snscrape_twitter/BUILD", "repo_id": "llama_index", "token_count": 6 }
1,432
# HTML to text >[html2text](https://github.com/Alir3z4/html2text/) is a Python package that converts a page of `HTML` into clean, easy-to-read plain `ASCII text`. The ASCII also happens to be a valid `Markdown` (a text-to-HTML format). ## Installation and Setup ```bash pip install html2text ``` ## Document Transf...
langchain/docs/docs/integrations/providers/html2text.mdx/0
{ "file_path": "langchain/docs/docs/integrations/providers/html2text.mdx", "repo_id": "langchain", "token_count": 154 }
135
# Testcases All-in-one # pod kill # standalone # 3 pods(standalone-ha-blabla, etcd, minio) # cluster-1-node # 11 pods(proxy, rootcoord, querynode, querycoord, datanode, datacoord, # indexnode, indexcoord, pulsar, etcd, minio) # cluster-n-nodes # 11 pods* n: kill one and kil...
milvus/tests/python_client/chaos/chaos_objects/network_partition/testcases.yaml/0
{ "file_path": "milvus/tests/python_client/chaos/chaos_objects/network_partition/testcases.yaml", "repo_id": "milvus", "token_count": 1775 }
1,990
import json from typing import List from langchain.tools.base import BaseTool FINISH_NAME = "finish" class PromptGenerator: """A class for generating custom prompt strings. Does this based on constraints, commands, resources, and performance evaluations. """ def __init__(self) -> None: """...
langchain/libs/experimental/langchain_experimental/autonomous_agents/autogpt/prompt_generator.py/0
{ "file_path": "langchain/libs/experimental/langchain_experimental/autonomous_agents/autogpt/prompt_generator.py", "repo_id": "langchain", "token_count": 2650 }
437
// Code generated by mockery v2.32.4. DO NOT EDIT. package mocks import ( context "context" commonpb "github.com/milvus-io/milvus-proto/go-api/v2/commonpb" datapb "github.com/milvus-io/milvus/internal/proto/datapb" grpc "google.golang.org/grpc" internalpb "github.com/milvus-io/milvus/internal/proto/internalp...
milvus/internal/mocks/mock_datanode_client.go/0
{ "file_path": "milvus/internal/mocks/mock_datanode_client.go", "repo_id": "milvus", "token_count": 21060 }
1,946
from llama_index.embeddings.openai.base import ( OpenAIEmbedding, OpenAIEmbeddingMode, OpenAIEmbeddingModelType, OpenAIEmbeddingModeModel, ) __all__ = [ "OpenAIEmbedding", "OpenAIEmbeddingMode", "OpenAIEmbeddingModelType", "OpenAIEmbeddingModeModel", ]
llama_index/llama-index-integrations/embeddings/llama-index-embeddings-openai/llama_index/embeddings/openai/__init__.py/0
{ "file_path": "llama_index/llama-index-integrations/embeddings/llama-index-embeddings-openai/llama_index/embeddings/openai/__init__.py", "repo_id": "llama_index", "token_count": 130 }
1,372
from typing import List import PIL.Image import PIL.ImageOps from packaging import version from PIL import Image if version.parse(version.parse(PIL.__version__).base_version) >= version.parse("9.1.0"): PIL_INTERPOLATION = { "linear": PIL.Image.Resampling.BILINEAR, "bilinear": PIL.Image.Resampling...
diffusers/src/diffusers/utils/pil_utils.py/0
{ "file_path": "diffusers/src/diffusers/utils/pil_utils.py", "repo_id": "diffusers", "token_count": 849 }
263
import { YandexGPT } from "@langchain/yandex/llms"; const model = new YandexGPT(); const res = await model.invoke(['Translate "I love programming" into French.']); console.log({ res });
langchainjs/examples/src/models/llm/yandex.ts/0
{ "file_path": "langchainjs/examples/src/models/llm/yandex.ts", "repo_id": "langchainjs", "token_count": 61 }
814
from langchain_community.callbacks.streamlit.streamlit_callback_handler import ( CHECKMARK_EMOJI, EXCEPTION_EMOJI, HISTORY_EMOJI, THINKING_EMOJI, LLMThought, LLMThoughtLabeler, LLMThoughtState, StreamlitCallbackHandler, ToolRecord, ) __all__ = [ "CHECKMARK_EMOJI", "THINKING_...
langchain/libs/langchain/langchain/callbacks/streamlit/streamlit_callback_handler.py/0
{ "file_path": "langchain/libs/langchain/langchain/callbacks/streamlit/streamlit_callback_handler.py", "repo_id": "langchain", "token_count": 236 }
476
<jupyter_start><jupyter_text>Google TrendsThis notebook goes over how to use the Google Trends Tool to fetch trends information.First, you need to sign up for an `SerpApi key` key at: https://serpapi.com/users/sign_up.Then you must install `google-search-results` with the command:`pip install google-search-results`Then...
langchain/docs/docs/integrations/tools/google_trends.ipynb/0
{ "file_path": "langchain/docs/docs/integrations/tools/google_trends.ipynb", "repo_id": "langchain", "token_count": 274 }
184
# Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
diffusers/src/diffusers/models/transformers/transformer_temporal.py/0
{ "file_path": "diffusers/src/diffusers/models/transformers/transformer_temporal.py", "repo_id": "diffusers", "token_count": 7286 }
230
import { GoogleAuth, GoogleAuthOptions } from "google-auth-library"; import { Embeddings, EmbeddingsParams } from "@langchain/core/embeddings"; import { AsyncCallerCallOptions } from "@langchain/core/utils/async_caller"; import { GoogleVertexAIBaseLLMInput, GoogleVertexAIBasePrediction, GoogleVertexAILLMPredictio...
langchainjs/langchain/src/experimental/multimodal_embeddings/googlevertexai.ts/0
{ "file_path": "langchainjs/langchain/src/experimental/multimodal_embeddings/googlevertexai.ts", "repo_id": "langchainjs", "token_count": 2172 }
939
python_sources()
llama_index/llama-index-legacy/llama_index/legacy/evaluation/retrieval/BUILD/0
{ "file_path": "llama_index/llama-index-legacy/llama_index/legacy/evaluation/retrieval/BUILD", "repo_id": "llama_index", "token_count": 6 }
1,590
# coding=utf-8 # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
transformers/tests/models/gpt_neox_japanese/test_tokenization_gpt_neox_japanese.py/0
{ "file_path": "transformers/tests/models/gpt_neox_japanese/test_tokenization_gpt_neox_japanese.py", "repo_id": "transformers", "token_count": 2273 }
728
python_sources()
llama_index/llama-index-legacy/llama_index/legacy/agent/legacy/BUILD/0
{ "file_path": "llama_index/llama-index-legacy/llama_index/legacy/agent/legacy/BUILD", "repo_id": "llama_index", "token_count": 6 }
1,547
# Meilisearch > [Meilisearch](https://meilisearch.com) is an open-source, lightning-fast, and hyper > relevant search engine. > It comes with great defaults to help developers build snappy search experiences. > > You can [self-host Meilisearch](https://www.meilisearch.com/docs/learn/getting_started/installation#loca...
langchain/docs/docs/integrations/providers/meilisearch.mdx/0
{ "file_path": "langchain/docs/docs/integrations/providers/meilisearch.mdx", "repo_id": "langchain", "token_count": 261 }
152
from langchain_community.tools.edenai.audio_text_to_speech import EdenAiTextToSpeechTool __all__ = ["EdenAiTextToSpeechTool"]
langchain/libs/langchain/langchain/tools/edenai/audio_text_to_speech.py/0
{ "file_path": "langchain/libs/langchain/langchain/tools/edenai/audio_text_to_speech.py", "repo_id": "langchain", "token_count": 45 }
548
""" HRNet Copied from https://github.com/HRNet/HRNet-Image-Classification Original header: Copyright (c) Microsoft Licensed under the MIT License. Written by Bin Xiao (Bin.Xiao@microsoft.com) Modified by Ke Sun (sunk@mail.ustc.edu.cn) """ import logging from typing import List import torch import torch.nn as...
pytorch-image-models/timm/models/hrnet.py/0
{ "file_path": "pytorch-image-models/timm/models/hrnet.py", "repo_id": "pytorch-image-models", "token_count": 17584 }
364
from langchain_core.utils.input import ( get_bolded_text, get_color_mapping, get_colored_text, print_text, ) __all__ = ["get_color_mapping", "get_colored_text", "get_bolded_text", "print_text"]
langchain/libs/langchain/langchain/utils/input.py/0
{ "file_path": "langchain/libs/langchain/langchain/utils/input.py", "repo_id": "langchain", "token_count": 88 }
577
# Updating to v0.10.0 With the introduction of LlamaIndex v0.10.0, there were several changes - integrations have separate `pip installs (See the [full registry](https://pretty-sodium-5e0.notion.site/ce81b247649a44e4b6b35dfb24af28a6?v=53b3c2ced7bb4c9996b81b83c9f01139)) - many imports changed - the service context was...
llama_index/docs/getting_started/v0_10_0_migration.md/0
{ "file_path": "llama_index/docs/getting_started/v0_10_0_migration.md", "repo_id": "llama_index", "token_count": 716 }
1,222
"""MultiOn Toolkit."""
langchain/libs/community/langchain_community/agent_toolkits/multion/__init__.py/0
{ "file_path": "langchain/libs/community/langchain_community/agent_toolkits/multion/__init__.py", "repo_id": "langchain", "token_count": 8 }
223
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/internal/rootcoord/quota_center.go/0
{ "file_path": "milvus/internal/rootcoord/quota_center.go", "repo_id": "milvus", "token_count": 12059 }
1,864
# flake8: noqa ENDPOINT_DESCRIPTION = "Ask questions about LangChain documentation!" ENDPOINT_NAME = "ask-langchain" INPUT_NAME = "query" OUTPUT_KEY = "result" NAME_FOR_MODEL = "langchainQABot" NAME_FOR_HUMAN = "LangChain QA Bot" DESCRIPTION_FOR_MODEL = "This plugin provides access to a LangChain QA Bot to answer quest...
langchain-aiplugin/retrieval_qa/constants.py/0
{ "file_path": "langchain-aiplugin/retrieval_qa/constants.py", "repo_id": "langchain-aiplugin", "token_count": 138 }
66
# Generated content DO NOT EDIT class Trainer: """ Base class for all trainers This class is not supposed to be instantiated directly. Instead, any implementation of a Trainer will return an instance of this class when instantiated. """ class BpeTrainer(Trainer): """ Trainer capable of tra...
tokenizers/bindings/python/py_src/tokenizers/trainers/__init__.pyi/0
{ "file_path": "tokenizers/bindings/python/py_src/tokenizers/trainers/__init__.pyi", "repo_id": "tokenizers", "token_count": 2178 }
423
# Run Milvus standalone through binary files To quickly install Milvus standalone without docker or kubernetes, this document provides a tutorial for installing Milvus and dependencies, etcd and MinIO, through binary files. Before installing, you can refer to [docker-compose.yml](https://github.com/milvus-io/milvus/b...
milvus/deployments/binary/README.md/0
{ "file_path": "milvus/deployments/binary/README.md", "repo_id": "milvus", "token_count": 587 }
1,636
from llama_index.multi_modal_llms.replicate.base import ReplicateMultiModal __all__ = ["ReplicateMultiModal"]
llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-replicate/llama_index/multi_modal_llms/replicate/__init__.py/0
{ "file_path": "llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-replicate/llama_index/multi_modal_llms/replicate/__init__.py", "repo_id": "llama_index", "token_count": 38 }
1,256
"""Init file.""" from llama_index.readers.openalex.base import OpenAlexReader __all__ = ["OpenAlexReader"]
llama_index/llama-index-integrations/readers/llama-index-readers-openalex/llama_index/readers/openalex/__init__.py/0
{ "file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-openalex/llama_index/readers/openalex/__init__.py", "repo_id": "llama_index", "token_count": 36 }
1,395
python_tests()
llama_index/llama-index-packs/llama-index-packs-neo4j-query-engine/tests/BUILD/0
{ "file_path": "llama_index/llama-index-packs/llama-index-packs-neo4j-query-engine/tests/BUILD", "repo_id": "llama_index", "token_count": 5 }
1,669
import _ from "lodash"; import { LLMOptions, Portkey as _Portkey } from "portkey-ai"; import { CallbackManagerForLLMRun } from "@langchain/core/callbacks/manager"; import { GenerationChunk, LLMResult } from "@langchain/core/outputs"; import { getEnvironmentVariable } from "@langchain/core/utils/env"; import { BaseLLM }...
langchainjs/libs/langchain-community/src/llms/portkey.ts/0
{ "file_path": "langchainjs/libs/langchain-community/src/llms/portkey.ts", "repo_id": "langchainjs", "token_count": 1719 }
958
from typing import TYPE_CHECKING from ...utils import DIFFUSERS_SLOW_IMPORT, _LazyModule _import_structure = {"pipeline_ddim": ["DDIMPipeline"]} if TYPE_CHECKING or DIFFUSERS_SLOW_IMPORT: from .pipeline_ddim import DDIMPipeline else: import sys sys.modules[__name__] = _LazyModule( __name__, ...
diffusers/src/diffusers/pipelines/ddim/__init__.py/0
{ "file_path": "diffusers/src/diffusers/pipelines/ddim/__init__.py", "repo_id": "diffusers", "token_count": 180 }
249
from typing import List from langchain.base_language import BaseLanguageModel from langchain.tools.base import BaseTool from langchain_experimental.autonomous_agents.hugginggpt.repsonse_generator import ( load_response_generator, ) from langchain_experimental.autonomous_agents.hugginggpt.task_executor import ( ...
langchain/libs/experimental/langchain_experimental/autonomous_agents/hugginggpt/hugginggpt.py/0
{ "file_path": "langchain/libs/experimental/langchain_experimental/autonomous_agents/hugginggpt/hugginggpt.py", "repo_id": "langchain", "token_count": 428 }
438
<jupyter_start><jupyter_text>Classification de token (TensorFlow) Installez les bibliothèques 🤗 *Datasets*, 🤗 *Transformers* et 🤗 *Accelerate* pour exécuter ce *notebook*.<jupyter_code>!pip install datasets transformers[sentencepiece] !apt install git-lfs<jupyter_output><empty_output><jupyter_text>Vous aurez besoin...
notebooks/course/fr/chapter7/section2_tf.ipynb/0
{ "file_path": "notebooks/course/fr/chapter7/section2_tf.ipynb", "repo_id": "notebooks", "token_count": 2616 }
314
insert_performance: collections: - milvus: db_config.primary_path: /test/milvus/db_data_011/cluster/sift_1m_128_l2 cache_config.cpu_cache_capacity: 4GB engine_config.use_blas_threshold: 1100 engine_config.gpu_search_threshold: 1 gpu_resource_config.enable: true...
milvus/tests/benchmark/milvus_benchmark/suites/2_insert_cluster.yaml/0
{ "file_path": "milvus/tests/benchmark/milvus_benchmark/suites/2_insert_cluster.yaml", "repo_id": "milvus", "token_count": 391 }
1,939
// Code generated by mockery v2.33.3. DO NOT EDIT. package mocks import ( dbmodel "github.com/chroma/chroma-coordinator/internal/metastore/db/dbmodel" mock "github.com/stretchr/testify/mock" ) // IDatabaseDb is an autogenerated mock type for the IDatabaseDb type type IDatabaseDb struct { mock.Mock } // DeleteAll...
chroma/go/coordinator/internal/metastore/db/dbmodel/mocks/IDatabaseDb.go/0
{ "file_path": "chroma/go/coordinator/internal/metastore/db/dbmodel/mocks/IDatabaseDb.go", "repo_id": "chroma", "token_count": 1016 }
52
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
transformers/tests/repo_utils/test_check_dummies.py/0
{ "file_path": "transformers/tests/repo_utils/test_check_dummies.py", "repo_id": "transformers", "token_count": 1800 }
753
# MixNet **MixNet** is a type of convolutional neural network discovered via AutoML that utilises [MixConvs](https://paperswithcode.com/method/mixconv) instead of regular [depthwise convolutions](https://paperswithcode.com/method/depthwise-convolution). {% include 'code_snippets.md' %} ## How do I train this model? ...
pytorch-image-models/docs/models/.templates/models/mixnet.md/0
{ "file_path": "pytorch-image-models/docs/models/.templates/models/mixnet.md", "repo_id": "pytorch-image-models", "token_count": 1878 }
345
use candle::{Result, Tensor}; pub fn apply_repeat_penalty(logits: &Tensor, penalty: f32, context: &[u32]) -> Result<Tensor> { let device = logits.device(); let mut logits = logits.to_vec1::<f32>()?; let context: std::collections::HashSet<_> = context.iter().collect(); for (token_id, logit) in logits.it...
candle/candle-transformers/src/utils.rs/0
{ "file_path": "candle/candle-transformers/src/utils.rs", "repo_id": "candle", "token_count": 299 }
75
import { load } from "@tensorflow-models/universal-sentence-encoder"; import * as tf from "@tensorflow/tfjs-core"; import { Embeddings, type EmbeddingsParams } from "@langchain/core/embeddings"; /** * Interface that extends EmbeddingsParams and defines additional * parameters specific to the TensorFlowEmbeddings cl...
langchainjs/libs/langchain-community/src/embeddings/tensorflow.ts/0
{ "file_path": "langchainjs/libs/langchain-community/src/embeddings/tensorflow.ts", "repo_id": "langchainjs", "token_count": 909 }
967
#!/usr/bin/env python # coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
transformers/examples/pytorch/token-classification/run_ner_no_trainer.py/0
{ "file_path": "transformers/examples/pytorch/token-classification/run_ner_no_trainer.py", "repo_id": "transformers", "token_count": 14733 }
592
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/internal/rootcoord/mock_test.go/0
{ "file_path": "milvus/internal/rootcoord/mock_test.go", "repo_id": "milvus", "token_count": 15420 }
2,058
#!/usr/bin/env python # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
accelerate/examples/by_feature/deepspeed_with_config_support.py/0
{ "file_path": "accelerate/examples/by_feature/deepspeed_with_config_support.py", "repo_id": "accelerate", "token_count": 12377 }
5
export { SerpAPI, type SerpAPIParameters } from "./serpapi.js"; export { DadJokeAPI } from "./dadjokeapi.js"; export { BingSerpAPI } from "./bingserpapi.js"; export { Tool, type ToolParams, StructuredTool } from "./base.js"; export { DynamicTool, type DynamicToolInput, DynamicStructuredTool, type DynamicStructu...
langchainjs/langchain/src/tools/index.ts/0
{ "file_path": "langchainjs/langchain/src/tools/index.ts", "repo_id": "langchainjs", "token_count": 560 }
957
poetry_requirements( name="poetry", )
llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-jaguar/BUILD/0
{ "file_path": "llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-jaguar/BUILD", "repo_id": "llama_index", "token_count": 18 }
1,657
# coding=utf-8 # Copyright 2022 The HuggingFace Team Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a clone of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
transformers/tests/quantization/bnb/test_mixed_int8.py/0
{ "file_path": "transformers/tests/quantization/bnb/test_mixed_int8.py", "repo_id": "transformers", "token_count": 15531 }
833
import { DataSource } from "typeorm"; import { SqlDatabase } from "langchain/sql_db"; import { ChatOpenAI } from "@langchain/openai"; import { PromptTemplate } from "@langchain/core/prompts"; import { RunnableSequence } from "@langchain/core/runnables"; import { StringOutputParser } from "@langchain/core/output_parsers...
langchainjs/examples/src/chains/sql_db.ts/0
{ "file_path": "langchainjs/examples/src/chains/sql_db.ts", "repo_id": "langchainjs", "token_count": 1040 }
794