text stringlengths 3 1.68M | id stringlengths 13 169 | metadata dict | __index_level_0__ int64 0 2.21k |
|---|---|---|---|
openapi: 3.0.0
info:
title: OpenAI API
description: APIs for sampling from and fine-tuning language models
version: '1.1.0'
servers:
- url: https://api.openai.com/v1
tags:
- name: OpenAI
description: The OpenAI REST API
paths:
/engines:
get:
operationId: listEngines
deprecated: true
ta... | langchainjs/examples/openai_openapi.yaml/0 | {
"file_path": "langchainjs/examples/openai_openapi.yaml",
"repo_id": "langchainjs",
"token_count": 54742
} | 807 |
# coding=utf-8
# Copyright 2018 LXMERT Authors, The Hugging Face Team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required b... | transformers/tests/models/lxmert/test_tokenization_lxmert.py/0 | {
"file_path": "transformers/tests/models/lxmert/test_tokenization_lxmert.py",
"repo_id": "transformers",
"token_count": 1286
} | 754 |
python_sources()
| llama_index/llama-index-packs/llama-index-packs-arize-phoenix-query-engine/examples/BUILD/0 | {
"file_path": "llama_index/llama-index-packs/llama-index-packs-arize-phoenix-query-engine/examples/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,643 |
import type { Runnable } from "@langchain/core/runnables";
import { BaseOutputParser } from "@langchain/core/output_parsers";
import type { AgentAction, AgentFinish } from "@langchain/core/agents";
import type { BaseMessage } from "@langchain/core/messages";
import type { ChainValues } from "@langchain/core/utils/types... | langchainjs/langchain/src/agents/types.ts/0 | {
"file_path": "langchainjs/langchain/src/agents/types.ts",
"repo_id": "langchainjs",
"token_count": 829
} | 856 |
from typing import TYPE_CHECKING
from ....utils import DIFFUSERS_SLOW_IMPORT, _LazyModule
_import_structure = {"pipeline_pndm": ["PNDMPipeline"]}
if TYPE_CHECKING or DIFFUSERS_SLOW_IMPORT:
from .pipeline_pndm import PNDMPipeline
else:
import sys
sys.modules[__name__] = _LazyModule(
__name__,
... | diffusers/src/diffusers/pipelines/deprecated/pndm/__init__.py/0 | {
"file_path": "diffusers/src/diffusers/pipelines/deprecated/pndm/__init__.py",
"repo_id": "diffusers",
"token_count": 182
} | 247 |
# coding=utf-8
# Copyright 2020 Microsoft and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required... | transformers/src/transformers/models/deberta/tokenization_deberta_fast.py/0 | {
"file_path": "transformers/src/transformers/models/deberta/tokenization_deberta_fast.py",
"repo_id": "transformers",
"token_count": 5211
} | 627 |
import sys
import unittest
from unittest.mock import MagicMock
import pytest
from llama_index.core.schema import NodeRelationship, RelatedNodeInfo, TextNode
from llama_index.core.vector_stores.types import VectorStoreQuery, VectorStoreQueryMode
from llama_index.vector_stores.cassandra import CassandraVectorStore
try:... | llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-cassandra/tests/test_cassandra.py/0 | {
"file_path": "llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-cassandra/tests/test_cassandra.py",
"repo_id": "llama_index",
"token_count": 1931
} | 1,513 |
poetry_requirements(
name="poetry",
)
| llama_index/llama-index-integrations/tools/llama-index-tools-wolfram-alpha/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/tools/llama-index-tools-wolfram-alpha/BUILD",
"repo_id": "llama_index",
"token_count": 18
} | 1,537 |
import { test } from "@jest/globals";
import { Cohere } from "../cohere.js";
test("Test Cohere", async () => {
const model = new Cohere({ maxTokens: 20 });
const res = await model.call("1 + 1 =");
console.log(res);
}, 50000);
| langchainjs/libs/langchain-community/src/llms/tests/cohere.int.test.ts/0 | {
"file_path": "langchainjs/libs/langchain-community/src/llms/tests/cohere.int.test.ts",
"repo_id": "langchainjs",
"token_count": 82
} | 993 |
from llama_index.core.multi_modal_llms.base import (
MultiModalLLM,
MultiModalLLMMetadata,
)
__all__ = [
"MultiModalLLMMetadata",
"MultiModalLLM",
]
| llama_index/llama-index-core/llama_index/core/multi_modal_llms/__init__.py/0 | {
"file_path": "llama_index/llama-index-core/llama_index/core/multi_modal_llms/__init__.py",
"repo_id": "llama_index",
"token_count": 77
} | 1,189 |
from __future__ import annotations
import tempfile
from copy import deepcopy
from pathlib import Path
from typing import TYPE_CHECKING, Any, Dict, List, Mapping, Optional, Sequence
from langchain_core.agents import AgentAction, AgentFinish
from langchain_core.callbacks import BaseCallbackHandler
from langchain_core.o... | langchain/libs/community/langchain_community/callbacks/clearml_callback.py/0 | {
"file_path": "langchain/libs/community/langchain_community/callbacks/clearml_callback.py",
"repo_id": "langchain",
"token_count": 9142
} | 212 |
//! Ancestral sampling with Euler method steps.
//!
//! Reference implementation in Rust:
//!
//! https://github.com/pykeio/diffusers/blob/250b9ad1898af41e76a74c0d8d4292652823338a/src/schedulers/euler_ancestral_discrete.rs
//!
//! Based on the original [`k-diffusion` implementation by Katherine Crowson][kd].
///
/// [k... | candle/candle-transformers/src/models/stable_diffusion/euler_ancestral_discrete.rs/0 | {
"file_path": "candle/candle-transformers/src/models/stable_diffusion/euler_ancestral_discrete.rs",
"repo_id": "candle",
"token_count": 4176
} | 68 |
import multiprocessing
from concurrent.futures import Future, ThreadPoolExecutor, wait
import random
import threading
from typing import Any, Dict, List, Optional, Set, Tuple, cast
import numpy as np
from chromadb.api import ServerAPI
import chromadb.test.property.invariants as invariants
from chromadb.test.property.s... | chroma/chromadb/test/test_multithreaded.py/0 | {
"file_path": "chroma/chromadb/test/test_multithreaded.py",
"repo_id": "chroma",
"token_count": 3671
} | 28 |
from pathlib import Path
from langchain_community.document_loaders import FacebookChatLoader
def test_facebook_chat_loader() -> None:
"""Test FacebookChatLoader."""
file_path = Path(__file__).parent.parent / "examples/facebook_chat.json"
loader = FacebookChatLoader(str(file_path))
docs = loader.load(... | langchain/libs/community/tests/integration_tests/document_loaders/test_facebook_chat.py/0 | {
"file_path": "langchain/libs/community/tests/integration_tests/document_loaders/test_facebook_chat.py",
"repo_id": "langchain",
"token_count": 564
} | 354 |
import * as uuid from "uuid";
import { ClickHouseClient, createClient } from "@clickhouse/client";
import type { EmbeddingsInterface } from "@langchain/core/embeddings";
import { VectorStore } from "@langchain/core/vectorstores";
import { Document } from "@langchain/core/documents";
/**
* Arguments for the MyScaleSt... | langchainjs/libs/langchain-community/src/vectorstores/myscale.ts/0 | {
"file_path": "langchainjs/libs/langchain-community/src/vectorstores/myscale.ts",
"repo_id": "langchainjs",
"token_count": 3231
} | 1,014 |
import { expect, test } from "@jest/globals";
import { PromptTemplate } from "@langchain/core/prompts";
import { PRINCIPLES } from "../../../chains/index.js";
import { ChatAnthropic } from "../../../chat_models/anthropic.js";
import { loadEvaluator } from "../../loader.js";
test("Test CriteriaEvalChain", async () => {... | langchainjs/langchain/src/evaluation/criteria/tests/criteria_eval_chain.int.test.ts/0 | {
"file_path": "langchainjs/langchain/src/evaluation/criteria/tests/criteria_eval_chain.int.test.ts",
"repo_id": "langchainjs",
"token_count": 1287
} | 919 |
# coding=utf-8
# Copyright 2023 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | alignment-handbook/src/alignment/model_utils.py/0 | {
"file_path": "alignment-handbook/src/alignment/model_utils.py",
"repo_id": "alignment-handbook",
"token_count": 1649
} | 21 |
## Appendix E. Statistics
| milvus/docs/developer_guides/appendix_e_statistics.md/0 | {
"file_path": "milvus/docs/developer_guides/appendix_e_statistics.md",
"repo_id": "milvus",
"token_count": 6
} | 1,645 |
// Copyright (C) 2019-2020 Zilliz. All rights reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable l... | milvus/internal/mq/mqimpl/rocksmq/server/rocksmq_impl.go/0 | {
"file_path": "milvus/internal/mq/mqimpl/rocksmq/server/rocksmq_impl.go",
"repo_id": "milvus",
"token_count": 12721
} | 2,004 |
package backend
import (
"fmt"
"github.com/milvus-io/milvus/cmd/tools/migration/configs"
"github.com/milvus-io/milvus/cmd/tools/migration/meta"
"github.com/milvus-io/milvus/internal/metastore/kv/querycoord"
"github.com/milvus-io/milvus/internal/metastore/kv/rootcoord"
"github.com/milvus-io/milvus/pkg/util"
)
/... | milvus/cmd/tools/migration/backend/etcd220.go/0 | {
"file_path": "milvus/cmd/tools/migration/backend/etcd220.go",
"repo_id": "milvus",
"token_count": 1221
} | 1,617 |
python_sources()
| llama_index/llama-index-integrations/readers/llama-index-readers-jaguar/llama_index/readers/jaguar/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-jaguar/llama_index/readers/jaguar/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,511 |
// Copyright (C) 2019-2020 Zilliz. All rights reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable l... | milvus/internal/core/src/indexbuilder/IndexCreatorBase.h/0 | {
"file_path": "milvus/internal/core/src/indexbuilder/IndexCreatorBase.h",
"repo_id": "milvus",
"token_count": 402
} | 1,756 |
"""Token splitter."""
import logging
from typing import Callable, List, Optional
from llama_index.core.bridge.pydantic import Field, PrivateAttr
from llama_index.core.callbacks.base import CallbackManager
from llama_index.core.callbacks.schema import CBEventType, EventPayload
from llama_index.core.constants import DEF... | llama_index/llama-index-core/llama_index/core/node_parser/text/token.py/0 | {
"file_path": "llama_index/llama-index-core/llama_index/core/node_parser/text/token.py",
"repo_id": "llama_index",
"token_count": 3799
} | 1,280 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/datacoord/cluster.go/0 | {
"file_path": "milvus/internal/datacoord/cluster.go",
"repo_id": "milvus",
"token_count": 1935
} | 1,960 |
# Copyright 2024 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | diffusers/src/diffusers/models/unets/unet_1d_blocks.py/0 | {
"file_path": "diffusers/src/diffusers/models/unets/unet_1d_blocks.py",
"repo_id": "diffusers",
"token_count": 12023
} | 231 |
from typing import cast
from llama_index.core.indices.vector_store.retrievers.auto_retriever.output_parser import (
VectorStoreQueryOutputParser,
)
from llama_index.core.output_parsers.base import StructuredOutput
from llama_index.core.vector_stores.types import (
ExactMatchFilter,
VectorStoreQuerySpec,
)
... | llama_index/llama-index-core/tests/indices/vector_store/auto_retriever/test_output_parser.py/0 | {
"file_path": "llama_index/llama-index-core/tests/indices/vector_store/auto_retriever/test_output_parser.py",
"repo_id": "llama_index",
"token_count": 548
} | 1,317 |
Documentation
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
The Rust API Reference is available directly on the `Docs.rs <https://docs.rs/tokenizers>`__
website.
| tokenizers/docs/source/api/rust.inc/0 | {
"file_path": "tokenizers/docs/source/api/rust.inc",
"repo_id": "tokenizers",
"token_count": 43
} | 413 |
from llama_index.legacy.evaluation.benchmarks.beir import BeirEvaluator
from llama_index.legacy.evaluation.benchmarks.hotpotqa import HotpotQAEvaluator
__all__ = ["BeirEvaluator", "HotpotQAEvaluator"]
| llama_index/llama-index-legacy/llama_index/legacy/evaluation/benchmarks/__init__.py/0 | {
"file_path": "llama_index/llama-index-legacy/llama_index/legacy/evaluation/benchmarks/__init__.py",
"repo_id": "llama_index",
"token_count": 74
} | 1,589 |
import ThumbsUpIcon from "../../assets/ThumbsUpIcon.svg?react";
import ThumbsDownIcon from "../../assets/ThumbsDownIcon.svg?react";
import CircleSpinIcon from "../../assets/CircleSpinIcon.svg?react";
import { resolveApiUrl } from "../../utils/url";
import { useState } from "react";
import { cn } from "../../utils/cn";
... | langserve/langserve/playground/src/components/feedback/CorrectnessFeedback.tsx/0 | {
"file_path": "langserve/langserve/playground/src/components/feedback/CorrectnessFeedback.tsx",
"repo_id": "langserve",
"token_count": 1268
} | 1,130 |
# Full-Stack Projects
We've created both tooling and a variety of example projects (all open-source) to help you get started building a full-stack LLM application.
## create-llama
`create-llama` is a command-line tool that will generate a full-stack application template for you. It supports both FastAPI, Vercel, and... | llama_index/docs/community/full_stack_projects.md/0 | {
"file_path": "llama_index/docs/community/full_stack_projects.md",
"repo_id": "llama_index",
"token_count": 535
} | 1,166 |
python_sources()
| llama_index/llama-index-core/llama_index/core/evaluation/retrieval/BUILD/0 | {
"file_path": "llama_index/llama-index-core/llama_index/core/evaluation/retrieval/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,186 |
import { getEnvironmentVariable } from "@langchain/core/utils/env";
import { Embeddings, EmbeddingsParams } from "@langchain/core/embeddings";
import { chunkArray } from "@langchain/core/utils/chunk_array";
import { ConfigurationParameters } from "../chat_models/minimax.js";
/**
* Interface for MinimaxEmbeddings para... | langchainjs/libs/langchain-community/src/embeddings/minimax.ts/0 | {
"file_path": "langchainjs/libs/langchain-community/src/embeddings/minimax.ts",
"repo_id": "langchainjs",
"token_count": 2428
} | 996 |
<jupyter_start><jupyter_code>!git clone https://huggingface.co/spaces/smangrul/peft-lora-sd-dreambooth
%cd "peft-lora-sd-dreambooth"
!pip install -r requirements.txt
!python colab.py<jupyter_output><empty_output> | peft/examples/lora_dreambooth/colab_notebook.ipynb/0 | {
"file_path": "peft/examples/lora_dreambooth/colab_notebook.ipynb",
"repo_id": "peft",
"token_count": 91
} | 301 |
"""Base reader class."""
from abc import ABC
from typing import TYPE_CHECKING, Any, Dict, Iterable, List
if TYPE_CHECKING:
from llama_index.core.bridge.langchain import Document as LCDocument
from llama_index.core.bridge.pydantic import Field
from llama_index.core.schema import BaseComponent, Document
class Base... | llama_index/llama-index-core/llama_index/core/readers/base.py/0 | {
"file_path": "llama_index/llama-index-core/llama_index/core/readers/base.py",
"repo_id": "llama_index",
"token_count": 919
} | 1,243 |
# LlamaIndex Vector_Stores Integration: Timescalevector
| llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-timescalevector/README.md/0 | {
"file_path": "llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-timescalevector/README.md",
"repo_id": "llama_index",
"token_count": 14
} | 1,540 |
import sys
from typing import Any, Dict, List, Optional
from unittest.mock import MagicMock
import numpy as np
from llama_index.legacy.storage.storage_context import StorageContext
from llama_index.legacy.vector_stores.pinecone import PineconeVectorStore
from tests.mock_utils.mock_utils import mock_tokenizer
class ... | llama_index/llama-index-legacy/tests/indices/vector_store/utils.py/0 | {
"file_path": "llama_index/llama-index-legacy/tests/indices/vector_store/utils.py",
"repo_id": "llama_index",
"token_count": 992
} | 1,742 |
import json
from typing import List
from langchain_core.documents import Document
from langchain_core.utils import stringify_dict
from langchain_community.document_loaders.base import BaseLoader
class AirbyteJSONLoader(BaseLoader):
"""Load local `Airbyte` json files."""
def __init__(self, file_path: str):
... | langchain/libs/community/langchain_community/document_loaders/airbyte_json.py/0 | {
"file_path": "langchain/libs/community/langchain_community/document_loaders/airbyte_json.py",
"repo_id": "langchain",
"token_count": 296
} | 240 |
from langchain_community.retrievers.llama_index import (
LlamaIndexGraphRetriever,
LlamaIndexRetriever,
)
__all__ = ["LlamaIndexRetriever", "LlamaIndexGraphRetriever"]
| langchain/libs/langchain/langchain/retrievers/llama_index.py/0 | {
"file_path": "langchain/libs/langchain/langchain/retrievers/llama_index.py",
"repo_id": "langchain",
"token_count": 67
} | 540 |
{
"candidates": [
{
"content": {
"parts": [
{
"text": "T"
}
],
"role": "model"
},
"finishReason": "SAFETY",
"index": 0,
"safetyRatings": [
{
"category": "HARM_CATEGORY_SEXUALLY_EXPLICIT",
"probability... | langchainjs/libs/langchain-google-common/src/tests/data/chat-1-mock.json/0 | {
"file_path": "langchainjs/libs/langchain-google-common/src/tests/data/chat-1-mock.json",
"repo_id": "langchainjs",
"token_count": 706
} | 989 |
# coding=utf-8
# Copyright 2024 HuggingFace Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or ag... | diffusers/examples/unconditional_image_generation/test_unconditional.py/0 | {
"file_path": "diffusers/examples/unconditional_image_generation/test_unconditional.py",
"repo_id": "diffusers",
"token_count": 2492
} | 227 |
from abc import ABC, abstractmethod
from typing import Any, Dict, List, Optional, Tuple
from langchain.callbacks.manager import CallbackManagerForChainRun
from langchain.chains.base import Chain
from langchain_experimental.tot.thought import ThoughtValidity
class ToTChecker(Chain, ABC):
"""
Tree of Thought ... | langchain/libs/experimental/langchain_experimental/tot/checker.py/0 | {
"file_path": "langchain/libs/experimental/langchain_experimental/tot/checker.py",
"repo_id": "langchain",
"token_count": 528
} | 439 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/pkg/util/lifetime/lifetime_test.go/0 | {
"file_path": "milvus/pkg/util/lifetime/lifetime_test.go",
"repo_id": "milvus",
"token_count": 562
} | 1,834 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/distributed/querycoord/client/client_test.go/0 | {
"file_path": "milvus/internal/distributed/querycoord/client/client_test.go",
"repo_id": "milvus",
"token_count": 2149
} | 1,817 |
# Ingestion Pipeline
An `IngestionPipeline` uses a concept of `Transformations` that are applied to input data. These `Transformations` are applied to your input data, and the resulting nodes are either returned or inserted into a vector database (if given). Each node+transformation pair is cached, so that subsequent ... | llama_index/docs/module_guides/loading/ingestion_pipeline/root.md/0 | {
"file_path": "llama_index/docs/module_guides/loading/ingestion_pipeline/root.md",
"repo_id": "llama_index",
"token_count": 2213
} | 1,178 |
<jupyter_start><jupyter_text>AsyncIn this example we will build a chat executor with native async implementations of the core logic. This enables taking advantage of Chat Models which have async clients, removing the need for calling the model in a separate thread. SetupFirst we need to install the packages required<j... | langgraph/examples/async.ipynb/0 | {
"file_path": "langgraph/examples/async.ipynb",
"repo_id": "langgraph",
"token_count": 3967
} | 1,028 |
# LlamaIndex Readers Integration: Steamship
| llama_index/llama-index-integrations/readers/llama-index-readers-steamship/README.md/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-steamship/README.md",
"repo_id": "llama_index",
"token_count": 11
} | 1,436 |
build_performance:
collections:
-
milvus:
db_config.primary_path: /test/milvus/db_data_011/sift_10m_100000_128_l2_sq8_4096
cache_config.cpu_cache_capacity: 32GB
engine_config.use_blas_threshold: 1100
engine_config.gpu_search_threshold: 1
gpu_resource_config.enable: tr... | milvus/tests/benchmark/milvus_benchmark/suites/011_gpu_build.yaml/0 | {
"file_path": "milvus/tests/benchmark/milvus_benchmark/suites/011_gpu_build.yaml",
"repo_id": "milvus",
"token_count": 328
} | 2,001 |
# PDF files
You can use this version of the popular PDFLoader in web environments.
By default, one document will be created for each page in the PDF file, you can change this behavior by setting the `splitPages` option to `false`.
## Setup
```bash npm2yarn
npm install pdf-parse
```
## Usage
import CodeBlock from "... | langchainjs/docs/core_docs/docs/integrations/document_loaders/web_loaders/pdf.mdx/0 | {
"file_path": "langchainjs/docs/core_docs/docs/integrations/document_loaders/web_loaders/pdf.mdx",
"repo_id": "langchainjs",
"token_count": 600
} | 781 |
from langchain_experimental.plan_and_execute.agent_executor import PlanAndExecute
from langchain_experimental.plan_and_execute.executors.agent_executor import (
load_agent_executor,
)
from langchain_experimental.plan_and_execute.planners.chat_planner import (
load_chat_planner,
)
__all__ = ["PlanAndExecute", "... | langchain/libs/experimental/langchain_experimental/plan_and_execute/__init__.py/0 | {
"file_path": "langchain/libs/experimental/langchain_experimental/plan_and_execute/__init__.py",
"repo_id": "langchain",
"token_count": 124
} | 445 |
[tool.poetry]
name = "rag-pinecone-multi-query"
version = "0.1.0"
description = "RAG with multi-query retriever using Pinecone"
authors = [
"Lance Martin <lance@langchain.dev>",
]
readme = "README.md"
[tool.poetry.dependencies]
python = ">=3.8.1,<4.0"
langchain = "^0.1"
openai = "<2"
tiktoken = ">=0.5.1"
pinecone-... | langchain/templates/rag-pinecone-multi-query/pyproject.toml/0 | {
"file_path": "langchain/templates/rag-pinecone-multi-query/pyproject.toml",
"repo_id": "langchain",
"token_count": 302
} | 662 |
## init
from llama_index.tools.azure_translate.base import (
ENDPOINT_BASE_URL,
AzureTranslateToolSpec,
)
__all__ = ["AzureTranslateToolSpec", "ENDPOINT_BASE_URL"]
| llama_index/llama-index-integrations/tools/llama-index-tools-azure-translate/llama_index/tools/azure_translate/__init__.py/0 | {
"file_path": "llama_index/llama-index-integrations/tools/llama-index-tools-azure-translate/llama_index/tools/azure_translate/__init__.py",
"repo_id": "llama_index",
"token_count": 72
} | 1,559 |
import dataclasses
import os
from typing import Any, Dict, List, Mapping, Optional, Union, cast
import requests
from langchain_core.callbacks import CallbackManagerForLLMRun
from langchain_core.language_models.llms import LLM
from langchain_core.pydantic_v1 import Extra, root_validator
from langchain_core.utils import... | langchain/libs/community/langchain_community/llms/aviary.py/0 | {
"file_path": "langchain/libs/community/langchain_community/llms/aviary.py",
"repo_id": "langchain",
"token_count": 2500
} | 265 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/kv/etcd/embed_etcd_kv.go/0 | {
"file_path": "milvus/internal/kv/etcd/embed_etcd_kv.go",
"repo_id": "milvus",
"token_count": 6951
} | 1,805 |
import type { AppOptions } from "firebase-admin";
import { getApps, initializeApp } from "firebase-admin/app";
import {
getFirestore,
DocumentData,
Firestore,
DocumentReference,
FieldValue,
} from "firebase-admin/firestore";
import { BaseListChatMessageHistory } from "@langchain/core/chat_history";
import {
... | langchainjs/libs/langchain-community/src/stores/message/firestore.ts/0 | {
"file_path": "langchainjs/libs/langchain-community/src/stores/message/firestore.ts",
"repo_id": "langchainjs",
"token_count": 2460
} | 1,001 |
from langchain_community.utilities.bibtex import BibtexparserWrapper
__all__ = ["BibtexparserWrapper"]
| langchain/libs/langchain/langchain/utilities/bibtex.py/0 | {
"file_path": "langchain/libs/langchain/langchain/utilities/bibtex.py",
"repo_id": "langchain",
"token_count": 33
} | 591 |
# This is the list of HuggingFace Datasets authors for copyright purposes.
#
# This does not necessarily list everyone who has contributed code, since in
# some cases, their employer may be the copyright holder. To see the full list
# of contributors, see the revision history in source control.
Google Inc.
HuggingFac... | datasets/AUTHORS/0 | {
"file_path": "datasets/AUTHORS",
"repo_id": "datasets",
"token_count": 78
} | 118 |
[tool.poetry]
name = "neo4j-advanced-rag"
version = "0.1.0"
description = "Balance precise embeddings and context retention with advanced strategies"
authors = [
"Tomaz Bratanic <tomaz.bratanic@neo4j.com>",
]
readme = "README.md"
[tool.poetry.dependencies]
python = ">=3.8.1,<4.0"
langchain = "^0.1"
tiktoken = "^0.... | langchain/templates/neo4j-advanced-rag/pyproject.toml/0 | {
"file_path": "langchain/templates/neo4j-advanced-rag/pyproject.toml",
"repo_id": "langchain",
"token_count": 322
} | 694 |
from llama_index.core.llama_pack import BaseLlamaPack
from llama_index.packs.nebulagraph_query_engine import NebulaGraphQueryEnginePack
def test_class():
names_of_base_classes = [b.__name__ for b in NebulaGraphQueryEnginePack.__mro__]
assert BaseLlamaPack.__name__ in names_of_base_classes
| llama_index/llama-index-packs/llama-index-packs-nebulagraph-query-engine/tests/test_packs_nebulagraph_query_engine.py/0 | {
"file_path": "llama_index/llama-index-packs/llama-index-packs-nebulagraph-query-engine/tests/test_packs_nebulagraph_query_engine.py",
"repo_id": "llama_index",
"token_count": 102
} | 1,579 |
# See: https://raw.githubusercontent.com/huggingface/tokenizers/main/bindings/python/stub.py
import argparse
import inspect
import os
from typing import Optional
import black
from pathlib import Path
import re
INDENT = " " * 4
GENERATED_COMMENT = "# Generated content DO NOT EDIT\n"
TYPING = """from typing import Any,... | candle/candle-pyo3/stub.py/0 | {
"file_path": "candle/candle-pyo3/stub.py",
"repo_id": "candle",
"token_count": 3852
} | 63 |
[metadata]
license_file = LICENSE
| trl/setup.cfg/0 | {
"file_path": "trl/setup.cfg",
"repo_id": "trl",
"token_count": 11
} | 777 |
import tempfile
from copy import deepcopy
from pathlib import Path
from typing import Any, Callable, Dict, List, Optional, Sequence
from langchain_core.agents import AgentAction, AgentFinish
from langchain_core.callbacks import BaseCallbackHandler
from langchain_core.outputs import Generation, LLMResult
import langch... | langchain/libs/community/langchain_community/callbacks/comet_ml_callback.py/0 | {
"file_path": "langchain/libs/community/langchain_community/callbacks/comet_ml_callback.py",
"repo_id": "langchain",
"token_count": 10833
} | 235 |
from enum import Enum
from typing import Any, Dict, List, Optional
from llama_index.core.base.base_retriever import BaseRetriever
from llama_index.core.indices.query.schema import QueryBundle
from llama_index.core.llama_pack import BaseLlamaPack
from llama_index.core.schema import NodeWithScore, TextNode
class Searc... | llama_index/llama-index-packs/llama-index-packs-agent-search-retriever/llama_index/packs/agent_search_retriever/base.py/0 | {
"file_path": "llama_index/llama-index-packs/llama-index-packs-agent-search-retriever/llama_index/packs/agent_search_retriever/base.py",
"repo_id": "llama_index",
"token_count": 1404
} | 1,638 |
# coding=utf-8
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | transformers/tests/models/bert/test_modeling_bert.py/0 | {
"file_path": "transformers/tests/models/bert/test_modeling_bert.py",
"repo_id": "transformers",
"token_count": 12408
} | 800 |
from __future__ import annotations
from typing import Any, Dict, Iterator, List, Mapping, Optional, cast
from langchain_core.callbacks import CallbackManagerForLLMRun
from langchain_core.language_models.chat_models import BaseChatModel
from langchain_core.messages import (
AIMessage,
AIMessageChunk,
BaseM... | langchain/libs/community/langchain_community/chat_models/volcengine_maas.py/0 | {
"file_path": "langchain/libs/community/langchain_community/chat_models/volcengine_maas.py",
"repo_id": "langchain",
"token_count": 2075
} | 227 |
from __future__ import annotations
import asyncio
import inspect
import warnings
from abc import ABC, abstractmethod
from typing import (
TYPE_CHECKING,
Any,
AsyncIterator,
Dict,
Iterator,
List,
Optional,
Sequence,
cast,
)
from langchain_core._api import deprecated
from langchain_c... | langchain/libs/core/langchain_core/language_models/chat_models.py/0 | {
"file_path": "langchain/libs/core/langchain_core/language_models/chat_models.py",
"repo_id": "langchain",
"token_count": 13973
} | 391 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/mq/msgstream/mqwrapper/rmq/rocksmq_msgstream_test.go/0 | {
"file_path": "milvus/internal/mq/msgstream/mqwrapper/rmq/rocksmq_msgstream_test.go",
"repo_id": "milvus",
"token_count": 8975
} | 1,869 |
"""
This module contains the ConneryAction Tool and ConneryService.
"""
from .service import ConneryService
from .tool import ConneryAction
__all__ = ["ConneryAction", "ConneryService"]
| langchain/libs/community/langchain_community/tools/connery/__init__.py/0 | {
"file_path": "langchain/libs/community/langchain_community/tools/connery/__init__.py",
"repo_id": "langchain",
"token_count": 51
} | 280 |
WEAVIATE_DOCS_INDEX_NAME = "LangChain_Combined_Docs_OpenAI_text_embedding_3_small"
| chat-langchain/constants.py/0 | {
"file_path": "chat-langchain/constants.py",
"repo_id": "chat-langchain",
"token_count": 37
} | 11 |
# candle-mamba-minimal: minimal implementation of Mamba
This is based on [mamba-minimal](https://github.com/johnma2006/mamba-minimal).
Compared to the mamba example, this version can handle training but is much
slower.
## Running the example
```bash
$ cargo run --example mamba-minimal --release -- --prompt "Mamba i... | candle/candle-examples/examples/mamba-minimal/README.md/0 | {
"file_path": "candle/candle-examples/examples/mamba-minimal/README.md",
"repo_id": "candle",
"token_count": 206
} | 38 |
"""Unittests for langchain.agents.chat package."""
from textwrap import dedent
from typing import Any, Tuple
from langchain_core.agents import AgentAction, AgentFinish
from langchain_core.prompts.chat import (
ChatPromptTemplate,
HumanMessagePromptTemplate,
SystemMessagePromptTemplate,
)
from langchain_cor... | langchain/libs/langchain/tests/unit_tests/agents/test_structured_chat.py/0 | {
"file_path": "langchain/libs/langchain/tests/unit_tests/agents/test_structured_chat.py",
"repo_id": "langchain",
"token_count": 3325
} | 648 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/core/src/index/ScalarIndex.cpp/0 | {
"file_path": "milvus/internal/core/src/index/ScalarIndex.cpp",
"repo_id": "milvus",
"token_count": 2485
} | 1,646 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/core/src/index/IndexFactory.h/0 | {
"file_path": "milvus/internal/core/src/index/IndexFactory.h",
"repo_id": "milvus",
"token_count": 1500
} | 1,871 |
from llama_index.llms.monsterapi.base import MonsterLLM
__all__ = ["MonsterLLM"]
| llama_index/llama-index-integrations/llms/llama-index-llms-monsterapi/llama_index/llms/monsterapi/__init__.py/0 | {
"file_path": "llama_index/llama-index-integrations/llms/llama-index-llms-monsterapi/llama_index/llms/monsterapi/__init__.py",
"repo_id": "llama_index",
"token_count": 29
} | 1,233 |
# Deprecated Terms
As LlamaIndex continues to evolve, many class names and APIs have been adjusted, improved, and deprecated.
The following is a list of previously popular terms that have been deprecated, with links to their replacements.
## GPTSimpleVectorIndex
This has been renamed to `VectorStoreIndex`, as well ... | llama_index/docs/changes/deprecated_terms.md/0 | {
"file_path": "llama_index/docs/changes/deprecated_terms.md",
"repo_id": "llama_index",
"token_count": 546
} | 1,165 |
<jupyter_start><jupyter_text>Incoporating semantic similarity in tabular databasesIn this notebook we will cover how to run semantic search over a specific table column within a single SQL query, combining tabular query with RAG. Overall workflow1. Generating embeddings for a specific column2. Storing the embeddings in... | langchain/cookbook/retrieval_in_sql.ipynb/0 | {
"file_path": "langchain/cookbook/retrieval_in_sql.ipynb",
"repo_id": "langchain",
"token_count": 3645
} | 82 |
# coding=utf-8
# Copyright 2024 HuggingFace Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or ag... | diffusers/tests/schedulers/test_schedulers.py/0 | {
"file_path": "diffusers/tests/schedulers/test_schedulers.py",
"repo_id": "diffusers",
"token_count": 17001
} | 283 |
"""Util that calls Lambda."""
import json
from typing import Any, Dict, Optional
from langchain_core.pydantic_v1 import BaseModel, Extra, root_validator
class LambdaWrapper(BaseModel):
"""Wrapper for AWS Lambda SDK.
To use, you should have the ``boto3`` package installed
and a lambda functions built from... | langchain/libs/community/langchain_community/utilities/awslambda.py/0 | {
"file_path": "langchain/libs/community/langchain_community/utilities/awslambda.py",
"repo_id": "langchain",
"token_count": 995
} | 320 |
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | accelerate/examples/by_feature/gradient_accumulation.py/0 | {
"file_path": "accelerate/examples/by_feature/gradient_accumulation.py",
"repo_id": "accelerate",
"token_count": 3352
} | 8 |
<jupyter_start><jupyter_text>LanceDB>[LanceDB](https://lancedb.com/) is an open-source database for vector-search built with persistent storage, which greatly simplifies retrevial, filtering and management of embeddings. Fully open source.This notebook shows how to use functionality related to the `LanceDB` vector data... | langchain/docs/docs/integrations/vectorstores/lancedb.ipynb/0 | {
"file_path": "langchain/docs/docs/integrations/vectorstores/lancedb.ipynb",
"repo_id": "langchain",
"token_count": 773
} | 195 |
import json
import os
from langchain_community.chat_message_histories import CosmosDBChatMessageHistory
from langchain_core.messages import message_to_dict
from langchain.memory import ConversationBufferMemory
# Replace these with your Azure Cosmos DB endpoint and key
endpoint = os.environ.get("COSMOS_DB_ENDPOINT", ... | langchain/libs/langchain/tests/integration_tests/memory/test_cosmos_db.py/0 | {
"file_path": "langchain/libs/langchain/tests/integration_tests/memory/test_cosmos_db.py",
"repo_id": "langchain",
"token_count": 540
} | 582 |
---
sidebar_position: 0
title: Get started
---
import CodeBlock from "@theme/CodeBlock";
import BasicExample from "@examples/guides/expression_language/get_started/basic.ts";
import BasicPromptExample from "@examples/guides/expression_language/get_started/prompt.ts";
import BasicChatModelExample from "@examples/guides... | langchainjs/docs/core_docs/docs/expression_language/get_started.mdx/0 | {
"file_path": "langchainjs/docs/core_docs/docs/expression_language/get_started.mdx",
"repo_id": "langchainjs",
"token_count": 1336
} | 725 |
# docker-compose to make it easier to spin up integration tests.
# Services should use NON standard ports to avoid collision with
version: "3"
name: langchain-tests
services:
redis:
image: redis/redis-stack-server:latest
# We use non standard ports since
# these instances are used for testing
# and ... | langchain/docker/docker-compose.yml/0 | {
"file_path": "langchain/docker/docker-compose.yml",
"repo_id": "langchain",
"token_count": 193
} | 83 |
#!/usr/bin/env bash
python run_asr.py \
--output_dir="./wav2vec2-large-xlsr-53-arabic-speech-corpus" \
--num_train_epochs="50" \
--per_device_train_batch_size="1" \
--per_device_eval_batch_size="1" \
--gradient_accumulation_steps="8" \
--evaluation_strategy="steps" \
--save_steps="500" \
--eval_steps="100" \
--logging_... | transformers/examples/research_projects/wav2vec2/finetune_large_xlsr_53_arabic_speech_corpus.sh/0 | {
"file_path": "transformers/examples/research_projects/wav2vec2/finetune_large_xlsr_53_arabic_speech_corpus.sh",
"repo_id": "transformers",
"token_count": 324
} | 581 |
version: "4"
services:
langchain-playground:
image: langchain/${_LANGSMITH_IMAGE_PREFIX-}langchainplus-playground:${_LANGSMITH_IMAGE_VERSION:-latest}
ports:
- 3001:3001
langchain-frontend:
image: langchain/${_LANGSMITH_IMAGE_PREFIX-}langchainplus-frontend-dynamic:${_LANGSMITH_IMAGE_VERSION:-latest... | langsmith-sdk/js/src/cli/docker-compose.yaml/0 | {
"file_path": "langsmith-sdk/js/src/cli/docker-compose.yaml",
"repo_id": "langsmith-sdk",
"token_count": 1795
} | 1,145 |
# LlamaIndex Readers Integration: Bagel
| llama_index/llama-index-integrations/readers/llama-index-readers-bagel/README.md/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-bagel/README.md",
"repo_id": "llama_index",
"token_count": 10
} | 1,469 |
from langchain_pinecone.vectorstores import Pinecone
__all__ = [
"Pinecone",
]
| langchain/libs/partners/pinecone/langchain_pinecone/__init__.py/0 | {
"file_path": "langchain/libs/partners/pinecone/langchain_pinecone/__init__.py",
"repo_id": "langchain",
"token_count": 31
} | 662 |
---
sidebar_class_name: node-only
---
# Gradient AI
The `GradientEmbeddings` class uses the Gradient AI API to generate embeddings for a given text.
## Setup
You'll need to install the official Gradient Node SDK as a peer dependency:
```bash npm2yarn
npm i @gradientai/nodejs-sdk
```
You will need to set the follo... | langchainjs/docs/core_docs/docs/integrations/text_embedding/gradient_ai.mdx/0 | {
"file_path": "langchainjs/docs/core_docs/docs/integrations/text_embedding/gradient_ai.mdx",
"repo_id": "langchainjs",
"token_count": 307
} | 789 |
"""Test Anthropic Chat API wrapper."""
from typing import List
from unittest.mock import MagicMock
import pytest
from langchain_core.messages import (
AIMessage,
BaseMessage,
HumanMessage,
SystemMessage,
)
from langchain_community.chat_models import BedrockChat
from langchain_community.chat_models.met... | langchain/libs/community/tests/unit_tests/chat_models/test_bedrock.py/0 | {
"file_path": "langchain/libs/community/tests/unit_tests/chat_models/test_bedrock.py",
"repo_id": "langchain",
"token_count": 772
} | 393 |
# coding=utf-8
# Copyright 2023 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | transformers/tests/models/blip/test_modeling_tf_blip.py/0 | {
"file_path": "transformers/tests/models/blip/test_modeling_tf_blip.py",
"repo_id": "transformers",
"token_count": 14940
} | 722 |
# SWSL ResNeXt
A **ResNeXt** repeats a [building block](https://paperswithcode.com/method/resnext-block) that aggregates a set of transformations with the same topology. Compared to a [ResNet](https://paperswithcode.com/method/resnet), it exposes a new dimension, *cardinality* (the size of the set of transformations)... | pytorch-image-models/docs/models/.templates/models/swsl-resnext.md/0 | {
"file_path": "pytorch-image-models/docs/models/.templates/models/swsl-resnext.md",
"repo_id": "pytorch-image-models",
"token_count": 2646
} | 349 |
"""Test token predictor."""
from typing import Any
from unittest.mock import patch
from llama_index.legacy.indices.keyword_table.base import KeywordTableIndex
from llama_index.legacy.indices.list.base import SummaryIndex
from llama_index.legacy.indices.tree.base import TreeIndex
from llama_index.legacy.llms.mock impo... | llama_index/llama-index-legacy/tests/token_predictor/test_base.py/0 | {
"file_path": "llama_index/llama-index-legacy/tests/token_predictor/test_base.py",
"repo_id": "llama_index",
"token_count": 634
} | 1,682 |
"""Test Elasticsearch functionality."""
import pytest
from langchain_community.vectorstores.elasticsearch import (
ApproxRetrievalStrategy,
ElasticsearchStore,
)
from tests.integration_tests.vectorstores.fake_embeddings import FakeEmbeddings
@pytest.mark.requires("elasticsearch")
def test_elasticsearch_hybri... | langchain/libs/community/tests/unit_tests/vectorstores/test_elasticsearch.py/0 | {
"file_path": "langchain/libs/community/tests/unit_tests/vectorstores/test_elasticsearch.py",
"repo_id": "langchain",
"token_count": 383
} | 384 |
use crate::{CpuStorage, Device, Result, Shape, Storage, Tensor};
use k_quants::*;
use std::borrow::Cow;
#[cfg(target_feature = "avx")]
pub mod avx;
mod dummy_metal;
pub mod ggml_file;
pub mod gguf_file;
pub mod k_quants;
#[cfg(feature = "metal")]
pub mod metal;
#[cfg(not(feature = "metal"))]
mod metal {
pub use su... | candle/candle-core/src/quantized/mod.rs/0 | {
"file_path": "candle/candle-core/src/quantized/mod.rs",
"repo_id": "candle",
"token_count": 7659
} | 30 |
import { z } from "zod";
import type { EmbeddingEndpoint, Embedding } from "../embeddingEndpoints";
import { chunk } from "$lib/utils/chunk";
export const embeddingEndpointTeiParametersSchema = z.object({
weight: z.number().int().positive().default(1),
model: z.any(),
type: z.literal("tei"),
url: z.string().url(),... | chat-ui/src/lib/server/embeddingEndpoints/tei/embeddingEndpoints.ts/0 | {
"file_path": "chat-ui/src/lib/server/embeddingEndpoints/tei/embeddingEndpoints.ts",
"repo_id": "chat-ui",
"token_count": 664
} | 90 |
{
"name": "tokenizers-linux-x64-musl",
"version": "0.13.4-rc1",
"os": [
"linux"
],
"cpu": [
"x64"
],
"main": "tokenizers.linux-x64-musl.node",
"files": [
"tokenizers.linux-x64-musl.node"
],
"description": "Tokenizers platform specific bindings",
"keywords": [
"napi-rs",
"NAPI",... | tokenizers/bindings/node/npm/linux-x64-musl/package.json/0 | {
"file_path": "tokenizers/bindings/node/npm/linux-x64-musl/package.json",
"repo_id": "tokenizers",
"token_count": 291
} | 397 |
# ArXiv Search Tool
This tool connects to ArXiv and allows an Agent to search for recent papers and their summaries to retrieve recent information on mathematical and scientific information
## Usage
This tool has a more extensive example usage documented in a Jupyter notebook [here](https://github.com/emptycrown/lla... | llama_index/llama-index-integrations/tools/llama-index-tools-arxiv/README.md/0 | {
"file_path": "llama_index/llama-index-integrations/tools/llama-index-tools-arxiv/README.md",
"repo_id": "llama_index",
"token_count": 291
} | 1,474 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/querycoordv2/task/utils.go/0 | {
"file_path": "milvus/internal/querycoordv2/task/utils.go",
"repo_id": "milvus",
"token_count": 2684
} | 1,866 |
"""Deprecated module for BaseLanguageModel class, kept for backwards compatibility."""
from __future__ import annotations
from langchain_core.language_models import BaseLanguageModel
__all__ = ["BaseLanguageModel"]
| langchain/libs/langchain/langchain/base_language.py/0 | {
"file_path": "langchain/libs/langchain/langchain/base_language.py",
"repo_id": "langchain",
"token_count": 50
} | 450 |
# Voyage Query Engine Pack
Create a query engine using GPT4 and [Voyage AI](https://docs.voyageai.com/embeddings/) Embeddings.
## CLI Usage
You can download llamapacks directly using `llamaindex-cli`, which comes installed with the `llama-index` python package:
```bash
llamaindex-cli download-llamapack VoyageQueryE... | llama_index/llama-index-packs/llama-index-packs-voyage-query-engine/README.md/0 | {
"file_path": "llama_index/llama-index-packs/llama-index-packs-voyage-query-engine/README.md",
"repo_id": "llama_index",
"token_count": 415
} | 1,879 |
package indexparamcheck
import (
"strconv"
"testing"
"github.com/stretchr/testify/assert"
"github.com/milvus-io/milvus/pkg/util/metric"
)
func Test_cagraChecker_CheckTrain(t *testing.T) {
p1 := map[string]string{
DIM: strconv.Itoa(128),
Metric: metric.L2,
}
p2 := map[string]string{
DIM: strconv.I... | milvus/pkg/util/indexparamcheck/cagra_checker_test.go/0 | {
"file_path": "milvus/pkg/util/indexparamcheck/cagra_checker_test.go",
"repo_id": "milvus",
"token_count": 1000
} | 1,838 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.