text stringlengths 3 1.68M | id stringlengths 13 169 | metadata dict | __index_level_0__ int64 0 2.21k |
|---|---|---|---|
# Serper - Google Search API
This page covers how to use the [Serper](https://serper.dev) Google Search API within LangChain. Serper is a low-cost Google Search API that can be used to add answer box, knowledge graph, and organic results data from Google Search.
It is broken into two parts: setup, and then references... | langchain/docs/docs/integrations/providers/google_serper.mdx/0 | {
"file_path": "langchain/docs/docs/integrations/providers/google_serper.mdx",
"repo_id": "langchain",
"token_count": 676
} | 147 |
PD_PREFIX = """
You are working with Xorbits dataframe object in Python.
Before importing Numpy or Pandas in the current script,
remember to import the xorbits version of the library instead.
To import the xorbits version of Numpy, replace the original import statement
`import pandas as pd` with `import xorbits.pandas ... | langchain/libs/experimental/langchain_experimental/agents/agent_toolkits/xorbits/prompt.py/0 | {
"file_path": "langchain/libs/experimental/langchain_experimental/agents/agent_toolkits/xorbits/prompt.py",
"repo_id": "langchain",
"token_count": 317
} | 414 |
// Presents an interface to a storage backend such as s3 or local disk.
// The interface is a simple key-value store, which maps to s3 well.
// For now the interface fetches a file and stores it at a specific
// location on disk. This is not ideal for s3, but it is a start.
// Ideally we would support streaming the fi... | chroma/rust/worker/src/storage/s3.rs/0 | {
"file_path": "chroma/rust/worker/src/storage/s3.rs",
"repo_id": "chroma",
"token_count": 4163
} | 59 |
# rag-astradb
This template will perform RAG using Astra DB (`AstraDB` vector store class)
## Environment Setup
An [Astra DB](https://astra.datastax.com) database is required; free tier is fine.
- You need the database **API endpoint** (such as `https://0123...-us-east1.apps.astra.datastax.com`) ...
- ... and a **... | langchain/templates/rag-astradb/README.md/0 | {
"file_path": "langchain/templates/rag-astradb/README.md",
"repo_id": "langchain",
"token_count": 781
} | 724 |
from copy import deepcopy
from typing import Any, Dict, List, Optional
from langchain_core.agents import AgentAction, AgentFinish
from langchain_core.callbacks import BaseCallbackHandler
from langchain_core.outputs import LLMResult
def import_aim() -> Any:
"""Import the aim python package and raise an error if i... | langchain/libs/community/langchain_community/callbacks/aim_callback.py/0 | {
"file_path": "langchain/libs/community/langchain_community/callbacks/aim_callback.py",
"repo_id": "langchain",
"token_count": 6637
} | 226 |
"""Common classes/functions for tree index operations."""
import asyncio
import logging
from typing import Dict, List, Optional, Sequence, Tuple
from llama_index.legacy.async_utils import run_async_tasks
from llama_index.legacy.callbacks.schema import CBEventType, EventPayload
from llama_index.legacy.data_structs.dat... | llama_index/llama-index-legacy/llama_index/legacy/indices/common_tree/base.py/0 | {
"file_path": "llama_index/llama-index-legacy/llama_index/legacy/indices/common_tree/base.py",
"repo_id": "llama_index",
"token_count": 4328
} | 1,495 |
import { logVersion010MigrationWarning } from "../util/entrypoint_deprecation.js";
/* #__PURE__ */ logVersion010MigrationWarning({
oldEntrypointName: "vectorstores/mongodb_atlas",
});
export * from "@langchain/community/vectorstores/mongodb_atlas";
| langchainjs/langchain/src/vectorstores/mongodb_atlas.ts/0 | {
"file_path": "langchainjs/langchain/src/vectorstores/mongodb_atlas.ts",
"repo_id": "langchainjs",
"token_count": 80
} | 966 |
use std::collections::HashMap;
use std::hash::Hash;
#[derive(Default)]
pub struct TrieBuilder<Label> {
trie: Trie<Label>,
}
impl<Label: Eq + Hash + Copy> TrieBuilder<Label> {
pub fn push(&mut self, element: &[Label]) {
self.trie.push(element);
}
pub fn build(self) -> Trie<Label> {
sel... | tokenizers/tokenizers/src/models/unigram/trie.rs/0 | {
"file_path": "tokenizers/tokenizers/src/models/unigram/trie.rs",
"repo_id": "tokenizers",
"token_count": 944
} | 436 |
#!/usr/bin/env python
# Copyright 2021 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unles... | accelerate/src/accelerate/commands/test.py/0 | {
"file_path": "accelerate/src/accelerate/commands/test.py",
"repo_id": "accelerate",
"token_count": 773
} | 12 |
from llama_index.core.base.embeddings.base import BaseEmbedding
from llama_index.embeddings.nomic import NomicEmbedding
def test_embedding_class():
names_of_base_classes = [b.__name__ for b in NomicEmbedding.__mro__]
assert BaseEmbedding.__name__ in names_of_base_classes
| llama_index/llama-index-integrations/embeddings/llama-index-embeddings-nomic/tests/test_embeddings_nomic.py/0 | {
"file_path": "llama_index/llama-index-integrations/embeddings/llama-index-embeddings-nomic/tests/test_embeddings_nomic.py",
"repo_id": "llama_index",
"token_count": 103
} | 1,200 |
import argparse
import json
import logging
import os
import subprocess
from pathlib import Path
from typing import Dict, List, Literal, Mapping, Optional, Union, cast
from langsmith import env as ls_env
from langsmith import utils as ls_utils
logging.basicConfig(level=logging.INFO, format="%(message)s")
logger = logg... | langsmith-sdk/python/langsmith/cli/main.py/0 | {
"file_path": "langsmith-sdk/python/langsmith/cli/main.py",
"repo_id": "langsmith-sdk",
"token_count": 5243
} | 1,070 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/querynodev2/mock_data.go/0 | {
"file_path": "milvus/internal/querynodev2/mock_data.go",
"repo_id": "milvus",
"token_count": 1865
} | 1,988 |
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by appl... | accelerate/examples/by_feature/cross_validation.py/0 | {
"file_path": "accelerate/examples/by_feature/cross_validation.py",
"repo_id": "accelerate",
"token_count": 4458
} | 5 |
"""Test Google PaLM Chat API wrapper."""
import pytest
from langchain_core.messages import AIMessage, HumanMessage, SystemMessage
from langchain_community.chat_models.google_palm import (
ChatGooglePalm,
ChatGooglePalmError,
_messages_to_prompt_dict,
)
def test_messages_to_prompt_dict_with_valid_message... | langchain/libs/community/tests/unit_tests/chat_models/test_google_palm.py/0 | {
"file_path": "langchain/libs/community/tests/unit_tests/chat_models/test_google_palm.py",
"repo_id": "langchain",
"token_count": 1707
} | 368 |
package predicates
import (
"testing"
"github.com/stretchr/testify/suite"
)
type PredicateSuite struct {
suite.Suite
}
func (s *PredicateSuite) TestValueEqual() {
p := ValueEqual("key", "value")
s.Equal("key", p.Key())
s.Equal("value", p.TargetValue())
s.Equal(PredTargetValue, p.Target())
s.Equal(PredTypeEq... | milvus/internal/kv/predicates/predicate_test.go/0 | {
"file_path": "milvus/internal/kv/predicates/predicate_test.go",
"repo_id": "milvus",
"token_count": 303
} | 1,994 |
"""Chain that calls SerpAPI.
Heavily borrowed from https://github.com/ofirpress/self-ask
"""
import os
import sys
from typing import Any, Dict, Optional, Tuple
import aiohttp
from langchain_core.pydantic_v1 import BaseModel, Extra, Field, root_validator
from langchain_core.utils import get_from_dict_or_env
class Hi... | langchain/libs/community/langchain_community/utilities/serpapi.py/0 | {
"file_path": "langchain/libs/community/langchain_community/utilities/serpapi.py",
"repo_id": "langchain",
"token_count": 4034
} | 304 |
from langchain_community.llms.fake import FakeListLLM, FakeStreamingListLLM
__all__ = ["FakeListLLM", "FakeStreamingListLLM"]
| langchain/libs/langchain/langchain/llms/fake.py/0 | {
"file_path": "langchain/libs/langchain/langchain/llms/fake.py",
"repo_id": "langchain",
"token_count": 43
} | 546 |
from llama_index.tools.metaphor.base import MetaphorToolSpec
__all__ = ["MetaphorToolSpec"]
| llama_index/llama-index-integrations/tools/llama-index-tools-metaphor/llama_index/tools/metaphor/__init__.py/0 | {
"file_path": "llama_index/llama-index-integrations/tools/llama-index-tools-metaphor/llama_index/tools/metaphor/__init__.py",
"repo_id": "llama_index",
"token_count": 33
} | 1,572 |
import { logVersion010MigrationWarning } from "../util/entrypoint_deprecation.js";
/* #__PURE__ */ logVersion010MigrationWarning({
oldEntrypointName: "llms/gradient_ai",
});
export * from "@langchain/community/llms/gradient_ai";
| langchainjs/langchain/src/llms/gradient_ai.ts/0 | {
"file_path": "langchainjs/langchain/src/llms/gradient_ai.ts",
"repo_id": "langchainjs",
"token_count": 74
} | 944 |
#!/usr/bin/env python
"""An example that shows how to create a custom agent executor like Runnable.
At the time of writing, there is a bug in the current AgentExecutor that
prevents it from correctly propagating configuration of the underlying
runnable. While that bug should be fixed, this is an example shows
how to c... | langserve/examples/configurable_agent_executor/server.py/0 | {
"file_path": "langserve/examples/configurable_agent_executor/server.py",
"repo_id": "langserve",
"token_count": 1757
} | 1,059 |
---
sidebar_label: Google AI
---
import CodeBlock from "@theme/CodeBlock";
# Google Generative AI
You can access Google's generative AI embeddings models through
`@langchain/google-genai` integration package.
Get an API key here: https://ai.google.dev/tutorials/setup
You'll need to install the `@langchain/google-g... | langchainjs/docs/core_docs/docs/integrations/text_embedding/google_generativeai.mdx/0 | {
"file_path": "langchainjs/docs/core_docs/docs/integrations/text_embedding/google_generativeai.mdx",
"repo_id": "langchainjs",
"token_count": 220
} | 736 |
# coding=utf-8
# Copyright 2023 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | transformers/tests/models/mixtral/test_modeling_mixtral.py/0 | {
"file_path": "transformers/tests/models/mixtral/test_modeling_mixtral.py",
"repo_id": "transformers",
"token_count": 10571
} | 737 |
"""Test formatting functionality."""
from langchain_core.language_models.base import _get_token_ids_default_method
class TestTokenCountingWithGPT2Tokenizer:
def test_tokenization(self) -> None:
# Check that the tokenization is consistent with the GPT-2 tokenizer
assert _get_token_ids_default_meth... | langchain/libs/langchain/tests/integration_tests/test_schema.py/0 | {
"file_path": "langchain/libs/langchain/tests/integration_tests/test_schema.py",
"repo_id": "langchain",
"token_count": 273
} | 615 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/rootcoord/dml_channels.go/0 | {
"file_path": "milvus/internal/rootcoord/dml_channels.go",
"repo_id": "milvus",
"token_count": 4438
} | 1,851 |
# Consistency Decoder
Consistency decoder can be used to decode the latents from the denoising UNet in the [`StableDiffusionPipeline`]. This decoder was introduced in the [DALL-E 3 technical report](https://openai.com/dall-e-3).
The original codebase can be found at [openai/consistencydecoder](https://github.com/ope... | diffusers/docs/source/en/api/models/consistency_decoder_vae.md/0 | {
"file_path": "diffusers/docs/source/en/api/models/consistency_decoder_vae.md",
"repo_id": "diffusers",
"token_count": 242
} | 178 |
<jupyter_start><jupyter_code>!pip install -q git+https://github.com/huggingface/transformers.git
!pip install -q git+https://github.com/huggingface/peft.git
!pip install -q git+https://github.com/huggingface/accelerate.git@main
!pip install huggingface_hub
!pip install bitsandbytes
!pip install SentencePiece
import os
... | peft/examples/multi_adapter_examples/PEFT_Multi_LoRA_Inference.ipynb/0 | {
"file_path": "peft/examples/multi_adapter_examples/PEFT_Multi_LoRA_Inference.ipynb",
"repo_id": "peft",
"token_count": 1328
} | 319 |
# AWS EC2 Basic Deployment
This is an example deployment to AWS EC2 Compute using [terraform](https://www.terraform.io/).
This deployment will do the following:
- Create a security group with required ports open (22 and 8000)
- Create EC2 instance with Ubuntu 22 and deploy Chroma using docker compose
- Create a data... | chroma/examples/deployments/aws-terraform/README.md/0 | {
"file_path": "chroma/examples/deployments/aws-terraform/README.md",
"repo_id": "chroma",
"token_count": 1515
} | 40 |
"""Data Connectors for LlamaIndex.
This module contains the data connectors for LlamaIndex. Each connector inherits
from a `BaseReader` class, connects to a data source, and loads Document objects
from that data source.
You may also choose to construct Document objects manually, for instance
in our `Insert How-To Gui... | llama_index/llama-index-legacy/llama_index/legacy/readers/__init__.py/0 | {
"file_path": "llama_index/llama-index-legacy/llama_index/legacy/readers/__init__.py",
"repo_id": "llama_index",
"token_count": 1329
} | 1,758 |
# flake8: noqa
PREFIX = """
You are working with a pandas dataframe in Python. The name of the dataframe is `df`.
You should use the tools below to answer the question posed of you:"""
MULTI_DF_PREFIX = """
You are working with {num_dfs} pandas dataframes in Python named df1, df2, etc. You
should use the tools below... | langchain/libs/experimental/langchain_experimental/agents/agent_toolkits/pandas/prompt.py/0 | {
"file_path": "langchain/libs/experimental/langchain_experimental/agents/agent_toolkits/pandas/prompt.py",
"repo_id": "langchain",
"token_count": 409
} | 435 |
from langchain_core.messages import HumanMessageChunk
from langchain_core.outputs import ChatGenerationChunk, GenerationChunk
def test_generation_chunk() -> None:
assert GenerationChunk(text="Hello, ") + GenerationChunk(
text="world!"
) == GenerationChunk(
text="Hello, world!"
), "Generati... | langchain/libs/core/tests/unit_tests/test_outputs.py/0 | {
"file_path": "langchain/libs/core/tests/unit_tests/test_outputs.py",
"repo_id": "langchain",
"token_count": 1005
} | 408 |
[build-system]
build-backend = "poetry.core.masonry.api"
requires = ["poetry-core"]
[tool.codespell]
check-filenames = true
check-hidden = true
skip = "*.csv,*.html,*.json,*.jsonl,*.pdf,*.txt,*.ipynb"
[tool.llamahub]
classes = ["HiveReader"]
contains_example = false
import_path = "llama_index.readers.hive"
[tool.myp... | llama_index/llama-index-integrations/readers/llama-index-readers-hive/pyproject.toml/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-hive/pyproject.toml",
"repo_id": "llama_index",
"token_count": 687
} | 1,457 |
from abc import ABC, abstractmethod
from typing import List, Optional
class Constraint(ABC):
r"""Abstract base class for all constraints that can be applied during generation.
It must define how the constraint can be satisfied.
All classes that inherit Constraint must follow the requirement that
```... | transformers/src/transformers/generation/beam_constraints.py/0 | {
"file_path": "transformers/src/transformers/generation/beam_constraints.py",
"repo_id": "transformers",
"token_count": 8310
} | 595 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/pkg/util/uniquegenerator/unique_int_generator_test.go/0 | {
"file_path": "milvus/pkg/util/uniquegenerator/unique_int_generator_test.go",
"repo_id": "milvus",
"token_count": 405
} | 1,856 |
import { withJsonFormsControlProps } from "@jsonforms/react";
import PlusIcon from "../assets/PlusIcon.svg?react";
import TrashIcon from "../assets/TrashIcon.svg?react";
import {
rankWith,
and,
schemaMatches,
Paths,
isControl,
} from "@jsonforms/core";
import { AutosizeTextarea } from "./AutosizeTextarea";
im... | langserve/langserve/playground/src/components/ChatMessageTuplesControlRenderer.tsx/0 | {
"file_path": "langserve/langserve/playground/src/components/ChatMessageTuplesControlRenderer.tsx",
"repo_id": "langserve",
"token_count": 2541
} | 1,129 |
import logging
import os
from typing import Any, Callable, Optional, Tuple, Union
from llama_index.core.llms.generic_utils import get_from_param_or_env
from tenacity import (
before_sleep_log,
retry,
retry_if_exception_type,
stop_after_attempt,
stop_after_delay,
wait_exponential,
wait_rando... | llama_index/llama-index-integrations/embeddings/llama-index-embeddings-openai/llama_index/embeddings/openai/utils.py/0 | {
"file_path": "llama_index/llama-index-integrations/embeddings/llama-index-embeddings-openai/llama_index/embeddings/openai/utils.py",
"repo_id": "llama_index",
"token_count": 1373
} | 1,284 |
"""Mbox parser.
Contains simple parser for mbox files.
"""
import logging
from pathlib import Path
from typing import Any, Dict, List, Optional
from llama_index.legacy.readers.base import BaseReader
from llama_index.legacy.schema import Document
logger = logging.getLogger(__name__)
class MboxReader(BaseReader):
... | llama_index/llama-index-legacy/llama_index/legacy/readers/file/mbox_reader.py/0 | {
"file_path": "llama_index/llama-index-legacy/llama_index/legacy/readers/file/mbox_reader.py",
"repo_id": "llama_index",
"token_count": 1676
} | 1,606 |
<!--Copyright 2024 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | diffusers/docs/source/ko/using-diffusers/reusing_seeds.md/0 | {
"file_path": "diffusers/docs/source/ko/using-diffusers/reusing_seeds.md",
"repo_id": "diffusers",
"token_count": 1922
} | 200 |
# Copyright 2023-present the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or... | peft/src/peft/tuners/lora/__init__.py/0 | {
"file_path": "peft/src/peft/tuners/lora/__init__.py",
"repo_id": "peft",
"token_count": 413
} | 349 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/core/src/index/VectorMemIndex.cpp/0 | {
"file_path": "milvus/internal/core/src/index/VectorMemIndex.cpp",
"repo_id": "milvus",
"token_count": 15202
} | 1,648 |
from typing import List, Optional
import pytest
from langchain_core.documents import Document
from langchain_community.vectorstores import SQLiteVSS
from tests.integration_tests.vectorstores.fake_embeddings import (
FakeEmbeddings,
fake_texts,
)
def _sqlite_vss_from_texts(
metadatas: Optional[List[dict]... | langchain/libs/community/tests/integration_tests/vectorstores/test_sqlitevss.py/0 | {
"file_path": "langchain/libs/community/tests/integration_tests/vectorstores/test_sqlitevss.py",
"repo_id": "langchain",
"token_count": 768
} | 360 |
# EfficientNet
**EfficientNet** is a convolutional neural network architecture and scaling method that uniformly scales all dimensions of depth/width/resolution using a *compound coefficient*. Unlike conventional practice that arbitrary scales these factors, the EfficientNet scaling method uniformly scales network wi... | pytorch-image-models/docs/models/efficientnet.md/0 | {
"file_path": "pytorch-image-models/docs/models/efficientnet.md",
"repo_id": "pytorch-image-models",
"token_count": 4897
} | 354 |
<jupyter_start><jupyter_text>OpenRouter OpenRouter provides a standardized API to access many LLMs at the best price offered. You can find out more on their [homepage](https://openrouter.ai).If you're opening this Notebook on colab, you will probably need to install LlamaIndex 🦙.<jupyter_code>%pip install llama-index-... | llama_index/docs/examples/llm/openrouter.ipynb/0 | {
"file_path": "llama_index/docs/examples/llm/openrouter.ipynb",
"repo_id": "llama_index",
"token_count": 1075
} | 1,090 |
""" BEiT: BERT Pre-Training of Image Transformers (https://arxiv.org/abs/2106.08254)
Model from official source: https://github.com/microsoft/unilm/tree/master/beit
@inproceedings{beit,
title={{BEiT}: {BERT} Pre-Training of Image Transformers},
author={Hangbo Bao and Li Dong and Songhao Piao and Furu Wei},
booktitle=... | pytorch-image-models/timm/models/beit.py/0 | {
"file_path": "pytorch-image-models/timm/models/beit.py",
"repo_id": "pytorch-image-models",
"token_count": 12467
} | 361 |
"""Dataset generation from documents."""
from __future__ import annotations
import asyncio
import re
from typing import List
from llama_index.legacy import Document, ServiceContext, SummaryIndex
from llama_index.legacy.async_utils import DEFAULT_NUM_WORKERS, run_jobs
from llama_index.legacy.core.response.schema impo... | llama_index/llama-index-legacy/llama_index/legacy/llama_dataset/generator.py/0 | {
"file_path": "llama_index/llama-index-legacy/llama_index/legacy/llama_dataset/generator.py",
"repo_id": "llama_index",
"token_count": 4596
} | 1,502 |
#!/usr/bin/env bash
python run_asr.py \
--output_dir="./wav2vec2-large-lv60-timit-asr" \
--num_train_epochs="30" \
--per_device_train_batch_size="2" \
--per_device_eval_batch_size="2" \
--gradient_accumulation_steps="4" \
--evaluation_strategy="steps" \
--save_steps="500" \
--eval_steps="100" \
--logging_steps="50" \
-... | transformers/examples/research_projects/wav2vec2/finetune_large_lv60_timit_asr.sh/0 | {
"file_path": "transformers/examples/research_projects/wav2vec2/finetune_large_lv60_timit_asr.sh",
"repo_id": "transformers",
"token_count": 276
} | 579 |
python_tests()
| llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-openai/tests/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-openai/tests/BUILD",
"repo_id": "llama_index",
"token_count": 5
} | 1,342 |
{
"compilerOptions": {
"target": "es5",
"lib": ["dom", "dom.iterable", "esnext"],
"allowJs": true,
"skipLibCheck": true,
"strict": true,
"forceConsistentCasingInFileNames": true,
"noEmit": true,
"esModuleInterop": true,
"module": "esnext",
"moduleResolution": "node",
"resol... | langchainjs/environment_tests/test-exports-vercel/tsconfig.json/0 | {
"file_path": "langchainjs/environment_tests/test-exports-vercel/tsconfig.json",
"repo_id": "langchainjs",
"token_count": 290
} | 835 |
# for backwards compatibility
from llama_index.legacy.schema import QueryBundle, QueryType
__all__ = ["QueryBundle", "QueryType"]
| llama_index/llama-index-legacy/llama_index/legacy/indices/query/schema.py/0 | {
"file_path": "llama_index/llama-index-legacy/llama_index/legacy/indices/query/schema.py",
"repo_id": "llama_index",
"token_count": 39
} | 1,604 |
import type { BaseLanguageModelInterface } from "@langchain/core/language_models/base";
import { z } from "zod";
import {
Example,
interpolateFString,
FewShotPromptTemplate,
} from "@langchain/core/prompts";
import { InputValues } from "@langchain/core/utils/types";
import { QueryTransformer, TraverseType } from ... | langchainjs/langchain/src/chains/query_constructor/index.ts/0 | {
"file_path": "langchainjs/langchain/src/chains/query_constructor/index.ts",
"repo_id": "langchainjs",
"token_count": 1759
} | 882 |
/* eslint-disable no-instanceof/no-instanceof */
import {
CacheDelete,
CacheListFetch,
CacheListPushBack,
ICacheClient,
InvalidArgumentError,
CollectionTtl,
} from "@gomomento/sdk-core";
import { BaseListChatMessageHistory } from "@langchain/core/chat_history";
import {
BaseMessage,
StoredMessage,
map... | langchainjs/libs/langchain-community/src/stores/message/momento.ts/0 | {
"file_path": "langchainjs/libs/langchain-community/src/stores/message/momento.ts",
"repo_id": "langchainjs",
"token_count": 2080
} | 1,070 |
# Scale Tests
## Goal
Scale tests are designed to check the scalability of Milvus.
For instance, if the dataNode pod expands from one to two:
- verify the consistency of existing data
- verify that the DDL and DML operation is working
## Prerequisite
- Kubernetes Cluster
- Milvus Operator (refer to [M... | milvus/tests/python_client/scale/README.md/0 | {
"file_path": "milvus/tests/python_client/scale/README.md",
"repo_id": "milvus",
"token_count": 421
} | 1,973 |
import os
import threading
import time
import glob
from chaos import constants
from yaml import full_load
from utils.util_log import test_log as log
from delayed_assert import expect
import pytest
def check_config(chaos_config):
if not chaos_config.get("kind", None):
raise Exception("kind must be specifie... | milvus/tests/python_client/chaos/chaos_commons.py/0 | {
"file_path": "milvus/tests/python_client/chaos/chaos_commons.py",
"repo_id": "milvus",
"token_count": 1836
} | 1,952 |
from typing import Any, List
from langchain_community.document_loaders.unstructured import (
UnstructuredFileLoader,
validate_unstructured_version,
)
class UnstructuredOrgModeLoader(UnstructuredFileLoader):
"""Load `Org-Mode` files using `Unstructured`.
You can run the loader in one of two modes: "s... | langchain/libs/community/langchain_community/document_loaders/org_mode.py/0 | {
"file_path": "langchain/libs/community/langchain_community/document_loaders/org_mode.py",
"repo_id": "langchain",
"token_count": 628
} | 255 |
---
title: chat-ui
emoji: 🔥
colorFrom: purple
colorTo: purple
sdk: docker
pinned: false
license: apache-2.0
base_path: /chat
app_port: 3000
failure_strategy: rollback
---
# Chat UI

A c... | chat-ui/README.md/0 | {
"file_path": "chat-ui/README.md",
"repo_id": "chat-ui",
"token_count": 8361
} | 83 |
python_sources()
| llama_index/llama-index-integrations/readers/llama-index-readers-pandas-ai/llama_index/readers/pandas_ai/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-pandas-ai/llama_index/readers/pandas_ai/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,422 |
python_sources()
| llama_index/llama-index-integrations/readers/llama-index-readers-snscrape-twitter/llama_index/readers/snscrape_twitter/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-snscrape-twitter/llama_index/readers/snscrape_twitter/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,432 |
# HTML to text
>[html2text](https://github.com/Alir3z4/html2text/) is a Python package that converts a page of `HTML` into clean, easy-to-read plain `ASCII text`.
The ASCII also happens to be a valid `Markdown` (a text-to-HTML format).
## Installation and Setup
```bash
pip install html2text
```
## Document Transf... | langchain/docs/docs/integrations/providers/html2text.mdx/0 | {
"file_path": "langchain/docs/docs/integrations/providers/html2text.mdx",
"repo_id": "langchain",
"token_count": 154
} | 135 |
# Testcases All-in-one
# pod kill
# standalone
# 3 pods(standalone-ha-blabla, etcd, minio)
# cluster-1-node
# 11 pods(proxy, rootcoord, querynode, querycoord, datanode, datacoord,
# indexnode, indexcoord, pulsar, etcd, minio)
# cluster-n-nodes
# 11 pods* n: kill one and kil... | milvus/tests/python_client/chaos/chaos_objects/network_partition/testcases.yaml/0 | {
"file_path": "milvus/tests/python_client/chaos/chaos_objects/network_partition/testcases.yaml",
"repo_id": "milvus",
"token_count": 1775
} | 1,990 |
import json
from typing import List
from langchain.tools.base import BaseTool
FINISH_NAME = "finish"
class PromptGenerator:
"""A class for generating custom prompt strings.
Does this based on constraints, commands, resources, and performance evaluations.
"""
def __init__(self) -> None:
"""... | langchain/libs/experimental/langchain_experimental/autonomous_agents/autogpt/prompt_generator.py/0 | {
"file_path": "langchain/libs/experimental/langchain_experimental/autonomous_agents/autogpt/prompt_generator.py",
"repo_id": "langchain",
"token_count": 2650
} | 437 |
// Code generated by mockery v2.32.4. DO NOT EDIT.
package mocks
import (
context "context"
commonpb "github.com/milvus-io/milvus-proto/go-api/v2/commonpb"
datapb "github.com/milvus-io/milvus/internal/proto/datapb"
grpc "google.golang.org/grpc"
internalpb "github.com/milvus-io/milvus/internal/proto/internalp... | milvus/internal/mocks/mock_datanode_client.go/0 | {
"file_path": "milvus/internal/mocks/mock_datanode_client.go",
"repo_id": "milvus",
"token_count": 21060
} | 1,946 |
from llama_index.embeddings.openai.base import (
OpenAIEmbedding,
OpenAIEmbeddingMode,
OpenAIEmbeddingModelType,
OpenAIEmbeddingModeModel,
)
__all__ = [
"OpenAIEmbedding",
"OpenAIEmbeddingMode",
"OpenAIEmbeddingModelType",
"OpenAIEmbeddingModeModel",
]
| llama_index/llama-index-integrations/embeddings/llama-index-embeddings-openai/llama_index/embeddings/openai/__init__.py/0 | {
"file_path": "llama_index/llama-index-integrations/embeddings/llama-index-embeddings-openai/llama_index/embeddings/openai/__init__.py",
"repo_id": "llama_index",
"token_count": 130
} | 1,372 |
from typing import List
import PIL.Image
import PIL.ImageOps
from packaging import version
from PIL import Image
if version.parse(version.parse(PIL.__version__).base_version) >= version.parse("9.1.0"):
PIL_INTERPOLATION = {
"linear": PIL.Image.Resampling.BILINEAR,
"bilinear": PIL.Image.Resampling... | diffusers/src/diffusers/utils/pil_utils.py/0 | {
"file_path": "diffusers/src/diffusers/utils/pil_utils.py",
"repo_id": "diffusers",
"token_count": 849
} | 263 |
import { YandexGPT } from "@langchain/yandex/llms";
const model = new YandexGPT();
const res = await model.invoke(['Translate "I love programming" into French.']);
console.log({ res });
| langchainjs/examples/src/models/llm/yandex.ts/0 | {
"file_path": "langchainjs/examples/src/models/llm/yandex.ts",
"repo_id": "langchainjs",
"token_count": 61
} | 814 |
from langchain_community.callbacks.streamlit.streamlit_callback_handler import (
CHECKMARK_EMOJI,
EXCEPTION_EMOJI,
HISTORY_EMOJI,
THINKING_EMOJI,
LLMThought,
LLMThoughtLabeler,
LLMThoughtState,
StreamlitCallbackHandler,
ToolRecord,
)
__all__ = [
"CHECKMARK_EMOJI",
"THINKING_... | langchain/libs/langchain/langchain/callbacks/streamlit/streamlit_callback_handler.py/0 | {
"file_path": "langchain/libs/langchain/langchain/callbacks/streamlit/streamlit_callback_handler.py",
"repo_id": "langchain",
"token_count": 236
} | 476 |
<jupyter_start><jupyter_text>Google TrendsThis notebook goes over how to use the Google Trends Tool to fetch trends information.First, you need to sign up for an `SerpApi key` key at: https://serpapi.com/users/sign_up.Then you must install `google-search-results` with the command:`pip install google-search-results`Then... | langchain/docs/docs/integrations/tools/google_trends.ipynb/0 | {
"file_path": "langchain/docs/docs/integrations/tools/google_trends.ipynb",
"repo_id": "langchain",
"token_count": 274
} | 184 |
# Copyright 2024 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | diffusers/src/diffusers/models/transformers/transformer_temporal.py/0 | {
"file_path": "diffusers/src/diffusers/models/transformers/transformer_temporal.py",
"repo_id": "diffusers",
"token_count": 7286
} | 230 |
import { GoogleAuth, GoogleAuthOptions } from "google-auth-library";
import { Embeddings, EmbeddingsParams } from "@langchain/core/embeddings";
import { AsyncCallerCallOptions } from "@langchain/core/utils/async_caller";
import {
GoogleVertexAIBaseLLMInput,
GoogleVertexAIBasePrediction,
GoogleVertexAILLMPredictio... | langchainjs/langchain/src/experimental/multimodal_embeddings/googlevertexai.ts/0 | {
"file_path": "langchainjs/langchain/src/experimental/multimodal_embeddings/googlevertexai.ts",
"repo_id": "langchainjs",
"token_count": 2172
} | 939 |
python_sources()
| llama_index/llama-index-legacy/llama_index/legacy/evaluation/retrieval/BUILD/0 | {
"file_path": "llama_index/llama-index-legacy/llama_index/legacy/evaluation/retrieval/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,590 |
# coding=utf-8
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | transformers/tests/models/gpt_neox_japanese/test_tokenization_gpt_neox_japanese.py/0 | {
"file_path": "transformers/tests/models/gpt_neox_japanese/test_tokenization_gpt_neox_japanese.py",
"repo_id": "transformers",
"token_count": 2273
} | 728 |
python_sources()
| llama_index/llama-index-legacy/llama_index/legacy/agent/legacy/BUILD/0 | {
"file_path": "llama_index/llama-index-legacy/llama_index/legacy/agent/legacy/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,547 |
# Meilisearch
> [Meilisearch](https://meilisearch.com) is an open-source, lightning-fast, and hyper
> relevant search engine.
> It comes with great defaults to help developers build snappy search experiences.
>
> You can [self-host Meilisearch](https://www.meilisearch.com/docs/learn/getting_started/installation#loca... | langchain/docs/docs/integrations/providers/meilisearch.mdx/0 | {
"file_path": "langchain/docs/docs/integrations/providers/meilisearch.mdx",
"repo_id": "langchain",
"token_count": 261
} | 152 |
from langchain_community.tools.edenai.audio_text_to_speech import EdenAiTextToSpeechTool
__all__ = ["EdenAiTextToSpeechTool"]
| langchain/libs/langchain/langchain/tools/edenai/audio_text_to_speech.py/0 | {
"file_path": "langchain/libs/langchain/langchain/tools/edenai/audio_text_to_speech.py",
"repo_id": "langchain",
"token_count": 45
} | 548 |
""" HRNet
Copied from https://github.com/HRNet/HRNet-Image-Classification
Original header:
Copyright (c) Microsoft
Licensed under the MIT License.
Written by Bin Xiao (Bin.Xiao@microsoft.com)
Modified by Ke Sun (sunk@mail.ustc.edu.cn)
"""
import logging
from typing import List
import torch
import torch.nn as... | pytorch-image-models/timm/models/hrnet.py/0 | {
"file_path": "pytorch-image-models/timm/models/hrnet.py",
"repo_id": "pytorch-image-models",
"token_count": 17584
} | 364 |
from langchain_core.utils.input import (
get_bolded_text,
get_color_mapping,
get_colored_text,
print_text,
)
__all__ = ["get_color_mapping", "get_colored_text", "get_bolded_text", "print_text"]
| langchain/libs/langchain/langchain/utils/input.py/0 | {
"file_path": "langchain/libs/langchain/langchain/utils/input.py",
"repo_id": "langchain",
"token_count": 88
} | 577 |
# Updating to v0.10.0
With the introduction of LlamaIndex v0.10.0, there were several changes
- integrations have separate `pip installs (See the [full registry](https://pretty-sodium-5e0.notion.site/ce81b247649a44e4b6b35dfb24af28a6?v=53b3c2ced7bb4c9996b81b83c9f01139))
- many imports changed
- the service context was... | llama_index/docs/getting_started/v0_10_0_migration.md/0 | {
"file_path": "llama_index/docs/getting_started/v0_10_0_migration.md",
"repo_id": "llama_index",
"token_count": 716
} | 1,222 |
"""MultiOn Toolkit."""
| langchain/libs/community/langchain_community/agent_toolkits/multion/__init__.py/0 | {
"file_path": "langchain/libs/community/langchain_community/agent_toolkits/multion/__init__.py",
"repo_id": "langchain",
"token_count": 8
} | 223 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/rootcoord/quota_center.go/0 | {
"file_path": "milvus/internal/rootcoord/quota_center.go",
"repo_id": "milvus",
"token_count": 12059
} | 1,864 |
# flake8: noqa
ENDPOINT_DESCRIPTION = "Ask questions about LangChain documentation!"
ENDPOINT_NAME = "ask-langchain"
INPUT_NAME = "query"
OUTPUT_KEY = "result"
NAME_FOR_MODEL = "langchainQABot"
NAME_FOR_HUMAN = "LangChain QA Bot"
DESCRIPTION_FOR_MODEL = "This plugin provides access to a LangChain QA Bot to answer quest... | langchain-aiplugin/retrieval_qa/constants.py/0 | {
"file_path": "langchain-aiplugin/retrieval_qa/constants.py",
"repo_id": "langchain-aiplugin",
"token_count": 138
} | 66 |
# Generated content DO NOT EDIT
class Trainer:
"""
Base class for all trainers
This class is not supposed to be instantiated directly. Instead, any implementation of a
Trainer will return an instance of this class when instantiated.
"""
class BpeTrainer(Trainer):
"""
Trainer capable of tra... | tokenizers/bindings/python/py_src/tokenizers/trainers/__init__.pyi/0 | {
"file_path": "tokenizers/bindings/python/py_src/tokenizers/trainers/__init__.pyi",
"repo_id": "tokenizers",
"token_count": 2178
} | 423 |
# Run Milvus standalone through binary files
To quickly install Milvus standalone without docker or kubernetes, this document provides a tutorial for installing Milvus and dependencies, etcd and MinIO, through binary files.
Before installing, you can refer to [docker-compose.yml](https://github.com/milvus-io/milvus/b... | milvus/deployments/binary/README.md/0 | {
"file_path": "milvus/deployments/binary/README.md",
"repo_id": "milvus",
"token_count": 587
} | 1,636 |
from llama_index.multi_modal_llms.replicate.base import ReplicateMultiModal
__all__ = ["ReplicateMultiModal"]
| llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-replicate/llama_index/multi_modal_llms/replicate/__init__.py/0 | {
"file_path": "llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-replicate/llama_index/multi_modal_llms/replicate/__init__.py",
"repo_id": "llama_index",
"token_count": 38
} | 1,256 |
"""Init file."""
from llama_index.readers.openalex.base import OpenAlexReader
__all__ = ["OpenAlexReader"]
| llama_index/llama-index-integrations/readers/llama-index-readers-openalex/llama_index/readers/openalex/__init__.py/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-openalex/llama_index/readers/openalex/__init__.py",
"repo_id": "llama_index",
"token_count": 36
} | 1,395 |
python_tests()
| llama_index/llama-index-packs/llama-index-packs-neo4j-query-engine/tests/BUILD/0 | {
"file_path": "llama_index/llama-index-packs/llama-index-packs-neo4j-query-engine/tests/BUILD",
"repo_id": "llama_index",
"token_count": 5
} | 1,669 |
import _ from "lodash";
import { LLMOptions, Portkey as _Portkey } from "portkey-ai";
import { CallbackManagerForLLMRun } from "@langchain/core/callbacks/manager";
import { GenerationChunk, LLMResult } from "@langchain/core/outputs";
import { getEnvironmentVariable } from "@langchain/core/utils/env";
import { BaseLLM }... | langchainjs/libs/langchain-community/src/llms/portkey.ts/0 | {
"file_path": "langchainjs/libs/langchain-community/src/llms/portkey.ts",
"repo_id": "langchainjs",
"token_count": 1719
} | 958 |
from typing import TYPE_CHECKING
from ...utils import DIFFUSERS_SLOW_IMPORT, _LazyModule
_import_structure = {"pipeline_ddim": ["DDIMPipeline"]}
if TYPE_CHECKING or DIFFUSERS_SLOW_IMPORT:
from .pipeline_ddim import DDIMPipeline
else:
import sys
sys.modules[__name__] = _LazyModule(
__name__,
... | diffusers/src/diffusers/pipelines/ddim/__init__.py/0 | {
"file_path": "diffusers/src/diffusers/pipelines/ddim/__init__.py",
"repo_id": "diffusers",
"token_count": 180
} | 249 |
from typing import List
from langchain.base_language import BaseLanguageModel
from langchain.tools.base import BaseTool
from langchain_experimental.autonomous_agents.hugginggpt.repsonse_generator import (
load_response_generator,
)
from langchain_experimental.autonomous_agents.hugginggpt.task_executor import (
... | langchain/libs/experimental/langchain_experimental/autonomous_agents/hugginggpt/hugginggpt.py/0 | {
"file_path": "langchain/libs/experimental/langchain_experimental/autonomous_agents/hugginggpt/hugginggpt.py",
"repo_id": "langchain",
"token_count": 428
} | 438 |
<jupyter_start><jupyter_text>Classification de token (TensorFlow) Installez les bibliothèques 🤗 *Datasets*, 🤗 *Transformers* et 🤗 *Accelerate* pour exécuter ce *notebook*.<jupyter_code>!pip install datasets transformers[sentencepiece]
!apt install git-lfs<jupyter_output><empty_output><jupyter_text>Vous aurez besoin... | notebooks/course/fr/chapter7/section2_tf.ipynb/0 | {
"file_path": "notebooks/course/fr/chapter7/section2_tf.ipynb",
"repo_id": "notebooks",
"token_count": 2616
} | 314 |
insert_performance:
collections:
-
milvus:
db_config.primary_path: /test/milvus/db_data_011/cluster/sift_1m_128_l2
cache_config.cpu_cache_capacity: 4GB
engine_config.use_blas_threshold: 1100
engine_config.gpu_search_threshold: 1
gpu_resource_config.enable: true... | milvus/tests/benchmark/milvus_benchmark/suites/2_insert_cluster.yaml/0 | {
"file_path": "milvus/tests/benchmark/milvus_benchmark/suites/2_insert_cluster.yaml",
"repo_id": "milvus",
"token_count": 391
} | 1,939 |
// Code generated by mockery v2.33.3. DO NOT EDIT.
package mocks
import (
dbmodel "github.com/chroma/chroma-coordinator/internal/metastore/db/dbmodel"
mock "github.com/stretchr/testify/mock"
)
// IDatabaseDb is an autogenerated mock type for the IDatabaseDb type
type IDatabaseDb struct {
mock.Mock
}
// DeleteAll... | chroma/go/coordinator/internal/metastore/db/dbmodel/mocks/IDatabaseDb.go/0 | {
"file_path": "chroma/go/coordinator/internal/metastore/db/dbmodel/mocks/IDatabaseDb.go",
"repo_id": "chroma",
"token_count": 1016
} | 52 |
# Copyright 2022 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | transformers/tests/repo_utils/test_check_dummies.py/0 | {
"file_path": "transformers/tests/repo_utils/test_check_dummies.py",
"repo_id": "transformers",
"token_count": 1800
} | 753 |
# MixNet
**MixNet** is a type of convolutional neural network discovered via AutoML that utilises [MixConvs](https://paperswithcode.com/method/mixconv) instead of regular [depthwise convolutions](https://paperswithcode.com/method/depthwise-convolution).
{% include 'code_snippets.md' %}
## How do I train this model?
... | pytorch-image-models/docs/models/.templates/models/mixnet.md/0 | {
"file_path": "pytorch-image-models/docs/models/.templates/models/mixnet.md",
"repo_id": "pytorch-image-models",
"token_count": 1878
} | 345 |
use candle::{Result, Tensor};
pub fn apply_repeat_penalty(logits: &Tensor, penalty: f32, context: &[u32]) -> Result<Tensor> {
let device = logits.device();
let mut logits = logits.to_vec1::<f32>()?;
let context: std::collections::HashSet<_> = context.iter().collect();
for (token_id, logit) in logits.it... | candle/candle-transformers/src/utils.rs/0 | {
"file_path": "candle/candle-transformers/src/utils.rs",
"repo_id": "candle",
"token_count": 299
} | 75 |
import { load } from "@tensorflow-models/universal-sentence-encoder";
import * as tf from "@tensorflow/tfjs-core";
import { Embeddings, type EmbeddingsParams } from "@langchain/core/embeddings";
/**
* Interface that extends EmbeddingsParams and defines additional
* parameters specific to the TensorFlowEmbeddings cl... | langchainjs/libs/langchain-community/src/embeddings/tensorflow.ts/0 | {
"file_path": "langchainjs/libs/langchain-community/src/embeddings/tensorflow.ts",
"repo_id": "langchainjs",
"token_count": 909
} | 967 |
#!/usr/bin/env python
# coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LI... | transformers/examples/pytorch/token-classification/run_ner_no_trainer.py/0 | {
"file_path": "transformers/examples/pytorch/token-classification/run_ner_no_trainer.py",
"repo_id": "transformers",
"token_count": 14733
} | 592 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/rootcoord/mock_test.go/0 | {
"file_path": "milvus/internal/rootcoord/mock_test.go",
"repo_id": "milvus",
"token_count": 15420
} | 2,058 |
#!/usr/bin/env python
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# U... | accelerate/examples/by_feature/deepspeed_with_config_support.py/0 | {
"file_path": "accelerate/examples/by_feature/deepspeed_with_config_support.py",
"repo_id": "accelerate",
"token_count": 12377
} | 5 |
export { SerpAPI, type SerpAPIParameters } from "./serpapi.js";
export { DadJokeAPI } from "./dadjokeapi.js";
export { BingSerpAPI } from "./bingserpapi.js";
export { Tool, type ToolParams, StructuredTool } from "./base.js";
export {
DynamicTool,
type DynamicToolInput,
DynamicStructuredTool,
type DynamicStructu... | langchainjs/langchain/src/tools/index.ts/0 | {
"file_path": "langchainjs/langchain/src/tools/index.ts",
"repo_id": "langchainjs",
"token_count": 560
} | 957 |
poetry_requirements(
name="poetry",
)
| llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-jaguar/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-jaguar/BUILD",
"repo_id": "llama_index",
"token_count": 18
} | 1,657 |
# coding=utf-8
# Copyright 2022 The HuggingFace Team Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a clone of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | transformers/tests/quantization/bnb/test_mixed_int8.py/0 | {
"file_path": "transformers/tests/quantization/bnb/test_mixed_int8.py",
"repo_id": "transformers",
"token_count": 15531
} | 833 |
import { DataSource } from "typeorm";
import { SqlDatabase } from "langchain/sql_db";
import { ChatOpenAI } from "@langchain/openai";
import { PromptTemplate } from "@langchain/core/prompts";
import { RunnableSequence } from "@langchain/core/runnables";
import { StringOutputParser } from "@langchain/core/output_parsers... | langchainjs/examples/src/chains/sql_db.ts/0 | {
"file_path": "langchainjs/examples/src/chains/sql_db.ts",
"repo_id": "langchainjs",
"token_count": 1040
} | 794 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.