text
stringlengths
3
1.68M
id
stringlengths
13
169
metadata
dict
__index_level_0__
int64
0
2.21k
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/internal/datanode/l0_compactor_test.go/0
{ "file_path": "milvus/internal/datanode/l0_compactor_test.go", "repo_id": "milvus", "token_count": 8944 }
1,704
import * as uuid from "uuid"; import { AsyncCaller, AsyncCallerParams } from "./utils/async_caller.js"; import { DataType, Dataset, DatasetShareSchema, Example, ExampleCreate, ExampleUpdate, Feedback, KVMap, LangChainBaseMessage, Run, RunCreate, RunUpdate, ScoreType, TracerSession, Tracer...
langsmith-sdk/js/src/client.ts/0
{ "file_path": "langsmith-sdk/js/src/client.ts", "repo_id": "langsmith-sdk", "token_count": 22133 }
1,066
<!--Copyright 2024 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
diffusers/docs/source/en/training/ddpo.md/0
{ "file_path": "diffusers/docs/source/en/training/ddpo.md", "repo_id": "diffusers", "token_count": 321 }
169
import { Cohere } from "@langchain/cohere"; const model = new Cohere({ maxTokens: 20, apiKey: "YOUR-API-KEY", // In Node.js defaults to process.env.COHERE_API_KEY }); const res = await model.invoke( "What would be a good company name a company that makes colorful socks?" ); console.log({ res });
langchainjs/examples/src/models/llm/cohere.ts/0
{ "file_path": "langchainjs/examples/src/models/llm/cohere.ts", "repo_id": "langchainjs", "token_count": 98 }
837
# coding=utf-8 # Copyright 2021 T5 Authors and HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by...
transformers/src/transformers/models/t5/modeling_flax_t5.py/0
{ "file_path": "transformers/src/transformers/models/t5/modeling_flax_t5.py", "repo_id": "transformers", "token_count": 32653 }
675
<jupyter_start><jupyter_text>IntroductionThis notebook is designed to run inference on the [Diffuser](https://arxiv.org/abs/2205.09991) planning model for model-based RL. The notebook is modified from the authors' [original](https://colab.research.google.com/drive/1YajKhu-CUIGBJeQPehjVPJcK_b38a8Nc?usp=sharingscrollTo=5...
notebooks/diffusers/reinforcement_learning_with_diffusers.ipynb/0
{ "file_path": "notebooks/diffusers/reinforcement_learning_with_diffusers.ipynb", "repo_id": "notebooks", "token_count": 8060 }
306
from llama_index.tools.waii.base import WaiiToolSpec __all__ = ["WaiiToolSpec"]
llama_index/llama-index-integrations/tools/llama-index-tools-waii/llama_index/tools/waii/__init__.py/0
{ "file_path": "llama_index/llama-index-integrations/tools/llama-index-tools-waii/llama_index/tools/waii/__init__.py", "repo_id": "llama_index", "token_count": 30 }
1,506
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a cop...
transformers/src/transformers/data/processors/utils.py/0
{ "file_path": "transformers/src/transformers/data/processors/utils.py", "repo_id": "transformers", "token_count": 5994 }
632
from langchain_community.vectorstores.tiledb import ( TileDB, ) __all__ = [ "TileDB", ]
langchain/libs/langchain/langchain/vectorstores/tiledb.py/0
{ "file_path": "langchain/libs/langchain/langchain/vectorstores/tiledb.py", "repo_id": "langchain", "token_count": 41 }
588
cluster: enabled: true image: all: repository: milvusdb/milvus tag: master-latest pullPolicy: IfNotPresent rootCoordinator: replicas: 2 activeStandby: enabled: true # Enable active-standby when you set multiple replicas for root coordinator queryCoordinator: replicas: 2 activeStandby: ...
milvus/tests/python_client/standby/cluster-values.yaml/0
{ "file_path": "milvus/tests/python_client/standby/cluster-values.yaml", "repo_id": "milvus", "token_count": 1664 }
1,980
package memberlist_manager import ( "context" "errors" "github.com/chroma/chroma-coordinator/internal/common" "github.com/pingcap/log" "go.uber.org/zap" "k8s.io/client-go/util/workqueue" ) // A memberlist manager is responsible for managing the memberlist for a // coordinator. A memberlist consists of a store ...
chroma/go/coordinator/internal/memberlist_manager/memberlist_manager.go/0
{ "file_path": "chroma/go/coordinator/internal/memberlist_manager/memberlist_manager.go", "repo_id": "chroma", "token_count": 1209 }
43
import pytest from llama_index.legacy.core.llms.types import ChatMessage from llama_index.legacy.llms.anthropic import Anthropic try: import anthropic except ImportError: anthropic = None # type: ignore @pytest.mark.skipif(anthropic is None, reason="anthropic not installed") def test_basic() -> None: ll...
llama_index/llama-index-legacy/tests/llms/test_anthropic.py/0
{ "file_path": "llama_index/llama-index-legacy/tests/llms/test_anthropic.py", "repo_id": "llama_index", "token_count": 858 }
1,625
from llama_index.llms.mistralai.base import MistralAI __all__ = ["MistralAI"]
llama_index/llama-index-integrations/llms/llama-index-llms-mistralai/llama_index/llms/mistralai/__init__.py/0
{ "file_path": "llama_index/llama-index-integrations/llms/llama-index-llms-mistralai/llama_index/llms/mistralai/__init__.py", "repo_id": "llama_index", "token_count": 30 }
1,361
[tool.poetry] name = "solo-performance-prompting-agent" version = "0.0.1" description = "Agent that transforms a single LLM into a cognitive synergist that has multiple points of view" authors = [] readme = "README.md" [tool.poetry.dependencies] python = ">=3.8.1,<4.0" langchain = "^0.1" openai = "<2" duckduckgo-searc...
langchain/templates/solo-performance-prompting-agent/pyproject.toml/0
{ "file_path": "langchain/templates/solo-performance-prompting-agent/pyproject.toml", "repo_id": "langchain", "token_count": 304 }
706
#!/usr/bin/env bash export PYTHONPATH="../":"${PYTHONPATH}" export WANDB_PROJECT=dmar export MAX_LEN=128 export m=sshleifer/student_marian_en_ro_6_1 python finetune.py \ --learning_rate=3e-4 \ --do_train \ --fp16 \ --data_dir wmt_en_ro \ --max_source_length $MAX_LEN --max_target_length $MAX_LEN --val_max_targ...
transformers/examples/research_projects/seq2seq-distillation/dynamic_bs_example.sh/0
{ "file_path": "transformers/examples/research_projects/seq2seq-distillation/dynamic_bs_example.sh", "repo_id": "transformers", "token_count": 267 }
544
<jupyter_start><jupyter_text>Handle parsing errorsOccasionally the LLM cannot determine what step to take because its outputs are not correctly formatted to be handled by the output parser. In this case, by default the agent errors. But you can easily control this functionality with `handle_parsing_errors`! Let's explo...
langchain/docs/docs/modules/agents/how_to/handle_parsing_errors.ipynb/0
{ "file_path": "langchain/docs/docs/modules/agents/how_to/handle_parsing_errors.ipynb", "repo_id": "langchain", "token_count": 1608 }
202
from llama_index.readers.rayyan.base import RayyanReader __all__ = ["RayyanReader"]
llama_index/llama-index-integrations/readers/llama-index-readers-rayyan/llama_index/readers/rayyan/__init__.py/0
{ "file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-rayyan/llama_index/readers/rayyan/__init__.py", "repo_id": "llama_index", "token_count": 29 }
1,432
<jupyter_start><jupyter_text>Astra DBDataStax [Astra DB](https://docs.datastax.com/en/astra/home/astra.html) is a serverless vector-capable database built on Cassandra and made conveniently available through an easy-to-use JSON API.In the walkthrough, we'll demo the `SelfQueryRetriever` with an `Astra DB` vector store....
langchain/docs/docs/integrations/retrievers/self_query/astradb.ipynb/0
{ "file_path": "langchain/docs/docs/integrations/retrievers/self_query/astradb.ipynb", "repo_id": "langchain", "token_count": 1941 }
158
# Structured chat :::info If you are using a functions-capable model like ChatOpenAI, we currently recommend that you use the [OpenAI Functions agent](/docs/modules/agents/agent_types/openai_functions_agent) for more complex tool calling. ::: The structured chat agent is capable of using multi-input tools. Older age...
langchainjs/docs/core_docs/docs/modules/agents/agent_types/structured_chat.mdx/0
{ "file_path": "langchainjs/docs/core_docs/docs/modules/agents/agent_types/structured_chat.mdx", "repo_id": "langchainjs", "token_count": 1110 }
733
# Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
transformers/src/transformers/models/bert/__init__.py/0
{ "file_path": "transformers/src/transformers/models/bert/__init__.py", "repo_id": "transformers", "token_count": 2639 }
647
from typing import List import datasets from datasets.tasks import AudioClassification from ..folder_based_builder import folder_based_builder logger = datasets.utils.logging.get_logger(__name__) class AudioFolderConfig(folder_based_builder.FolderBasedBuilderConfig): """Builder Config for AudioFolder.""" ...
datasets/src/datasets/packaged_modules/audiofolder/audiofolder.py/0
{ "file_path": "datasets/src/datasets/packaged_modules/audiofolder/audiofolder.py", "repo_id": "datasets", "token_count": 618 }
135
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/internal/datanode/metacache/meta_cache_test.go/0
{ "file_path": "milvus/internal/datanode/metacache/meta_cache_test.go", "repo_id": "milvus", "token_count": 2666 }
1,699
from langchain_community.tools.playwright.base import ( BaseBrowserTool, ) __all__ = ["BaseBrowserTool"]
langchain/libs/langchain/langchain/tools/playwright/base.py/0
{ "file_path": "langchain/libs/langchain/langchain/tools/playwright/base.py", "repo_id": "langchain", "token_count": 36 }
557
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
transformers/src/transformers/models/bigbird_pegasus/convert_bigbird_pegasus_tf_to_pytorch.py/0
{ "file_path": "transformers/src/transformers/models/bigbird_pegasus/convert_bigbird_pegasus_tf_to_pytorch.py", "repo_id": "transformers", "token_count": 2618 }
586
# Detoxifying a Language Model using PPO Language models (LMs) are known to sometimes generate toxic outputs. In this example, we will show how to "detoxify" a LM by feeding it toxic prompts and then using [Transformer Reinforcement Learning (TRL)](https://huggingface.co/docs/trl/index) and Proximal Policy Optimizatio...
trl/docs/source/detoxifying_a_lm.mdx/0
{ "file_path": "trl/docs/source/detoxifying_a_lm.mdx", "repo_id": "trl", "token_count": 3783 }
806
/* eslint-disable no-process-env */ /* eslint-disable @typescript-eslint/no-non-null-assertion */ import { test } from "@jest/globals"; import { OpenAI } from "@langchain/openai"; import { StartExecutionAWSSfnTool } from "@langchain/community/tools/aws_sfn"; import { AWSSfnToolkit } from "@langchain/community/agents/t...
langchainjs/langchain/src/agents/tests/aws_sfn.int.test.ts/0
{ "file_path": "langchainjs/langchain/src/agents/tests/aws_sfn.int.test.ts", "repo_id": "langchainjs", "token_count": 751 }
895
# Copyright 2023 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
trl/tests/test_dpo_trainer.py/0
{ "file_path": "trl/tests/test_dpo_trainer.py", "repo_id": "trl", "token_count": 11138 }
824
poetry_requirements( name="poetry", )
llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-pinecone/BUILD/0
{ "file_path": "llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-pinecone/BUILD", "repo_id": "llama_index", "token_count": 18 }
1,615
from llama_index.core.multi_modal_llms.base import MultiModalLLM from llama_index.multi_modal_llms.replicate_multi_modal import ReplicateMultiModal def test_embedding_class(): names_of_base_classes = [b.__name__ for b in ReplicateMultiModal.__mro__] assert MultiModalLLM.__name__ in names_of_base_classes
llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-replicate-multi-modal/tests/test_multi-modal-llms_replicate_multi_modal.py/0
{ "file_path": "llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-replicate-multi-modal/tests/test_multi-modal-llms_replicate_multi_modal.py", "repo_id": "llama_index", "token_count": 116 }
1,431
# MIT License # Copyright (c) 2019 Yang Liu and the HuggingFace team # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, c...
transformers/examples/research_projects/bertabs/modeling_bertabs.py/0
{ "file_path": "transformers/examples/research_projects/bertabs/modeling_bertabs.py", "repo_id": "transformers", "token_count": 17886 }
531
"""Base interface for chains combining documents.""" from abc import ABC, abstractmethod from typing import Any, Dict, List, Optional, Tuple, Type from langchain_core.callbacks import ( AsyncCallbackManagerForChainRun, CallbackManagerForChainRun, ) from langchain_core.documents import Document from langchain_...
langchain/libs/langchain/langchain/chains/combine_documents/base.py/0
{ "file_path": "langchain/libs/langchain/langchain/chains/combine_documents/base.py", "repo_id": "langchain", "token_count": 3024 }
463
import CodeBlock from "@theme/CodeBlock"; import CreateClientExample from "@examples/indexes/vector_stores/upstash/create_client.ts"; import IndexQueryExample from "@examples/indexes/vector_stores/upstash/index_and_query_docs.ts"; import DeleteExample from "@examples/indexes/vector_stores/upstash/delete_docs.ts"; impor...
langchainjs/docs/core_docs/docs/integrations/vectorstores/upstash.mdx/0
{ "file_path": "langchainjs/docs/core_docs/docs/integrations/vectorstores/upstash.mdx", "repo_id": "langchainjs", "token_count": 497 }
746
--- sidebar_position: 5 --- # Contribute Integrations To begin, make sure you have all the dependencies outlined in guide on [Contributing Code](./code). There are a few different places you can contribute integrations for LangChain: - **Community**: For lighter-weight integrations that are primarily maintained by L...
langchain/docs/docs/contributing/integrations.mdx/0
{ "file_path": "langchain/docs/docs/contributing/integrations.mdx", "repo_id": "langchain", "token_count": 1416 }
86
from langchain.memory import ConversationBufferMemory from langchain.output_parsers.list import CommaSeparatedListOutputParser from langchain.prompts import PromptTemplate from langchain.sql_database import SQLDatabase from langchain_experimental.sql.base import SQLDatabaseChain, SQLDatabaseSequentialChain from tests....
langchain/libs/experimental/tests/unit_tests/test_sql.py/0
{ "file_path": "langchain/libs/experimental/tests/unit_tests/test_sql.py", "repo_id": "langchain", "token_count": 1515 }
475
version: '3.5' services: etcd: image: quay.io/coreos/etcd:v3.5.5 environment: - ETCD_AUTO_COMPACTION_MODE=revision - ETCD_AUTO_COMPACTION_RETENTION=1000 - ETCD_QUOTA_BACKEND_BYTES=4294967296 - ETCD_SNAPSHOT_COUNT=50000 volumes: - ${DOCKER_VOLUME_DIRECTORY:-.}/volumes/etcd:/e...
milvus/deployments/docker/dev/docker-compose-apple-silicon.yml/0
{ "file_path": "milvus/deployments/docker/dev/docker-compose-apple-silicon.yml", "repo_id": "milvus", "token_count": 1140 }
1,903
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
transformers/src/transformers/models/deit/convert_deit_timm_to_pytorch.py/0
{ "file_path": "transformers/src/transformers/models/deit/convert_deit_timm_to_pytorch.py", "repo_id": "transformers", "token_count": 3875 }
589
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/internal/querynodev2/pipeline/pipeline.go/0
{ "file_path": "milvus/internal/querynodev2/pipeline/pipeline.go", "repo_id": "milvus", "token_count": 941 }
1,990
# Microsoft All functionality related to `Microsoft Azure` and other `Microsoft` products. ## Chat Models ### Azure OpenAI >[Microsoft Azure](https://en.wikipedia.org/wiki/Microsoft_Azure), often referred to as `Azure` is a cloud computing platform run by `Microsoft`, which offers access, management, and development...
langchain/docs/docs/integrations/platforms/microsoft.mdx/0
{ "file_path": "langchain/docs/docs/integrations/platforms/microsoft.mdx", "repo_id": "langchain", "token_count": 3312 }
124
"""Test TritonTensorRT Chat API wrapper.""" from langchain_nvidia_trt import TritonTensorRTLLM def test_initialization() -> None: """Test integration initialization.""" TritonTensorRTLLM(model_name="ensemble", server_url="http://localhost:8001")
langchain/libs/partners/nvidia-trt/tests/unit_tests/test_llms.py/0
{ "file_path": "langchain/libs/partners/nvidia-trt/tests/unit_tests/test_llms.py", "repo_id": "langchain", "token_count": 85 }
636
import type { Conversation } from "$lib/types/Conversation"; import type { Message } from "$lib/types/Message"; export function addChildren( conv: Pick<Conversation, "messages" | "rootMessageId">, message: Omit<Message, "id">, parentId?: Message["id"] ): Message["id"] { // if this is the first message we just push...
chat-ui/src/lib/utils/tree/addChildren.ts/0
{ "file_path": "chat-ui/src/lib/utils/tree/addChildren.ts", "repo_id": "chat-ui", "token_count": 496 }
97
python_sources()
llama_index/experimental/splitter_playground/BUILD/0
{ "file_path": "llama_index/experimental/splitter_playground/BUILD", "repo_id": "llama_index", "token_count": 6 }
1,156
import { Source } from "./SourceBubble"; export function InlineCitation(props: { source: Source; sourceNumber: number; highlighted: boolean; onMouseEnter: () => any; onMouseLeave: () => any; }) { const { source, sourceNumber, highlighted, onMouseEnter, onMouseLeave } = props; return ( <a hr...
chat-langchain/chat-langchain/app/components/InlineCitation.tsx/0
{ "file_path": "chat-langchain/chat-langchain/app/components/InlineCitation.tsx", "repo_id": "chat-langchain", "token_count": 254 }
9
// Adapted from turboderp exllama: https://github.com/turboderp/exllama #ifndef _matrix_cuh #define _matrix_cuh #include <cuda_runtime.h> #include <cuda_fp16.h> class MatrixView_half { public: const half* data; const int height; const int width; __device__ __forceinline__ MatrixView_half(const half*...
text-generation-inference/server/exllama_kernels/exllama_kernels/matrix.cuh/0
{ "file_path": "text-generation-inference/server/exllama_kernels/exllama_kernels/matrix.cuh", "repo_id": "text-generation-inference", "token_count": 5380 }
397
python_sources()
llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-singlestoredb/llama_index/vector_stores/singlestoredb/BUILD/0
{ "file_path": "llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-singlestoredb/llama_index/vector_stores/singlestoredb/BUILD", "repo_id": "llama_index", "token_count": 6 }
1,536
"""Code to support various indexing workflows. Provides code to: * Create knowledge graphs from data. * Support indexing workflows from LangChain data loaders to vectorstores. For indexing workflows, this code is used to avoid writing duplicated content into the vectostore and to avoid over-writing content if it's ...
langchain/libs/langchain/langchain/indexes/__init__.py/0
{ "file_path": "langchain/libs/langchain/langchain/indexes/__init__.py", "repo_id": "langchain", "token_count": 268 }
532
package indexparamcheck import ( "strconv" "testing" "github.com/stretchr/testify/assert" "github.com/milvus-io/milvus-proto/go-api/v2/schemapb" "github.com/milvus-io/milvus/pkg/util/metric" ) func Test_binIVFFlatChecker_CheckTrain(t *testing.T) { validParams := map[string]string{ DIM: strconv.Itoa(128),...
milvus/pkg/util/indexparamcheck/bin_ivf_flat_checker_test.go/0
{ "file_path": "milvus/pkg/util/indexparamcheck/bin_ivf_flat_checker_test.go", "repo_id": "milvus", "token_count": 2025 }
1,969
# Attribution-NonCommercial 4.0 International (CC BY-NC 4.0) # William Peebles and Saining Xie # # Copyright (c) 2021 OpenAI # MIT License # # Copyright 2024 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance wi...
diffusers/src/diffusers/pipelines/dit/pipeline_dit.py/0
{ "file_path": "diffusers/src/diffusers/pipelines/dit/pipeline_dit.py", "repo_id": "diffusers", "token_count": 4179 }
256
python_tests()
llama_index/llama-index-integrations/readers/llama-index-readers-jaguar/tests/BUILD/0
{ "file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-jaguar/tests/BUILD", "repo_id": "llama_index", "token_count": 5 }
1,393
"""Test LLM program.""" import json from unittest.mock import MagicMock from llama_index.core.base.llms.types import ( ChatMessage, ChatResponse, CompletionResponse, LLMMetadata, MessageRole, ) from llama_index.core.bridge.pydantic import BaseModel from llama_index.core.output_parsers.pydantic imp...
llama_index/llama-index-core/tests/program/test_llm_program.py/0
{ "file_path": "llama_index/llama-index-core/tests/program/test_llm_program.py", "repo_id": "llama_index", "token_count": 1103 }
1,228
<!--- Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or ...
transformers/docs/source/zh/installation.md/0
{ "file_path": "transformers/docs/source/zh/installation.md", "repo_id": "transformers", "token_count": 4826 }
514
[build-system] build-backend = "poetry.core.masonry.api" requires = ["poetry-core"] [tool.codespell] check-filenames = true check-hidden = true skip = "*.csv,*.html,*.json,*.jsonl,*.pdf,*.txt,*.ipynb" [tool.llamahub] classes = ["BedrockEmbedding", "Models"] contains_example = false import_path = "llama_index.embeddin...
llama_index/llama-index-integrations/embeddings/llama-index-embeddings-bedrock/pyproject.toml/0
{ "file_path": "llama_index/llama-index-integrations/embeddings/llama-index-embeddings-bedrock/pyproject.toml", "repo_id": "llama_index", "token_count": 654 }
1,221
#![allow(unused)] /// A fast implementation of mamba for inference only. /// This is based on: https://github.com/LaurentMazare/mamba.rs use crate::models::with_tracing::{linear, linear_no_bias, Linear}; use candle::{DType, Device, IndexOp, Module, Result, Tensor, D}; use candle_nn::{RmsNorm, VarBuilder}; const D_CONV...
candle/candle-transformers/src/models/mamba.rs/0
{ "file_path": "candle/candle-transformers/src/models/mamba.rs", "repo_id": "candle", "token_count": 3784 }
72
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/internal/proxy/accesslog/minio_handler_test.go/0
{ "file_path": "milvus/internal/proxy/accesslog/minio_handler_test.go", "repo_id": "milvus", "token_count": 1759 }
1,734
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/pkg/config/event_dispatcher.go/0
{ "file_path": "milvus/pkg/config/event_dispatcher.go", "repo_id": "milvus", "token_count": 991 }
1,881
const load = require("langchain/load"); const load_serializable = require("langchain/load/serializable"); const agents = require("langchain/agents"); const agents_toolkits = require("langchain/agents/toolkits"); const agents_toolkits_connery = require("langchain/agents/toolkits/connery"); const agents_format_scratchpad...
langchainjs/environment_tests/test-exports-cjs/src/entrypoints.js/0
{ "file_path": "langchainjs/environment_tests/test-exports-cjs/src/entrypoints.js", "repo_id": "langchainjs", "token_count": 2439 }
831
#!/bin/bash # author emaill will be jenkins's email, when the pr branch is not fast forward compared with master branch # Exit immediately for non zero status set -e function get_author_email(){ email=$(git --no-pager show -s --format=\'%ae\' HEAD ) # Get last commit author when Jenkins submit a merge commit ...
milvus/tests/scripts/get_author_email.sh/0
{ "file_path": "milvus/tests/scripts/get_author_email.sh", "repo_id": "milvus", "token_count": 174 }
1,908
import { expect, test } from "@jest/globals"; import { DataSourceOptions } from "typeorm"; import { OpenAIEmbeddings } from "@langchain/openai"; import { TypeORMVectorStore } from "../typeorm.js"; test.skip("Test embeddings creation", async () => { const args = { postgresConnectionOptions: { type: "postgre...
langchainjs/libs/langchain-community/src/vectorstores/tests/typeorm.int.test.ts/0
{ "file_path": "langchainjs/libs/langchain-community/src/vectorstores/tests/typeorm.int.test.ts", "repo_id": "langchainjs", "token_count": 459 }
1,030
# coding=utf-8 # Copyright 2022 Microsoft Research and The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless...
transformers/src/transformers/models/layoutlmv3/modeling_tf_layoutlmv3.py/0
{ "file_path": "transformers/src/transformers/models/layoutlmv3/modeling_tf_layoutlmv3.py", "repo_id": "transformers", "token_count": 34245 }
620
import os from pymilvus import connections, Index, MilvusException from utils.util_log import test_log as log from base.collection_wrapper import ApiCollectionWrapper from common import common_func as cf from common import common_type as ct def e2e_milvus(host, c_name): """ e2e milvus """ log.debug(f'pid: {...
milvus/tests/python_client/scale/scale_common.py/0
{ "file_path": "milvus/tests/python_client/scale/scale_common.py", "repo_id": "milvus", "token_count": 825 }
1,902
# Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
transformers/tests/models/wav2vec2_with_lm/test_processor_wav2vec2_with_lm.py/0
{ "file_path": "transformers/tests/models/wav2vec2_with_lm/test_processor_wav2vec2_with_lm.py", "repo_id": "transformers", "token_count": 8780 }
757
import html import inspect import re import urllib.parse as ul from typing import Any, Callable, Dict, List, Optional, Union import numpy as np import PIL.Image import torch import torch.nn.functional as F from transformers import CLIPImageProcessor, T5EncoderModel, T5Tokenizer from ...loaders import LoraLoaderMixin ...
diffusers/src/diffusers/pipelines/deepfloyd_if/pipeline_if_superresolution.py/0
{ "file_path": "diffusers/src/diffusers/pipelines/deepfloyd_if/pipeline_if_superresolution.py", "repo_id": "diffusers", "token_count": 18577 }
246
"""Test FastEmbed embeddings.""" import pytest from langchain_community.embeddings.fastembed import FastEmbedEmbeddings @pytest.mark.parametrize( "model_name", ["sentence-transformers/all-MiniLM-L6-v2", "BAAI/bge-small-en-v1.5"] ) @pytest.mark.parametrize("max_length", [50, 512]) @pytest.mark.parametrize("doc_em...
langchain/libs/community/tests/integration_tests/embeddings/test_fastembed.py/0
{ "file_path": "langchain/libs/community/tests/integration_tests/embeddings/test_fastembed.py", "repo_id": "langchain", "token_count": 1081 }
355
python_sources()
llama_index/llama-index-integrations/readers/llama-index-readers-readwise/llama_index/readers/readwise/BUILD/0
{ "file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-readwise/llama_index/readers/readwise/BUILD", "repo_id": "llama_index", "token_count": 6 }
1,450
{ "compilerOptions": { "composite": true, "skipLibCheck": true, "module": "ESNext", "moduleResolution": "bundler", "allowSyntheticDefaultImports": true }, "include": ["vite.config.ts"] }
opengpts/frontend/tsconfig.node.json/0
{ "file_path": "opengpts/frontend/tsconfig.node.json", "repo_id": "opengpts", "token_count": 89 }
2,024
"""Test Base Schema of documents.""" from typing import Iterator from langchain_community.document_loaders.base import BaseBlobParser from langchain_community.document_loaders.blob_loaders import Blob from langchain_core.documents import Document def test_base_blob_parser() -> None: """Verify that the eager meth...
langchain/libs/langchain/tests/unit_tests/document_loaders/test_base.py/0
{ "file_path": "langchain/libs/langchain/tests/unit_tests/document_loaders/test_base.py", "repo_id": "langchain", "token_count": 332 }
605
import numpy as np import pandas as pd results = { 'results-imagenet.csv': [ 'results-imagenet-real.csv', 'results-imagenetv2-matched-frequency.csv', 'results-sketch.csv' ], 'results-imagenet-a-clean.csv': [ 'results-imagenet-a.csv', ], 'results-imagenet-r-clean.csv...
pytorch-image-models/results/generate_csv_results.py/0
{ "file_path": "pytorch-image-models/results/generate_csv_results.py", "repo_id": "pytorch-image-models", "token_count": 1346 }
336
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/internal/querynodev2/pkoracle/key.go/0
{ "file_path": "milvus/internal/querynodev2/pkoracle/key.go", "repo_id": "milvus", "token_count": 528 }
1,767
"""Tool for the Google search API.""" from typing import Optional from langchain_core.callbacks import CallbackManagerForToolRun from langchain_core.tools import BaseTool from langchain_community.utilities.google_search import GoogleSearchAPIWrapper class GoogleSearchRun(BaseTool): """Tool that queries the Goo...
langchain/libs/community/langchain_community/tools/google_search/tool.py/0
{ "file_path": "langchain/libs/community/langchain_community/tools/google_search/tool.py", "repo_id": "langchain", "token_count": 527 }
293
# coding=utf-8 # Copyright 2018, Hao Tan, Mohit Bansal # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law...
transformers/src/transformers/models/lxmert/configuration_lxmert.py/0
{ "file_path": "transformers/src/transformers/models/lxmert/configuration_lxmert.py", "repo_id": "transformers", "token_count": 3447 }
653
package proxy import ( "context" "fmt" "math" "regexp" "strconv" "github.com/cockroachdb/errors" "github.com/golang/protobuf/proto" "github.com/samber/lo" "go.opentelemetry.io/otel" "go.uber.org/zap" "github.com/milvus-io/milvus-proto/go-api/v2/commonpb" "github.com/milvus-io/milvus-proto/go-api/v2/milvu...
milvus/internal/proxy/task_search.go/0
{ "file_path": "milvus/internal/proxy/task_search.go", "repo_id": "milvus", "token_count": 11698 }
1,824
# Moderation chain This notebook walks through examples of how to use a moderation chain, and several common ways for doing so. Moderation chains are useful for detecting text that could be hateful, violent, etc. This can be useful to apply on both user input, but also on the output of a Language Model. Some API pro...
langchain/docs/docs/guides/safety/moderation.mdx/0
{ "file_path": "langchain/docs/docs/guides/safety/moderation.mdx", "repo_id": "langchain", "token_count": 2427 }
92
python_sources()
llama_index/llama-index-integrations/readers/llama-index-readers-youtube-transcript/llama_index/readers/youtube_transcript/BUILD/0
{ "file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-youtube-transcript/llama_index/readers/youtube_transcript/BUILD", "repo_id": "llama_index", "token_count": 6 }
1,577
# init file from llama_index.tools.python_file.base import ( PythonFileToolSpec, ) __all__ = ["PythonFileToolSpec"]
llama_index/llama-index-integrations/tools/llama-index-tools-python-file/llama_index/tools/python_file/__init__.py/0
{ "file_path": "llama_index/llama-index-integrations/tools/llama-index-tools-python-file/llama_index/tools/python_file/__init__.py", "repo_id": "llama_index", "token_count": 43 }
1,495
# Epsilla This page covers how to use [Epsilla](https://github.com/epsilla-cloud/vectordb) within LangChain. It is broken into two parts: installation and setup, and then references to specific Epsilla wrappers. ## Installation and Setup - Install the Python SDK with `pip/pip3 install pyepsilla` ## Wrappers ### Ve...
langchain/docs/docs/integrations/providers/epsilla.mdx/0
{ "file_path": "langchain/docs/docs/integrations/providers/epsilla.mdx", "repo_id": "langchain", "token_count": 184 }
136
from langchain_core.vectorstores import VST, VectorStore, VectorStoreRetriever __all__ = ["VectorStore", "VectorStoreRetriever", "VST"]
langchain/libs/langchain/langchain/schema/vectorstore.py/0
{ "file_path": "langchain/libs/langchain/langchain/schema/vectorstore.py", "repo_id": "langchain", "token_count": 42 }
541
# Upstash Redis This example demonstrates how to setup chat history storage using the `UpstashRedisStore` `BaseStore` integration. ## Setup ```bash npm2yarn npm install @upstash/redis ``` ## Usage import CodeBlock from "@theme/CodeBlock"; import Example from "@examples/stores/upstash_redis_storage.ts"; <CodeBlock...
langchainjs/docs/core_docs/docs/integrations/stores/upstash_redis_storage.mdx/0
{ "file_path": "langchainjs/docs/core_docs/docs/integrations/stores/upstash_redis_storage.mdx", "repo_id": "langchainjs", "token_count": 116 }
713
<jupyter_start><jupyter_text>Volc Engine MaasThis notebook provides you with a guide on how to get started with Volc Engine's MaaS llm models.<jupyter_code># Install the package %pip install --upgrade --quiet volcengine from langchain.prompts import PromptTemplate from langchain_community.llms import VolcEngineMaasLLM...
langchain/docs/docs/integrations/llms/volcengine_maas.ipynb/0
{ "file_path": "langchain/docs/docs/integrations/llms/volcengine_maas.ipynb", "repo_id": "langchain", "token_count": 274 }
124
from typing import Any, Callable, Dict, Optional, Sequence from ibm_watson_machine_learning.foundation_models.model import Model from llama_index.core.base.llms.types import ( ChatMessage, ChatResponse, ChatResponseAsyncGen, ChatResponseGen, CompletionResponse, CompletionResponseAsyncGen, C...
llama_index/llama-index-integrations/llms/llama-index-llms-watsonx/llama_index/llms/watsonx/base.py/0
{ "file_path": "llama_index/llama-index-integrations/llms/llama-index-llms-watsonx/llama_index/llms/watsonx/base.py", "repo_id": "llama_index", "token_count": 3185 }
1,320
#!/bin/bash set -e set -x echo "check os env" platform='Linux' unamestr=$(uname) if [[ "$unamestr" == 'Linux' ]]; then platform='Linux' elif [[ "$unamestr" == 'Darwin' ]]; then platform='Mac' fi echo "platform: $platform" ns="chaos-testing" # switch namespace # kubectl config set-context --current --namespace...
milvus/tests/python_client/chaos/chaos_test.sh/0
{ "file_path": "milvus/tests/python_client/chaos/chaos_test.sh", "repo_id": "milvus", "token_count": 1527 }
2,172
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
transformers/src/transformers/models/bit/configuration_bit.py/0
{ "file_path": "transformers/src/transformers/models/bit/configuration_bit.py", "repo_id": "transformers", "token_count": 2409 }
587
# coding=utf-8 # Copyright 2022 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
transformers/tests/models/oneformer/test_image_processing_oneformer.py/0
{ "file_path": "transformers/tests/models/oneformer/test_image_processing_oneformer.py", "repo_id": "transformers", "token_count": 6134 }
815
import { Gradient } from "@gradientai/nodejs-sdk"; import { type BaseLLMCallOptions, type BaseLLMParams, LLM, } from "@langchain/core/language_models/llms"; import { getEnvironmentVariable } from "@langchain/core/utils/env"; /** * The GradientLLMParams interface defines the input parameters for * the GradientL...
langchainjs/libs/langchain-community/src/llms/gradient_ai.ts/0
{ "file_path": "langchainjs/libs/langchain-community/src/llms/gradient_ai.ts", "repo_id": "langchainjs", "token_count": 1269 }
1,001
from enum import Enum class ElevenLabsModel(str, Enum): """Models available for Eleven Labs Text2Speech.""" MULTI_LINGUAL = "eleven_multilingual_v1" MONO_LINGUAL = "eleven_monolingual_v1"
langchain/libs/community/langchain_community/tools/eleven_labs/models.py/0
{ "file_path": "langchain/libs/community/langchain_community/tools/eleven_labs/models.py", "repo_id": "langchain", "token_count": 80 }
288
from llama_index.core.vector_stores.types import BasePydanticVectorStore from llama_index.vector_stores.redis import RedisVectorStore def test_class(): names_of_base_classes = [b.__name__ for b in RedisVectorStore.__mro__] assert BasePydanticVectorStore.__name__ in names_of_base_classes
llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-redis/tests/test_vector_stores_redis.py/0
{ "file_path": "llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-redis/tests/test_vector_stores_redis.py", "repo_id": "llama_index", "token_count": 100 }
1,535
# SelecSLS **SelecSLS** uses novel selective long and short range skip connections to improve the information flow allowing for a drastically faster network without compromising accuracy. ## How do I use this model on an image? To load a pretrained model: ```py >>> import timm >>> model = timm.create_model('selecsl...
pytorch-image-models/hfdocs/source/models/selecsls.mdx/0
{ "file_path": "pytorch-image-models/hfdocs/source/models/selecsls.mdx", "repo_id": "pytorch-image-models", "token_count": 2420 }
381
from __future__ import annotations from abc import ABC, abstractmethod from typing import TYPE_CHECKING, Any, Sequence from langchain_core.runnables.config import run_in_executor if TYPE_CHECKING: from langchain_core.documents import Document class BaseDocumentTransformer(ABC): """Abstract base class for d...
langchain/libs/core/langchain_core/documents/transformers.py/0
{ "file_path": "langchain/libs/core/langchain_core/documents/transformers.py", "repo_id": "langchain", "token_count": 1075 }
388
<!--⚠️ Note that this file is in Markdown but contain specific syntax for our doc-builder (similar to MDX) that may not be rendered properly in your Markdown viewer. --> # Fully Sharded Data Parallel [Fully sharded data parallel](https://pytorch.org/docs/stable/fsdp.html) (FSDP) is developed for distributed training ...
peft/docs/source/accelerate/fsdp.md/0
{ "file_path": "peft/docs/source/accelerate/fsdp.md", "repo_id": "peft", "token_count": 2180 }
321
"""Query Rewriting Retriever Pack.""" from typing import Any, Dict, List from llama_index.core.indices.service_context import ServiceContext from llama_index.core.indices.vector_store import VectorStoreIndex from llama_index.core.llama_pack.base import BaseLlamaPack from llama_index.core.query_engine import Retriever...
llama_index/llama-index-packs/llama-index-packs-fusion-retriever/llama_index/packs/fusion_retriever/query_rewrite/base.py/0
{ "file_path": "llama_index/llama-index-packs/llama-index-packs-fusion-retriever/llama_index/packs/fusion_retriever/query_rewrite/base.py", "repo_id": "llama_index", "token_count": 959 }
1,654
from llama_index.core.llms.base import BaseLLM from llama_index.llms.gemini import Gemini def test_embedding_class(): names_of_base_classes = [b.__name__ for b in Gemini.__mro__] assert BaseLLM.__name__ in names_of_base_classes
llama_index/llama-index-integrations/llms/llama-index-llms-gemini/tests/test_llms_gemini.py/0
{ "file_path": "llama_index/llama-index-integrations/llms/llama-index-llms-gemini/tests/test_llms_gemini.py", "repo_id": "llama_index", "token_count": 89 }
1,224
<jupyter_start><jupyter_code>import argparse import gc import hashlib import itertools import logging import math import os import threading import warnings from pathlib import Path from typing import Optional import psutil import json import torch import torch.nn.functional as F import torch.utils.checkpoint from tor...
peft/examples/lora_dreambooth/lora_dreambooth_inference.ipynb/0
{ "file_path": "peft/examples/lora_dreambooth/lora_dreambooth_inference.ipynb", "repo_id": "peft", "token_count": 2282 }
340
from langchain.schema import AgentAction, AgentFinish def parse_output(message: str): FINAL_ANSWER_ACTION = "<final_answer>" includes_answer = FINAL_ANSWER_ACTION in message if includes_answer: answer = message.split(FINAL_ANSWER_ACTION)[1].strip() if "</final_answer>" in answer: ...
langchain/templates/solo-performance-prompting-agent/solo_performance_prompting_agent/parser.py/0
{ "file_path": "langchain/templates/solo-performance-prompting-agent/solo_performance_prompting_agent/parser.py", "repo_id": "langchain", "token_count": 332 }
678
import { z } from "zod"; import { zodToJsonSchema } from "zod-to-json-schema"; import { OllamaFunctions } from "langchain/experimental/chat_models/ollama_functions"; import { JsonOutputFunctionsParser } from "langchain/output_parsers"; import { PromptTemplate } from "@langchain/core/prompts"; const EXTRACTION_TEMPLAT...
langchainjs/examples/src/models/chat/ollama_functions/extraction.ts/0
{ "file_path": "langchainjs/examples/src/models/chat/ollama_functions/extraction.ts", "repo_id": "langchainjs", "token_count": 600 }
899
python_tests()
llama_index/llama-index-integrations/readers/llama-index-readers-airbyte-stripe/tests/BUILD/0
{ "file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-airbyte-stripe/tests/BUILD", "repo_id": "llama_index", "token_count": 5 }
1,370
<script lang="ts"> import { onDestroy } from "svelte"; import IconCopy from "./icons/IconCopy.svelte"; import Tooltip from "./Tooltip.svelte"; export let classNames = ""; export let value: string; let isSuccess = false; let timeout: ReturnType<typeof setTimeout>; const handleClick = async () => { // write...
chat-ui/src/lib/components/CopyToClipBoardBtn.svelte/0
{ "file_path": "chat-ui/src/lib/components/CopyToClipBoardBtn.svelte", "repo_id": "chat-ui", "token_count": 433 }
85
"""Tabular parser. Contains parsers for tabular data files. """ from pathlib import Path from typing import Any, Dict, List, Optional import pandas as pd from llama_index.core.readers.base import BaseReader from llama_index.core.schema import Document class CSVReader(BaseReader): """CSV parser. Args: ...
llama_index/llama-index-integrations/readers/llama-index-readers-file/llama_index/readers/file/tabular/base.py/0
{ "file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-file/llama_index/readers/file/tabular/base.py", "repo_id": "llama_index", "token_count": 1604 }
1,370
**TEMPLATE** ===================================== *search & replace the following keywords, e.g.:* `:%s/\[name of model\]/brand_new_bert/g` -[lowercase name of model] # e.g. brand_new_bert -[camelcase name of model] # e.g. BrandNewBert -[name of mentor] # e.g. [Peter](https://github.com/peter) -[link to origin...
transformers/templates/adding_a_new_model/ADD_NEW_MODEL_PROPOSAL_TEMPLATE.md/0
{ "file_path": "transformers/templates/adding_a_new_model/ADD_NEW_MODEL_PROPOSAL_TEMPLATE.md", "repo_id": "transformers", "token_count": 14136 }
789
<jupyter_start><jupyter_text>acreom [acreom](https://acreom.com) is a dev-first knowledge base with tasks running on local markdown files.Below is an example on how to load a local acreom vault into Langchain. As the local vault in acreom is a folder of plain text .md files, the loader requires the path to the director...
langchain/docs/docs/integrations/document_loaders/acreom.ipynb/0
{ "file_path": "langchain/docs/docs/integrations/document_loaders/acreom.ipynb", "repo_id": "langchain", "token_count": 193 }
95
import { expect, test } from "@jest/globals"; import { PuppeteerWebBaseLoader } from "../web/puppeteer.js"; test.skip("Test puppeteer web scraper loader", async () => { const loader = new PuppeteerWebBaseLoader("https://www.google.com/"); const result = await loader.load(); expect(result).toBeDefined(); expec...
langchainjs/langchain/src/document_loaders/tests/puppeteer.int.test.ts/0
{ "file_path": "langchainjs/langchain/src/document_loaders/tests/puppeteer.int.test.ts", "repo_id": "langchainjs", "token_count": 665 }
973
// Licensed to the LF AI & Data foundation under one // or more contributor license agreements. See the NOTICE file // distributed with this work for additional information // regarding copyright ownership. The ASF licenses this file // to you under the Apache License, Version 2.0 (the // "License"); you may not use th...
milvus/internal/util/importutil/binlog_file_test.go/0
{ "file_path": "milvus/internal/util/importutil/binlog_file_test.go", "repo_id": "milvus", "token_count": 12905 }
2,017
// Copyright (C) 2019-2020 Zilliz. All rights reserved. // // Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance // with the License. You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable l...
milvus/pkg/util/indexparamcheck/conf_adapter_mgr_test.go/0
{ "file_path": "milvus/pkg/util/indexparamcheck/conf_adapter_mgr_test.go", "repo_id": "milvus", "token_count": 1779 }
2,108