text stringlengths 3 1.68M | id stringlengths 13 169 | metadata dict | __index_level_0__ int64 0 2.21k |
|---|---|---|---|
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/datanode/l0_compactor_test.go/0 | {
"file_path": "milvus/internal/datanode/l0_compactor_test.go",
"repo_id": "milvus",
"token_count": 8944
} | 1,704 |
import * as uuid from "uuid";
import { AsyncCaller, AsyncCallerParams } from "./utils/async_caller.js";
import {
DataType,
Dataset,
DatasetShareSchema,
Example,
ExampleCreate,
ExampleUpdate,
Feedback,
KVMap,
LangChainBaseMessage,
Run,
RunCreate,
RunUpdate,
ScoreType,
TracerSession,
Tracer... | langsmith-sdk/js/src/client.ts/0 | {
"file_path": "langsmith-sdk/js/src/client.ts",
"repo_id": "langsmith-sdk",
"token_count": 22133
} | 1,066 |
<!--Copyright 2024 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | diffusers/docs/source/en/training/ddpo.md/0 | {
"file_path": "diffusers/docs/source/en/training/ddpo.md",
"repo_id": "diffusers",
"token_count": 321
} | 169 |
import { Cohere } from "@langchain/cohere";
const model = new Cohere({
maxTokens: 20,
apiKey: "YOUR-API-KEY", // In Node.js defaults to process.env.COHERE_API_KEY
});
const res = await model.invoke(
"What would be a good company name a company that makes colorful socks?"
);
console.log({ res });
| langchainjs/examples/src/models/llm/cohere.ts/0 | {
"file_path": "langchainjs/examples/src/models/llm/cohere.ts",
"repo_id": "langchainjs",
"token_count": 98
} | 837 |
# coding=utf-8
# Copyright 2021 T5 Authors and HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by... | transformers/src/transformers/models/t5/modeling_flax_t5.py/0 | {
"file_path": "transformers/src/transformers/models/t5/modeling_flax_t5.py",
"repo_id": "transformers",
"token_count": 32653
} | 675 |
<jupyter_start><jupyter_text>IntroductionThis notebook is designed to run inference on the [Diffuser](https://arxiv.org/abs/2205.09991) planning model for model-based RL. The notebook is modified from the authors' [original](https://colab.research.google.com/drive/1YajKhu-CUIGBJeQPehjVPJcK_b38a8Nc?usp=sharingscrollTo=5... | notebooks/diffusers/reinforcement_learning_with_diffusers.ipynb/0 | {
"file_path": "notebooks/diffusers/reinforcement_learning_with_diffusers.ipynb",
"repo_id": "notebooks",
"token_count": 8060
} | 306 |
from llama_index.tools.waii.base import WaiiToolSpec
__all__ = ["WaiiToolSpec"]
| llama_index/llama-index-integrations/tools/llama-index-tools-waii/llama_index/tools/waii/__init__.py/0 | {
"file_path": "llama_index/llama-index-integrations/tools/llama-index-tools-waii/llama_index/tools/waii/__init__.py",
"repo_id": "llama_index",
"token_count": 30
} | 1,506 |
# coding=utf-8
# Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a cop... | transformers/src/transformers/data/processors/utils.py/0 | {
"file_path": "transformers/src/transformers/data/processors/utils.py",
"repo_id": "transformers",
"token_count": 5994
} | 632 |
from langchain_community.vectorstores.tiledb import (
TileDB,
)
__all__ = [
"TileDB",
]
| langchain/libs/langchain/langchain/vectorstores/tiledb.py/0 | {
"file_path": "langchain/libs/langchain/langchain/vectorstores/tiledb.py",
"repo_id": "langchain",
"token_count": 41
} | 588 |
cluster:
enabled: true
image:
all:
repository: milvusdb/milvus
tag: master-latest
pullPolicy: IfNotPresent
rootCoordinator:
replicas: 2
activeStandby:
enabled: true # Enable active-standby when you set multiple replicas for root coordinator
queryCoordinator:
replicas: 2
activeStandby:
... | milvus/tests/python_client/standby/cluster-values.yaml/0 | {
"file_path": "milvus/tests/python_client/standby/cluster-values.yaml",
"repo_id": "milvus",
"token_count": 1664
} | 1,980 |
package memberlist_manager
import (
"context"
"errors"
"github.com/chroma/chroma-coordinator/internal/common"
"github.com/pingcap/log"
"go.uber.org/zap"
"k8s.io/client-go/util/workqueue"
)
// A memberlist manager is responsible for managing the memberlist for a
// coordinator. A memberlist consists of a store ... | chroma/go/coordinator/internal/memberlist_manager/memberlist_manager.go/0 | {
"file_path": "chroma/go/coordinator/internal/memberlist_manager/memberlist_manager.go",
"repo_id": "chroma",
"token_count": 1209
} | 43 |
import pytest
from llama_index.legacy.core.llms.types import ChatMessage
from llama_index.legacy.llms.anthropic import Anthropic
try:
import anthropic
except ImportError:
anthropic = None # type: ignore
@pytest.mark.skipif(anthropic is None, reason="anthropic not installed")
def test_basic() -> None:
ll... | llama_index/llama-index-legacy/tests/llms/test_anthropic.py/0 | {
"file_path": "llama_index/llama-index-legacy/tests/llms/test_anthropic.py",
"repo_id": "llama_index",
"token_count": 858
} | 1,625 |
from llama_index.llms.mistralai.base import MistralAI
__all__ = ["MistralAI"]
| llama_index/llama-index-integrations/llms/llama-index-llms-mistralai/llama_index/llms/mistralai/__init__.py/0 | {
"file_path": "llama_index/llama-index-integrations/llms/llama-index-llms-mistralai/llama_index/llms/mistralai/__init__.py",
"repo_id": "llama_index",
"token_count": 30
} | 1,361 |
[tool.poetry]
name = "solo-performance-prompting-agent"
version = "0.0.1"
description = "Agent that transforms a single LLM into a cognitive synergist that has multiple points of view"
authors = []
readme = "README.md"
[tool.poetry.dependencies]
python = ">=3.8.1,<4.0"
langchain = "^0.1"
openai = "<2"
duckduckgo-searc... | langchain/templates/solo-performance-prompting-agent/pyproject.toml/0 | {
"file_path": "langchain/templates/solo-performance-prompting-agent/pyproject.toml",
"repo_id": "langchain",
"token_count": 304
} | 706 |
#!/usr/bin/env bash
export PYTHONPATH="../":"${PYTHONPATH}"
export WANDB_PROJECT=dmar
export MAX_LEN=128
export m=sshleifer/student_marian_en_ro_6_1
python finetune.py \
--learning_rate=3e-4 \
--do_train \
--fp16 \
--data_dir wmt_en_ro \
--max_source_length $MAX_LEN --max_target_length $MAX_LEN --val_max_targ... | transformers/examples/research_projects/seq2seq-distillation/dynamic_bs_example.sh/0 | {
"file_path": "transformers/examples/research_projects/seq2seq-distillation/dynamic_bs_example.sh",
"repo_id": "transformers",
"token_count": 267
} | 544 |
<jupyter_start><jupyter_text>Handle parsing errorsOccasionally the LLM cannot determine what step to take because its outputs are not correctly formatted to be handled by the output parser. In this case, by default the agent errors. But you can easily control this functionality with `handle_parsing_errors`! Let's explo... | langchain/docs/docs/modules/agents/how_to/handle_parsing_errors.ipynb/0 | {
"file_path": "langchain/docs/docs/modules/agents/how_to/handle_parsing_errors.ipynb",
"repo_id": "langchain",
"token_count": 1608
} | 202 |
from llama_index.readers.rayyan.base import RayyanReader
__all__ = ["RayyanReader"]
| llama_index/llama-index-integrations/readers/llama-index-readers-rayyan/llama_index/readers/rayyan/__init__.py/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-rayyan/llama_index/readers/rayyan/__init__.py",
"repo_id": "llama_index",
"token_count": 29
} | 1,432 |
<jupyter_start><jupyter_text>Astra DBDataStax [Astra DB](https://docs.datastax.com/en/astra/home/astra.html) is a serverless vector-capable database built on Cassandra and made conveniently available through an easy-to-use JSON API.In the walkthrough, we'll demo the `SelfQueryRetriever` with an `Astra DB` vector store.... | langchain/docs/docs/integrations/retrievers/self_query/astradb.ipynb/0 | {
"file_path": "langchain/docs/docs/integrations/retrievers/self_query/astradb.ipynb",
"repo_id": "langchain",
"token_count": 1941
} | 158 |
# Structured chat
:::info
If you are using a functions-capable model like ChatOpenAI, we currently recommend that you use the [OpenAI Functions agent](/docs/modules/agents/agent_types/openai_functions_agent)
for more complex tool calling.
:::
The structured chat agent is capable of using multi-input tools.
Older age... | langchainjs/docs/core_docs/docs/modules/agents/agent_types/structured_chat.mdx/0 | {
"file_path": "langchainjs/docs/core_docs/docs/modules/agents/agent_types/structured_chat.mdx",
"repo_id": "langchainjs",
"token_count": 1110
} | 733 |
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | transformers/src/transformers/models/bert/__init__.py/0 | {
"file_path": "transformers/src/transformers/models/bert/__init__.py",
"repo_id": "transformers",
"token_count": 2639
} | 647 |
from typing import List
import datasets
from datasets.tasks import AudioClassification
from ..folder_based_builder import folder_based_builder
logger = datasets.utils.logging.get_logger(__name__)
class AudioFolderConfig(folder_based_builder.FolderBasedBuilderConfig):
"""Builder Config for AudioFolder."""
... | datasets/src/datasets/packaged_modules/audiofolder/audiofolder.py/0 | {
"file_path": "datasets/src/datasets/packaged_modules/audiofolder/audiofolder.py",
"repo_id": "datasets",
"token_count": 618
} | 135 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/datanode/metacache/meta_cache_test.go/0 | {
"file_path": "milvus/internal/datanode/metacache/meta_cache_test.go",
"repo_id": "milvus",
"token_count": 2666
} | 1,699 |
from langchain_community.tools.playwright.base import (
BaseBrowserTool,
)
__all__ = ["BaseBrowserTool"]
| langchain/libs/langchain/langchain/tools/playwright/base.py/0 | {
"file_path": "langchain/libs/langchain/langchain/tools/playwright/base.py",
"repo_id": "langchain",
"token_count": 36
} | 557 |
# coding=utf-8
# Copyright 2021 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | transformers/src/transformers/models/bigbird_pegasus/convert_bigbird_pegasus_tf_to_pytorch.py/0 | {
"file_path": "transformers/src/transformers/models/bigbird_pegasus/convert_bigbird_pegasus_tf_to_pytorch.py",
"repo_id": "transformers",
"token_count": 2618
} | 586 |
# Detoxifying a Language Model using PPO
Language models (LMs) are known to sometimes generate toxic outputs. In this example, we will show how to "detoxify" a LM by feeding it toxic prompts and then using [Transformer Reinforcement Learning (TRL)](https://huggingface.co/docs/trl/index) and Proximal Policy Optimizatio... | trl/docs/source/detoxifying_a_lm.mdx/0 | {
"file_path": "trl/docs/source/detoxifying_a_lm.mdx",
"repo_id": "trl",
"token_count": 3783
} | 806 |
/* eslint-disable no-process-env */
/* eslint-disable @typescript-eslint/no-non-null-assertion */
import { test } from "@jest/globals";
import { OpenAI } from "@langchain/openai";
import { StartExecutionAWSSfnTool } from "@langchain/community/tools/aws_sfn";
import { AWSSfnToolkit } from "@langchain/community/agents/t... | langchainjs/langchain/src/agents/tests/aws_sfn.int.test.ts/0 | {
"file_path": "langchainjs/langchain/src/agents/tests/aws_sfn.int.test.ts",
"repo_id": "langchainjs",
"token_count": 751
} | 895 |
# Copyright 2023 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | trl/tests/test_dpo_trainer.py/0 | {
"file_path": "trl/tests/test_dpo_trainer.py",
"repo_id": "trl",
"token_count": 11138
} | 824 |
poetry_requirements(
name="poetry",
)
| llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-pinecone/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-pinecone/BUILD",
"repo_id": "llama_index",
"token_count": 18
} | 1,615 |
from llama_index.core.multi_modal_llms.base import MultiModalLLM
from llama_index.multi_modal_llms.replicate_multi_modal import ReplicateMultiModal
def test_embedding_class():
names_of_base_classes = [b.__name__ for b in ReplicateMultiModal.__mro__]
assert MultiModalLLM.__name__ in names_of_base_classes
| llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-replicate-multi-modal/tests/test_multi-modal-llms_replicate_multi_modal.py/0 | {
"file_path": "llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-replicate-multi-modal/tests/test_multi-modal-llms_replicate_multi_modal.py",
"repo_id": "llama_index",
"token_count": 116
} | 1,431 |
# MIT License
# Copyright (c) 2019 Yang Liu and the HuggingFace team
# Permission is hereby granted, free of charge, to any person obtaining a copy
# of this software and associated documentation files (the "Software"), to deal
# in the Software without restriction, including without limitation the rights
# to use, c... | transformers/examples/research_projects/bertabs/modeling_bertabs.py/0 | {
"file_path": "transformers/examples/research_projects/bertabs/modeling_bertabs.py",
"repo_id": "transformers",
"token_count": 17886
} | 531 |
"""Base interface for chains combining documents."""
from abc import ABC, abstractmethod
from typing import Any, Dict, List, Optional, Tuple, Type
from langchain_core.callbacks import (
AsyncCallbackManagerForChainRun,
CallbackManagerForChainRun,
)
from langchain_core.documents import Document
from langchain_... | langchain/libs/langchain/langchain/chains/combine_documents/base.py/0 | {
"file_path": "langchain/libs/langchain/langchain/chains/combine_documents/base.py",
"repo_id": "langchain",
"token_count": 3024
} | 463 |
import CodeBlock from "@theme/CodeBlock";
import CreateClientExample from "@examples/indexes/vector_stores/upstash/create_client.ts";
import IndexQueryExample from "@examples/indexes/vector_stores/upstash/index_and_query_docs.ts";
import DeleteExample from "@examples/indexes/vector_stores/upstash/delete_docs.ts";
impor... | langchainjs/docs/core_docs/docs/integrations/vectorstores/upstash.mdx/0 | {
"file_path": "langchainjs/docs/core_docs/docs/integrations/vectorstores/upstash.mdx",
"repo_id": "langchainjs",
"token_count": 497
} | 746 |
---
sidebar_position: 5
---
# Contribute Integrations
To begin, make sure you have all the dependencies outlined in guide on [Contributing Code](./code).
There are a few different places you can contribute integrations for LangChain:
- **Community**: For lighter-weight integrations that are primarily maintained by L... | langchain/docs/docs/contributing/integrations.mdx/0 | {
"file_path": "langchain/docs/docs/contributing/integrations.mdx",
"repo_id": "langchain",
"token_count": 1416
} | 86 |
from langchain.memory import ConversationBufferMemory
from langchain.output_parsers.list import CommaSeparatedListOutputParser
from langchain.prompts import PromptTemplate
from langchain.sql_database import SQLDatabase
from langchain_experimental.sql.base import SQLDatabaseChain, SQLDatabaseSequentialChain
from tests.... | langchain/libs/experimental/tests/unit_tests/test_sql.py/0 | {
"file_path": "langchain/libs/experimental/tests/unit_tests/test_sql.py",
"repo_id": "langchain",
"token_count": 1515
} | 475 |
version: '3.5'
services:
etcd:
image: quay.io/coreos/etcd:v3.5.5
environment:
- ETCD_AUTO_COMPACTION_MODE=revision
- ETCD_AUTO_COMPACTION_RETENTION=1000
- ETCD_QUOTA_BACKEND_BYTES=4294967296
- ETCD_SNAPSHOT_COUNT=50000
volumes:
- ${DOCKER_VOLUME_DIRECTORY:-.}/volumes/etcd:/e... | milvus/deployments/docker/dev/docker-compose-apple-silicon.yml/0 | {
"file_path": "milvus/deployments/docker/dev/docker-compose-apple-silicon.yml",
"repo_id": "milvus",
"token_count": 1140
} | 1,903 |
# coding=utf-8
# Copyright 2021 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | transformers/src/transformers/models/deit/convert_deit_timm_to_pytorch.py/0 | {
"file_path": "transformers/src/transformers/models/deit/convert_deit_timm_to_pytorch.py",
"repo_id": "transformers",
"token_count": 3875
} | 589 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/querynodev2/pipeline/pipeline.go/0 | {
"file_path": "milvus/internal/querynodev2/pipeline/pipeline.go",
"repo_id": "milvus",
"token_count": 941
} | 1,990 |
# Microsoft
All functionality related to `Microsoft Azure` and other `Microsoft` products.
## Chat Models
### Azure OpenAI
>[Microsoft Azure](https://en.wikipedia.org/wiki/Microsoft_Azure), often referred to as `Azure` is a cloud computing platform run by `Microsoft`, which offers access, management, and development... | langchain/docs/docs/integrations/platforms/microsoft.mdx/0 | {
"file_path": "langchain/docs/docs/integrations/platforms/microsoft.mdx",
"repo_id": "langchain",
"token_count": 3312
} | 124 |
"""Test TritonTensorRT Chat API wrapper."""
from langchain_nvidia_trt import TritonTensorRTLLM
def test_initialization() -> None:
"""Test integration initialization."""
TritonTensorRTLLM(model_name="ensemble", server_url="http://localhost:8001")
| langchain/libs/partners/nvidia-trt/tests/unit_tests/test_llms.py/0 | {
"file_path": "langchain/libs/partners/nvidia-trt/tests/unit_tests/test_llms.py",
"repo_id": "langchain",
"token_count": 85
} | 636 |
import type { Conversation } from "$lib/types/Conversation";
import type { Message } from "$lib/types/Message";
export function addChildren(
conv: Pick<Conversation, "messages" | "rootMessageId">,
message: Omit<Message, "id">,
parentId?: Message["id"]
): Message["id"] {
// if this is the first message we just push... | chat-ui/src/lib/utils/tree/addChildren.ts/0 | {
"file_path": "chat-ui/src/lib/utils/tree/addChildren.ts",
"repo_id": "chat-ui",
"token_count": 496
} | 97 |
python_sources()
| llama_index/experimental/splitter_playground/BUILD/0 | {
"file_path": "llama_index/experimental/splitter_playground/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,156 |
import { Source } from "./SourceBubble";
export function InlineCitation(props: {
source: Source;
sourceNumber: number;
highlighted: boolean;
onMouseEnter: () => any;
onMouseLeave: () => any;
}) {
const { source, sourceNumber, highlighted, onMouseEnter, onMouseLeave } =
props;
return (
<a
hr... | chat-langchain/chat-langchain/app/components/InlineCitation.tsx/0 | {
"file_path": "chat-langchain/chat-langchain/app/components/InlineCitation.tsx",
"repo_id": "chat-langchain",
"token_count": 254
} | 9 |
// Adapted from turboderp exllama: https://github.com/turboderp/exllama
#ifndef _matrix_cuh
#define _matrix_cuh
#include <cuda_runtime.h>
#include <cuda_fp16.h>
class MatrixView_half
{
public:
const half* data;
const int height;
const int width;
__device__ __forceinline__ MatrixView_half(const half*... | text-generation-inference/server/exllama_kernels/exllama_kernels/matrix.cuh/0 | {
"file_path": "text-generation-inference/server/exllama_kernels/exllama_kernels/matrix.cuh",
"repo_id": "text-generation-inference",
"token_count": 5380
} | 397 |
python_sources()
| llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-singlestoredb/llama_index/vector_stores/singlestoredb/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-singlestoredb/llama_index/vector_stores/singlestoredb/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,536 |
"""Code to support various indexing workflows.
Provides code to:
* Create knowledge graphs from data.
* Support indexing workflows from LangChain data loaders to vectorstores.
For indexing workflows, this code is used to avoid writing duplicated content
into the vectostore and to avoid over-writing content if it's ... | langchain/libs/langchain/langchain/indexes/__init__.py/0 | {
"file_path": "langchain/libs/langchain/langchain/indexes/__init__.py",
"repo_id": "langchain",
"token_count": 268
} | 532 |
package indexparamcheck
import (
"strconv"
"testing"
"github.com/stretchr/testify/assert"
"github.com/milvus-io/milvus-proto/go-api/v2/schemapb"
"github.com/milvus-io/milvus/pkg/util/metric"
)
func Test_binIVFFlatChecker_CheckTrain(t *testing.T) {
validParams := map[string]string{
DIM: strconv.Itoa(128),... | milvus/pkg/util/indexparamcheck/bin_ivf_flat_checker_test.go/0 | {
"file_path": "milvus/pkg/util/indexparamcheck/bin_ivf_flat_checker_test.go",
"repo_id": "milvus",
"token_count": 2025
} | 1,969 |
# Attribution-NonCommercial 4.0 International (CC BY-NC 4.0)
# William Peebles and Saining Xie
#
# Copyright (c) 2021 OpenAI
# MIT License
#
# Copyright 2024 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance wi... | diffusers/src/diffusers/pipelines/dit/pipeline_dit.py/0 | {
"file_path": "diffusers/src/diffusers/pipelines/dit/pipeline_dit.py",
"repo_id": "diffusers",
"token_count": 4179
} | 256 |
python_tests()
| llama_index/llama-index-integrations/readers/llama-index-readers-jaguar/tests/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-jaguar/tests/BUILD",
"repo_id": "llama_index",
"token_count": 5
} | 1,393 |
"""Test LLM program."""
import json
from unittest.mock import MagicMock
from llama_index.core.base.llms.types import (
ChatMessage,
ChatResponse,
CompletionResponse,
LLMMetadata,
MessageRole,
)
from llama_index.core.bridge.pydantic import BaseModel
from llama_index.core.output_parsers.pydantic imp... | llama_index/llama-index-core/tests/program/test_llm_program.py/0 | {
"file_path": "llama_index/llama-index-core/tests/program/test_llm_program.py",
"repo_id": "llama_index",
"token_count": 1103
} | 1,228 |
<!---
Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or ... | transformers/docs/source/zh/installation.md/0 | {
"file_path": "transformers/docs/source/zh/installation.md",
"repo_id": "transformers",
"token_count": 4826
} | 514 |
[build-system]
build-backend = "poetry.core.masonry.api"
requires = ["poetry-core"]
[tool.codespell]
check-filenames = true
check-hidden = true
skip = "*.csv,*.html,*.json,*.jsonl,*.pdf,*.txt,*.ipynb"
[tool.llamahub]
classes = ["BedrockEmbedding", "Models"]
contains_example = false
import_path = "llama_index.embeddin... | llama_index/llama-index-integrations/embeddings/llama-index-embeddings-bedrock/pyproject.toml/0 | {
"file_path": "llama_index/llama-index-integrations/embeddings/llama-index-embeddings-bedrock/pyproject.toml",
"repo_id": "llama_index",
"token_count": 654
} | 1,221 |
#![allow(unused)]
/// A fast implementation of mamba for inference only.
/// This is based on: https://github.com/LaurentMazare/mamba.rs
use crate::models::with_tracing::{linear, linear_no_bias, Linear};
use candle::{DType, Device, IndexOp, Module, Result, Tensor, D};
use candle_nn::{RmsNorm, VarBuilder};
const D_CONV... | candle/candle-transformers/src/models/mamba.rs/0 | {
"file_path": "candle/candle-transformers/src/models/mamba.rs",
"repo_id": "candle",
"token_count": 3784
} | 72 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/proxy/accesslog/minio_handler_test.go/0 | {
"file_path": "milvus/internal/proxy/accesslog/minio_handler_test.go",
"repo_id": "milvus",
"token_count": 1759
} | 1,734 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/pkg/config/event_dispatcher.go/0 | {
"file_path": "milvus/pkg/config/event_dispatcher.go",
"repo_id": "milvus",
"token_count": 991
} | 1,881 |
const load = require("langchain/load");
const load_serializable = require("langchain/load/serializable");
const agents = require("langchain/agents");
const agents_toolkits = require("langchain/agents/toolkits");
const agents_toolkits_connery = require("langchain/agents/toolkits/connery");
const agents_format_scratchpad... | langchainjs/environment_tests/test-exports-cjs/src/entrypoints.js/0 | {
"file_path": "langchainjs/environment_tests/test-exports-cjs/src/entrypoints.js",
"repo_id": "langchainjs",
"token_count": 2439
} | 831 |
#!/bin/bash
# author emaill will be jenkins's email, when the pr branch is not fast forward compared with master branch
# Exit immediately for non zero status
set -e
function get_author_email(){
email=$(git --no-pager show -s --format=\'%ae\' HEAD )
# Get last commit author when Jenkins submit a merge commit
... | milvus/tests/scripts/get_author_email.sh/0 | {
"file_path": "milvus/tests/scripts/get_author_email.sh",
"repo_id": "milvus",
"token_count": 174
} | 1,908 |
import { expect, test } from "@jest/globals";
import { DataSourceOptions } from "typeorm";
import { OpenAIEmbeddings } from "@langchain/openai";
import { TypeORMVectorStore } from "../typeorm.js";
test.skip("Test embeddings creation", async () => {
const args = {
postgresConnectionOptions: {
type: "postgre... | langchainjs/libs/langchain-community/src/vectorstores/tests/typeorm.int.test.ts/0 | {
"file_path": "langchainjs/libs/langchain-community/src/vectorstores/tests/typeorm.int.test.ts",
"repo_id": "langchainjs",
"token_count": 459
} | 1,030 |
# coding=utf-8
# Copyright 2022 Microsoft Research and The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless... | transformers/src/transformers/models/layoutlmv3/modeling_tf_layoutlmv3.py/0 | {
"file_path": "transformers/src/transformers/models/layoutlmv3/modeling_tf_layoutlmv3.py",
"repo_id": "transformers",
"token_count": 34245
} | 620 |
import os
from pymilvus import connections, Index, MilvusException
from utils.util_log import test_log as log
from base.collection_wrapper import ApiCollectionWrapper
from common import common_func as cf
from common import common_type as ct
def e2e_milvus(host, c_name):
""" e2e milvus """
log.debug(f'pid: {... | milvus/tests/python_client/scale/scale_common.py/0 | {
"file_path": "milvus/tests/python_client/scale/scale_common.py",
"repo_id": "milvus",
"token_count": 825
} | 1,902 |
# Copyright 2021 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | transformers/tests/models/wav2vec2_with_lm/test_processor_wav2vec2_with_lm.py/0 | {
"file_path": "transformers/tests/models/wav2vec2_with_lm/test_processor_wav2vec2_with_lm.py",
"repo_id": "transformers",
"token_count": 8780
} | 757 |
import html
import inspect
import re
import urllib.parse as ul
from typing import Any, Callable, Dict, List, Optional, Union
import numpy as np
import PIL.Image
import torch
import torch.nn.functional as F
from transformers import CLIPImageProcessor, T5EncoderModel, T5Tokenizer
from ...loaders import LoraLoaderMixin
... | diffusers/src/diffusers/pipelines/deepfloyd_if/pipeline_if_superresolution.py/0 | {
"file_path": "diffusers/src/diffusers/pipelines/deepfloyd_if/pipeline_if_superresolution.py",
"repo_id": "diffusers",
"token_count": 18577
} | 246 |
"""Test FastEmbed embeddings."""
import pytest
from langchain_community.embeddings.fastembed import FastEmbedEmbeddings
@pytest.mark.parametrize(
"model_name", ["sentence-transformers/all-MiniLM-L6-v2", "BAAI/bge-small-en-v1.5"]
)
@pytest.mark.parametrize("max_length", [50, 512])
@pytest.mark.parametrize("doc_em... | langchain/libs/community/tests/integration_tests/embeddings/test_fastembed.py/0 | {
"file_path": "langchain/libs/community/tests/integration_tests/embeddings/test_fastembed.py",
"repo_id": "langchain",
"token_count": 1081
} | 355 |
python_sources()
| llama_index/llama-index-integrations/readers/llama-index-readers-readwise/llama_index/readers/readwise/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-readwise/llama_index/readers/readwise/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,450 |
{
"compilerOptions": {
"composite": true,
"skipLibCheck": true,
"module": "ESNext",
"moduleResolution": "bundler",
"allowSyntheticDefaultImports": true
},
"include": ["vite.config.ts"]
}
| opengpts/frontend/tsconfig.node.json/0 | {
"file_path": "opengpts/frontend/tsconfig.node.json",
"repo_id": "opengpts",
"token_count": 89
} | 2,024 |
"""Test Base Schema of documents."""
from typing import Iterator
from langchain_community.document_loaders.base import BaseBlobParser
from langchain_community.document_loaders.blob_loaders import Blob
from langchain_core.documents import Document
def test_base_blob_parser() -> None:
"""Verify that the eager meth... | langchain/libs/langchain/tests/unit_tests/document_loaders/test_base.py/0 | {
"file_path": "langchain/libs/langchain/tests/unit_tests/document_loaders/test_base.py",
"repo_id": "langchain",
"token_count": 332
} | 605 |
import numpy as np
import pandas as pd
results = {
'results-imagenet.csv': [
'results-imagenet-real.csv',
'results-imagenetv2-matched-frequency.csv',
'results-sketch.csv'
],
'results-imagenet-a-clean.csv': [
'results-imagenet-a.csv',
],
'results-imagenet-r-clean.csv... | pytorch-image-models/results/generate_csv_results.py/0 | {
"file_path": "pytorch-image-models/results/generate_csv_results.py",
"repo_id": "pytorch-image-models",
"token_count": 1346
} | 336 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/querynodev2/pkoracle/key.go/0 | {
"file_path": "milvus/internal/querynodev2/pkoracle/key.go",
"repo_id": "milvus",
"token_count": 528
} | 1,767 |
"""Tool for the Google search API."""
from typing import Optional
from langchain_core.callbacks import CallbackManagerForToolRun
from langchain_core.tools import BaseTool
from langchain_community.utilities.google_search import GoogleSearchAPIWrapper
class GoogleSearchRun(BaseTool):
"""Tool that queries the Goo... | langchain/libs/community/langchain_community/tools/google_search/tool.py/0 | {
"file_path": "langchain/libs/community/langchain_community/tools/google_search/tool.py",
"repo_id": "langchain",
"token_count": 527
} | 293 |
# coding=utf-8
# Copyright 2018, Hao Tan, Mohit Bansal
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law... | transformers/src/transformers/models/lxmert/configuration_lxmert.py/0 | {
"file_path": "transformers/src/transformers/models/lxmert/configuration_lxmert.py",
"repo_id": "transformers",
"token_count": 3447
} | 653 |
package proxy
import (
"context"
"fmt"
"math"
"regexp"
"strconv"
"github.com/cockroachdb/errors"
"github.com/golang/protobuf/proto"
"github.com/samber/lo"
"go.opentelemetry.io/otel"
"go.uber.org/zap"
"github.com/milvus-io/milvus-proto/go-api/v2/commonpb"
"github.com/milvus-io/milvus-proto/go-api/v2/milvu... | milvus/internal/proxy/task_search.go/0 | {
"file_path": "milvus/internal/proxy/task_search.go",
"repo_id": "milvus",
"token_count": 11698
} | 1,824 |
# Moderation chain
This notebook walks through examples of how to use a moderation chain, and several common ways for doing so.
Moderation chains are useful for detecting text that could be hateful, violent, etc. This can be useful to apply on both user input, but also on the output of a Language Model.
Some API pro... | langchain/docs/docs/guides/safety/moderation.mdx/0 | {
"file_path": "langchain/docs/docs/guides/safety/moderation.mdx",
"repo_id": "langchain",
"token_count": 2427
} | 92 |
python_sources()
| llama_index/llama-index-integrations/readers/llama-index-readers-youtube-transcript/llama_index/readers/youtube_transcript/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-youtube-transcript/llama_index/readers/youtube_transcript/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,577 |
# init file
from llama_index.tools.python_file.base import (
PythonFileToolSpec,
)
__all__ = ["PythonFileToolSpec"]
| llama_index/llama-index-integrations/tools/llama-index-tools-python-file/llama_index/tools/python_file/__init__.py/0 | {
"file_path": "llama_index/llama-index-integrations/tools/llama-index-tools-python-file/llama_index/tools/python_file/__init__.py",
"repo_id": "llama_index",
"token_count": 43
} | 1,495 |
# Epsilla
This page covers how to use [Epsilla](https://github.com/epsilla-cloud/vectordb) within LangChain.
It is broken into two parts: installation and setup, and then references to specific Epsilla wrappers.
## Installation and Setup
- Install the Python SDK with `pip/pip3 install pyepsilla`
## Wrappers
### Ve... | langchain/docs/docs/integrations/providers/epsilla.mdx/0 | {
"file_path": "langchain/docs/docs/integrations/providers/epsilla.mdx",
"repo_id": "langchain",
"token_count": 184
} | 136 |
from langchain_core.vectorstores import VST, VectorStore, VectorStoreRetriever
__all__ = ["VectorStore", "VectorStoreRetriever", "VST"]
| langchain/libs/langchain/langchain/schema/vectorstore.py/0 | {
"file_path": "langchain/libs/langchain/langchain/schema/vectorstore.py",
"repo_id": "langchain",
"token_count": 42
} | 541 |
# Upstash Redis
This example demonstrates how to setup chat history storage using the `UpstashRedisStore` `BaseStore` integration.
## Setup
```bash npm2yarn
npm install @upstash/redis
```
## Usage
import CodeBlock from "@theme/CodeBlock";
import Example from "@examples/stores/upstash_redis_storage.ts";
<CodeBlock... | langchainjs/docs/core_docs/docs/integrations/stores/upstash_redis_storage.mdx/0 | {
"file_path": "langchainjs/docs/core_docs/docs/integrations/stores/upstash_redis_storage.mdx",
"repo_id": "langchainjs",
"token_count": 116
} | 713 |
<jupyter_start><jupyter_text>Volc Engine MaasThis notebook provides you with a guide on how to get started with Volc Engine's MaaS llm models.<jupyter_code># Install the package
%pip install --upgrade --quiet volcengine
from langchain.prompts import PromptTemplate
from langchain_community.llms import VolcEngineMaasLLM... | langchain/docs/docs/integrations/llms/volcengine_maas.ipynb/0 | {
"file_path": "langchain/docs/docs/integrations/llms/volcengine_maas.ipynb",
"repo_id": "langchain",
"token_count": 274
} | 124 |
from typing import Any, Callable, Dict, Optional, Sequence
from ibm_watson_machine_learning.foundation_models.model import Model
from llama_index.core.base.llms.types import (
ChatMessage,
ChatResponse,
ChatResponseAsyncGen,
ChatResponseGen,
CompletionResponse,
CompletionResponseAsyncGen,
C... | llama_index/llama-index-integrations/llms/llama-index-llms-watsonx/llama_index/llms/watsonx/base.py/0 | {
"file_path": "llama_index/llama-index-integrations/llms/llama-index-llms-watsonx/llama_index/llms/watsonx/base.py",
"repo_id": "llama_index",
"token_count": 3185
} | 1,320 |
#!/bin/bash
set -e
set -x
echo "check os env"
platform='Linux'
unamestr=$(uname)
if [[ "$unamestr" == 'Linux' ]]; then
platform='Linux'
elif [[ "$unamestr" == 'Darwin' ]]; then
platform='Mac'
fi
echo "platform: $platform"
ns="chaos-testing"
# switch namespace
# kubectl config set-context --current --namespace... | milvus/tests/python_client/chaos/chaos_test.sh/0 | {
"file_path": "milvus/tests/python_client/chaos/chaos_test.sh",
"repo_id": "milvus",
"token_count": 1527
} | 2,172 |
# coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | transformers/src/transformers/models/bit/configuration_bit.py/0 | {
"file_path": "transformers/src/transformers/models/bit/configuration_bit.py",
"repo_id": "transformers",
"token_count": 2409
} | 587 |
# coding=utf-8
# Copyright 2022 HuggingFace Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or ag... | transformers/tests/models/oneformer/test_image_processing_oneformer.py/0 | {
"file_path": "transformers/tests/models/oneformer/test_image_processing_oneformer.py",
"repo_id": "transformers",
"token_count": 6134
} | 815 |
import { Gradient } from "@gradientai/nodejs-sdk";
import {
type BaseLLMCallOptions,
type BaseLLMParams,
LLM,
} from "@langchain/core/language_models/llms";
import { getEnvironmentVariable } from "@langchain/core/utils/env";
/**
* The GradientLLMParams interface defines the input parameters for
* the GradientL... | langchainjs/libs/langchain-community/src/llms/gradient_ai.ts/0 | {
"file_path": "langchainjs/libs/langchain-community/src/llms/gradient_ai.ts",
"repo_id": "langchainjs",
"token_count": 1269
} | 1,001 |
from enum import Enum
class ElevenLabsModel(str, Enum):
"""Models available for Eleven Labs Text2Speech."""
MULTI_LINGUAL = "eleven_multilingual_v1"
MONO_LINGUAL = "eleven_monolingual_v1"
| langchain/libs/community/langchain_community/tools/eleven_labs/models.py/0 | {
"file_path": "langchain/libs/community/langchain_community/tools/eleven_labs/models.py",
"repo_id": "langchain",
"token_count": 80
} | 288 |
from llama_index.core.vector_stores.types import BasePydanticVectorStore
from llama_index.vector_stores.redis import RedisVectorStore
def test_class():
names_of_base_classes = [b.__name__ for b in RedisVectorStore.__mro__]
assert BasePydanticVectorStore.__name__ in names_of_base_classes
| llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-redis/tests/test_vector_stores_redis.py/0 | {
"file_path": "llama_index/llama-index-integrations/vector_stores/llama-index-vector-stores-redis/tests/test_vector_stores_redis.py",
"repo_id": "llama_index",
"token_count": 100
} | 1,535 |
# SelecSLS
**SelecSLS** uses novel selective long and short range skip connections to improve the information flow allowing for a drastically faster network without compromising accuracy.
## How do I use this model on an image?
To load a pretrained model:
```py
>>> import timm
>>> model = timm.create_model('selecsl... | pytorch-image-models/hfdocs/source/models/selecsls.mdx/0 | {
"file_path": "pytorch-image-models/hfdocs/source/models/selecsls.mdx",
"repo_id": "pytorch-image-models",
"token_count": 2420
} | 381 |
from __future__ import annotations
from abc import ABC, abstractmethod
from typing import TYPE_CHECKING, Any, Sequence
from langchain_core.runnables.config import run_in_executor
if TYPE_CHECKING:
from langchain_core.documents import Document
class BaseDocumentTransformer(ABC):
"""Abstract base class for d... | langchain/libs/core/langchain_core/documents/transformers.py/0 | {
"file_path": "langchain/libs/core/langchain_core/documents/transformers.py",
"repo_id": "langchain",
"token_count": 1075
} | 388 |
<!--⚠️ Note that this file is in Markdown but contain specific syntax for our doc-builder (similar to MDX) that may not be
rendered properly in your Markdown viewer.
-->
# Fully Sharded Data Parallel
[Fully sharded data parallel](https://pytorch.org/docs/stable/fsdp.html) (FSDP) is developed for distributed training ... | peft/docs/source/accelerate/fsdp.md/0 | {
"file_path": "peft/docs/source/accelerate/fsdp.md",
"repo_id": "peft",
"token_count": 2180
} | 321 |
"""Query Rewriting Retriever Pack."""
from typing import Any, Dict, List
from llama_index.core.indices.service_context import ServiceContext
from llama_index.core.indices.vector_store import VectorStoreIndex
from llama_index.core.llama_pack.base import BaseLlamaPack
from llama_index.core.query_engine import Retriever... | llama_index/llama-index-packs/llama-index-packs-fusion-retriever/llama_index/packs/fusion_retriever/query_rewrite/base.py/0 | {
"file_path": "llama_index/llama-index-packs/llama-index-packs-fusion-retriever/llama_index/packs/fusion_retriever/query_rewrite/base.py",
"repo_id": "llama_index",
"token_count": 959
} | 1,654 |
from llama_index.core.llms.base import BaseLLM
from llama_index.llms.gemini import Gemini
def test_embedding_class():
names_of_base_classes = [b.__name__ for b in Gemini.__mro__]
assert BaseLLM.__name__ in names_of_base_classes
| llama_index/llama-index-integrations/llms/llama-index-llms-gemini/tests/test_llms_gemini.py/0 | {
"file_path": "llama_index/llama-index-integrations/llms/llama-index-llms-gemini/tests/test_llms_gemini.py",
"repo_id": "llama_index",
"token_count": 89
} | 1,224 |
<jupyter_start><jupyter_code>import argparse
import gc
import hashlib
import itertools
import logging
import math
import os
import threading
import warnings
from pathlib import Path
from typing import Optional
import psutil
import json
import torch
import torch.nn.functional as F
import torch.utils.checkpoint
from tor... | peft/examples/lora_dreambooth/lora_dreambooth_inference.ipynb/0 | {
"file_path": "peft/examples/lora_dreambooth/lora_dreambooth_inference.ipynb",
"repo_id": "peft",
"token_count": 2282
} | 340 |
from langchain.schema import AgentAction, AgentFinish
def parse_output(message: str):
FINAL_ANSWER_ACTION = "<final_answer>"
includes_answer = FINAL_ANSWER_ACTION in message
if includes_answer:
answer = message.split(FINAL_ANSWER_ACTION)[1].strip()
if "</final_answer>" in answer:
... | langchain/templates/solo-performance-prompting-agent/solo_performance_prompting_agent/parser.py/0 | {
"file_path": "langchain/templates/solo-performance-prompting-agent/solo_performance_prompting_agent/parser.py",
"repo_id": "langchain",
"token_count": 332
} | 678 |
import { z } from "zod";
import { zodToJsonSchema } from "zod-to-json-schema";
import { OllamaFunctions } from "langchain/experimental/chat_models/ollama_functions";
import { JsonOutputFunctionsParser } from "langchain/output_parsers";
import { PromptTemplate } from "@langchain/core/prompts";
const EXTRACTION_TEMPLAT... | langchainjs/examples/src/models/chat/ollama_functions/extraction.ts/0 | {
"file_path": "langchainjs/examples/src/models/chat/ollama_functions/extraction.ts",
"repo_id": "langchainjs",
"token_count": 600
} | 899 |
python_tests()
| llama_index/llama-index-integrations/readers/llama-index-readers-airbyte-stripe/tests/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-airbyte-stripe/tests/BUILD",
"repo_id": "llama_index",
"token_count": 5
} | 1,370 |
<script lang="ts">
import { onDestroy } from "svelte";
import IconCopy from "./icons/IconCopy.svelte";
import Tooltip from "./Tooltip.svelte";
export let classNames = "";
export let value: string;
let isSuccess = false;
let timeout: ReturnType<typeof setTimeout>;
const handleClick = async () => {
// write... | chat-ui/src/lib/components/CopyToClipBoardBtn.svelte/0 | {
"file_path": "chat-ui/src/lib/components/CopyToClipBoardBtn.svelte",
"repo_id": "chat-ui",
"token_count": 433
} | 85 |
"""Tabular parser.
Contains parsers for tabular data files.
"""
from pathlib import Path
from typing import Any, Dict, List, Optional
import pandas as pd
from llama_index.core.readers.base import BaseReader
from llama_index.core.schema import Document
class CSVReader(BaseReader):
"""CSV parser.
Args:
... | llama_index/llama-index-integrations/readers/llama-index-readers-file/llama_index/readers/file/tabular/base.py/0 | {
"file_path": "llama_index/llama-index-integrations/readers/llama-index-readers-file/llama_index/readers/file/tabular/base.py",
"repo_id": "llama_index",
"token_count": 1604
} | 1,370 |
**TEMPLATE**
=====================================
*search & replace the following keywords, e.g.:*
`:%s/\[name of model\]/brand_new_bert/g`
-[lowercase name of model] # e.g. brand_new_bert
-[camelcase name of model] # e.g. BrandNewBert
-[name of mentor] # e.g. [Peter](https://github.com/peter)
-[link to origin... | transformers/templates/adding_a_new_model/ADD_NEW_MODEL_PROPOSAL_TEMPLATE.md/0 | {
"file_path": "transformers/templates/adding_a_new_model/ADD_NEW_MODEL_PROPOSAL_TEMPLATE.md",
"repo_id": "transformers",
"token_count": 14136
} | 789 |
<jupyter_start><jupyter_text>acreom [acreom](https://acreom.com) is a dev-first knowledge base with tasks running on local markdown files.Below is an example on how to load a local acreom vault into Langchain. As the local vault in acreom is a folder of plain text .md files, the loader requires the path to the director... | langchain/docs/docs/integrations/document_loaders/acreom.ipynb/0 | {
"file_path": "langchain/docs/docs/integrations/document_loaders/acreom.ipynb",
"repo_id": "langchain",
"token_count": 193
} | 95 |
import { expect, test } from "@jest/globals";
import { PuppeteerWebBaseLoader } from "../web/puppeteer.js";
test.skip("Test puppeteer web scraper loader", async () => {
const loader = new PuppeteerWebBaseLoader("https://www.google.com/");
const result = await loader.load();
expect(result).toBeDefined();
expec... | langchainjs/langchain/src/document_loaders/tests/puppeteer.int.test.ts/0 | {
"file_path": "langchainjs/langchain/src/document_loaders/tests/puppeteer.int.test.ts",
"repo_id": "langchainjs",
"token_count": 665
} | 973 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/util/importutil/binlog_file_test.go/0 | {
"file_path": "milvus/internal/util/importutil/binlog_file_test.go",
"repo_id": "milvus",
"token_count": 12905
} | 2,017 |
// Copyright (C) 2019-2020 Zilliz. All rights reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable l... | milvus/pkg/util/indexparamcheck/conf_adapter_mgr_test.go/0 | {
"file_path": "milvus/pkg/util/indexparamcheck/conf_adapter_mgr_test.go",
"repo_id": "milvus",
"token_count": 1779
} | 2,108 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.