text stringlengths 3 1.68M | id stringlengths 13 169 | metadata dict | __index_level_0__ int64 0 2.21k |
|---|---|---|---|
python_sources()
| llama_index/llama-index-legacy/llama_index/legacy/indices/multi_modal/BUILD/0 | {
"file_path": "llama_index/llama-index-legacy/llama_index/legacy/indices/multi_modal/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,574 |
import { BaseChain } from "./base.js";
import { loadFromHub } from "../util/hub.js";
import { FileLoader, LoadValues, loadFromFile } from "../util/load.js";
import { parseFileConfig } from "../util/parse.js";
const loadChainFromFile: FileLoader<BaseChain> = async (
file: string,
path: string,
values: LoadValues ... | langchainjs/langchain/src/chains/load.ts/0 | {
"file_path": "langchainjs/langchain/src/chains/load.ts",
"repo_id": "langchainjs",
"token_count": 475
} | 890 |
"""Methods for creating function specs in the style of OpenAI Functions"""
from __future__ import annotations
import inspect
from typing import (
TYPE_CHECKING,
Any,
Callable,
Dict,
List,
Literal,
Optional,
Tuple,
Type,
Union,
cast,
)
from typing_extensions import TypedDict... | langchain/libs/core/langchain_core/utils/function_calling.py/0 | {
"file_path": "langchain/libs/core/langchain_core/utils/function_calling.py",
"repo_id": "langchain",
"token_count": 4230
} | 401 |
<!--Copyright 2020 de The HuggingFace Team. Todos los derechos reservados
Con licencia bajo la Licencia Apache, Versión 2.0 (la "Licencia"); No puedes usar este archivo excepto de conformidad con la Licencia.
Puedes obtener una copia de la Licencia en
http://www.apache.org/licenses/LICENSE-2.0
Al menos que sea requr... | transformers/docs/source/es/philosophy.md/0 | {
"file_path": "transformers/docs/source/es/philosophy.md",
"repo_id": "transformers",
"token_count": 1964
} | 534 |
FROM golang:1.20.4-alpine3.17
RUN apk add --no-cache make bash | milvus/build/docker/meta-migration/builder/Dockerfile/0 | {
"file_path": "milvus/build/docker/meta-migration/builder/Dockerfile",
"repo_id": "milvus",
"token_count": 28
} | 1,611 |
from abc import abstractmethod
from typing import Any, List, Sequence, Union
from llama_index.core.base.query_pipeline.query import (
ChainableMixin,
QueryComponent,
)
from llama_index.core.bridge.pydantic import BaseModel
from llama_index.core.prompts.mixin import PromptMixin, PromptMixinType
from llama_index... | llama_index/llama-index-core/llama_index/core/selectors/types.py/0 | {
"file_path": "llama_index/llama-index-core/llama_index/core/selectors/types.py",
"repo_id": "llama_index",
"token_count": 1303
} | 1,247 |
# coding=utf-8
# Copyright 2019-present, the HuggingFace Inc. team and Facebook, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Un... | transformers/examples/research_projects/distillation/lm_seqs_dataset.py/0 | {
"file_path": "transformers/examples/research_projects/distillation/lm_seqs_dataset.py",
"repo_id": "transformers",
"token_count": 2821
} | 523 |
"""RAKE keyword-table based index.
Similar to KeywordTableIndex, but uses RAKE instead of GPT.
"""
from typing import Any, Set, Union
from llama_index.core.base.base_retriever import BaseRetriever
from llama_index.core.indices.keyword_table.base import (
BaseKeywordTableIndex,
KeywordTableRetrieverMode,
)
f... | llama_index/llama-index-core/llama_index/core/indices/keyword_table/rake_base.py/0 | {
"file_path": "llama_index/llama-index-core/llama_index/core/indices/keyword_table/rake_base.py",
"repo_id": "llama_index",
"token_count": 417
} | 1,178 |
# coding=utf-8
# Copyright 2020-present the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by ap... | transformers/src/transformers/trainer_pt_utils.py/0 | {
"file_path": "transformers/src/transformers/trainer_pt_utils.py",
"repo_id": "transformers",
"token_count": 21037
} | 706 |
"""DashScope api utils."""
from http import HTTPStatus
from typing import Any, Dict, List, Sequence
from llama_index.core.base.llms.types import (
ChatMessage,
ChatResponse,
CompletionResponse,
)
from llama_index.core.schema import ImageDocument
def dashscope_response_to_completion_response(response: An... | llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-dashscope/llama_index/multi_modal_llms/dashscope/utils.py/0 | {
"file_path": "llama_index/llama-index-integrations/multi_modal_llms/llama-index-multi-modal-llms-dashscope/llama_index/multi_modal_llms/dashscope/utils.py",
"repo_id": "llama_index",
"token_count": 1011
} | 1,340 |
"""Prompts for ChatGPT."""
from llama_index.core.base.llms.types import ChatMessage, MessageRole
from llama_index.core.prompts.base import ChatPromptTemplate
# text qa prompt
TEXT_QA_SYSTEM_PROMPT = ChatMessage(
content=(
"You are an expert Q&A system that is trusted around the world.\n"
"Always a... | llama_index/llama-index-core/llama_index/core/prompts/chat_prompts.py/0 | {
"file_path": "llama_index/llama-index-core/llama_index/core/prompts/chat_prompts.py",
"repo_id": "llama_index",
"token_count": 1579
} | 1,140 |
<jupyter_start><jupyter_text>Petals`Petals` runs 100B+ language models at home, BitTorrent-style.This notebook goes over how to use Langchain with [Petals](https://github.com/bigscience-workshop/petals). Install petalsThe `petals` package is required to use the Petals API. Install `petals` using `pip3 install petals`.... | langchain/docs/docs/integrations/llms/petals.ipynb/0 | {
"file_path": "langchain/docs/docs/integrations/llms/petals.ipynb",
"repo_id": "langchain",
"token_count": 721
} | 118 |
<!--Copyright 2020 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | transformers/docs/source/ja/model_doc/cpm.md/0 | {
"file_path": "transformers/docs/source/ja/model_doc/cpm.md",
"repo_id": "transformers",
"token_count": 1254
} | 517 |
/* eslint-disable no-process-env */
/* eslint-disable no-promise-executor-return */
import { test, expect } from "@jest/globals";
import { MongoClient } from "mongodb";
import { setTimeout } from "timers/promises";
import { OpenAIEmbeddings } from "@langchain/openai";
import { Document } from "@langchain/core/document... | langchainjs/libs/langchain-community/src/vectorstores/tests/mongodb_atlas.int.test.ts/0 | {
"file_path": "langchainjs/libs/langchain-community/src/vectorstores/tests/mongodb_atlas.int.test.ts",
"repo_id": "langchainjs",
"token_count": 2654
} | 999 |
# coding=utf-8
# Copyright 2021 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | transformers/src/transformers/dynamic_module_utils.py/0 | {
"file_path": "transformers/src/transformers/dynamic_module_utils.py",
"repo_id": "transformers",
"token_count": 10973
} | 553 |
"""**Memory** maintains Chain state, incorporating context from past runs.
**Class hierarchy for Memory:**
.. code-block::
BaseMemory --> BaseChatMemory --> <name>Memory # Examples: ZepMemory, MotorheadMemory
**Main helpers:**
.. code-block::
BaseChatMessageHistory
**Chat Message History** stores the ch... | langchain/libs/langchain/langchain/memory/__init__.py/0 | {
"file_path": "langchain/libs/langchain/langchain/memory/__init__.py",
"repo_id": "langchain",
"token_count": 1000
} | 520 |
# Copyright 2024 DiffEdit Authors and Pix2Pix Zero Authors and The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/license... | diffusers/src/diffusers/pipelines/stable_diffusion_diffedit/pipeline_stable_diffusion_diffedit.py/0 | {
"file_path": "diffusers/src/diffusers/pipelines/stable_diffusion_diffedit/pipeline_stable_diffusion_diffedit.py",
"repo_id": "diffusers",
"token_count": 34828
} | 233 |
import { NodeHandler, ASTParser } from "./base.js";
import { CallExpressionType, MemberExpressionType } from "./types.js";
/**
* Handles call expressions in the AST parsed by the `ASTParser`. This
* class is part of the LangChain Expression Language (LCEL), a
* declarative way to compose chains together in LangChai... | langchainjs/langchain/src/output_parsers/expression_type_handlers/call_expression_handler.ts/0 | {
"file_path": "langchainjs/langchain/src/output_parsers/expression_type_handlers/call_expression_handler.ts",
"repo_id": "langchainjs",
"token_count": 1007
} | 921 |
from typing import Any, Dict, List, Tuple
import numpy as np
class MockTxtaiIndex:
"""Mock txtai index."""
def __init__(self, *args: Any, **kwargs: Any) -> None:
"""Initialize params."""
self._index: Dict[int, np.ndarray] = {}
self.backend = None
def count(self) -> int:
... | llama_index/llama-index-legacy/tests/indices/vector_store/mock_txtai.py/0 | {
"file_path": "llama_index/llama-index-legacy/tests/indices/vector_store/mock_txtai.py",
"repo_id": "llama_index",
"token_count": 578
} | 1,741 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/core/src/index/ScalarIndex.h/0 | {
"file_path": "milvus/internal/core/src/index/ScalarIndex.h",
"repo_id": "milvus",
"token_count": 858
} | 1,927 |
<!DOCTYPE html>
<html lang="en">
<head>
<meta charset="utf-8" />
<title>Welcome to Candle!</title>
<link data-trunk rel="copy-file" href="mel_filters.safetensors" />
<!-- samples -->
<link data-trunk rel="copy-dir" href="audios" />
<!-- tiny.en -->
<link data-trunk rel="copy-dir" href="whi... | candle/candle-wasm-examples/whisper/index.html/0 | {
"file_path": "candle/candle-wasm-examples/whisper/index.html",
"repo_id": "candle",
"token_count": 523
} | 78 |
<jupyter_start><jupyter_text>Memory in the Multi-Input ChainMost memory objects assume a single input. In this notebook, we go over how to add memory to a chain that has multiple inputs. We will add memory to a question/answering chain. This chain takes as inputs both related documents and a user question.<jupyter_code... | langchain/docs/docs/modules/memory/adding_memory_chain_multiple_inputs.ipynb/0 | {
"file_path": "langchain/docs/docs/modules/memory/adding_memory_chain_multiple_inputs.ipynb",
"repo_id": "langchain",
"token_count": 617
} | 194 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/rootcoord/list_db_task_test.go/0 | {
"file_path": "milvus/internal/rootcoord/list_db_task_test.go",
"repo_id": "milvus",
"token_count": 955
} | 1,863 |
poetry_requirements(
name="poetry",
)
| llama_index/llama-index-integrations/program/llama-index-program-guidance/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/program/llama-index-program-guidance/BUILD",
"repo_id": "llama_index",
"token_count": 18
} | 1,300 |
from llama_index.packs.infer_retrieve_rerank.base import InferRetrieveRerankPack
__all__ = ["InferRetrieveRerankPack"]
| llama_index/llama-index-packs/llama-index-packs-infer-retrieve-rerank/llama_index/packs/infer_retrieve_rerank/__init__.py/0 | {
"file_path": "llama_index/llama-index-packs/llama-index-packs-infer-retrieve-rerank/llama_index/packs/infer_retrieve_rerank/__init__.py",
"repo_id": "llama_index",
"token_count": 44
} | 1,707 |
from __future__ import annotations
import logging
import os
import warnings
from typing import (
Any,
Dict,
Iterable,
List,
Literal,
Mapping,
Optional,
Sequence,
Set,
Tuple,
Union,
cast,
)
import numpy as np
import openai
import tiktoken
from langchain_core.embeddings i... | langchain/libs/partners/openai/langchain_openai/embeddings/base.py/0 | {
"file_path": "langchain/libs/partners/openai/langchain_openai/embeddings/base.py",
"repo_id": "langchain",
"token_count": 9604
} | 693 |
// Copyright (C) 2019-2020 Zilliz. All rights reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable l... | milvus/pkg/util/metricsinfo/metrics_info_test.go/0 | {
"file_path": "milvus/pkg/util/metricsinfo/metrics_info_test.go",
"repo_id": "milvus",
"token_count": 4255
} | 2,062 |
python_tests(
interpreter_constraints=["==3.9.*", "==3.10.*"],
)
| llama_index/llama-index-packs/llama-index-packs-panel-chatbot/tests/BUILD/0 | {
"file_path": "llama_index/llama-index-packs/llama-index-packs-panel-chatbot/tests/BUILD",
"repo_id": "llama_index",
"token_count": 29
} | 1,716 |
"""Test finetuning engine."""
import pkgutil
import pytest
def test_torch_imports() -> None:
"""Test that torch is an optional dependency."""
# importing fine-tuning modules should be ok
from llama_index.finetuning import EmbeddingAdapterFinetuneEngine # noqa
from llama_index.finetuning import OpenA... | llama_index/llama-index-finetuning/tests/test_base.py/0 | {
"file_path": "llama_index/llama-index-finetuning/tests/test_base.py",
"repo_id": "llama_index",
"token_count": 271
} | 1,286 |
from langchain_core.agents import AgentAction, AgentFinish
from langchain.agents.output_parsers.xml import XMLAgentOutputParser
def test_tool_usage() -> None:
parser = XMLAgentOutputParser()
# Test when final closing </tool_input> is included
_input = """<tool>search</tool><tool_input>foo</tool_input>"""... | langchain/libs/langchain/tests/unit_tests/agents/output_parsers/test_xml.py/0 | {
"file_path": "langchain/libs/langchain/tests/unit_tests/agents/output_parsers/test_xml.py",
"repo_id": "langchain",
"token_count": 459
} | 630 |
// Copyright (C) 2019-2020 Zilliz. All rights reserved.
//
// Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance
// with the License. You may obtain a copy of the License at
//
// http://www.apache.org/licenses/LICENSE-2.0
//
// Unless required by applicable l... | milvus/pkg/util/metricsinfo/topology_test.go/0 | {
"file_path": "milvus/pkg/util/metricsinfo/topology_test.go",
"repo_id": "milvus",
"token_count": 4381
} | 1,920 |
{
"input_variables": [
"question",
"student_answer"
],
"output_parser": {
"regex": "(.*?)\nScore: (.*)",
"output_keys": [
"answer",
"score"
],
"default_output_key": null,
"_type": "regex_parser"
},
"partial_variables": {... | langchain/libs/core/tests/unit_tests/examples/prompt_with_output_parser.json/0 | {
"file_path": "langchain/libs/core/tests/unit_tests/examples/prompt_with_output_parser.json",
"repo_id": "langchain",
"token_count": 252
} | 414 |
from __future__ import annotations
from typing import Any, Dict, List
from langchain_core.output_parsers import BaseOutputParser
from langchain_core.pydantic_v1 import root_validator
class CombiningOutputParser(BaseOutputParser):
"""Combine multiple output parsers into one."""
parsers: List[BaseOutputParse... | langchain/libs/langchain/langchain/output_parsers/combining.py/0 | {
"file_path": "langchain/libs/langchain/langchain/output_parsers/combining.py",
"repo_id": "langchain",
"token_count": 733
} | 573 |
// @ts-nocheck
// Inlined because of ESM import issues
/*!
* https://github.com/Starcounter-Jack/JSON-Patch
* (c) 2013-2021 Joachim Wester
* MIT license
*/
import {
_deepClone,
_objectKeys,
escapePathComponent,
hasOwnProperty,
} from "./helpers.js";
import { applyPatch, Operation } from "./core.js";
expo... | langchainjs/langchain-core/src/utils/fast-json-patch/src/duplex.ts/0 | {
"file_path": "langchainjs/langchain-core/src/utils/fast-json-patch/src/duplex.ts",
"repo_id": "langchainjs",
"token_count": 2545
} | 874 |
/* eslint-disable no-promise-executor-return */
/* eslint-disable no-process-env */
/* eslint-disable @typescript-eslint/no-non-null-assertion */
import { test, expect, describe } from "@jest/globals";
import { HumanMessage, AIMessage } from "@langchain/core/messages";
import { UpstashRedisChatMessageHistory } from "... | langchainjs/libs/langchain-community/src/stores/tests/redis_upstash.int.test.ts/0 | {
"file_path": "langchainjs/libs/langchain-community/src/stores/tests/redis_upstash.int.test.ts",
"repo_id": "langchainjs",
"token_count": 936
} | 994 |
"""Firestore Chat Message History."""
from __future__ import annotations
import logging
from typing import TYPE_CHECKING, List, Optional
from langchain_core.chat_history import BaseChatMessageHistory
from langchain_core.messages import (
BaseMessage,
messages_from_dict,
messages_to_dict,
)
logger = loggi... | langchain/libs/community/langchain_community/chat_message_histories/firestore.py/0 | {
"file_path": "langchain/libs/community/langchain_community/chat_message_histories/firestore.py",
"repo_id": "langchain",
"token_count": 1356
} | 240 |
poetry_requirements(
name="poetry",
)
python_requirements(
name="reqs",
)
| llama_index/llama-index-packs/llama-index-packs-deeplake-deepmemory-retriever/BUILD/0 | {
"file_path": "llama_index/llama-index-packs/llama-index-packs-deeplake-deepmemory-retriever/BUILD",
"repo_id": "llama_index",
"token_count": 36
} | 1,838 |
"""Test the ReadFile tool."""
from pathlib import Path
from tempfile import TemporaryDirectory
from langchain_community.tools.file_management.read import ReadFileTool
def test_read_file_with_root_dir() -> None:
"""Test the ReadFile tool."""
with TemporaryDirectory() as temp_dir:
with (Path(temp_dir)... | langchain/libs/community/tests/unit_tests/tools/file_management/test_read.py/0 | {
"file_path": "langchain/libs/community/tests/unit_tests/tools/file_management/test_read.py",
"repo_id": "langchain",
"token_count": 404
} | 424 |
use super::{super::OrderedVocabIter, convert_merges_to_hashmap, BpeBuilder, Pair, BPE};
use serde::{
de::{Error, MapAccess, Visitor},
ser::SerializeStruct,
Deserialize, Deserializer, Serialize, Serializer,
};
use std::collections::HashMap;
impl Serialize for BPE {
fn serialize<S>(&self, serializer: S) ... | tokenizers/tokenizers/src/models/bpe/serialization.rs/0 | {
"file_path": "tokenizers/tokenizers/src/models/bpe/serialization.rs",
"repo_id": "tokenizers",
"token_count": 2739
} | 457 |
from typing import Iterator, List
from langchain_core.documents import Document
from langchain_community.document_loaders.base import BaseLoader
class AirtableLoader(BaseLoader):
"""Load the `Airtable` tables."""
def __init__(self, api_token: str, table_id: str, base_id: str):
"""Initialize with AP... | langchain/libs/community/langchain_community/document_loaders/airtable.py/0 | {
"file_path": "langchain/libs/community/langchain_community/document_loaders/airtable.py",
"repo_id": "langchain",
"token_count": 573
} | 230 |
<!--Copyright 2023 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | transformers/docs/source/en/tflite.md/0 | {
"file_path": "transformers/docs/source/en/tflite.md",
"repo_id": "transformers",
"token_count": 871
} | 480 |
"""written under MIT Licence, Michael Feil 2023."""
import asyncio
from concurrent.futures import ThreadPoolExecutor
from typing import Any, Callable, Dict, List, Optional, Tuple
import aiohttp
import numpy as np
import requests
from langchain_core.embeddings import Embeddings
from langchain_core.pydantic_v1 import B... | langchain/libs/community/langchain_community/embeddings/infinity.py/0 | {
"file_path": "langchain/libs/community/langchain_community/embeddings/infinity.py",
"repo_id": "langchain",
"token_count": 4778
} | 270 |
<html>
<head>
<meta content="text/html;charset=utf-8" http-equiv="Content-Type" />
<title>Candle Phi 1.5 / Phi 2.0 Rust/WASM</title>
</head>
<body></body>
</html>
<!DOCTYPE html>
<html>
<head>
<meta charset="UTF-8" />
<meta name="viewport" content="width=device-width, initial-scale=1.0" />
... | candle/candle-wasm-examples/phi/index.html/0 | {
"file_path": "candle/candle-wasm-examples/phi/index.html",
"repo_id": "candle",
"token_count": 9818
} | 75 |
package planparserv2
import (
"encoding/json"
"go.uber.org/zap"
"github.com/milvus-io/milvus/internal/proto/planpb"
"github.com/milvus-io/milvus/pkg/log"
)
type ShowExprVisitor struct{}
func extractColumnInfo(info *planpb.ColumnInfo) interface{} {
js := make(map[string]interface{})
js["field_id"] = info.GetF... | milvus/internal/parser/planparserv2/show_visitor.go/0 | {
"file_path": "milvus/internal/parser/planparserv2/show_visitor.go",
"repo_id": "milvus",
"token_count": 2163
} | 1,736 |
# coding=utf-8
# Copyright 2021 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | transformers/src/transformers/models/speech_to_text_2/processing_speech_to_text_2.py/0 | {
"file_path": "transformers/src/transformers/models/speech_to_text_2/processing_speech_to_text_2.py",
"repo_id": "transformers",
"token_count": 1789
} | 661 |
<jupyter_start><jupyter_text>Run TemplateIn `server.py`, set -```add_routes(app, chain, path="/sql_llama2")```This template includes an example DB of 2023 NBA rosters.We can ask questions related to NBA players.<jupyter_code>from langserve.client import RemoteRunnable
sql_app = RemoteRunnable("http://0.0.0.0:8001/sql... | langchain/templates/sql-llama2/sql_llama2.ipynb/0 | {
"file_path": "langchain/templates/sql-llama2/sql_llama2.ipynb",
"repo_id": "langchain",
"token_count": 149
} | 690 |
<jupyter_start><jupyter_text>Text-to-SQL Guide (Query Engine + Retriever)This is a basic guide to LlamaIndex's Text-to-SQL capabilities. 1. We first show how to perform text-to-SQL over a toy dataset: this will do "retrieval" (sql query over db) and "synthesis".2. We then show how to buid a TableIndex over the schema t... | llama_index/docs/examples/index_structs/struct_indices/SQLIndexDemo.ipynb/0 | {
"file_path": "llama_index/docs/examples/index_structs/struct_indices/SQLIndexDemo.ipynb",
"repo_id": "llama_index",
"token_count": 2673
} | 1,084 |
import { ChatOpenAI, OpenAIEmbeddings } from "@langchain/openai";
import { MemoryVectorStore } from "langchain/vectorstores/memory";
import { SerpAPILoader } from "langchain/document_loaders/web/serpapi";
import { ChatPromptTemplate } from "@langchain/core/prompts";
import { createStuffDocumentsChain } from "langchain/... | langchainjs/examples/src/document_loaders/serpapi.ts/0 | {
"file_path": "langchainjs/examples/src/document_loaders/serpapi.ts",
"repo_id": "langchainjs",
"token_count": 435
} | 768 |
# PromptLayer
>[PromptLayer](https://docs.promptlayer.com/introduction) is a platform for prompt engineering.
> It also helps with the LLM observability to visualize requests, version prompts, and track usage.
>
>While `PromptLayer` does have LLMs that integrate directly with LangChain (e.g.
> [`PromptLayerOpenAI`](... | langchain/docs/docs/integrations/providers/promptlayer.mdx/0 | {
"file_path": "langchain/docs/docs/integrations/providers/promptlayer.mdx",
"repo_id": "langchain",
"token_count": 368
} | 157 |
# coding=utf-8
# Copyright 2019-present, the HuggingFace Inc. team, The Google AI Language Team and Facebook, Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.or... | transformers/src/transformers/models/deprecated/retribert/modeling_retribert.py/0 | {
"file_path": "transformers/src/transformers/models/deprecated/retribert/modeling_retribert.py",
"repo_id": "transformers",
"token_count": 3872
} | 630 |
{
"openapi": "3.1.0",
"info": {
"title": "Robocorp Actions Server",
"description": "Robocorp Actions Server",
"version": "0.1.0"
},
"paths": {
"/api/actions/test/test-action/run": {
"post": {
"summary": "Test Action",
"description": "Greets an employee",
"operationI... | langchain/libs/partners/robocorp/tests/unit_tests/_openapi.fixture.json/0 | {
"file_path": "langchain/libs/partners/robocorp/tests/unit_tests/_openapi.fixture.json",
"repo_id": "langchain",
"token_count": 6984
} | 642 |
<jupyter_start><jupyter_text>If you're opening this Notebook on colab, you will probably need to install 🤗 Transformers as well as some other libraries. Uncomment the following cell and run it.<jupyter_code># Install
!pip install -q biopython transformers datasets huggingface_hub accelerate<jupyter_output><empty_outpu... | notebooks/examples/nucleotide_transformer_dna_sequence_modelling.ipynb/0 | {
"file_path": "notebooks/examples/nucleotide_transformer_dna_sequence_modelling.ipynb",
"repo_id": "notebooks",
"token_count": 6637
} | 320 |
import math
import torch
from torch.optim.optimizer import Optimizer
class Nadam(Optimizer):
"""Implements Nadam algorithm (a variant of Adam based on Nesterov momentum).
It has been proposed in `Incorporating Nesterov Momentum into Adam`__.
Arguments:
params (iterable): iterable of parameters ... | pytorch-image-models/timm/optim/nadam.py/0 | {
"file_path": "pytorch-image-models/timm/optim/nadam.py",
"repo_id": "pytorch-image-models",
"token_count": 1921
} | 372 |
python_sources()
| llama_index/llama-index-integrations/storage/index_store/llama-index-storage-index-store-dynamodb/llama_index/storage/index_store/dynamodb/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/storage/index_store/llama-index-storage-index-store-dynamodb/llama_index/storage/index_store/dynamodb/BUILD",
"repo_id": "llama_index",
"token_count": 6
} | 1,404 |
""" Linear layer (alternate definition)
"""
import torch
import torch.nn.functional as F
from torch import nn as nn
class Linear(nn.Linear):
r"""Applies a linear transformation to the incoming data: :math:`y = xA^T + b`
Wraps torch.nn.Linear to support AMP + torchscript usage by manually casting
weight &... | pytorch-image-models/timm/layers/linear.py/0 | {
"file_path": "pytorch-image-models/timm/layers/linear.py",
"repo_id": "pytorch-image-models",
"token_count": 282
} | 368 |
import json
import logging
from datetime import datetime
from typing import List, Optional
from langchain_core.chat_history import BaseChatMessageHistory
from langchain_core.messages import BaseMessage, message_to_dict, messages_from_dict
from sqlalchemy import create_engine, text
from sqlalchemy.exc import SQLAlchemy... | langchain/libs/community/langchain_community/chat_message_histories/tidb.py/0 | {
"file_path": "langchain/libs/community/langchain_community/chat_message_histories/tidb.py",
"repo_id": "langchain",
"token_count": 2341
} | 233 |
"""Base interfaces for tracing runs."""
from langchain_core.tracers.base import BaseTracer, TracerException
__all__ = ["BaseTracer", "TracerException"]
| langchain/libs/langchain/langchain/callbacks/tracers/base.py/0 | {
"file_path": "langchain/libs/langchain/langchain/callbacks/tracers/base.py",
"repo_id": "langchain",
"token_count": 45
} | 454 |
from langchain_community.document_loaders.recursive_url_loader import RecursiveUrlLoader
def test_async_recursive_url_loader() -> None:
url = "https://docs.python.org/3.9/"
loader = RecursiveUrlLoader(
url,
extractor=lambda _: "placeholder",
use_async=True,
max_depth=3,
... | langchain/libs/community/tests/integration_tests/document_loaders/test_recursive_url_loader.py/0 | {
"file_path": "langchain/libs/community/tests/integration_tests/document_loaders/test_recursive_url_loader.py",
"repo_id": "langchain",
"token_count": 752
} | 344 |
# coding=utf-8
# Copyright 2020 The Facebook AI Research Team Authors and The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LIC... | transformers/src/transformers/models/mbart/tokenization_mbart_fast.py/0 | {
"file_path": "transformers/src/transformers/models/mbart/tokenization_mbart_fast.py",
"repo_id": "transformers",
"token_count": 5169
} | 656 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/pkg/mq/msgstream/mqwrapper/pulsar/pulsar_id.go/0 | {
"file_path": "milvus/pkg/mq/msgstream/mqwrapper/pulsar/pulsar_id.go",
"repo_id": "milvus",
"token_count": 943
} | 2,097 |
import { PromptTemplate } from "@langchain/core/prompts";
const _DEFAULT_SUMMARIZER_TEMPLATE = `Progressively summarize the lines of conversation provided, adding onto the previous summary returning a new summary.
EXAMPLE
Current summary:
The human asks what the AI thinks of artificial intelligence. The AI thinks art... | langchainjs/langchain/src/memory/prompt.ts/0 | {
"file_path": "langchainjs/langchain/src/memory/prompt.ts",
"repo_id": "langchainjs",
"token_count": 1550
} | 906 |
<!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | transformers/docs/source/ja/multilingual.md/0 | {
"file_path": "transformers/docs/source/ja/multilingual.md",
"repo_id": "transformers",
"token_count": 4086
} | 498 |
"""Wrapper functions around an LLM chain."""
import logging
from abc import ABC, abstractmethod
from collections import ChainMap
from typing import Any, Dict, List, Optional, Union
from llama_index.core.base.llms.types import (
ChatMessage,
LLMMetadata,
MessageRole,
)
from llama_index.core.bridge.pydantic... | llama_index/llama-index-core/llama_index/core/service_context_elements/llm_predictor.py/0 | {
"file_path": "llama_index/llama-index-core/llama_index/core/service_context_elements/llm_predictor.py",
"repo_id": "llama_index",
"token_count": 5147
} | 1,141 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/internal/querycoordv2/balance/score_based_balancer_test.go/0 | {
"file_path": "milvus/internal/querycoordv2/balance/score_based_balancer_test.go",
"repo_id": "milvus",
"token_count": 10808
} | 1,972 |
#[cfg(feature = "mkl")]
extern crate intel_mkl_src;
#[cfg(feature = "accelerate")]
extern crate accelerate_src;
use candle_transformers::models::jina_bert::{BertModel, Config};
use anyhow::Error as E;
use candle::{DType, Module, Tensor};
use candle_nn::VarBuilder;
use clap::Parser;
#[derive(Parser, Debug)]
#[comman... | candle/candle-examples/examples/jina-bert/main.rs/0 | {
"file_path": "candle/candle-examples/examples/jina-bert/main.rs",
"repo_id": "candle",
"token_count": 3088
} | 49 |
- name: querynode
docker_container:
name: querynode
image: "{{image}}"
command: "milvus run querynode"
env:
ETCD_ENDPOINTS: "{{ETCD_ENDPOINTS}}"
MINIO_ADDRESS: "{{MINIO_ADDRESS}}"
PULSAR_ADDRESS: "{{PULSAR_ADDRESS}}"
METRICS_PORT: "9091"
network_mode: "{{nodes_network}}" #U... | milvus/deployments/docker/cluster-distributed-deployment/roles/deploy-querynode/tasks/main.yml/0 | {
"file_path": "milvus/deployments/docker/cluster-distributed-deployment/roles/deploy-querynode/tasks/main.yml",
"repo_id": "milvus",
"token_count": 170
} | 1,718 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/tests/integration/util_insert.go/0 | {
"file_path": "milvus/tests/integration/util_insert.go",
"repo_id": "milvus",
"token_count": 2254
} | 1,982 |
from langchain_experimental.autonomous_agents.autogpt.agent import AutoGPT
from langchain_experimental.autonomous_agents.baby_agi.baby_agi import BabyAGI
from langchain_experimental.autonomous_agents.hugginggpt.hugginggpt import HuggingGPT
__all__ = ["BabyAGI", "AutoGPT", "HuggingGPT"]
| langchain/libs/experimental/langchain_experimental/autonomous_agents/__init__.py/0 | {
"file_path": "langchain/libs/experimental/langchain_experimental/autonomous_agents/__init__.py",
"repo_id": "langchain",
"token_count": 97
} | 425 |
# Metric Card for Pearson Correlation Coefficient (pearsonr)
## Metric Description
Pearson correlation coefficient and p-value for testing non-correlation.
The Pearson correlation coefficient measures the linear relationship between two datasets. The calculation of the p-value relies on the assumption that each data... | datasets/metrics/pearsonr/README.md/0 | {
"file_path": "datasets/metrics/pearsonr/README.md",
"repo_id": "datasets",
"token_count": 1387
} | 142 |
from __future__ import annotations
import json
import logging
import re
from typing import Optional, Union
from langchain_core.agents import AgentAction, AgentFinish
from langchain_core.exceptions import OutputParserException
from langchain_core.language_models import BaseLanguageModel
from langchain_core.pydantic_v1... | langchain/libs/langchain/langchain/agents/structured_chat/output_parser.py/0 | {
"file_path": "langchain/libs/langchain/langchain/agents/structured_chat/output_parser.py",
"repo_id": "langchain",
"token_count": 1527
} | 449 |
[package]
name = "candle-wasm-tests"
version.workspace = true
edition.workspace = true
description = "WASM tests for candle"
keywords.workspace = true
categories.workspace = true
[dependencies]
candle = { workspace = true }
rand = { workspace = true }
getrandom = { version = "0.2", features = ["js"] }
[dev-dependenci... | candle/candle-wasm-tests/Cargo.toml/0 | {
"file_path": "candle/candle-wasm-tests/Cargo.toml",
"repo_id": "candle",
"token_count": 122
} | 89 |
package datacoord
import (
"fmt"
"sync"
"github.com/samber/lo"
"go.uber.org/atomic"
"go.uber.org/zap"
"github.com/milvus-io/milvus/internal/proto/datapb"
"github.com/milvus-io/milvus/pkg/log"
"github.com/milvus-io/milvus/pkg/metrics"
"github.com/milvus-io/milvus/pkg/util/typeutil"
)
type Scheduler interfac... | milvus/internal/datacoord/compaction_scheduler.go/0 | {
"file_path": "milvus/internal/datacoord/compaction_scheduler.go",
"repo_id": "milvus",
"token_count": 2390
} | 1,763 |
import { LangChainTracer } from "./tracer_langchain.js";
import { LangChainTracerV1 } from "./tracer_langchain_v1.js";
/**
* @deprecated Use the V2 handler instead.
*
* Function that returns an instance of `LangChainTracerV1`. If a session
* is provided, it loads that session into the tracer; otherwise, it loads
... | langchainjs/langchain-core/src/tracers/initialize.ts/0 | {
"file_path": "langchainjs/langchain-core/src/tracers/initialize.ts",
"repo_id": "langchainjs",
"token_count": 304
} | 834 |
# coding=utf-8
# Copyright 2022 Microsoft Research and The HuggingFace Inc. team.
# All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses... | transformers/src/transformers/models/git/modeling_git.py/0 | {
"file_path": "transformers/src/transformers/models/git/modeling_git.py",
"repo_id": "transformers",
"token_count": 29391
} | 608 |
package indexnode
import (
"context"
"fmt"
"sync"
"testing"
"time"
"github.com/stretchr/testify/assert"
"github.com/milvus-io/milvus-proto/go-api/v2/commonpb"
"github.com/milvus-io/milvus/pkg/util/paramtable"
)
type fakeTaskState int
const (
fakeTaskInited = iota
fakeTaskEnqueued
fakeTaskPrepared
fakeT... | milvus/internal/indexnode/task_scheduler_test.go/0 | {
"file_path": "milvus/internal/indexnode/task_scheduler_test.go",
"repo_id": "milvus",
"token_count": 1997
} | 1,804 |
import asyncio
from typing import Any, Optional, Sequence
from llama_index.legacy.async_utils import run_async_tasks
from llama_index.legacy.prompts import BasePromptTemplate
from llama_index.legacy.prompts.default_prompt_selectors import (
DEFAULT_TREE_SUMMARIZE_PROMPT_SEL,
)
from llama_index.legacy.prompts.mixin... | llama_index/llama-index-legacy/llama_index/legacy/response_synthesizers/tree_summarize.py/0 | {
"file_path": "llama_index/llama-index-legacy/llama_index/legacy/response_synthesizers/tree_summarize.py",
"repo_id": "llama_index",
"token_count": 4614
} | 1,770 |
package datacoord
import (
"fmt"
"testing"
"github.com/samber/lo"
"github.com/stretchr/testify/suite"
"github.com/milvus-io/milvus/internal/proto/datapb"
"github.com/milvus-io/milvus/pkg/metrics"
"github.com/milvus-io/milvus/pkg/util/testutils"
)
func TestSchedulerSuite(t *testing.T) {
suite.Run(t, new(Sche... | milvus/internal/datacoord/compaction_scheduler_test.go/0 | {
"file_path": "milvus/internal/datacoord/compaction_scheduler_test.go",
"repo_id": "milvus",
"token_count": 3335
} | 1,690 |
import { HumanMessage, AIMessage } from "@langchain/core/messages";
import { ChatMessageHistory } from "langchain/stores/message/in_memory";
const history = new ChatMessageHistory();
await history.addMessage(new HumanMessage("hi"));
await history.addMessage(new AIMessage("what is up?"));
console.log(await history.g... | langchainjs/examples/src/memory/chat_messages.ts/0 | {
"file_path": "langchainjs/examples/src/memory/chat_messages.ts",
"repo_id": "langchainjs",
"token_count": 184
} | 814 |
# Text Generation
The Hugging Face Text Generation Python library provides a convenient way of interfacing with a
`text-generation-inference` instance running on
[Hugging Face Inference Endpoints](https://huggingface.co/inference-endpoints) or on the Hugging Face Hub.
## Get Started
### Install
```shell
pip install... | text-generation-inference/clients/python/README.md/0 | {
"file_path": "text-generation-inference/clients/python/README.md",
"repo_id": "text-generation-inference",
"token_count": 2195
} | 369 |
# LlamaIndex Embeddings Integration: Sagemaker Endpoint
| llama_index/llama-index-integrations/embeddings/llama-index-embeddings-sagemaker-endpoint/README.md/0 | {
"file_path": "llama_index/llama-index-integrations/embeddings/llama-index-embeddings-sagemaker-endpoint/README.md",
"repo_id": "llama_index",
"token_count": 14
} | 1,238 |
# Live 1: How the course work, Q&A, and playing with Huggy
In this first live stream, we explained how the course work (scope, units, challenges, and more) and answered your questions.
And finally, we saw some LunarLander agents you've trained and play with your Huggies 🐶
<Youtube id="JeJIswxyrsM" />
To know when ... | deep-rl-class/units/en/live1/live1.mdx/0 | {
"file_path": "deep-rl-class/units/en/live1/live1.mdx",
"repo_id": "deep-rl-class",
"token_count": 131
} | 149 |
# coding=utf-8
# Copyright 2019-present, Facebook, Inc and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Un... | transformers/src/transformers/models/xlm/modeling_tf_xlm.py/0 | {
"file_path": "transformers/src/transformers/models/xlm/modeling_tf_xlm.py",
"repo_id": "transformers",
"token_count": 25192
} | 678 |
#!/usr/bin/env python
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless... | transformers/examples/legacy/seq2seq/pack_dataset.py/0 | {
"file_path": "transformers/examples/legacy/seq2seq/pack_dataset.py",
"repo_id": "transformers",
"token_count": 1358
} | 544 |
package planparserv2
import (
"fmt"
"strconv"
"strings"
"github.com/antlr/antlr4/runtime/Go/antlr"
"github.com/milvus-io/milvus-proto/go-api/v2/schemapb"
parser "github.com/milvus-io/milvus/internal/parser/planparserv2/generated"
"github.com/milvus-io/milvus/internal/proto/planpb"
"github.com/milvus-io/milvu... | milvus/internal/parser/planparserv2/parser_visitor.go/0 | {
"file_path": "milvus/internal/parser/planparserv2/parser_visitor.go",
"repo_id": "milvus",
"token_count": 14583
} | 1,730 |
import { test, expect } from "@jest/globals";
import { MinimaxEmbeddings } from "../minimax.js";
test.skip("Test MinimaxEmbeddings.embedQuery", async () => {
const embeddings = new MinimaxEmbeddings();
const res = await embeddings.embedQuery("Hello world");
expect(typeof res[0]).toBe("number");
});
test.skip("T... | langchainjs/libs/langchain-community/src/embeddings/tests/minimax.int.test.ts/0 | {
"file_path": "langchainjs/libs/langchain-community/src/embeddings/tests/minimax.int.test.ts",
"repo_id": "langchainjs",
"token_count": 386
} | 968 |
name: "" # Filename comes here
allow_empty: false
allow_empty_text: true
subsections:
- name: "Dataset Card for X" # First-level markdown heading
allow_empty: false
allow_empty_text: true
subsections:
- name: "Table of Contents"
allow_empty: false
allow_empty_text: false
subs... | datasets/src/datasets/utils/resources/readme_structure.yaml/0 | {
"file_path": "datasets/src/datasets/utils/resources/readme_structure.yaml",
"repo_id": "datasets",
"token_count": 1924
} | 142 |
# coding=utf-8
# Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a cop... | transformers/src/transformers/models/bert/configuration_bert.py/0 | {
"file_path": "transformers/src/transformers/models/bert/configuration_bert.py",
"repo_id": "transformers",
"token_count": 4005
} | 636 |
<jupyter_start><jupyter_text>10K AnalysisIn this demo, we explore answering complex queries by decomposing them into simpler sub-queries. If you're opening this Notebook on colab, you will probably need to install LlamaIndex 🦙.<jupyter_code>%pip install llama-index-llms-openai
!pip install llama-index
import nest_asyn... | llama_index/docs/examples/usecases/10k_sub_question.ipynb/0 | {
"file_path": "llama_index/docs/examples/usecases/10k_sub_question.ipynb",
"repo_id": "llama_index",
"token_count": 1230
} | 1,132 |
# coding=utf-8
# Copyright 2024 The HuggingFace Inc. team.
# Copyright (c) 2022, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.a... | diffusers/src/diffusers/configuration_utils.py/0 | {
"file_path": "diffusers/src/diffusers/configuration_utils.py",
"repo_id": "diffusers",
"token_count": 13483
} | 231 |
poetry_requirements(
name="poetry",
)
| llama_index/llama-index-integrations/tools/llama-index-tools-exa/BUILD/0 | {
"file_path": "llama_index/llama-index-integrations/tools/llama-index-tools-exa/BUILD",
"repo_id": "llama_index",
"token_count": 18
} | 1,423 |
package coordinator
import (
"context"
"sort"
"testing"
"github.com/chroma/chroma-coordinator/internal/common"
"github.com/chroma/chroma-coordinator/internal/metastore/db/dbcore"
"github.com/chroma/chroma-coordinator/internal/model"
"github.com/chroma/chroma-coordinator/internal/types"
"github.com/google/uuid... | chroma/go/coordinator/internal/coordinator/apis_test.go/0 | {
"file_path": "chroma/go/coordinator/internal/coordinator/apis_test.go",
"repo_id": "chroma",
"token_count": 12884
} | 44 |
import asyncio
from typing import Any, Callable, List, Optional, Sequence
from llama_index.core.async_utils import run_async_tasks
from llama_index.core.callbacks.base import CallbackManager
from llama_index.core.indices.prompt_helper import PromptHelper
from llama_index.core.prompts import BasePromptTemplate
from lla... | llama_index/llama-index-core/llama_index/core/response_synthesizers/accumulate.py/0 | {
"file_path": "llama_index/llama-index-core/llama_index/core/response_synthesizers/accumulate.py",
"repo_id": "llama_index",
"token_count": 2426
} | 1,294 |
:root {
font-family: Inter, system-ui, Avenir, Helvetica, Arial, sans-serif;
line-height: 1.5;
font-weight: 400;
color-scheme: light dark;
color: rgba(255, 255, 255, 0.87);
background-color: #242424;
font-synthesis: none;
text-rendering: optimizeLegibility;
-webkit-font-smoothing: antialiased;
-mo... | langchainjs/environment_tests/test-exports-vite/src/style.css/0 | {
"file_path": "langchainjs/environment_tests/test-exports-vite/src/style.css",
"repo_id": "langchainjs",
"token_count": 682
} | 836 |
// Licensed to the LF AI & Data foundation under one
// or more contributor license agreements. See the NOTICE file
// distributed with this work for additional information
// regarding copyright ownership. The ASF licenses this file
// to you under the Apache License, Version 2.0 (the
// "License"); you may not use th... | milvus/cmd/components/proxy.go/0 | {
"file_path": "milvus/cmd/components/proxy.go",
"repo_id": "milvus",
"token_count": 732
} | 1,892 |
export * from "@langchain/core/utils/tiktoken";
| langchainjs/langchain/src/util/tiktoken.ts/0 | {
"file_path": "langchainjs/langchain/src/util/tiktoken.ts",
"repo_id": "langchainjs",
"token_count": 16
} | 934 |
# Model arguments
model_name_or_path: mistralai/Mistral-7B-v0.1
model_revision: main
torch_dtype: bfloat16
use_flash_attention_2: true
# Data training arguments
chat_template: "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'... | alignment-handbook/recipes/zephyr-7b-beta/sft/config_full.yaml/0 | {
"file_path": "alignment-handbook/recipes/zephyr-7b-beta/sft/config_full.yaml",
"repo_id": "alignment-handbook",
"token_count": 568
} | 23 |
"""Query transform prompts."""
from llama_index.legacy.prompts.base import PromptTemplate
from llama_index.legacy.prompts.prompt_type import PromptType
# deprecated, kept for backwards compatibility
"""Decompose prompt for query transformation.
PromptTemplate to "decompose" a query into another query
given the exist... | llama_index/llama-index-legacy/llama_index/legacy/indices/query/query_transform/prompts.py/0 | {
"file_path": "llama_index/llama-index-legacy/llama_index/legacy/indices/query/query_transform/prompts.py",
"repo_id": "llama_index",
"token_count": 1620
} | 1,667 |
from typing import Any, Dict, List, Optional
import requests
from langchain_core.messages import get_buffer_string
from langchain.memory.chat_memory import BaseChatMemory
MANAGED_URL = "https://api.getmetal.io/v1/motorhead"
# LOCAL_URL = "http://localhost:8080"
class MotorheadMemory(BaseChatMemory):
"""Chat me... | langchain/libs/langchain/langchain/memory/motorhead_memory.py/0 | {
"file_path": "langchain/libs/langchain/langchain/memory/motorhead_memory.py",
"repo_id": "langchain",
"token_count": 1411
} | 548 |
"""Test Clarifai API wrapper.
In order to run this test, you need to have an account on Clarifai.
You can sign up for free at https://clarifai.com/signup.
pip install clarifai
You'll need to set env variable CLARIFAI_PAT_KEY to your personal access token key.
"""
from langchain_community.llms.clarifai import Clarifai... | langchain/libs/community/tests/integration_tests/llms/test_clarifai.py/0 | {
"file_path": "langchain/libs/community/tests/integration_tests/llms/test_clarifai.py",
"repo_id": "langchain",
"token_count": 374
} | 336 |
<!--Copyright 2024 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | diffusers/docs/source/en/optimization/habana.md/0 | {
"file_path": "diffusers/docs/source/en/optimization/habana.md",
"repo_id": "diffusers",
"token_count": 1399
} | 177 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.