01. LLM models available
LLM models available
Copy
# Configuration file for managing API KEY as environment variable
from dotenv import load_dotenv
# Load API KEY information
load_dotenv()Copy
TrueCopy
# Set up LangSmith tracking. https://smith.langchain.com
# !pip install langchain-teddynote
from langchain_teddynote import logging
from langchain_teddynote.messages import stream_response
# Enter a project name.
logging.langsmith("CH04-Models")Copy
OpenAI
Summary
OpenAI offers a Large Language Model (LLM) dedicated to chatting. When creating this model, you can specify various options, and these options affect how the model works.
Option details
temperature
Option to set sampling temperature. Values can be selected between 0 and 2. High values (e.g. 0.8) make the output more random, and lower values (e.g. 0.2) make the output more focused and deterministic.
max_tokens
Specifies the maximum number of tokens to be created in the chat completion. This option controls the length of the text that the model can generate at once.
model_name
Option to select an applicable model. For more information OpenAI model documentation You can check in.
Model specification
Link: https://platform.openai.com/docs/models/gpt-4o
Model name
Explanation
Context length
Max output token
Learning data
gpt-4o
The latest multi-mode flagship model cheaper and faster than the GPT-4 turbo
128,000 tokens
16,384 tokens
Until October 2023
gpt-4-turbo
The latest GPT-4 turbo model. Vision function, JSON mode, function call support
128,000 tokens
4,096 tokens
Until December 2023
gpt-4o-mini
Small model with better performance than the GPT-3.5 turbo
128,000 tokens
16,384 tokens
Until October 2023
o1-preview
Reasoning model for solving difficult problems in various domains
128,000 tokens
32,768 tokens
Until October 2023
o1-mini
Quick reasoning model specialized in coding, math and science
128,000 tokens
65,536 tokens
Until October 2023
gpt-4o-realtime
Audio and text input processing model for real-time API (beta)
128,000 tokens
4,096 tokens
Until October 2023
Copy
Copy
Anthropic
Anthropic is an American startup focused on artificial (AI) safety and research. The main information is:
Year established : 2021 years
location : San Francisco, USA
Founder : OpenAI employees (Daniela Amodei and Dario Amodei, etc.)
Corporate form : Registered as a Public Benefit Corporation
Claude
Claude is Anthropic's representative large-scale language model (LLM) family.
API key issuance : https://console.anthropic.com/settings/keys
model_name
model
Anthropic API
AWS Bedrock
GCP Vertex AI
Claude 3.5 Opus
Coming soon
Coming soon
Coming soon
Coming soon
Claude 3.5 Sonnet
claude-3-5-sonnet-20241022
claude-3-5-sonnet-20241022
anthropic.claude-3-5-sonnet-20241022-v2:0
claude-3-5-sonnet-v2@20241022
Claude 3.5 Haiku
Coming soon
Coming soon
Coming soon
Coming soon
Claude 3 Opus
claude-3-opus-20240229
claude-3-opus-20240229
anthropic.claude-3-opus-20240229-v1:0
claude-3-opus@20240229
Claude 3 Sonnet
claude-3-sonnet-20240229
claude-3-sonnet-20240229
anthropic.claude-3-sonnet-20240229-v1:0
claude-3-sonnet@20240229
Claude 3 Haiku
claude-3-haiku-20240307
claude-3-haiku-20240307
anthropic.claude-3-haiku-20240307-v1:0
claude-3-haiku@20240307
Copy
Copy
Perplexity
Link: https://www.perplexity.ai/
Seonhang : 2022 years
Main investors : Jeff Bezos, Nvidia, Databricks, Bessemer Venture Partners, IVP, Wayra, etc.
Recent funding : $500 million (October 2024)
Corporate value : About 90 billion dollars (as of November 2024)
Active monthly users : 15 million people
Perplexity Pro exact features
Daily Pro Search : 300 rounds
AI model selection : GPT-4 Omni, Claude 3 Sonnet/Haiku, Sonar Large 32k[5]
File analysis : PDF, CSV, image file support
price : month In additionhasyear20In additionhasyear 200
How to use Perplexity API
price
API credit acquisition
Perplexity Pro subscription API credits equivalent to $5 per month
API model options
Llama 3 based model
Perplexity Online LLM
Citation function included
API key issuance: API console
After issuing API key .env Save key to file
Copy
hump
Copy
Reference
Support model
Model
Parameter Count
Context Length
Model Type
llama-3.1-sonar-small-128k-online
8B
127,072
Chat Completion
llama-3.1-sonar-large-128k-online
70B
127,072
Chat Completion
llama-3.1-sonar-huge-128k-online
405B
127,072
Chat Completion
ChatPerplexity parameter
model Specify the language model to use (eg "llama-3.1-sonar-small-128k-online")-determine basic performance and abilities.
temperature Adjust the randomness of the responses (0.0-1.0), 0 is decisive, 1 is the most unlawful response generation.
top_p Setting the probability threshold for token sampling (0.0-1.0), the higher the more various outputs allowed.
search_domain_filter Limit search results to specified domains, provide in list form (eg ["perplexity.ai"]).
return_images A disadvantage flag that determines whether an image is included in the response.
return_related_questions Bulle flag to enable/disable related question suggestion function.
top_k Limit the number of search results to use (meaning 0 is unlimited).
streaming A disadvantage flag that determines whether a response will be received by streaming or in a finished form.
presence_penalty Penalty for token repetition (-2.0 to 2.0), higher to suppress reuse.
frequency_penalty General/rare token preference adjustment (-2.0 to 2.0), higher, rare token preference.
Copy
Output the response. ChatPerplexity The source of knowledge information citations Save to property.
Copy
Copy
Streaming output
Copy
Copy
Cohere
Cohere is a leading AI company that provides enterprise-wide artificial intelligence solutions, developing a large language model (LLM) to help companies easily introduce and leverage AI technology.
Cohere overview
Seonhang : 2020 years
Main investors : Inovia Capital, NVIDIA, Oracle, Salesforce Ventures
Series C funding : $270 million childish
Corporate mission : Providing an AI platform for businesses
Main products
Command R+
Command R+ is Cohere's latest LLM optimized for business use.
Main features
Long context window : 128k token support
Advanced RAG function : Provides search enhancement creation function
Multilingual support : Support for 10 major business languages
Automation tool usage function : Automating complex business processes
Aya
Aya is an open source multilingual LLM developed by Cohere for AI, a non-profit research institute in Cohere.
Main features
Language support : 101 languages supported (more than twice the existing open source model)
Training dataset : Revealing a large multilingual training dataset containing 500 million data points
Copy
Copy
Upstage
Upstage is a domestic startup specialized in artificial intelligence (AI) technology, especially large language models (LLM) and document AI.
Main products and technologies
Solar LLM
Main features : Upstage's flagship large-scale language model, noted for its fast performance and cost effectiveness.
Technical approach : Maximize performance by applying Depth-Up Scaling (DUS) technology.
Platform integration : Integrated into API through various platforms such as Amazon SageMaker JumpStart.
Document AI Pack
function : A document processing solution based on OCR technology, which accurately extracts and digitizes what is needed from complex documents.
AskUp Seargest
Features : Provides personalized search and referral services, and is an upgraded version of the existing ChatGPT integrated free chatbot AskUp.
API key issuance
API key issuance here Available in.
Copy
Copy
Copy
Xionic
Sionic AI is a promising geothermal startup in South Korea, developing enterprise-generated AI solutions. Here are the main information about this company:
Main products
STORM Platform : A platform that allows companies to apply generated AI right away without technical concerns
STORM Answer : Aiming to improve business productivity with enterprise-optimized, productive AI solutions
Xionic : Korean AI model with commercial license
Korean model of commercial license
Link: https://github.com/sionic-ai/xionic (Note) 2024.11.29 Standard Usage Value (API server error)
Copy
LogicKor
LogicKor is a benchmark leaderboard created to evaluate the multi-sector thinking power of Korean language models.
purpose
Benchmarks for measuring thinking in various fields of Korean language model
Evaluation area
Korean reasoning
Math
writing
Coding
comprehension
Main features
Various model evaluation : You can compare the performance of various language models at home and abroad
Objective performance measurement : Provide objective indicators by evaluating the actual performance of the model in varying degrees
Open source : An open platform that anyone can access and check the results
LogicKor Leaderboard is an important tool for the development of the Korean AI model, and continuous improvement and development are expected.
link: LogicKor leaderboard
Last updated