gpt4free
The official gpt4free repository | various collection of powerful language models
Top Related Projects
Reverse engineered ChatGPT API
AI agent stdlib that works with any LLM and TypeScript AI SDK.
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
The official Python library for the OpenAI API
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Quick Overview
gpt4free is an open-source project that provides free access to various AI models, including GPT-4, through reverse-engineered APIs. It aims to make advanced language models accessible to developers and researchers without the need for paid subscriptions or API keys.
Pros
- Free access to powerful AI models
- Multiple providers and models available
- Active community and frequent updates
- Useful for testing and prototyping AI applications
Cons
- Potential legal and ethical concerns regarding API usage
- Reliability issues due to dependence on third-party services
- May not be suitable for production environments
- Limited support and documentation compared to official APIs
Code Examples
- Using the Forefront provider:
import g4f
response = g4f.ChatCompletion.create(
model="gpt-3.5-turbo",
provider=g4f.Provider.Forefront,
messages=[{"role": "user", "content": "Hello, how are you?"}],
stream=True,
)
for message in response:
print(message, flush=True, end='')
- Using the You provider:
import g4f
response = g4f.ChatCompletion.create(
model=g4f.models.gpt_35_turbo,
messages=[{"role": "user", "content": "Write a poem about AI"}],
provider=g4f.Provider.You,
)
print(response)
- Using the Bing provider:
import g4f
response = g4f.ChatCompletion.create(
model="gpt-4",
provider=g4f.Provider.Bing,
messages=[{"role": "user", "content": "Explain quantum computing"}],
cookies=g4f.get_cookies(".bing.com"),
)
print(response)
Getting Started
To get started with gpt4free, follow these steps:
- Install the library:
pip install -U g4f
- Import the library and use a provider:
import g4f
response = g4f.ChatCompletion.create(
model="gpt-3.5-turbo",
provider=g4f.Provider.OpenaiChat,
messages=[{"role": "user", "content": "Hello, world!"}],
)
print(response)
Note: Make sure to check the project's documentation for the latest updates and provider-specific instructions.
Competitor Comparisons
Reverse engineered ChatGPT API
Pros of ChatGPT
- More established project with a larger community and longer development history
- Offers a wider range of features, including support for multiple ChatGPT models
- Better documentation and more comprehensive setup instructions
Cons of ChatGPT
- Requires authentication and API keys, which may be less accessible for some users
- More complex setup process compared to gpt4free
- May have higher usage costs due to reliance on official OpenAI APIs
Code Comparison
ChatGPT:
from revChatGPT.V3 import Chatbot
chatbot = Chatbot(api_key="your_api_key")
response = chatbot.ask("Hello, how are you?")
print(response)
gpt4free:
import g4f
response = g4f.ChatCompletion.create(
model="gpt-3.5-turbo",
messages=[{"role": "user", "content": "Hello, how are you?"}]
)
print(response)
Both repositories provide access to ChatGPT-like functionality, but gpt4free aims to offer free access without authentication, while ChatGPT focuses on providing a more robust and official integration with OpenAI's services. The choice between them depends on the user's needs, budget, and ethical considerations regarding API usage.
AI agent stdlib that works with any LLM and TypeScript AI SDK.
Pros of agentic
- Focuses on building autonomous AI agents, offering a more specialized and advanced approach
- Provides a framework for creating complex, goal-oriented AI systems
- Emphasizes ethical considerations and responsible AI development
Cons of agentic
- Less accessible for users seeking simple, ready-to-use GPT-like functionality
- Requires more technical knowledge and setup compared to gpt4free
- Smaller community and fewer contributors, potentially leading to slower development
Code Comparison
gpt4free:
from g4f import ChatCompletion
response = ChatCompletion.create(
model="gpt-3.5-turbo",
messages=[{"role": "user", "content": "Hello, how are you?"}]
)
print(response)
agentic:
from agentic import Agent, Task
agent = Agent()
task = Task("Greet the user and ask how they are")
result = agent.run(task)
print(result)
The code comparison shows that gpt4free provides a more straightforward interface for generating responses, while agentic focuses on creating autonomous agents to perform tasks. gpt4free is better suited for quick, chat-like interactions, whereas agentic is designed for more complex, goal-oriented AI applications.
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
Pros of AutoGPT
- Autonomous task completion with minimal human intervention
- Versatile application across various domains (e.g., coding, research, analysis)
- Active development and community support
Cons of AutoGPT
- Requires API key and potentially higher costs for extended use
- More complex setup and configuration process
- May produce inconsistent or unexpected results due to its autonomous nature
Code Comparison
AutoGPT:
def start_interaction_loop(self):
# Interaction loop
while True:
# Get user input
user_input = input("Human: ")
if user_input.lower() == "exit":
break
gpt4free:
def create_chat(self, model="gpt-3.5-turbo", messages=None, **kwargs):
if messages is None:
messages = []
return ChatCompletion.create(
model=model, messages=messages, **kwargs
)
AutoGPT focuses on creating an autonomous agent that can perform tasks with minimal human intervention, while gpt4free aims to provide free access to various language models. AutoGPT offers more advanced features but requires more setup, while gpt4free is simpler to use but may have limitations in terms of available models and functionality.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
Pros of DeepSpeed
- Highly optimized for large-scale distributed training of deep learning models
- Supports a wide range of AI models and frameworks (PyTorch, TensorFlow, etc.)
- Backed by Microsoft, ensuring ongoing development and support
Cons of DeepSpeed
- Steeper learning curve due to its complexity and advanced features
- Primarily focused on model training, not inference or API access
- Requires more computational resources for optimal performance
Code Comparison
DeepSpeed:
import deepspeed
model_engine, optimizer, _, _ = deepspeed.initialize(args=args,
model=model,
model_parameters=params)
for step, batch in enumerate(data_loader):
loss = model_engine(batch)
model_engine.backward(loss)
model_engine.step()
gpt4free:
import g4f
response = g4f.ChatCompletion.create(
model="gpt-3.5-turbo",
messages=[{"role": "user", "content": "Hello, how are you?"}]
)
print(response)
Summary
DeepSpeed is a powerful tool for optimizing large-scale AI model training, while gpt4free focuses on providing free access to GPT-like models. DeepSpeed offers more advanced features and better performance for training, but requires more expertise and resources. gpt4free is simpler to use and provides easy access to AI models, but may have limitations in terms of model quality and legal considerations.
The official Python library for the OpenAI API
Pros of openai-python
- Official library maintained by OpenAI, ensuring reliability and up-to-date features
- Comprehensive documentation and support from OpenAI
- Seamless integration with OpenAI's API and services
Cons of openai-python
- Requires an API key and associated costs for usage
- Limited to OpenAI's models and services
Code Comparison
openai-python:
import openai
openai.api_key = "your-api-key"
response = openai.Completion.create(engine="text-davinci-002", prompt="Hello, world!")
print(response.choices[0].text)
gpt4free:
import g4f
response = g4f.ChatCompletion.create(model="gpt-3.5-turbo", messages=[{"role": "user", "content": "Hello, world!"}])
print(response)
Key Differences
- openai-python is the official library, while gpt4free is a third-party alternative
- gpt4free aims to provide free access to AI models, while openai-python requires an API key and associated costs
- openai-python offers more extensive features and model options, while gpt4free focuses on providing free alternatives
- gpt4free may have potential legal and ethical concerns due to its nature of bypassing official APIs
Use Cases
- openai-python: Ideal for professional and commercial applications requiring reliable and official API access
- gpt4free: Suitable for personal projects, experimentation, or scenarios where API costs are a concern, but with potential limitations and risks
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Pros of transformers
- Comprehensive library with support for numerous pre-trained models
- Extensive documentation and community support
- Seamless integration with PyTorch and TensorFlow
Cons of transformers
- Steeper learning curve for beginners
- Larger library size and potentially higher resource requirements
Code Comparison
transformers:
from transformers import pipeline
classifier = pipeline("sentiment-analysis")
result = classifier("I love this library!")[0]
print(f"Label: {result['label']}, Score: {result['score']:.4f}")
gpt4free:
import g4f
response = g4f.ChatCompletion.create(
model="gpt-3.5-turbo",
messages=[{"role": "user", "content": "Hello, how are you?"}]
)
print(response)
Summary
transformers is a robust, well-documented library for working with various pre-trained models, offering extensive functionality and integration with popular deep learning frameworks. It's ideal for advanced users and large-scale projects.
gpt4free, on the other hand, provides a simpler interface for accessing GPT models, making it more accessible for quick implementations and experimentation. However, it may lack the comprehensive features and community support of transformers.
Convert designs to code with AI
Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.
Try Visual CopilotREADME
Written by @xtekky & maintained by @hlohaus
By using this repository or any code related to it, you agree to the legal notice. The author is not responsible for the usage of this repository nor endorses it, nor is the author responsible for any copies, forks, re-uploads made by other users, or anything else related to GPT4Free. This is the author's only account and repository. To prevent impersonation or irresponsible actions, please comply with the GNU GPL license this Repository uses.
[!Warning] > "gpt4free" serves as a PoC (proof of concept), demonstrating the development of an API package with multi-provider requests, with features like timeouts, load balance and flow control.
pip install -U g4f
docker pull hlohaus789/g4f
ð What's New
- Added
gpt-4o
, simply usegpt-4o
inchat.completion.create
. - Installation Guide for Windows (.exe): ð» #installation-guide-for-windows
- Join our Telegram Channel: ð¨ telegram.me/g4f_channel
- Join our Discord Group: ð¬ discord.gg/XfybzPXPH5
g4f
now supports 100% local inference: ð§ local-docs
ð» Site Takedown
Is your site on this repository and you want to take it down? Send an email to takedown@g4f.ai with proof it is yours and it will be removed as fast as possible. To prevent reproduction please secure your API. ð
ð Feedback and Todo
You can always leave some feedback here: https://forms.gle/FeWV9RLEedfdkmFN6
As per the survey, here is a list of improvements to come
- Update the repository to include the new openai library syntax (ex:
Openai()
class) | completed, useg4f.client.Client
- Golang implementation
- ð§ Improve Documentation (in /docs & Guides, Howtos, & Do video tutorials)
- Improve the provider status list & updates
- Tutorials on how to reverse sites to write your own wrapper (PoC only ofc)
- Improve the Bing wrapper. (Wait and Retry or reuse conversation)
- ð§ Write a standard provider performance test to improve the stability
- Potential support and development of local models
- ð§ Improve compatibility and error handling
ð Table of Contents
- ð What's New
- ð Table of Contents
- ð ï¸ Getting Started
- ð¡ Usage
- ð Providers and Models
- ð Powered by gpt4free
- ð¤ Contribute
- ð Contributors
- Â©ï¸ Copyright
- â Star History
- ð License
ð ï¸ Getting Started
Docker Container Guide
Getting Started Quickly:
-
Install Docker: Begin by downloading and installing Docker.
-
Set Up the Container: Use the following commands to pull the latest image and start the container:
docker pull hlohaus789/g4f
docker run \
-p 8080:8080 -p 1337:1337 -p 7900:7900 \
--shm-size="2g" \
-v ${PWD}/har_and_cookies:/app/har_and_cookies \
-v ${PWD}/generated_images:/app/generated_images \
hlohaus789/g4f:latest
-
Access the Client:
- To use the included client, navigate to: http://localhost:8080/chat/
- Or set the API base for your client to: http://localhost:1337/v1
-
(Optional) Provider Login: If required, you can access the container's desktop here: http://localhost:7900/?autoconnect=1&resize=scale&password=secret for provider login purposes.
Installation Guide for Windows (.exe)
To ensure the seamless operation of our application, please follow the instructions below. These steps are designed to guide you through the installation process on Windows operating systems.
Installation Steps
- Download the Application: Visit our releases page and download the most recent version of the application, named
g4f.exe.zip
. - File Placement: After downloading, locate the
.zip
file in your Downloads folder. Unpack it to a directory of your choice on your system, then execute theg4f.exe
file to run the app. - Open GUI: The app starts a web server with the GUI. Open your favorite browser and navigate to
http://localhost:8080/chat/
to access the application interface. - Firewall Configuration (Hotfix): Upon installation, it may be necessary to adjust your Windows Firewall settings to allow the application to operate correctly. To do this, access your Windows Firewall settings and allow the application.
By following these steps, you should be able to successfully install and run the application on your Windows system. If you encounter any issues during the installation process, please refer to our Issue Tracker or try to get contact over Discord for assistance.
Run the Webview UI on other Platfroms:
Use your smartphone:
Run the Web UI on Your Smartphone:
Use python
Prerequisites:
- Download and install Python (Version 3.10+ is recommended).
- Install Google Chrome for providers with webdriver
Install using PyPI package:
pip install -U g4f[all]
How do I install only parts or do disable parts? Use partial requirements: /docs/requirements
Install from source:
How do I load the project using git and installing the project requirements? Read this tutorial and follow it step by step: /docs/git
Install using Docker:
How do I build and run composer image from source? Use docker-compose: /docs/docker
ð¡ Usage
Text Generation
from g4f.client import Client
client = Client()
response = client.chat.completions.create(
model="gpt-3.5-turbo",
messages=[{"role": "user", "content": "Hello"}],
...
)
print(response.choices[0].message.content)
Hello! How can I assist you today?
Image Generation
from g4f.client import Client
client = Client()
response = client.images.generate(
model="gemini",
prompt="a white siamese cat",
...
)
image_url = response.data[0].url
Full Documentation for Python API
- New AsyncClient API from G4F: /docs/async_client
- Client API like the OpenAI Python library: /docs/client
- Legacy API with python modules: /docs/legacy
Web UI
To start the web interface, type the following codes in python:
from g4f.gui import run_gui
run_gui()
or execute the following command:
python -m g4f.cli gui -port 8080 -debug
Interference API
You can use the Interference API to serve other OpenAI integrations with G4F.
See docs: /docs/interference
Access with: http://localhost:1337/v1
Configuration
Cookies
Cookies are essential for using Meta AI and Microsoft Designer to create images. Additionally, cookies are required for the Google Gemini and WhiteRabbitNeo Provider. From Bing, ensure you have the "_U" cookie, and from Google, all cookies starting with "__Secure-1PSID" are needed.
You can pass these cookies directly to the create function or set them using the set_cookies
method before running G4F:
from g4f.cookies import set_cookies
set_cookies(".bing.com", {
"_U": "cookie value"
})
set_cookies(".google.com", {
"__Secure-1PSID": "cookie value"
})
Using .har and Cookie Files
You can place .har
and cookie files in the default ./har_and_cookies
directory. To export a cookie file, use the EditThisCookie Extension available on the Chrome Web Store.
Creating .har Files to Capture Cookies
To capture cookies, you can also create .har
files. For more details, refer to the next section.
Changing the Cookies Directory and Loading Cookie Files in Python
You can change the cookies directory and load cookie files in your Python environment. To set the cookies directory relative to your Python file, use the following code:
import os.path
from g4f.cookies import set_cookies_dir, read_cookie_files
import g4f.debug
g4f.debug.logging = True
cookies_dir = os.path.join(os.path.dirname(__file__), "har_and_cookies")
set_cookies_dir(cookies_dir)
read_cookie_files(cookies_dir)
Debug Mode
If you enable debug mode, you will see logs similar to the following:
Read .har file: ./har_and_cookies/you.com.har
Cookies added: 10 from .you.com
Read cookie file: ./har_and_cookies/google.json
Cookies added: 16 from .google.com
.HAR File for OpenaiChat Provider
Generating a .HAR File
To utilize the OpenaiChat provider, a .har file is required from https://chatgpt.com/. Follow the steps below to create a valid .har file:
- Navigate to https://chatgpt.com/ using your preferred web browser and log in with your credentials.
- Access the Developer Tools in your browser. This can typically be done by right-clicking the page and selecting "Inspect," or by pressing F12 or Ctrl+Shift+I (Cmd+Option+I on a Mac).
- With the Developer Tools open, switch to the "Network" tab.
- Reload the website to capture the loading process within the Network tab.
- Initiate an action in the chat which can be captured in the .har file.
- Right-click any of the network activities listed and select "Save all as HAR with content" to export the .har file.
Storing the .HAR File
- Place the exported .har file in the
./har_and_cookies
directory if you are using Docker. Alternatively, you can store it in any preferred location within your current working directory.
Note: Ensure that your .har file is stored securely, as it may contain sensitive information.
Using Proxy
If you want to hide or change your IP address for the providers, you can set a proxy globally via an environment variable:
- On macOS and Linux:
export G4F_PROXY="http://host:port"
- On Windows:
set G4F_PROXY=http://host:port
ð Providers and Models
GPT-4
Website | Provider | GPT-3.5 | GPT-4 | Stream | Status | Auth |
---|---|---|---|---|---|---|
bing.com | g4f.Provider.Bing | â | âï¸ | âï¸ | â | |
chatgpt.ai | g4f.Provider.ChatgptAi | â | âï¸ | âï¸ | â | |
liaobots.site | g4f.Provider.Liaobots | âï¸ | âï¸ | âï¸ | â | |
chatgpt.com | g4f.Provider.OpenaiChat | âï¸ | âï¸ | âï¸ | â+âï¸ | |
raycast.com | g4f.Provider.Raycast | âï¸ | âï¸ | âï¸ | âï¸ | |
beta.theb.ai | g4f.Provider.Theb | âï¸ | âï¸ | âï¸ | â | |
you.com | g4f.Provider.You | âï¸ | âï¸ | âï¸ | â |
Best OpenSource Models
While we wait for gpt-5, here is a list of new models that are at least better than gpt-3.5-turbo. Some are better than gpt-4. Expect this list to grow.
Website | Provider | parameters | better than |
---|---|---|---|
claude-3-opus | g4f.Provider.You | ?B | gpt-4-0125-preview |
command-r+ | g4f.Provider.HuggingChat | 104B | gpt-4-0314 |
llama-3-70b | g4f.Provider.Llama or DeepInfra | 70B | gpt-4-0314 |
claude-3-sonnet | g4f.Provider.You | ?B | gpt-4-0314 |
reka-core | g4f.Provider.Reka | 21B | gpt-4-vision |
dbrx-instruct | g4f.Provider.DeepInfra | 132B / 36B active | gpt-3.5-turbo |
mixtral-8x22b | g4f.Provider.DeepInfra | 176B / 44b active | gpt-3.5-turbo |
GPT-3.5
Website | Provider | GPT-3.5 | GPT-4 | Stream | Status | Auth |
---|---|---|---|---|---|---|
chat3.aiyunos.top | g4f.Provider.AItianhuSpace | âï¸ | â | âï¸ | â | |
chat10.aichatos.xyz | g4f.Provider.Aichatos | âï¸ | â | âï¸ | â | |
chatforai.store | g4f.Provider.ChatForAi | âï¸ | â | âï¸ | â | |
chatgpt4online.org | g4f.Provider.Chatgpt4Online | âï¸ | â | âï¸ | â | |
chatgpt-free.cc | g4f.Provider.ChatgptNext | âï¸ | â | âï¸ | â | |
chatgptx.de | g4f.Provider.ChatgptX | âï¸ | â | âï¸ | â | |
duckduckgo.com | g4f.Provider.DDG | âï¸ | â | âï¸ | â | |
feedough.com | g4f.Provider.Feedough | âï¸ | â | âï¸ | â | |
flowgpt.com | g4f.Provider.FlowGpt | âï¸ | â | âï¸ | â | |
freegptsnav.aifree.site | g4f.Provider.FreeGpt | âï¸ | â | âï¸ | â | |
gpttalk.ru | g4f.Provider.GptTalkRu | âï¸ | â | âï¸ | â | |
koala.sh | g4f.Provider.Koala | âï¸ | â | âï¸ | â | |
app.myshell.ai | g4f.Provider.MyShell | âï¸ | â | âï¸ | â | |
perplexity.ai | g4f.Provider.PerplexityAi | âï¸ | â | âï¸ | â | |
poe.com | g4f.Provider.Poe | âï¸ | â | âï¸ | âï¸ | |
talkai.info | g4f.Provider.TalkAi | âï¸ | â | âï¸ | â | |
chat.vercel.ai | g4f.Provider.Vercel | âï¸ | â | âï¸ | â | |
aitianhu.com | g4f.Provider.AItianhu | âï¸ | â | âï¸ | â | |
chatgpt.bestim.org | g4f.Provider.Bestim | âï¸ | â | âï¸ | â | |
chatbase.co | g4f.Provider.ChatBase | âï¸ | â | âï¸ | â | |
chatgptdemo.info | g4f.Provider.ChatgptDemo | âï¸ | â | âï¸ | â | |
chat.chatgptdemo.ai | g4f.Provider.ChatgptDemoAi | âï¸ | â | âï¸ | â | |
chatgptfree.ai | g4f.Provider.ChatgptFree | âï¸ | â | â | â | |
chatgptlogin.ai | g4f.Provider.ChatgptLogin | âï¸ | â | âï¸ | â | |
chat.3211000.xyz | g4f.Provider.Chatxyz | âï¸ | â | âï¸ | â | |
gpt6.ai | g4f.Provider.Gpt6 | âï¸ | â | âï¸ | â | |
gptchatly.com | g4f.Provider.GptChatly | âï¸ | â | â | â | |
ai18.gptforlove.com | g4f.Provider.GptForLove | âï¸ | â | âï¸ | â | |
gptgo.ai | g4f.Provider.GptGo | âï¸ | â | âï¸ | â | |
gptgod.site | g4f.Provider.GptGod | âï¸ | â | âï¸ | â | |
onlinegpt.org | g4f.Provider.OnlineGpt | âï¸ | â | âï¸ | â |
Other
Website | Provider | Stream | Status | Auth |
---|---|---|---|---|
openchat.team | g4f.Provider.Aura | âï¸ | â | |
blackbox.ai | g4f.Provider.Blackbox | âï¸ | â | |
cohereforai-c4ai-command-r-plus.hf.space | g4f.Provider.Cohere | âï¸ | â | |
deepinfra.com | g4f.Provider.DeepInfra | âï¸ | â | |
free.chatgpt.org.uk | g4f.Provider.FreeChatgpt | âï¸ | â | |
gemini.google.com | g4f.Provider.Gemini | âï¸ | âï¸ | |
ai.google.dev | g4f.Provider.GeminiPro | âï¸ | âï¸ | |
gemini-chatbot-sigma.vercel.app | g4f.Provider.GeminiProChat | âï¸ | â | |
developers.sber.ru | g4f.Provider.GigaChat | âï¸ | âï¸ | |
console.groq.com | g4f.Provider.Groq | âï¸ | âï¸ | |
huggingface.co | g4f.Provider.HuggingChat | âï¸ | â | |
huggingface.co | g4f.Provider.HuggingFace | âï¸ | â | |
llama2.ai | g4f.Provider.Llama | âï¸ | â | |
meta.ai | g4f.Provider.MetaAI | âï¸ | â | |
openrouter.ai | g4f.Provider.OpenRouter | âï¸ | âï¸ | |
labs.perplexity.ai | g4f.Provider.PerplexityLabs | âï¸ | â | |
pi.ai | g4f.Provider.Pi | âï¸ | â | |
replicate.com | g4f.Provider.Replicate | âï¸ | â | |
theb.ai | g4f.Provider.ThebApi | âï¸ | âï¸ | |
whiterabbitneo.com | g4f.Provider.WhiteRabbitNeo | âï¸ | âï¸ | |
bard.google.com | g4f.Provider.Bard | â | âï¸ |
Models
Model | Base Provider | Provider | Website |
---|---|---|---|
gpt-3.5-turbo | OpenAI | 8+ Providers | openai.com |
gpt-4 | OpenAI | 2+ Providers | openai.com |
gpt-4-turbo | OpenAI | g4f.Provider.Bing | openai.com |
Llama-2-7b-chat-hf | Meta | 2+ Providers | llama.meta.com |
Llama-2-13b-chat-hf | Meta | 2+ Providers | llama.meta.com |
Llama-2-70b-chat-hf | Meta | 3+ Providers | llama.meta.com |
Meta-Llama-3-8b-instruct | Meta | 1+ Providers | llama.meta.com |
Meta-Llama-3-70b-instruct | Meta | 2+ Providers | llama.meta.com |
CodeLlama-34b-Instruct-hf | Meta | g4f.Provider.HuggingChat | llama.meta.com |
CodeLlama-70b-Instruct-hf | Meta | 2+ Providers | llama.meta.com |
Mixtral-8x7B-Instruct-v0.1 | Huggingface | 4+ Providers | huggingface.co |
Mistral-7B-Instruct-v0.1 | Huggingface | 3+ Providers | huggingface.co |
Mistral-7B-Instruct-v0.2 | Huggingface | g4f.Provider.DeepInfra | huggingface.co |
zephyr-orpo-141b-A35b-v0.1 | Huggingface | 2+ Providers | huggingface.co |
dolphin-2.6-mixtral-8x7b | Huggingface | g4f.Provider.DeepInfra | huggingface.co |
gemini | g4f.Provider.Gemini | gemini.google.com | |
gemini-pro | 2+ Providers | gemini.google.com | |
claude-v2 | Anthropic | 1+ Providers | anthropic.com |
claude-3-opus | Anthropic | g4f.Provider.You | anthropic.com |
claude-3-sonnet | Anthropic | g4f.Provider.You | anthropic.com |
lzlv_70b_fp16_hf | Huggingface | g4f.Provider.DeepInfra | huggingface.co |
airoboros-70b | Huggingface | g4f.Provider.DeepInfra | huggingface.co |
openchat_3.5 | Huggingface | 2+ Providers | huggingface.co |
pi | Inflection | g4f.Provider.Pi | inflection.ai |
Image and Vision Models
Label | Provider | Image Model | Vision Model | Website |
---|---|---|---|---|
Microsoft Copilot in Bing | g4f.Provider.Bing | dall-e-3 | gpt-4-vision | bing.com |
DeepInfra | g4f.Provider.DeepInfra | stability-ai/sdxl | llava-1.5-7b-hf | deepinfra.com |
Gemini | g4f.Provider.Gemini | âï¸ | âï¸ | gemini.google.com |
Gemini API | g4f.Provider.GeminiPro | â | gemini-1.5-pro | ai.google.dev |
Meta AI | g4f.Provider.MetaAI | âï¸ | â | meta.ai |
OpenAI ChatGPT | g4f.Provider.OpenaiChat | dall-e-3 | gpt-4-vision | chatgpt.com |
Reka | g4f.Provider.Reka | â | âï¸ | chat.reka.ai |
Replicate | g4f.Provider.Replicate | stability-ai/sdxl | llava-v1.6-34b | replicate.com |
You.com | g4f.Provider.You | dall-e-3 | âï¸ | you.com |
ð Powered by gpt4free
ð Projects | â Stars | ð Forks | ð Issues | ð¬ Pull requests |
gpt4free | ||||
gpt4free-ts | ||||
Free AI API's & Potential Providers List | ||||
ChatGPT-Clone | ||||
Ai agent | ||||
ChatGpt Discord Bot | ||||
chatGPT-discord-bot | ||||
Nyx-Bot (Discord) | ||||
LangChain gpt4free | ||||
ChatGpt Telegram Bot | ||||
ChatGpt Line Bot | ||||
Action Translate Readme | ||||
Langchain Document GPT | ||||
python-tgpt | ||||
GPT4js |
ð¤ Contribute
We welcome contributions from the community. Whether you're adding new providers or features, or simply fixing typos and making small improvements, your input is valued. Creating a pull request is all it takes â our co-pilot will handle the code review process. Once all changes have been addressed, we'll merge the pull request into the main branch and release the updates at a later time.
Guide: How do i create a new Provider?
Guide: How can AI help me with writing code?
- Read: /docs/guides/help_me
ð Contributors
A list of all contributors is available here
- The
Vercel.py
file contains code from vercel-llm-api by @ading2210 - The
har_file.py
has input from xqdoo00o/ChatGPT-to-API - The
PerplexityLabs.py
has input from nathanrchn/perplexityai - The
Gemini.py
has input from dsdanielpark/Gemini-API - The
MetaAI.py
file contains code from meta-ai-api by @Strvm - The
proofofwork.py
has input from missuo/FreeGPT35
Having input implies that the AI's code generation utilized it as one of many sources.
Â©ï¸ Copyright
This program is licensed under the GNU GPL v3
xtekky/gpt4free: Copyright (C) 2023 xtekky
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <https://www.gnu.org/licenses/>.
â Star History
ð License
|
This project is licensed under GNU_GPL_v3.0. |
Top Related Projects
Reverse engineered ChatGPT API
AI agent stdlib that works with any LLM and TypeScript AI SDK.
AutoGPT is the vision of accessible AI for everyone, to use and to build on. Our mission is to provide the tools, so that you can focus on what matters.
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
The official Python library for the OpenAI API
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
Convert designs to code with AI
Introducing Visual Copilot: A new AI model to turn Figma designs to high quality code using your components.
Try Visual Copilot