www.phind.com Open in urlscan Pro
2606:4700::6812:536  Public Scan

Submitted URL: https://phind.com/blog/phind-model-beats-gpt4-fast
Effective URL: https://www.phind.com/blog/phind-model-beats-gpt4-fast
Submission Tags: 0xscam
Submission: On September 19 via api from US — Scanned from DE

Form analysis 0 forms found in the DOM

Text Content

Sign In

TOOLS

Try Phind in VS Code
Set as Default
Mobile App
Hotkeys



PHIND PRO

Plans



WHO WE ARE

About
Tutorial
Blog
Privacy
Terms



CONTACT

Join the Discord Community
Follow us on Twitter



THEME

Dark Mode


PHIND MODEL BEATS GPT-4 AT CODING, WITH GPT-3.5-LIKE SPEED AND 16K CONTEXT


WE'RE EXCITED TO ANNOUNCE THAT PHIND NOW DEFAULTS TO OUR OWN MODEL THAT MATCHES
AND EXCEEDS GPT-4'S CODING ABILITIES WHILE RUNNING 5X FASTER. YOU CAN NOW GET
HIGH QUALITY ANSWERS FOR TECHNICAL QUESTIONS IN 10 SECONDS INSTEAD OF 50.


The current 7th-generation Phind Model is built on top of our open-source
CodeLlama-34B fine-tunes that were the first models to beat GPT-4's score on
HumanEval and are still the best open source coding models overall by a wide
margin.

 * The Phind Model V7 achieves 74.7% pass@1 on HumanEval

This new model has been fine-tuned on an additional 70B+ tokens of high quality
code and reasoning problems and exhibits a HumanEval score of 74.7%. However,
we've found that HumanEval is a poor indicator of real-world helpfulness. After
deploying previous iterations of the Phind Model on our service, we've collected
detailed feedback and noticed that our model matches or exceeds GPT-4's
helpfulness most of the time on real-world questions. Many in our Discord
community have begun using Phind exclusively with the Phind Model despite also
having unlimited access to GPT-4.

One of the Phind Model's key advantages is that's very fast. We've been able to
achieve a 5x speedup over GPT-4 by running our model on H100s using the new
TensorRT-LLM library from NVIDIA, reaching 100 tokens per second single-stream.

Another key advantage of the Phind Model is context – it supports up to 16k
tokens. We currently allow inputs of up to 12k tokens on the website and reserve
the remaining 4k for web results.

There are still some rough edges with the Phind Model and we'll continue
improving it constantly. One area where it still suffers is consistency — on
certain challenging questions where it is capable of getting the right answer,
the Phind Model might take more generations to get to the right answer than
GPT-4.

Try the Phind Model