Sunday, 8 Feb 2026
Subscribe
logo
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Font ResizerAa
Data Center NewsData Center News
Search
  • Global
  • AI
  • Cloud Computing
  • Edge Computing
  • Security
  • Investment
  • Sustainability
  • More
    • Colocation
    • Quantum Computing
    • Regulation & Policy
    • Infrastructure
    • Power & Cooling
    • Design
    • Innovations
    • Blog
Have an existing account? Sign In
Follow US
© 2022 Foxiz News Network. Ruby Design Company. All Rights Reserved.
Data Center News > Blog > AI > Microsoft unveils serverless fine-tuning for its Phi-3 small language model
AI

Microsoft unveils serverless fine-tuning for its Phi-3 small language model

Last updated: July 26, 2024 12:57 am
Published July 26, 2024
Share
Microsoft unveils serverless fine-tuning for its Phi-3 small language model
SHARE

Be a part of our every day and weekly newsletters for the most recent updates and unique content material on industry-leading AI protection. Be taught Extra


Microsoft is a significant backer and associate of OpenAI, however that doesn’t imply it needs to let the latter firm run away with the generative AI ballgame.

As proof of that, at present Microsoft introduced a new way to fine-tune its Phi-3 small language model with out builders having to handle their very own servers, and without spending a dime (initially).

Fine-tuning refers to the process of adapting an AI model via system prompts or adjusting its underlying weights (parameters) to make it behave in numerous and extra optimum methods for particular use circumstances and finish customers, even including new capabilities.

What’s Phi-3?

The corporate unveiled Phi-3, a 3 billion parameter mannequin, again in April as a low-cost, enterprise grade choice for third-party builders to construct new purposes and software program atop of.

Whereas considerably smaller than most different main language fashions (Meta’s Llama 3.1 as an example, is available in a 405 billion parameter taste — parameters being the “settings” that information the neural community’s processing and responses), Phi-3 carried out on the extent of OpenAI’s GPT-3.5 mannequin, in accordance with feedback offered at the moment to VentureBeat by Sébastien Bubeck, Vice President of Microsoft generative AI.

Particularly, Phi-3 was designed to supply reasonably priced efficiency on coding, widespread sense reasoning, and common data.

It’s now an entire household consisting of 6 separate fashions with completely different numbers of parameters and context lengths (the quantity of tokens, or numerical representations of knowledge) the consumer can present in a single enter, the latter starting from 4,000 to 128,000 — with prices starting from $0.0003 USD per 1,000 enter tokens to $0.0005 USD/1K enter tokens.

See also  Reshaping Data Centres and the Digital Landscape

Nonetheless, put into the extra typical “per million” token pricing, it comes out to $0.3/$0.9 per 1 million tokens to begin, precisely double OpenAI’s new GPT-4o mini pricing for enter and about 1.5 instances as costly for output tokens.

Phi-3 was designed to be secure for enterprises to make use of with guardrails to cut back bias and toxicity. Even again when it was first introduced, Microsoft’s Bubeck promoted its functionality to be fine-tuned for particular enterprise use circumstances.

“You may usher in your information and fine-tune this common mannequin, and get wonderful efficiency on slim verticals,” he informed us.

However at that time, there was no serverless choice to fine-tune it: in case you wished to do it, you needed to arrange your individual Microsoft Azure server or obtain the mannequin and run it by yourself native machine, which can not have sufficient house.

Serverless fine-tuning unlocks new choices

At this time, nevertheless, Microsoft introduced most of the people availability of its “Fashions-as-a-Service (serverless endpoint)” in its Azure AI development platform.

It additionally introduced that “Phi-3-small is now accessible through a serverless endpoint so builders can rapidly and simply get began with AI growth with out having to handle underlying infrastructure.”

Phi-3-vision, which might deal with imagery inputs “will quickly be accessible through a serverless endpoint” as nicely, in accordance with Microsoft’s weblog publish.

However these fashions are merely accessible “as is” via Microsoft’s Azure AI growth platform. Builders can construct apps atop them, however they’ll’t create their very own variations of the fashions tuned to their very own use circumstances.

See also  Microsoft tops scorecard on nature in European data centres

For builders trying to try this, Microsoft says they need to flip to the Phi-3-mini and Phi-3-medium, which may be fine-tuned with third-party “information to construct AI experiences which are extra related to their customers, safely, and economically.”

“Given their small compute footprint, cloud and edge compatibility, Phi-3 fashions are nicely fitted to fine-tuning to enhance base mannequin efficiency throughout a wide range of situations together with studying a brand new ability or a job (e.g. tutoring) or enhancing consistency and high quality of the response (e.g. tone or model of responses in chat/Q&A),” the corporate writes.

Particularly, Microsoft states that the tutorial software program firm Khan Academy is already utilizing a fine-tuned Phi-3 to benchmark the efficiency of its Khanmigo for Lecturers powered by Microsoft’s Azure OpenAI Service.

A brand new value and functionality battle for enterprise AI builders

The pricing for serverless fine-tuning of Phi-3-mini-4k-instruct begins at $0.004 per 1,000 tokens ($4 per 1 million tokens), whereas no pricing has been listed but for the medium mannequin.

Whereas it’s a transparent win for builders trying to keep within the Microsoft ecosystem, it’s additionally a notable competitor to Microsoft’s personal ally OpenAI’s efforts to seize enterprise AI builders.

And OpenAI simply days in the past introduced free fine-tuning of GPT-4o mini as much as 2 million tokens per day via September twenty third, for so-called “Tier 4 and 5” users of its application programming interface (API), or those that spend a minimum of $250 or $1000 on API credit.

Coming additionally on the heels of Meta’s launch of the open supply Llama 3.1 household and Mistral’s new Mistral Giant 2 mannequin, each of which can be superb tuned for various makes use of, it’s clear the race to supply compelling AI choices for enterprise growth is in full swing — and AI suppliers are courting builders with each small and large fashions.

See also  How Chevron is using gen AI to strike oil

Source link
TAGGED: finetuning, language, Microsoft, Model, Phi3, serverless, small, unveils
Share This Article
Twitter Email Copy Link Print
Previous Article CSA Releases Comprehensive AI Model Risk Management Framework CSA Releases Comprehensive AI Model Risk Management Framework
Next Article Scopio Labs Scopio Labs Raises $42M in Funding
Leave a comment

Leave a Reply Cancel reply

Your email address will not be published. Required fields are marked *

Your Trusted Source for Accurate and Timely Updates!

Our commitment to accuracy, impartiality, and delivering breaking news as it happens has earned us the trust of a vast audience. Stay ahead with real-time updates on the latest events, trends.
FacebookLike
TwitterFollow
InstagramFollow
YoutubeSubscribe
LinkedInFollow
MediumFollow
- Advertisement -
Ad image

Popular Posts

Jabali Raises $5M in Seed Funding

Jabali, a San Francisco, CA- and Raleigh, NC-based AI firm advancing a generative AI recreation…

March 7, 2024

ECL reveals 1 GW off-grid, hydrogen-powered AI Factory data centre

Information Heart-as-a-Service pioneer ECL, which unveiled the world’s first off-grid, sustainable, modular, built-to-suit information heart…

September 27, 2024

runQL Raises $1.6M Pre-Seed Funding Round

runQL, a Waterloo, Ontario, Canada-based tech startup enhancing how knowledge professionals and builders work with…

November 16, 2024

Half of EMEA cloud costs going to fees, but most plan to increase capacity

European organisations are nonetheless beleaguered by storage payments as 50% of all cloud storage prices…

March 15, 2024

Data Centers & AI Are Sucking Up Huge Amounts Of Renewable Energy

Join daily news updates from CleanTechnica on electronic mail. Or follow us on Google News!…

June 4, 2024

You Might Also Like

SuperCool review: Evaluating the reality of autonomous creation
AI

SuperCool review: Evaluating the reality of autonomous creation

By saad
Top 7 best AI penetration testing companies in 2026
AI

Top 7 best AI penetration testing companies in 2026

By saad
Intuit, Uber, and State Farm trial AI agents inside enterprise workflows
AI

Intuit, Uber, and State Farm trial enterprise AI agents

By saad
How separating logic and search boosts AI agent scalability
AI

How separating logic and search boosts AI agent scalability

By saad
Data Center News
Facebook Twitter Youtube Instagram Linkedin

About US

Data Center News: Stay informed on the pulse of data centers. Latest updates, tech trends, and industry insights—all in one place. Elevate your data infrastructure knowledge.

Top Categories
  • Global Market
  • Infrastructure
  • Innovations
  • Investments
Usefull Links
  • Home
  • Contact
  • Privacy Policy
  • Terms & Conditions

© 2024 – datacenternews.tech – All rights reserved

Welcome Back!

Sign in to your account

Lost your password?
We use cookies to ensure that we give you the best experience on our website. If you continue to use this site we will assume that you are happy with it.
You can revoke your consent any time using the Revoke consent button.