The AI Book
    Facebook Twitter Instagram
    The AI BookThe AI Book
    • Home
    • Categories
      • AI Media Processing
      • AI Language processing (NLP)
      • AI Marketing
      • AI Business Applications
    • Guides
    • Contact
    Subscribe
    Facebook Twitter Instagram
    The AI Book
    Daily AI News

    Meta releases Code Llama, a new LLM geared for programming

    24 August 2023No Comments3 Mins Read

    [ad_1]

    Head over to our on-demand library to view sessions from VB Transform 2023. Register Here


    True to the rumors and advance reports, Meta Platforms, the company formerly known as Facebook, today unveiled Code Llama, its new generative AI large language model (LLM) designed specifically for programming — and like the more general-purpose LLaMA 2, it’s open source and licensed for commercial use.

    Code Llama is “designed to support software engineers in all sectors — including research, industry, open source projects, NGOs, and businesses,” Meta says in its blog post announcing the models.

    The tool immediately becomes a major rival to OpenAI’s Codex (powered by a modified GPT-3), the Codex-powered Github Copilot from Microsoft, and other coding-specific LLM assistants such as Stack Overflow’s OverflowAI.

    In its blog post, Meta explains that Code LlaMA is a “code-specialized” version of LLaMA 2 that can generate code, complete code, create developer notes and documentation, be used for debugging, and more. It supports Python, C++, Java, PHP, Typescript (Javascript), C# and Bash. You can read the full research paper from Meta about its performance here, which describes Code LlaMA as a “family” of LLMs for code.

    Event

    VB Transform 2023 On-Demand

    Did you miss a session from VB Transform 2023? Register to access the on-demand library for all of our featured sessions.

     

    Register Now

    Building on that analogy, the family includes three main members: a 7-billion, a 13-billion and a 34-billion parameter model, each trained on 500 billion tokens. The smaller models are designed to run on fewer GPUs (the 7-billion model can run on a single one), a beneficial attribute given the rumored scarcity in this critical piece of hardware at the moment, and Meta says both are faster than its 34-billion big model.

    All models support up to 100,000 tokens for their prompts. This means “users can provide the model with more context from their codebase to make the generations more relevant,” according to Meta.

    The LLaMA extended family also includes two fine-tuned models, one for Python and one for Instruct, the latter of which “has [been] fine-tuned to generate helpful and safe answers in natural language,” and therefore, Meta says, should be used when generating new code from natural language prompts. That is, it returns safer, more expected and perhaps less creative responses.

    You can download Code LlaMA directly from Meta here and find the source code on Github here.

    VentureBeat’s mission is to be a digital town square for technical decision-makers to gain knowledge about transformative enterprise technology and transact. Discover our Briefings.

    [ad_2]

    Source link

    Previous ArticleModular looks to boost AI mojo with $100M funding raise
    Next Article The Myth of ‘Open Source’ AI
    The AI Book

    Related Posts

    Daily AI News

    Adobe Previews New GenAI Tools for Video Workflows

    16 April 2024
    Daily AI News

    Exciting Updates From Stanford HAI’s Seventh Annual AI Index Report

    15 April 2024
    Daily AI News

    8 Reasons to Make the Switch

    15 April 2024
    Add A Comment

    Leave A Reply Cancel Reply

    • Privacy Policy
    • Terms and Conditions
    • About Us
    • Contact Form
    © 2025 The AI Book.

    Type above and press Enter to search. Press Esc to cancel.