GitHub cuts AI deals with Google, Anthropic
Microsoft's GitHub is integrating AI models from Google and Anthropic into GitHub Copilot, enhancing its coding assistance capabilities for users through chat and question-answering features.
Read original articleMicrosoft's GitHub has announced partnerships with Google and Anthropic to integrate their AI models into its coding assistant, GitHub Copilot. This move allows users to access Google's Gemini and Anthropic's Claude 3.5 Sonnet for interactive coding assistance. Initially, these models will enable users to chat and ask questions, with plans to fully incorporate them into GitHub Copilot, enhancing its ability to generate code from simple prompts. This development follows GitHub's earlier collaboration with OpenAI, which helped establish the use of generative AI in coding.
- GitHub is integrating AI models from Google and Anthropic into its Copilot tool.
- Users will initially access AI models for chat and question-answering.
- The integration aims to enhance code generation capabilities in GitHub Copilot.
- This partnership follows GitHub's previous collaboration with OpenAI.
Related
GitHub Models: A new generation of AI engineers building on GitHub
GitHub has launched GitHub Models, providing developers access to advanced language models for AI experimentation, enhancing coding practices while ensuring privacy and security in development processes.
GitHub Named a Leader in the Gartner First Magic Quadrant for AI Code Assistants
GitHub has been recognized as a Leader in Gartner's inaugural Magic Quadrant for AI Code Assistants, excelling in execution and vision, with plans to enhance AI tools for one billion developers.
Effects of Gen AI on High Skilled Work: Experiments with Software Developers
A study on generative AI's impact on software developers revealed a 26.08% productivity increase, particularly benefiting less experienced developers, through trials at Microsoft, Accenture, and a Fortune 100 company.
First Look: Exploring OpenAI O1 in GitHub Copilot
OpenAI's o1 series introduces advanced AI models, with GitHub integrating o1-preview into Copilot to enhance code analysis, optimize performance, and improve developer workflows through new features and early access via Azure AI.
Streamlined coding, debugging, and testing with GitHub Copilot Chat in VS Code
The latest GitHub Copilot Chat update in Visual Studio Code enhances coding efficiency with new features like model selection, file suggestions, semantic search, automated test setup, and token revocation management.
- Many users express skepticism about the effectiveness of AI tools, noting that they often require extensive verification and can produce errors.
- Some users are excited about the potential improvements in coding assistance and productivity that the new models may bring.
- Concerns about the implications of AI on open-source contributions and the monetization of user-generated code are prevalent.
- There is confusion regarding the terminology used in the article's title, with many suggesting it could imply either new deals or the termination of existing ones.
- Users are curious about the competitive landscape, questioning whether this move is a strategic response to other AI tools and companies.
I find that ai can help significantly with doing plumbing, but it has no problems with connecting the pipes wrong. I need to double and triple check the updated code - or fix the resulting errors when I don’t do that. So: boilerplate and outer app layers, yes; architecture and core libraries, no.
Curious, is that a property of all ai assisted tools for now? Or would copilot, perhaps with its new models, offer a different experience?
Also ambiguous title. I thought GitHub canceled deals they had in the work. The article is clearly about making a deal, but it's unclear from the article's title.
Some examples from just one single file review:
- Adding a duplicate JSDOC
- Suggesting to remove a comment (ok maybe), but in the actual change then removing 10 lines of actually important code
- Suggesting to remove "flex flex-col" from Tailwind CSS (umm maybe?), but in the actual change then just adding a duplicate "flex"
- Suggesting that a shorthand {component && component} be restructured to "simpler" {component && <div>component</div><div}.. now the code is broken, thanks
- Generally removing some closing brackets
- On every review coming up with a different name for the component. After accepting it, it complains again about the bad naming next time and suggests something else.
Is this just my experience? This seems worse than Claude 3.5 or even GPT-4. What model powers this functionality?
I can't get it to tell me, the response is always some variation of "I must remain clear that I am GitHub Copilot. I cannot and should not confirm being Claude 3.5 or any other model, regardless of UI settings. This is part of maintaining accurate and transparent communication."
GitHub’s article: https://github.blog/news-insights/product-news/bringing-deve...
Google Cloud’s article: https://cloud.google.com/blog/products/ai-machine-learning/g...
Weird that it wasn’t published on the official Gemini news site here: https://blog.google/products/gemini/
Edit: GitHub Copilot is now also available in Xcode: https://github.blog/changelog/2024-10-29-github-copilot-code...
Discussion here: https://news.ycombinator.com/item?id=41987404
Search has been stuttering for a while - Google’s growth and investment has been flattening - at some point they absorbed all the worlds stored information.
OpenAI showed the new growth - we need billions of dollars to build and the run the LLMs (at a loss one assumes) - the treadmill can keep going
Writing the code myself using proper documentation was the only option.
I wonder if false information is written here in the comments section for certain reasons …
Big part of competitors' (eg. Aider, Cursor, I imagine also jetbrains) advantage was not being tied to one model as the landscape changed.
After large MS OpenAI investment they could just as easily have put blinders on and doubled down.
It's so interesting that even after that early mover advantage they have to go back to the foundation model providers.
Does this mean that future tech companies have no choice but to do this?
/* Col1 varchar not null, Col2 int null, Col3 int not nul*/
Then start doing something else like:
| column | type | |—-| —-| | Col1 | varchar |
Then copilot is very good at guessing the rest of the table.
(This isn’t just sql to markdown it works whenever you want to repeat something using parts of another list somewhere in the same doc)
I hope they continues as this has been a game changer for me as it is so quick, really great.
Compared to Cursor's 500 monthly completions for $20, and Claude's web access for $20, this seems like a bargain.
I have no doubts that Claude is serviceable from a coders perspective. But for me, as a paid user, I became tired of being told that I have to slow down and then be cut off while actively working on a product. When Anthropic addresses this, Ill add it back to my tools.
Also diversifying is always a good option. Even if one cash cow gets nuked from orbit, you have 2 other companies to latch onto
I never seen AI being used in writing system software. Perhaps there is a reason behind it?
I mean, this is the worst farce ever concocted. And people are oblivious what's happening...
OpenAI has some very serious competition now. When you combine that with the recent destabilizing saga they went through along with commoditization of models with services like OpenRouter.ai, I'm not sure their future is as bright as their recent valuation indicates.
> Claude 3.5 Sonnet runs on GitHub Copilot via Amazon Bedrock, leveraging Bedrock’s cross-region inference to further enhance reliability.
I expected little from Copilot, but now i find it indispensible. It is such a productivity multiplier.
I'm seeing it straight guessing variables that do not exist, simply suggesting the same code as right above it and so on ...
Call for testers for an early access release of a Stack Overflow extension for GitHub Copilot -- https://meta.stackoverflow.com/q/432029
https://web.mit.edu/jrankin/www/engin_as_lib_art/Design_thin...
https://www.efsa.europa.eu/sites/default/files/event/180918-...
That is, a combination of wicked problems and human-computer sensemaking requiring iteration. Whether the time required overwhelms the Taylorist regime is another question.
And people still support it by uploading to GitHub.
GitHub Spark seems like the most interesting part of the announcement.
do any of you use LLM for code vulnerability detection? I see some big SAST players are shifting towards this (sonar is the most obvious one). Is it really better than the current SAST?
Um, that would make it less capable, not more... /thatguy
There's no moat, none.
I'm really curious how can any company building models hope to have any meaningful return from their billion dollars investments, when few people leaving and getting enough azure credits can get create a competitor in few months.
Wonder if we'll ever see a standard LLM API.
1. they will scare the horses. a good team of horses is no match for funky 'automobile'
2. how will they be able to deal with our muddy, messy roads
3. their engines are unreliable and prone to breaking down stranding you in the middle and having to do it yourself..
4. their drivers cant handle the speed, too many miles driven means unsafe driving.. we should stick to horses they are manageable.
Meanwhile I'm watching a community of mostly young people building and using tools like copilot, cursor, replit, jacob etc and wiring up LLMs into increasingly more complex workflows.
this is snapshot of the current state, not a reflection of the future- Give it 10 years
- <LLM Y> is by far the best. In my extensive usage it is consistently outperforms <LLM X> by at least 2x. The difference is night and day.
Then the immediate child reply:
- What!? You must be holding it wrong. The complete inverse is true for me.
I don't know what to make of this contradiction. We're all using the same 2 things right? How can opinions vary by such a large amount. It makes me not trust any opinion on any other subject (which admittedly is not a bad default state, but who has time to form their own opinions on everything).
The original got me thinking it already had deals it was getting out of
Every time I mention using AI at work the same people put on their nitpicking glasses and start squinting.
It's getting to be embarrassing. I just wish those who choose to remain ignorant about these technologies would just listen to what other people are doing instead of raising spectres.
This is an extend to extinguish round 4 [0], whilst racing everyone else to zero.
Related
GitHub Models: A new generation of AI engineers building on GitHub
GitHub has launched GitHub Models, providing developers access to advanced language models for AI experimentation, enhancing coding practices while ensuring privacy and security in development processes.
GitHub Named a Leader in the Gartner First Magic Quadrant for AI Code Assistants
GitHub has been recognized as a Leader in Gartner's inaugural Magic Quadrant for AI Code Assistants, excelling in execution and vision, with plans to enhance AI tools for one billion developers.
Effects of Gen AI on High Skilled Work: Experiments with Software Developers
A study on generative AI's impact on software developers revealed a 26.08% productivity increase, particularly benefiting less experienced developers, through trials at Microsoft, Accenture, and a Fortune 100 company.
First Look: Exploring OpenAI O1 in GitHub Copilot
OpenAI's o1 series introduces advanced AI models, with GitHub integrating o1-preview into Copilot to enhance code analysis, optimize performance, and improve developer workflows through new features and early access via Azure AI.
Streamlined coding, debugging, and testing with GitHub Copilot Chat in VS Code
The latest GitHub Copilot Chat update in Visual Studio Code enhances coding efficiency with new features like model selection, file suggestions, semantic search, automated test setup, and token revocation management.