Gemini

Discussion of anything and everything relating to chess playing software and machines.

Moderator: Ras

User avatar
towforce
Posts: 12277
Joined: Thu Mar 09, 2006 12:57 am
Location: Birmingham UK
Full name: Graham Laight

Re: Gemini

Post by towforce »

Werewolf wrote: Thu Oct 24, 2024 10:44 pmChatGPT 4O - failed.
ChatGPT 40 with canvas - failed.
ChatGPT-01 Mini - passed first time and gave -4. I can't test ChatGPT-01 (Preview) but presumably it would also pass.

Claude 3.5 Sonnet (New) - failed.

Thanks for those test results.

Chatbot versioning numbering seems to be a bit mad (but that will be the humans - not the bot).

I thought the following progression was mad: ChatGPT 3.5 => ChatGPT 4o => GPT o1

Claude has just topped it! Sonnet 3.5 => Sonnet 3.5 (New)

People say the new Sonnet is a lot better than the old one - so why doesn't it get a new version number??? :?
Want to attract exceptional people? Be exceptional.
Werewolf
Posts: 1983
Joined: Thu Sep 18, 2008 10:24 pm

Re: Gemini

Post by Werewolf »

I've now tested the problem "What is the smallest integer whose square is between 5 and 17?" thoroughly.
If any AI failed, rather than give leading prompts I just said "this is incorrect" and let it try again.

These are my (somewhat surprising) results:

ChatGPT-01 (preview) - after thinking for 3 seconds it answered 3. After being told it was wrong, it thought for 21 seconds and answered -4.
ChatGPT-01 Mini - correct first time.
ChatGPT-4O - failed despite multiple goes.

Gemini Advanced - failed despite multiple goes.

Claude 3.5 Sonnet (New) - failed despite multiple goes.

Nvidia's 70B model - failed despite multiple goes.

Grok 2 - not tested.

Meanwhile this news just out...maybe I'll get to have working Connect 4 program by Christmas and Chess program in 2025!
https://www.tomshardware.com/tech-indus ... than-gpt-4
User avatar
towforce
Posts: 12277
Joined: Thu Mar 09, 2006 12:57 am
Location: Birmingham UK
Full name: Graham Laight

Re: Gemini

Post by towforce »

Werewolf wrote: Fri Oct 25, 2024 8:57 pmMeanwhile this news just out...maybe I'll get to have working Connect 4 program by Christmas and Chess program in 2025!
https://www.tomshardware.com/tech-indus ... than-gpt-4

Very interesting. If it's 100x larger that ChatGPTo4, then we might possibly get a hint as to whether we're beginning to see diminishing returns for more computing resource yet.

Some people are saying that it won't be available for free, and even that it might be somewhat expensive (probably more than today's standard of $20 per month for the top chatbots). For companies that can use it to replace staff, it will be well worth the cost, though.

The rate of change is increasing rapidly, and was already "too fast" in 1970, when "Future Shock" was published, the global best-selling book, was published. The effects described in that book are now on steroids! Here's the documentary film Orson Welles made about the book:


Want to attract exceptional people? Be exceptional.
User avatar
towforce
Posts: 12277
Joined: Thu Mar 09, 2006 12:57 am
Location: Birmingham UK
Full name: Graham Laight

Re: Gemini

Post by towforce »

Gemini is about to be enhanced from "chatbot" to "agent working for you":


Want to attract exceptional people? Be exceptional.
User avatar
towforce
Posts: 12277
Joined: Thu Mar 09, 2006 12:57 am
Location: Birmingham UK
Full name: Graham Laight

Re: Gemini

Post by towforce »

@Werewolf - here's another one you could try if you wish - Bolt: builds and deploys entire web apps.

https://bolt.new/


Want to attract exceptional people? Be exceptional.
User avatar
towforce
Posts: 12277
Joined: Thu Mar 09, 2006 12:57 am
Location: Birmingham UK
Full name: Graham Laight

Re: Gemini

Post by towforce »

Video summarises research into how knowledge is organised in an LLM. The second half of the video goes waaaaay too far in comparing LLMs with a human brain (human brains have an inbuilt mechanisms for learning through play and experimentation, and humans learn a small amount of knowledge in depth while LLMs and chess nets are learning a very large amount of shallow knowledge). However, I still strongly recommend watching to understand the similarities of the emergent knowledge structures:


Want to attract exceptional people? Be exceptional.
Pedro
Posts: 29
Joined: Mon Oct 26, 2020 3:05 pm
Full name: Pedro

Re: Gemini

Post by Pedro »

Elon Musk is doubling the world's largest AI GPU cluster — expanding Colossus GPU cluster to 200,000 'soon,' has floated 300,000 in the past
News


By Mark Tyson published October 29, 2024
xAI Colossus AI supercomputer continues to grow at a very fast pace

https://www.tomshardware.com/pc-compone ... n-the-past


Billionaire Elon Musk has taken to Twitter / X to boast that his remarkable xAI data center is set to double its firepower “soon.” He was commenting on the recent video exposé of his xAI Colossus AI supercomputer. In the highlighted video, TechTuber ServeTheHome was stunned when he saw the gleaming rows of Supermicro servers packed with 100,000 state-of-the-art Nvidia enterprise GPUs.

So, the xAI Colossus AI supercomputer is on course “Soon to become a 200k H100/H200 training cluster in a single building.” Its 100,000 GPU incarnation, which only just started AI training about two weeks ago, was already notable. While we think “soon” might indeed be soon in this case. However, Musk’s prior tech timing slippages (e.g., Tesla's full self-driving, Hyperloop delays, SolarCity struggles) mean we should be generally cautious about his forward-looking boasts.

The xAI Colossus has already been dubbed an engineering marvel. Importantly, praise for the supercomputer’s prowess isn’t limited to the usual Musk toadies. Nvidia CEO Jensen Huang also described this supercomputer project as a “superhuman” feat that had “never been done before.” xAI engineers must have worked very hard and long hours to set up the xAI Colossus AI supercomputer in 19 days. Typically, projects of this scale and complexity can take up to four years to get running, indicated Huang.

What will the 200,000 H100/H200 GPUs be used for? This very considerable computing resource will probably not be tasked with making scientific breakthroughs for the benefit of mankind. Instead, the 200,000 power-hungry GPUs are likely destined to train AI models and chatbots like Grok 3, ramping up the potency of its machine learning distilled ‘anti-woke’ retorts.

This isn’t the hardware endgame for xAI Collosus hardware expansion, far from it. Musk previously touted a Colossus packing 300,000 Nvidia H200 GPUs throbbing within.

At the current pace of upgrades, we could even see Musk Tweeting about reaching this 300,000 goal before 2024 is out. Perhaps, if anything delays ‘Grok 300,000,’ it could be factors outside of Musk’s control, like GPU supplies. We have also previously reported that on-site power generation had to be beefed up to cope even with stage 1 of xAI's Colossus, so that’s another hurdle - alongside complex liquid cooling and networking hardware.
smatovic
Posts: 3175
Joined: Wed Mar 10, 2010 10:18 pm
Location: Hamburg, Germany
Full name: Srdja Matovic

Re: Gemini

Post by smatovic »

Pedro wrote: Mon Nov 04, 2024 8:36 pm Elon Musk is doubling the world's largest AI GPU cluster — expanding Colossus GPU cluster to 200,000 'soon,' has floated 300,000 in the past
News

[...]
Oh boy,

Titan, supercomputer from 2012 to 2019 had 18,688 K20X GPUs:
https://en.wikipedia.org/wiki/Titan_(supercomputer)

Summit, supercomputer from 2019 has 27,648 V100 GPUs:
https://en.wikipedia.org/wiki/Summit_(supercomputer)

Frontier, first exaFLOP, #1 (top500) supercomputer from 2022 has 37,888 MI250X GPUs:
https://en.wikipedia.org/wiki/Frontier_(supercomputer)

...these AI guys are going strong :)

--
Srdja
User avatar
towforce
Posts: 12277
Joined: Thu Mar 09, 2006 12:57 am
Location: Birmingham UK
Full name: Graham Laight

Re: Gemini

Post by towforce »

Pedro wrote: Mon Nov 04, 2024 8:36 pm Elon Musk is doubling the world's largest AI GPU cluster — expanding Colossus GPU cluster to 200,000 'soon,' has floated 300,000 in the past
News

If Mr Musk building this to get ahead in the chatbot and humanoid robot game, then presumably GPUs must offer more bang per buck than TPUs - even when the arithmetic precision is of no value.

This is surprising.
Want to attract exceptional people? Be exceptional.
Pedro
Posts: 29
Joined: Mon Oct 26, 2020 3:05 pm
Full name: Pedro

Re: Gemini

Post by Pedro »

towforce wrote: Tue Nov 05, 2024 10:03 am
Pedro wrote: Mon Nov 04, 2024 8:36 pm Elon Musk is doubling the world's largest AI GPU cluster — expanding Colossus GPU cluster to 200,000 'soon,' has floated 300,000 in the past
News

If Mr Musk building this to get ahead in the chatbot and humanoid robot game, then presumably GPUs must offer more bang per buck than TPUs - even when the arithmetic precision is of no value.

This is surprising.
Skynet is coming...

:D

Image