Calgarypuck Forums - The Unofficial Calgary Flames Fan Community

Go Back   Calgarypuck Forums - The Unofficial Calgary Flames Fan Community > Main Forums > The Off Topic Forum
Register Forum Rules FAQ Community Calendar Today's Posts Search

Reply
 
Thread Tools Search this Thread
Old 12-16-2024, 09:19 AM   #521
I-Hate-Hulse
Franchise Player
 
I-Hate-Hulse's Avatar
 
Join Date: Jul 2003
Location: Sector 7-G
Exp:
Default

Is there an AI application out there where you can upload several PDF's (about 100 pages long each) on a topic, and have AI complete a 200 page PDF questionnaire based on the previously uploaded PDF files? The questionnaire to be answered is part textboxes, and part radio buttons.

I've been trying notebook.lm, which seems to ingest the incoming PDF files OK. But it doesn't seem to support the latter half of this OK - the completion of the PDF questionnaire.

I could feed it the questions to be answered one by one into the text input box, but that's not efficient.
I-Hate-Hulse is offline   Reply With Quote
Old 01-27-2025, 07:40 AM   #522
Firebot
#1 Goaltender
 
Join Date: Jul 2011
Exp:
Default

Some pretty massive news in AI land that has some pretty incredible impact worthy of a bump, especially to American AI giants who have convinced investors that the only way to get to AGI was by throwing insane amounts of cash at the problem while keeping all the research private to profit from.

A small team in China called Deepseek, on a pet project, built an LLM thinking model that rivals o1 from Open AI with Deepseek R1. That in itself is quite the feat to see a Chinese company suddenly at the top of the charts, but what is truly remarkable is they built the model with only 6 million dollars in training cost, AND have released the papers detailing the method AND released the model as an open source model. Everyone can now have their own open source LLM thinking model if they have the computing power for it. They are also doing it on inferior chips due to the ban imposed by the US on Nvidia providing chips to China. And anyone can also use their API at a fraction of cost compared to o1 and o3 models.

https://www.reddit.com/r/LocalLLaMA/...ane/?rdt=33050

This is an absolute game changer, and the market is in full panic mode.

NVDA is down 12% in one blow. NASDAQ down 3% so far already. This may well just crash the AI bubble (for the better IMO) as investors wake up to realize they were sold snake oil by salesmen who just lost their secret.

I tried it for myself very briefly, it built a working version of Tetris in one simple prompt in Python.

https://www.deepseek.com/

If you want to try it out. (yes it's heavily censored if you are asking it for China sensitive political topics, but that should be given).

Last edited by Firebot; 01-27-2025 at 07:52 AM.
Firebot is offline   Reply With Quote
The Following 11 Users Say Thank You to Firebot For This Useful Post:
Old 01-27-2025, 08:01 AM   #523
Fuzz
Franchise Player
 
Fuzz's Avatar
 
Join Date: Mar 2015
Location: Pickle Jar Lake
Exp:
Default

Here's an article with a lot more detail(and DIY instructions):

https://www.theregister.com/2025/01/...eek_r1_ai_cot/

EDIT: sorry, text got removed.

You can test it here if you don't want to use a Chinese site:
https://lmarena.ai/?arena
Click the "direct chat" tab, then select deepseek-r1 from the first dropdown box.

It did the best job yet of my test question.
Spoiler!

Last edited by Fuzz; 01-27-2025 at 08:50 AM.
Fuzz is offline   Reply With Quote
The Following 2 Users Say Thank You to Fuzz For This Useful Post:
Old 01-27-2025, 09:29 AM   #524
Old Yeller
First Line Centre
 
Join Date: Mar 2004
Exp:
Default

Nvidia down 16% now... basically every time I refresh it goes down again.
Old Yeller is online now   Reply With Quote
Old 01-27-2025, 09:33 AM   #525
photon
The new goggles also do nothing.
 
photon's Avatar
 
Join Date: Oct 2001
Location: Calgary
Exp:
Default

I guess the question is even if true is there enough AI work out there for this to be more an adjustment for nvidia than a permanent change. Will this change things fundamentally or does this just slow nvidia's growth as everyone finds more things to do with their hardware?

I.e. buying opportunity or not.
__________________
Uncertainty is an uncomfortable position.
But certainty is an absurd one.
photon is offline   Reply With Quote
Old 01-27-2025, 10:04 AM   #526
sa226
#1 Goaltender
 
sa226's Avatar
 
Join Date: Apr 2009
Location: Back in Calgary!!
Exp:
Default

I sold my AVGO position this morning. It was a long term growth hold for me, and I was already up a whole bunch with it so I'm okay with that.

Generally its a good idea to lock in profits and average out, but I'm not a trader and don't have time to watch it that closely.

Theres just too much emotion out there right now. If it was an over reaction of the market I'll re-evaluate and see if I want to jump back in. If it keeps free falling, it could be a good buying opportunity. I'd just rather watch whatever this is from the sidelines.

Just realized this wasn't the stock market thread. Whatevs.
sa226 is online now   Reply With Quote
The Following 2 Users Say Thank You to sa226 For This Useful Post:
Old 01-27-2025, 10:25 AM   #527
Azure
Had an idea!
 
Azure's Avatar
 
Join Date: Oct 2005
Exp:
Default

Quote:
Originally Posted by Firebot View Post
Some pretty massive news in AI land that has some pretty incredible impact worthy of a bump, especially to American AI giants who have convinced investors that the only way to get to AGI was by throwing insane amounts of cash at the problem while keeping all the research private to profit from.

A small team in China called Deepseek, on a pet project, built an LLM thinking model that rivals o1 from Open AI with Deepseek R1. That in itself is quite the feat to see a Chinese company suddenly at the top of the charts, but what is truly remarkable is they built the model with only 6 million dollars in training cost, AND have released the papers detailing the method AND released the model as an open source model. Everyone can now have their own open source LLM thinking model if they have the computing power for it. They are also doing it on inferior chips due to the ban imposed by the US on Nvidia providing chips to China. And anyone can also use their API at a fraction of cost compared to o1 and o3 models.

https://www.reddit.com/r/LocalLLaMA/...ane/?rdt=33050

This is an absolute game changer, and the market is in full panic mode.

NVDA is down 12% in one blow. NASDAQ down 3% so far already. This may well just crash the AI bubble (for the better IMO) as investors wake up to realize they were sold snake oil by salesmen who just lost their secret.

I tried it for myself very briefly, it built a working version of Tetris in one simple prompt in Python.

https://www.deepseek.com/

If you want to try it out. (yes it's heavily censored if you are asking it for China sensitive political topics, but that should be given).
lol, sure.
Azure is offline   Reply With Quote
Old 01-27-2025, 10:33 AM   #528
Russic
Dances with Wolves
 
Russic's Avatar
 
Join Date: Jun 2006
Location: Section 304
Exp:
Default

Quote:
Originally Posted by Azure View Post
lol, sure.
By virtue of it being open source, it's fairly straightforward to check their work (as I understand it). I don't see too many people questioning the validity of the claims.
Russic is offline   Reply With Quote
Old 01-27-2025, 12:34 PM   #529
Azure
Had an idea!
 
Azure's Avatar
 
Join Date: Oct 2005
Exp:
Default

Quote:
Originally Posted by Russic View Post
By virtue of it being open source, it's fairly straightforward to check their work (as I understand it). I don't see too many people questioning the validity of the claims.
Quote:
DeepSeek's impact on the AI technology supply chain "all depends on the veracity of DeepSeek's claims," Truist Securities analyst William Stein said in a client note Monday.

DeepSeek alleges that it built its AI model with "with inferior chips and at a fraction of the cost" of U.S. models, Stein said. "We cannot determine the veracity of DeepSeek's claims. If they're true, it magnifies the already-known risk of an AI spending slowdown."

DeepSeek claims it used 2,048 Nvidia H800 chips, a downgraded version of Nvidia's H100 chips designed to comply with U.S. export restrictions. However, some industry officials are skeptical of that claim, saying DeepSeek has more Nvidia processors and more advanced versions than it is letting on.
https://www.investors.com/news/techn...deepseek-news/

I mean the training might be true, but seems pretty dumb to think that they are this far advanced with far lessor GPUs.
Azure is offline   Reply With Quote
The Following User Says Thank You to Azure For This Useful Post:
Old 01-27-2025, 01:13 PM   #530
opendoor
Franchise Player
 
Join Date: Apr 2007
Exp:
Default

Quote:
Originally Posted by Azure View Post
https://www.investors.com/news/techn...deepseek-news/

I mean the training might be true, but seems pretty dumb to think that they are this far advanced with far lessor GPUs.
I don't know, I don't find it that implausible, particularly with how a bunch of people in the US are panicking and talking about the CCP and Tiananmen Square to distract from it. If it wasn't a real threat to US dominance, the reaction would probably be different. I mean, look at innovation in EVs. China has a bunch of automakers all producing affordable EVs while the US has basically one company that can actually produce them, and then a bunch of dinosaurs who can't innovate. I don't see why this space would be any different given how bloated US tech companies are.

And even if you set aside how it was developed, it's clearly more efficient in terms of resources when in use. And anyone can run it locally. Hopefully this progress toward open source models continues.
opendoor is offline   Reply With Quote
The Following User Says Thank You to opendoor For This Useful Post:
Old 01-27-2025, 01:17 PM   #531
Fuzz
Franchise Player
 
Fuzz's Avatar
 
Join Date: Mar 2015
Location: Pickle Jar Lake
Exp:
Default

This was always going to happen at some point. Current models are just brute force(more data, more compute), it shouldn't be shocking that someone developed a more elegant method. It seems pretty dumb to think that the current path was going to continue to be the correct one.
Fuzz is offline   Reply With Quote
Old 01-27-2025, 01:26 PM   #532
Shazam
Franchise Player
 
Shazam's Avatar
 
Join Date: Aug 2005
Location: Memento Mori
Exp:
Default

Quote:
Originally Posted by Fuzz View Post
This was always going to happen at some point. Current models are just brute force(more data, more compute), it shouldn't be shocking that someone developed a more elegant method. It seems pretty dumb to think that the current path was going to continue to be the correct one.
There is nothing novel about DS.

It is cheaper because there is no fine tuning done by humans.

https://www.seangoedecke.com/deepseek-r1
__________________
If you don't pass this sig to ten of your friends, you will become an Oilers fan.
Shazam is offline   Reply With Quote
Old 01-27-2025, 01:33 PM   #533
Fuzz
Franchise Player
 
Fuzz's Avatar
 
Join Date: Mar 2015
Location: Pickle Jar Lake
Exp:
Default

Quote:
Originally Posted by Shazam View Post
There is nothing novel about DS.

It is cheaper because there is no fine tuning done by humans.

https://www.seangoedecke.com/deepseek-r1
Quote:
In short, this is a reinforcement learning approach, not a fine-tuning approach. There’s no need to generate a huge body of chain-of-thought data ahead of time, and there’s no need to run an expensive answer-checking model. Instead, the model generates its own chains-of-thought as it goes2. There are other points made in the DeepSeek-R1 paper, but I think this is by far the most important.


Quote:
Addendum: this is a relatively straightforward approach that others must have thought of. Why did it happen now and not a year ago? The most compelling answer is probably this: open-source base models had to get good enough at reasoning that they could be RL-ed into becoming reasoning models. It’s plausible that a year ago that wasn’t the case. A less compelling answer: the quality of reasoning-based benchmarks is much higher now than it was. For this approach to work, you need to be able to feed the model a ton of problems that require reasoning to solve (otherwise it’ll jump straight to the solution). Maybe those problems have only recently become available.

Thanks for the link. Sounds like it is a bit novel, as they used a different feedback mechanism that doesn't appear to have caused any major issues. This will save a lot of training/money and I'd expect it to get refined further.
Fuzz is offline   Reply With Quote
Old 01-27-2025, 01:57 PM   #534
Jason14h
Franchise Player
 
Join Date: Oct 2005
Exp:
Default

It’s not just this model - it’s the breakthrough that people thought was months or years away minimum

This new method will be improved and iterated on now which will reduce the resources required to train and run . Thus reducing the chips people need to buy

This stuff was already moving exponentially fast , so a breakthrough like this can cause exponential ripples
Jason14h is offline   Reply With Quote
Old 01-27-2025, 02:45 PM   #535
activeStick
Franchise Player
 
activeStick's Avatar
 
Join Date: Jan 2014
Exp:
Default

And it's been hit by a cyber attack - my guess is by the US given its impact on the American AI industry - within hours of its release.

https://www.kron4.com/news/national/...-cyber-attack/

Quote:
Chinese AI startup DeepSeek announced Monday that it had been hit by a large-scale cyberattack. The one-year-old startup launched its R1#AI model last week and has since drawn comparisons to OpenAI.

On Monday, DeepSeek announced that it would be temporarily limiting registrations to “ensure continued service.” Existing users, according to DeepSeek, are still able to log in as usual.
activeStick is offline   Reply With Quote
The Following 2 Users Say Thank You to activeStick For This Useful Post:
Old 01-27-2025, 02:58 PM   #536
Fuzz
Franchise Player
 
Fuzz's Avatar
 
Join Date: Mar 2015
Location: Pickle Jar Lake
Exp:
Default

I wonder if it was a cyber attack, or just everyone trying to check it out.
Fuzz is offline   Reply With Quote
Old 01-28-2025, 09:23 AM   #537
Russic
Dances with Wolves
 
Russic's Avatar
 
Join Date: Jun 2006
Location: Section 304
Exp:
Default

To me what's most interesting about this is that it was ultimately USA's restriction on chips that ended up sparking the necessity of China finding a different way around the problem. Life finds a way.
Russic is offline   Reply With Quote
The Following 2 Users Say Thank You to Russic For This Useful Post:
Old 01-28-2025, 10:38 AM   #538
The Yen Man
Franchise Player
 
The Yen Man's Avatar
 
Join Date: Feb 2006
Location: Calgary
Exp:
Default

I totally get questioning the actual $6.5M price tag, but even if it was say 10X that price, that's still way more efficient than what current AI models were saying it would cost.

China new exactly what they were doing putting this out as open source. They wanted to make sure their model was accessible to anyone, and screw over the big dog US tech companies as much as possible. They basically threw a grenade into the US tech stock market yesterday.

Let's see if this is a temporary blip or something permanent. I'll be the first to admit I don't know enough about AI to know how serious of a threat this is to the US tech giants. But anything that disrupts all those mega corps is ok in my books. They are getting way too big in size and influence.
The Yen Man is offline   Reply With Quote
The Following User Says Thank You to The Yen Man For This Useful Post:
Old 01-28-2025, 10:53 AM   #539
Fuzz
Franchise Player
 
Fuzz's Avatar
 
Join Date: Mar 2015
Location: Pickle Jar Lake
Exp:
Default

There is also no reason the other models can't adopt this technique, and still leverage their extra power and model sizes. Will be interesting to see if that is another step improvement.
Fuzz is offline   Reply With Quote
Old 01-28-2025, 11:32 AM   #540
BoLevi
First Line Centre
 
Join Date: Mar 2019
Exp:
Default

I think people are overstating the geopolitical aspects of this.

AI's biggest innovation will be getting commoditization. The Chinese are pretty good at that.
BoLevi is offline   Reply With Quote
Reply

Tags
they will overtake us


Posting Rules
You may not post new threads
You may not post replies
You may not post attachments
You may not edit your posts

BB code is On
Smilies are On
[IMG] code is On
HTML code is Off

Forum Jump


All times are GMT -6. The time now is 02:42 PM.

Calgary Flames
2024-25




Powered by vBulletin® Version 3.8.4
Copyright ©2000 - 2025, Jelsoft Enterprises Ltd.
Copyright Calgarypuck 2021 | See Our Privacy Policy