Prices of the electricity we use to charge

Peter.Bridge

Esteemed Pedelecer
Apr 19, 2023
1,663
749
Thanks, I try to watch anything by her on TV and read her books, I haven't read Emperor of Rome yet. I listened to 12 Caesars as an audiobook, but I think it is a more degree level academic book and assumes a better academic knowledge than I have got ! Even so it was very interesting as to how history is written and how things get attributed - nothing is set in stone ! I've read SPQR and Women & Power : a manifesto and she argues very persuasively
 
G

Ghost1951

Guest
Thanks, I try to watch anything by her on TV and read her books, I haven't read Emperor of Rome yet. I listened to 12 Caesars as an audiobook, but I think it is a more degree level academic book and assumes a better academic knowledge than I have got ! Even so it was very interesting as to how history is written and how things get attributed - nothing is set in stone ! I've read SPQR and Women & Power : a manifesto and she argues very persuasively
Emperor of Rome is the first of her books I have owned. It is a signed copy too :). The story of how I came by it is up above in the thread, and I brought it into the thread because of the parallels with ancient and modern men of Imperial bent. When I have read it I will likely get others. I've read a lot of her articles and watched every series I can find of hers. This one is new to me. I'm not sure how many of them are on Youtube, I am part way into the second, and only stopped watching it this morning because I had to get on the road and travel.

I have said many times on here that where I spend most of my time has deep Roman connections. I can walk out of my back door and walk a mile along a track and arrive at Hadrian's Wall. The Museum and dig at Vindolanda is only a short e-bike ride. The museum there is stuffed with fascinating things they have dug up: worn out shoes and boots, women's sandals that look like they could have been bought on a modern high street and discarded after wear and tear. The most striking stuff can be found in the recovered builders tool section. A builder's trowel is a builder's trowel whether it was bought at B&Q or dug up from a rubbish dump in Vindolanda. No change - literally none. Hammers and pry bars the same. Pincers too.

63366

This Roman women's shoe found in a well in Germany

63367

In that second vid - the one linked above, Beard spends some time looking at roads and aqueducts. They built amazing things that still stand. Elsewhere, I have read about the Pantheon - the building with what I think is still the largest concrete dome in the world.

63364


63365

63369

I wonder how long our architecture will last by comparison.


When she was getting the book signed, my partner said M Beard was a very accessible person who listened carefully to what people had to say and showed an interest in them. Not a bit of arrogance or self importance. Rare - I'd say.
 
Last edited by a moderator:
  • Like
Reactions: Peter.Bridge

guerney

Esteemed Pedelecer
Sep 7, 2021
12,022
3,367
My Rysen 7 processor never gets above about 60c. It has good, silent fans though. I don't have a gpu, and even so, an average query might mean i need to wait ten seconds for it to understand and begin answering. It produces output faster than I can read it.

I think it is true that Rysen orocessors are less power hungry than equivalent Intel ones.
Which Ryzen 7? How much RAM? Which Deepseek R1? How many tokens per second? I might be becoming curious enough to assemble an AMD system with RTX 3060 12GB - it'd be useful for video rendering, AI tinkerings wouldn't be my main reason.
 
Last edited:

Woosh

Trade Member
May 19, 2012
21,258
17,268
Southend on Sea
wooshbikes.co.uk
Which Ryzen 7? How much RAM? Which Deepseek R1? How many tokens per second? I might be becoming curious enough to assemble an AMD system with RTX 3060 12GB - it'd be useful for video rendering, AI tinkerings wouldn't be my main reason.
If you have not started, don't. You will need tools, installing a home kit is time consuming, even for someone who knows docker well. Just go online. You can get chatGPT to write a shell script for the installation of your home kit but in general, those scripts never work without tinkering.
 

guerney

Esteemed Pedelecer
Sep 7, 2021
12,022
3,367
Might be fun in the idle hours. The 32b model looks potentially useful. I've got to crack the stock market to make millions somehow.
 
G

Ghost1951

Guest
Which Ryzen 7? How much RAM? Which Deepseek R1? How many tokens per second? I might be becoming curious enough to assemble an AMD system with RTX 3060 12GB.
This is the machine I bought. Rysen 7, 5825U, 32 Gbyts DDR4 Ram, 1Tbyt Hdd. The Nucbox was about £350 when I bought it, but the price on Amazon has risen significantly. They do go up and down, maybe related to your cookies. Dynamic pricing, it is called. This is what I bought at Amazon:


There are options on that page. I don't recommend less than 32Gbyts of Ram. You might not require a terabyte of memory. I have used nowhere near that, but have it anyway. The models are pretty massive. If you wanted a lot of them, you might use all of that.

I have a deepseek model on it, but I don't use that one, much preferring the Gemma3 Google models. I don't want to wait while the machine produces reams of reasoning which deepseek seems to do. Maybe that can be turned off. The Gemma 3 model seems more capable and accurate for the queries I have asked of it. It does creative writing well for a start. Other people have different requirements and prefer other models.

The mini PC will handle the mid sized Gemma model quite well, I think around 12 Bn parameters. The 27 Bn one is too big to run at acceptable speed. I've tried a few models and they have different specialities in their training so what I like, may not suit you.

Ollama has loads of models of various sizes that you can download - all free. The free stuff is a genuine Pro bono gift from the big tech companies. These are mot monetised at all.

There are models from META, MIcroSoft, Deepseek, and Google and maybe more. Huggingface is another massive receptacle for free models of all kinds.

I DO NOT CLAIM that this size of PC is the best way to run AI models. It certainly isn't. One of my sons runs 27Bn parameter models on a games machine with a reasonable GPU. The price of the GPU alone would surpass the price of this mini PC.

That setup he has, is fast and furious. It gives answers quicker than ChatGPT even when accessed remotely. The reason is of course that ChatGPT is dealing with thousands of queries a second and his isn't. Also, the ChatGPT models run in the unpaid versions with 175 billion parameters, not 27Bn and if you pay they are MUCH bigger and more capable.

With my set up, you need to wait a few seconds after sending a query while the machine works out what you want. The initial query in a session comes back quickly - probably around four or five seconds, depending how you access it. Once it starts answering your query, the data comes at faster than my comfortable reading speed. It sort of streams onto your screen like an old telytype machine.

If you run Ollama in a terminal window, everything is faster. I tend to use it from my chromebook, via a front end programme which gives an interface like the Chat GPT one. This adds overhead.

For networking the AI, as front ends go, OpenWeb UI is faster than the Page Assist UI, because it seems to send your query character by character as you are typing it (that's what ChatGPT does) and this gives the AI model a head start in working out what you want it to do, as opposed to a UI which waits until you have typed the whole query and then sends the whole thing.

I found that Windows Defender fought against Openwebui and the Page assist UI was easier to get going on the network at home. A linux environment is probably better in that respect.

In a particular session of question and answer, the response time gradually increases because what is known as the 'context' builds up. This is all part of the new questions that are submitted to the AI model in a long session. It is how it keeps hold of the whole discussion you are having with it. So it needs each time you send a question, to look back over the whole conversation and start again trying to contextualise and understand what has been said. The latest question or response you send is only a part of the whole discussion, so it needs to re-process that too, and that will soon mean you are waiting maybe a minute from adding some question to the session for an answer to start coming back. You likely know all of this, but is might be an issue with a smaller machine like this, if you need a long dialogue with the AI rather than a quick single question or instruction to do something. This is much less apparent if I access my son's machine remotely. I don't notice the impact of the context building up.

I found an easy way around the buildup of context and slowing down. Just ask a single question in a particular instance of the model, unless you need to build up context over a longer session of dialogue.

I got into this because I wanted to get a grip on how these models do what they do and this is why I bought the windows Mini computer. I wasn't prepared to spend £1500, so this was my compromise. It was a learning opportunity for me. This technology will revolutionise work and much of our employment economy in the next decade so I thought I should understand what it is, how it works, and what it can do.

Windows caused me no end of trouble in networking access. Windows Defender obstructs attempts to come into the machine from outside. I think you are an IT professional Guerney and if so, you won't be phased, but I had to put some time into sorting this stuff out. My son runs his Ollama set up in a Linux environment. If you have background in that, I think it would be easier to get going than mine was, but I have a lot of spare time which you don't. It was only networking the AI around my home network that was a problem in Windows and it wasn't insurmountable. I easily loaded Ollama, got models downloaded and could run them in a terminal window on the PC. It was only getting it to work with my chromebook that was a bit of a trouble.

The number of tokens per second it handles varies according to the size of the models. Big models run slower than small ones. Ollama has models of all kinds of sizes. I had to find a compromise between quality with speed. 4Bn parameter Gemma 3 runs about 15 tokens per second and 12Bn drops to about 7 tk per second. The full games machine with dedicated GPU my son uses is very quick. The bigger models are less likely to give hallucinating answers or factual errors. 8bn parameters is a good compromise.

The attraction of running models on your own machine is that once you download the model, you have it forever and can run entirely away from the internet. Some commercial users have concern about the privacy of their data and information in queries and reports. My son sent me redacted samples of reports he is running on his system, with attached private client database and library of commercially sensitive documents, and client system data which relating to his role as a tech consultant. For work like that, an offline system is essential because of client sensitivity. He reckons the system he has constructed using qwen3:14b and qdrant vector store, can save him days of work per week in responding to client queries in his job. This makes him much more productive. He works for a big tech consultancy firm and deals with very large companies. He gets paid a lot. They expect a lot in return.

The box I bought is quite small.

63368
 
Last edited by a moderator:
G

Ghost1951

Guest
If you have not started, don't. You will need tools, installing a home kit is time consuming, even for someone who knows docker well. Just go online. You can get chatGPT to write a shell script for the installation of your home kit but in general, those scripts never work without tinkering.
It just installed no problem for me on a Windows system and my son found the same on a Linux one. Getting Ollama to run was VERY easy. Networking it all from Windows 11 - not so much.

Start here if you want to :

 

Woosh

Trade Member
May 19, 2012
21,258
17,268
Southend on Sea
wooshbikes.co.uk
It just installed no problem for me on a Windows system and my son found the same on a Linux one. Getting Ollama to run was VERY easy. Networking it all from Windows 11 - not so much.

Start here if you want to :

Ollama isn't a problem, it's just the frame. To use ai like online, you have to install tools, Web scraping, RAG vector database, email reader etc. Also, reasoning models are large. If you use it for research ( my wife uses it for her PhD), you need tools to access academic archives. Perplexity lets you use their stuff for free. I use aistudio.google. for video and images, you'll need to install cuda, pytorch, diffusion engine etc.
 
Last edited:
G

Ghost1951

Guest
If you have not started, don't. You will need tools, installing a home kit is time consuming, even for someone who knows docker well. Just go online. You can get chatGPT to write a shell script for the installation of your home kit but in general, those scripts never work without tinkering.
Docker does seem to be a bugg er to mess with. Ollama loaded it and it just worked with the models. I don't know what it is doing other than allowing Ollama to run in windows in some sort of virtual linux environment.... I know it loads every time I boot that mini pc and opens a window. I just minimise it and let it do its thing. I stopped trying to interact with it when I gave up the idea of allowing remote connections from outside the home network. I travel about a fair bit so had at one time thought I could just access my home system through cloudflare. Too mind bending for the amount of use it would get and I went off to do other things.
 
Last edited by a moderator:
  • Like
Reactions: Woosh
G

Ghost1951

Guest
Ollama isn't a problem, it's just the frame. To use ai like online, you have to install tools, Web scraping, RAG vector database, email reader etc. Also, reasoning models are large. If you use it for research ( my wife uses it for her PhD), you need tools to access academic archives. Perplexity lets you use their stuff for free. I use aistudio.google. for video and images, you'll need to install cuda, pytorch, diffusion engine etc.
My needs are obviously much less demanding than the ones you are describing.

My son uses Perplexity.When I was doing the due diligence before buying that mini pc, I asked him if he thought it would be usable. He sent me a perplexity report which was quite amazing.



EDIT:

Just noted Mikel's post above - wondering what 'Conformational bias' is.....

I know about confirmation bias and have done for fifty years, but I never heard of that. Better watch his video and find out. I like to keep up to date...

Oh no - that video was all about confirmation bias - the presenter's. He doesn't like AI in case it will maybe tend to put him out of business. Well here's the thing - he is shouting into the hurricane. AI will put a lot of people out of business and a lot of new people will have new businesses because of it. That's pretty much certain.
 
Last edited by a moderator:
  • Like
Reactions: Woosh

Woosh

Trade Member
May 19, 2012
21,258
17,268
Southend on Sea
wooshbikes.co.uk
I might be becoming curious enough to assemble an AMD system with RTX 3060 12GB - it'd be useful for video rendering, AI tinkerings wouldn't be my main reason.
This is an AI script for your projected PC to do generative videos (very slowly). You would need more VRAM but it will run with just 12GB.

# AI Video Generation Installation Script
# Requires Windows 10/11 with NVIDIA GPU

# 1. Verify NVIDIA GPU requirements
$gpuInfo = nvidia-smi --query-gpu=name,driver_version,memory.total --format=csv
if (-not $gpuInfo) {
Write-Host "Error: NVIDIA GPU not detected. Installation aborted." -ForegroundColor Red
exit
}

# 2. Install prerequisites
winget install --id Python.Python.3.10
winget install --id Git.Git

# 3. Download and install Pinokio
$pinokioUrl = "https://github.com/pinokio/pinokio/releases/latest/download/pinokio.exe"
$installPath = "$env:USERPROFILE\Pinokio"
Invoke-WebRequest -Uri $pinokioUrl -OutFile "$env:TEMP\pinokio.exe"
Start-Process -Wait -FilePath "$env:TEMP\pinokio.exe" -ArgumentList "/S /D=$installPath"

# 4. Configure Stable Video Diffusion
$config = @{
"name" = "stable-video-diffusion"
"version" = "1.1"
"dependencies" = @(
@{
"type" = "git"
"url" = "https://github.com/Stability-AI/generative-models"
}
)
}

$configPath = "$installPath\configs\stable-video.json"
$config | ConvertTo-Json | Out-File $configPath

# 5. Run installation
Start-Process -FilePath "$installPath\pinokio.exe" -ArgumentList "install $configPath"

Write-Host "`nInstallation complete! Launch Pinokio from Start Menu to begin video generation."


Or with Sora:

# For Linux/Windows WSL users with 24GB+ VRAM
conda create -n opensora python=3.10 -y
conda activate opensora
git clone https://github.com/hpcaitech/Open-Sora
cd Open-Sora
pip install -v . xformers==0.0.27.post2 flash-attn
huggingface-cli download hpcai-tech/Open-Sora-v2 --local-dir ./ckpts
 

guerney

Esteemed Pedelecer
Sep 7, 2021
12,022
3,367
This is an AI script for your projected PC to do generative videos (very slowly). You would need more VRAM but it will run with just 12GB.

# AI Video Generation Installation Script
# Requires Windows 10/11 with NVIDIA GPU

# 1. Verify NVIDIA GPU requirements
$gpuInfo = nvidia-smi --query-gpu=name,driver_version,memory.total --format=csv
if (-not $gpuInfo) {
Write-Host "Error: NVIDIA GPU not detected. Installation aborted." -ForegroundColor Red
exit
}

# 2. Install prerequisites
winget install --id Python.Python.3.10
winget install --id Git.Git

# 3. Download and install Pinokio
$pinokioUrl = "https://github.com/pinokio/pinokio/releases/latest/download/pinokio.exe"
$installPath = "$env:USERPROFILE\Pinokio"
Invoke-WebRequest -Uri $pinokioUrl -OutFile "$env:TEMP\pinokio.exe"
Start-Process -Wait -FilePath "$env:TEMP\pinokio.exe" -ArgumentList "/S /D=$installPath"

# 4. Configure Stable Video Diffusion
$config = @{
"name" = "stable-video-diffusion"
"version" = "1.1"
"dependencies" = @(
@{
"type" = "git"
"url" = "https://github.com/Stability-AI/generative-models"
}
)
}

$configPath = "$installPath\configs\stable-video.json"
$config | ConvertTo-Json | Out-File $configPath

# 5. Run installation
Start-Process -FilePath "$installPath\pinokio.exe" -ArgumentList "install $configPath"

Write-Host "`nInstallation complete! Launch Pinokio from Start Menu to begin video generation."


Or with Sora:

# For Linux/Windows WSL users with 24GB+ VRAM
conda create -n opensora python=3.10 -y
conda activate opensora
git clone https://github.com/hpcaitech/Open-Sora
cd Open-Sora
pip install -v . xformers==0.0.27.post2 flash-attn
huggingface-cli download hpcai-tech/Open-Sora-v2 --local-dir ./ckpts
72GB VRAM or more would speed that up a bit. Some AMD graphics cards have 24GB => 144GB VRAM. ASRock motherboards used to be short lived trash. I had one with both PCI-E and AGP for 4 monitors, 'twas the shortest lived brand new motherboard I've ever owned. They claim to have improved quality since.

https://www.asrock.com/MB/Intel/Q270%20Pro%20BTC+/

 
Last edited:

Woosh

Trade Member
May 19, 2012
21,258
17,268
Southend on Sea
wooshbikes.co.uk
72GB VRAM or more would speed that up a bit. ASRock motherboards used to be short lived
You don't need an extreme cpu but two RTX 5090 32GB to have 72gb vram. That's serious money even in shenzhen. £5k-£6k would do it.
 

guerney

Esteemed Pedelecer
Sep 7, 2021
12,022
3,367
You don't need an extreme cpu but two RTX 5090 32GB to have 72gb vram. That's serious money even in shenzhen. £5k-£6k would do it.
Deepseek telling me what I forgot to remember when I've walked into a room, doesn't seem worth the cost. 6 X 5090 => 192GB VRAM would be a fun side project for someone, but I doubt Youtube views will pay enough.



Ollama isn't a problem, it's just the frame. To use ai like online, you have to install tools, Web scraping, RAG vector database, email reader etc. Also, reasoning models are large. If you use it for research ( my wife uses it for her PhD), you need tools to access academic archives. Perplexity lets you use their stuff for free. I use aistudio.google. for video and images, you'll need to install cuda, pytorch, diffusion engine etc.
You won't get it in the neck from your missus, if you're not responsible for running it locally and it ruins her PhD with hallucinations.
 
Last edited: