Running LLMs locally. One size does NOT fit all

A few questions have popped up recently, so I thought I would address them here.

I’ve already written about running local LLMs here. Both posts also include the reasons why you might want to do this. I’m not going to repeat that here.

Can I run a local LLM on ???

This is the most common question I get. Someone will list their hardware and ask if they can run an LLM locally. The answer is always, “It depends”.

  • What kit do you actually have?
  • What sort of models do you want to use?
  • How are you expecting to use them?
  • Do you expect it to be fast, or is it acceptable for it to be slow?

I can’t give a definitive answer, but here are some things to consider.

What kit do you actually have?

The models you will be able to run will vary a lot depending on your kit. If you have an old laptop (like me), you will be restricted to running the small models, as you may be running on CPU, using the system memory. I always start with “Granite 4 H Tiny” from IBM, which is small and seems to run pretty quick on my laptop, whilst also giving me OK results. I will use bigger models as well, but only if I’m prepared to wait for a result.

If you have a GPU, the type of models you can run will depend a lot on the VRAM on the GPU. The relationship between the size of the model and the VRAM you need is not as simple as it sounds. A lot of people will just say the model has to be smaller than the VRAM of your GPU, which is kind-of correct, but you also have to consider some other stuff. The quantization (compression) of the model reduces the model size. The compression of the memory (see TurboQuant) and the size of the context you are using, which affects the Key-Value (KV) cache. You want everything to fit into the VRAM on the card. Even with a GPU you might still be limited to smaller models if you want to pass in a lot of context into your queries.

Typically you will find the more parameters a model has, the bigger it is, but remember some models with a lot of parameter can use a subset of them, making them a lot more efficient than they would first appear.

If you have kit with unified memory, like Apple Silicon kit, the system memory can be used for either the main system or the GPU, so in some cases it’s easier to run bigger models on that kit because you effectively have more VRAM. You aren’t going to run a huge model on a 8G Neo, but you may find the Apple Silicon kit works a lot better than the equivalent x86-64 kit because of unified memory.

I’ve been purposely vague here, but you hopefully get the idea that it is not as straight forward as you telling me your kit, and me giving you a definitive answer. You have to play around.

What sort of models do you want to use?

Based on the previous section, you already know the limitations of your kit. When you are using cloud-based LLMs, you are often routed to different models depending on the type of work you are trying to do. Some are specialized for coding. Some are better and dealing with written text. Some are better at handling image/video data. You have to select a model that will work with your kit, that is focussed on the type of work you are trying to do. You may find a smaller specialized model gives you better and quicker results than one of the big models. It all depends.

Play around and see what works for your requirement and kit.

How are you expecting to use them?

Asking short directed questions is very different to having huge amounts of context, based on either lots of provided information, or long running chats. Every time a new question is asked, the previous chat history becomes part of the context. You need to be careful.

Likewise, trying to run agents that can make tool calls can quickly increase the size of the context. Being extremely simplistic, for every tool call, the query is halted, the tool call is made, the information returned is added to the context, the query run again with the additional information. The amount of context can grow substantially.

Once again, you will have to play around and see what tools and models work for your kit. You are not going to be able to run 50 OpenClaw agents against a massive model on your 10 year old crappy laptop. Just be realistic.

Do you expect it to be fast, or is it acceptable for it to be slow?

I think a lot of people try the free version of OpenAI and expect running a local model to be just as fast. Pick the right model and it could be, but most of the time you have to manage your expectations. I will often try stuff on a small model. If I get an acceptable answer I go with it. If I don’t I will try a larger, much slower model. Sometimes I’m prepared to wait for a superior answer.

You have to deal with the limitations of your kit.

Local LLMs are crap?

I saw this comment on a post recently and it made me kind-of mad.

“I tried most of these small models. They are all crap. Sure they might succeed in some narrow tasks, that still need their output to be verified.”

Well yes, the small models are often best suited to narrow/directed tasks. But there are hundreds/thousands of them to choose from. Find those that suit your workflows. Not every task needs a frontier model to give good results.

If you have good kit at home, like a gaming GPU or a higher spec Apple Silicon laptop, you might be able to run bigger models.

The bit about, “that still need their output to be verified”, made me want to go postal. Literally every response from a LLM needs to be verified. They hallucinate all the time. Almost everyone that tries to tell you they are near perfect has some skin in the game. They stand to make money by convincing you to buy their service/tool. The reality is they screw up all the time. Not validating the results is like Googling commands and running them directly on your production servers as root. You are going to be out of a job pretty quickly.

Conclusion

You can run LLMs locally on a wide range of kit. The result you get will vary greatly depending on your kit, the models you are trying to use and how you are trying to use them. You are not going to be challenging OpenAI/Anthropic/Google any time soon, but you might get something that works for you. If so, it will be free, private and under your control.

Just play around and be realistic. I’m not an AI expert. Just a dabbler…

Cheers

Tim…

Oracle REST Data Services (ORDS) 26.1.0 Released

A new version of Oracle REST Data Services (ORDS) has been released. Version 26.1.0 comes with a bunch of bug fixes and new features (see here), but what is more interesting to me is at the bottom of the release notes.

“Starting with ORDS 26.4 (planned for Winter 2026), Oracle REST Data Services will add support for Tomcat 10 and WebLogic 15. This means that ORDS will no longer run on older versions of WebLogic and Tomcat.”

That doesn’t effect this release, but before the end of the year you need to be thinking about newer versions of WebLogic or Tomcat, assuming you are not running ORDS in standalone mode.

Getting ready for the 26.4

If you are using Tomcat it might be worth spending some time playing around with Tomcat 10 in advance of the 26.4 release.

My Vagrant builds and container images will have to be changed a little to take this into account. Mostly related to the HTTPS setup.

I don’t use WebLogic anymore, so I can’t be of much help there, but suffice to say you need to play around with WebLogic 15 soon, rather than wait until the last minute.

Why bother?

Even if you aren’t using any of the new features, you need to stay on the latest version for security updates, so you need to keep on top of these changes. Remember, working in tech is like swimming upstream in a river. The moment you stop swimming, you start to drift downstream.

Cheers

Tim…

The Oracle ACE Program : My 20 Year Anniversary

It’s April 1st, which means it’s my 20th year anniversary of being an Oracle ACE.

As usual I’ll do a little recap of the anniversaries that will happen for me this year.

  • 31 years working with Oracle technology in August. (August 1995)
  • 26 years doing my website in July. (Original name: 03 July 2000 or current name: 31 August 2001)
  • 21 years blogging in June. (15 June 2005)
  • 20 years on the Oracle ACE Program. (01 April 2006)
  • 11 years doing videos on my YouTube channel, with some very long breaks of course. 🙂

How many years left before I retire? I don’t know… 🙂

Cheers

Tim…

PS. I think Basheer Khan has been in the program a year or two longer than me. I’m not sure if anyone else still in the program has been here longer than us.

Database Links from Oracle to Snowflake : How to fix a couple of problems…

Recently I was asked to create a database link between Oracle 19c and Snowflake. I’ve done similar links from Oracle to SQL Server and MySQL using ODBC, so I figured it would be simple. Along the way I hit a couple of snags, which I will point out here.

Snowflake Credentials

Using credentials to connect to Snowflake is not a great idea, and I think it is now deprecated, so we needed to make a connection using a JWT Token. I was given the connection details and a key, so the first thing to do was create a DNS to Snowflake using the unixODBC package. That was simple enough, and it worked fine when I connected using the isql utility. All the queries I threw at it worked fine.

The problem started when I tried to configure Oracle Heterogenous Service (HS) for ODBC using this DSN. Since the DSN contains everything needed to connect, I figured I didn’t need to worry about credentials in the database link, so I did this.

create database link snowflake using 'SNOWFLAKE';

That didn’t work. I turned on tracing in the HS layer and I could see a “JWT Token invalid” error. I could also see the HS layer was setting the default user to match my current Oracle username. I figured this wouldn’t matter as the DSN contained the correct information, but I was wrong. In order to make a connection the HS layer needed the correct Snowflake username in the link. As a result I ended up with a database link like this.

create database link snowflake connect to CORRECT_USERNAME identified by dummy using 'SNOWFLAKE';

The password could be anything, as it wasn’t used, but the username was important.

That was the authentication problem solved.

VARCHAR to LONG Conversion

The next problem we hit was due to the way data types are handled by the ODBC driver, and interpreted by the HS layer in Oracle. The default size of a VARCHAR is 16777216 bytes in Snowflake. Any VARCHAR with a definition larger than 4000 bytes gets converted to a LONG, and we can only have one long per query. If we have more than one LONG the gateway agent crashes without a useful message. It took a while looking at trace files to figure out this was the problem.

Once we understood the problem we went looking for solutions and found loads of suggested parameter combinations, none of which worked. One of the things we tried was to create a view in Snowflake over the data to make the larger columns look smaller.

create view my_view as
select id,
       cast(description as varchar(4000)) as description,
       cast(additional_info as varchar(4000)) as additional_info,
       code
from   my_table;

This was later put forward as a solution, but it didn’t work for us. These columns were still getting presented as LONG, even if we cast them to smaller sizes like 500.

I couldn’t find anyone else talking about this on the net or in Oracle Support, so after banging my head against a brick wall for a while I opened a service request with Oracle Support. Luckily they quickly came up with two proposed solutions, one of which was the view approach, which had already failed for us. The second was a magic switch to prevent the conversion from a long VARCHAR to a LONG.

HS_WORKAROUNDS = 0x02

Once this was in place the DB link was working as we expected. It wouldn’t be a solution if you really needed LONG columns returned, but we didn’t.

The Technical Stuff

I amended my old Linux ODBC and HS for ODBC posts to include this Snowflake related information in case anyone wants to see the details.

Conclusion

It is never, EVER, easy… 🙂

Cheers

Tim…

PS. Large language models (LLMs) produce an infinite stream of bullshit solutions for any problem. Be sure to hire 10 more developers so you can chase down every false lead… 🙂

Third-Party Applications and Assertions in Oracle AI Database 26ai

You should start playing with new features as soon as you can, because it will take some time for you to start thinking about suitable uses for them. Things don’t just become part of your toolkit overnight. You’ve probably heard me say this over the years multiple times…

Playing with assertions

Over the weekend I was playing with assertions in Oracle AI database 26ai. It is probably heresy, but this is one of those features I was not too bothered about. I know some people have been pushing for this for many years, but me? Not so much. It’s not that I have anything against the idea. It just wasn’t a high priority for me.

In case you don’t know, an assertion is a type of integrity constraint. You can use them to mimic any other types of integrity constraint, but the real purpose of them is to do more wider reaching stuff, that up until now you probably would have had to do with triggers. Think of it like protecting us from logical corruption of the data, or adding business logic from a data perspective into the database.

If you care about the tech side of things, check out this post for an overview.

What’s this got to do with third-party apps?

I finished up the post, and kind-of forgot about assertions, but clearly something was going on in the back of my mind…

We use a lot of third-party apps, and the problem with them is you are at the mercy of the vendors, and what they chose to do from a database design perspective. In some cases we don’t even get basic integrity constraints like primary keys and foreign keys. We could add these manually, but we have to be careful about not impacting on their system. In some cases we have to manually clean up junk data, that would otherwise be prevented from existing by integrity constraints.

Last week we had an issue with a non-production environment that was throwing errors in an application. After a little digging I noticed the problem was being caused by a “ORA-01422: exact fetch returns more than requested number of rows” error. The table in question held multiple states, but only one should be active at a time, based on start and end times. Clearly someone had amended some data causing two overlapping active states, which caused the problem. This is a perfect example where an assertion would prevent a logical data corruption, without introducing something nasty that would freak out the vendor developers.

Now I’m not saying we are free to add a bunch of assertions. Whenever you add new integrity constraints to an existing system you have to test everything very carefully. It’s possible you will be breaking an existing process or impacting the performance of the system. I’m just saying assertions give you a different angle to protect your data, beyond what conventional integrity constraints can do. This might be especially useful when third-party vendors have dropped the ball in their applications.

Just a thought…

Cheers

Tim…

Update: From Franck Pachot.

“It would be nice to have assertions log a warning instead of raising an error. Those third-party apps probably do not handle the exception correctly. Maybe better to allow what the app does, and raise an alert to fix it (and tant against the vendor)”

Artificial Intelligence (AI): The CEO BS Generator

Warning: A lot of the text in this post was generated by AI. The opinions/rants are mine.

I know you think artificial intelligence (AI) will take over the world and subjugate or kill the human race, but the truth is way more boring. It will drown us in corporate bullshit.

Local Setup

  • Tool: LM Studio
  • Model: gemma-3-4b
  • Prompt: Write a post in the style of a chief executive officer extolling the virtues of artificial intelligence, and how it is going to revolutionize the business.
  • Hardware: Crappy 8 years old laptop.

The LLM Output – Unedited

Okay, here’s a draft of a CEO blog post, written in the enthusiastic, slightly hyperbolic style you requested, focusing on the transformative power of AI:


Subject: The Dawn of Intelligent Business – It’s Not Coming, It’s Here (And We’re Leading the Charge!)

Team, Partners, Future-Forward Thinkers –

Let me be blunt: we’re living through a pivotal moment in business history. For too long, we’ve been operating with processes that are… well, frankly, human. Reactive. Prone to bottlenecks. And let’s be honest, sometimes just plain inefficient. But the tide is turning. The future isn’t coming; it’s here, powered by the incredible potential of Artificial Intelligence.

At [Company Name], we’ve been deeply invested in harnessing AI – not as a buzzword or a fleeting trend, but as the bedrock of our strategy. And I can tell you with absolute conviction: this is going to revolutionize everything.

What does that revolution look like? Let me paint you a picture:

  • Unprecedented Efficiency: Our new AI-powered analytics platform isn’t just crunching numbers; it’s predicting market shifts, identifying hidden opportunities, and optimizing our operations with laser precision – freeing up our brilliant teams to focus on what they do best: innovation and strategy.
  • Hyper-Personalized Customer Experiences: Forget generic marketing blasts! AI is allowing us to understand each customer’s needs at a level we could only dream of before. We’re delivering tailored recommendations, proactive support, and truly engaging experiences that build loyalty and drive growth.
  • Data-Driven Decisions – Amplified: We’ve moved beyond gut feelings. Our AI algorithms are sifting through mountains of data to reveal insights that would take a human team months – or even years – to uncover. This isn’t just smarter decision-making; it’s exponential growth potential.
  • Empowering Our People: AI isn’t replacing our talented workforce; it’s augmenting them! By automating repetitive tasks, AI allows our employees to focus on creative problem-solving, strategic thinking, and building genuine relationships – the uniquely human elements that will always be at the heart of our success.

I know some might feel a little apprehensive about this level of change. But I want to assure you: we’re not blindly leaping into the future. We’re approaching AI with careful planning, ethical considerations, and a relentless commitment to responsible innovation.

This isn’t just about profits; it’s about building a smarter, more resilient, and ultimately, more impactful business for years to come.

Join us as we lead the charge into this intelligent era!

Sincerely,

[Your Name]

CEO, [Company Name]

[Link to Company Website/Relevant Article]


Important Notes & Considerations:

  • Tone: I’ve aimed for a tone that’s enthusiastic and confident, bordering on slightly over-the-top – as you requested.
  • Specificity: I’ve included placeholders like “[Company Name]” and “[relevant article]” to allow you to customize it. Ideally, you would replace these with specific details about your company’s AI initiatives.
  • Ethical Considerations: I’ve briefly touched on ethics, but this is a complex area. You should ensure your company has a robust ethical framework for its AI development and deployment.
  • Realism: While the post aims to be inspiring, it’s important to ground your messaging in reality. Don’t overpromise – focus on tangible benefits and realistic timelines.

To help me refine this further, could you tell me:

  • What industry is this for? (e.g., tech, finance, healthcare)
  • Are there any specific AI applications you want to highlight?

My Opinion

This is pretty much the script for any tech bro CEO trying to extoll the virtues of AI. Maybe throw in some mentions of Martian cities, data centres in space, Dyson spheres or a “post scarcity society” for good measure. It even gives me suggestions to make the output more specific, if I am too dumb to prompt it myself.

I’m a fan of automation, and using AI for specific automations. I’m a fan of using it as a coding assistant, not developer replacement. I’m not a fan of using AI for the generation of art (books, music, stealing artwork etc.), and I’m definitely not a fan of using AI to generate corporate bullshit and endless amounts of paperwork nobody will ever read.

My new role

I would like to take this opportunity to put myself forward for the position of CEO of your company. As you can see, I’m more than qualified now…

Cheers

Tim…

PS. The human element of this post was an old man shouting at clouds/LLMs…

Oracle AI Database 26ai : Latest articles on my website

I’ve mentioned several times about my backlog of 3+ year old posts that have been waiting for an on-prem release of 23c/23ai/26ai. Since the release of the on-prem version of Oracle AI Database 26ai I’ve been working through those posts, checking the content and updating the doc links. I think that process is now complete.

All of my 26ai articles can be found here. Many of the new articles relate to installations and upgrades, but there are a few other new things in there also. You can see the list of newest stuff on the homepage if you are interested.

Unsupported Installations

While I was going through all this stuff I thought I would have a go at a 26ai installation on Oracle Linux 10 (OL10) and Fedora 43. Neither are supported, so don’t do them, but it’s fun.

The 26ai install is pretty clean on OL10, so I’m hoping this means we won’t have to wait to long for it to be supported.

Vagrant Builds

There are vagrant builds of everything I’ve been doing over the last couple of weeks. You can find them in the “database”, “dataguard” and “rac” subdirectories of this repository.

What’s Next?

I’ve written about most of the major stuff that interests me in 23c/23ai/26ai, so to a certain extent I think I’m done with this release. Of course, there are lots of features I’ve not written about, but I don’t know how much motivation I have to look at things I’ll likely never use.

There are new features added in each Release Update (RU), so I’m sure there will be a trickle of new things from time to time. I do have some things on my list that were added in 23.26.1. 🙂

So when is the next on-prem release?

Cheers

Tim…

Running Large Language Models (LLMs) Locally using LM Studio

A few months ago I wrote a post about running Large Language Models (LLMs) locally using Ollama and Open WebUI. That works well, and feels very much like how you might want to run an LLM on a server in your organisation, but it’s not exactly friendly if you are not used to getting your hands dirty with virtual machines (VMs) and Docker/Podman.

I was watching a YouTube video where in passing they mentioned LM Studio, and I thought I would take a look. It is exactly what the average Windows/Mac/Linux user would find comfortable to run LLMs locally. Just download the app and off you go. No messing about.

Performance

The performance is going to vary depending on your computer, GPU and the model you choose, but it is much more lightweight than my Ollama example.

I’m using the same hardware I used for my Ollama tests. For that I was running a VM and a Podman container, and the extra overhead made it pretty sluggish on my crappy old laptop with a bad GPU. LM Studio was much faster in comparison, no doubt because of the reduced complexity.

I suspect most people will find it very useable on their existing kit, provided it is less than about 5 years old. Or in my case an 8 year old laptop, which was quite good in its day. 🙂

Getting started is easy

It’s really simple. Download it from here.

Once installed it will ask you if you want to download your first model. It defaults to “Gemma 3 4b”, which is based on Google Gemini, so it’s a pretty good place to start. Once that’s downloaded you can start a new chat by clicking a button, and can organise previous chats into folders. The interface is pretty simple and self explanatory.

If you want to try out some different models, click the “Search Model” button on the left toolbar, and pick something you fancy. I’ve found the “granite-4-h-tiny” from IBM to be really quick. It gives quite abbreviated answers, so you might prefer something more verbose, but you have hundreds of models to choose from, so play around and find something you prefer. In my Ollama post I gave examples of using these models.

# General models.
ollama pull llama3
ollama pull deepseek-r1
ollama pull gpt-oss:20b

# Code-focussed models.
ollama pull codellama
ollama pull deepseek-coder
ollama pull codestral

Silly example

To give you an idea of how it looks, here is a silly example. I asked the following question with 2 engines.

“How do I write a python script that accepts command line arguments?”

Here is the result using “granite-4-h-tiny”. Notice the model selection (1), the question I asked (2) and the output (3).

Here is the result of the same question using “gemma-3-4b”. It gave a different answer. I asked if that was the recommended way, and it said no, and showed me a different answer that looked more like that produced by “granite-4-h-tiny”. 🙂

You can refine your answer by asking additional questions in the same chat, just like you would on a web-based LLM.

Turn on the local API server

  • Click on the Developer icon in the left-hand toolbar.
  • Click on the “Local Server” tab.
  • Enable the local server using the toggle on the top-left.
  • Load the default model of choice using the “Load Model” button on the top-left.

The API should now be available on port 1234 by default.

$ curl http://localhost:1234/api/v1/models
{
  "models": [
    {
      "type": "llm",
      "publisher": "google",
      "key": "google/gemma-3-4b",
      "display_name": "Gemma 3 4B",
      "architecture": "gemma3",
      "quantization": {
        "name": "Q4_K_M",
        "bits_per_weight": 4
      },
      "size_bytes": 3341081486,
      "params_string": "4B",
      "loaded_instances": [
        {
          "id": "google/gemma-3-4b",
          "config": {
            "context_length": 4096,
            "eval_batch_size": 512,
            "flash_attention": true,
            "offload_kv_cache_to_gpu": true
          }
        }
      ],
      "max_context_length": 131072,
      "format": "gguf",
      "capabilities": {
        "vision": true,
        "trained_for_tool_use": false
      },
      "description": null,
      "variants": [
        "google/gemma-3-4b@q4_k_m"
      ],
      "selected_variant": "google/gemma-3-4b@q4_k_m"
    },
    {
      "type": "llm",
      "publisher": "ibm",
      "key": "ibm/granite-4-h-tiny",
      "display_name": "Granite 4 H Tiny",
      "architecture": "granitehybrid",
      "quantization": {
        "name": "Q4_K_M",
        "bits_per_weight": 4
      },
      "size_bytes": 4231029544,
      "params_string": "7B",
      "loaded_instances": [],
      "max_context_length": 1048576,
      "format": "gguf",
      "capabilities": {
        "vision": false,
        "trained_for_tool_use": true
      },
      "description": null,
      "variants": [
        "ibm/granite-4-h-tiny@q4_k_m"
      ],
      "selected_variant": "ibm/granite-4-h-tiny@q4_k_m"
    },
    {
      "type": "embedding",
      "publisher": "nomic-ai",
      "key": "text-embedding-nomic-embed-text-v1.5",
      "display_name": "Nomic Embed Text v1.5",
      "quantization": {
        "name": "Q4_K_M",
        "bits_per_weight": 4
      },
      "size_bytes": 84106624,
      "params_string": null,
      "loaded_instances": [],
      "max_context_length": 2048,
      "format": "gguf"
    }
  ]
}
$

Why bother running LLMs locally?

To repeat what I said in my previous post…

There are potential issues with using a cloud-based LLM.

  • Security: Passing sensitive data or your company intellectual property to the cloud could be a problem. Using a local LLM means you can do pretty much what you want with it. It also means your prompts are not being stored and available to the outside world when the provider of your cloud-based LLM inevitably gets hacked, or sells your details.
  • Consistency: The cloud vendors update their models from time to time, and you may find a “new and improved” version of a model works less well for what you are doing. Just look what the “upgrade” to ChatGPT 5 broke. Using a local LLM means you are in control of if/when you update models.
  • Cost: You have the cost of the local kit, but after that it’s all “free”. Running a model requires a lot less kit than training a model. Depending on your needs, maybe running something locally will make more sense than paying for a cloud subscription. It can get really expensive if you require a large call volume on a cloud subscription.
  • Choice: There are lots of open models to choose from (see here). Maybe one of the more targeted models will work better for you than an all purpose model.

Overall

Your experience will vary depending your kit and the model you pick. There are some really big models, but they will need a lot more memory and a good GPU to function. Even these small models are really useful though, and importantly will run on crappy kit.

I have no intention of paying for a cloud subscription when I can run these for free and not be spied on. Have fun!

Cheers

Tim…

Oracle VirtualBox 7.2.6

Oracle has released VirtualBox 7.2.6.

The downloads and changelog are in the usual places.

I’ve done an install on a Windows 10 and attempted one on a Windows 11 PC.

As mentioned in my post about the previous release (here), the updates are not as straight forward as they used to be. In both cases I had to do the following to get things working.

  • Uninstall VirtualBox.
  • Make sure all “VirtualBox Host-Only Ethernet Adapter” adapters in Device Manager had been removed.
  • Reboot.
  • Install VirtualBox using “run as administrator”.
  • Reboot.

That process worked for Windows 10, but not for Windows 11. I tried a couple of times, but it doesn’t seem happy. I’m not sure why Windows 11 seems to be getting worse, not better…

I’ve run a bunch of Vagrant builds for 26ai on my Windows 10 machine and it seems to be working as normal, so fingers crossed…

I think I’m going to give up on my Windows 11 PC. It’s going to get converted to Linux at some point, and that will hopefully make is usable again…

Cheers

Tim…

Oracle AI Database 26ai On-Prem Installations

A couple of days ago Oracle released Oracle AI Database 26ai On-Prem for Linux x64. You can download it from here.

This opens the floodgates for a bunch of articles I’ve been sitting on for over 3 years. First things first are the installations, so here goes.

Installation Articles

Here are the usual suspects.

Vagrant Builds

Each of these articles is associated with one or more Vagrant builds.

Single Instance:

RPM Single Instance:

RAC:

Data Guard:

Those should get you up and running pretty quickly, assuming you don’t want to use the Oracle provided builds.

What’s with the version number?

The version 23.26.1.0.0 seems to freak people out. This was explained when 26ai was first released in the cloud. You can read about it here.

What Next?

With these out of the way I can start working through the remaining articles that have been sitting idle for over three years. There is some testing and checking of doc links to do, but they will filter out over the coming days/weeks, along with some new stuff that was released as part of the move from 23.26.0 to 23.26.1.

Happy installing/upgrading…

Cheers

Tim…