I think the only way right now is to duplicate ignite3db file and edit it,
as it's hardcoded now.
Line 30 of the ignite3db file is calling the vars.env, so you'd want to
change it to something like . ${IGNITE_HOME}/etc/vars.env.cluster2Also I'd like to reiterate that you should not mix the cluster and node. What I mean is that if you have node1, node2, node3 in the cluster1 and node1, node2, node3 in the cluster2, you'd need to have 6 different env files and 6 different start scripts. This is only one way to have the nodes start as you want, I could imaging modifying the vars.env file in such a way that it takes the values from the environment and so you don't need to duplicate start scripts as well. On Thu, Mar 12, 2026 at 5:14 PM Charlin S <[email protected]> wrote: > Hi Vadim, > > Thank you for the update. > > I have duplicated the files as suggested (ignite3db-cluster2 and > var.env.cluster2) and successfully started the second cluster. > > Could you please advise on how to correctly pass the specific var.env file > when starting the cluster using the command > ./ignite3-db-3.1.0/bin/ignite3db? > > Thanks & Regards, > Charlin S > > On Thu, 12 Mar 2026 at 17:44, Vadim Pakhnushev <[email protected]> > wrote: > >> Looks like the LOG_DIR should also be changed, I think we had a ticket >> somewhere to change the log file name to contain the node name but it's not >> the case now and simplest workaround would be to assign a separate log >> directory for each node. >> >> On Thu, Mar 12, 2026 at 4:07 PM Vadim Pakhnushev <[email protected]> >> wrote: >> >>> Hi! >>> >>> First of all, Ignite 3 uses separate node and cluster configurations. I >>> presume you're talking about node configuration? >>> Let's say you want to run two nodes from a single distribution. >>> Unfortunately, Ignite 3 is not very flexible in this case and you will >>> need to manually edit configuration files. >>> >>> I would probably copy-paste vars.env files to something like vars1.env >>> and vars2.env. >>> Then in each file I'd change NODE_NAME to unique name, let's say node1 >>> and node2. >>> Then as you already did - modify CONFIG_FILE to point to the >>> configuration file for each node. >>> Then I would copy-paste ignite3db files as well and change the . >>> ${IGNITE_HOME}/etc/vars.env line to point to the vars.env file for the >>> specific node. >>> >>> I think we could do better in this regard, let me create a ticket for >>> improvement and let me know if the above suggestion end up working. >>> >>> Thanks, >>> Vadim >>> >>> On Thu, Mar 12, 2026 at 3:08 PM Charlin S <[email protected]> >>> wrote: >>> >>>> Hi All, >>>> >>>> I am currently working with two different cluster setup configurations >>>> (using different port numbers) for Ignite 2.17. I am able to start both >>>> clusters sequentially so they run simultaneously using the following >>>> commands: >>>> >>>> 1. ignite.sh configfile1.xml >>>> 2. ignite.sh configfile2.xml >>>> >>>> For Ignite 3.1, I have migrated these files to configfile1.conf and >>>> configfile2.conf using the Ignite migration tool and placed them in the >>>> ignite3-db-3.1.0/etc/ folder. I successfully started the first cluster by >>>> editing vars.env to set CONFIG_FILE=${CONF_DIR}/configfile1.conf. >>>> >>>> Could you please advise on how to start the second cluster using >>>> configfile2.conf in this new environment? >>>> >>>> Thanks & Regards, >>>> Charlin S >>>> >>>
