I was using non-GUI option to perform a distributed load testing with Jmeter from master server A (Linux) to slave server B (Linux). Here's what I did:
./jmeter -n -t xxx.jmx -l results.log -R xxx.xxx.xxx.xxx
By default, Jmeter will write the results file to the master server A. I'm wondering if there's a way to directly store the results file on remote server B.
Thanks in advance!
The idea is they are dumb slaves, possibly without even the means to store data locally. The methods are invoked on a remote processor (RMI) and run from there, but all other resources are kept at the client side.
You could set up a CI server like Jenkins to run jmeter clients in a distributed manner. That will allow you more control over how tests are distributed. I have a jenkins set up that runs jmeter client on a master, with distributed tests to remote (jmeter) slaves, which are also jenkins slaves, tied to jobs that control the workspaces. This means they are never used by jenkins, but jenkins can get results files out of the workspace, aggregate them, and produce graphs using jmeter plugins CMDrunner.
The other possibility is to prefix your sample results with the unique part of the IP address of the current slave. This will allow you to display them aggregated or separately in your results viewer, and potentially write scripts to separate the results out and store them back on the slave they came from.
You could also use samba or similar to share the remote slave drive, and store the results directly there from the master.
If you have only one slave, any of these would be easy to implement, but get more complex as you add more slaves.
If you love us? You can donate to us via Paypal or buy me a coffee so we can maintain and grow! Thank you!
Donate Us With